-
Notifications
You must be signed in to change notification settings - Fork 63
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bump Inference APIs to Neuron 2.13 #206
Conversation
JingyaHuang
commented
Aug 29, 2023
•
edited
Loading
edited
- Unblock VAE encoder
- Remove optimized cross-attention score workaround
- Unblock dynamic batching test
- Update documentation
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
@@ -226,6 +231,7 @@ optimum-cli export neuron --model stabilityai/stable-diffusion-2-1-base \ | |||
--batch_size 1 \ | |||
--height 512 `# height in pixels of generated image, eg. 512, 768` \ | |||
--width 512 `# width in pixels of generated image, eg. 512, 768` \ | |||
--num_image_per_prompt 4 `# number of images to generate per prompt, defaults to 1` \ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
is this needed with dynamic batching support now?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If we turn on dynamic batching, we can input any batch size / number images per prompt, but by default the dynamic batch size is turned off, so we can keep the snippet this way I think.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM