-
Notifications
You must be signed in to change notification settings - Fork 642
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug] Launching Llama-3.2-11B-Vision-Instruct just hangs on generation #2619
Comments
I found out that when I use |
I had to roll back to v0.4.0 for 11b vision to work again. It errors out on 0.4.1 for me. |
@bluenevus does |
OK. I thought that hanging issue is connected with lack of VRAM of GPU. So I had rented H100 with 80 GB VRAM in order to launch Llama-3.2-11B-Vision-Instruct and ran this simple script: @sgl.function
def caption_image(s, image_file):
s += "You are very smart image captioning service"
s += "Given this image: " + sgl.image(image_file)
s += "Overall style of this image is: " + sgl.select("global_style", choices=["cinematic", "animated", "anime", "3d", "cartoon", "digital art"])
sgl.set_default_backend(sgl.RuntimeEndpoint(base_url))
image_path = "./examples/image.png"
state = caption_image.run(image_file=image_path)
print(state["global_style"]) And it still hangs with these logs:
Then I have changed I think overall |
Not sure what that means but you can see the compose components here deploy: |
Checklist
Describe the bug
I have rented RTX 6000Ada with 48.0 GB VRAM GPU via vast.ai.
Specs:
Then I have installed flashinfer by this command:
Then installed this lib with this command:
Then downloaded Llama-3.2-11B-Vision-Instruct and launched it like this:
Then I have used this simple code to infer an image:
However, when I launched this code just to check simple image, it just hangs and I receive no response or even error message.
Logs:
I don't understand why this is happening?
Reproduction
I have written it in description
Environment
Specs:
The text was updated successfully, but these errors were encountered: