-
Notifications
You must be signed in to change notification settings - Fork 248
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Mac M2 build error #716
Comments
@trickster thanks for reporting this. I think #719 should fix this, can you please run:
And try to build again? |
It did not fix it for me |
@jawshoeadan do you have the same error? |
@EricLBuehler |
So, after I fixed it locally,
This works and loads all the layers, but it's not on GPU and the inference is insanely slow. Do you know why? The gguf models work, but for some reason chat_template is not applied?
|
@trickster thanks for letting me know. #719 has been updated accordingly.
So, it's not using the GPU at all? Can you please attach the full output of running the model?
Perhaps the chat template is incorrect? I just merged #734 which adds the Phi 3.5 chat template, but perhaps using the GGUF builtin template or sourcing the tokenizer and chat template from the official HF repository would be beter. The following works:
But I can reproduce the issue with the GGUF tokenizer. |
@trickster I'm closing this as the build error is incorrect. Can you please open a separate issue with the GGUF template behavior? |
Minimum reproducible example
The minium example to reproduce the error. Simpler examples make it easier and faster to fix!
Error
Latest commit or version
8ddf258
The text was updated successfully, but these errors were encountered: