You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I think this is fixed in here. Not sure if the latest version includes it though.
For a quick fix, I recommend just proving n_ctx and n_gpu_layers directly as keyword arguments instead of under model_kwargs
Describe the issue as clearly as possible:
i have used llamacpp on itself or via langchain with no problems. When loading the same gguf model using outlines' llamacpp wrapper, i get a keyerror
Steps/code to reproduce the bug:
Expected result:
Error message:
Outlines/Python version information:
Version information
Context for the issue:
No response
The text was updated successfully, but these errors were encountered: