Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

can't use mmap because of ggml? #190

Closed
aicoder2048 opened this issue May 16, 2023 · 4 comments
Closed

can't use mmap because of ggml? #190

aicoder2048 opened this issue May 16, 2023 · 4 comments
Labels
primordial Related to the primordial version of PrivateGPT, which is now frozen in favour of the new PrivateGPT

Comments

@aicoder2048
Copy link

llama.cpp: can't use mmap because tensors are not aligned; convert to new format to avoid this
llama_model_load_internal: format = 'ggml' (old version with low tokenizer quality and no mmap support)

@imfurkaann
Copy link

.

@dennis-gonzales
Copy link

same error

@lrodriguezcim
Copy link

same!

@nikhil-xb
Copy link

I am facing the same problem while loading the llama embeddings. After a bit of exploration, I found that there has been a new quantization format(speed improvement) introduced in the llama-cpp-python supporting repository. Refer this pull request.

Now this project privateGPT and many others are uses old quantization formats. As of now, there don't seem to be a way to convert old quantized format to new ones other than retrieving source models f16 and quantizing them to new formats.

Would recommend following this thread for more info: ggerganov/llama.cpp#1408

@imartinez imartinez added the primordial Related to the primordial version of PrivateGPT, which is now frozen in favour of the new PrivateGPT label Oct 19, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
primordial Related to the primordial version of PrivateGPT, which is now frozen in favour of the new PrivateGPT
Projects
None yet
Development

No branches or pull requests

6 participants