Replies: 4 comments 31 replies
-
ggml already supports gpt-j, you should just be able to convert and quantize them. |
Beta Was this translation helpful? Give feedback.
-
I did not succeed with the convert-h5-to-ggml.py script and the model Dolly_GPT-J-6b. Script thinks for a while and crashes with the status "Killed". |
Beta Was this translation helpful? Give feedback.
-
Prob out-of-memory killer played his part? Check journalctl / syslog and try to monitor system resources while running it |
Beta Was this translation helpful? Give feedback.
-
I will investigate this as I would like to see foundation models and their variants outside of llama be blessed with the gift of ggml. |
Beta Was this translation helpful? Give feedback.
-
I just watched the latest video of my favorite youtuber - https://www.youtube.com/watch?v=AWAo4iyNWGc&t=14s and was wondering, if someone has already quantized & converted one of these to be compatible with llama.cpp?
The beauty of Dolly-like models is that they're based on open source gpt-j-6B from EleutherAI, so noone will be hunting us for using them without an ask.
Beta Was this translation helpful? Give feedback.
All reactions