You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
this is a question to anyone who has tried EAGLE with LLAMA3, I was wondering which LLAMA3 model exactly you were using? I.e. I assume a quantized version since the original one from Meta is huge, which quantization gave a good ratio of quality and performance in combination with EAGLE? I would also appreciate if someone could point me to a quantized LLAMA3 model repo which is known to work with EAGLE, so far I have found GGUF versions which are not supported and it seems I am not able to quantize the original LLAMA3 myself due to insufficient RAM. Any hints would be greatly appreciated, thank you!
The text was updated successfully, but these errors were encountered:
Hi,
this is a question to anyone who has tried EAGLE with LLAMA3, I was wondering which LLAMA3 model exactly you were using? I.e. I assume a quantized version since the original one from Meta is huge, which quantization gave a good ratio of quality and performance in combination with EAGLE? I would also appreciate if someone could point me to a quantized LLAMA3 model repo which is known to work with EAGLE, so far I have found GGUF versions which are not supported and it seems I am not able to quantize the original LLAMA3 myself due to insufficient RAM. Any hints would be greatly appreciated, thank you!
The text was updated successfully, but these errors were encountered: