I keep running out of memory. What's the biggest model, and most context, I can run on 3060 12gb? With decent speed? : r/LocalLLaMA #457
Labels
linux
Linux notes tools links
llm-experiments
experiments with large language models
llm-inference-engines
Software to run inference on large language models
llm-quantization
All about Quantized LLM models and serving
llm-serving-optimisations
Tips, tricks and tools to speedup inference of large language models
Planning
Planning and organizing tips and tools
shell-script
shell scripting in Bash, ZSH, POSIX etc
Here's the reformatted text in Markdown format:
Suggested labels
{ "label-name": "memory-optimization", "description": "Strategies for optimizing memory usage when running models on 3060 12gb GPU.", "confidence": 94.88 }
The text was updated successfully, but these errors were encountered: