Tags
7 pages
GGUF
Local LLM Models Recommended for an RTX 3060 GPU
Running Qwen3.6 Locally: VRAM Requirements for 27B and 35B-A3B Quantized Models
Running Gemma 4 Locally: VRAM Requirements for E2B, E4B, 26B, and 31B Quantized Models
How to Use llama-quantize for GGUF Models
How to Get GGUF Models from Hugging Face with llama.cpp
Choosing Llama GGUF Quantization on Hugging Face: Practical Advice from Q8 to Q2
How to Download a GGUF Model from Hugging Face and Import It into Ollama