A comprehensive toolkit for training and running lightweight adapters for GGUF-based language models (ERNIE, Llama, Mistral, Phi-3, etc.) without modifying the base model.
-
Updated
Feb 23, 2026 - Python
A comprehensive toolkit for training and running lightweight adapters for GGUF-based language models (ERNIE, Llama, Mistral, Phi-3, etc.) without modifying the base model.
Accurate VRAM calculator for Local LLMs (Llama 4, DeepSeek V3, Qwen 2.5). Calculates GGUF quantization, GQA context overhead, and offloading limits
Add a description, image, and links to the local-llm-gguf topic page so that developers can more easily learn about it.
To associate your repository with the local-llm-gguf topic, visit your repo's landing page and select "manage topics."