Option 1
Ollama
Friendly wrapper around llama.cpp, OpenAI-compatible API.
Best for
Most users, default daily driver, server mode for tools like Cursor.
Pros
- One-line install + model pull
- Curated model registry
- OpenAI-compatible API on :11434
- Active community + great docs
- Cross-platform: macOS, Linux, Windows
Cons
- Less control over quantization tuning
- Curated registry can lag latest releases
- Some advanced llama.cpp features hidden