llama.cpp
Why people choose llama.cpp
Section titled “Why people choose llama.cpp”llama.cpp is the path for users who want:
- clear control over runtime behavior,
- predictable local inference,
- and a setup that is easier to reason about when something breaks.
Best fit
Section titled “Best fit”- Mac and desktop users
- users who care about repeatability
- users who have already outgrown one-click apps
Not the best first stop if
Section titled “Not the best first stop if”- you are brand new to local AI,
- or your goal is just to see Gemma 4 run once
In that case, start with LM Studio or AI Edge Gallery.