llama.cpp logo

llama.cpp

Verified

High-performance C/C++ LLM inference with minimal setup

|Share:

About llama.cpp

llama.cpp screenshot

High-performance C/C++ LLM inference with minimal setup.

llama.cpp works 100% offline, is open source, is completely free to use, runs on CPU without a dedicated GPU.

Platform Support

Available for: Windows, macOS, Linux

System Requirements

  • Minimum RAM: 8 GB
  • GPU: Not required — runs on CPU

Links

GitHub Repository

Full description coming soon. Check the official website or GitHub for more details.

Frequently Asked Questions

What is llama.cpp?

High-performance C/C++ LLM inference with minimal setup ## About llama.cpp High-performance C/C++ LLM inference with minimal setup. llama.cpp works 100% offline, is open source, is completely free to use, runs on CPU without a dedicated GPU. ### Platfor...

Is llama.cpp free?

Yes, llama.cpp is completely free to use. It's also open source.

Does llama.cpp work offline?

Yes, llama.cpp works 100% offline once installed.

What platforms does llama.cpp support?

llama.cpp is available for Windows, macOS, Linux.

Stats

Stars94,642
Last committoday
Self-hostedYes
View Repository

Requirements

Platforms
windows
macos
linux
Offline capable
Yes
Minimum RAM8 GB
GPU requiredNo