Hackernews posts about Llama.cpp
Related:
M2 Max
- Gregory Gerganov and llama.cpp team joining HF (huggingface.co)
- Zluda update Q4 2025 – ROCm7, Windows, full llama.cpp and more (vosen.github.io)
- Show HN: Notebook page on llama.cpp official webui (github.com)
- Show HN: Director-AI – token-level NLI+RAG (github.com)
- Show HN: Agentic – Vesta AI Explorer (kruks.ai)
- Yet another reminder why you should not use Ollama (github.com)
- Vision Now Available in Llama.cpp (github.com)
- Llama.cpp guide – Running LLMs locally on any hardware, from scratch (steelph0enix.github.io)
- Heap-overflowing Llama.cpp to RCE (retr0.blog)
- Llama.cpp supports Vulkan. why doesn't Ollama? (github.com)
- Ollama violating llama.cpp license for over a year (github.com)
- Show HN: Open-source load balancer for llama.cpp (github.com)
- Llama.cpp AI Performance with the GeForce RTX 5090 Review (www.phoronix.com)
- Mistral Integration Improved in Llama.cpp (github.com)
- Llama.cpp: Add GPT-OSS (github.com)
- Llama.cpp Now Part of the Nvidia RTX AI Toolkit (developer.nvidia.com)
- Grok-1 Support for Llama.cpp (github.com)
- Llama.cpp Working on Support for Llama3 (github.com)
- DeepSeek-R1 speeds up llama.cpp code by x2 (github.com)
- Llama.cpp AI Performance with the GeForce RTX 5090 (www.phoronix.com)
- Tinker with LLMs in the privacy of your own home using Llama.cpp (www.theregister.com)
- Llama.cpp AI Performance with the GeForce RTX 5090 (www.phoronix.com)
- Llama's Paradox – Exploiting Llama.cpp (retr0.blog)