Hackernews posts about Llama 3
Llama 3 is an AI-powered chatbot that uses pure NumPy to generate human-like conversations in a browser-based interface.
- World AI Agents–35 AI Models (Claude, GPT, Llama)via One OpenAIcompatible API (world-ai-agents.com)
- Gemini-CLI, zeroclaw, Llama.cpp, Qwen3.5 on TK1 (forums.developer.nvidia.com)
- Show HN: Bonsai 1.7B ternary model at 442T/s on M4 Max (agents2agents.ai)
- Show HN: I read Replika's privacy policy and then built a competitor (apps.apple.com)
- Running Gemma 4 31B on Mac with Ollama (sammyrulez.github.io)
- Show HN: Local RAG Pipeline with Weaviate and Ollama (www.storyblok.com)
- Show HN: Helix – open-source self-healing back end for production crashes (88hours.github.io)
- Llama3 implemented from scratch (github.com)
- Llama 3 implemented in pure NumPy (docs.likejazz.com)
- Llama 3-V: Matching GPT4-V with a 100x smaller model and 500 dollars (aksh-garg.medium.com)
- Llama 3.1 (llama.meta.com)
- Llama-3.3-70B-Instruct (huggingface.co)
- Llama 3.1 Omni Model (github.com)
- Cost of self hosting Llama-3 8B-Instruct (blog.lytix.co)
- llama-fs: A self-organizing file system with llama 3 (github.com)
- Llama 3.1 in C (github.com)
- Show HN: Llama 3.3 70B Sparse Autoencoders with API access (www.goodfire.ai)
- Meta's Llama 3.1 can recall 42 percent of the first Harry Potter book (www.understandingai.org)
- Show HN: Tune LLaMa3.1 on Google Cloud TPUs (github.com)
- Implementing LLaMA3 in 100 Lines of Pure Jax (saurabhalone.com)
- Longwriter – Increase llama3.1 output to 10k words (github.com)
- Hermes 3: The First Fine-Tuned Llama 3.1 405B Model (lambdalabs.com)
- Llama 3.2 released: Multimodal, 1B to 90B sizes (www.llama.com)
- Karpathy/Nano-Llama31 (github.com)
- How to Run Llama 3 405B on Home Devices? Build AI Cluster (b4rtaz.medium.com)