Hackernews posts about Llama 2
Llama 2 is a large language model AI designed to replace earlier models like GPT-3.5/4 and provide advanced natural language processing capabilities.
- Show HN: MLX.zig – Llama 3.2 in Zig (github.com)
- Llama 3.2 from-scratch implementation focused on code readability (huggingface.co)
- Show HN: I Developed AI Memory Booster: Self-Hosted AI with Long Term Memory (aimemorybooster.com)
- Training Llama Using LibGen: Hack, a Theft, or Just Fair Use? (www.behind-the-enemy-lines.com)
- Meta must defend claim it stripped copyright info from Llama's training fodder (www.theregister.com)
- Meta's Llama models hit 1B downloads (techcrunch.com)
- Show, Don't Tell: A Llama PM's Guide to Writing GenAI Evals (www.ddmckinnon.com)
- Show HN: 1.32 Petaflops hardware for local prototyping (www.autonomous.ai)
- Show HN: Text an LLM at 61493035885 (benkaiser.dev)
- 90T/s on my iPhone llama3.2-1B-fp16 (www.reddit.com)
- Ollama 0.6.2 released with support for AMD Strix Halo (www.phoronix.com)
- Ollama 0.6.2 Released with Support for AMD Strix Halo (www.phoronix.com)
- Show HN: Create a local RAG AI in 2 minutes (github.com)
- Building Local AI Agents: Semantic Kernel and Ollama in C# (laurentkempe.com)
- Llama 2 (ai.meta.com)
- Run Llama 2 uncensored locally (ollama.ai)
- Llama2.c: Inference llama 2 in one file of pure C (github.com)
- Guide to running Llama 2 locally (replicate.com)
- LLaMA2 Chat 70B outperformed ChatGPT (tatsu-lab.github.io)
- Fast and Portable Llama2 Inference on the Heterogeneous Edge (www.secondstate.io)
- Fine-Tuning Llama-2: A Comprehensive Case Study for Tailoring Custom Models (www.anyscale.com)
- A simple guide to fine-tuning Llama 2 (brev.dev)
- JetMoE: Reaching LLaMA2 performance with 0.1M dollars (research.myshell.ai)
- Fast Llama 2 on CPUs with Sparse Fine-Tuning and DeepSparse (neuralmagic.com)
- Accessing Llama 2 from the command-line with the LLM-replicate plugin (simonwillison.net)
- WebLLM: Llama2 in the Browser (webllm.mlc.ai)
- Llama 2 on ONNX runs locally (github.com)
- Show HN: Llama2 Embeddings FastAPI Server (github.com)
- What's new in Llama 2 and how to run it locally (agi-sphere.com)
- Understanding Llama 2 and the New Code Llama LLMs (magazine.sebastianraschka.com)
- Llama 2 Long (arxiv.org)
- SEQUOIA: Exact Llama2-70B on an RTX4090 with half-second per-token latency (infini-ai-lab.github.io)
- Cookbook: Finetuning Llama 2 in your own cloud environment, privately (blog.skypilot.co)
- LoRA Fine-Tuning Efficiently Undoes Safety Training from Llama 2-Chat 70B (www.lesswrong.com)