Hackernews posts about Llama 2
Llama 2 is a large language model AI designed to replace earlier models like GPT-3.5/4 and provide advanced natural language processing capabilities.
- From llama.vim to Qt Creator using AI (cristianadam.eu)
- D2F – We made dLLMs 2.5x faster than LLaMA3 (arxiv.org)
- A Python RAG tutorial with Pinecone and Ollama 3.2 with a code example (blog.yasuflores.me)
- Show HN: OWhisper – Ollama for realtime speech-to-text (docs.hyprnote.com)
- Show HN: My Agentic Newsletter Project (iliareingold.com)
- Fast and Portable Llama2 Inference on the Heterogeneous Edge (www.secondstate.io)
- JetMoE: Reaching LLaMA2 performance with 0.1M dollars (research.myshell.ai)
- Fast Llama 2 on CPUs with Sparse Fine-Tuning and DeepSparse (neuralmagic.com)
- WebLLM: Llama2 in the Browser (webllm.mlc.ai)
- Understanding Llama 2 and the New Code Llama LLMs (magazine.sebastianraschka.com)
- Llama 2 Long (arxiv.org)
- SEQUOIA: Exact Llama2-70B on an RTX4090 with half-second per-token latency (infini-ai-lab.github.io)
- LoRA Fine-Tuning Efficiently Undoes Safety Training from Llama 2-Chat 70B (www.lesswrong.com)
- Run Llama 2 uncensored locally (ollama.com)
- Cloudflare and Meta Collaborate to Make Llama 2 Available Globally (www.cloudflare.com)
- A poor man's guide to fine-tuning Llama 2 (duarteocarmo.com)
- GPT 3.5 vs. Llama 2 fine-tuning: A Comprehensive Comparison (ragntune.com)
- Show HN: Finetune Llama-3 2x faster in a Colab notebook (colab.research.google.com)
- Llama 2 70B on M2 Max at 7 tokens/sec (twitter.com)
- Yi-34B, Llama 2, and common practices in LLM training (blog.eleuther.ai)
- EagleX 1.7T: Soaring past LLaMA 7B 2T in both English and Multi-lang evals (substack.recursal.ai)