Hackernews posts about llama2-70B
- SEQUOIA: Exact Llama2-70B on an RTX4090 with half-second per-token latency (infini-ai-lab.github.io)
- Scaling LLama2-70B with Multiple Nvidia/AMD GPU (blog.mlc.ai)
- Run Llama2-70B in Web Browser with WebGPU Acceleration (webllm.mlc.ai)
- Groq reaches 100T/s per user on Llama2-70B (www.prnewswire.com)
- Llama2-70B with MosaicML Inference (www.mosaicml.com)
- Meta AI releases Code Llama 70B (twitter.com)
- The AI Hype Fraud of Reflection-Llama-70B (twitter.com)
- Show HN: Perplexity (llama3 70B) Inline Bot on Telegram (www.telegrambots.ai)
- Nvidia has published a competitive llama3-70B QA/RAG fine tune (old.reddit.com)
- Editing Files at 1000 tokens/s with llama-70B (www.cursor.com)
- OpenAI Lifeboat – Proxy OpenAI Code to Llama 70B on Replicate (lifeboat.replicate.dev)
- A Llama 70B finetune that has reflection baked into it's weights (huggingface.co)
- Fine-tuning Llama3-70B for GradPilot – easily beats Claude Sonnet (www.alignedhq.ai)
- LLM Boxing – Llama 70B-chat vs. GPT3.5 blind test (llmboxing.com)
- DPO fine-tuned Mistral 7B beats Llama 70B on MT Bench (huggingface.co)
- Running LLaMA-70B in the browser with MLC (www.latent.space)
- Llama3-70B mentions "sentience" independently (twitter.com)
- Higgsfield AI: Anyone Can Train Llama 70B or Mistral for Free (higgsfield.ai)