Hackernews posts about Llama-3.1 70B
- Reflection-Llama-3.1-70B is Llama-3 with LoRA (old.reddit.com)
- Nvidia releases weights for Llama-3.1-Nemotron-70B-Instruct (huggingface.co)
- Nvidia Llama-3.1-Nemotron-70B-Instruct (build.nvidia.com)
- Llama 3.1 70B compressed by 6.4x using AQLM-PV, now released (huggingface.co)
- cerebras: 450 tokens/sec llama 3.1 70B (www.theregister.com)
- Reflection Llama-3.1 70B: Testing and Summary of What We Know (www.datacamp.com)
- HuggingChat: Chat with Llama 3.1 (70B and 405B) (huggingface.co)
- Llama-3.1-Nemotron-70B-Instruct Model (build.nvidia.com)
- Reflection Llama 3.1 70B (artificialanalysis.ai)
- Show HN: Open-source study to measure end user satisfaction levels with LLMs (open-llm-initiative.com)
- Show HN: I built a website where you can easily fine-tune Llama 3.1 models (www.tunellama.com)
- Show HN: Mixlayer – code and deploy LLM prompts using JavaScript (www.mixlayer.com)
- Open-source 70B model surpass GPT-4o and Claude 3.5 on Arena Hard (huggingface.co)
- Nvidia Nemotron (huggingface.co)