Hackernews posts about Cerebras Inference
- Cerebras Inference: AI at Instant Speed (cerebras.ai)
- Cerebras Inference (twitter.com)
- Cerebras Inference – Voice Mode (cerebras.vercel.app)
- cerebras: 450 tokens/sec llama 3.1 70B (www.theregister.com)
- Cerebras Enters AI Inference Blows Away Tiny Nvidia H100 GPUs by Besting HBM (www.servethehome.com)
- Cerebras Launches the Fastest AI Inference (inference.cerebras.ai)
- Llama 8B at 1800 tokens per second on Cerebras (inference.cerebras.ai)
- Cerebras Launches the Fastest AI Inference (old.reddit.com)
- Nvidia MLPerf Inference v4.0 is Out (www.servethehome.com)
- Nvidia MLPerf Inference v4.0 is Out (www.servethehome.com)