Hackernews posts about Cerebras Inference
- Cerebras now supports OpenAI GPT-OSS-120B at 3k Tokens Per SEC (www.cerebras.ai)
- Cerebras Inference: AI at Instant Speed (cerebras.ai)
- Cerebras Inference (twitter.com)
- Cerebras Inference – Voice Mode (cerebras.vercel.app)
- AWS Marketplace: Cerebras Inference Cloud (aws.amazon.com)
- Cerebras Launches Cerebras Inference Cloud Availability in AWS Marketplace (www.businesswire.com)
- cerebras: 450 tokens/sec llama 3.1 70B (www.theregister.com)
- Cerebras brings instant inference to Mistral Le Chat (cerebras.ai)
- Cerebras Enters AI Inference Blows Away Tiny Nvidia H100 GPUs by Besting HBM (www.servethehome.com)
- Cerebras Launches the Fastest AI Inference (inference.cerebras.ai)
- Llama 8B at 1800 tokens per second on Cerebras (inference.cerebras.ai)
- Meta Collaborates with Cerebras in New Llama API (www.cerebras.ai)
- Cerebras Launches the Fastest AI Inference (old.reddit.com)
- Nvidia MLPerf Inference v4.0 is Out (www.servethehome.com)
- Nvidia MLPerf Inference v4.0 is Out (www.servethehome.com)