Hackernews posts about gpt-oss
- Show HN: GPT-OSS Flash Attention via Native PyTorch SDPA (gist.github.com)
- Show HN: AIDictation – zero data retention dictation app (aidictation.com)
- Show HN: The Port Augusta Times – "All the news that's fit to generate" (henrygabriels.github.io)
- GPT-OSS vs. Qwen3 and a detailed look how things evolved since GPT-2 (magazine.sebastianraschka.com)
- What GPT-OSS leaks about OpenAI's training data (fi-le.net)
- GPT-OSS-120B runs on just 8GB VRAM & 64GB+ system RAM (old.reddit.com)
- Running GPT-OSS-120B at 500 tokens per second on Nvidia GPUs (www.baseten.co)
- GPT-OSS Reinforcement Learning (docs.unsloth.ai)
- Introducing gpt-oss (openai.com)
- Visualizing GPT-OSS-20B embeddings (melonmars.github.io)
- GPT-OSS 120B Runs at 3000 tokens/sec on Cerebras (www.cerebras.ai)
- Llama.cpp: Add GPT-OSS (github.com)
- Cerebras now supports OpenAI GPT-OSS-120B at 3k Tokens Per SEC (www.cerebras.ai)
- OpenAI GPT-OSS models use MXFP4 to cut inference costs (www.theregister.com)
- GPT-OSS Reasoning and Any-LLM: Nuances of OpenAI API Compatibility (blog.mozilla.ai)
- OpenAI/GPT-OSS-120B · Hugging Face (huggingface.co)
- Show HN: Run gpt-oss-20b on 8GB GPUs (github.com)
- OpenAI GPT-OSS (github.com)
- OpenAI GPT-OSS Model Card [pdf] (cdn.openai.com)
- Using Codex CLI with GPT-OSS:120B on an Nvidia DGX Spark via Tailscale (til.simonwillison.net)
- OpenAI GPT-OSS LLMs use MXFP4 floating point data type: smaller, faster, cheaper (www.theregister.com)
- GPT-OSS is a big deal (twitter.com)
- How to Run Your Own OpenAI GPT OSS Server for Fun and Profit (northcodie.blogspot.com)
- GPT-OSS-Safeguard (openai.com)
- Long context GPT-OSS fine-tuning (unsloth.ai)
- Using Codex CLI with GPT-OSS:120B on an Nvidia DGX Spark via Tailscale (til.simonwillison.net)
- Tricks from OpenAI GPT-OSS you can use with transformers (huggingface.co)
- gpt-oss is a great model (twitter.com)