Hackernews posts about FP16
- TTS engines: WebSocket vs. sync is 5.5x, INT8 slower than fp16 on M4 (ai.gopubby.com)
- Bolt Graphics Targets FP64 HPC Workloads with Zeus GPU (www.hpcwire.com)
- Everyone Wants Servers and Nobody Wants Servers (connectedplaces.online)
- ONNX Runtime and CoreML May Silently Convert Your Model to FP16 (ym2132.github.io)
- Running the Deepseek-R1 671B Model at FP16 Fidelity on AMD EPYC CPUs (www.servethehome.com)
- 90T/s on my iPhone llama3.2-1B-fp16 (www.reddit.com)
- PyTorch 2.6 Delivers FP16 Support for x86 CPUs, Better Intel GPU Experience (www.phoronix.com)
- PyTorch 2.6 Delivers FP16 Support for x86 CPUs, Better Intel GPU Experience (www.phoronix.com)
- Show HN: OpenGraviton – Run 500B+ parameter models on a consumer Mac Mini (opengraviton.github.io)
- Show HN: I made Qwen3.5-4B 13% smarter by compressing it to 4-bit (huggingface.co)