Hackernews posts about VLLM
- Observability Stack for vLLM Inference (github.com)
- Deploy DeepSeek‑R1 with VLLM and Ray Serve on Kubernetes (www.anyscale.com)
- An Open Source Stack for AI Compute: Kubernetes and Ray and PyTorch and VLLM (www.anyscale.com)
- Why LLMs can't really build software (zed.dev)
- Coding with LLMs in the summer of 2025 – an update (antirez.com)
- Tao on “blue team” vs. “red team” LLMs (mathstodon.xyz)
- AccountingBench: Evaluating LLMs on real long-horizon business tasks (accounting.penrose.com)
- LLM Embeddings Explained: A Visual and Intuitive Guide (huggingface.co)
- LLM architecture comparison (magazine.sebastianraschka.com)
- LLMs aren't world models (yosefk.com)
- Show HN: Price Per Token – LLM API Pricing Data (pricepertoken.com)
- The current state of LLM-driven development (blog.tolki.dev)
- Llama-Scan: Convert PDFs to Text W Local LLMs (github.com)
- LLM Inflation (tratt.net)
- Lack of intent is what makes reading LLM-generated text exhausting (lambdaland.org)
- Show HN: Yet another memory system for LLMs (github.com)
- Is Chain-of-Thought Reasoning of LLMs a Mirage? A Data Distribution Lens (arstechnica.com)
- LLMs and coding agents are a security nightmare (garymarcus.substack.com)
- Show HN: Mcp-use – Connect any LLM to any MCP (github.com)
- Show HN: WebGPU enables local LLM in the browser – demo site with AI chat (andreinwald.github.io)