Hackernews posts about VLLM
- Performance vs. Practicality: A Comparison of VLLM and Ollama (robert-mcdermott.medium.com)
- Scaling VLLM for Embeddings: 16x Throughput and Cost Reduction (www.snowflake.com)
- The next generation of AI inference, powered by LLM-d (cloud.google.com)
- Show HN: We made an AI QA tester that uses VLMs to test your front-end (www.fracten.com)
- Show HN: Vision AI Checkup, an Optometrist for VLMs (visioncheckup.com)
- Vision Language Models Are Biased (vlmsarebiased.github.io)
- Vision Language Models (Better, Faster, Stronger) (huggingface.co)
- Vision Language Models (Better, Faster, Stronger) (huggingface.co)
- The last six months in LLMs, illustrated by pelicans on bicycles (simonwillison.net)
- Human coders are still better than LLMs (antirez.com)
- Show HN: My LLM CLI tool can run tools now, from Python code or plugins (simonwillison.net)
- LLMs get lost in multi-turn conversation (arxiv.org)
- After months of coding with LLMs, I'm going back to using my brain (albertofortin.com)
- LLM function calls don't scale; code orchestration is simpler, more effective (jngiam.bearblog.dev)
- LLMs are cheap (www.snellman.net)
- LLMs and Elixir: Windfall or deathblow? (www.zachdaniel.dev)
- Tokasaurus: An LLM inference engine for high-throughput workloads (scalingintelligence.stanford.edu)
- Peer Programming with LLMs, for Senior+ Engineers (pmbanugo.me)
- The behavior of LLMs in hiring decisions: Systemic biases in candidate selection (davidrozado.substack.com)
- Build real-time knowledge graph for documents with LLM (cocoindex.io)