Hackernews posts about GPT-2
- Replicating GPT-2, the most dangerous NLP model (2019) (datascienceathome.com)
- Homeworkers get 24 more minutes of sleep a day (www.bbc.co.uk)
- Show HN: RowMaker – Turn Images and Words into Spreadsheets (www.simpltech.ai)
- Show HN: Rund – Get and filter places around you (apps.apple.com)
- Reproducing GPT-2 in llm.c (github.com)
- Bumblebee: GPT2, Stable Diffusion, and More in Elixir (news.livebook.dev)
- We Found an Neuron in GPT-2 (clementneo.com)
- The Illustrated GPT-2: Visualizing Transformer Language Models (2019) (jalammar.github.io)
- Talk = GPT-2 and Whisper and WASM (github.com)
- Show HN: Fully client-side GPT2 prediction visualizer (perplexity.vercel.app)
- Build and train GPT-2 from scratch using PyTorch (differ.blog)
- C++ GPT-2 inference engine (github.com)
- Fast GPT-2 inference written in Fortran (github.com)
- Compare how GPT-2, 3, 3.5 and 4 answer the same questions (theaidigest.org)
- GPT-2 as step toward general intelligence (2019) (slatestarcodex.com)
- Spreadsheets are all you need: Understanding GPT2 and Transformers (spreadsheets-are-all-you-need.ai)
- Full forward pass of GPT-2 in one file of pure CUDA (github.com)
- Let's reproduce GPT-2 (124M) [video] (www.youtube.com)
- Gpt2-Chatbot Removed from Lmsys (lmsys.org)
- We Found an Neuron in GPT-2 (www.lesswrong.com)
- What will GPT-2030 look like? (www.lesswrong.com)
- Let's reproduce GPT-2 (124M) (twitter.com)
- Steering GPT-2-XL by adding an activation vector (www.lesswrong.com)
- WebGPT: Run GPT2 on the Browser with WebGPU (github.com)