Hackernews posts about Transformer
Transformer is a type of neural network architecture designed for natural language processing tasks that relies on self-attention mechanisms to process input sequences in parallel.
Related:
Apple
- Understanding Transformers via N-gram Statistics (arxiv.org)
- Transformer: The Deep Chemistry of Life and Death (nick-lane.net)
- You could have invented Transformers (gwern.net)
- Embeddings: Decoder-Only Transformers are SoTA Encoders (with some fine-tuning) (miguel-martin.com)
- Show HN: I made a number transformer where you evolve digits into target forms (numshift.n4ze3m.com)
- From Typewriters to Transformers: AI Is Just the Next Tools Abstraction (hardcoresoftware.learningbyshipping.com)
- From Typewriters to Transformers: AI Is Just the Next Tools Abstraction (hardcoresoftware.learningbyshipping.com)
- Understanding AI Large Language Model Transformers: An Analogy for Mere Mortals (jamesmccaffrey.wordpress.com)
- Transformer Lab Now Works with AMD GPUs (old.reddit.com)
- What is the Transformers' Context Window? (and how to make it LONG) (www.youtube.com)
- Transformer Progress (draft.lczero.org)
- Transactions as Transformers (will62794.github.io)
- Transactions as Transformers (will62794.github.io)
- Show HN: Run 30B model in 4GB Active Memory (github.com)
- Show HN: Tinker with Meta's "tokenizer-free" patcher (huggingface.co)