Hacker News
- Mamba Explained: The State Space Model Taking On Transformers https://www.kolaayonrinde.com/blog/2024/02/11/mamba.html 93 comments
Linking pages
Linked pages
- GPT-4 https://openai.com/research/gpt-4 5744 comments
- "Focusing is about saying no" - Steve Jobs (WWDC'97) - YouTube https://www.youtube.com/watch?v=H8eP99neOVs 100 comments
- [2312.00752] Mamba: Linear-Time Sequence Modeling with Selective State Spaces https://arxiv.org/abs/2312.00752 42 comments
- The Annotated S4 https://srush.github.io/annotated-s4/ 12 comments
- Highly accurate protein structure prediction with AlphaFold | Nature https://www.nature.com/articles/s41586-021-03819-2 9 comments
- [2205.14135] FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness https://arxiv.org/abs/2205.14135 3 comments
- https://storage.googleapis.com/deepmind-media/AlphaCode2/AlphaCode2_Tech_Report.pdf 3 comments
- https://proceedings.neurips.cc/paper_files/paper/2022/file/c529dba08a146ea8d6cf715ae8930cbe-Paper-Conference.pdf 1 comment
- [2309.10668] Language Modeling Is Compression https://arxiv.org/abs/2309.10668 1 comment
- [2203.15556] Training Compute-Optimal Large Language Models https://arxiv.org/abs/2203.15556 0 comments
- Is Attention All You Need? http://www.isattentionallyouneed.com/ 0 comments
- In-Context Learning, In Context https://thegradient.pub/in-context-learning-in-context/ 0 comments
- In-context Learning and Induction Heads https://transformer-circuits.pub/2022/in-context-learning-and-induction-heads/index.html 0 comments
- Dictionary Learning with Sparse AutoEncoders | Kola Ayonrinde https://www.kolaayonrinde.com/blog/2023/11/03/dictionary-learning.html 0 comments
Related searches:
Search whole site: site:kolaayonrinde.com
Search title: Mamba Explained | Kola Ayonrinde
See how to search.