Hacker News
- Ring Attention Explained – Unlocking Near Infinite Context Window https://coconut-mode.com/posts/ring-attention/ 2 comments
Linked pages
- [2312.00752] Mamba: Linear-Time Sequence Modeling with Selective State Spaces https://arxiv.org/abs/2312.00752 42 comments
- [2112.05682] Self-attention Does Not Need $O(n^2)$ Memory https://arxiv.org/abs/2112.05682 37 comments
- [2310.01889] Ring Attention with Blockwise Transformers for Near-Infinite Context https://arxiv.org/abs/2310.01889 20 comments
- [2205.14135] FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness https://arxiv.org/abs/2205.14135 3 comments
Related searches:
Search whole site: site:coconut-mode.com
Search title: Ring Attention Explained | Coconut Mode
See how to search.