Hacker News
- Flex Attention – How to Scale Attention Models to a Billion Users? https://yash-sri.xyz/blog/flex_attention 1 comment
Linked pages
- [1706.03762] Attention Is All You Need https://arxiv.org/abs/1706.03762 145 comments
- FlexAttention: The Flexibility of PyTorch with the Performance of FlashAttention | PyTorch https://pytorch.org/blog/flexattention/ 24 comments
- [2009.06489] The Hardware Lottery https://arxiv.org/abs/2009.06489 16 comments
- A friendly introduction to machine learning compilers and optimizers https://huyenchip.com/2021/09/07/a-friendly-introduction-to-machine-learning-compilers-and-optimizers.html 4 comments
- [2205.14135] FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness https://arxiv.org/abs/2205.14135 3 comments
Related searches:
Search whole site: site:yash-sri.xyz
Search title: Yashovardhan Srivastava | A genius, shy and broke bloke.
See how to search.