Hacker News
- Show HN: Flash Attention in ~100 lines of CUDA https://github.com/tspeterkim/flash-attention-minimal 40 comments
- [P] Flash Attention in ~100 lines of CUDA https://github.com/tspeterkim/flash-attention-minimal 2 comments machinelearning
Linking pages
Linked pages
Would you like to stay up to date with Computer science? Checkout Computer science
Weekly.
Related searches:
Search whole site: site:github.com
Search title: GitHub - tspeterkim/flash-attention-minimal: Flash Attention in ~100 lines of CUDA (forward pass only)
See how to search.