Hacker News
- Qwen2.5-1M: Deploy your own Qwen with context length up to 1M tokens https://qwenlm.github.io/blog/qwen2.5-1m/ 107 comments
Linking pages
Linked pages
- [2309.16039] Effective Long-Context Scaling of Foundation Models https://arxiv.org/abs/2309.16039 31 comments
- GitHub - vllm-project/vllm: A high-throughput and memory-efficient inference and serving engine for LLMs https://github.com/vllm-project/vllm 0 comments
- [2402.17463] Training-Free Long-Context Scaling of Large Language Models https://arxiv.org/abs/2402.17463 0 comments
- GitHub - hsiehjackson/RULER: This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models? https://github.com/hsiehjackson/RULER 0 comments
Related searches:
Search whole site: site:qwenlm.github.io
Search title: Qwen2.5-1M: Deploy Your Own Qwen with Context Length up to 1M Tokens | Qwen
See how to search.