Linking pages
- Research Papers in February 2024 https://sebastianraschka.com/blog/2024/research-papers-in-february-2024.html 7 comments
- Open Language Models (OLMos) and the LLM landscape https://www.interconnects.ai/p/olmo 0 comments
- Dolma, OLMo, and the Future of Open-Source LLMs https://cameronrwolfe.substack.com/p/dolma-olmo-and-the-future-of-open 0 comments
- Installing and Developing vLLM with Ease | vLLM Blog https://blog.vllm.ai/2025/01/10/dev-experience.html 0 comments
- DeepSeek R1's recipe to replicate o1 and the future of reasoning LMs https://www.interconnects.ai/p/deepseek-r1-recipe-for-o1 0 comments
Linked pages
- [2305.14314] QLoRA: Efficient Finetuning of Quantized LLMs https://arxiv.org/abs/2305.14314 129 comments
- GitHub - huggingface/transformers: 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX. https://github.com/huggingface/transformers 26 comments
- [2106.09685] LoRA: Low-Rank Adaptation of Large Language Models https://arxiv.org/abs/2106.09685 8 comments
- GitHub - nlpxucan/WizardLM: WizardLM: Empowering Large Pre-Trained Language Models to Follow Complex Instructions https://github.com/nlpxucan/WizardLM 1 comment
- GitHub - eric-mitchell/direct-preference-optimization: Reference implementation for DPO (Direct Preference Optimization) https://github.com/eric-mitchell/direct-preference-optimization 0 comments
- GitHub - vllm-project/vllm: A high-throughput and memory-efficient inference and serving engine for LLMs https://github.com/vllm-project/vllm 0 comments
- [2406.09279] Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback https://arxiv.org/abs/2406.09279 0 comments
Related searches:
Search whole site: site:github.com
Search title: GitHub - allenai/open-instruct
See how to search.