Hacker News
- Optimizing LLMs from a Dataset Perspective https://sebastianraschka.com/blog/2023/optimizing-LLMs-dataset-perspective.html 24 comments
Linked pages
- GitHub - openlm-research/open_llama: OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset https://github.com/openlm-research/open_llama 183 comments
- [2305.15717] The False Promise of Imitating Proprietary LLMs https://arxiv.org/abs/2305.15717 119 comments
- Announcing StableCode — Stability AI https://stability.ai/blog/stablecode-llm-generative-ai-coding 107 comments
- Falcon LLM - Home https://falconllm.tii.ae/ 87 comments
- How Long Can Open-Source LLMs Truly Promise on Context Length? | LMSYS Org https://lmsys.org/blog/2023-06-29-longchat/ 62 comments
- [2305.11206] LIMA: Less Is More for Alignment https://arxiv.org/abs/2305.11206 44 comments
- LLM Training: RLHF and Its Alternatives https://magazine.sebastianraschka.com/p/llm-training-rlhf-and-its-alternatives 14 comments
- Machine Learning Q… by Sebastian Raschka, PhD [PDF/iPad/Kindle] https://leanpub.com/machine-learning-q-and-ai 12 comments
- [2304.01373] Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling https://arxiv.org/abs/2304.01373 7 comments
- Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality | LMSYS Org https://lmsys.org/blog/2023-03-30-vicuna/ 7 comments
- NeurIPS Large Language Model Efficiency Challenge:1 LLM + 1GPU + 1Day | NeurIPS 2023 Challenge https://llm-efficiency-challenge.github.io/index 5 comments
- GitHub - Stability-AI/StableLM: StableLM: Stability AI Language Models https://github.com/Stability-AI/StableLM 4 comments
- [2110.08207] Multitask Prompted Training Enables Zero-Shot Task Generalization https://arxiv.org/abs/2110.08207 2 comments
- [2212.10560] Self-Instruct: Aligning Language Model with Self Generated Instructions https://arxiv.org/abs/2212.10560 1 comment
- GitHub - gururise/AlpacaDataCleaned: Alpaca dataset from Stanford, cleaned and curated https://github.com/gururise/AlpacaDataCleaned 0 comments
- [2307.08701] AlpaGasus: Training A Better Alpaca with Fewer Data https://arxiv.org/abs/2307.08701#samsung 0 comments
- [2307.09288] Llama 2: Open Foundation and Fine-Tuned Chat Models https://arxiv.org/abs/2307.09288 0 comments
- Meet Stable Beluga 1 and Stable Beluga 2, Our Large and Mighty Instruction Fine-Tuned Language Models — Stability AI https://stability.ai/blog/stable-beluga-large-instruction-fine-tuned-models 0 comments
- [2308.06259] Self-Alignment with Instruction Backtranslation https://arxiv.org/abs/2308.06259 0 comments
- [2308.12950] Code Llama: Open Foundation Models for Code https://arxiv.org/abs/2308.12950 0 comments
Related searches:
Search whole site: site:sebastianraschka.com
Search title: Optimizing LLMs From a Dataset Perspective
See how to search.