Hacker News
- Mixture-of-Experts (Moe) LLMs https://cameronrwolfe.substack.com/p/moe-llms 0 comments
- Mixture-of-Experts (MoE) LLMs https://cameronrwolfe.substack.com/p/moe-llms 0 comments
Linked pages
- https://openai.com/index/learning-to-reason-with-llms/ 1525 comments
- [2501.12948] DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning https://arxiv.org/abs/2501.12948 1054 comments
- Introducing Code Llama, a state-of-the-art large language model for coding https://ai.meta.com/blog/code-llama-large-language-model-coding/ 527 comments
- xAI: Understand the Universe https://x.ai/ 437 comments
- Introducing DBRX: A New State-of-the-Art Open LLM | Databricks https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm 343 comments
- Grok-2 Beta Release https://x.ai/blog/grok-2 333 comments
- The Pile http://pile.eleuther.ai/ 294 comments
- Cheaper, Better, Faster, Stronger | Mistral AI | Frontier AI in your hands https://mistral.ai/news/mixtral-8x22b/ 243 comments
- [2401.04088] Mixtral of Experts https://arxiv.org/abs/2401.04088 150 comments
- [1701.06538] Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer https://arxiv.org/abs/1701.06538 125 comments
- [2310.06825] Mistral 7B https://arxiv.org/abs/2310.06825 124 comments
- GitHub - openai/tiktoken https://github.com/openai/tiktoken 74 comments
- Coefficient of variation - Wikipedia https://en.wikipedia.org/wiki/Coefficient_of_variation 21 comments
- Announcing Grok-1.5 https://x.ai/blog/grok-1.5 6 comments
- DeepSeek V3 and the cost of frontier AI models https://www.interconnects.ai/p/deepseek-v3-and-the-actual-cost-of 5 comments
- Open Release of Grok-1 https://x.ai/blog/grok-os 2 comments
- [1911.02150] Fast Transformer Decoding: One Write-Head is All You Need https://arxiv.org/abs/1911.02150 1 comment
- [2403.04132] Chatbot Arena: An Open Platform for Evaluating LLMs by Human Preference https://arxiv.org/abs/2403.04132 1 comment
- Open Source Licenses 101: Apache License 2.0 - FOSSA https://fossa.com/blog/open-source-licenses-101-apache-license-2-0/ 0 comments
- GitHub - mosaicml/llm-foundry https://github.com/mosaicml/llm-foundry 0 comments
Related searches:
Search whole site: site:cameronrwolfe.substack.com
Search title: Mixture-of-Experts (MoE) LLMs - by Cameron R. Wolfe, Ph.D.
See how to search.