Hacker News
Linking pages
- Building a fully local AI smart home assistant | John's Website https://johnthenerd.com/blog/local-llm-assistant/ 186 comments
- Towards 1-bit Machine Learning Models https://mobiusml.github.io/1bit_blog/ 157 comments
- Everything I've learned so far about running local LLMs https://nullprogram.com/blog/2024/11/10/ 84 comments
- 10 Noteworthy AI Research Papers of 2023 https://magazine.sebastianraschka.com/p/10-ai-research-papers-2023 24 comments
- Getting Started with Mixtral-8x7B https://www.secondstate.io/articles/mixtral-8-7b/ 3 comments
- Sholto Douglas & Trenton Bricken - How to Build & Understand GPT-7's Mind https://www.dwarkeshpatel.com/p/sholto-douglas-trenton-bricken 3 comments
- NeuralByte's weekly AI rundown - by Tomas Maixner https://open.substack.com/pub/neuralbyte/p/neuralbytes-weekly-ai-rundown-304?r=33qj5t&showWelcome=true 2 comments
- The Story of Arthur Mensch: How to Build a Startup with $2.1B Value in 9 Months https://thecreatorsai.com/p/the-story-of-arthur-mensch-how-to 2 comments
- Aman's AI Journal • Primers • Overview of Large Language Models https://aman.ai/primers/ai/LLM/ 1 comment
- Mistral AI, a Paris-based OpenAI rival, closed its $415 million funding round | TechCrunch https://techcrunch.com/2023/12/11/mistral-ai-a-paris-based-openai-rival-closed-its-415-million-funding-round/ 1 comment
- Faster Mixtral inference with TensorRT-LLM and quantization https://www.baseten.co/blog/faster-mixtral-inference-with-tensorrt-llm-and-quantization/ 1 comment
- The GPU Poor strike back - by Omar Sanseviero https://thehackerllama.substack.com/p/the-gpu-poor-strike-back 0 comments
- Scaling Mixtral LLM Serving with the High GPU Availability and Cost Efficiency | SkyPilot Blog https://blog.skypilot.co/scaling-mixtral-llm-serving-with-the-best-gpu-availability-and-cost-efficiency/ 0 comments
- The Reasoning Engine - by Rogerio Chaves https://reasoningengine.ai/p/the-reasoning-engine 0 comments
- Fast Inference of Mixture-of-Experts Language Models with Offloading https://browse.arxiv.org/html/2312.17238v1 0 comments
- GitHub - dvmazur/mixtral-offloading: Run Mixtral-8x7B models in Colab or consumer desktops https://github.com/dvmazur/mixtral-offloading 0 comments
- The Four Wars of the AI Stack (Dec 2023 Recap) https://www.latent.space/p/dec-2023 0 comments
- Using Jina Embeddings v2 with Haystack 2.0 pipelines to summarize legal documents | Haystack https://haystack.deepset.ai/blog/using-jina-embeddings-haystack 0 comments
- The Four Wars of the AI Stack (Dec 2023 Recap) https://www.latent.space/i/140396949/mixtral-sparks-a-gpuinference-war 0 comments
- [Paper Review] Efficient Memory Management for Large Language Model Serving with PagedAttention https://newsletter.micahlerner.com/p/paper-review-efficient-memory-management 0 comments
Linked pages
Related searches:
Search whole site: site:mistral.ai
Search title: Mixtral of experts | Mistral AI | Open source models
See how to search.