Hacker News
Linking pages
Linked pages
- [2402.17764] The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits https://arxiv.org/abs/2402.17764 575 comments
- Mixtral of experts | Mistral AI | Open source models https://mistral.ai/news/mixtral-of-experts/ 300 comments
- Llama https://llama.meta.com/ 270 comments
- [2305.14314] QLoRA: Efficient Finetuning of Quantized LLMs https://arxiv.org/abs/2305.14314 129 comments
- [2310.11453] BitNet: Scaling 1-bit Transformers for Large Language Models https://arxiv.org/abs/2310.11453 21 comments
- [2106.09685] LoRA: Low-Rank Adaptation of Large Language Models https://arxiv.org/abs/2106.09685 8 comments
- Low-Rank Pruning of Llama2 https://mobiusml.github.io/low-rank-llama2/ 3 comments
- HQQ quantization https://mobiusml.github.io/hqq_blog/ 2 comments
- [2402.04396] QuIP#: Even Better LLM Quantization with Hadamard Incoherence and Lattice Codebooks https://arxiv.org/abs/2402.04396 0 comments
Related searches:
Search whole site: site:mobiusml.github.io
Search title: Towards 1-bit Machine Learning Models
See how to search.