Hacker News
Linked pages
- Cursor - The AI-first Code Editor https://cursor.so 25 comments
- GPT-4 Architecture, Infrastructure, Training Dataset, Costs, Vision, MoE https://www.semianalysis.com/p/gpt-4-architecture-infrastructure 10 comments
- Transformer Inference Arithmetic | kipply's blog https://kipp.ly/blog/transformer-inference-arithmetic/ 4 comments
- [2001.08361] Scaling Laws for Neural Language Models https://arxiv.org/abs/2001.08361 0 comments
- [2302.01318] Accelerating Large Language Model Decoding with Speculative Sampling https://arxiv.org/abs/2302.01318 0 comments
- A Gentle Introduction to 8-bit Matrix Multiplication for transformers at scale using transformers, accelerate and bitsandbytes https://huggingface.co/blog/hf-bitsandbytes-integration 0 comments
- [2305.19370] Blockwise Parallel Transformer for Long Context Large Models https://arxiv.org/abs/2305.19370 0 comments
- Llama 2: Open Foundation and Fine-Tuned Chat Models | Meta AI Research https://ai.meta.com/research/publications/llama-2-open-foundation-and-fine-tuned-chat-models/ 0 comments
Related searches:
Search whole site: site:www.cursor.so
Search title: Why GPT-3.5 is (mostly) cheaper than Llama 2
See how to search.