Hacker News
- Loading LLM (Llama-2 70B) 20x faster with Anyscale Endpoints https://www.anyscale.com/blog/loading-llama-2-70b-20x-faster-with-anyscale-endpoints 2 comments
- Fine-Tuning LLMs: LoRA or Full-Parameter? An In-Depth Analysis with Llama 2 https://www.anyscale.com/blog/fine-tuning-llms-lora-or-full-parameter-an-in-depth-analysis-with-llama-2 2 comments
- ThirdAI Uses Ray for Parallel Training of Billion-Parameter NN on Commodity CPUs https://www.anyscale.com/blog/how-thirdai-uses-ray-for-parallel-training-of-billion-parameter-neural-networks-on-commodity-cpus 15 comments
- Llama 2 is about as factually accurate as GPT-4 for summaries and is 30X cheaper https://www.anyscale.com/blog/llama-2-is-about-as-factually-accurate-as-gpt-4-for-summaries-and-is-30x-cheaper 54 comments
- Continuous batching to increase LLM inference throughput and reduce p50 latency https://www.anyscale.com/blog/continuous-batching-llm-inference 20 comments
- Numbers every LLM Developer should know https://www.anyscale.com/blog/num-every-llm-developer-should-know 18 comments
- Fine-Tuning Llama-2: A Comprehensive Case Study for Tailoring Custom Models https://www.anyscale.com/blog/fine-tuning-llama-2-a-comprehensive-case-study-for-tailoring-models-to-unique-applications 59 comments
- Ray breaks the $1/TB barrier as the world’s most cost-efficient sorting system https://www.anyscale.com/blog/ray-breaks-the-usd1-tb-barrier-as-the-worlds-most-cost-efficient-sorting 8 comments
- LightGBM vs. XGBoost: Which distributed version is faster? https://www.anyscale.com/blog/introducing-distributed-lightgbm-training-with-ray 2 comments