Hacker News
- How to evaluate performance of LLM inference frameworks https://www.lamini.ai/blog/evaluate-performance-llm-inference-frameworks 2 comments
- Lamini Memory Tuning: 10x Fewer Hallucinations https://www.lamini.ai/blog/lamini-memory-tuning 57 comments
- Lamini LLM Finetuning on AMD ROCm: A Technical Recipe https://www.lamini.ai/blog/lamini-llm-finetuning-on-amd-rocm-a-technical-recipe 4 comments