Hacker News
Lobsters
Linking pages
- GitHub - huggingface/candle: Minimalist ML framework for Rust https://github.com/huggingface/candle 205 comments
- GitHub - google/gemma.cpp: lightweight, standalone C++ inference engine for Google's Gemma models. https://github.com/google/gemma.cpp 130 comments
- GitHub - mikeroyal/Self-Hosting-Guide: Self-Hosting Guide. Learn all about locally hosting (on premises & private web servers) and managing software applications by yourself or your organization. Including Cloud, LLMs, WireGuard, Automation, Home Assistant, and Networking. https://github.com/mikeroyal/Self-Hosting-Guide 108 comments
- GitHub - samuel-vitorino/lm.rs: Minimal LLM inference in Rust https://github.com/samuel-vitorino/lm.rs 79 comments
- llama2.c/README.md at master · trholding/llama2.c · GitHub https://github.com/trholding/llama2.c 60 comments
- GitHub - okuvshynov/slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization https://github.com/okuvshynov/slowllama 54 comments
- GitHub - pytorch/torchchat: Run PyTorch LLMs locally on servers, desktop and mobile https://github.com/pytorch/torchchat 41 comments
- Ahead of AI #11: New Foundation Models https://magazine.sebastianraschka.com/p/ahead-of-ai-11-new-foundation-models 34 comments
- GitHub - mukel/llama2.java: Inference Llama 2 in one file of pure Java https://github.com/mukel/llama2.java 18 comments
- GitHub - zml/zml: High performance AI inference stack. Built for production. @ziglang / @openxla / MLIR / @bazelbuild https://github.com/zml/zml 13 comments
- GitHub - snunez1/llama.cl: Inference Llama in Common Lisp https://github.com/snunez1/llama.cl 11 comments
- GitHub - srush/llama2.rs https://github.com/srush/llama2.rs 10 comments
- GitHub - nikolaydubina/llama2.go: LLAMA-2 in pure Go https://github.com/nikolaydubina/llama2.go 8 comments
- GitHub - mikepapadim/llama-shepherd-cli: A CLI to manage install and configure llama inference implemenation in multiple languages https://github.com/mikepapadim/llama-shepherd-cli 8 comments
- GitHub - turingmotors/swan: This project aims to enable language model inference on FPGAs, supporting AI applications in edge devices and environments with limited resources. https://github.com/turingmotors/swan 7 comments
- GitHub - maxbbraun/llama4micro: A "large" language model running on a microcontroller https://github.com/maxbbraun/llama4micro 5 comments
- GitHub - mukel/llama3.java: Practical Llama 3 inference in Java https://github.com/mukel/llama3.java 4 comments
- OpenAI’s Andrej Karpathy Launches Baby Llama 2 https://www.theinsaneapp.com/2023/07/openai-karpathy-launches-baby-llama-2.html 3 comments
- GitHub - donge/llama2.zig: This is a Zig copycat version of llama2.c https://github.com/donge/llama2.zig 2 comments
- GitHub - rbitr/llm.f90: LLM inference in Fortran https://github.com/rbitr/llm.f90 2 comments
Linked pages
Would you like to stay up to date with Computer science? Checkout Computer science
Weekly.
Related searches:
Search whole site: site:github.com
Search title: GitHub - karpathy/llama2.c: Inference Llama 2 in pure C, single file, fp32, haha
See how to search.