- Self-host StableLM-2-Zephyr-1.6B with Wasm runtime. Portable across GPUs CPUs OSes https://www.secondstate.io/articles/stablelm-2-zephyr-1.6b/ 4 comments webassembly
Linked pages
- Fast and Portable Llama2 Inference on the Heterogeneous Edge https://www.secondstate.io/articles/fast-llm-inference/ 98 comments
- GitHub - WasmEdge/WasmEdge: WasmEdge is a lightweight, high-performance, and extensible WebAssembly runtime for cloud native, edge, and decentralized applications. It powers serverless apps, embedded functions, microservices, smart contracts, and IoT devices. https://github.com/WasmEdge/WasmEdge 33 comments
- https://localhost:8080/ 26 comments
- GitHub - second-state/LlamaEdge: The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge https://github.com/second-state/LlamaEdge 5 comments
- stabilityai/stablelm-2-zephyr-1_6b · Hugging Face https://huggingface.co/stabilityai/stablelm-2-zephyr-1_6b 0 comments
- https://twitter.com/realwasmedge/status/1752514157545476457 0 comments
Would you like to stay up to date with Web Development? Checkout Web Development
Weekly.
Related searches:
Search whole site: site:www.secondstate.io
Search title: Getting Started with StableLM-2-Zephyr-1.6B
See how to search.