Linking pages
- GitHub - NVIDIA/TransformerEngine: A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference. https://github.com/NVIDIA/TransformerEngine 0 comments
Linked pages
- Llama 2 - Meta AI https://ai.meta.com/llama/ 820 comments
- fsspec: Filesystem interfaces for Python — fsspec 2024.6.0.post1+g8be9763.d20240613 documentation https://filesystem-spec.readthedocs.io/en/latest/index.html 28 comments
- GitHub - patrick-kidger/equinox: Elegant easy-to-use neural networks in JAX. https://docs.kidger.site/equinox/ https://github.com/patrick-kidger/equinox 25 comments
- Weights & Biases – Developer tools for ML https://wandb.ai/site 11 comments
- Introducing MPT-7B: A New Standard for Open-Source, Commercially Usable LLMs https://www.mosaicml.com/blog/mpt-7b 11 comments
- [2305.14342] Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training https://arxiv.org/abs/2305.14342 8 comments
- GitHub - huggingface/safetensors: Simple, safe way to store and distribute tensors https://github.com/huggingface/safetensors 1 comment
- TensorStore https://google.github.io/tensorstore/ 0 comments
- Backpack Models | Backpack Language Models are sequence models with an interface for interpretability. http://backpackmodels.science/ 0 comments
- GitHub - stanford-crfm/haliax: Named Tensors for Legible Deep Learning in JAX https://github.com/stanford-crfm/haliax 0 comments
Related searches:
Search whole site: site:github.com
Search title: GitHub - stanford-crfm/levanter: Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax
See how to search.