Linking pages
- Pathways Language Model (PaLM): Scaling to 540 Billion Parameters for Breakthrough Performance – Google AI Blog https://ai.googleblog.com/2022/04/pathways-language-model-palm-scaling-to.html 279 comments
- Google Research: Themes from 2021 and Beyond – Google AI Blog https://ai.googleblog.com/2022/01/google-research-themes-from-2021-and.html 52 comments
- MaMMUT: A simple vision-encoder text-decoder architecture for multimodal tasks – Google AI Blog https://ai.googleblog.com/2023/05/mammut-simple-vision-encoder-text.html 33 comments
- Good News About the Carbon Footprint of Machine Learning Training – Google AI Blog https://ai.googleblog.com/2022/02/good-news-about-carbon-footprint-of.html 0 comments
- PaLI: Scaling Language-Image Learning in 100+ Languages – Google AI Blog https://ai.googleblog.com/2022/09/pali-scaling-language-image-learning-in.html 0 comments
- LIMoE: Learning Multiple Modalities with One Sparse Mixture-of-Experts Model – Google AI Blog https://ai.googleblog.com/2022/06/limoe-learning-multiple-modalities-with.html 0 comments
- Mixture-of-Experts with Expert Choice Routing – Google AI Blog https://ai.googleblog.com/2022/11/mixture-of-experts-with-expert-choice.html 0 comments
- GitHub - tomohideshibata/BERT-related-papers: BERT-related papers https://github.com/tomohideshibata/BERT-related-papers 0 comments
- Google is Leading the AGI race. But can it win? https://sergey.substack.com/p/google-is-leading-the-agi-race-but 0 comments
- Google Research, 2022 & beyond: Algorithms for efficient deep learning – Google AI Blog https://ai.googleblog.com/2023/02/google-research-2022-beyond-algorithms.html 0 comments
- Timeline of AI and language models – Dr Alan D. Thompson – Life Architect https://lifearchitect.ai/timeline/ 0 comments
- Mixture-of-Experts with Expert Choice Routing – Google Research Blog https://blog.research.google/2022/11/mixture-of-experts-with-expert-choice.html 0 comments
Linked pages
- Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, the World’s Largest and Most Powerful Generative Language Model | NVIDIA Technical Blog https://developer.nvidia.com/blog/using-deepspeed-and-megatron-to-train-megatron-turing-nlg-530b-the-worlds-largest-and-most-powerful-generative-language-model/ 124 comments
- Introducing Pathways: A next-generation AI architecture https://blog.google/technology/ai/introducing-pathways-next-generation-ai-architecture/ 33 comments
- Transformer: A Novel Neural Network Architecture for Language Understanding – Google AI Blog https://ai.googleblog.com/2017/08/transformer-novel-neural-network.html 3 comments
- [2112.06905] GLaM: Efficient Scaling of Language Models with Mixture-of-Experts https://arxiv.org/abs/2112.06905 1 comment
- Google wins MLPerf benchmarks with TPU v4 | Google Cloud Blog https://cloud.google.com/blog/products/ai-machine-learning/google-wins-mlperf-benchmarks-with-tpu-v4 0 comments
- General and Scalable Parallelization for Neural Networks – Google AI Blog https://ai.googleblog.com/2021/12/general-and-scalable-parallelization.html 0 comments
Related searches:
Search whole site: site:ai.googleblog.com
Search title: More Efficient In-Context Learning with GLaM – Google AI Blog
See how to search.