Hacker News
- Turing-NLG: A 17B-parameter language model https://www.microsoft.com/en-us/research/blog/turing-nlg-a-17-billion-parameter-language-model-by-microsoft/ 139 comments
Linking pages
- Does GPT-2 Know Your Phone Number? – The Berkeley Artificial Intelligence Research Blog https://bair.berkeley.edu/blog/2020/12/20/lmmem/ 155 comments
- The Mathematics of Training LLMs — with Quentin Anthony of Eleuther AI https://www.latent.space/p/transformers-math#details 66 comments
- Wu Dao 2.0: A Monster of 1.75 Trillion Parameters | by Alberto Romero | Medium | Towards Data Science https://towardsdatascience.com/gpt-3-scared-you-meet-wu-dao-2-0-a-monster-of-1-75-trillion-parameters-832cd83db484 10 comments
- Microsoft announces new supercomputer, lays out vision for future AI work - Source https://news.microsoft.com/source/features/innovation/openai-azure-supercomputer/ 5 comments
- NVIDIA and Microsoft Join Forces on Massive Cloud AI Computer - News https://www.allaboutcircuits.com/news/nvidia-and-microsoft-join-forces-massive-cloud-ai-computer/ 3 comments
- Watson's Creator Wants to Teach AI a New Trick: Common Sense | WIRED https://www.wired.com/story/watsons-creator-teach-ai-new-trick-common-sense/ 2 comments
- GitHub - microsoft/DeepSpeed: DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective. https://github.com/microsoft/DeepSpeed 1 comment
- Why Release a Large Language Model? | EleutherAI Blog https://blog.eleuther.ai/why-release-a-large-language-model/ 0 comments
- The Imperative for Sustainable AI Systems https://thegradient.pub/sustainable-ai/ 0 comments
- Latest News - DeepSpeed https://www.deepspeed.ai/ 0 comments
- GPT-3: We’re at the very beginning of a new app ecosystem | VentureBeat https://venturebeat.com/2021/02/27/gpt-3-were-at-the-very-beginning-of-a-new-app-ecosystem/ 0 comments
- Reflections on Foundation Models https://thegradient.pub/reflections-on-foundation-models/ 0 comments
- NLP Research Highlights — Issue #1 – DAIR.AI https://dair.ai/NLP_Research_Highlights_-_Issue_-1/ 0 comments
- Microsoft announces new supercomputer, lays out vision for future AI work - Source https://blogs.microsoft.com/ai/openai-azure-supercomputer/ 0 comments
- GitHub - accelerated-text/awesome-nlg: A curated list of resources dedicated to Natural Language Generation (NLG) https://github.com/tokenmill/awesome-nlg 0 comments
- Common Problems When Reproducing A Machine Learning Paper https://derekchia.com/common-problems-when-reproducing-a-machine-learning-paper/ 0 comments
- What is DeepSpeed? - by Michael Spencer https://datasciencelearningcenter.substack.com/p/what-is-deepspeed 0 comments
- GPT-3: Language Models are Few-Shot Learners | by Grigory Sapunov | Intento https://blog.inten.to/gpt-3-language-models-are-few-shot-learners-a13d1ae8b1f9 0 comments
- Microsoft's Massive New Language AI Is Triple the Size of OpenAI’s GPT-3 https://singularityhub.com/2021/10/13/microsofts-massive-new-language-ai-is-triple-the-size-of-openais-gpt-3/ 0 comments
- TorchServe and [TorchElastic for Kubernetes], new PyTorch libraries for serving and training models at scale | by PyTorch | PyTorch | Medium https://medium.com/pytorch/torchserve-and-torchelastic-for-kubernetes-new-pytorch-libraries-for-serving-and-training-models-2efd12e09adc 0 comments
Related searches:
Search whole site: site:microsoft.com
Search title: Turing-NLG: A 17-billion-parameter language model by Microsoft - Microsoft Research
See how to search.