Linking pages
- World-first research dissects an AI's mind, and starts editing its thoughts https://newatlas.com/technology/ai-thinking-patterns/ 360 comments
- Dario Amodei — Machines of Loving Grace https://darioamodei.com/machines-of-loving-grace 143 comments
- Here comes the Muybridge camera moment but for text. Photoshop too (Interconnected) https://interconnected.org/home/2024/05/31/camera 115 comments
- Golden Gate Claude \ Anthropic https://www.anthropic.com/news/golden-gate-claude 66 comments
- GitHub - SalvatoreRa/ML-news-of-the-week: A collection of the the best ML and AI news every week (research, news, resources) https://github.com/SalvatoreRa/ML-news-of-the-week 8 comments
- A.I.’s Black Boxes Just Got a Little Less Mysterious - The New York Times https://www.nytimes.com/2024/05/21/technology/ai-language-models-anthropic.html 2 comments
- AI #65: I Spy With My AI - by Zvi Mowshowitz https://thezvi.substack.com/p/ai-65-i-spy-with-my-ai 0 comments
- Golden Gate Claude \ Anthropic https://www.anthropic.com/news/golden-gate-claude?p=2 0 comments
- The Schumer Report on AI (RTFB) - by Zvi Mowshowitz https://thezvi.substack.com/p/the-schumer-report-rtfb 0 comments
- I am the Golden Gate Bridge - by Zvi Mowshowitz https://thezvi.substack.com/p/i-am-the-golden-gate-bridge 0 comments
- Golden Gate Claude: What is it? - Claude101 https://claude101.com/golden-gate-claude/ 0 comments
Linked pages
- Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet https://transformer-circuits.pub/2024/scaling-monosemanticity/ 135 comments
- GitHub - openai/transformer-debugger https://github.com/openai/transformer-debugger 120 comments
- [2310.13548] Towards Understanding Sycophancy in Language Models https://arxiv.org/abs/2310.13548 72 comments
- Claude https://claude.ai/ 55 comments
- https://www-cdn.anthropic.com/de8ba9b01c9ab7cbabf5c33b80b7bbc618857627/Model_Card_Claude_3.pdf 3 comments
- [2001.08361] Scaling Laws for Neural Language Models https://arxiv.org/abs/2001.08361 0 comments
- [2310.03693] Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To! https://arxiv.org/abs/2310.03693 0 comments
- [2404.16014] Improving Dictionary Learning with Gated Sparse Autoencoders https://arxiv.org/abs/2404.16014 0 comments
Related searches:
Search whole site: site:www.anthropic.com
Search title: Mapping the Mind of a Large Language Model \ Anthropic
See how to search.