id stringlengths 7 35 | source stringclasses 8 values | text stringlengths 0 1.73M | num_tokens int64 0 386k | __index_level_0__ int64 0 146 |
|---|---|---|---|---|
ai_2727 | ai_2027 | "# AI\n\n# 2027\n\n## Daniel Kokotajlo\n\n## Scott Alexander\n\n## Thomas Larsen\n\n## Eli Lifland\n(...TRUNCATED) | 55,231 | 0 |
intro_ai_safety_textbook | intro_ai_safety_textbook | "## Introduction to AI Safety,\n\n## Ethics, and Society\n\nThis is a draft manuscript. The final pr(...TRUNCATED) | 386,132 | 1 |
situational_awarness_superalignment | situational_awarness | "### IIIc. Superalignment\n\n**Reliably controlling AI systems much smarter than we are\nis an unsol(...TRUNCATED) | 10,812 | 2 |
anthropic-blog-0 | anthropic_blog | "Evaluating honesty and lie detection techniques on a diverse suite of dishonest models\nRowan WangN(...TRUNCATED) | 15,590 | 3 |
anthropic-blog-1 | anthropic_blog | "Strengthening Red Teams: A Modular Scaffold for Control Evaluations\nChloe Loughridge1, Paul Cologn(...TRUNCATED) | 3,540 | 4 |
anthropic-blog-2 | anthropic_blog | "Anthropic's Pilot Sabotage Risk Report\nMain Report:\nSamuel R. Bowman, Misha Wagner, Fabien Roger,(...TRUNCATED) | 1,181 | 5 |
anthropic-blog-3 | anthropic_blog | "Stress-testing model specs reveals character differences among language models\n\n\nJifan Zhang1, H(...TRUNCATED) | 3,363 | 6 |
anthropic-blog-4 | anthropic_blog | "Believe It or Not: How Deeply do LLMs Believe Implanted Facts?\nStewart Slocum1,\nOctober 21, 2025\(...TRUNCATED) | 2,013 | 7 |
anthropic-blog-5 | anthropic_blog | "Inoculation Prompting: Instructing LLMs to misbehave at train-time improves test-time alignment\nNe(...TRUNCATED) | 1,283 | 8 |
anthropic-blog-6 | anthropic_blog | "Training fails to elicit subtle reasoning in current language models\ntl;dr\nWhile recent AI system(...TRUNCATED) | 27,434 | 9 |
End of preview. Expand
in Data Studio
README.md exists but content is empty.
- Downloads last month
- 18