AI Safety Newsletter
Episodes
AISN #69: Department of War, Anthropic, and National Security
13 Mar 2026
Contributed by Lukas
Also, Anthropic Removes a Core Safety Commitment. Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and...
AISN #68: Moltbook Exposes Risky AI Behavior
02 Feb 2026
Contributed by Lukas
Plus: The Pentagon Accelerates AI and GPT-5.2 solves open mathematics problems.. Welcome to the AI Safety Newsletter by the Center for AI Safety. We...
AISN #67: Trump’s preemption order, H200s go to China, and new frontier AI from OpenAI and DeepSeek
17 Dec 2025
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.. Wel...
AISN #67: Trump’s preemption order, H200s go to China, and new frontier AI from OpenAI and DeepSeek
17 Dec 2025
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.. Wel...
AISN #66: AISN #66: Evaluating Frontier Models, New Gemini and Claude, Preemption is Back
02 Dec 2025
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.. Wel...
AISN #65: Measuring Automation and Superintelligence Moratorium Letter
29 Oct 2025
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. In th...
AISN #63: New AGI Definition and Senate Bill Would Establish Liability for AI Harms
16 Oct 2025
Contributed by Lukas
In this edition: A new bill in the Senate would hold AI companies liable for harms their products create; China tightens its export controls on rare ...
AISN #63: California’s SB-53 Passes the Legislature
24 Sep 2025
Contributed by Lukas
In this edition: California's legislature sent SB-53—the ‘Transparency in Frontier Artificial Intelligence Act’—to Governor Newsom's desk. If...
AISN #62: Big Tech Launches $100 Million pro-AI Super PAC
27 Aug 2025
Contributed by Lukas
Also: Meta's Chatbot Policies Prompt Backlash Amid AI Reorganization; China Reverses Course on Nvidia H20 Purchases. In this edition: Big tech launc...
AISN #61: OpenAI Releases GPT-5
12 Aug 2025
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. In th...
AISN #60: The AI Action Plan
31 Jul 2025
Contributed by Lukas
Also: ChatGPT Agent and IMO Gold. In this edition: The Trump Administration publishes its AI Action Plan; OpenAI released ChatGPT Agent and announce...
AISN #59: EU Publishes General-Purpose AI Code of Practice
15 Jul 2025
Contributed by Lukas
Plus: Meta Superintelligence Labs. Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No ...
AISN #58: Senate Removes State AI Regulation Moratorium
03 Jul 2025
Contributed by Lukas
Plus: Judges Split on Whether Training AI on Copyrighted Material is Fair Use. In this edition: The Senate removes a provision from Republican's “...
AISN #57: The RAISE Act
17 Jun 2025
Contributed by Lukas
In this edition: The New York Legislature passes an act regulating frontier AI—but it may not be signed into law for some time. Listen to the AI Sa...
AISN #56: Google Releases Veo 3
28 May 2025
Contributed by Lukas
Plus, Opus 4 Demonstrates the Fragility of Voluntary Governance. In this edition: Google released a frontier video generation model at its annual de...
AISN #55: Trump Administration Rescinds AI Diffusion Rule, Allows Chip Sales to Gulf States
20 May 2025
Contributed by Lukas
Plus, Bills on Whistleblower Protections, Chip Location Verification, and State Preemption. In this edition: The Trump Administration rescinds the B...
AISN #54: OpenAI Updates Restructure Plan
13 May 2025
Contributed by Lukas
Plus, AI Safety Collaboration in Singapore. In this edition: OpenAI claims an updated restructure plan would preserve nonprofit control; A global co...
AISN #53: An Open Letter Attempts to Block OpenAI Restructuring
29 Apr 2025
Contributed by Lukas
Plus, SafeBench Winners. Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical ...
AISN #52: An Expert Virology Benchmark
22 Apr 2025
Contributed by Lukas
Plus, AI-Enabled Coups. In this edition: AI now outperforms human experts in specialized virology knowledge in a new benchmark; A new report explore...
AISN #51: AI Frontiers
15 Apr 2025
Contributed by Lukas
Plus, AI 2027. In this newsletter, we cover the launch of AI Frontiers, a new forum for expert commentary on the future of AI. We also discuss AI 20...
AISN #50: AI Action Plan Responses
31 Mar 2025
Contributed by Lukas
Plus, Detecting Misbehavior in Reasoning Models. In this newsletter, we cover AI companies’ responses to the federal government's request for info...
AISN #49: AI Action Plan Responses
31 Mar 2025
Contributed by Lukas
Plus, Detecting Misbehavior in Reasoning Models. In this newsletter, we cover AI companies’ responses to the federal government's request for info...
AISN
06 Mar 2025
Contributed by Lukas
Plus, Measuring AI Honesty. Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technic...
Superintelligence Strategy: Expert Version
05 Mar 2025
Contributed by Lukas
Superintelligence is destabilizing since it threatens other states’ survival—it could be weaponized, or states may lose control of it. Attempts to...
Superintelligence Strategy: Standard Version
05 Mar 2025
Contributed by Lukas
Superintelligence is destabilizing since it threatens other states’ survival—it could be weaponized, or states may lose control of it. Attempts to...
AISN #48: Utility Engineering and EnigmaEval
18 Feb 2025
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Liste...
AISN #47: Reasoning Models
06 Feb 2025
Contributed by Lukas
Plus, State-Sponsored AI Cyberattacks. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. Reasoning Models DeepSeek-R1 has be...
AISN #46: The Transition
23 Jan 2025
Contributed by Lukas
Plus, Humanity's Last Exam, and the AI Safety, Ethics, and Society Course. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts....
AISN #45: Center for AI Safety 2024 Year in Review
19 Dec 2024
Contributed by Lukas
As 2024 draws to a close, we want to thank you for your continued support for AI safety and review what we’ve been able to accomplish. In this spec...
AISN #44: The Trump Circle on AI Safety
19 Nov 2024
Contributed by Lukas
Plus, Chinese researchers used Llama to create a military tool for the PLA, a Google AI system discovered a zero-day cybersecurity vulnerability, and...
AISN #43: White House Issues First National Security Memo on AI
28 Oct 2024
Contributed by Lukas
Plus, AI and Job Displacement, and AI Takes Over the Nobels. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. White House I...
AISN #42: Newsom Vetoes SB 1047
01 Oct 2024
Contributed by Lukas
Plus, OpenAI's o1, and AI Governance Summary. Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI ...
AISN #41: The Next Generation of Compute Scale
11 Sep 2024
Contributed by Lukas
Plus, Ranking Models by Susceptibility to Jailbreaking, and Machine Ethics. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts...
AISN #40: California AI Legislation
21 Aug 2024
Contributed by Lukas
Plus, NVIDIA Delays Chip Production, and Do AI Safety Benchmarks Actually Measure Safety?. Listen to the AI Safety Newsletter for free on Spotify or...
AISN #39: Implications of a Trump Administration for AI Policy
29 Jul 2024
Contributed by Lukas
Plus, Safety Engineering Overview. Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts. Implications of a Trump administration ...
AISN #38: Supreme Court Decision Could Limit Federal Ability to Regulate AI
09 Jul 2024
Contributed by Lukas
Plus, “Circuit Breakers” for AI systems, and updates on China's AI industry. Listen to the AI Safety Newsletter for free on Spotify or Apple Pod...
AISN #37: US Launches Antitrust Investigations
18 Jun 2024
Contributed by Lukas
US Launches Antitrust Investigations The U.S. Government has launched antitrust investigations into Nvidia, OpenAI, and Microsoft. The U.S. Departmen...
AISN #36: Voluntary Commitments are Insufficient
30 May 2024
Contributed by Lukas
Voluntary Commitments are Insufficient AI companies agree to RSPs in Seoul. Following the second AI Global Summit held in Seoul, the UK and Republic ...
AISN #35: Lobbying on AI Regulation
16 May 2024
Contributed by Lukas
OpenAI and Google Announce New Multimodal Models In the current paradigm of AI development, there are long delays between the release of successive m...
AISN #34: New Military AI Systems
01 May 2024
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. AI La...
AISN #33: Reassessing AI and Biorisk
11 Apr 2024
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. This ...
AISN #32: Measuring and Reducing Hazardous Knowledge in LLMs
07 Mar 2024
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Measu...
AISN #31: A New AI Policy Bill in California
21 Feb 2024
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. This ...
AISN #30: Investments in Compute and Military AI
24 Jan 2024
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Compu...
AISN #29: Progress on the EU AI Act
04 Jan 2024
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. A Pro...
The Landscape of US AI Legislation
29 Dec 2023
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. This ...
AISN #28: Center for AI Safety 2023 Year in Review
21 Dec 2023
Contributed by Lukas
As 2023 comes to a close, we want to thank you for your continued support for AI safety. This has been a big year for AI and for the Center for AI Sa...
AISN #27: Defensive Accelerationism
07 Dec 2023
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required. Defen...
AISN #26: National Institutions for AI Safety
15 Nov 2023
Contributed by Lukas
Also, Results From the UK Summit, and New Releases From OpenAI and xAI. Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss de...
AISN #25: White House Executive Order on AI, UK AI Safety Summit, and Progress on Voluntary Evaluations of AI Risks.
31 Oct 2023
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.White H...
AISN #24: Kissinger Urges US-China Cooperation on AI, China’s New AI Law, US Export Controls, International Institutions, and Open Source AI.
18 Oct 2023
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.China's...
AISN #23: New OpenAI Models, News from Anthropic, and Representation Engineering.
04 Oct 2023
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.OpenAI ...
AISN #21: Google DeepMind’s GPT-4 Competitor, Military Investments in Autonomous Drones, The UK AI Safety Summit, and Case Studies in AI Policy.
05 Sep 2023
Contributed by Lukas
Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.Google ...
AISN #20: LLM Proliferation, AI Deception, and Continuing Drivers of AI Capabilities.
29 Aug 2023
Contributed by Lukas
AI Deception: Examples, Risks, SolutionsAI deception is the topic of a new paper from researchers at and affiliated with the Center for AI Safety. It ...
[Paper] “An Overview of Catastrophic AI Risks” by Dan Hendrycks, Mantas Mazeika and Thomas Woodside
21 Aug 2023
Contributed by Lukas
Rapid advancements in artificial intelligence (AI) have sparked growing concerns among experts, policymakers, and world leaders regarding the potentia...
[Paper] “X-Risk Analysis for AI Research” by Dan Hendrycks and Mantas Mazeika
21 Aug 2023
Contributed by Lukas
Artificial intelligence (AI) has the potential to greatly improve society, but as with any powerful technology, it comes with heightened risks and res...
[Paper] “Unsolved Problems in ML Safety” by Dan Hendrycks, Nicholas Carlini, John Schulman and Jacob Steinhardt
21 Aug 2023
Contributed by Lukas
Machine learning (ML) systems are rapidly increasing in size, are acquiring new capabilities, and are increasingly deployed in high-stakes settings. A...
AISN #19: US-China Competition on AI Chips, Measuring Language Agent Developments, Economic Analysis of Language Model Propaganda, and White House AI Cyber Challenge.
15 Aug 2023
Contributed by Lukas
US-China Competition on AI ChipsModern AI systems are trained on advanced computer chips which are designed and fabricated by only a handful of compan...
AISN #18: Challenges of Reinforcement Learning from Human Feedback, Microsoft’s Security Breach, and Conceptual Research on AI Safety.
08 Aug 2023
Contributed by Lukas
Challenges of Reinforcement Learning from Human FeedbackIf you’ve used ChatGPT, you might’ve noticed the “thumbs up” and “thumbs down” but...
AISN #17: Automatically Circumventing LLM Guardrails, the Frontier Model Forum, and Senate Hearing on AI Oversight.
01 Aug 2023
Contributed by Lukas
Automatically Circumventing LLM GuardrailsLarge language models (LLMs) can generate hazardous information, such as step-by-step instructions on how to...
AISN #16: White House Secures Voluntary Commitments from Leading AI Labs, and Lessons from Oppenheimer .
25 Jul 2023
Contributed by Lukas
White House Unveils Voluntary Commitments to AI Safety from Leading AI LabsLast Friday, the White House announced a series of voluntary commitments fr...
AISN #15: China and the US take action to regulate AI, results from a tournament forecasting AI risk, updates on xAI’s plan, and Meta releases its open-source and commercially available Llama 2.
19 Jul 2023
Contributed by Lukas
Both China and the US take action to regulate AILast week, regulators in both China and the US took aim at generative AI services. These actions show ...
AISN #14: OpenAI’s ‘Superalignment’ team, Musk’s xAI launches, and developments in military AI use .
12 Jul 2023
Contributed by Lukas
OpenAI announces a ‘superalignment’ teamOn July 5th, OpenAI announced the ‘Superalignment’ team: a new research team given the goal of alignin...
AISN #13: An interdisciplinary perspective on AI proxy failures, new competitors to ChatGPT, and prompting language models to misbehave.
05 Jul 2023
Contributed by Lukas
Interdisciplinary Perspective on AI Proxy FailuresIn this story, we discuss a recent paper on why proxy goals fail. First, we introduce proxy gaming, ...
AISN #12: Policy Proposals from NTIA’s Request for Comment, and Reconsidering Instrumental Convergence.
27 Jun 2023
Contributed by Lukas
Policy Proposals from NTIA’s Request for CommentThe National Telecommunications and Information Administration publicly requested comments on the ma...
AISN #11: An Overview of Catastrophic AI Risks.
22 Jun 2023
Contributed by Lukas
An Overview of Catastrophic AI RisksGlobal leaders are concerned that artificial intelligence could pose catastrophic risks. 42% of CEOs polled at the...
AISN #10: How AI could enable bioterrorism, and policymakers continue to focus on AI .
13 Jun 2023
Contributed by Lukas
How AI could enable bioterrorismOnly a hundred years ago, no person could have single handedly destroyed humanity. Nuclear weapons changed this situat...
AISN #9: Statement on Extinction Risks, Competitive Pressures, and When Will AI Reach Human-Level? .
06 Jun 2023
Contributed by Lukas
Top Scientists Warn of Extinction Risks from AILast week, hundreds of AI scientists and notable public figures signed a public statement on AI risks w...
AISN #8: Why AI could go rogue, how to screen for AI risks, and grants for research on democratic governance of AI.
30 May 2023
Contributed by Lukas
Yoshua Bengio makes the case for rogue AIAI systems pose a variety of different risks. Renowned AI scientist Yoshua Bengio recently argued for one par...
AISN #7: Disinformation, recommendations for AI labs, and Senate hearings on AI.
23 May 2023
Contributed by Lukas
How AI enables disinformationYesterday, a fake photo generated by an AI tool showed an explosion at the Pentagon. The photo was falsely attributed to ...
AISN #6: Examples of AI safety progress, Yoshua Bengio proposes a ban on AI agents, and lessons from nuclear arms control .
16 May 2023
Contributed by Lukas
Examples of AI safety progressTraining AIs to behave safely and beneficially is difficult. They might learn to game their reward function, deceive hum...
AISN #5: Geoffrey Hinton speaks out on AI risk, the White House meets with AI labs, and Trojan attacks on language models.
09 May 2023
Contributed by Lukas
Geoffrey Hinton is concerned about existential risks from AIGeoffrey Hinton won the Turing Award for his work on AI. Now he says that part of him regr...
AISN #4: AI and cybersecurity, persuasive AIs, weaponization, and Hinton talks AI risks.
02 May 2023
Contributed by Lukas
Cybersecurity Challenges in AI SafetyMeta accidentally leaks a language model to the public. Meta’s newest language model, LLaMa, was publicly leake...
AISN #3: AI policy proposals and a new challenger approaches.
25 Apr 2023
Contributed by Lukas
Policy Proposals for AI SafetyCritical industries rely on the government to protect consumer safety. The FAA approves new airplane designs, the FDA te...
AISN #2: ChaosGPT and the rise of language model agents, evolutionary pressures and AI, AI safety in the media.
18 Apr 2023
Contributed by Lukas
ChaosGPT and the Rise of Language AgentsChatbots like ChatGPT usually only respond to one prompt at a time, and a human user must provide a new prompt...
AISN #1: Public opinion on AI, plugging ChatGPT into the internet, and the economic impacts of language models..
10 Apr 2023
Contributed by Lukas
Growing concerns about rapid AI progressRecent advancements in AI have thrust it into the center of attention. What do people think about the risks of...