risk
30 articles about risk in AI news
OpenAI Shelves 'Adult Mode' Chatbot Indefinitely, Citing Safety Risks and Strategic Refocus
OpenAI has canceled its planned erotic chatbot feature after internal pushback over risks to minors and technical safety challenges. The move is part of a broader shift away from experimental 'side quests' toward core productivity tools.
Judge Questions Legality of Pentagon's 'Supply Chain Risk' Designation Against Anthropic, Calls Actions 'Troubling'
A U.S. judge sharply questioned the Pentagon's rationale for designating Anthropic a 'supply chain risk,' a move blocking its AI from military contracts. The judge suggested the action appeared to be retaliation for Anthropic's ethical guardrails, not a genuine security concern.
Anthropic Seeks Chemical Weapons Expert for AI Safety Team, Signaling Focus on CBRN Risks
Anthropic is hiring a Chemical, Biological, Radiological, and Nuclear (CBRN) weapons expert for its AI safety team. The role focuses on assessing and mitigating catastrophic risks from frontier AI models.
OpenAI Delays 'Adult Mode' for ChatGPT Amid Internal Backlash Over Safety Risks
OpenAI has delayed a proposed 'adult mode' for ChatGPT following internal warnings about risks including emotional dependency, compulsive use, and inadequate age verification with a ~12% error rate.
JPMorgan CEO Jamie Dimon: AI Could Enable 4-Day Work Week, Already Used for Risk, Marketing, Underwriting
JPMorgan Chase CEO Jamie Dimon stated AI could enable a 4-day work week. He detailed current uses in risk calculation, marketing, and underwriting.
NVIDIA VP Kari Briski to Discuss Nemotron 3 Super Development in Upcoming Interview
NVIDIA VP Kari Briski will be interviewed on Thursday about the company's Nemotron models, specifically the recent Nemotron 3 Super. The recorded conversation will be published by NVIDIA.
Andrej Karpathy Analysis: AI Poses High Risk to 57 Million US Jobs, ~40% of Workforce
Andrej Karpathy's analysis concludes AI puts 57 million US workers at high to very high risk of negative job impact. This ~40% figure contextualizes recent tech layoffs and discussions around universal high income.
Amazon's AI Agent Incident Highlights Critical Risks of Unsupervised Automation in Retail
Amazon's retail website suffered multiple high-severity outages linked to an engineer acting on inaccurate advice from an AI agent that sourced information from an outdated internal wiki. This incident underscores the operational risks of deploying autonomous AI agents without proper human oversight and data governance in critical retail systems.
Anthropic Takes Legal Stand: AI Company Sues Pentagon Over 'Supply Chain Risk' Designation
AI safety company Anthropic has filed two lawsuits against the Pentagon after being labeled a 'supply chain risk'—a designation typically applied to foreign adversaries. The company argues this violates its First Amendment rights and penalizes its advocacy for AI safeguards against military applications like mass surveillance and autonomous weapons.
Safety Gap: OpenAI's Most Powerful AI Models Released Without Critical Risk Assessments
OpenAI's GPT-5.4 Pro, potentially the world's most capable AI for high-risk tasks like bioweapons research and cyber operations, has been released without published safety evaluations or system cards, continuing a concerning pattern with 'Pro' model releases.
AI Deciphers Patient Language to Predict Stroke Risk with Unprecedented Precision
Researchers have developed an AI system that analyzes patient-reported symptoms to detect early stroke risk in diabetic individuals. Using graph neural networks and patient-centered language, the system achieves near-perfect predictive accuracy while minimizing false alarms.
FT's AI Risk Chart Sparks Debate: 50% Chance of Human Extinction Versus Abundance
A Financial Times chart showing AI could lead to either human extinction or unprecedented abundance has ignited debate about mainstream recognition of existential risks. The visualization presents a stark 50/50 probability between catastrophic and utopian outcomes.
Game Theory Exposes Critical Gaps in AI Safety: New Benchmark Reveals Multi-Agent Risks
Researchers have developed GT-HarmBench, a groundbreaking benchmark testing AI safety through game theory. The study reveals frontier models choose socially beneficial actions only 62% of time in multi-agent scenarios, highlighting significant coordination risks.
U.S. AI Data Center Builds Face 50% Delay Risk on China Power Gear
Electrical infrastructure, not chips or capital, is becoming the critical bottleneck for AI data center deployment. U.S. projects face 5-year transformer lead times while depending on China for 30-40% of key components.
Claude-to-IM Skill: Get Claude Code in Your Team Chat (Without OpenClaw's Security Risks)
Open-source bridge brings Claude Code to Telegram/Discord with permission prompts, streaming, and persistent sessions—safer alternative to OpenClaw.
Andrej Karpathy's Deleted Tool: AI Exposure Scores for 342 Jobs, Finds $3.7T in High-Risk Wages
Andrej Karpathy briefly released a tool scoring 342 job types for AI exposure using an LLM, finding an average score of 5.3/10. The analysis identified $3.7 trillion in annual wages at high exposure (7+), with software developers at 9/10 and medical transcriptionists at 10/10.
How to Audit Your CLAUDE.md to Prevent 'Ghost File' Security Risks
A security researcher demonstrated how vague CLAUDE.md instructions can create hidden 'ghost files'—here's how to audit your prompts to prevent this.
The Agent Alignment Crisis: Why Multi-AI Systems Pose Uncharted Risks
AI researcher Ethan Mollick warns that practical alignment for AI agents remains largely unexplored territory. Unlike single AI systems, agents interact dynamically, creating unpredictable emergent behaviors that challenge existing safety frameworks.
The Hidden Risk in Your AI Agent's Instruction Manual: When More Context Backfires
New research reveals that overloading AI coding agents with excessive context in AGENTS.md files can actually degrade their performance. The study challenges the assumption that more information always leads to better results, highlighting a critical optimization point for developers.
Anthropic CEO Warns of Military AI Risks: The Accountability Crisis in Autonomous Warfare
Anthropic CEO Dario Amodei raises alarms about selling unreliable AI technology for military use, warning of civilian harm and accountability gaps in concentrated drone fleets. He calls for urgent oversight conversations.
Meta Halts Mercor Work After Supply Chain Breach Exposes AI Training Secrets
A supply chain attack via compromised software updates at data-labeling vendor Mercor has forced Meta to pause collaboration, risking exposure of core AI training pipelines and quality metrics used by top labs.
Insider Knowledge: How Much Can RAG Systems Gain from Evaluation Secrets?
New research warns that RAG systems can be gamed to achieve near-perfect evaluation scores if they have access to the evaluation criteria, creating a risk of mistaking metric overfitting for genuine progress. This highlights a critical vulnerability in the dominant LLM-judge evaluation paradigm.
Neko Health Launches $400 AI-Powered Full-Body Health Scans in New York This Spring
Neko Health, the $1.8B startup founded by Spotify's Daniel Ek, is launching its AI-driven full-body health screening service in the US. The $400 scan uses imaging and blood tests to screen for cancer, heart disease, and diabetes risk, though medical experts are divided on its efficacy.
How to Use Claude Code's New 'Auto Mode' for Safer Desktop Automation
Claude Code's new 'Auto Mode' lets you delegate tasks to run autonomously on your desktop, but you must configure it correctly to avoid security risks.
Why Cheaper LLMs Can Cost More: The Hidden Economics of AI Inference in 2026
A Medium article outlines a practical framework for balancing performance, cost, and operational risk in real-world LLM deployment, arguing that focusing solely on model cost can lead to higher total expenses.
Anthropic's Opus 5 and OpenAI's 'Spud' Rumored as Major AI Leaps, Prompting Security Concerns
A Fortune report, cited on social media, claims Anthropic's upcoming Opus 5 model is a 'massive leap' from Claude 3.5 Sonnet, posing significant security risks. OpenAI is also rumored to have a similarly advanced model, 'Spud,' in development.
Claude Code's New Auto Mode: Run Commands Without Constant Permission Prompts
Claude Code's new Auto Mode uses a safety classifier to autonomously execute safe actions while blocking risky ones, eliminating constant permission prompts for routine tasks.
Anthropic Launches Claude Code Auto Mode Preview, a Safety Classifier to Prevent Mass File Deletions
Anthropic is previewing 'auto mode' for Claude Code, a classifier that autonomously executes safe actions while blocking risky ones like mass deletions. The feature, rolling out to Team, Enterprise, and API users, follows high-profile incidents like a recent AWS outage linked to an AI tool.
Claude Code's New 'Auto Mode' Preview: What's Allowed, What's Blocked, and How to Get Access
Anthropic's new safety classifier for Claude Code autonomously executes safe actions while blocking risky ones. Here's how it works and how to use it.
Atlassian's Official MCP Server vs. The Community Version: Which Should You Connect to Claude Code?
Atlassian's official MCP server is GA, but critical bugs and a more powerful community alternative mean your choice depends on your stack and tolerance for risk.