research policy

30 articles about research policy in AI news

The Digital Twin Revolution: How LLMs Are Creating Virtual Testbeds for Social Media Policy

Researchers have developed an LLM-augmented digital twin system that simulates short-video platforms like TikTok to test policy changes before implementation. This four-twin architecture allows platforms to study long-term effects of AI tools and content policies in realistic closed-loop simulations.

79% relevant

One Policy to Rule Them All: AI Robot Masters Unseen Tools with Zero-Shot Generalization

Researchers have developed a single robot policy capable of manipulating diverse, never-before-seen tools using sim-to-real reinforcement learning. The system achieves zero-shot generalization across 24 tasks, 12 objects, and 6 tool categories without object-specific training.

85% relevant

Anthropic Forms Corporate PAC to Influence AI Policy Ahead of Midterms

Anthropic is forming a corporate PAC to lobby on AI policy, signaling a strategic shift towards direct political engagement as regulatory debates intensify in Washington. This move follows similar efforts by OpenAI and Google.

85% relevant

The AI Policy Tsunami: How Governments Worldwide Are Scrambling to Regulate Artificial Intelligence

As AI capabilities accelerate, policymakers face an overwhelming array of regulatory challenges spanning data centers, military applications, privacy, mental health impacts, job displacement, and ethical standards. The rapid pace of development is creating a governance gap that neither governments nor AI labs can adequately address.

85% relevant

The AI Policy Gap: Why Governments Are Struggling to Keep Pace with Rapid Technological Change

AI expert Ethan Mollick warns that rapid AI advancements combined with knowledge gaps and uncertain futures are leading to reactive, scattered policy responses rather than coherent governance frameworks.

85% relevant

Google's Cookie Policy Update and the Challenge of AI-Powered Personalization

Google has updated its user-facing cookie and data consent interface, emphasizing its use of data for personalization and ad measurement. This reflects the ongoing tension between data-driven AI services and user privacy, a critical issue for luxury retail's digital transformation.

82% relevant

ChatGPT's Android App Hints at Future 'Naughty Chats' Feature, Signaling a Potential Shift in AI Content Policy

A recent update to the ChatGPT Android app includes code referencing 'Naughty chats,' suggesting OpenAI may be developing an adult-themed, 18+ mode. This discovery hints at a potential strategic expansion into less restricted conversational AI.

85% relevant

The Self Driving Portfolio: Agentic Architecture for Institutional Asset Management

Researchers propose an 'agentic strategic asset allocation pipeline' using ~50 specialized AI agents to forecast markets, construct portfolios, and self-improve. The system is governed by a traditional Investment Policy Statement, aiming to automate high-level asset management.

88% relevant

CanViT: First Active-Vision Foundation Model Hits 45.9% mIoU on ADE20K with Sequential Glimpses

Researchers introduce CanViT, the first task- and policy-agnostic Active-Vision Foundation Model (AVFM). It achieves 38.5% mIoU on ADE20K segmentation with a single low-resolution glimpse, outperforming prior active models while using 19.5x fewer FLOPs.

91% relevant

AIGQ: Taobao's End-to-End Generative Architecture for E-commerce Query Recommendation

Alibaba researchers propose AIGQ, a hybrid generative framework for pre-search query recommendations. It uses list-level fine-tuning, a novel policy optimization algorithm, and a hybrid deployment architecture to overcome traditional limitations, showing substantial online improvements on Taobao.

100% relevant

SPREAD Framework Solves AI's 'Catastrophic Forgetting' Problem in Lifelong Learning

Researchers have developed SPREAD, a new AI framework that preserves learned skills across sequential tasks by aligning policy representations in low-rank subspaces. This breakthrough addresses catastrophic forgetting in lifelong imitation learning, enabling more stable and robust AI agents.

75% relevant

Mapping the Minefield: New Study Charts Five-Stage Taxonomy of LLM Harms

A new research paper systematically categorizes the potential harms of large language models across five lifecycle stages—from training to deployment—and argues that only multi-layered technical and policy safeguards can manage the risks.

95% relevant

MLLMRec-R1: A New Framework for Efficient Multimodal Sequential Recommendation with LLMs

Researchers propose MLLMRec-R1, a framework that makes Group Relative Policy Optimization (GRPO) practical for multimodal sequential recommendation by addressing computational cost and reward inflation issues. This enables more explainable, reasoning-based recommendations.

90% relevant

Beyond the Simplex: How Hilbert Space Geometry is Revolutionizing AI Alignment

Researchers have developed GOPO, a new alignment algorithm that reframes policy optimization as orthogonal projection in Hilbert space, offering stable gradients and intrinsic sparsity without heuristic clipping. This geometric approach addresses fundamental limitations in current reinforcement learning methods.

80% relevant

The Digital Detox Effect: How Phone-Free Schools Are Boosting Academic Performance

A landmark study reveals that banning mobile phones in schools significantly improves academic performance, particularly for struggling students. The research provides compelling evidence for educational policy changes worldwide.

85% relevant

Microsoft's EMPO²: A Memory-Augmented RL Framework That Supercharges LLM Agent Exploration

Microsoft has unveiled EMPO², a hybrid reinforcement learning framework that enhances LLM agents with augmented memory for true exploration. The system combines on- and off-policy optimization to discover novel states, achieving 128.6% performance gains over existing methods on ScienceWorld benchmarks.

85% relevant

AI Meets Infrastructure: OpenAI's New Tool Could Slash Federal Permitting Time by 15%

OpenAI has partnered with Pacific Northwest National Laboratory to launch DraftNEPABench, a benchmark showing AI coding agents can reduce National Environmental Policy Act drafting time by up to 15%. This collaboration signals AI's growing role in modernizing government processes.

75% relevant

Anthropic Abandons Core Safety Commitment Amid Intensifying AI Race

Anthropic has quietly removed a key safety pledge from its Responsible Scaling Policy, no longer committing to pause AI training without guaranteed safety protections. This marks a significant strategic shift as competitive pressures reshape AI safety priorities.

95% relevant

From Dismissed Warnings to Economic Reality: How AI's Job Disruption Forecasts Are Gaining Urgency

After two years of largely ignored warnings from AI lab CEOs about massive job displacement, workers and policymakers are beginning to take these predictions seriously as AI capabilities accelerate, creating new pressures on the industry.

85% relevant

GDPval Benchmark Reveals AI's Professional Competence: A New Tool for Economic Planning

A new interactive demonstration using OpenAI's GDPval benchmark shows current AI capabilities across economically valuable professional tasks. The project aims to make AI's real-world impact tangible for policymakers and civil society organizations, bridging the gap between technical assessments and practical economic decisions.

75% relevant

AI Researcher Kimmonismus Predicts AGI Within 6-12 Months, Widespread Worker Replacement in 1-2 Years

Independent AI researcher Kimmonismus predicts AGI will arrive within 6-12 months, with widespread worker displacement following in 1-2 years. The forecast, shared on X, adds to a growing chorus of near-term AGI predictions from industry figures.

85% relevant

New Research Proposes FilterRAG and ML-FilterRAG to Defend Against Knowledge Poisoning Attacks in RAG Systems

Researchers propose two novel defense methods, FilterRAG and ML-FilterRAG, to mitigate 'PoisonedRAG' attacks where adversaries inject malicious texts into a knowledge source to manipulate an LLM's output. The defenses identify and filter adversarial content, maintaining performance close to clean RAG systems.

92% relevant

China Surpasses US in AI Research Authorship with 2,152 First-Author Researchers in 2024

China now leads the US in first-author AI research contributions, with 2,152 researchers versus 1,810. This marks the first time China has overtaken the US in this key metric of research leadership.

87% relevant

Ex-OpenAI Researcher Daniel Kokotajlo Puts 70% Probability on AI-Caused Human Extinction by 2029

Former OpenAI governance researcher Daniel Kokotajlo publicly estimates a 70% chance of AI leading to human extinction within approximately five years. The claim, made in a recent interview, adds a stark numerical prediction to ongoing AI safety debates.

87% relevant

Google Researchers Challenge Singularity Narrative: Intelligence Emerges from Social Systems, Not Individual Minds

Google researchers argue AI's intelligence explosion will be social, not individual, observing frontier models like DeepSeek-R1 spontaneously develop internal 'societies of thought.' This reframes scaling strategy from bigger models to richer multi-agent systems.

87% relevant

AI Superintelligence Could Make Humans 'Obsolete as Baboons,' Warns Former OpenAI Researcher

Former OpenAI researcher Scott Aaronson warns that AI superintelligence could render humans obsolete within 25 years, comparing our potential future to baboons in zoos. He says global leadership is unprepared for this existential shift.

87% relevant

The Diversity Dilemma: New Research Challenges Assumptions About AI Alignment

A groundbreaking study reveals that moral reasoning in AI alignment may not require diversity-preserving algorithms as previously assumed. Researchers found reward-maximizing methods perform equally well, challenging conventional wisdom about how to align language models with human values.

86% relevant

AI Researchers Solve Critical LLM Confidence Problem with Novel Decoupling Technique

Researchers have identified and solved a fundamental conflict in how large language models learn reasoning versus confidence calibration. Their new DCPO framework preserves reasoning accuracy while dramatically reducing overconfidence in incorrect answers, addressing a major reliability concern for AI deployment.

75% relevant

New AI Research: Cluster-Aware Attention-Based Deep RL for Pickup and Delivery Problems

Researchers propose CAADRL, a deep reinforcement learning framework that explicitly models clustered spatial layouts to solve complex pickup and delivery routing problems more efficiently. It matches state-of-the-art performance with significantly lower inference latency.

79% relevant

Anthropic Challenges U.S. Government in Dual Lawsuits Over AI Research Restrictions

AI safety company Anthropic has filed lawsuits in two separate federal courts challenging U.S. government restrictions that have placed its research lab on an export blacklist. The legal action represents a significant confrontation between AI developers and regulatory authorities over research transparency and national security concerns.

85% relevant