google research

30 articles about google research in AI news

Google Research Publishes TurboQuant Paper, Claiming 80% AI Cost Reduction

Google Research has published a technical paper introducing TurboQuant, a new AI model quantization method that reportedly reduces memory usage by 6x and could cut AI inference costs by 80%. The research suggests significant implications for AI infrastructure economics and hardware investment strategies.

85% relevant

Google Researchers Challenge Singularity Narrative: Intelligence Emerges from Social Systems, Not Individual Minds

Google researchers argue AI's intelligence explosion will be social, not individual, observing frontier models like DeepSeek-R1 spontaneously develop internal 'societies of thought.' This reframes scaling strategy from bigger models to richer multi-agent systems.

87% relevant

Google Research's TurboQuant Achieves 6x LLM Compression Without Accuracy Loss, 8x Speedup on H100

Google Research introduced TurboQuant, a novel compression algorithm that shrinks LLM memory footprint by 6x without retraining or accuracy drop. Its 4-bit version delivers 8x faster processing on H100 GPUs while matching full-precision quality.

95% relevant

Google Quantum AI Team Reduces Bitcoin-Cracking Qubit Estimate to ~500k, Enabling 9-Minute Key Derivation

Google researchers have compiled Shor's algorithm to solve Bitcoin's 256-bit elliptic curve problem with ~1.2k logical qubits, translating to <500k physical qubits—a 20x reduction from 2023 estimates. This makes 'on-spend' attacks against unconfirmed transactions theoretically plausible with fast-clock quantum hardware.

95% relevant

Google's TurboQuant Compresses LLM KV Cache 6x with Zero Accuracy Loss, Cutting GPU Memory by 80%

Google researchers introduced TurboQuant, a method that compresses LLM KV cache from 32-bit to 3-bit precision without accuracy degradation. This reduces GPU memory consumption by over 80% and speeds up inference 8x on H100 GPUs.

97% relevant

Google's Bayesian Breakthrough: Teaching AI to Think with Uncertainty

Google researchers have developed a new training method that teaches large language models to reason probabilistically, addressing a fundamental weakness in current AI systems. This 'Bayesian upgrade' enables models to update beliefs with new evidence rather than relying on static training data.

80% relevant

Google's 'Deep-Thinking Ratio' Breakthrough: Smarter AI Reasoning at Half the Cost

Google researchers have developed a 'Deep-Thinking Ratio' metric that identifies when AI models are genuinely reasoning versus just generating longer text. This breakthrough improves accuracy while cutting inference costs by approximately 50% through early halting of unpromising computations.

85% relevant

Google's TimesFM Foundation Model: A New Paradigm for Time Series Forecasting

Google Research has open-sourced TimesFM, a 200 million parameter foundation model for time series forecasting. Trained on 100 billion real-world time points, it demonstrates remarkable zero-shot forecasting capabilities across diverse domains without task-specific training.

85% relevant

Zero-Shot Cross-Domain Knowledge Distillation: A YouTube-to-Music Case Study

Google researchers detail a case study transferring knowledge from YouTube's massive video recommender to a smaller music app, using zero-shot cross-domain distillation to boost ranking models without training a dedicated teacher. This offers a practical blueprint for improving low-traffic AI systems.

96% relevant

Google's TurboQuant AI Research Report Sparks Sell-Off in Micron, Samsung, and SK Hynix Memory Stocks

Google's TurboQuant research blog publication triggered immediate market reaction, with shares of major memory manufacturers dropping 2-4% as investors anticipate AI-driven efficiency gains reducing future memory demand.

85% relevant

Sergey Brin Returns to Google AI Research, Citing 'Exciting' Technical Progress

Google co-founder Sergey Brin has resumed a hands-on role in AI research, attending weekly meetings and reviewing technical documents. His return is driven by the 'exciting' pace of progress in the field.

87% relevant

Stanford, Google, MIT Paper Claims LLMs Can Self-Improve Prompts

A collaborative paper from Stanford, Google, and MIT researchers indicates large language models can self-improve their prompts via iterative refinement. This could automate a core task currently performed by human prompt engineers.

87% relevant

Google DeepMind Maps Six 'AI Agent Traps' That Can Hijack Autonomous Systems in the Wild

Google DeepMind has published a framework identifying six categories of 'traps'—from hidden web instructions to poisoned memory—that can exploit autonomous AI agents. This research provides the first systematic taxonomy for a growing attack surface as agents gain web access and tool-use capabilities.

95% relevant

Google Lyria 3 Pro Music AI Demoed: Generates '1990s Boy Band' Version of Rilke Poetry

A researcher gained early access to Google's Lyria 3 Pro music generation AI, demonstrating its ability to transform Rainer Maria Rilke's 'First Elegy' into a 1990s boy band track. The demo highlights rapid stylistic remixing capabilities not yet publicly available.

85% relevant

Google DeepMind's 'Learning Through Conversation' Paper Shows LLMs Can Improve with Real-Time Feedback

Google DeepMind researchers have published a paper demonstrating that large language models can be trained to learn and improve their responses during a conversation by incorporating user feedback, moving beyond static pre-training.

85% relevant

Google DeepMind Proposes 'Intelligent AI Delegation' Framework for Dynamic Task Handoffs with Verifiable Trust

Google DeepMind researchers propose a formal framework for delegating tasks to AI agents, treating delegation as a structured process with dynamic trust models, verifiable proofs, and failure management. The system is designed to prevent over- or under-delegation and enable AI-to-AI task handoffs with clear accountability.

97% relevant

Google's Groundsource: Using AI to Mine Historical Disaster Data from Global News

Google AI Research has unveiled Groundsource, a novel methodology using the Gemini model to transform unstructured global news reports into structured historical datasets. The system addresses critical data gaps in disaster management, starting with 2.6 million urban flash flood events.

75% relevant

Spine Swarms: How an 8-Person Team Outperformed AI Giants in Deep Research

A small team of engineers has developed Spine Swarms, an AI system that reportedly outperforms Google, Perplexity, Claude, and GPT-5.2 in deep research tasks. This breakthrough demonstrates how agile teams can compete with tech giants in specialized AI applications.

95% relevant

NotebookLM's PowerPoint Integration: AI Research Assistant Evolves into Presentation Creator

Google's NotebookLM has expanded beyond research summarization to include slide generation and editing capabilities with direct PowerPoint export. This transforms the AI research assistant into a complete presentation workflow tool.

85% relevant

Google DeepMind Reveals Fundamental Flaw in Diffusion Model Training

Google DeepMind researchers have identified a critical weakness in how diffusion models are trained, challenging the standard approach of borrowing KL penalties from VAEs. Their new paper reveals this method lacks principled control over latent information, potentially limiting model performance.

85% relevant

Google DeepMind's Breakthrough: LLMs Now Designing Their Own Multi-Agent Learning Algorithms

Google DeepMind researchers have demonstrated that large language models can autonomously discover novel multi-agent learning algorithms, potentially revolutionizing how we approach complex AI coordination problems. This represents a significant shift toward AI systems that can design their own learning strategies.

85% relevant

Google's RT-X Project Establishes New Robot Learning Standard

Google's RT-X project has established a new standard for robot learning by creating a unified dataset of detailed human demonstrations across 22 institutions and 30+ robot types. This enables large-scale cross-robot training previously impossible with fragmented data.

85% relevant

PhD Researcher Replaces Notion & Email Tools with AI Agent 'Muse'

A researcher has reportedly replaced multiple productivity tools (Notion, note-taking apps, inbox triage) with a custom AI agent named 'Muse'. This highlights a growing trend of using specialized AI agents to consolidate workflows.

87% relevant

OpenAI President Teases 'Spud' Model, Two Years of Research

OpenAI President Greg Brockman briefly mentioned an upcoming model codenamed 'Spud', stating it represents 'two years worth of research that is coming to fruition.' No technical details or release timeline were provided.

85% relevant

Google News Feed Shows AI Virtual Try-On as Active Retail Trend

A Google News feed item highlights 'Fashion Retailers Adopt AI Virtual Try-On' as a topic. This indicates the technology has reached a threshold of news volume and engagement to be surfaced by algorithms as a significant trend, not a niche experiment.

72% relevant

Sam Altman Outlines 3 AI Futures: Research, Operations, Personal Agents

OpenAI CEO Sam Altman outlined three potential outcomes for AI development: systems that conduct scientific research, accelerate company operations, and serve as trusted personal agents. This vision frames the strategic direction for OpenAI and the broader industry.

85% relevant

AI Research Loop Paper Claims Automated Experimentation Can Accelerate AI Development

A shared paper highlights research into using AI to run a mostly automated loop of experiments, suggesting a method to speed up AI research itself. The source notes a potential problem with the approach but does not specify details.

85% relevant

ASI-Evolve Automates AI Research Loop, Discovers 105 Better Linear Attention Designs and Boosts AMC32 Scores by 12.5 Points

Researchers developed ASI-Evolve, an AI system that automates experimental loops in AI research. It discovered 105 improved linear attention variants and boosted AMC32 scores by 12.5 points, demonstrating automated research acceleration.

95% relevant

OpenAI Reallocates Compute and Talent Toward 'Automated Researchers' and Agent Systems

OpenAI is reallocating significant compute resources and engineering talent toward developing 'automated researchers' and agent-based systems capable of executing complex tasks end-to-end, signaling a strategic pivot away from some existing projects.

89% relevant

Google Launches Gemini API 'Flex' & 'Turbo' Tiers, Cuts Standard Pricing by 50%

Google has added 'Flex' and 'Turbo' service tiers to its Gemini API, with Flex offering a 50% reduction in cost compared to Standard. This move provides developers with more granular control over cost versus latency for their AI applications.

87% relevant