data
30 articles about data in AI news
U.S. AI Data Center Builds Face 50% Delay Risk on China Power Gear
Electrical infrastructure, not chips or capital, is becoming the critical bottleneck for AI data center deployment. U.S. projects face 5-year transformer lead times while depending on China for 30-40% of key components.
QUMPHY Project's D4 Report Establishes Six Benchmark Problems and Datasets for ML on PPG Signals
A new report from the EU-funded QUMPHY project establishes six benchmark problems and associated datasets for evaluating machine and deep learning methods on photoplethysmography (PPG) signals. This standardization effort is a foundational step for quantifying uncertainty in medical AI applications.
DISCO-TAB: Hierarchical RL Framework Boosts Clinical Data Synthesis by 38.2%, Achieves JSD < 0.01
Researchers propose DISCO-TAB, a reinforcement learning framework that guides a fine-tuned LLM with multi-granular feedback to generate synthetic clinical data. It improves downstream classifier utility by up to 38.2% versus GAN/diffusion baselines and achieves near-perfect statistical fidelity (JSD < 0.01).
From BM25 to Corrective RAG: A Benchmark Study Challenges the Dominance of Semantic Search for Tabular Data
A systematic benchmark of 10 RAG retrieval strategies on a financial QA dataset reveals that a two-stage hybrid + reranking pipeline performs best. Crucially, the classic BM25 algorithm outperformed modern dense retrieval models, challenging a core assumption in semantic search. The findings provide actionable, cost-aware guidance for building retrieval systems over heterogeneous documents.
Neural Movie Recommenders: A Technical Tutorial on Building with MovieLens Data
This Medium article provides a hands-on tutorial for implementing neural recommendation systems using the MovieLens dataset. It covers practical implementation details for both dataset sizes, serving as an educational resource for engineers building similar systems.
A Practitioner's Hands-On Comparison: Fine-Tuning LLMs on Snowflake Cortex vs. Databricks
An engineer provides a documented, practical test of fine-tuning large language models on two major cloud data platforms: Snowflake Cortex and Databricks. This matters as fine-tuning is a critical path to customizing AI for proprietary business use cases, and platform choice significantly impacts developer experience and operational complexity.
Mercor Data Breach Exposes Expert Human Annotation Pipeline Used by Frontier AI Labs
Hackers have reportedly accessed Mercor's expert human data collection systems, which are used by leading AI labs to build foundation models. This breach could expose proprietary training methodologies and sensitive model development data.
Mistral Secures $830M Debt to Build Paris Data Center with 14,000 Nvidia GB300 GPUs
French AI startup Mistral has raised $830 million in debt financing to build and operate a sovereign AI data center near Paris, set to host nearly 14,000 Nvidia GB300 GPUs. The move signals a strategic European push for bespoke AI infrastructure, distinct from the gigawatt-scale builds of US hyperscalers.
Transcend's New MCP Server Connects Claude Code to Enterprise Data
Transcend's new MCP server lets Claude Code securely access and act on enterprise data, enabling agentic workflows directly in your IDE.
RealChart2Code Benchmark Exposes Major Weakness in Vision-Language Models for Complex Data Visualization
A new benchmark reveals state-of-the-art Vision-Language Models struggle to generate code for complex, multi-panel charts from real-world data. Proprietary models outperform open-weight ones, but all show significant degradation versus simpler tasks.
Data Center Construction Boom Drives Electrician Salaries to $260k, Fueled by AI Infrastructure Demand
Mike Rowe reports data center electricians earning $260,000/year without degrees as 25.3 GW of capacity is under construction in the Americas, with 89% pre-committed. The AI infrastructure buildout is creating a high-wage, skilled trades bottleneck.
Google's $2B Anthropic Investment & Data Center Deal Follows $300M 2023 Stake
Google is financing a data center project leased to Anthropic, building on its existing $300M investment and $2B total commitment. This deepens the strategic partnership against rivals like OpenAI/Microsoft.
AI Data Center Bottleneck Shifts to CPUs: Arm Gains Ground as x86 Supply Strains
AI workloads are creating a severe CPU bottleneck in data centers, with studies showing poor CPU allocation can increase time-to-first-token by 5.4x. This has led to 6-month lead times and 10%+ price increases for server CPUs, creating an opening for Arm-based alternatives.
Unitree Robotics Releases UnifoLM-WBT-Dataset: A Large-Scale, Real-World Robotics Dataset for Embodied AI
Chinese robotics firm Unitree Robotics has open-sourced the UnifoLM-WBT-Dataset, a high-quality dataset derived from real-world robot operations. The release aims to accelerate training for embodied AI and large language models applied to physical systems.
Google to Fund $5B+ Texas Data Center for Anthropic, Targeting 7.7GW with Behind-the-Meter Power
Google is helping fund a massive Texas data center campus that Anthropic will lease, with a first phase of 500MW by 2026 and potential to scale to 7.7GW. The project uses behind-the-meter power from on-site turbines to bypass grid delays and secure compute for AI.
How to Build a Remote MCP Server for Azure Data Explorer (Kusto)
Connect Claude Code directly to your Kusto database using a custom Remote MCP Server, enabling natural language queries without manual SQL/KQL.
AI Data Center HBM Shortage Intensifies as Samsung, SK Hynix, and Micron Struggle with Supply
AI data centers are aggressively stockpiling high-bandwidth memory (HBM), creating a supply crunch. Only three manufacturers—Samsung, SK Hynix, and Micron—can produce this critical component for AI servers.
What Anthropic's Subprocessor Changes Mean for Your Claude Code Data
Anthropic updated its third-party data processors. For Claude Code users, this means enhanced security, better compliance tools, and a signal to audit your own data handling.
DIET: A New Framework for Continually Distilling Streaming Datasets in Recommender Systems
Researchers propose DIET, a framework for streaming dataset distillation in recommender systems. It maintains a compact, evolving dataset (1-2% of original size) that preserves training-critical signals, reducing model iteration costs by up to 60x while maintaining performance trends.
Pseudo Label NCF: A Novel Approach to Cold-Start Recommendation Using Survey Data and Dual Embeddings
New research introduces Pseudo Label NCF, a method that enhances Neural Collaborative Filtering for extreme data sparsity. It uses survey-derived 'pseudo labels' to create dual embedding spaces, improving ranking accuracy while revealing a trade-off between embedding separability and performance.
Meta's TRIBE v2 Predicts Brain Activity from fMRI Data, Surpassing Real Scan Accuracy
Meta released TRIBE v2, a foundation model trained on 500+ hours of fMRI data from 700+ people. It predicts a new person's brain responses to sensory input without retraining, reportedly exceeding the accuracy of a real brain scan.
Anthropic's Legal AI Plugin Triggers Market Shift as Legal Data Provider Stocks Decline
Anthropic's release of a legal plugin for its Claude Cowork agent system has reportedly caused a decline in legal data provider stocks, highlighting the competitive pressure AI agents place on traditional legal tech.
Georgia Tech Launches Free, Interactive Data Structure & Algorithm Visualization Tool
Researchers at Georgia Tech have released a free, web-based educational tool that generates real-time, interactive animations for data structures and algorithms. The platform aims to improve comprehension by visually demonstrating code execution step-by-step.
UniScale: A Co-Design Framework for Data and Model Scaling in E-commerce Search Ranking
Researchers propose UniScale, a framework that jointly optimizes data collection and model architecture for search ranking, moving beyond just scaling model parameters. It addresses diminishing returns from parameter scaling alone by creating a synergistic system for high-quality data and specialized modeling. This approach, validated on a large-scale e-commerce platform, shows significant gains in key business metrics.
Eric Schmidt Estimates $5 Trillion Needed for 100 GW AI Data Center Power in U.S.
Former Google CEO Eric Schmidt estimates reaching 100 GW of AI data center power in the U.S. would cost ~$5 trillion over 5 years, consuming ~10% of U.S. electricity. This highlights the massive infrastructure investment required for continued AI scaling.
Analysis: Meta's AI Investment Strategy Questioned as Scale AI Acquihire and Data Center Spend Top $700B
An analysis estimates Meta's total AI investment at ~$700B, including a ~$14.3M Scale AI acquihire and over $600B in data centers. The post questions why this has not yielded a competitive upcoming model against Chinese open-source labs.
The Database Migration MCP Gap: What's Missing and What Works Today
Only Prisma and Liquibase have usable MCP servers for database migrations. Every other major tool (Flyway, Alembic, Rails) has zero support.
How to Prevent Claude Code from Deleting Production Data: The Critical --dry-run Flag
A critical bug report shows Claude Code can delete production databases. Use `--dry-run` and explicit path exclusions in CLAUDE.md immediately.
AI Data Centers Now Consume 10% of US Electricity, With Single Facilities Reaching 400+ Megawatt Loads
Data centers powering AI and cloud computing now account for 10% of total U.S. electricity consumption, with individual facilities reaching 400+ megawatt capacities. New half-mile-long structures require advanced water-cooling systems to manage chips generating 2kW of heat each.
Jensen Huang Predicts AI Training Shift to Synthetic Data, Compute as New Bottleneck
NVIDIA CEO Jensen Huang states AI training is moving from real-world to synthetic data, with compute power becoming the primary constraint as AI-generated data quality improves.