3d reconstruction
15 articles about 3d reconstruction in AI news
Radar Meets AI: How RF Signals Are Revolutionizing 3D Scene Reconstruction
Researchers have developed a multimodal approach combining radio-frequency sensing with Gaussian Splatting to create robust 3D scene rendering that works in challenging conditions where vision alone fails. This breakthrough enables high-fidelity reconstruction in adverse weather, low light, and through occlusions.
NVIDIA Releases NVPanoptix-3D on Hugging Face: Single-Image 3D Indoor Scene Reconstruction
NVIDIA has open-sourced NVPanoptix-3D, a model that reconstructs complete 3D indoor scenes—including panoptic segmentation, depth, and geometry—from a single RGB image in one forward pass.
NVIDIA DLSS 5 Demo Shows 3D Guided Neural Rendering for Next-Gen Upscaling
A leaked demo of NVIDIA's upcoming DLSS 5 technology showcases 3D guided neural rendering, promising a significant leap in image reconstruction quality for real-time graphics.
Meshcraft Democratizes 3D Creation: Multi-Engine AI Platform Bridges Text-to-3D Gap
Meshcraft emerges as a web-based platform offering text-to-3D and image-to-3D generation with selectable AI engines. The tool provides both free and premium options, addressing quality bottlenecks in 3D generation through engine optimization rather than image model refinement.
Utonia AI Breakthrough: A Single Transformer Model Unifies All 3D Point Cloud Data
Researchers have developed Utonia, a single self-supervised transformer that learns unified 3D representations across diverse point cloud data types including LiDAR, CAD models, indoor scans, and video-lifted data. This breakthrough enables unprecedented cross-domain transfer and emergent behaviors in 3D AI.
VGGT-Det: How AI Is Learning to See in 3D Without Camera Calibration
Researchers have developed VGGT-Det, a breakthrough framework for multi-view 3D object detection that works without calibrated camera poses. The system mines internal geometric priors through attention mechanisms, outperforming traditional methods in indoor environments.
Beyond the Loss Function: New AI Architecture Embeds Physics Directly into Neural Networks for 10x Faster Wave Modeling
Researchers have developed a novel Physics-Embedded PINN that integrates wave physics directly into neural network architecture, achieving 10x faster convergence and dramatically reduced memory usage compared to traditional methods. This breakthrough enables large-scale 3D wave field reconstruction for applications from wireless communications to room acoustics.
KitchenTwin: VLM-Guided Scale Recovery Fuses Global Point Clouds with Object Meshes for Metric Digital Twins
Researchers propose KitchenTwin, a scale-aware 3D fusion framework that registers object meshes with transformer-predicted global point clouds using VLM-guided geometric anchors. The method resolves fundamental coordinate mismatches to build metrically consistent digital twins for embodied AI, and releases an open-source dataset.
Sparse Sensors, Rich Views: How Minimal Radar Data Supercharges AI Scene Generation
Researchers have developed a novel approach that combines single images with extremely sparse radar or LiDAR data to dramatically improve AI's ability to generate realistic 3D views from 2D photos. This multimodal technique overcomes fundamental limitations of vision-only systems in challenging conditions like bad weather and low texture.
Geometric Latent Diffusion (GLD) Achieves SOTA Novel View Synthesis, Trains 4.4× Faster Than VAE
GLD repurposes features from geometric foundation models like Depth Anything 3 as a latent space for multi-view diffusion. It trains significantly faster than VAE-based approaches and achieves state-of-the-art novel view synthesis without text-to-image pretraining.
CoRe Framework Integrates Equivariant Contrastive Learning for Medical Image Registration, Surpassing Baseline Methods
Researchers propose CoRe, a medical image registration framework that jointly optimizes an equivariant contrastive learning objective with the registration task. The method learns deformation-invariant feature representations, improving performance on abdominal and thoracic registration tasks.
NVIDIA Employees Clarify DLSS5 Does Not Alter Character Models or Assets, Only Lighting
NVIDIA employees confirmed at a press conference that DLSS5 makes no changes to character models or game assets, countering speculation about AI filters. The visual differences are attributed solely to lighting changes.
New Benchmark Exposes Critical Weakness in Multimodal AI: Object Orientation
A new AI benchmark, DORI, reveals that state-of-the-art vision-language models perform near-randomly on object orientation tasks. This fundamental spatial reasoning gap has direct implications for retail applications like virtual try-on and visual search.
TimeGS: How Computer Graphics Techniques Are Revolutionizing Time Series Forecasting
Researchers have introduced TimeGS, a novel AI framework that treats time series forecasting as a 2D rendering problem. By adapting Gaussian splatting techniques from computer graphics, the approach achieves state-of-the-art performance while maintaining temporal continuity.
MIRAGE AI Framework Bridges Critical Gap in Alzheimer's Diagnosis by Synthesizing MRI Insights from Health Records
Researchers have developed MIRAGE, a novel AI framework that uses knowledge graphs to synthesize diagnostic MRI information from electronic health records, potentially revolutionizing Alzheimer's disease assessment in resource-limited settings by bridging the missing-modality gap.