y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#3d-reconstruction News & Analysis

18 articles tagged with #3d-reconstruction. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

18 articles
AIBullisharXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

3D-LFM: Lifting Foundation Model

Researchers have developed the first 3D Lifting Foundation Model (3D-LFM) that can reconstruct 3D structures from 2D landmarks without requiring correspondence across training data. The model uses transformer architecture to achieve state-of-the-art performance across various object categories with resilience to occlusions and noise.

AIBullisharXiv โ€“ CS AI ยท Mar 117/10
๐Ÿง 

World2Mind: Cognition Toolkit for Allocentric Spatial Reasoning in Foundation Models

Researchers introduce World2Mind, a training-free spatial intelligence toolkit that enhances foundation models' 3D spatial reasoning capabilities by up to 18%. The system uses 3D reconstruction and cognitive mapping to create structured spatial representations, enabling text-only models to perform complex spatial reasoning tasks.

๐Ÿง  GPT-5
AIBullisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

ZipMap: Linear-Time Stateful 3D Reconstruction with Test-Time Training

Researchers introduce ZipMap, a new AI model for 3D reconstruction that achieves linear-time processing while maintaining accuracy comparable to slower quadratic-time methods. The system can reconstruct over 700 frames in under 10 seconds on a single H100 GPU, making it more than 20x faster than current state-of-the-art approaches like VGGT.

AIBullisharXiv โ€“ CS AI ยท Apr 66/10
๐Ÿง 

NavCrafter: Exploring 3D Scenes from a Single Image

NavCrafter is a new AI framework that creates flexible 3D scenes from a single image by generating novel-view video sequences with controllable camera movement. The system uses video diffusion models and enhanced 3D Gaussian Splatting to achieve superior 3D reconstruction and novel-view synthesis under large viewpoint changes.

AINeutralarXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

EgoGrasp: World-Space Hand-Object Interaction Estimation from Egocentric Videos

EgoGrasp introduces the first method to reconstruct world-space hand-object interactions from egocentric videos using open-vocabulary objects. The multi-stage framework combines vision foundation models with body-guided diffusion models to achieve state-of-the-art performance in 3D scene reconstruction and hand pose estimation.

AIBullisharXiv โ€“ CS AI ยท Mar 37/106
๐Ÿง 

M-Gaussian: An Magnetic Gaussian Framework for Efficient Multi-Stack MRI Reconstruction

Researchers developed M-Gaussian, a new AI framework that adapts 3D Gaussian Splatting for efficient multi-stack MRI reconstruction. The method achieves 40.31 dB PSNR while being 14 times faster than existing implicit neural representation methods, offering improved balance between quality and computational efficiency.

AIBullisharXiv โ€“ CS AI ยท Mar 36/107
๐Ÿง 

ArtiFixer: Enhancing and Extending 3D Reconstruction with Auto-Regressive Diffusion Models

Researchers propose ArtiFixer, a two-stage pipeline using auto-regressive diffusion models to enhance 3D reconstruction quality. The method addresses scalability and quality issues in existing approaches by training a bidirectional generative model with opacity mixing, then distilling it into a causal auto-regressive model that generates hundreds of frames in a single pass.

AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

MAP-Diff: Multi-Anchor Guided Diffusion for Progressive 3D Whole-Body Low-Dose PET Denoising

Researchers developed MAP-Diff, a multi-anchor guided diffusion framework that improves 3D whole-body PET scan denoising by using intermediate-dose scans as trajectory anchors. The method achieves significant improvements in image quality metrics, increasing PSNR from 42.48 dB to 43.71 dB while reducing radiation exposure for patients.

AIBullisharXiv โ€“ CS AI ยท Mar 26/1017
๐Ÿง 

LiteReality: Graphics-Ready 3D Scene Reconstruction from RGB-D Scans

Researchers have developed LiteReality, a novel pipeline that converts RGB-D scans of indoor environments into compact, realistic 3D virtual replicas suitable for AR/VR, gaming, robotics, and digital twins. The system features scene understanding, object retrieval, material painting, and physics integration to create graphics-ready environments that support object individuality and physically-based rendering.

AIBullisharXiv โ€“ CS AI ยท Feb 276/107
๐Ÿง 

AeroDGS: Physically Consistent Dynamic Gaussian Splatting for Single-Sequence Aerial 4D Reconstruction

Researchers have developed AeroDGS, a physics-guided 4D Gaussian splatting framework that enables accurate dynamic scene reconstruction from single-view aerial UAV footage. The system addresses key challenges in monocular aerial reconstruction by incorporating physics-based optimization and geometric constraints to resolve depth ambiguity and improve motion estimation.

AIBullisharXiv โ€“ CS AI ยท Feb 276/108
๐Ÿง 

Latent Gaussian Splatting for 4D Panoptic Occupancy Tracking

Researchers have developed LaGS (Latent Gaussian Splatting), a new AI method for 4D panoptic occupancy tracking that enables robots to better understand dynamic environments. The approach combines camera-based tracking with 3D occupancy prediction, achieving state-of-the-art performance on industry-standard datasets.

$UNI
AINeutralarXiv โ€“ CS AI ยท Apr 74/10
๐Ÿง 

TreeGaussian: Tree-Guided Cascaded Contrastive Learning for Hierarchical Consistent 3D Gaussian Scene Segmentation and Understanding

TreeGaussian introduces a new framework for 3D scene understanding that uses tree-guided cascaded contrastive learning to better capture hierarchical semantic relationships in complex 3D environments. The method addresses limitations in existing 3D Gaussian Splatting approaches by implementing structured learning across object-part hierarchies and improving segmentation consistency.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

Field imaging framework for morphological characterization of aggregates with computer vision: Algorithms and applications

Researchers developed a comprehensive field imaging framework using computer vision and AI to automatically characterize construction aggregates like sand, gravel, and stone. The system uses 2D image analysis and 3D point cloud reconstruction with machine learning to replace manual inspection methods in construction material assessment.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

Improving Multi-View Reconstruction via Texture-Guided Gaussian-Mesh Joint Optimization

Researchers propose a novel framework for 3D object reconstruction from multi-view images that simultaneously optimizes mesh geometry and appearance through Gaussian-guided rendering. The unified approach addresses limitations of existing methods that separate geometry and appearance optimization, enabling better downstream editing tasks like relighting and shape deformation.

AINeutralarXiv โ€“ CS AI ยท Mar 34/103
๐Ÿง 

CloDS: Visual-Only Unsupervised Cloth Dynamics Learning in Unknown Conditions

Researchers introduce CloDS (Cloth Dynamics Splatting), an unsupervised AI framework that learns cloth dynamics from visual observations without requiring known physical properties. The system uses a three-stage pipeline with dual-position opacity modulation to handle complex cloth deformations and self-occlusions through mesh-based Gaussian splatting.

AIBullisharXiv โ€“ CS AI ยท Mar 34/105
๐Ÿง 

PPC-MT: Parallel Point Cloud Completion with Mamba-Transformer Hybrid Architecture

Researchers propose PPC-MT, a hybrid Mamba-Transformer architecture for point cloud completion that uses parallel processing guided by Principal Component Analysis. The framework outperforms existing methods on benchmark datasets while maintaining computational efficiency by combining Mamba's linear complexity with Transformer's fine-grained modeling capabilities.

AINeutralarXiv โ€“ CS AI ยท Mar 24/106
๐Ÿง 

Uncertainty Matters in Dynamic Gaussian Splatting for Monocular 4D Reconstruction

Researchers introduce USplat4D, a new uncertainty-aware dynamic Gaussian Splatting framework that improves 3D scene reconstruction from monocular video by modeling per-Gaussian uncertainty. The approach addresses motion drift and poor synthesis quality by treating well-observed Gaussians as reliable anchors while handling poorly observed ones as less reliable.