Apple Unveils AI Model That Builds 3D Scenes Using Images

In a significant leap for 3D computer vision, Apple has introduced a new AI system called Matrix3D that can reconstruct rich, photorealistic 3D objects and scenes from just a handful of 2D images. The model, developed in collaboration with researchers from Nanjing University and The Hong Kong University of Science and Technology, offers a unified and efficient alternative to traditional photogrammetry methods, which have long relied on large image sets and a fragmented processing pipeline.
Rethinking Photogrammetry for the AI Era
Photogrammetry, a foundational technique in 3D reconstruction, works by using multiple 2D photos—often in the hundreds—to recreate physical environments digitally. The conventional approach involves multiple stages, each handled by separate algorithms: determining camera orientation, predicting depth information, and finally generating the 3D structure. This process, while robust, is both time- and resource-intensive.
Matrix3D turns that model on its head. Rather than requiring extensive photographic input and a patchwork of specialized components, the new system compresses the entire photogrammetry pipeline into a single, end-to-end model. The result is a streamlined architecture capable of working with far fewer images while delivering comparable or even superior results.
Also read: Apple Expands AI Features with iOS 18.4 and Vision Pro Support
A Single Model, Multiple Functions
Central to Matrix3D’s capabilities is its foundation in diffusion transformers, a generative AI framework similar to those powering large-scale language and image models. The system is trained using a masked learning strategy, where key pieces of data are deliberately obscured during training, forcing the model to infer missing information. This method enables Matrix3D to generalize well even when given limited visual input, effectively reconstructing depth, estimating camera positions, and generating new perspectives from just two or three images.
Such versatility represents a meaningful step forward in how AI can simplify and enhance complex visual workflows. With its compact requirements and broad functionality, Matrix3D offers new possibilities for developers, content creators, and researchers working with 3D data.
Open Access With Future Applications in Sight
Apple’s researchers have shared the full details of Matrix3D in a paper published on arXiv and made the source code available via GitHub. The release also includes interactive demonstrations, allowing users to explore the model’s reconstructions firsthand.
Although Apple has not officially disclosed where Matrix3D might appear in its product ecosystem, its potential for integration is clear. Devices like the Vision Pro headset, which aim to bridge the gap between digital and physical environments, could benefit significantly from the model’s ability to turn ordinary 2D photos into immersive 3D scenes. In that context, Matrix3D could lay the groundwork for a new generation of spatial computing experiences built on minimal input and maximum fidelity.