GRM: Large Gaussian Reconstruction Model for Efficient 3D Reconstruction and Generation
Abstract
We introduce GRM, a large-scale reconstructor capable of recovering a 3D asset from sparse-view images in around 0.1s. GRM is a feed-forward transformer-based model that efficiently incorporates multi-view information to translate the input pixels into pixel-aligned Gaussians, which are unprojected to create a set of densely distributed 3D Gaussians representing a scene. Together, our transformer architecture and the use of 3D Gaussians unlock a scalable and efficient reconstruction framework. Extensive experimental results demonstrate the superiority of our method over alternatives regarding both reconstruction quality and efficiency. We also showcase the potential of GRM in generative tasks, i.e., text-to-3D and image-to-3D, by integrating it with existing multi-view diffusion models. Our project website is at: https://justimyhxu.github.io/projects/grm/.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation (2024)
- CRM: Single Image to 3D Textured Mesh with Convolutional Reconstruction Model (2024)
- VideoMV: Consistent Multi-View Generation Based on Large Video Generative Model (2024)
- SPAD : Spatially Aware Multiview Diffusers (2024)
- Controllable Text-to-3D Generation via Surface-Aligned Gaussian Splatting (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Rapid 3D Scene Generation with GRM: A Game Changer in Graphics!
Links ๐:
๐ Subscribe: https://www.youtube.com/@Arxflix
๐ Twitter: https://x.com/arxflix
๐ LMNT (Partner): https://lmnt.com/
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper