Arxiv 2024
Qijun Feng, Zhen Xing, Zuxuan Wu, Yu-Gang Jiang
Fudan Univerisity
Abstract
We introduce GeoGS3D, a novel two-stage framework for reconstructing detailed 3D objects from single-view images. Inspired by the success of pre-trained 2D diffusion models, our method incorporates an orthogonal plane decomposition mechanism to extract 3D geometric features from the 2D input, facilitating the generation of multi-view consistent images. During the following Gaussian Splatting, these images are fused with epipolar attention, fully utilizing the geometric correlations across views. Moreover, we propose a novel metric, Gaussian Divergence Significance (GDS), to prune unnecessary operations during optimization, significantly accelerating the reconstruction process. Extensive experiments demonstrate that \methodname~generates images with high consistency across views and reconstructs high-quality 3D objects, both qualitatively and quantitatively.
Novel View Generation and 3D Reconstruction
Qualitative Comparison
Interactive Results
Text-to-image-to-3D
A grey double sofa.
A basket of bread.
A telephone in vintage style.
Citation
@misc{feng2024fdgaussian,
title={FDGaussian: Fast Gaussian Splatting from Single Image via Geometric-aware Diffusion Model},
author={Qijun Feng and Zhen Xing and Zuxuan Wu and Yu-Gang Jiang},
year={2024},
eprint={2403.10242},
archivePrefix={arXiv},
primaryClass={cs.CV}}