GeoGS3D: Single-view 3D Reconstruction via Geometric-aware Diffusion Model and Gaussian Splatting

Arxiv 2024

Qijun Feng, Zhen Xing, Zuxuan Wu, Yu-Gang Jiang

Fudan Univerisity  

Abstract

We introduce GeoGS3D, a novel two-stage framework for reconstructing detailed 3D objects from single-view images. Inspired by the success of pre-trained 2D diffusion models, our method incorporates an orthogonal plane decomposition mechanism to extract 3D geometric features from the 2D input, facilitating the generation of multi-view consistent images. During the following Gaussian Splatting, these images are fused with epipolar attention, fully utilizing the geometric correlations across views. Moreover, we propose a novel metric, Gaussian Divergence Significance (GDS), to prune unnecessary operations during optimization, significantly accelerating the reconstruction process. Extensive experiments demonstrate that \methodname~generates images with high consistency across views and reconstructs high-quality 3D objects, both qualitatively and quantitatively.

Novel View Generation and 3D Reconstruction

Qualitative Comparison

Interactive Results

panda
windmill
Image 3
Image 1
Image 2
Image 3
Image 1
Image 2
Image 3

Text-to-image-to-3D

A grey double sofa.

Image 1

A basket of bread.

Image 1

A telephone in vintage style.

Image 1

Citation

@misc{feng2024fdgaussian,
          title={FDGaussian: Fast Gaussian Splatting from Single Image via Geometric-aware Diffusion Model}, 
          author={Qijun Feng and Zhen Xing and Zuxuan Wu and Yu-Gang Jiang},
          year={2024},
          eprint={2403.10242},
          archivePrefix={arXiv},
          primaryClass={cs.CV}}