IPoD: Implicit Field Learning with Point Diffusion for Generalizable 3D Object Reconstruction from Single RGB-D Images
arxiv(2024)
摘要
Generalizable 3D object reconstruction from single-view RGB-D images remains
a challenging task, particularly with real-world data. Current state-of-the-art
methods develop Transformer-based implicit field learning, necessitating an
intensive learning paradigm that requires dense query-supervision uniformly
sampled throughout the entire space. We propose a novel approach, IPoD, which
harmonizes implicit field learning with point diffusion. This approach treats
the query points for implicit field learning as a noisy point cloud for
iterative denoising, allowing for their dynamic adaptation to the target object
shape. Such adaptive query points harness diffusion learning's capability for
coarse shape recovery and also enhances the implicit representation's ability
to delineate finer details. Besides, an additional self-conditioning mechanism
is designed to use implicit predictions as the guidance of diffusion learning,
leading to a cooperative system. Experiments conducted on the CO3D-v2 dataset
affirm the superiority of IPoD, achieving 7.8
in Chamfer distance over existing methods. The generalizability of IPoD is also
demonstrated on the MVImgNet dataset. Our project page is at
https://yushuang-wu.github.io/IPoD.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要