作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2026, Vol. 52 ›› Issue (5): 239-249. doi: 10.19678/j.issn.1000-3428.0070753

• 计算机视觉与图形图像处理 • 上一篇    下一篇

基于旋转不变区域一致性的双视图点云重建方法

高雨菲, 贾鑫*(), 黄张驰, 许志男, 霍鹏飞, 陆芷茵   

  1. 天津理工大学工程训练中心, 天津 300384
  • 收稿日期:2024-12-26 修回日期:2025-03-09 出版日期:2026-05-15 发布日期:2025-04-21
  • 通讯作者: 贾鑫
  • 作者简介:

    高雨菲, 女, 学士, 主研方向为计算机视觉、三维重建、点云处理

    贾鑫(CCF会员、通信作者), 讲师、博士、硕士生导师

    黄张弛, 学士

    许志男, 学士

    霍鹏飞, 学士

    陆芷茵, 学士

  • 基金资助:
    国家自然科学基金青年基金(62302335); 天津市大学生创新创业项目(202310060026)

Dual-View Point Cloud Reconstruction Method Based on Rotation-Invariant Regional Consistency

GAO Yufei, JIA Xin*(), HUANG Zhangchi, XU Zhinan, HUO Pengfei, LU Zhiyin   

  1. Engineering Training Center, Tianjin University of Technology, Tianjin 300384, China
  • Received:2024-12-26 Revised:2025-03-09 Online:2026-05-15 Published:2025-04-21
  • Contact: JIA Xin

摘要:

多视图三维重建旨在通过多张二维图像恢复给定对象的三维形状。然而, 现有方法忽略了学习对象的旋转不变性以及区域一致性, 难以准确聚合多视图特征, 造成重建结果细节丢失。为了解决该问题, 提出一种基于旋转不变区域一致性的双视图点云重建方法(DPR2)。DPR2以两张RGB图像作为输入, 在探索对象区域旋转不变性的基础上, 学习跨视图对象的区域一致性, 促进多视图特征聚合, 并重建给定对象的精细点云。在编码阶段, 首先引入点云初始化网络, 为每个视图初始化一个粗糙点云; 其次, 提出区域级旋转不变特征提取网络, 通过计算两点之间的欧氏距离来捕捉粗糙点云不同区域的旋转不变特征。在解码阶段, 设计双阶段交叉注意力机制, 它可以构建跨视图点云的高质量区域一致性, 从而准确实现多视图特征聚合。另外, 设计一种点云细化网络, 利用被聚合的特征, 将粗糙点云细化为具有细粒度细节和光滑表面的点云。在ShapeNet和Pix3D数据集上的大量实验结果表明, DPR2的重建性能优于现有先进方法, 与最新方法P2M++、MVP2M++相比, DPR2的倒角距离(CD)分别改善了23.62%和9.06%。

关键词: 多视图三维重建, 点云, 区域一致性, 旋转不变性, 特征聚合

Abstract:

Multi-view 3D reconstruction aims to recover the 3D shape of a given object from multiple 2D images. However, existing methods neglect to learn the rotational invariance and regional consistency of objects, making it difficult to accurately aggregate multi-view features, resulting in the loss of reconstruction details. To address this issue, this study first proposes Dual-view Point cloud reconstruction based on Rotation-invariant Regional consistency (DPR2). DPR2 takes two RGB images as input, explores the rotational invariance of object regions, learns the regional consistency of objects across views, promotes the aggregation of multi-view features, and reconstructs the fine point cloud of the given object. In the encoding stage, a point-cloud initialization network is first introduced to initialize a coarse point cloud for each view. The study also proposes a region-level rotational invariant feature extraction network that captures the rotational invariant features of different regions of the coarse point cloud by calculating the Euclidean distance between two points. In the decoding stage, a two-stage cross-attention mechanism is designed to construct high-quality regional consistency of cross-view point clouds, thereby accurately achieving multi-view feature aggregation. Additionally, a point-cloud refinement network is designed that utilizes aggregated features to refine the coarse point cloud into one with fine-grained details and smooth surfaces. Extensive experimental results on the ShapeNet and Pix3D datasets show that DPR2 outperforms existing state-of-the-art methods. Compared with the latest methods, P2M++ and MVP2M++, DPR2 improves the Chamfer Distance (CD) by 23.62% and 9.06%, respectively.

Key words: multi-view 3D reconstruction, point cloud, regional consistency, rotation invariance, feature aggregation