论文标题
LCPFormer:通过变压器中的局部环境传播迈向有效的3D点云分析
LCPFormer: Towards Effective 3D Point Cloud Analysis via Local Context Propagation in Transformers
论文作者
论文摘要
具有潜在的注意机制的变压器及其捕获长期依赖性的能力使其成为无序点云数据的自然选择。但是,将本地区域与一般采样体系结构分开损害了实例的结构信息,而相邻的本地区域之间的固有关系缺乏探索,而本地结构信息在基于变压器的3D点云模型中至关重要。因此,在本文中,我们提出了一个名为“局部环境传播”(LCP)的新型模块,以利用邻近本地区域之间传递的信息,并使他们的表示更具信息性和歧视性。更具体地说,我们使用相邻本地区域的重叠点(统计上表明是普遍的),然后将这些共享点的这些共享点的特征重新进行,然后再将它们传递给下一层。在两个变压器层之间插入LCP模块会导致网络表现力的显着提高。最后,我们设计了配备LCP模块的灵活的LCPFormer体系结构。所提出的方法适用于不同的任务,并在基准中胜过基于变压器的各种方法,包括3D形状分类和密集的预测任务,例如3D对象检测和语义分割。代码将发布以进行生殖。
Transformer with its underlying attention mechanism and the ability to capture long-range dependencies makes it become a natural choice for unordered point cloud data. However, separated local regions from the general sampling architecture corrupt the structural information of the instances, and the inherent relationships between adjacent local regions lack exploration, while local structural information is crucial in a transformer-based 3D point cloud model. Therefore, in this paper, we propose a novel module named Local Context Propagation (LCP) to exploit the message passing between neighboring local regions and make their representations more informative and discriminative. More specifically, we use the overlap points of adjacent local regions (which statistically show to be prevalent) as intermediaries, then re-weight the features of these shared points from different local regions before passing them to the next layers. Inserting the LCP module between two transformer layers results in a significant improvement in network expressiveness. Finally, we design a flexible LCPFormer architecture equipped with the LCP module. The proposed method is applicable to different tasks and outperforms various transformer-based methods in benchmarks including 3D shape classification and dense prediction tasks such as 3D object detection and semantic segmentation. Code will be released for reproduction.