Fast Point Transformer

12/09/2021
by   Chunghyun Park, et al.
0

The recent success of neural networks enables a better interpretation of 3D point clouds, but processing a large-scale 3D scene remains a challenging problem. Most current approaches divide a large-scale scene into small regions and combine the local predictions together. However, this scheme inevitably involves additional stages for pre- and post-processing and may also degrade the final output due to predictions in a local perspective. This paper introduces Fast Point Transformer that consists of a new lightweight self-attention layer. Our approach encodes continuous 3D coordinates, and the voxel hashing-based architecture boosts computational efficiency. The proposed method is demonstrated with 3D semantic segmentation and 3D detection. The accuracy of our approach is competitive to the best voxel-based method, and our network achieves 136 times faster inference time than the state-of-the-art, Point Transformer, with a reasonable accuracy trade-off.

READ FULL TEXT

page 1

page 14

page 15

research
03/19/2022

Voxel Set Transformer: A Set-to-Set Approach to 3D Object Detection from Point Clouds

Transformer has demonstrated promising performance in many 2D vision tas...
research
09/06/2021

Voxel Transformer for 3D Object Detection

We present Voxel Transformer (VoTr), a novel and effective voxel-based T...
research
05/23/2023

Hierarchical Adaptive Voxel-guided Sampling for Real-time Applications in Large-scale Point Clouds

While point-based neural architectures have demonstrated their efficacy,...
research
11/30/2017

3DContextNet: K-d Tree Guided Hierarchical Learning of Point Clouds Using Local Contextual Cues

3D data such as point clouds and meshes are becoming more and more avail...
research
01/15/2023

DSVT: Dynamic Sparse Voxel Transformer with Rotated Sets

Designing an efficient yet deployment-friendly 3D backbone to handle spa...
research
07/18/2021

Dynamic Convolution for 3D Point Cloud Instance Segmentation

We propose an approach to instance segmentation from 3D point clouds bas...
research
06/18/2020

SE(3)-Transformers: 3D Roto-Translation Equivariant Attention Networks

We introduce the SE(3)-Transformer, a variant of the self-attention modu...

Please sign up or login with your details

Forgot password? Click here to reset