TorchSparse: Efficient Point Cloud Inference Engine
Speedup
Benchmark (surveying)
Convolution (computer science)
FLOPS
DOI:
10.48550/arxiv.2204.10319
Publication Date:
2022-01-01
AUTHORS (5)
ABSTRACT
Deep learning on point clouds has received increased attention thanks to its wide applications in AR/VR and autonomous driving. These require low latency high accuracy provide real-time user experience ensure safety. Unlike conventional dense workloads, the sparse irregular nature of poses severe challenges running CNNs efficiently general-purpose hardware. Furthermore, existing acceleration techniques for 2D images do not translate 3D clouds. In this paper, we introduce TorchSparse, a high-performance cloud inference engine that accelerates convolution computation GPUs. TorchSparse directly optimizes two bottlenecks convolution: data movement. It applies adaptive matrix multiplication grouping trade better regularity, achieving 1.4-1.5x speedup multiplication. also movement by adopting vectorized, quantized fused locality-aware memory access, reducing cost 2.7x. Evaluated seven representative models across three benchmark datasets, achieves 1.6x 1.5x measured end-to-end over state-of-the-art MinkowskiEngine SpConv, respectively.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....