TraIL-Det: Transformation-Invariant Local Feature Networks for 3D LiDAR Object Detection with Unsupervised Pre-Training

Li Li, Tanqiu Qiao, Hubert P. H. Shum and Toby P. Breckon
Proceedings of the 2024 British Machine Vision Conference (BMVC), 2024

 H5-Index: 65#

TraIL-Det: Transformation-Invariant Local Feature Networks for 3D LiDAR Object Detection with Unsupervised Pre-Training
# According to Google Scholar 2024

Abstract

3D point clouds are essential for perceiving outdoor scenes, especially within the realm of autonomous driving. Recent advances in 3D LiDAR Object Detection focus primarily on the spatial positioning and distribution of points to ensure accurate detection. However, despite their robust performance in variable conditions, these methods are hindered by their sole reliance on coordinates and point intensity, resulting in inadequate isometric invariance and suboptimal detection outcomes. To tackle this challenge, our work introduces Transformation-Invariant Local (TraIL) features and the associated TraIL-Det architecture. Our TraIL features exhibit rigid transformation invariance and effectively adapt to variations in point density, with a design focus on capturing the localized geometry of neighboring structures. They utilize the inherent isotropic radiation of LiDAR to enhance local representation, improve computational efficiency, and boost detection performance. To effectively process the geometric relations among points within each proposal, we propose a Multi-head self-Attention Encoder (MAE) with asymmetric geometric features to encode high-dimensional TraIL features into manageable representations. Our method outperforms contemporary self-supervised 3D object detection approaches in terms of mAP on KITTI (67.8, 20% label, moderate) and Waymo (68.9, 20% label, moderate) datasets under various label ratios (20%, 50%, and 100%).


Downloads


YouTube


Cite This Research

Plain Text

Li Li, Tanqiu Qiao, Hubert P. H. Shum and Toby P. Breckon, "TraIL-Det: Transformation-Invariant Local Feature Networks for 3D LiDAR Object Detection with Unsupervised Pre-Training," in BMVC '24: Proceedings of the 2024 British Machine Vision Conference, Glasgow, UK, 2024.

BibTeX

@inproceedings{li24traildet,
 author={Li, Li and Qiao, Tanqiu and Shum, Hubert P. H. and Breckon, Toby P.},
 booktitle={Proceedings of the 2024 British Machine Vision Conference},
 series={BMVC '24},
 title={TraIL-Det: Transformation-Invariant Local Feature Networks for 3D LiDAR Object Detection with Unsupervised Pre-Training},
 year={2024},
 location={Glasgow, UK},
}

RIS

TY  - CONF
AU  - Li, Li
AU  - Qiao, Tanqiu
AU  - Shum, Hubert P. H.
AU  - Breckon, Toby P.
T2  - Proceedings of the 2024 British Machine Vision Conference
TI  - TraIL-Det: Transformation-Invariant Local Feature Networks for 3D LiDAR Object Detection with Unsupervised Pre-Training
PY  - 2024
ER  - 


Supporting Grants


Similar Research

Li Li, Hubert P. H. Shum and Toby P. Breckon, "RAPiD-Seg: Range-Aware Pointwise Distance Distribution Networks for 3D LiDAR Segmentation", Proceedings of the 2024 European Conference on Computer Vision (ECCV), 2024
Li Li, Hubert P. H. Shum and Toby P. Breckon, "Less is More: Reducing Task and Model Complexity for 3D Point Cloud Semantic Segmentation", Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023
Jiaxu Liu, Zhengdi Yu, Toby P. Breckon and Hubert P. H. Shum, "U3DS3: Unsupervised 3D Semantic Scene Segmentation", Proceedings of the 2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2024
Li Li, Khalid N. Ismail, Hubert P. H. Shum and Toby P. Breckon, "DurLAR: A High-fidelity 128-Channel LiDAR Dataset with Panoramic Ambientand Reflectivity Imagery for Multi-Modal Autonomous Driving Applications", Proceedings of the 2021 International Conference on 3D Vision (3DV), 2021

 

Last updated on 7 September 2024
RSS Feed