@inproceedings{9730fac280a44930a3eeda67d98f74ff,
title = "BEVContrast: Self-Supervision in BEV Space for Automotive Lidar Point Clouds",
abstract = "We present a surprisingly simple and efficient method for self-supervision of 3D backbone on automotive Lidar point clouds. We design a contrastive loss between features of Lidar scans captured in the same scene. Several such approaches have been proposed in the literature from PointConstrast [40], which uses a contrast at the level of points, to the state-of-the-art TARL [30], which uses a contrast at the level of segments, roughly corresponding to objects. While the former enjoys a great simplicity of implementation, it is surpassed by the latter, which however requires a costly pre-processing. In BEVContrast, we define our contrast at the level of 2D cells in the Bird's Eye View plane. Resulting cell-level representations offer a good trade-off between the point-level representations exploited in PointContrast and segment-level representations exploited in TARL: we retain the simplicity of PointContrast (cell representations are cheap to compute) while surpassing the performance of TARL in downstream semantic segmentation. The code is available at github.com/valeoai/BEVContrast",
keywords = "Lidar, Self-supervision, automotive, object detection, semantic segmentation, unsupervised",
author = "Corentin Sautier and Gilles Puy and Alexandre Boulch and Renaud Marlet and Vincent Lepetit",
note = "Publisher Copyright: {\textcopyright} 2024 IEEE.; 11th International Conference on 3D Vision, 3DV 2024 ; Conference date: 18-03-2024 Through 21-03-2024",
year = "2024",
month = jan,
day = "1",
doi = "10.1109/3DV62453.2024.00017",
language = "English",
series = "Proceedings - 2024 International Conference on 3D Vision, 3DV 2024",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "559--568",
booktitle = "Proceedings - 2024 International Conference on 3D Vision, 3DV 2024",
}