TY - JOUR
T1 - 3D Adversarial Augmentations for Robust Out-of-Domain Predictions
AU - Lehner, Alexander
AU - Gasperini, Stefano
AU - Marcos-Ramiro, Alvaro
AU - Schmidt, Michael
AU - Navab, Nassir
AU - Busam, Benjamin
AU - Tombari, Federico
N1 - Publisher Copyright:
© The Author(s) 2023.
PY - 2024/3
Y1 - 2024/3
N2 - Since real-world training datasets cannot properly sample the long tail of the underlying data distribution, corner cases and rare out-of-domain samples can severely hinder the performance of state-of-the-art models. This problem becomes even more severe for dense tasks, such as 3D semantic segmentation, where points of non-standard objects can be confidently associated to the wrong class. In this work, we focus on improving the generalization to out-of-domain data. We achieve this by augmenting the training set with adversarial examples. First, we learn a set of vectors that deform the objects in an adversarial fashion. To prevent the adversarial examples from being too far from the existing data distribution, we preserve their plausibility through a series of constraints, ensuring sensor-awareness and shapes smoothness. Then, we perform adversarial augmentation by applying the learned sample-independent vectors to the available objects when training a model. We conduct extensive experiments across a variety of scenarios on data from KITTI, Waymo, and CrashD for 3D object detection, and on data from SemanticKITTI, Waymo, and nuScenes for 3D semantic segmentation. Despite training on a standard single dataset, our approach substantially improves the robustness and generalization of both 3D object detection and 3D semantic segmentation methods to out-of-domain data.
AB - Since real-world training datasets cannot properly sample the long tail of the underlying data distribution, corner cases and rare out-of-domain samples can severely hinder the performance of state-of-the-art models. This problem becomes even more severe for dense tasks, such as 3D semantic segmentation, where points of non-standard objects can be confidently associated to the wrong class. In this work, we focus on improving the generalization to out-of-domain data. We achieve this by augmenting the training set with adversarial examples. First, we learn a set of vectors that deform the objects in an adversarial fashion. To prevent the adversarial examples from being too far from the existing data distribution, we preserve their plausibility through a series of constraints, ensuring sensor-awareness and shapes smoothness. Then, we perform adversarial augmentation by applying the learned sample-independent vectors to the available objects when training a model. We conduct extensive experiments across a variety of scenarios on data from KITTI, Waymo, and CrashD for 3D object detection, and on data from SemanticKITTI, Waymo, and nuScenes for 3D semantic segmentation. Despite training on a standard single dataset, our approach substantially improves the robustness and generalization of both 3D object detection and 3D semantic segmentation methods to out-of-domain data.
KW - 3D point cloud
KW - Adversarial augmentation
KW - Domain generalization
KW - Out-of-domain
KW - Robustness
UR - http://www.scopus.com/inward/record.url?scp=85174395896&partnerID=8YFLogxK
U2 - 10.1007/s11263-023-01914-7
DO - 10.1007/s11263-023-01914-7
M3 - Article
AN - SCOPUS:85174395896
SN - 0920-5691
VL - 132
SP - 931
EP - 963
JO - International Journal of Computer Vision
JF - International Journal of Computer Vision
IS - 3
ER -