Lack of Robustness of LIDAR-Based Deep Learning Systems to Small Adversarial Perturbations

Konferenz: ISR 2018 - 50th International Symposium on Robotics
20.06.2018 - 21.06.2016 in München, Germany

Tagungsband: ISR 2018

Seiten: 7Sprache: EnglischTyp: PDF

Persönliche VDE-Mitglieder erhalten auf diesen Artikel 10% Rabatt

Patel, Naman; Liu, Kang; Krishnamurthy, Prashanth; Garg, Siddharth; Khorrami, Farshad (ntrol/Robotics Research Laboratory (CRRL), Department of Electrical & Computer Engineering, NYU Tandon School of Engineering, Brooklyn, NY 11201, USA)

In this paper, we investigate the robustness of LIDAR-based autonomous navigation for unmanned vehicles using Deep Neural Networks (DNN) to adversarial perturbations. A well-trained network robust to sensor noise can yield an undesirable network response (e.g., steering the vehicle in a wrong direction) by maliciously crafted perturbations in sensor data. We show through experimental evaluations on our unmanned ground vehicle (UGV) that small perturbations in some of the LIDAR sensor data (even perturbations smaller than the sensor accuracy) can lead the DNN to generate incorrect outputs. This is somewhat unexpected from a sensor such as LIDAR, which provides very well-defined structural/geometrical information about the environment.