Defense Against Adversarial and Backdoor ML Attacks in Autonomous CPS

Unified defenses for trustworthy perception in autonomous cyber-physical systems

Thrust II: Defense Against Adversarial and Backdoor ML Attacks in Autonomous CPS

As autonomous CPS increasingly rely on high-capacity ML models to interpret multimodal sensor data and guide real-time decisions, these models become prime targets for adversarial and backdoor attacks. Even small, carefully crafted perturbations—often imperceptible to humans—can break perception pipelines, distort situational awareness, and trigger dangerous control actions. These risks are amplified in autonomous vehicles, robots, and UAVs, where the physical consequences of misclassification can be immediate and severe.

My research in this thrust focuses on understanding, modeling, and defending against these ML-targeted manipulations within safety-critical autonomous systems. I study how adversarial and backdoor attacks exploit the internal representations of perception models, how these vulnerabilities manifest across sensing modalities (Shahriar et al., 2026), and how to design defenses that remain effective under real-world constraints such as limited compute, dynamic environments, and online decision-making.

A key component of this research direction is developing unified, modality-agnostic defenses that protect the perception stack against a wide spectrum of ML attacks. Rather than treating adversarial and backdoor attacks as isolated problems, my work examines the shared structural characteristics of malicious perturbations and leverages these insights to build generalizable, efficient, and interpretable detection mechanisms (Shahriar et al., 2025). I also explore how noise geometry, representation alignment, and system-level context can be used to create lightweight defenses suitable for real-time CPS and resource-constrained platforms.

Broadly, this thrust aims to establish a principled foundation for trustworthy perception in autonomous CPS by ensuring that ML models remain reliable, secure, and tamper-resistant even in adversarial environments.

References

2026

  1. SaTML
    Temporal Misalignment Attacks against Multimodal Perception in Autonomous Driving
    Md Hasan Shahriar, Md Mohaimin Al Barat, Harshavardhan Sundar, and 4 more authors
    In IEEE Conference on Secure and Trustworthy Machine Learning (SaTML), 2026
    Accepted

2025

  1. ESORICS
    Let the Noise Speak: Harnessing Noise for a Unified Defense Against Adversarial and Backdoor Attacks
    Md Hasan Shahriar, Ning Wang, Naren Ramakrishnan, and 2 more authors
    In European Symposium on Research in Computer Security, 2025