Adversarial Machine Learning Attack on Modulation Classification

26 Sep 2019  ·  Usama Muhammad, Asim Muhammad, Qadir Junaid, Al-Fuqaha Ala, Imran Muhammad Ali ·

Modulation classification is an important component of cognitive self-driving networks. Recently many ML-based modulation classification methods have been proposed. We have evaluated the robustness of 9 ML-based modulation classifiers against the powerful Carlini \& Wagner (C-W) attack and showed that the current ML-based modulation classifiers do not provide any deterrence against adversarial ML examples. To the best of our knowledge, we are the first to report the results of the application of the C-W attack for creating adversarial examples against various ML models for modulation classification.

PDF Abstract
No code implementations yet. Submit your code now

Categories


Cryptography and Security Networking and Internet Architecture

Datasets


  Add Datasets introduced or used in this paper