Author(s)

Tarjanee Vyas, Prof. Dhaval R. Chandarana

  • Manuscript ID: 140087
  • Volume: 2
  • Issue: 1
  • Pages: 228–250

Subject Area: Computer Science

Abstract

Adversarial Machine Learning (AML) aims to enhance the security and robustness of machine learning models deployed in critical domains such as healthcare, finance, and autonomous technologies. Despite their impressive performance, these models are susceptible to adversarial interventions—carefully crafted inputs designed to manipulate model outputs. Key attack vectors include evasion attacks, data poisoning, and model extraction, each targeting different phases of the machine learning workflow. Numerous defense mechanisms, including adversarial training, preprocessing techniques, and robust optimization methods, have been proposed; however, no single approach provides complete protection, as attackers continue to develop increasingly sophisticated strategies. This review systematically examines prominent adversarial attack methods, analyzes existing defensive techniques, and outlines future research directions such as achieving provable robustness, creating adaptive defense frameworks, enhancing model interpretability, and addressing ethical considerations to ensure trustworthy and secure AI systems.

Keywords
Adversarial Machine LearningAdversarial AttacksDefense MechanismsRobust Machine LearningAI SecurityAdversarial TrainingModel Robustness.