Tạp chí Khoa học Đại học Đà Lạt (Aug 2024)

SURVEY AND PROPOSED METHOD TO DETECT ADVERSARIAL EXAMPLES USING AN ADVERSARIAL RETRAINING MODEL

  • Thanh Son Phan,
  • Quang Hua Ta,
  • Duy Trung Pham,
  • Phi Ho Truong

DOI
https://doi.org/10.37569/DalatUniversity.14.3.1150(2024)
Journal volume & issue
Vol. 14, no. 3

Abstract

Read online

Artificial intelligence (AI) has found applications across various sectors and industries, offering numerous advantages to human beings. One prominent area where AI has made significant contributions is in machine learning models. These models have revolutionized various fields, benefiting society in numerous ways, from self-driving cars and intelligent chatbots to automated facial authentication systems. However, in recent years, machine learning models have been the target of various attack methods. One common and dangerous attack method is adversarial attack, where modified input images can cause misclassification or erroneous predictions by the models. To confront that challenge, we present a novel approach called adversarial retraining that uses adversarial examples to train machine learning and deep learning models. This technique aims to enhance the robustness and performance of these models by subjecting them to adversarial scenarios during the training process. In this paper, we survey detection methods and propose a method to detect adversarial examples using YOLOv7, a commonly used intensive research model. By training adversarial retraining and conducting experiments, we show that the proposed method is an effective solution for helping deep learning models detect certain cases of adversarial examples.

Keywords