IEEE Access (Jan 2025)

Imitation Game for Adversarial Disillusion With Chain-of-Thought Reasoning in Generative AI

  • Ching-Chun Chang,
  • Fan-Yun Chen,
  • Shih-Hong Gu,
  • Kai Gao,
  • Hanrui Wang,
  • Isao Echizen

DOI
https://doi.org/10.1109/access.2025.3574016
Journal volume & issue
Vol. 13
pp. 95085 – 95093

Abstract

Read online

As the cornerstone of artificial intelligence, machine perception confronts a fundamental threat posed by adversarial illusions. These adversarial attacks manifest in two primary forms: deductive illusion, where specific stimuli are crafted based on the victim model’s general decision logic, and inductive illusion, where the victim model’s general decision logic is shaped by specific stimuli. The former exploits the model’s decision boundaries to create a stimulus that, when applied, interferes with its decision-making process. The latter reinforces a conditioned reflex in the model, embedding a backdoor during its learning phase that, when triggered by a stimulus, causes aberrant behaviors. The multifaceted nature of adversarial illusions calls for a unified defence framework, addressing vulnerabilities across various forms of attack. In this study, we propose a disillusion paradigm based on the concept of an imitation game. At the heart of the imitation game lies a multimodal generative agent, steered by chain-of-thought reasoning, which observes, internalizes and reconstructs the semantic essence of a sample, liberated from the classic pursuit of reversing the sample to its original state. As a proof of concept, we conduct experimental simulations using a multimodal generative dialogue agent and evaluates the methodology under a variety of attack scenarios. Experimental results demonstrate that the proposed framework consistently neutralizes both deductive and inductive adversarial illusions across diverse white-box and black-box attack scenarios.

Keywords