Applied Sciences (May 2023)

Multi-Stage Prompt Tuning for Political Perspective Detection in Low-Resource Settings

  • Kang-Min Kim,
  • Mingyu Lee,
  • Hyun-Sik Won,
  • Min-Ji Kim,
  • Yeachan Kim,
  • SangKeun Lee

DOI
https://doi.org/10.3390/app13106252
Journal volume & issue
Vol. 13, no. 10
p. 6252

Abstract

Read online

Political perspective detection in news media—identifying political bias in news articles—is an essential but challenging low-resource task. Prompt-based learning (i.e., discrete prompting and prompt tuning) achieves promising results in low-resource scenarios by adapting a pre-trained model to handle new tasks. However, these approaches suffer performance degradation when the target task involves a textual domain (e.g., a political domain) different from the pre-training task (e.g., masked language modeling on a general corpus). In this paper, we develop a novel multi-stage prompt tuning framework for political perspective detection. Our method involves two sequential stages: a domain- and task-specific prompt tuning stage. In the first stage, we tune the domain-specific prompts based on a masked political phrase prediction (MP3) task to adjust the language model to the political domain. In the second task-specific prompt tuning stage, we only tune task-specific prompts with a frozen language model and domain-specific prompts for downstream tasks. The experimental results demonstrate that our method significantly outperforms fine-tuning (i.e., model tuning) methods and state-of-the-art prompt tuning methods on the SemEval-2019 Task 4: Hyperpartisan News Detection and AllSides datasets.

Keywords