Journal of Computing and Information Technology (Jan 2024)

A Brief Survey on Safety of Large Language Models

  • Zhengjie Gao,
  • Xuanzi Liu,
  • Yuanshuai Lan,
  • Zheng Yang

DOI
https://doi.org/10.20532/cit.2024.1005778
Journal volume & issue
Vol. 32, no. 1
pp. 47 – 64

Abstract

Read online

Large Language Models (LLMs) have revolutionized Natural Language Processing (NLP) and have been widely adopted in various applications such as machine translation, chatbots, text summarization, and so on. However, the use of LLMs has raised concerns about their potential safety and security risks. In this survey, we explore the safety implications of LLMs, including ethical considerations, hallucination, and prompt injection. We also discuss current research efforts to mitigate these risks and identify areas for future research. Our survey provides a comprehensive overview of the safety concerns related to LLMs, which can help researchers and practitioners in the NLP community develop more safe and ethical applications of LLMs.

Keywords