Large Language Model Firewall for AIGC Protection with Intelligent Detection Policy
Author
Abstract

LLMs face content security risks such as prompt information injection, insecure output processing, sensitive information leakage, and over-dependence, etc. By constructing a firewall for LLMs with intelligent detection strategies and introducing multi-engine detection capabilities such as rule matching, semantic computing, and AI models, we can intelligently detect and dispose of inputs and outputs of the LLMs, and realize the full-time on-line security protection of LLM applications. The system is tested on open-source LLMs, and there is a significant improvement in terms of the detection rate of insecure content.

Year of Publication
2024
Date Published
apr
URL
https://ieeexplore.ieee.org/document/10625752
DOI
10.1109/MICCIS63508.2024.00047
Google Scholar | BibTeX | DOI