With the rapid development of artificial intelligence (AI) technology, more and more companies are applying it to various aspects of daily life. However, the security and ethical issues of this technology have also attracted widespread attention. Recently, researchers from Intel, Boise State University, and the University of Illinois found that chatbots may be misled when facing large amounts of information, thereby violating their safety rules.
This study proposes a new type of attack method called "information overload." The researchers developed an automated system called "InfoFlood," which uses this method to trick large language models (LLMs) such as ChatGPT and Gemini into saying things they should not. According to the research, traditional protective measures mainly rely on identifying specific keywords, but when faced with an information flood, the model may become confused and bypass these protections.
Image source note: The image is AI-generated, and the image licensing service provider is Midjourney
The research team explained that when chatbots receive excessive information, they may misinterpret the user's intent, leading to incorrect answers. This method creates a standardized prompt template to overwhelm the AI model with more information, thus causing interference and confusion. For example, when the model refuses to answer a question, InfoFlood adds false citations or irrelevant ethical statements in the prompt, putting the model in a dilemma.
More worrying is that this discovery reveals that even with safety filters in place, malicious users can still manipulate the model through information overload to implant harmful content. The researchers plan to send a report of this finding to companies that have large AI models, reminding them to strengthen their security measures.
As technology continues to advance, how to reasonably use AI while protecting users remains a pressing challenge. Researchers hope that this study will spark more discussions about AI safety and ethics and encourage technology companies to take more effective protective measures.