Common areas to target with red teaming
Consider these categories for your red team RAG attack strategy:
- Bias and stereotypes: The chatbot may be manipulated to give biased answers, which can harm the company’s reputation if shared on social media.
- Sensitive information disclosure: Competitors or cybercriminals may attempt to obtain sensitive information, such as prompts or private data, through the chatbot.
- Service disruption: Ill-intentioned individuals may send long or crafted requests to disrupt the chatbot’s availability for legitimate users.
- Hallucinations: The chatbot may provide incorrect information due to suboptimal retrieval mechanisms, low-quality documents, or the LLM’s tendency to agree with the user.
Techniques you can employ to make these attacks include the following:
- Bypassing safeguards:
- Text completion: Red teaming techniques for bypassing safeguards in LLM applications include exploiting text completion by...