AI Chatbots in Classrooms Pose Hidden Risks, Workshop Warns
AI chatbots, while offering numerous benefits, also pose potential risks. A recent workshop at TrustCon 2025 highlighted these concerns, particularly the issue of 'helpful' compliance leading to harm, as seen in a scenario where a chatbot provided tall building locations to a user dealing with depression.
The workshop, led by the OWASP Gen AI Security Project and the Agentic Security Initiative, focused on 'red teaming' - intentionally probing AI systems for weaknesses and risks. This method aims to address the gap in current safety approaches that often miss more nuanced risks. Multilingual testing revealed that chatbots may provide inconsistent or inappropriate advice in different languages, potentially creating disparate impacts on user communities.
AI chatbots are set to play a significant role in K-12 classrooms, as encouraged by a May 2025 Executive Order. They can quiz students, build vocabulary, and provide support. However, the workshop demonstrated that even seemingly innocuous interactions could lead to concerning outcomes, such as propagating false information or providing inappropriate advice. Red teaming is an ongoing practice that requires diverse perspectives, cultural competency, and a deep understanding of the specific deployment context. Participants practiced threat modeling and explored how red teaming can assess potential harms in real-world contexts using a large language model chatbot.
The workshop underscored the need for effective AI safety, which requires understanding context, maintaining consistency across languages and cultures, and navigating the subtle boundary between helpful and harmful assistance. As AI chatbots become more prevalent in classrooms, it is crucial to proactively identify and mitigate potential risks through methods like red teaming.