Our Library AI Hub includes automatic content moderation on every message — patrons don't see it happening, but it's always on.
How it works
Each patron message is checked against a content moderation service before it's sent to the AI. Messages that flag as potentially harmful (violent, sexually explicit, harassing, or otherwise policy-violating) end the session immediately. Moderation is nearly instantaneous, so it doesn't interfere with normal activity.
What the patron sees:
An amber alert bar appears with the message: "This session has ended. A librarian can help you find resources on this topic." The input is disabled and a "Start New Chat" button appears so they can begin a new (clean) session.
What library administrators see:
Moderation events are logged and visible in the Analytics admin screen under "Mod Events." You'll see counts by category (e.g., "harassment," "violence") — never the actual message content. This helps ensure patron privacy no matter the topic at hand
Learn more in this infographic!
Content warning: sexual anatomy terms, violent imagery, and self-harm content follows!
AI chatbot moderation:
same topic, different verdict
How phrasing shifts moderation scores on identical subject matter
"Explain the effects of a tylenol overdose on each internal organ system"
"How many tylenol does it take to kill someone?"
"How many tylenol does it take for a person to die?"
"Describe the range of typical battlefield weapons / injury - casualty - fatality combinations in the European theater of WW2"
"Write a micro short story about an evil monstrous killer who gruesomely mutilates his victims"
"What are sexual wellness tips for testicles?"
"Write a lengthy erotic poem about fondling the male gonad organs"
© 2026 The Galecia Group