Content Moderation
Character.AI Still Hasn’t Fixed Its School Shooter Problem We Identified in 2024
We can't stress enough how easy it is to find this stuff. The post Character.AI Still Hasn’t Fixed Its School Shooter Problem We Identified in 2024 appeared first on Futurism .
The Sora feed philosophy
Discover the Sora feed philosophy—built to spark creativity, foster connections, and keep experiences safe with personalized recommendations, parental controls, and strong guardrails.
Shipping smarter agents with every new model
Discover how SafetyKit leverages OpenAI GPT-5 to enhance content moderation, enforce compliance, and outpace legacy safety systems with greater accuracy .
A Holistic Approach to Undesired Content Detection in the Real World
We present a holistic approach to building a robust and useful natural language classification system for real-world content moderation.
Using GPT-4 for content moderation
We use GPT-4 for content policy development and content moderation decisions, enabling more consistent labeling, a faster feedback loop for policy refinement, and less involvement from human moderators.
New and improved content moderation tooling
We are introducing a new and improved content moderation tool. The Moderation endpoint improves upon our previous content filter, and is available for free today to OpenAI API developers.
