← Back to OpenAI updates ← Terug naar OpenAI-updates
OpenAI ARTICLE ARTIKEL 10 August 2022 10 augustus 2022

New and improved content moderation tooling New and improved content moderation tooling

Read documentation(opens in a new window) Read documentation(opens in a new window)

Article details Artikelgegevens
AI maker AI-maker OpenAI Type Type Article Artikel Published Gepubliceerd 10 August 2022 10 augustus 2022 Updates Updates Videos Video's View original article Bekijk origineel artikel
Why it matters Waarom dit telt

Quick editorial signal Snelle redactionele duiding

2 min
Impact Impact

Relevant if you build with AI tools, APIs, or coding agents. Relevant als je bouwt met AI-tools, API's of coding agents.

Audience Voor wie Developers Developers
Level Niveau Expert Expert
  • Track this as a OpenAI update, not just a standalone headline. Bekijk dit als OpenAI-update, niet alleen als losse headline.
  • Useful for builders who need to understand API, coding, or workflow changes. Nuttig voor bouwers die API-, code- of workflowwijzigingen willen begrijpen.
  • Likely worth revisiting after people have used the release in practice. Waarschijnlijk de moeite waard om opnieuw te bekijken zodra mensen het in praktijk gebruiken.
model apps developers safety

Listen to article

To help developers protect their applications against possible misuse, we are introducing the faster and more accurateModeration endpoint⁠(opens in a new window). This endpoint provides OpenAI API developers with free access toGPT‑based⁠classifiers that detect undesired content—an instance ofusing AI systems⁠to assist with human supervision of these systems. We have also released both atechnical paper⁠(opens in a new window)describing our methodology and thedataset⁠(opens in a new window)used for evaluation.

When given a text input, the Moderation endpoint assesses whether the content is sexual, hateful, violent, or promotes self-harm—content prohibited by ourcontent policy⁠(opens in a new window). The endpoint has been trained to be quick, accurate, and to perform robustly across a range of applications. Importantly, this reduces the chances of products “saying” the wrong thing, even when deployed to users at-scale. As a consequence, AI can unlock benefits in sensitive settings, like education, where it could not otherwise be used with confidence.

input text

Violence

Self-harm

Hate

Sexual

Moderation endpoint

Flagged

The Moderation endpoint helps developers to benefit from our infrastructure investments. Rather than build and maintain their own classifiers—an extensive process, as we document in ourpaper⁠(opens in a new window)—they can instead access accurate classifiers through a single API call.

As part of OpenAI’scommitment⁠tomaking the AI ecosystem safer⁠, we are providing this endpoint to allow free moderation of all OpenAI API-generated content. For instance,Inworld⁠(opens in a new window), an OpenAI API customer, uses the Moderation endpoint to help their AI-based virtual characters remain appropriate for their audiences. By leveraging OpenAI’s technology, Inworld can focus on their core product: creating memorable characters. We currently do not support monitoring of third-party traffic.

Get started with the Moderation endpoint by checking outthe documentation⁠(opens in a new window). More details of the training process and model performance are available in ourpaper⁠(opens in a new window). We have also released anevaluation dataset⁠(opens in a new window), featuring Common Crawl data labeled within these categories, which we hope will spur further research in this area.

Get started with the Moderation endpoint by checking outthe documentation⁠(opens in a new window). More details of the training process and model performance are available in ourpaper⁠(opens in a new window). We have also released anevaluation dataset⁠(opens in a new window), featuring Common Crawl data labeled within these categories, which we hope will spur further research in this area.

* View documentation(opens in a new window)

* API Platform

Authors

Todor Markov, Chong Zhang, Sandhini Agarwal, Tyna Eloundou, Teddy Lee, Steven Adler, Angela Jiang, Lilian Weng

Related articles

View all

Global news partnerships: Le Monde and Prisa Media Company Mar 13, 2024

Review completed & Altman, Brockman to continue to lead OpenAI Company Mar 8, 2024

OpenAI announces new members to board of directors Company Mar 8, 2024

OpenAI announces new members to board of directors Company Mar 8, 2024

Help shape what we cover next Help bepalen wat we hierna volgen

Anonymous feedback, no frontend account needed. Anonieme feedback, zonder front-end account.

More from OpenAI Meer van OpenAI

All updates Alle updates

Gemini komt eraan