Announcing | TLM (Trustworthy Language Model) for reliable LLM outputs.Learn more.

Content Moderation

Use Cleanlab to quickly improve your content moderation datasets, models, and processes.

Find labeling errors; Decide when another review is appropriate; Discover good/bad moderators; Deploy robust ML models with 1-click.
Hero Picture

Case StudyToxic Language Detection @ VAST-OSINT

VAST-OSINT used Cleanlab to identify incorrect labels to improve toxic language detection models.

Quote from David Knickerbocker CTO of VAST-OSINT
A while back, I made a toxic language classifier. However, I was unsatisfied with the training data, […] I split the text by sentences while retaining the original label, hoping I'd be able to quickly clean-up, but that didn't work well.

I took the sentence-labeled training data and threw it at cleanlab to see how well confident learning could identify the incorrect labels. These results look amazing to me.

If nothing else, this can help identify training data to TOSS if you don't want to automate correction.
VAST-OSINT is on a quest to tame the web into safe, secure and on-demand data streams to help customers isolate and remediate misinformation, detect influence operations, and keep your companies and customers safer.
Company Logo

HOW CLEANLAB HELPS YOU BETTER MODERATE CONTENT

Help Section Icon
Determine which of your moderators is performing best/worst overall. Read More
Image
Help Section Icon
Videos on using Cleanlab Studio to find and fix incorrect values in:
Help Section Icon
Train and deploy state-of-the-art content moderation/categorization models (with well-calibrated uncertainty estimates) in 1-click. Cleanlab Studio automatically applies the most suitable Foundation/LLM models and AutoML systems for your content. Learn more
Help Section Icon
Quickly find and fix issues in a content dataset (categorization errors, outliers, ambiguous examples, near duplicates) — and then easily deploy a more reliable ML model. Read More
Help Section Icon
Confidently make model-assisted moderation decisions in real-time, deciding when to flag content for human review, and when to request a second/third review (for hard examples). Read More
Help Section Icon
Read about analyzing politeness labels provided by multiple data annotators.
Help Section Icon
Read about automatic error detection for image/text tagging datasets.


Cleanlab Studio auto-corrects raw data to ensure reliable predictions so you can provide safe user experiences.

Image
Case Study
SHARECHAT USES CLEANLAB TO IMPROVE CONTENT CLASSIFICATION


Cleanlab automatically identified an error rate of 3% in the concept categorization process for content in the Moj video-sharing app. Shown are a couple mis-categorized examples that Cleanlab detected in the app.

Image

For this dataset, Cleanlab Studio’s AutoML automatically produced a more accurate visual concept classification model (56.7% accuracy) than ShareChat’s in-house 3D ResNet model (52.6% accuracy). Auto-correcting the dataset immediately boosted Cleanlab AutoML accuracy up to 58.9% (see chart).

Image

ShareChat is India’s largest native social media app with over 300 million users. The company employs large teams of content moderators to categorize user video content in many ways.

Logo of the company ShareChat. A multicolor rounded square with the words ShareChat below.