Skip to main content
The Classifier tool analyzes content and flags it as potentially harmful across several predefined categories. Use it to review user-submitted text, moderate content at scale, or screen inputs before publishing.

Using the Classifier

1

Open Classifier in your dashboard

Click Classifier in the sidebar of your Monorail AI dashboard.
2

Enter your text

Paste or type the text you want to analyze into the input field.
3

Click Classify

Submit the text and let the AI evaluate it against the harm categories.
4

Review the results

Examine the category breakdown returned by the AI, which indicates potential harm levels across each dimension.

Classification categories

The Classifier evaluates content across the following categories:
  • Hate speech and discrimination
  • Violence and graphic content
  • Self-harm or dangerous activities
  • Sexual content
  • Spam or deceptive content
The Classifier provides probabilistic scores — it is an AI tool and may not catch all harmful content or may produce false positives. Always combine automated classification with human review for critical use cases.

Common use cases

  • Screening user-submitted content on platforms
  • Pre-publishing review for brand-safety compliance
  • Research and content analysis

Credit usage

Classifier uses text models and charges per token processed. The cost scales with the length of the content you submit. See the Credits page for current rates.