Toxicity classifier
The toxicity model detects whether text contains toxic content such as threatening language, insults, obscenities, identity-based hate, or sexually explicit language. The model was trained on the civil comments dataset: https://figshare.com/articles/data_json/7376747 which contains ~2 million comments labeled for toxicity. The model is built on top of the Universal Sentence Encoder (Cer et al., 2018).
More information about how the toxicity labels were calibrated can be found here.

Check out our demo, which uses the toxicity model to predict the toxicity of several sentences taken from this Kaggle dataset. Users can also input their own text for classification.
Installation
Using yarn:
$ yarn add @tensorflow/tfjs @tensorflow-models/toxicity
Using npm:
$ npm install @tensorflow/tfjs @tensorflow-models/toxicity
Usage
To import in npm:
import * as toxicity from '@tensorflow-models/toxicity';
or as a standalone script tag:
<script src="https://cdn.jsdelivr.net/npm/@tensorflow/tfjs"></script>
<script src="https://cdn.jsdelivr.net/npm/@tensorflow-models/toxicity"></script>
Then:
const threshold = 0.9;
toxicity.load(threshold).then(model => {
const sentences = ['you suck'];
model.classify(sentences).then(predictions => {
console.log(predictions);
});
});