The moderation object
Represents if a given input is potentially harmful.
id string
The unique identifier for the moderation request.
model string
The model used to generate the moderation results.
results array
A list of moderation objects.
flaggedboolean
Whether any of the below categories are flagged.
categoriesobject
A list of the categories, and whether they are flagged or not.
harassmentboolean
Content that expresses, incites, or promotes harassing language towards any target.
harassment/threateningboolean
Harassment content that also includes violence or serious harm towards any target.
sexualboolean
Content meant to arouse sexual excitement, such as the description of sexual activity, or that promotes sexual services (excluding sex education and wellness).
hateboolean
Content that expresses, incites, or promotes hate based on race, gender, ethnicity, religion, nationality, sexual orientation, disability status, or caste. Hateful content aimed at non-protected groups (e.g., chess players) is harassment.
hate/threateningboolean
Hateful content that also includes violence or serious harm towards the targeted group based on race, gender, ethnicity, religion, nationality, sexual orientation, disability status, or caste.
illicitboolean or null
Content that includes instructions or advice that facilitate the planning or execution of wrongdoing, or that gives advice or instruction on how to commit illicit acts.
illicit/violentboolean or null
Content that includes instructions or advice that facilitate the planning or execution of wrongdoing that also includes violence, or that gives advice or instruction on the procurement of any weapon.
self-harmboolean
Content that promotes, encourages, or depicts acts of self-harm, such as suicide, cutting, and eating disorders.
self-harm/intentboolean
Content where the speaker expresses that they are engaging or intend to engage in acts of self-harm, such as suicide, cutting, and eating disorders.
self-harm/instructionsboolean
Content that encourages performing acts of self-harm, such as suicide, cutting, and eating disorders, or that gives instructions or advice on how to commit such acts.
sexual/minorsboolean
Sexual content that includes an individual who is under 18 years old.
violenceboolean
Content that depicts death, violence, or physical injury.
violence/graphicboolean
Content that depicts death, violence, or physical injury in graphic detail.
category_scoresobject
A list of the categories along with their scores as predicted by the model.
harassmentnumber
The score for the categoryharassment.
harassment/threateningnumber
The score for the categoryharassment/threatening.
sexualnumber
The score for the categorysexual.
hatenumber
The score for the categoryhate.
hate/threateningnumber
The score for the categoryhate/threatening.
illicitnumber
The score for the categoryillicit.
illicit/violentnumber
The score for the categoryillicit/violent.
self-harmnumber
The score for the categoryself-harm.
self-harm/intentnumber
The score for the categoryself-harm/intent.
self-harm/instructionsnumber
The score for the categoryself-harm/instructions.
sexual/minorsnumber
The score for the categorysexual/minors.
violencenumber
The score for the categoryviolence.
violence/graphicnumber
The score for the categoryviolence/graphic.
category_applied_input_typesobject
A list of the categories along with the input type(s) that the score applies to.
harassmentarray
The applied input type(s) for the categoryharassment.
harassment/threateningarray
The applied input type(s) for the categoryharassment/threatening.
sexualarray
The applied input type(s) for the categorysexual.
hatearray
The applied input type(s) for the categoryhate.
hate/threateningarray
The applied input type(s) for the categoryhate/threatening.
illicitarray
The applied input type(s) for the categoryillicit.
illicit/violentarray
The applied input type(s) for the categoryillicit/violent.
self-harmarray
The applied input type(s) for the categoryself-harm.
self-harm/intentarray
The applied input type(s) for the categoryself-harm/intent.
self-harm/instructionsarray
The applied input type(s) for the categoryself-harm/instructions.
sexual/minorsarray
The applied input type(s) for the categorysexual/minors.
violencearray
The applied input type(s) for the categoryviolence.
violence/graphicarray
The applied input type(s) for the categoryviolence/graphic.
{
"id": "modr-0d9740456c391e43c445bf0f010940c7",
"model": "omni-moderation-latest",
"results": [
{
"flagged": true,
"categories": {
"harassment": true,
"harassment/threatening": true,
"sexual": false,
"hate": false,
"hate/threatening": false,
"illicit": false,
"illicit/violent": false,
"self-harm/intent": false,
"self-harm/instructions": false,
"self-harm": false,
"sexual/minors": false,
"violence": true,
"violence/graphic": true
},
"category_scores": {
"harassment": 0.8189693396524255,
"harassment/threatening": 0.804985420696006,
"sexual": 1.573112165348997e-6,
"hate": 0.007562942636942845,
"hate/threatening": 0.004208854591835476,
"illicit": 0.030535955153511665,
"illicit/violent": 0.008925306722380033,
"self-harm/intent": 0.00023023930975076432,
"self-harm/instructions": 0.0002293869201073356,
"self-harm": 0.012598046106750154,
"sexual/minors": 2.212566909570261e-8,
"violence": 0.9999992735124786,
"violence/graphic": 0.843064871157054
},
"category_applied_input_types": {
"harassment": [
"text"
],
"harassment/threatening": [
"text"
],
"sexual": [
"text",
"image"
],
"hate": [
"text"
],
"hate/threatening": [
"text"
],
"illicit": [
"text"
],
"illicit/violent": [
"text"
],
"self-harm/intent": [
"text",
"image"
],
"self-harm/instructions": [
"text",
"image"
],
"self-harm": [
"text",
"image"
],
"sexual/minors": [
"text"
],
"violence": [
"text",
"image"
],
"violence/graphic": [
"text",
"image"
]
}
}
]
}