Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Azure AI Content Safety is a content moderation service developed by Microsoft that helps you detect harmful content from different modalities and languages. The Content Safety (Text) tool is a wrapper for the Azure AI Content Safety Text API, which allows you to detect text content and get moderation results. For more information, see Azure AI Content Safety.
Prerequisites
- Create an Azure AI Content Safety resource.
- Add an
Azure Content Safetyconnection in prompt flow. Fill theAPI keyfield withPrimary keyfrom theKeys and Endpointsection of the created resource.
Inputs
You can use the following parameters as inputs for this tool:
| Name | Type | Description | Required |
|---|---|---|---|
| text | string | Text that needs to be moderated. | Yes |
| hate_category | string | Moderation sensitivity for the Hate category. Choose from four options: disable, low_sensitivity, medium_sensitivity, or high_sensitivity. The disable option means no moderation for the Hate category. The other three options mean different degrees of strictness in filtering out hate content. The default is medium_sensitivity. |
Yes |
| sexual_category | string | Moderation sensitivity for the Sexual category. Choose from four options: disable, low_sensitivity, medium_sensitivity, or high_sensitivity. The disable option means no moderation for the Sexual category. The other three options mean different degrees of strictness in filtering out sexual content. The default is medium_sensitivity. |
Yes |
| self_harm_category | string | Moderation sensitivity for the Self-harm category. Choose from four options: disable, low_sensitivity, medium_sensitivity, or high_sensitivity. The disable option means no moderation for the Self-harm category. The other three options mean different degrees of strictness in filtering out self-harm content. The default is medium_sensitivity. |
Yes |
| violence_category | string | Moderation sensitivity for the Violence category. Choose from four options: disable, low_sensitivity, medium_sensitivity, or high_sensitivity. The disable option means no moderation for the Violence category. The other three options mean different degrees of strictness in filtering out violence content. The default is medium_sensitivity. |
Yes |
For more information, see Azure AI Content Safety.
Outputs
The following sample is an example JSON format response returned by the tool:
{
"action_by_category": {
"Hate": "Accept",
"SelfHarm": "Accept",
"Sexual": "Accept",
"Violence": "Accept"
},
"suggested_action": "Accept"
}
The action_by_category field gives you a binary value for each category: Accept or Reject. This value shows if the text meets the sensitivity level that you set in the request parameters for that category.
The suggested_action field gives you an overall recommendation based on the four categories. If any category has a Reject value, suggested_action is also Reject.