Comprehend / Client / detect_toxic_content

detect_toxic_content#

Comprehend.Client.detect_toxic_content(**kwargs)#

Performs toxicity analysis on the list of text strings that you provide as input. The API response contains a results list that matches the size of the input list. For more information about toxicity detection, see Toxicity detection in the Amazon Comprehend Developer Guide.

See also: AWS API Documentation

Request Syntax

response = client.detect_toxic_content(
    TextSegments=[
        {
            'Text': 'string'
        },
    ],
    LanguageCode='en'|'es'|'fr'|'de'|'it'|'pt'|'ar'|'hi'|'ja'|'ko'|'zh'|'zh-TW'
)
Parameters:
  • TextSegments (list) –

    [REQUIRED]

    A list of up to 10 text strings. Each string has a maximum size of 1 KB, and the maximum size of the list is 10 KB.

    • (dict) –

      One of the of text strings. Each string has a size limit of 1KB.

      • Text (string) – [REQUIRED]

        The text content.

  • LanguageCode (string) –

    [REQUIRED]

    The language of the input text. Currently, English is the only supported language.

Return type:

dict

Returns:

Response Syntax

{
    'ResultList': [
        {
            'Labels': [
                {
                    'Name': 'GRAPHIC'|'HARASSMENT_OR_ABUSE'|'HATE_SPEECH'|'INSULT'|'PROFANITY'|'SEXUAL'|'VIOLENCE_OR_THREAT',
                    'Score': ...
                },
            ],
            'Toxicity': ...
        },
    ]
}

Response Structure

  • (dict) –

    • ResultList (list) –

      Results of the content moderation analysis. Each entry in the results list contains a list of toxic content types identified in the text, along with a confidence score for each content type. The results list also includes a toxicity score for each entry in the results list.

      • (dict) –

        Toxicity analysis result for one string. For more information about toxicity detection, see Toxicity detection in the Amazon Comprehend Developer Guide.

        • Labels (list) –

          Array of toxic content types identified in the string.

          • (dict) –

            Toxic content analysis result for one string. For more information about toxicity detection, see Toxicity detection in the Amazon Comprehend Developer Guide

            • Name (string) –

              The name of the toxic content type.

            • Score (float) –

              Model confidence in the detected content type. Value range is zero to one, where one is highest confidence.

        • Toxicity (float) –

          Overall toxicity score for the string. Value range is zero to one, where one is the highest confidence.

Exceptions