How to identify part of prompt triggering "Azure AI Content Safety"

Lucas Blin 0 Reputation points
2025-01-29T14:04:45.4266667+00:00

Reaching to you in order to understand why the following exchange is being filter as "sexual" through the Azure AI Content Safety :

  • Description of the flagged content.

      User input : "Je souhaite vendre et j'ai envie de le faire rapidement." (french)

      Translation : "I want to sell and to do it quickly"

  • Context in which the content was posted.

      As first message input to our chatbot.

  • Reason given by Azure AI Content Safety for the flagging (if positive).

      Sexual, severity : medium. Not much more information on why its being filtered out

  • Explanation of why the content is a false positive or negative.

      There is no mention of sexuality in the user input and this false positive is restrictive to our target work domain. My only guess is that it would be semantically close to something sexual in english ?

      The same sentence in English seems to work, however.

  • Any adjustments already attempted by adjusting severity settings or using custom categories.

      Couldn't not adjust severity as I'm in Europe and it doesn't seem possible and I can't get more info on which part of the prompt triggered the filter.

  • Screenshots or logs of the flagged content and system responses.

Reaching to you in order to understand why the following exchange is being filter as "sexual" through the Azure AI Content Safety :

  • Description of the flagged content.

      User input : "Je souhaite vendre et j'ai envie de le faire rapidement." (french)

      Translation : "I want to sell and to do it quickly"

  • Context in which the content was posted.

      As first message input to our chatbot.

  • Reason given by Azure AI Content Safety for the flagging (if positive).

      Sexual, severity : medium. Not much more information on why its being filtered out

  • Explanation of why the content is a false positive or negative.

      There is no mention of sexuality in the user input and this false positive is restrictive to our target work domain. My only guess is that it would be semantically close to something sexual in english ?

      The same sentence in English seems to work, however.

  • Any adjustments already attempted by adjusting severity settings or using custom categories.

      Couldn't not adjust severity as I'm in Europe and it doesn't seem possible and I can't get more info on which part of the prompt triggered the filter.

  • Screenshots or logs of the flagged content and system responses.

User's image

User's image

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
3,617 questions
Azure AI Content Safety
Azure AI Content Safety
An Azure service that enables users to identify content that is potentially offensive, risky, or otherwise undesirable. Previously known as Azure Content Moderator.
32 questions
0 comments No comments
{count} votes

1 answer

Sort by: Most helpful
  1. Pavankumar Purilla 3,005 Reputation points Microsoft Vendor
    2025-01-29T22:05:15.55+00:00

    Hi Lucas Blin,
    Greetings & Welcome to Microsoft Q&A forum! Thanks for posting your query!

    Thank you for bringing this to our attention. We understand your concern regarding the flagged content, and we want to assure you that this was purely a mistake by our content moderation model. There is no mention of any inappropriate content in your message, and we apologize for any inconvenience this may have caused.

    To address this issue and prevent similar occurrences in the future, we kindly ask you to raise a support request with Azure AI Content Safety. Your feedback is valuable to us, and we are committed to improving our services to better meet your needs.
    Thank you!

    0 comments No comments

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.