Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Responsible AI improvement with Content Safety/Content Filtering (multiple items) #129

Open
4 of 6 tasks
gbecerra1982 opened this issue Oct 24, 2023 · 2 comments
Open
4 of 6 tasks
Assignees

Comments

@gbecerra1982
Copy link
Collaborator

gbecerra1982 commented Oct 24, 2023

List of tasks: (see this item description below)

Item description

User should be able to define what functions from Responsible AI plugin he/she wants to use as a guardrail when receiving the ask from the user and before sending the response back to the user and its thresholds.

List of functions:

  • Unfairness
  • Harm detection Text moderation (harm categories)
  • Prompt attacks Prompt shield
  • Protected material
  • Groundedness check (migrate to CS)
  • Blocked words (migrate to CS)

Notes:

  • Users can configure what functions they want to use and thresholds in gpt-rag configuration.
  • Items that can be met using native Azure OpenAI content filtering should do it so we save API calls.
  • Orchestator responses should contain metadata information about guardrails responses so future APIM or Security function can check them to enforce.

Out of scope items to be handled in a separated item:

  1. IaaS (bicep) update to create and configure content safety service

  2. Architecture redesign:
    Create a new Azure Function "Custom Security Policy" that will receive the text from the Orchestrator and validate the content does not have violence, sexual, etc.
    This function is the beggining of Security Function to add controls of security to the platform, further will be introduced additional security controls.
    Image

We need to prepare this function so the Security Team can add additional controls (i.e. Microsoft Purview, etc)

References:
https://techcommunity.microsoft.com/t5/ai-azure-ai-services-blog/azure-ai-announces-prompt-shields-for-jailbreak-and-indirect/ba-p/4099140

https://techcommunity.microsoft.com/t5/ai-azure-ai-services-blog/detect-and-mitigate-ungrounded-model-outputs/ba-p/4099261%23:~:text=Today%2520Azure%2520AI%2520makes%2520this,Copilots%2520and%2520document%2520summarization%2520applications.

@placerda placerda transferred this issue from Azure/GPT-RAG Nov 20, 2023
@placerda placerda changed the title Add Content Safety Integrated Add Content Safety Integrated (Orchestration) Apr 2, 2024
@gbecerra1982 gbecerra1982 transferred this issue from Azure/gpt-rag-orchestrator Apr 8, 2024
@placerda placerda changed the title Add Content Safety Integrated (Orchestration) Responsible AI improvement with Content Safety Apr 9, 2024
@placerda placerda changed the title Responsible AI improvement with Content Safety Responsible AI improvement with Content Safety (multiple items) Apr 9, 2024
@placerda placerda changed the title Responsible AI improvement with Content Safety (multiple items) Responsible AI improvement with Content Safety/Content Filtering (multiple items) May 7, 2024
@vladborys
Copy link
Collaborator

@placerda
Copy link
Collaborator

I'll pull this to my fork and then create the pull request after doing the bash/sh adjustments

@placerda placerda self-assigned this May 21, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants