Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Create Threat Model and Discuss RAG with its security risks for LLM #241

Open
jsotiro opened this issue Nov 4, 2023 · 1 comment
Open
Assignees
Labels
diagram Issues related to the Top 10 diagram discuss Indicates that this issue requires a deeper discussion v2 A topic for v2 discussion

Comments

@jsotiro
Copy link
Collaborator

jsotiro commented Nov 4, 2023

Retrieval augmented generation (RAG) is a technique to enrich LLMs with the apps/org own data. It has become very popular as it lowers the complexity entry to enriching input in LLM apps, allows for better access controls as opposed to fine-tuning, and is known to reduce hallucination (see https://www.securityweek.com/vector-embeddings-antidote-to-psychotic-llms-and-a-cure-for-alert-fatigue/) see also the excellent Samsung paper on enterprise use of GenAI and the role of RAG.

Currently, we have occasional references to RAG but we should create a threat model and discuss this on its own and the impact on the LLM top 10. As RAG becomes a distinct enterprise PATTERN it also creates its own security risks and expands the attack surface with:
This includes

  • a second process of generating embeddings that can vary in complexity and can be using the main or secondary LLM
  • vector database
  • the use of embeddings in prompts and responses
  • use of other services (eg Azure Cognitive Services) or specialised plugins, for instance, the OpenAI upsert retrieval plugin for semantic search plugin(https://github.com/openai/chatgpt-retrieval-plugin)

Some useful links**

architectural approaches
Azure: https://github.com/Azure/GPT-RAG
AWS SageMaker: https://docs.aws.amazon.com/sagemaker/latest/dg/jumpstart-foundation-models-customize-rag.html
AWS Bedrock RAG workshop: https://github.com/aws-samples/amazon-bedrock-rag-workshop

security concerns
Security of AI Embeddings explained
Anonymity at Risk? Assessing Re-Identification Capabilities of Large Language Models
Embedding Layer: AI (Brace For These Hidden GPT Dangers)

@jsotiro jsotiro added discuss Indicates that this issue requires a deeper discussion v2 A topic for v2 discussion labels Nov 4, 2023
@GangGreenTemperTatum GangGreenTemperTatum self-assigned this Apr 14, 2024
@GangGreenTemperTatum
Copy link
Collaborator

heya @jsotiro , doing some housekeeping on the repo, did you get chance to bring this up and is anything outstanding needed? i also have this issue which basically superseeds i believe as i also mentioned different API architectures and mediums

@GangGreenTemperTatum GangGreenTemperTatum added the diagram Issues related to the Top 10 diagram label Apr 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
diagram Issues related to the Top 10 diagram discuss Indicates that this issue requires a deeper discussion v2 A topic for v2 discussion
Projects
None yet
Development

No branches or pull requests

2 participants