Skip to content
Discussion options

You must be logged in to vote

I’m building a production-grade Retrieval-Augmented Generation (RAG) system using LangChain and a vector database (e.g. Pinecone or Chroma) to power an internal knowledge assistant.

The system requirements are:

  • Ingest and index large document sets (PDFs, Markdown, internal docs)
  • Support semantic search with embeddings
  • Maintain good response latency at scale
  • Ensure reliable updates when documents change
  • Be deployable in a cloud-native environment (Docker / Kubernetes)

I’m trying to decide on best practices for:

  1. Chunking strategy

    • Optimal chunk size and overlap for long documents
    • How to handle structured vs unstructured content
  2. Vector store design

    • When to use managed services (Pinec…

Replies: 5 comments

Comment options

You must be logged in to vote
0 replies
Answer selected by techdev-loop
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants