Skip to content


Agentic Documents Assistant

License: MIT

The Agentic Documents Assistant is an LLM assistant that allows users to answer complex questions from their business documents through natural conversations. It supports answering factual questions by retrieving information directly from documents using semantic search with the popular RAG design pattern. Additionally, it answers analytical questions such as which contracts will expire in the next 3 months? by translating user questions into SQL queries and running them against a database of entities extracted from the documents using a batch process. It is also able to answer complex multi-step questions by combining retrieval, analytical, and other tools and data sources using an LLM agent design pattern.

To learn more about the design and architecture of this solution, check the accompanying AWS ML blog post: Boosting RAG-based intelligent document assistants using entity extraction, SQL querying, and agents with Amazon Bedrock.

Key Features

  • Semantic search to augment response generation with relevant documents
  • Structured metadata & entities extraction and SQL queries for analytical reasoning
  • An agent built with the Reason and Act (ReAct) instruction format that determines whether to use search or SQL to answer a given question.

Architecture Overview

The following architecture diagrams depicts the design of the solution.

Architecture of the agentic AI documents assistant on AWS


Below an outline of the main folders included in this asset.

Folder Description
backend Includes a Typescript CDK project implementing IaaC to setup the backend infrastructure.
frontend A Typescript CDK project to setup infrastructure for deploying and hosting the frontend app with AWS Amplify.
frontend/chat-app A Next.js app with AWS Cognito Authentication and secured backend connectivity.
data-pipelines Notebooks implementing SageMaker Jobs and Pipeline to process the data in batch.
experiments Notebooks and code showcasing different modules of the solution as standalone experiments for research and development.

Getting Started

Follow the insturctions below to setup the solution on your account.



To install the solution in your AWS account:

  1. Clone this repository.
  2. Install the backend CDK app, as follows:
    1. Go inside the backend folder.
    2. Run npm install to install the dependencies.
    3. If you have never used CDK in the current account and region, run bootstrapping with npx cdk bootstrap.
    4. Run npx cdk deploy to deploy the stack.
    5. Take note of the SageMaker IAM Policy ARN found in the CDK stack output.
  3. Deploy the Next.js frontend on AWS Amplify:
    1. Go inside the frontend folder.
    2. Run npm install to install the dependencies.
    3. Run npx cdk deploy to deploy a stack that builds an Amplify CI/CD
    4. Once the CI/CD is ready go to the Amplify console and trigger a build.
    5. Once the app is built, click the hosting link to view the app. You can now create a new account and login to view the assistant.
  4. To load the underlying data, run the SageMaker Pipeline in data-pipelines/04-sagemaker-pipeline-for-documents-processing.ipynb. This will load two files: (1) a pre-created csv file to load into the SQL tables, and (2) a json file containing preprocessed Amazon financial documents which will be used to create the semantic search index used by the LLM assistant.

After running the above steps successfully, you can start interacting with the assistant and ask questions. If you want to update the underlying documents and extract new entities, explore the notebooks 1 to 5 under the experiments folder.

Clean up

To remove the resources of the solution:

  1. Remove the stack inside the backend folder by running npx cdk destroy.
  2. Remove the stack inside the frontend folder by running npx cdk destroy.


The authors of this asset are:

  • Mohamed Ali Jamaoui: Solution designer/Core maintainer.
  • Giuseppe Hannen: Extensive contribution to the data extraction modules.
  • Laurens ten Cate: Contributed to extending the agent with SQL tool and early streamlit UI deployments.


See CONTRIBUTING for more information.


Future improvements

  • Improve the overall-inference speed.


This library is licensed under the MIT-0 License. See the LICENSE file.