diff --git a/README.md b/README.md index 02ee5f0..a2896f2 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,6 @@ ![Banner](https://raw.githubusercontent.com/pavanjava/bootstrap-rag/refs/heads/main/assets/bootstrap-rag.png) # bootstrap-rag -this project will bootstrap and scaffold the projects for specific semantic search and RAG applications along with regular boiler plate code. +This project will bootstrap and scaffold the projects for specific semantic search and RAG applications along with regular boilerplate code. ### Architecture ![Arch](assets/architecture.png) diff --git a/bootstraprag/templates/evaluations/phoenix_evals/readme.md b/bootstraprag/templates/evaluations/phoenix_evals/readme.md index 15beb66..be589ce 100644 --- a/bootstraprag/templates/evaluations/phoenix_evals/readme.md +++ b/bootstraprag/templates/evaluations/phoenix_evals/readme.md @@ -1,2 +1,23 @@ ## Phoenix Evaluations -- Under development \ No newline at end of file + +This repository provides a script for evaluating model-generated responses using Phoenix's `HallucinationEvaluator` and `QAEvaluator`. +To start evaluating run `bootstraprag create phoenix_evals` +select the specific template shown on cli + +```text +? Which technology would you like to use? standalone-evaluations +? Which template would you like to use? + deep-evals + mlflow-evals +❯ phoenix-evals + ragas-evals +``` + +just replace the `input_data.csv` with your own data, the file has following columns +`id,reference,query,response`. + +### How to execute? +- run `python basic_evaluations.py` + +### What to expect? +- At the end of process, you can see the `evaluation_report.csv` is created and kept in the parent folder where you can see different aspects of evaluations carried on your input data. \ No newline at end of file