Skip to content

redhat-ai-dev/developer-lightspeed-evaluation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

9 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Developer Lightspeed Evaluation

This branch contains the evaluation resources for Developer Lightspeed 1.8, including the source RAG documentation, the generated datasets, and the performance results across multiple LLMs.

πŸ“‚ Repository Structure

Path Description
πŸ“‚ rhdh-product-docs The Red Hat Developer Hub (RHDH) product documentation (v1.8) used as the source knowledge base for this evaluation.
πŸ“‚ dataset Contains raw and processed Q&A pairs, formatted specifically for the lightspeed-evaluation tool.
πŸ“‚ evaluation-result Detailed metrics and outcome reports from the model evaluations.
πŸ“„ categories_rhdh.yaml Manually defined topic groups used to classify and organize the Q&A pairs.

πŸ§ͺ Evaluation Overview

For the Developer Lightspeed 1.8 release, we ran evaluation against five distinct models.

Models Evaluated:

  • Gemini: Gemini-2.5-pro, Gemini-2.5-flash-lite
  • GPT: gpt-oss-120b
  • Llama: llama3.1:8b, llama4:scout

πŸ“Š View Results: For a deep dive into the performance metrics, please refer to the Evaluation Results directory.


βš™οΈ Methodology & Generation

The dataset in this repository was constructed using a synthetic generation pipeline to ensure comprehensive coverage of the documentation.

  • Source Material: The dataset is derived entirely from the RHDH 1.8 Product Docs.
  • Generation Tool: We used Ragas (Testset Generation for RAG) to generate diverse Q&A pairs.
  • Evaluation Tool: The evaluation was executed using the lightspeed-evaluation tool, which consumes the dataset and calculates performance metrics.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published