Skip to content

Feature/implement the mentalchat16k dataset support for clinical evaluation#1218

Merged
chakravarthik27 merged 56 commits intorelease/2.7.0from
feature/implement-the-mentalchat16k-dataset-support-for-clinical-evaluation
Sep 9, 2025
Merged

Feature/implement the mentalchat16k dataset support for clinical evaluation#1218
chakravarthik27 merged 56 commits intorelease/2.7.0from
feature/implement-the-mentalchat16k-dataset-support-for-clinical-evaluation

Conversation

@chakravarthik27
Copy link
Collaborator

This pull request introduces a new mental health evaluation capability to the codebase, enabling the assessment of AI-generated mental health counseling responses using a set of clinical consultation metrics. It adds a new evaluation prompt and schema, a corresponding evaluation class, and integrates this functionality into the clinical test transformation pipeline. The changes also include a new SimplePrompt sample type to support these evaluations and ensure results are parsed and scored appropriately.

Mental Health Evaluation Integration

  • Added MENTAL_HEALTH_EVAL_PROMPT and MHCEvaluation schema in eval_prompts.py to define a structured prompt and scoring rubric for mental health counseling response evaluation.
  • Introduced the RatingEval class in llm_eval.py, which uses the new prompt and schema to parse and score AI responses, including batch evaluation support.
  • Updated the clinical test transformation logic to support a new mental_health test type, with a dedicated MentalHealth class that loads data, transforms samples, and runs evaluations using the new prompt and scoring system. [1] [2]

Sample Type Extension

  • Added the SimplePrompt class to sample.py, designed for prompt-response pairs, with methods for evaluation, scoring, and pass/fail determination using the mental health metrics and the new evaluation pipeline.

Internal Imports and Wiring

  • Registered and imported the new prompt and evaluation classes in relevant modules to enable their use throughout the codebase. [1] [2]

…n Snow Labs to Pacific AI Corp across documentation.
…n Snow Labs to Pacific AI Corp across documentation.
…corp' of https://github.com/pacific-ai-corp/langtest into refactor/replace-links-from-johnsnowlabs-to-pacific-ai-corp
Copy link

@blidiselalin blidiselalin left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should be ok

@chakravarthik27 chakravarthik27 merged commit 9490563 into release/2.7.0 Sep 9, 2025
3 checks passed
@chakravarthik27 chakravarthik27 deleted the feature/implement-the-mentalchat16k-dataset-support-for-clinical-evaluation branch September 27, 2025 08:04
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants