Skip to content

Bad score for answer_correctness metric #1385

@Simeon-Krah

Description

@Simeon-Krah

[ ] I have checked the documentation and related resources and couldn't resolve my bug.

Describe the bug
The answer generated by my model is similar to the ground_truths, but my answer_correctness score is very poor eventhough I have good score in the other metrics like context_recall, precision and faithfulness.

Ragas version:0.1.18
Python version:3.12

Code to Reproduce
scores = evaluate(
eval_dataset,
metrics=[context_recall, context_precision, faithfulness, answer_correctness],
llm = llm,
embeddings=GoogleGenerativeAIEmbeddings(google_api_key="AIzaSyCJFrruyyyuuuytuuiysJnFcQZJa-E-QJOxyjK_X8rpBkIQ", model="models/embedding-001"),
)

Error trace
No error but poor answer correctness metric score

Expected behavior
To get a good score for the answer_correctness like I got for the others
image
Screenshot 2024-09-30 110333

Additional context
Add any other context about the problem here.

Metadata

Metadata

Assignees

No one assigned

    Labels

    answered🤖 The question has been answered. Will be closed automatically if no new commentsbugSomething isn't workingmodule-metricsthis is part of metrics module

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions