Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

🗺️ Evaluation Experiment Tracking #2220

Closed
mikeldking opened this issue Feb 6, 2024 · 1 comment
Closed

🗺️ Evaluation Experiment Tracking #2220

mikeldking opened this issue Feb 6, 2024 · 1 comment
Labels
c/evals documentation Improvements or additions to documentation roadmap

Comments

@mikeldking
Copy link
Contributor

As a user I would like to:

Ship LLM apps with confidence
Golden datasets for QA, experimentation, and fine-tuning
Evaluate changes to LLMs, prompts, retrieval,
Track experiment runs during development and production
Production data provides the critical feedback loop, keeping benchmarks up to date.

@dosubot dosubot bot added c/evals c/metrics documentation Improvements or additions to documentation enhancement New feature or request labels Feb 6, 2024
@Arize-ai Arize-ai deleted a comment from dosubot bot Feb 7, 2024
@mikeldking mikeldking removed the enhancement New feature or request label May 13, 2024
@mikeldking
Copy link
Contributor Author

merging with #2017

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
c/evals documentation Improvements or additions to documentation roadmap
Projects
Status: Done
Development

No branches or pull requests

1 participant