Industrial-level evaluation benchmarks for Coding LLMs in the full life-cycle of AI native software developing.企业级代码大模型评测体系,持续开放中
-
Updated
Jan 19, 2024 - Python
Industrial-level evaluation benchmarks for Coding LLMs in the full life-cycle of AI native software developing.企业级代码大模型评测体系,持续开放中
Pip compatible CodeBLEU metric implementation available for linux/macos/win
Backend for automated evaluation of programming tasks in higher education
The SF Code Evaluator
Python library to interact synchronously and asynchronously with tio.run
Add a description, image, and links to the code-evaluation topic page so that developers can more easily learn about it.
To associate your repository with the code-evaluation topic, visit your repo's landing page and select "manage topics."