Skip to content

Commit

Permalink
[Docs] add humanevalx dataset link in config (#559)
Browse files Browse the repository at this point in the history
* [Docs] add humanevalx dataset link in config

* [Docs] add humanevalx dataset link in config

* minor fix
  • Loading branch information
yingfhu committed Nov 10, 2023
1 parent 32884f2 commit 95e0da0
Show file tree
Hide file tree
Showing 4 changed files with 15 additions and 1 deletion.
5 changes: 4 additions & 1 deletion configs/datasets/humanevalx/humanevalx_gen_0af626.py
Original file line number Diff line number Diff line change
Expand Up @@ -44,12 +44,15 @@
] # do not support rust now
}

# Please download the needed `xx.jsonl.gz` from
# https://github.com/THUDM/CodeGeeX2/tree/main/benchmark/humanevalx
# and move them into `data/humanevalx/` folder
humanevalx_datasets = [
dict(
type=HumanevalXDataset,
abbr=f'humanevalx-{lang}',
language=lang,
path='./backup_data/humanevalx',
path='./data/humanevalx',
reader_cfg=humanevalx_reader_cfg,
infer_cfg=humanevalx_infer_cfg[lang],
eval_cfg=humanevalx_eval_cfg_dict[lang])
Expand Down
3 changes: 3 additions & 0 deletions configs/datasets/humanevalx/humanevalx_gen_620cfa.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,9 @@
for lang in ['python', 'cpp', 'go', 'java', 'js'] # do not support rust now
}

# Please download the needed `xx.jsonl.gz` from
# https://github.com/THUDM/CodeGeeX2/tree/main/benchmark/humanevalx
# and move them into `data/humanevalx/` folder
humanevalx_datasets = [
dict(
type=HumanevalXDataset,
Expand Down
4 changes: 4 additions & 0 deletions docs/en/advanced_guides/code_eval_service.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,10 @@

To complete LLM code capability evaluation, we need to set up an independent evaluation environment to avoid executing erroneous codes on development environments which would cause unavoidable losses. The current Code Evaluation Service used in OpenCompass refers to the project [code-evaluator](https://github.com/open-compass/code-evaluator.git), which has already supported evaluating datasets for multiple programming languages [humaneval-x](https://huggingface.co/datasets/THUDM/humaneval-x). The following tutorials will introduce how to conduct code review services under different requirements.

Dataset [download address](https://github.com/THUDM/CodeGeeX2/tree/main/benchmark/humanevalx). Please download the needed files (xx.jsonl.gz) into `./data/humanevalx` folder.

Supported languages are `python`, `cpp`, `go`, `java`, `js`.

## Launching the Code Evaluation Service

1. Ensure you have installed Docker, please refer to [Docker installation document](https://docs.docker.com/engine/install/).
Expand Down
4 changes: 4 additions & 0 deletions docs/zh_cn/advanced_guides/code_eval_service.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,10 @@

为了完成LLM代码能力评测,我们需要搭建一套独立的评测环境,避免在开发环境执行错误代码从而造成不可避免的损失。目前 OpenCompass 使用的代码评测服务可参考[code-evaluator](https://github.com/open-compass/code-evaluator)项目,并已经支持评测多编程语言的数据集 [humaneval-x](https://huggingface.co/datasets/THUDM/humaneval-x)。接下来将围绕代码评测服务介绍不同需要下的评测教程。

数据集[下载地址](https://github.com/THUDM/CodeGeeX2/tree/main/benchmark/humanevalx),请下载需要评测的语言(××.jsonl.gz)文件,并放入`./data/humanevalx`文件夹。

目前支持的语言有`python`, `cpp`, `go`, `java`, `js`

## 启动代码评测服务

1. 确保您已经安装了 docker,可参考[安装docker文档](https://docs.docker.com/engine/install/)
Expand Down

0 comments on commit 95e0da0

Please sign in to comment.