diff --git a/README.md b/README.md index 99646d029..f2197f25a 100755 --- a/README.md +++ b/README.md @@ -120,22 +120,22 @@ Supported models list: First, download the image we provide: ```bash # A2 x86 -docker pull xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-x86 +docker pull xllm/xllm-ai:xllm-dev-hb-rc2-x86 # A2 arm -docker pull xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-arm +docker pull xllm/xllm-ai:xllm-dev-hb-rc2-arm # A3 arm -docker pull xllm/xllm-ai:xllm-0.7.1-dev-hc-rc2-arm +docker pull xllm/xllm-ai:xllm-dev-hc-rc2-arm # or # A2 x86 -docker pull quay.io/jd_xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-x86 +docker pull quay.io/jd_xllm/xllm-ai:xllm-dev-hb-rc2-x86 # A2 arm -docker pull quay.io/jd_xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-arm +docker pull quay.io/jd_xllm/xllm-ai:xllm-dev-hb-rc2-arm # A3 arm -docker pull quay.io/jd_xllm/xllm-ai:xllm-0.7.1-dev-hc-rc2-arm +docker pull quay.io/jd_xllm/xllm-ai:xllm-dev-hc-rc2-arm ``` Then create the corresponding container: ```bash -sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ -v /etc/hccn.conf:/etc/hccn.conf xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-x86 +sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ -v /etc/hccn.conf:/etc/hccn.conf xllm/xllm-ai:xllm-dev-hb-rc2-x86 ``` Install official repo and submodules: diff --git a/README_zh.md b/README_zh.md index 68319229d..40b1bbf93 100755 --- a/README_zh.md +++ b/README_zh.md @@ -116,22 +116,22 @@ xLLM 提供了强大的智能计算能力,通过硬件系统的算力优化与 首先下载我们提供的镜像: ```bash # A2 x86 -docker pull quay.io/jd_xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-x86 +docker pull quay.io/jd_xllm/xllm-ai:xllm-dev-hb-rc2-x86 # A2 arm -docker pull quay.io/jd_xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-arm +docker pull quay.io/jd_xllm/xllm-ai:xllm-dev-hb-rc2-arm # A3 arm -docker pull quay.io/jd_xllm/xllm-ai:xllm-0.7.1-dev-hc-rc2-arm +docker pull quay.io/jd_xllm/xllm-ai:xllm-dev-hc-rc2-arm # 或者 # A2 x86 -docker pull xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-x86 +docker pull xllm/xllm-ai:xllm-dev-hb-rc2-x86 # A2 arm -docker pull xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-arm +docker pull xllm/xllm-ai:xllm-dev-hb-rc2-arm # A3 arm -docker pull xllm/xllm-ai:xllm-0.7.1-dev-hc-rc2-arm +docker pull xllm/xllm-ai:xllm-dev-hc-rc2-arm ``` 然后创建对应的容器 ```bash -sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ -v /etc/hccn.conf:/etc/hccn.conf quay.io/jd_xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-x86 +sudo docker run -it --ipc=host -u 0 --privileged --name mydocker --network=host --device=/dev/davinci0 --device=/dev/davinci_manager --device=/dev/devmm_svm --device=/dev/hisi_hdc -v /var/queue_schedule:/var/queue_schedule -v /usr/local/Ascend/driver:/usr/local/Ascend/driver -v /usr/local/Ascend/add-ons/:/usr/local/Ascend/add-ons/ -v /usr/local/sbin/npu-smi:/usr/local/sbin/npu-smi -v /usr/local/sbin/:/usr/local/sbin/ -v /var/log/npu/conf/slog/slog.conf:/var/log/npu/conf/slog/slog.conf -v /var/log/npu/slog/:/var/log/npu/slog -v /export/home:/export/home -w /export/home -v ~/.ssh:/root/.ssh -v /var/log/npu/profiling/:/var/log/npu/profiling -v /var/log/npu/dump/:/var/log/npu/dump -v /home/:/home/ -v /runtime/:/runtime/ -v /etc/hccn.conf:/etc/hccn.conf xllm/xllm-ai:xllm-dev-hb-rc2-x86 ``` 下载官方仓库与模块依赖: diff --git a/RELEASE.md b/RELEASE.md index 77ebf46d1..1d881753c 100644 --- a/RELEASE.md +++ b/RELEASE.md @@ -1,30 +1,3 @@ -# Release xllm 0.7.1 - -## **Major Features and Improvements** - -### Model Support - -- Support GLM-4.5-Air. -- Support Qwen3-VL-Moe. - -### Feature - -- Support scheduler overlap when enable chunked prefill and MTP. -- Enable multi-process mode when running VLM model. -- Support AclGraph for GLM-4.5. - -### Bugfix - -- Reslove core dump of qwen embedding 0.6B. -- Resolve duplicate content in multi-turn tool call conversations. -- Support sampler parameters for MTP. -- Enable MTP and schedule overlap to work simultaneously. -- Resolve google.protobuf.Struct parsing failures which broke tool_call and think toggle functionality. -- Fix the precision issue in the Qwen2 model caused by model_type is not be assigned. -- Fix core dump of GLM 4.5 when enable MTP. -- Temporarily use heap allocation for VLM backend. -- Reslove core dump of stream chat completion request for VLM. - # Release xllm 0.7.0 ## **Major Features and Improvements** diff --git a/cibuild/build_npu.sh b/cibuild/build_npu.sh index 13eed9110..d4ea3477d 100644 --- a/cibuild/build_npu.sh +++ b/cibuild/build_npu.sh @@ -6,7 +6,7 @@ function error() { exit 1 } -IMAGE="quay.io/jd_xllm/xllm-ai:xllm-0.7.1-dev-hb-rc2-x86" +IMAGE="quay.io/jd_xllm/xllm-ai:xllm-dev-hb-rc2-x86" RUN_OPTS=( --rm diff --git a/version.txt b/version.txt index 39e898a4f..faef31a43 100644 --- a/version.txt +++ b/version.txt @@ -1 +1 @@ -0.7.1 +0.7.0