-
Notifications
You must be signed in to change notification settings - Fork 49
[fix] Adapt all sparse-attention methods to the new connector. #441
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
hek14
merged 2 commits into
ModelEngine-Group:dev-ucm-v1
from
wangwenxin0312:dev_sparse_fix
Dec 1, 2025
Merged
[fix] Adapt all sparse-attention methods to the new connector. #441
hek14
merged 2 commits into
ModelEngine-Group:dev-ucm-v1
from
wangwenxin0312:dev_sparse_fix
Dec 1, 2025
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
eea1826 to
f1e807a
Compare
5e89688 to
6276613
Compare
hek14
approved these changes
Dec 1, 2025
Lijiachen1018
pushed a commit
to Lijiachen1018/unified-cache-management
that referenced
this pull request
Dec 1, 2025
…Engine-Group#441) * sparse to adapt new connector * Adapt the YAML configuration
ygwpz
added a commit
that referenced
this pull request
Dec 2, 2025
* [opt] refactor uc connector (#364) refactor ucm_connector * [Feat] Implement kv cache broadcast in MLA (#367) * [Feat] Implement kv cache broadcast in MLA in ucm_connector * [Style] Change wait for broadcast into single task method * [feature] add ucm mock connector (#375) * add ucm mock connector * fix chunk prefill bug * [Feat] Support get launch config from yaml (#377) * [Feat] Support launch from config file * [Docs] Update documents for launch with yaml * [Fix] Change load only on first rank into configuration * [Feat] Add support for hit ratio in yaml * [Fix] Fix load only first rank in non mla scene * [fix] refuse monkey patch (#383) refuse monkey patch * [bugfix] fix gqa bug (#384) fix gqa bug * [bugfix] end == 0 bug (#385) fix end == 0 bug * [feature] optimize generate_tensor (#396) optimize generate_tensor * [Fix] fix mla bug when no broadcast in wait for save (#398) * [feat]adapt GQA & modify config.yaml (#407) * adapt GQA & modify config.yaml * move process to UCMDirectConnector * fix comment * modify hash function * fix style * code style and modify hash * init parent_block_hash_value * [feat]Adapt vllm_ascend_0110 and Add configurable options (#415) * Adapt vllm_ascend_0110 and Add configurable options * avoid type conversion in init kvcache * [patch]seprate sparse patch (#417) seprate spase patch Co-authored-by: lijiachen19 <lijiachen19@huawei.com> * [bugfix]Support tensor parallelism across servers (#420) Support tensor parallelism across servers * [Feat] UCM supports metrics display online via Grafana and Promethues (#414) * [Feat] Build metrics frame * [Feat]add metrics(ucm_obser.py + metrics_configs.yaml) * [Feat] Implementation of metrics logger on the C++ side for storing and retrieving stats * [Fix] Provide simple grafana and fix bugs * [feat] change the log position of UCM metrics * [fix]modify grafana.json * [Feat] UCM supports metrics display online via Grafana and Promethues * [Fix] Remove configs to examples and add liscense --------- Co-authored-by: flesher0813 <1208954694@qq.com> Co-authored-by: hero<tianxuehan@huawei.com> * [feat]Merge develop to dev-ucm-v1 and fix code style (#428) * [fix] fix sparse attention (#397) fix ascend attention Co-authored-by: lijiachen19 <lijiachen19@huawei.com> * [opt] Share Infra implementation and unify status codes (#399) share infra module Co-authored-by: Fang Run <Fang_Run@126.com> * [bugfix] Fix ESA to be compatible with the latest NFSStore. (#401) fix esa to adapt latest NFSStore * release v0.1.0rc4 (#402) Co-authored-by: lijiachen19 <lijiachen19@huawei.com> * [opt] Remove unused cc impl of dramstore (#406) remove unused cc impl of dramstore * [Fix]remove dram docs and modify quick-start doc (#411) * [Fix]remove dram docs and modify quick-start doc * modify index.md --------- Co-authored-by: t00939662 <tianxuehan@huawei.com> * [Feature] Added performance testing tool based on the PyTest testing framework (#295) Performance testing tool based on the PyTest testing framework. * [Misc] Add cpp-linter.yml (#422) * [docs]add metrics doc (#416) * [docs]add metrics doc * modify metrics.md * modify metrics.md --------- Co-authored-by: t00939662 <tianxuehan@huawei.com> * [perf] Modify CUDA SIMD and add Triton hash encoder (#408) * fix cpp code style --------- Co-authored-by: Lijiachen1018 <30387633+Lijiachen1018@users.noreply.github.com> Co-authored-by: lijiachen19 <lijiachen19@huawei.com> Co-authored-by: Mag1c.H <hemajun815@163.com> Co-authored-by: Fang Run <Fang_Run@126.com> Co-authored-by: MaxWang <wangwenxin21@huawei.com> Co-authored-by: hero0307 <tianxuehan0307@163.com> Co-authored-by: t00939662 <tianxuehan@huawei.com> Co-authored-by: ML <85485147+Menglths@users.noreply.github.com> Co-authored-by: ShiXiaolei <indirashi@163.com> * add env variable ENABLE_SPARSE (#430) Co-authored-by: lijiachen19 <lijiachen19@huawei.com> * Fix(patch): fix patch for vllm-ascend (#433) Fix(patch): fix patch for vllm-ascend volcengine/verl#2564 Co-authored-by: lijiachen19 <lijiachen19@huawei.com> * [bugfix] fix accuracy problem when chunked prefill (#438) * fix accuracy problem when chunked prefill * [bugfix]fix num_schedule-tokens=1 (#442) * fix num_schedule-tokens=1 * Simplify the code * [fix]: Fix sparse patch (#444) Fix sparse patch Co-authored-by: lijiachen <lijiachen19@huawei.com> * [bugfix] The Metrics module uses a non-existent variable self.rank (#445) * [Feature]Add an access bandwidth test script for ucm_connector (#418) * Add an access bandwidth test script for 'ucm_connector' * [bugfix]adapt vllm0.9.1 (#446) adapt vllm0.9.1 * [Fix]Set the multiprocessing start method of the test tool to 'spawn'. (#447) Set the multiprocessing start method of the test tool to 'spawn' and add NPU cleanup * [fix] Adapt all sparse-attention methods to the new connector. (#441) * sparse to adapt new connector * Adapt the YAML configuration * [docs] renew docs for v1 (#448) renew docs for v1 Co-authored-by: lijiachen19 <lijiachen19@huawei.com> * set version to 0.1.0 (#450) * [Feature] GSA adapt nfsStore (#451) * adapt nfsstore * fix codestyle --------- Co-authored-by: ygwpz <543529648@qq.com> Co-authored-by: harrisonyhq <harrisonyhq@gmail.com> Co-authored-by: qyh111 <qiuyuhao1@huawei.com> Co-authored-by: lijiachen19 <lijiachen19@huawei.com> Co-authored-by: sumingZero <58885253+sumingZero@users.noreply.github.com> Co-authored-by: flesher0813 <1208954694@qq.com> Co-authored-by: Mag1c.H <hemajun815@163.com> Co-authored-by: Fang Run <Fang_Run@126.com> Co-authored-by: MaxWang <wangwenxin21@huawei.com> Co-authored-by: hero0307 <tianxuehan0307@163.com> Co-authored-by: t00939662 <tianxuehan@huawei.com> Co-authored-by: ML <85485147+Menglths@users.noreply.github.com> Co-authored-by: ShiXiaolei <indirashi@163.com> Co-authored-by: zhou-haitao <74044944+zhou-haitao@users.noreply.github.com> Co-authored-by: zbb200819 <1130072360@qq.com>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Purpose
What this PR does / why we need it?
Adapt to the new connector by updating the hash generation and layer-offset computation.
Modifications
Does this PR introduce any user-facing change?
Test
How was this patch tested?

python examples/offline_inference_esa.py