Skip to content

[CUDA] Add SparseAttention kernel for sm=75 #29156

[CUDA] Add SparseAttention kernel for sm=75

[CUDA] Add SparseAttention kernel for sm=75 #29156

Triggered via pull request May 1, 2024 23:47
Status Failure
Total duration 23m 59s
Artifacts

lint.yml

on: pull_request
Fit to window
Zoom out
Zoom in

Annotations

1 error and 17 warnings
Lint C++
reviewdog: Too many results (annotations) in diff. You may miss some annotations due to GitHub limitation for annotation created by logging command. Please check GitHub Actions log console to see all results. Limitation: - 10 warning annotations and 10 error annotations per step - 50 annotations per job (sum of annotations from all the steps) - 50 annotations per run (separate from the job annotations, these annotations aren't created by users) Source: https://github.com/orgs/community/discussions/26680#discussioncomment-3252835
Python format
Node.js 16 actions are deprecated. Please update the following actions to use Node.js 20: actions-rs/toolchain@v1. For more information see: https://github.blog/changelog/2023-09-22-github-actions-transitioning-from-node-16-to-node-20/.
Python format
The following actions uses node12 which is deprecated and will be forced to run on node16: actions-rs/toolchain@v1. For more info: https://github.blog/changelog/2023-06-13-github-actions-all-actions-will-run-on-node16-instead-of-node12-by-default/
Python format
The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/
Python format
The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/
Python format
The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/
Python format
The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/
Python format
This run of the CodeQL Action does not have permission to access Code Scanning API endpoints. As a result, it will not be opted into any experimental features. This could be because the Action is running on a pull request from a fork. If not, please ensure the Action has the 'security-events: write' permission. Details: Resource not accessible by integration
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_dispatcher_fp16_sm75.h#L179
[cpplint] reported by reviewdog 🐶 Using C-style cast. Use static_cast<int>(...) instead [readability/casting] [4] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_dispatcher_fp16_sm75.h:179: Using C-style cast. Use static_cast<int>(...) instead [readability/casting] [4]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_dispatcher_fp16_sm75.h#L183
[cpplint] reported by reviewdog 🐶 Using C-style cast. Use static_cast<int>(...) instead [readability/casting] [4] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_dispatcher_fp16_sm75.h:183: Using C-style cast. Use static_cast<int>(...) instead [readability/casting] [4]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_0_d64_2_sm75.cc#L18
[cpplint] reported by reviewdog 🐶 Lines should be <= 120 characters long [whitespace/line_length] [2] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_0_d64_2_sm75.cc:18: Lines should be <= 120 characters long [whitespace/line_length] [2]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_0_d64_2_sm75.cc#L29
[cpplint] reported by reviewdog 🐶 Using C-style cast. Use reinterpret_cast<void*>(...) instead [readability/casting] [4] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_0_d64_2_sm75.cc:29: Using C-style cast. Use reinterpret_cast<void*>(...) instead [readability/casting] [4]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_0_d64_2_sm75.cc#L32
[cpplint] reported by reviewdog 🐶 Lines should be <= 120 characters long [whitespace/line_length] [2] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_0_d64_2_sm75.cc:32: Lines should be <= 120 characters long [whitespace/line_length] [2]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_1_d64_2_sm75.cc#L18
[cpplint] reported by reviewdog 🐶 Lines should be <= 120 characters long [whitespace/line_length] [2] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_1_d64_2_sm75.cc:18: Lines should be <= 120 characters long [whitespace/line_length] [2]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_1_d64_2_sm75.cc#L29
[cpplint] reported by reviewdog 🐶 Using C-style cast. Use reinterpret_cast<void*>(...) instead [readability/casting] [4] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_1_d64_2_sm75.cc:29: Using C-style cast. Use reinterpret_cast<void*>(...) instead [readability/casting] [4]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_1_d64_2_sm75.cc#L32
[cpplint] reported by reviewdog 🐶 Lines should be <= 120 characters long [whitespace/line_length] [2] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_0_n64_1_d64_2_sm75.cc:32: Lines should be <= 120 characters long [whitespace/line_length] [2]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_1_n64_0_d64_2_sm75.cc#L18
[cpplint] reported by reviewdog 🐶 Lines should be <= 120 characters long [whitespace/line_length] [2] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_1_n64_0_d64_2_sm75.cc:18: Lines should be <= 120 characters long [whitespace/line_length] [2]
Lint C++: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_1_n64_0_d64_2_sm75.cc#L29
[cpplint] reported by reviewdog 🐶 Using C-style cast. Use reinterpret_cast<void*>(...) instead [readability/casting] [4] Raw Output: onnxruntime/contrib_ops/cuda/sparse/sparse_attention_v1/sparse_attention_v1_fp16_m16_1_n64_0_d64_2_sm75.cc:29: Using C-style cast. Use reinterpret_cast<void*>(...) instead [readability/casting] [4]