Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
174 changes: 0 additions & 174 deletions .azure-pipelines/model-test.yml

This file was deleted.

162 changes: 0 additions & 162 deletions .azure-pipelines/ut-basic.yml
Original file line number Diff line number Diff line change
Expand Up @@ -130,165 +130,3 @@ stages:
utScriptFileName: "run_basic_others"
uploadPath: $(UPLOAD_PATH)
utArtifact: "ut_others"

- stage: Adaptor_base
displayName: Unit Test FWKs adaptor baseline
dependsOn: []
jobs:
- job:
displayName: Test FWKs adaptor baseline
continueOnError: true
steps:
- template: template/ut-template.yml
parameters:
dockerConfigName: "gitCloneDockerConfig"
utScriptFileName: "run_basic_adaptor"
uploadPath: $(UPLOAD_PATH)
utArtifact: "ut-base_adaptor"
repo: $(REPO)

- stage: API_base
displayName: Unit Test User facing API baseline
dependsOn: []
jobs:
- job:
displayName: Test User facing API baseline
continueOnError: true
steps:
- template: template/ut-template.yml
parameters:
dockerConfigName: "gitCloneDockerConfig"
utScriptFileName: "run_basic_api"
uploadPath: $(UPLOAD_PATH)
utArtifact: "ut-base_api"
repo: $(REPO)

- stage: Pruning_base
displayName: Unit Test Pruning baseline
dependsOn: []
jobs:
- job:
displayName: Test PyTorch Pruning baseline
continueOnError: true
steps:
- template: template/ut-template.yml
parameters:
dockerConfigName: "gitCloneDockerConfig"
utScriptFileName: "run_basic_pt_pruning"
uploadPath: $(UPLOAD_PATH)
utArtifact: "ut-base_pt-pruning"
repo: $(REPO)
- job:
displayName: Test TensorFlow Pruning baseline
continueOnError: true
steps:
- template: template/ut-template.yml
parameters:
dockerConfigName: "gitCloneDockerConfig"
utScriptFileName: "run_basic_tf_pruning"
uploadPath: $(UPLOAD_PATH)
utArtifact: "ut-base_tf-pruning"
repo: $(REPO)

- stage: TFNewAPI_base
displayName: Unit Test TF newAPI baseline
dependsOn: []
jobs:
- job:
displayName: Test TF newAPI baseline
continueOnError: true
steps:
- template: template/ut-template.yml
parameters:
dockerConfigName: "gitCloneDockerConfig"
utScriptFileName: "run_basic_adaptor_tfnewapi"
uploadPath: $(UPLOAD_PATH)
utArtifact: "ut-base_tfnewapi"
repo: $(REPO)

- stage: ITEX_base
displayName: Unit Test ITEX baseline
dependsOn: []
jobs:
- job:
displayName: Test ITEX baseline
continueOnError: true
steps:
- template: template/ut-template.yml
parameters:
dockerConfigName: "gitCloneDockerConfig"
utScriptFileName: "run_basic_itex"
uploadPath: $(UPLOAD_PATH)
utArtifact: "ut-base_itex"
repo: $(REPO)

- stage: Others_base
displayName: Unit Test other cases baseline
dependsOn: []
jobs:
- job:
displayName: Test other cases baseline
continueOnError: true
steps:
- template: template/ut-template.yml
parameters:
dockerConfigName: "gitCloneDockerConfig"
utScriptFileName: "run_basic_others"
uploadPath: $(UPLOAD_PATH)
utArtifact: "ut-base_others"
repo: $(REPO)

- stage: Coverage
displayName: "Coverage Compare"
pool:
vmImage: "ubuntu-latest"
dependsOn: [Adaptor, API, Pruning, TFNewAPI, ITEX, Others, Adaptor_base, API_base, Pruning_base, TFNewAPI_base, ITEX_base, Others_base]
jobs:
- job: CollectDatafiles
steps:
- script: |
if [[ ! $(docker images | grep -i ${IMAGE_NAME}:${IMAGE_TAG}) ]]; then
docker build -f ${BUILD_SOURCESDIRECTORY}/.azure-pipelines/docker/Dockerfile.devel -t ${IMAGE_NAME}:${IMAGE_TAG} .
fi
docker images | grep -i ${IMAGE_NAME}
if [[ $? -ne 0 ]]; then
echo "NO Such Repo"
exit 1
fi
displayName: "Build develop docker image"

- task: DownloadPipelineArtifact@2
inputs:
artifact:
patterns: '*_coverage/.coverage.*'
path: $(DOWNLOAD_PATH)

- script: |
echo "--- create container ---"
docker run -d -it --name="collectLogs" -v ${BUILD_SOURCESDIRECTORY}:/neural-compressor ${IMAGE_NAME}:${IMAGE_TAG} /bin/bash
echo "--- docker ps ---"
docker ps
echo "--- collect logs ---"
docker exec collectLogs /bin/bash +x -c "cd /neural-compressor/.azure-pipelines/scripts \
&& bash install_nc.sh \
&& bash ut/collect_log.sh"
displayName: "Collect UT Coverage"

- task: PublishCodeCoverageResults@2
inputs:
summaryFileLocation: $(Build.SourcesDirectory)/log_dir/coverage_PR/coverage.xml

- task: PublishPipelineArtifact@1
condition: succeededOrFailed()
inputs:
targetPath: $(UPLOAD_PATH)
artifact: $(ARTIFACT_NAME)
publishLocation: "pipeline"

- task: Bash@3
condition: always()
inputs:
targetType: "inline"
script: |
docker exec collectLogs bash -c "rm -fr /neural-compressor/* && rm -fr /neural-compressor/.* || true"
displayName: "Docker clean up"
17 changes: 16 additions & 1 deletion examples/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,22 @@ Intel® Neural Compressor validated examples with multiple compression technique
</tr>
</thead>
<tbody>

<tr>
<td>Llama-4-Scout-17B-16E-Instruct</td>
<td>Multimodal Modeling</td>
<td>Quantization (MXFP4)</td>
<td><a href="./pytorch/multimodal-modeling/quantization/auto_round/llama4">link</a></td>
</tr>
<tr>
<td rowspan="2">Llama-3.3-70B-Instruct</td>
<td rowspan="2">Natural Language Processing</td>
<td>Mixed Precision (MXFP4+MXFP8)</td>
<td><a href="./pytorch/nlp/huggingface_models/language-modeling/quantization/mix-precision#mix-precision-quantization-mxfp4--mxfp8">link</a></td>
</tr>
<tr>
<td>Quantization (MXFP4/MXFP8/NVFP4)</td>
<td><a href="./pytorch/nlp/huggingface_models/language-modeling/quantization/mix-precision#mxfp4--mxfp8">link</a></td>
</tr>
<tr>
<td rowspan="2">gpt_j</td>
<td rowspan="2">Natural Language Processing</td>
Expand Down