Skip to content

[BugFix]add int8 cache dtype when using attention quantization #126

[BugFix]add int8 cache dtype when using attention quantization

[BugFix]add int8 cache dtype when using attention quantization #126

Triggered via pull request February 21, 2025 10:47
Status Success
Total duration 32m 9s
Artifacts 1

image.yml

on: pull_request
vllm-ascend image
31m 58s
vllm-ascend image
Fit to window
Zoom out
Zoom in

Artifacts

Produced during runtime
Name Size
vllm-project~vllm-ascend~9UYGZ2.dockerbuild
125 KB