Skip to content

[BugFix]Add int8 cache dtype when using ascend attention quantization #109

[BugFix]Add int8 cache dtype when using ascend attention quantization

[BugFix]Add int8 cache dtype when using ascend attention quantization #109

Triggered via pull request February 21, 2025 01:47
Status Success
Total duration 27m 44s
Artifacts 1

image.yml

on: pull_request
vllm-ascend image
27m 37s
vllm-ascend image
Fit to window
Zoom out
Zoom in

Artifacts

Produced during runtime
Name Size
vllm-project~vllm-ascend~XZ4DJU.dockerbuild
244 KB