Skip to content

[BugFix]Add int8 cache dtype when using ascend attention quantization #108

[BugFix]Add int8 cache dtype when using ascend attention quantization

[BugFix]Add int8 cache dtype when using ascend attention quantization #108

Triggered via pull request February 21, 2025 01:44
Status Success
Total duration 30m 29s
Artifacts 1

image.yml

on: pull_request
vllm-ascend image
30m 20s
vllm-ascend image
Fit to window
Zoom out
Zoom in

Artifacts

Produced during runtime
Name Size
vllm-project~vllm-ascend~TA1VPH.dockerbuild
249 KB