Skip to content

[BugFix]add int8 cache dtype when using attention quantization #112

[BugFix]add int8 cache dtype when using attention quantization

[BugFix]add int8 cache dtype when using attention quantization #112

Triggered via pull request February 21, 2025 02:11
Status Success
Total duration 36m 16s
Artifacts 1

image.yml

on: pull_request
vllm-ascend image
36m 7s
vllm-ascend image
Fit to window
Zoom out
Zoom in

Artifacts

Produced during runtime
Name Size
vllm-project~vllm-ascend~W1VCGO.dockerbuild
175 KB