Skip to content

[BugFix]add int8 cache dtype when using attention quantization #113

[BugFix]add int8 cache dtype when using attention quantization

[BugFix]add int8 cache dtype when using attention quantization #113

Triggered via pull request February 21, 2025 02:20
Status Success
Total duration 42m 5s
Artifacts 1

image.yml

on: pull_request
vllm-ascend image
41m 57s
vllm-ascend image
Fit to window
Zoom out
Zoom in

Artifacts

Produced during runtime
Name Size
vllm-project~vllm-ascend~9D5GX2.dockerbuild
175 KB