From 07dcfaacf396a15ed58070ae2b78b52c01f52b94 Mon Sep 17 00:00:00 2001 From: Dipika Date: Thu, 20 Feb 2025 19:58:16 +0000 Subject: [PATCH] update example --- .../llama7b_sparse_w4a16.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/examples/quantization_2of4_sparse_w4a16/llama7b_sparse_w4a16.py b/examples/quantization_2of4_sparse_w4a16/llama7b_sparse_w4a16.py index 884952f5e..2cca58815 100644 --- a/examples/quantization_2of4_sparse_w4a16/llama7b_sparse_w4a16.py +++ b/examples/quantization_2of4_sparse_w4a16/llama7b_sparse_w4a16.py @@ -54,8 +54,8 @@ warmup_ratio=warmup_ratio, ) logger.info( - "Note: llcompressor does not currently support running ", - "compressed models in the marlin-24 format. The model ", - "produced from this example can be run on vLLM with ", - "dtype=torch.float16", + "llmcompressor does not currently support running compressed models in the marlin24 format." # noqa +) +logger.info( + "The model produced from this example can be run on vLLM with dtype=torch.float16" )