Skip to content

Commit

Permalink
rename and add tests
Browse files Browse the repository at this point in the history
  • Loading branch information
kylesayrs committed Aug 30, 2024
1 parent 143ec4b commit bfa3ff6
Show file tree
Hide file tree
Showing 4 changed files with 26 additions and 2 deletions.
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
cadence: "nightly"
test_type: "regression"
model_stub: "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T"
new_recipe: "tests/llmcompressor/transformers/compression/recipes/new_quant_actorder.yaml"
new_recipe: "tests/llmcompressor/transformers/compression/recipes/new_quant_actorder_group.yaml"
ppl_threshold: 20
Original file line number Diff line number Diff line change
@@ -0,0 +1,5 @@
cadence: "nightly"
test_type: "regression"
model_stub: "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T"
new_recipe: "tests/llmcompressor/transformers/compression/recipes/new_quant_actorder_weight.yaml"
ppl_threshold: 20
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ test_stage:
symmetric: False
strategy: "group"
group_size: 128
actorder: True
actorder: "group"
input_activations: null
output_activations: null
targets: ["Linear"]
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,19 @@
test_stage:
quant_modifiers:
QuantizationModifier:
ignore: ["lm_head", "model.layers.0.mlp.down_proj"]
config_groups:
group_0:
weights:
num_bits: 4
type: "int"
symmetric: False
strategy: "group"
group_size: 128
actorder: "weight"
input_activations: null
output_activations: null
targets: ["Linear"]
GPTQModifier:
block_size: 128
sequential_update: False

0 comments on commit bfa3ff6

Please sign in to comment.