Skip to content

Commit 4399503

Browse files
Sara Adkinsdbogunowicz
Sara Adkins
andauthored
Update Test Recipes for Latest Modifier Changes (#2326)
* update recipes for new modifier * fixes --------- Co-authored-by: dbogunowicz <[email protected]>
1 parent 813033e commit 4399503

File tree

4 files changed

+10
-63
lines changed

4 files changed

+10
-63
lines changed

Diff for: tests/sparseml/transformers/obcq/recipes/quant.yaml

+1-26
Original file line numberDiff line numberDiff line change
@@ -6,32 +6,7 @@ test_stage:
66
[["re:.*q_proj", "re:.*k_proj", "re:.*v_proj"], "re:.*input_layernorm"],
77
[["re:.*gate_proj", "re:.*up_proj"], "re:.*post_attention_layernorm"]
88
]
9-
LegacyQuantizationModifier:
10-
ignore:
11-
- LlamaRotaryEmbedding
12-
- LlamaRMSNorm
13-
- SiLU
14-
- model.layers.0.mlp.down_proj
15-
- model.layers.1.mlp.down_proj
16-
- model.layers.2.mlp.down_proj
17-
- model.layers.3.mlp.down_proj
18-
- model.layers.4.mlp.down_proj
19-
- model.layers.5.mlp.down_proj
20-
scheme_overrides:
21-
Embedding:
22-
input_activations: null
23-
weights:
24-
num_bits: 8
25-
symmetric: False
269
GPTQModifier:
2710
block_size: 128
2811
sequential_update: False
29-
percdamp: 0.01
30-
targets: [
31-
"model.layers.0",
32-
"model.layers.1",
33-
"model.layers.2",
34-
"model.layers.3",
35-
"model.layers.4",
36-
"model.layers.5"
37-
]
12+
percdamp: 0.01
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,11 @@
11
test_stage:
22
obcq_modifiers:
3+
SparseGPTModifier:
4+
sparsity: 0.5
5+
block_size: 128
6+
sequential_update: False
7+
percdamp: 0.01
8+
mask_structure: "0:0"
39
SmoothQuantModifier:
410
smoothing_strength: 0.5
511
mappings: [
@@ -11,13 +17,6 @@ test_stage:
1117
- LlamaRotaryEmbedding
1218
- LlamaRMSNorm
1319
- SiLU
14-
- model.layers.0.mlp.down_proj
15-
- model.layers.1.mlp.down_proj
16-
- model.layers.2.mlp.down_proj
17-
- model.layers.3.mlp.down_proj
18-
- model.layers.4.mlp.down_proj
19-
- model.layers.5.mlp.down_proj
20-
post_oneshot_calibration: True
2120
scheme_overrides:
2221
Embedding:
2322
input_activations: null
@@ -27,26 +26,4 @@ test_stage:
2726
GPTQModifier:
2827
block_size: 128
2928
sequential_update: False
30-
percdamp: 0.01
31-
targets: [
32-
"model.layers.0",
33-
"model.layers.1",
34-
"model.layers.2",
35-
"model.layers.3",
36-
"model.layers.4",
37-
"model.layers.5"
38-
]
39-
SparseGPTModifier:
40-
sparsity: 0.5
41-
block_size: 128
42-
sequential_update: False
43-
percdamp: 0.01
44-
mask_structure: "0:0"
45-
targets: [
46-
"model.layers.0",
47-
"model.layers.1",
48-
"model.layers.2",
49-
"model.layers.3",
50-
"model.layers.4",
51-
"model.layers.5"
52-
]
29+
percdamp: 0.01

Diff for: tests/sparseml/transformers/obcq/recipes/sparse.yaml

+2-3
Original file line numberDiff line numberDiff line change
@@ -5,9 +5,8 @@ test_stage:
55
block_size: 128
66
sequential_update: False
77
percdamp: 0.01
8-
mask_structure: "0:0"
98
targets: [
109
"model.layers.0",
1110
"model.layers.1",
12-
"lm_head"
13-
]
11+
]
12+
mask_structure: "0:0"

Diff for: tests/sparseml/transformers/obcq/test_obcq_sparsity.py

-4
Original file line numberDiff line numberDiff line change
@@ -60,8 +60,6 @@ def test_sparsities(self):
6060

6161
model = get_session_model()
6262

63-
lm_head_sparsity = tensor_sparsity(model.lm_head.weight)
64-
assert math.isclose(lm_head_sparsity.item(), self.sparsity, rel_tol=1e-4)
6563
layer_1_sparse = tensor_sparsity(model.model.layers[1].self_attn.k_proj.weight)
6664
assert math.isclose(layer_1_sparse.item(), self.sparsity, rel_tol=1e-4)
6765
layer_2_dense = tensor_sparsity(model.model.layers[2].self_attn.k_proj.weight)
@@ -118,8 +116,6 @@ def test_sparsities_gpu(self):
118116

119117
model = get_session_model()
120118

121-
lm_head_sparsity = tensor_sparsity(model.lm_head.weight)
122-
assert math.isclose(lm_head_sparsity.item(), self.sparsity, rel_tol=1e-4)
123119
layer_1_sparse = tensor_sparsity(model.model.layers[1].self_attn.k_proj.weight)
124120
assert math.isclose(layer_1_sparse.item(), self.sparsity, rel_tol=1e-4)
125121
layer_2_dense = tensor_sparsity(model.model.layers[2].self_attn.k_proj.weight)

0 commit comments

Comments
 (0)