Skip to content

Commit 33dee65

Browse files
fix test_exporters_cli_int4_with_local_model_and_default_config
1 parent 6c4d667 commit 33dee65

File tree

2 files changed

+20
-11
lines changed

2 files changed

+20
-11
lines changed

optimum/intel/openvino/configuration.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -90,8 +90,8 @@ class OVQuantizationMethod(str, Enum):
9090
"bigscience/bloomz-560m": {
9191
"bits": 4,
9292
"sym": True,
93-
"group_size": 64,
94-
"ratio": 0.8,
93+
"group_size": 128,
94+
"ratio": 1.0,
9595
"dataset": "wikitext2",
9696
"quant_method": OVQuantizationMethod.AWQ,
9797
},

tests/openvino/test_exporters_cli.py

Lines changed: 18 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717
from tempfile import TemporaryDirectory
1818

1919
from parameterized import parameterized
20-
from transformers import AutoModelForCausalLM
20+
from transformers import AutoModelForCausalLM, AutoTokenizer
2121
from utils_tests import (
2222
_ARCHITECTURES_TO_EXPECTED_INT8,
2323
MODEL_NAMES,
@@ -253,10 +253,12 @@ def test_exporters_cli_int4(self, task: str, model_type: str, option: str, expec
253253

254254
def test_exporters_cli_int4_with_local_model_and_default_config(self):
255255
with TemporaryDirectory() as tmpdir:
256-
pt_model = AutoModelForCausalLM.from_pretrained(MODEL_NAMES["bloom"])
257-
# overload for matching with default configuration
258-
pt_model.config._name_or_path = "bigscience/bloomz-7b1"
256+
model_id = "bigscience/bloomz-560m"
257+
tokenizer = AutoTokenizer.from_pretrained(model_id)
258+
pt_model = AutoModelForCausalLM.from_pretrained(model_id)
259+
tokenizer.save_pretrained(tmpdir)
259260
pt_model.save_pretrained(tmpdir)
261+
260262
subprocess.run(
261263
f"optimum-cli export openvino --model {tmpdir} --task text-generation-with-past --weight-format int4 {tmpdir}",
262264
shell=True,
@@ -267,16 +269,23 @@ def test_exporters_cli_int4_with_local_model_and_default_config(self):
267269
rt_info = model.model.get_rt_info()
268270
self.assertTrue("nncf" in rt_info)
269271
self.assertTrue("weight_compression" in rt_info["nncf"])
270-
default_config = _DEFAULT_4BIT_CONFIGS["bigscience/bloomz-7b1"]
271272
model_weight_compression_config = rt_info["nncf"]["weight_compression"]
272-
sym = default_config.pop("sym", False)
273+
274+
default_config = _DEFAULT_4BIT_CONFIGS[model_id]
273275
bits = default_config.pop("bits", None)
274276
self.assertEqual(bits, 4)
275277

276-
mode = f'int{bits}_{"sym" if sym else "asym"}'
277-
default_config["mode"] = mode
278+
sym = default_config.pop("sym", False)
279+
default_config["mode"] = f'int{bits}_{"sym" if sym else "asym"}'
280+
281+
quant_method = default_config.pop("quant_method", None)
282+
default_config["awq"] = quant_method == "awq"
283+
default_config["gptq"] = quant_method == "gptq"
284+
285+
default_config.pop("dataset", None)
286+
278287
for key, value in default_config.items():
279-
self.assertTrue(key in model_weight_compression_config)
288+
self.assertIn(key, model_weight_compression_config)
280289
self.assertEqual(
281290
model_weight_compression_config[key].value,
282291
str(value),

0 commit comments

Comments
 (0)