From 3479a7f1f40c60e4d4b8b5fb045ea072d0b73691 Mon Sep 17 00:00:00 2001 From: dbogunowicz <97082108+dbogunowicz@users.noreply.github.com> Date: Mon, 10 Jun 2024 18:19:04 +0200 Subject: [PATCH 1/3] Update tiny_llama_consec_runs.yaml --- .../obcq/obcq_configs/consec_runs/tiny_llama_consec_runs.yaml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tests/sparseml/transformers/obcq/obcq_configs/consec_runs/tiny_llama_consec_runs.yaml b/tests/sparseml/transformers/obcq/obcq_configs/consec_runs/tiny_llama_consec_runs.yaml index e1073aae904..44d50d2fdd0 100644 --- a/tests/sparseml/transformers/obcq/obcq_configs/consec_runs/tiny_llama_consec_runs.yaml +++ b/tests/sparseml/transformers/obcq/obcq_configs/consec_runs/tiny_llama_consec_runs.yaml @@ -1,6 +1,6 @@ cadence: "commit" test_type: "sanity" -model: "Xenova/llama2.c-stories15M" +model: "HuggingFaceM4/tiny-random-LlamaForCausalLM" dataset: open_platypus first_recipe: "tests/sparseml/transformers/obcq/recipes/quant_and_sparse.yaml" -second_recipe: "tests/sparseml/transformers/obcq/recipes/additional_sparsity.yaml" \ No newline at end of file +second_recipe: "tests/sparseml/transformers/obcq/recipes/additional_sparsity.yaml" From aa8bbd4d1e895c8203b9b768bca6a71a0e120aef Mon Sep 17 00:00:00 2001 From: dbogunowicz <97082108+dbogunowicz@users.noreply.github.com> Date: Mon, 10 Jun 2024 18:20:06 +0200 Subject: [PATCH 2/3] Update tiny_llama_repeat_quant.yaml --- .../obcq_configs/repeat_quants/tiny_llama_repeat_quant.yaml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tests/sparseml/transformers/obcq/obcq_configs/repeat_quants/tiny_llama_repeat_quant.yaml b/tests/sparseml/transformers/obcq/obcq_configs/repeat_quants/tiny_llama_repeat_quant.yaml index a91b7b4d56a..7545d40512e 100644 --- a/tests/sparseml/transformers/obcq/obcq_configs/repeat_quants/tiny_llama_repeat_quant.yaml +++ b/tests/sparseml/transformers/obcq/obcq_configs/repeat_quants/tiny_llama_repeat_quant.yaml @@ -1,6 +1,6 @@ cadence: "commit" test_type: "sanity" -model: "Xenova/llama2.c-stories15M" +model: "HuggingFaceM4/tiny-random-LlamaForCausalLM" dataset: open_platypus first_recipe: | first_stage: @@ -22,4 +22,4 @@ second_recipe: | - LlamaRotaryEmbedding - LlamaRMSNorm - SiLU - - Embedding \ No newline at end of file + - Embedding From 46f11f8462956ae3d6dc18bf8a8a8b8c2d6bc1b1 Mon Sep 17 00:00:00 2001 From: dbogunowicz <97082108+dbogunowicz@users.noreply.github.com> Date: Tue, 11 Jun 2024 11:59:21 +0200 Subject: [PATCH 3/3] Update test_consecutive_runs.py --- tests/sparseml/transformers/obcq/test_consecutive_runs.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/sparseml/transformers/obcq/test_consecutive_runs.py b/tests/sparseml/transformers/obcq/test_consecutive_runs.py index 7bcfc8b7efe..04b78ec82b8 100644 --- a/tests/sparseml/transformers/obcq/test_consecutive_runs.py +++ b/tests/sparseml/transformers/obcq/test_consecutive_runs.py @@ -114,7 +114,7 @@ def setUp(self): self.output_second = Path(self.output) / "test_2" def test_consecutive_runs_small(self): - self._test_consecutive_runs(tolerance=1e-3) + self._test_consecutive_runs(tolerance=1e-1) @requires_gpu