Skip to content

Commit

Permalink
Merge branch 'HabanaAI:habana_main' into habana_main
Browse files Browse the repository at this point in the history
  • Loading branch information
nageshdn authored Jan 7, 2025
2 parents 84d6620 + 2d24be7 commit f67e5bc
Show file tree
Hide file tree
Showing 12 changed files with 379 additions and 240 deletions.
2 changes: 1 addition & 1 deletion .github/CODEOWNERS
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# See https://help.github.com/articles/about-codeowners/
# for more info about CODEOWNERS file

* @kzawora-intel @madamczykhabana @michalkuligowski @mgawarkiewicz
* @kzawora-intel @madamczykhabana @michalkuligowski @mgawarkiewicz @vivekgoe
2 changes: 1 addition & 1 deletion .jenkins/test_config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -57,4 +57,4 @@ stages:
command: TORCH_COMPILE_DISABLE=true VLLM_CONTIGUOUS_PA=false VLLM_SKIP_WARMUP=True pytest -v tests/spec_decode/e2e/test_medusa_correctness.py::test_medusa_e2e_greedy_correctness
- name: gsm8k_small_g2_tp1_eagle_spec_decode
flavor: g2
command: TORCH_COMPILE_DISABLE=true VLLM_CONTIGUOUS_PA=false VLLM_SKIP_WARMUP=True pytest -v tests/spec_decode/e2e/test_eagle_correctness.py::test_eagle_e2e_greedy_correctness
command: VLLM_COS_SIN_RECOMPUTE=true TORCH_COMPILE_DISABLE=true VLLM_CONTIGUOUS_PA=false VLLM_SKIP_WARMUP=True pytest -v tests/spec_decode/e2e/test_eagle_correctness.py::test_eagle_e2e_greedy_correctness
1 change: 1 addition & 0 deletions tests/lora/conftest.py
Original file line number Diff line number Diff line change
Expand Up @@ -64,6 +64,7 @@ def cleanup_fixture(should_do_global_cleanup_after_test: bool):

@pytest.fixture
def dist_init():
import habana_frameworks.torch.hpu # noqa: F401
temp_file = tempfile.mkstemp()[1]
backend_type = "hccl" if current_platform.is_hpu() else "nccl"
init_distributed_environment(
Expand Down
Loading

0 comments on commit f67e5bc

Please sign in to comment.