diff --git a/tests/e2e/vLLM/test_vllm.py b/tests/e2e/vLLM/test_vllm.py index 5b475a5b7..f524ff10a 100644 --- a/tests/e2e/vLLM/test_vllm.py +++ b/tests/e2e/vLLM/test_vllm.py @@ -13,7 +13,6 @@ from tests.e2e.e2e_utils import run_oneshot_for_e2e_testing from tests.examples.utils import requires_gpu_count -""" try: from vllm import LLM, SamplingParams @@ -21,7 +20,7 @@ except ImportError: vllm_installed = False logger.warning("vllm is not installed. This test will be skipped") -""" + HF_MODEL_HUB_NAME = "nm-testing" TEST_DATA_FILE = os.environ.get("TEST_DATA_FILE", "") @@ -43,7 +42,7 @@ def record_config_file(record_testsuite_property: Callable[[str, object], None]) # Will run each test case in its own process through run_tests.sh # emulating vLLM CI testing @requires_gpu_count(1) -# @pytest.mark.skipif(not vllm_installed, reason="vLLM is not installed, skipping test") +@pytest.mark.skipif(not vllm_installed, reason="vLLM is not installed, skipping test") class TestvLLM: """ The following test quantizes a model using a preset scheme or recipe, @@ -139,7 +138,6 @@ def test_vllm(self): folder_path=self.save_dir, ) - """ logger.info("================= RUNNING vLLM =========================") sampling_params = SamplingParams(temperature=0.80, top_p=0.95) @@ -162,7 +160,6 @@ def test_vllm(self): logger.info(generated_text) self.tear_down() - """ def tear_down(self): if self.save_dir is not None: