Skip to content

Commit

Permalink
remove
Browse files Browse the repository at this point in the history
  • Loading branch information
dsikka committed Jan 14, 2025
1 parent 2d56538 commit 6e2bfa4
Showing 1 changed file with 2 additions and 5 deletions.
7 changes: 2 additions & 5 deletions tests/e2e/vLLM/test_vllm.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,15 +13,14 @@
from tests.e2e.e2e_utils import run_oneshot_for_e2e_testing
from tests.examples.utils import requires_gpu_count

"""
try:
from vllm import LLM, SamplingParams

vllm_installed = True
except ImportError:
vllm_installed = False
logger.warning("vllm is not installed. This test will be skipped")
"""


HF_MODEL_HUB_NAME = "nm-testing"
TEST_DATA_FILE = os.environ.get("TEST_DATA_FILE", "")
Expand All @@ -43,7 +42,7 @@ def record_config_file(record_testsuite_property: Callable[[str, object], None])
# Will run each test case in its own process through run_tests.sh
# emulating vLLM CI testing
@requires_gpu_count(1)
# @pytest.mark.skipif(not vllm_installed, reason="vLLM is not installed, skipping test")
@pytest.mark.skipif(not vllm_installed, reason="vLLM is not installed, skipping test")
class TestvLLM:
"""
The following test quantizes a model using a preset scheme or recipe,
Expand Down Expand Up @@ -139,7 +138,6 @@ def test_vllm(self):
folder_path=self.save_dir,
)

"""
logger.info("================= RUNNING vLLM =========================")

sampling_params = SamplingParams(temperature=0.80, top_p=0.95)
Expand All @@ -162,7 +160,6 @@ def test_vllm(self):
logger.info(generated_text)

self.tear_down()
"""

def tear_down(self):
if self.save_dir is not None:
Expand Down

0 comments on commit 6e2bfa4

Please sign in to comment.