From 957dae18fe6b88cc59877fccf926151bb35e0453 Mon Sep 17 00:00:00 2001 From: xrdaukar Date: Wed, 12 Feb 2025 17:56:52 -0800 Subject: [PATCH] Update VLM sample `oumi infer -i` commands (#1428) --- .../recipes/vision/llama3_2_vision/inference/11b_infer.yaml | 3 ++- .../vision/llama3_2_vision/inference/11b_rvllm_infer.yaml | 3 ++- .../vision/llama3_2_vision/inference/11b_sglang_infer.yaml | 3 ++- .../vision/llama3_2_vision/inference/11b_vllm_infer.yaml | 3 ++- configs/recipes/vision/llava_7b/inference/infer.yaml | 3 ++- configs/recipes/vision/llava_7b/inference/vllm_infer.yaml | 3 ++- configs/recipes/vision/phi3/inference/vllm_infer.yaml | 3 ++- configs/recipes/vision/qwen2_5_vl_3b/inference/infer.yaml | 3 ++- configs/recipes/vision/qwen2_5_vl_3b/inference/vllm_infer.yaml | 3 ++- configs/recipes/vision/qwen2_vl_2b/inference/infer.yaml | 3 ++- configs/recipes/vision/qwen2_vl_2b/inference/sglang_infer.yaml | 3 ++- configs/recipes/vision/qwen2_vl_2b/inference/vllm_infer.yaml | 3 ++- 12 files changed, 24 insertions(+), 12 deletions(-) diff --git a/configs/recipes/vision/llama3_2_vision/inference/11b_infer.yaml b/configs/recipes/vision/llama3_2_vision/inference/11b_infer.yaml index 2d1194137..151ed93c8 100644 --- a/configs/recipes/vision/llama3_2_vision/inference/11b_infer.yaml +++ b/configs/recipes/vision/llama3_2_vision/inference/11b_infer.yaml @@ -5,7 +5,8 @@ # - Request access to Llama 3.2: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct # # Usage: -# oumi infer -i -c configs/recipes/vision/llama3_2_vision/inference/11b_infer.yaml +# oumi infer -i -c configs/recipes/vision/llama3_2_vision/inference/11b_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html diff --git a/configs/recipes/vision/llama3_2_vision/inference/11b_rvllm_infer.yaml b/configs/recipes/vision/llama3_2_vision/inference/11b_rvllm_infer.yaml index 161a756fb..a6afd4214 100644 --- a/configs/recipes/vision/llama3_2_vision/inference/11b_rvllm_infer.yaml +++ b/configs/recipes/vision/llama3_2_vision/inference/11b_rvllm_infer.yaml @@ -6,7 +6,8 @@ # - Request access to Llama 3.2: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct # # Usage: -# oumi infer -i -c configs/recipes/vision/llama3_2_vision/inference/11b_rvllm_infer.yaml +# oumi infer -i -c configs/recipes/vision/llama3_2_vision/inference/11b_rvllm_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # Sample python command to start vLLM server: # diff --git a/configs/recipes/vision/llama3_2_vision/inference/11b_sglang_infer.yaml b/configs/recipes/vision/llama3_2_vision/inference/11b_sglang_infer.yaml index c1679d118..0a5b73e48 100644 --- a/configs/recipes/vision/llama3_2_vision/inference/11b_sglang_infer.yaml +++ b/configs/recipes/vision/llama3_2_vision/inference/11b_sglang_infer.yaml @@ -6,7 +6,8 @@ # - Request access to Llama 3.2: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct # # Usage: -# oumi infer -i -c configs/recipes/vision/llama3_2_vision/inference/11b_sglang_infer.yaml +# oumi infer -i -c configs/recipes/vision/llama3_2_vision/inference/11b_sglang_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # Sample command to start SGLang server: # CUDA_VISIBLE_DEVICES=0 python -m sglang.launch_server \ diff --git a/configs/recipes/vision/llama3_2_vision/inference/11b_vllm_infer.yaml b/configs/recipes/vision/llama3_2_vision/inference/11b_vllm_infer.yaml index 45c2feb85..160f91d8b 100644 --- a/configs/recipes/vision/llama3_2_vision/inference/11b_vllm_infer.yaml +++ b/configs/recipes/vision/llama3_2_vision/inference/11b_vllm_infer.yaml @@ -6,7 +6,8 @@ # - Request access to Llama 3.2: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct # # Usage: -# oumi infer -i -c configs/recipes/vision/llama3_2_vision/inference/11b_vllm_infer.yaml +# oumi infer -i -c configs/recipes/vision/llama3_2_vision/inference/11b_vllm_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html diff --git a/configs/recipes/vision/llava_7b/inference/infer.yaml b/configs/recipes/vision/llava_7b/inference/infer.yaml index a00a07d0e..da450b374 100644 --- a/configs/recipes/vision/llava_7b/inference/infer.yaml +++ b/configs/recipes/vision/llava_7b/inference/infer.yaml @@ -1,7 +1,8 @@ # Llava 7B inference config. # # Usage: -# oumi infer -i -c configs/recipes/vision/llava_7b/inference/infer.yaml +# oumi infer -i -c configs/recipes/vision/llava_7b/inference/infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html diff --git a/configs/recipes/vision/llava_7b/inference/vllm_infer.yaml b/configs/recipes/vision/llava_7b/inference/vllm_infer.yaml index e3ed5890b..dcbb3231a 100644 --- a/configs/recipes/vision/llava_7b/inference/vllm_infer.yaml +++ b/configs/recipes/vision/llava_7b/inference/vllm_infer.yaml @@ -4,7 +4,8 @@ # - Run `pip install vllm` # # Usage: -# oumi infer -i -c configs/recipes/vision/llava_7b/inference/vllm_infer.yaml +# oumi infer -i -c configs/recipes/vision/llava_7b/inference/vllm_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html diff --git a/configs/recipes/vision/phi3/inference/vllm_infer.yaml b/configs/recipes/vision/phi3/inference/vllm_infer.yaml index 879db9443..af59fd1b7 100644 --- a/configs/recipes/vision/phi3/inference/vllm_infer.yaml +++ b/configs/recipes/vision/phi3/inference/vllm_infer.yaml @@ -4,7 +4,8 @@ # - Run `pip install vllm` # # Usage: -# oumi infer -i -c configs/recipes/vision/phi3/inference/vllm_infer.yaml +# oumi infer -i -c configs/recipes/vision/phi3/inference/vllm_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html diff --git a/configs/recipes/vision/qwen2_5_vl_3b/inference/infer.yaml b/configs/recipes/vision/qwen2_5_vl_3b/inference/infer.yaml index 9da278a70..c41dc42fd 100644 --- a/configs/recipes/vision/qwen2_5_vl_3b/inference/infer.yaml +++ b/configs/recipes/vision/qwen2_5_vl_3b/inference/infer.yaml @@ -1,7 +1,8 @@ # Qwen 2.5 VL 3B inference config. # # Usage: -# oumi infer -i -c configs/recipes/vision/qwen2_5_vl_3b/inference/infer.yaml +# oumi infer -i -c configs/recipes/vision/qwen2_5_vl_3b/inference/infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html diff --git a/configs/recipes/vision/qwen2_5_vl_3b/inference/vllm_infer.yaml b/configs/recipes/vision/qwen2_5_vl_3b/inference/vllm_infer.yaml index 6051cc702..c92c9ca9f 100644 --- a/configs/recipes/vision/qwen2_5_vl_3b/inference/vllm_infer.yaml +++ b/configs/recipes/vision/qwen2_5_vl_3b/inference/vllm_infer.yaml @@ -4,7 +4,8 @@ # - Run `pip install vllm` # # Usage: -# oumi infer -i -c configs/recipes/vision/qwen2_5_vl_3b/inference/vllm_infer.yaml +# oumi infer -i -c configs/recipes/vision/qwen2_5_vl_3b/inference/vllm_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html diff --git a/configs/recipes/vision/qwen2_vl_2b/inference/infer.yaml b/configs/recipes/vision/qwen2_vl_2b/inference/infer.yaml index d46bf5e60..29c340c6f 100644 --- a/configs/recipes/vision/qwen2_vl_2b/inference/infer.yaml +++ b/configs/recipes/vision/qwen2_vl_2b/inference/infer.yaml @@ -1,7 +1,8 @@ # Inference config for Qwen2 VL 2B Instruct. # # Usage: -# oumi infer -i -c configs/recipes/vision/qwen2_vl_2b/inference/infer.yaml +# oumi infer -i -c configs/recipes/vision/qwen2_vl_2b/inference/infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html diff --git a/configs/recipes/vision/qwen2_vl_2b/inference/sglang_infer.yaml b/configs/recipes/vision/qwen2_vl_2b/inference/sglang_infer.yaml index 3187edda2..8ffa4b81a 100644 --- a/configs/recipes/vision/qwen2_vl_2b/inference/sglang_infer.yaml +++ b/configs/recipes/vision/qwen2_vl_2b/inference/sglang_infer.yaml @@ -4,7 +4,8 @@ # - Install SGLang: https://docs.sglang.ai/start/install.html # # Usage: -# oumi infer -i -c configs/recipes/vision/qwen2_vl_2b/inference/sglang_infer.yaml +# oumi infer -i -c configs/recipes/vision/qwen2_vl_2b/inference/sglang_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # Sample command to start SGLang server: # CUDA_VISIBLE_DEVICES=0 python -m sglang.launch_server \ diff --git a/configs/recipes/vision/qwen2_vl_2b/inference/vllm_infer.yaml b/configs/recipes/vision/qwen2_vl_2b/inference/vllm_infer.yaml index 1b2804ccd..49e2eb27a 100644 --- a/configs/recipes/vision/qwen2_vl_2b/inference/vllm_infer.yaml +++ b/configs/recipes/vision/qwen2_vl_2b/inference/vllm_infer.yaml @@ -4,7 +4,8 @@ # - Run `pip install vllm` # # Usage: -# oumi infer -i -c configs/recipes/vision/qwen2_vl_2b/inference/vllm_infer.yaml +# oumi infer -i -c configs/recipes/vision/qwen2_vl_2b/inference/vllm_infer.yaml \ +# --image "tests/testdata/images/the_great_wave_off_kanagawa.jpg" # # See Also: # - Documentation: https://oumi.ai/docs/en/latest/user_guides/infer/infer.html