Skip to content

Commit 608909d

Browse files
jrplatinJacob Platin
andauthored
[Bug fixes] Fix a few more vLLM imports + Dockerfile typo (#953)
Signed-off-by: Jacob Platin <[email protected]> Co-authored-by: Jacob Platin <[email protected]>
1 parent c11e5aa commit 608909d

File tree

7 files changed

+7
-7
lines changed

7 files changed

+7
-7
lines changed

docker/Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ RUN python3 -m pip install --no-cache-dir \
2727
pytest-asyncio \
2828
git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d#egg=lm-eval[api] \
2929
pytest-cov \
30-
tblib \
30+
tblib
3131

3232
# Install tpu_inference
3333
WORKDIR /workspace/tpu_inference

examples/multi_modal_inference.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
from vllm import LLM, EngineArgs, SamplingParams
2222
from vllm.assets.image import ImageAsset
2323
from vllm.multimodal.image import convert_image_mode
24-
from vllm.utils import FlexibleArgumentParser
24+
from vllm.utils.argparse_utils import FlexibleArgumentParser
2525

2626

2727
class ModelRequestData(NamedTuple):

examples/offline_lora_inference.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@
77
import vllm.envs as envs
88
from vllm import LLM, EngineArgs
99
from vllm.lora.request import LoRARequest
10-
from vllm.utils import FlexibleArgumentParser
10+
from vllm.utils.argparse_utils import FlexibleArgumentParser
1111

1212

1313
def create_parser():

examples/tpu_profiling.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@
1919
from vllm import LLM, SamplingParams
2020
from vllm.engine.arg_utils import EngineArgs
2121
from vllm.inputs import PromptType
22-
from vllm.utils import FlexibleArgumentParser
22+
from vllm.utils.argparse_utils import FlexibleArgumentParser
2323

2424
DURATION_MS = int(os.getenv("VLLM_TPU_PROFILE_DURATION_MS", 3000))
2525
DELAY_MS = int(os.getenv("VLLM_TPU_PROFILE_DELAY_MS", 0))

scripts/vllm/benchmarking/benchmark_serving.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@
4747
from backend_request_func import get_tokenizer
4848

4949
try:
50-
from vllm.utils import FlexibleArgumentParser
50+
from vllm.utils.argparse_utils import FlexibleArgumentParser
5151
except ImportError:
5252
from argparse import ArgumentParser as FlexibleArgumentParser
5353

tests/e2e/benchmarking/mlperf.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -295,7 +295,7 @@ for model_name in $model_list; do
295295
--dataset-name "$dataset_name" \
296296
--dataset-path "$dataset_path" \
297297
--num-prompts "$num_prompts" \
298-
--run_eval 2>&1 | tee -a "$BENCHMARK_LOG_FILE"
298+
--run-eval 2>&1 | tee -a "$BENCHMARK_LOG_FILE"
299299

300300
# TODO (jacobplatin): probably want to add an option to skip this in the future
301301
if [ "$dataset_name" == "mlperf" ]; then

tpu_inference/core/disagg_executor.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,11 +6,11 @@
66
from vllm.logger import init_logger
77
from vllm.multimodal import MULTIMODAL_REGISTRY
88
from vllm.multimodal.cache import worker_receiver_cache_from_config
9-
from vllm.utils import run_method
109
from vllm.utils.network_utils import (get_distributed_init_method, get_ip,
1110
get_open_port)
1211
from vllm.v1.executor.abstract import Executor
1312
from vllm.v1.outputs import AsyncModelRunnerOutput
13+
from vllm.v1.serial_utils import run_method
1414
from vllm.v1.worker.worker_base import WorkerWrapperBase
1515

1616
logger = init_logger(__name__)

0 commit comments

Comments
 (0)