From 80111d12d30d92a657e24194212a5fbfcac2af50 Mon Sep 17 00:00:00 2001 From: Yunnglin Date: Thu, 19 Dec 2024 15:11:53 +0800 Subject: [PATCH] add spawn env --- evalscope/perf/arguments.py | 1 + evalscope/perf/utils/local_server.py | 1 + 2 files changed, 2 insertions(+) diff --git a/evalscope/perf/arguments.py b/evalscope/perf/arguments.py index 604fe4d..25602a3 100644 --- a/evalscope/perf/arguments.py +++ b/evalscope/perf/arguments.py @@ -68,6 +68,7 @@ def from_args(args): model=args.model, attn_implementation=args.attn_implementation, url=args.url, + port=args.port, api_key=args.api_key, connect_timeout=args.connect_timeout, read_timeout=args.read_timeout, diff --git a/evalscope/perf/utils/local_server.py b/evalscope/perf/utils/local_server.py index 5afb02d..c145117 100644 --- a/evalscope/perf/utils/local_server.py +++ b/evalscope/perf/utils/local_server.py @@ -103,6 +103,7 @@ def start_app(args: Arguments): elif args.api == 'local_vllm': os.environ['VLLM_USE_MODELSCOPE'] = 'True' os.environ['VLLM_ALLOW_LONG_MAX_MODEL_LEN'] = '1' + os.environ['VLLM_WORKER_MULTIPROC_METHOD'] = 'spawn' # yapf: disable proc = subprocess.Popen([ 'python', '-m', 'vllm.entrypoints.openai.api_server',