From 6885fde317433eec52e00c14329270d742f0630d Mon Sep 17 00:00:00 2001 From: Pooya Davoodi Date: Fri, 23 Aug 2024 13:58:26 -0700 Subject: [PATCH] [Bugfix] Fix run_batch logger (#7640) --- vllm/entrypoints/openai/run_batch.py | 7 ++----- 1 file changed, 2 insertions(+), 5 deletions(-) diff --git a/vllm/entrypoints/openai/run_batch.py b/vllm/entrypoints/openai/run_batch.py index af8d95ea..764712fd 100644 --- a/vllm/entrypoints/openai/run_batch.py +++ b/vllm/entrypoints/openai/run_batch.py @@ -6,7 +6,7 @@ import aiohttp from vllm.engine.arg_utils import AsyncEngineArgs, nullable_str from vllm.engine.async_llm_engine import AsyncLLMEngine -from vllm.entrypoints.logger import RequestLogger +from vllm.entrypoints.logger import RequestLogger, logger # yapf: disable from vllm.entrypoints.openai.protocol import (BatchRequestInput, BatchRequestOutput, @@ -16,13 +16,10 @@ from vllm.entrypoints.openai.protocol import (BatchRequestInput, # yapf: enable from vllm.entrypoints.openai.serving_chat import OpenAIServingChat from vllm.entrypoints.openai.serving_embedding import OpenAIServingEmbedding -from vllm.logger import init_logger from vllm.usage.usage_lib import UsageContext from vllm.utils import FlexibleArgumentParser, random_uuid from vllm.version import __version__ as VLLM_VERSION -logger = init_logger(__name__) - def parse_args(): parser = FlexibleArgumentParser( @@ -184,7 +181,7 @@ async def main(args): if __name__ == "__main__": args = parse_args() - logger.info("vLLM API server version %s", VLLM_VERSION) + logger.info("vLLM batch processing API version %s", VLLM_VERSION) logger.info("args: %s", args) asyncio.run(main(args))