Skip to content

Commit

Permalink
[Bugfix] Fix run_batch logger (vllm-project#7640)
Browse files Browse the repository at this point in the history
  • Loading branch information
pooyadavoodi authored Aug 23, 2024
1 parent 9db93de commit 6885fde
Showing 1 changed file with 2 additions and 5 deletions.
7 changes: 2 additions & 5 deletions vllm/entrypoints/openai/run_batch.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@

from vllm.engine.arg_utils import AsyncEngineArgs, nullable_str
from vllm.engine.async_llm_engine import AsyncLLMEngine
from vllm.entrypoints.logger import RequestLogger
from vllm.entrypoints.logger import RequestLogger, logger
# yapf: disable
from vllm.entrypoints.openai.protocol import (BatchRequestInput,
BatchRequestOutput,
Expand All @@ -16,13 +16,10 @@
# yapf: enable
from vllm.entrypoints.openai.serving_chat import OpenAIServingChat
from vllm.entrypoints.openai.serving_embedding import OpenAIServingEmbedding
from vllm.logger import init_logger
from vllm.usage.usage_lib import UsageContext
from vllm.utils import FlexibleArgumentParser, random_uuid
from vllm.version import __version__ as VLLM_VERSION

logger = init_logger(__name__)


def parse_args():
parser = FlexibleArgumentParser(
Expand Down Expand Up @@ -184,7 +181,7 @@ async def main(args):
if __name__ == "__main__":
args = parse_args()

logger.info("vLLM API server version %s", VLLM_VERSION)
logger.info("vLLM batch processing API version %s", VLLM_VERSION)
logger.info("args: %s", args)

asyncio.run(main(args))

0 comments on commit 6885fde

Please sign in to comment.