diff --git a/.github/workflows/yapf.yml b/.github/workflows/yapf.yml index b77a1a47..b163c960 100644 --- a/.github/workflows/yapf.yml +++ b/.github/workflows/yapf.yml @@ -28,4 +28,4 @@ jobs: pip install toml==0.10.2 - name: Running yapf run: | - yapf --diff --recursive vllm tests + yapf --diff --recursive . diff --git a/benchmarks/benchmark_serving.py b/benchmarks/benchmark_serving.py index 28faa96f..1a36d9d6 100644 --- a/benchmarks/benchmark_serving.py +++ b/benchmarks/benchmark_serving.py @@ -92,17 +92,9 @@ async def get_request( await asyncio.sleep(interval) -async def send_request( - backend: str, - model: str, - api_url: str, - prompt: str, - prompt_len: int, - output_len: int, - best_of: int, - use_beam_search: bool, - pbar: tqdm -) -> None: +async def send_request(backend: str, model: str, api_url: str, prompt: str, + prompt_len: int, output_len: int, best_of: int, + use_beam_search: bool, pbar: tqdm) -> None: request_start_time = time.perf_counter() headers = {"User-Agent": "Benchmark Client"} @@ -155,7 +147,6 @@ async def send_request( pbar.update(1) - async def benchmark( backend: str, model: str, @@ -217,7 +208,10 @@ if __name__ == "__main__": type=str, default="vllm", choices=["vllm", "tgi"]) - parser.add_argument("--protocol", type=str, default="http", choices=["http", "https"]) + parser.add_argument("--protocol", + type=str, + default="http", + choices=["http", "https"]) parser.add_argument("--host", type=str, default="localhost") parser.add_argument("--port", type=int, default=8000) parser.add_argument("--endpoint", type=str, default="/generate") diff --git a/examples/openai_chatcompletion_client.py b/examples/openai_chatcompletion_client.py index 0b8e4b86..bbada389 100644 --- a/examples/openai_chatcompletion_client.py +++ b/examples/openai_chatcompletion_client.py @@ -32,6 +32,5 @@ chat_completion = client.chat.completions.create( model=model, ) - print("Chat completion results:") print(chat_completion) diff --git a/examples/openai_completion_client.py b/examples/openai_completion_client.py index 7a80c4ac..58519f97 100644 --- a/examples/openai_completion_client.py +++ b/examples/openai_completion_client.py @@ -21,8 +21,7 @@ completion = client.completions.create( echo=False, n=2, stream=stream, - logprobs=3 -) + logprobs=3) print("Completion results:") if stream: diff --git a/format.sh b/format.sh index 25183989..c7810886 100755 --- a/format.sh +++ b/format.sh @@ -71,7 +71,7 @@ format_changed() { # Format all files format_all() { - yapf --in-place "${YAPF_FLAGS[@]}" "${YAPF_EXCLUDES[@]}" vllm tests + yapf --in-place "${YAPF_FLAGS[@]}" "${YAPF_EXCLUDES[@]}" . } ## This flag formats individual files. --files *must* be the first command line