Skip to content

Commit

Permalink
lint
Browse files Browse the repository at this point in the history
  • Loading branch information
damccorm committed Sep 23, 2024
1 parent 7703f53 commit 93bbd49
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions sdks/python/apache_beam/ml/inference/vllm_inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,7 @@ def __init__(self, model_name: str, vllm_server_kwargs: Dict[str, str]):
self._vllm_server_kwargs = vllm_server_kwargs
self._server_started = False
self._server_process = None
self._server_port = None
self._server_port: int = -1

self.start_server()

Expand Down Expand Up @@ -218,7 +218,7 @@ def __init__(
self,
model_name: str,
chat_template_path: Optional[str] = None,
vllm_server_kwargs: Dict[str, str] = None):
vllm_server_kwargs: Optional[Dict[str, str]] = None):
""" Implementation of the ModelHandler interface for vLLM using previous
messages as input.
Expand Down

0 comments on commit 93bbd49

Please sign in to comment.