Skip to content

Commit

Permalink
Remove bad exceptions
Browse files Browse the repository at this point in the history
  • Loading branch information
damccorm committed Oct 8, 2024
1 parent 1346b09 commit fa1fd55
Showing 1 changed file with 0 additions and 3 deletions.
3 changes: 0 additions & 3 deletions sdks/python/apache_beam/ml/inference/vllm_inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -152,7 +152,6 @@ def check_connectivity(self, retries=3):
time.sleep(5)

if retries == 0:
raise Exception('test')
self._server_started = False
raise Exception(
"Failed to start vLLM server, polling process exited with code " +
Expand All @@ -161,8 +160,6 @@ def check_connectivity(self, retries=3):
else:
self.start_server(retries - 1)

raise Exception('test2')


class VLLMCompletionsModelHandler(ModelHandler[str,
PredictionResult,
Expand Down

0 comments on commit fa1fd55

Please sign in to comment.