You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For a single model instance, even a small number of concurrent requests significantly increases response time. From what I've observed, the model only accepts one text input at a time and returns the corresponding audio response. Is it possible to make some simple changes to allow it to process multiple texts simultaneously without slowing down?
The text was updated successfully, but these errors were encountered:
How can the model support parallel inference?
For a single model instance, even a small number of concurrent requests significantly increases response time. From what I've observed, the model only accepts one text input at a time and returns the corresponding audio response. Is it possible to make some simple changes to allow it to process multiple texts simultaneously without slowing down?
The text was updated successfully, but these errors were encountered: