You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Since the birth of Backend.AI Model Service, main concern of the feature is that it is too hard to utilize for majority of users those who want to serve their own model. To overcome this problem we decided to add several new features on both Core and WebUI, which will potentially enhance overall experience of the Model Service feature.
Core: New "Dry Run" API
This new API should validate actual whole lifecycle of the inference session. Its request schema will be identical with the model service creation API. The implementation should first read model-definition.yml, create a new inference session accordingly but without the bound routing and endpoint, wait until model server loads, perform a health check (if defined at model definition), and finally terminate the created session. The API should report the whole progress to callee with the help of SSE.
Alternative ideas
No response
Anything else?
No response
The content you are editing has changed. Please copy your edits and refresh the page.
Main idea
Since the birth of Backend.AI Model Service, main concern of the feature is that it is too hard to utilize for majority of users those who want to serve their own model. To overcome this problem we decided to add several new features on both Core and WebUI, which will potentially enhance overall experience of the Model Service feature.
This new API should validate actual whole lifecycle of the inference session. Its request schema will be identical with the model service creation API. The implementation should first read
model-definition.yml
, create a new inference session accordingly but without the bound routing and endpoint, wait until model server loads, perform a health check (if defined at model definition), and finally terminate the created session. The API should report the whole progress to callee with the help of SSE.Alternative ideas
No response
Anything else?
No response
Tasks
try-start
model service API #1754The text was updated successfully, but these errors were encountered: