-
Notifications
You must be signed in to change notification settings - Fork 61
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[feature] Provider unified offline batch inference interface #47
Comments
It is related to our python SDK. Ref #49 |
There are roughly two ways to consider:
|
I think we are using this approach in model registry (triton inference server). But we wanna support offline inference here. |
We use the converted model for offline inference. But we need convert out model first ,maybe use model registry。 |
Personally prefer the latter. If we can unify the API on top of models, we can support multiple framework formats. If we wanna support the offline inference, we always need an SDK, I think. |
/assign |
Is this a BUG REPORT or FEATURE REQUEST?:
What happened:
Investigate if we can use https://github.com/uber/neuropod to provide a unified offline batch inference interface for users. They can use ormb python sdk to download the model first then use neuropod to run offline inference.
Thank @terrytangyuan for introducing the project.
What you expected to happen:
How to reproduce it (as minimally and precisely as possible):
Anything else we need to know?:
The text was updated successfully, but these errors were encountered: