-
Notifications
You must be signed in to change notification settings - Fork 155
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: Add parameters support to InferResponse #394
Conversation
7b63cc9
to
c3a9c1b
Compare
* Infer response to track parameters * Add parameters to binding infer response * Rank parameters argument up among InferResponse constructor arguments * Add setting parameters to Triton response * Send response parameters only on non-error * Fix double declaration * Unify py dictionary parameters to json str
c3a9c1b
to
816d4ac
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
General look&feel LGTM - left a couple questions. Mostly defer to @krishung5 for an expert opinion
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🚀
Thanks Jacky! |
@@ -803,6 +803,9 @@ You can read more about the inference response parameters in the [parameters | |||
extension](https://github.com/triton-inference-server/server/blob/main/docs/protocol/extension_parameters.md) | |||
documentation. | |||
|
|||
Inference response parameters is currently not supported on BLS inference |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
not sure this is clear ....
What does the PR do?
Add support for setting response parameters in regular and decoupled Python backend model response(s).
Checklist
<commit_type>: <Title>
Commit Type:
Check the conventional commit type
box here and add the label to the github PR.
Related PRs:
triton-inference-server/server#7964
Where should the reviewer start?
Start with the test cases on the related server PR, and then
pb_stub.cc
->infer_response.cc
->python_be.cc
.Test plan:
New tests are added to the related server PR.
Caveats:
Responses to BLS models are not populated with the response parameters, if any. (DLIS-7864)
Background
N/A
Related Issues: (use one of the action keywords Closes / Fixes / Resolves / Relates to)
N/A