Matthew Kotila
Matthew Kotila
Unfortunately we don't support supplying binary files for more than one request, but you should be able to convert the binary data into b64 representation and include that in an...
> @kzelias: @matthewkotila > If I use b64 + json, I will need to change the logic of the triton service, right? Would need to decode b64. The decoding of...
@siretru you can find information about the GPU utilization metric that Perf Analyzer offers here: https://github.com/triton-inference-server/client/blob/main/src/c%2B%2B/perf_analyzer/docs/measurements_metrics.md#server-side-prometheus-metrics
Hi @RRoundTable, thanks for making these changes. Can you fill out and email the CLA as described here?: https://github.com/triton-inference-server/server/blob/main/CONTRIBUTING.md#contributor-license-agreement-cla
> > @matthewkotila: Hi @RRoundTable, thanks for making these changes. Can you fill out and email the CLA as described here?: > > https://github.com/triton-inference-server/server/blob/main/CONTRIBUTING.md#contributor-license-agreement-cla > > @RRoundTable: Sorry. I checked...
@RRoundTable I see the PR was closed--do you still want to contribute these changes (or anything to the [triton-inference-server](https://github.com/triton-inference-server) organization)?
Yeah, I would advise running the above command. Check these instructions for a more complete list of dependencies/etc: https://github.com/triton-inference-server/client/blob/main/src/c%2B%2B/perf_analyzer/docs/install.md#build-from-source
Converted to draft pending Slack conversation about whether there are still stakeholders for this feature.
Closing due to inactivity.
> @kthui: Hi @matthewkotila @nv-hwoo, wondering if you have worked with a BERT model in the past? If so, would you be able to share the model config? I am...