Geoffrey Angus

Results 22 comments of Geoffrey Angus
trafficstars

From the visualization standpoint, you can take inspiration from some of our existing visualization functionality: https://github.com/ludwig-ai/ludwig/blob/49b4c79cc89c5369a37d3cb4922893898004ce13/ludwig/visualize.py#L2388 In order to change the threshold in `predict`, you can build off of functionality...

Hi @fire, sorry for the late response here. To your points: 1. NVIDIA Triton does support C++ backends: https://github.com/triton-inference-server/pytorch_backend 2. I do believe that it needs a CUDA runtime, but...

Primarily reviewed the TorchScript-relevant parts of the PR– it's starting to look good. A couple of comments then I think those parts look good to me.

Superseded by https://github.com/ludwig-ai/ludwig/pull/2733.

Hi @aarnphm, can you clarify this point? > you can provide the adapter_name via the swagger UI, under the same level as prompt key. I've deployed an OPT model using...

Got it– I'm also trying now by using the vanilla `openllm` CLI tool. Here's how I deployed the model: ``` openllm start falcon --model-id tiiuae/falcon-7b-instruct ``` Here's what I submitted:...

Hi @philippe-solodov-wd– can you describe your use case in more detail? Are you trying to train on a GPU-enabled machine, without using the GPU?

Hello @chayanray– Yes, it looks like you may have to download the `llm` subpackage for Ludwig. I see that you are on `0.8.dev`. If you therefore installed by git cloning...

Hi @GlacierPurpleBison, taking a look. Do you know if this bug is a SageMaker-specific bug, or if this occurs when initializing a vanilla LoRAX docker container as well?