Traian Rebedea

Results 6 comments of Traian Rebedea

Hi, @habanoz ! Unfortunately, if the `command-r` model does not work for generating intents, it does not really seem useful for Guardrails for now. I am sure that with a...

Hi @imintollms , For HF models you should use `nemoguardrails.llm.providers.HuggingFacePipelineCompatible` instead of the vanilla `HuggingFacePipeline`, similar to this: https://github.com/NVIDIA/NeMo-Guardrails/blob/8bb50afc68060e6c46d6fc557bcdb97cffec0fc8/examples/configs/llm/hf_pipeline_dolly/config.py#L60 This will solve your problem.

Hi @nashugame , I was able to run your sample code, I assume you are using `HuggingFaceH4/zephyr-7b-beta`. For me, the model required about 16-18 GB GPU RAM and the latency...

HI @ahsan3219 , There is an VertexAI Langchain wrapper that you can use. Is there a reason for not using it? I don't fully understand the code you shared above....

Hi @maxyousif15 ! I have not been able to replicate this bug. I have run with the configs you provided and the streaming example demo that I also assume you...

The message `I'm not sure what to say.` in `single_call` mode appears when the LLM is not able to generate all the necessary steps `(user intent, ... , bot intent,...