Víctor Martínez
Víctor Martínez
well looks like it no longer works :( tried imap using app password and it does not sync after adding account :)
needed to run with `google-chrome-stable --enable-unsafe-webgpu`, the chrome://flags alone did not work
@codingjlu just a warning that webgpu is experimental and unsafe and that you are at your own risk
@ltdrdata great! keeping an eye, will take a chance at implementing the fine-tuning loop when these issues are resolved
@Wingie toolpaca is finetuned around toolformer prompts which are completely different in both syntax and operation from what langchain provides. see https://github.com/lucidrains/toolformer-pytorch/blob/main/toolformer_pytorch/prompts.py vs https://github.com/hwchase17/langchain/blob/master/langchain/agents/conversational/prompt.py the proper way would be to...
@oobabooga langchain is essentially a prompt generation and execution framework. allows to do things like re-writting and re-evaluating the conversation history to perform external data ingestion or auto-summarizing the history...
also includes the newest SOTA 2bit quant support :D
> https://github.com/ollama/ollama/blob/ecc133d843c8567b27ff3bdc9ff811ecad99281a/docs/faq.md?plain=1#L189 > > use keep_alive param * any negative number which will keep the model loaded in memory (e.g. -1 or "-1m") Does the opposite actually.
> Hey @knoopx You can actually do this by calling `curl http://localhost:11434/api/generate -d '{"model": "llama2", "keep_alive": 0}'` (not with `-1` which will always leave the model loaded). That will immediately...
fyi: you can load vicuna model through huggingface transformers by installing it from their git repo. then just load the tokenizer and model via `LlamaTokenizer.from_pretrained(...)` and `LlamaForCausalLM.from_pretrained(...)` and pass them...