Zephyr-7B-beta-RAG-Demo
Zephyr-7B-beta-RAG-Demo copied to clipboard
Zephyr 7B beta RAG Demo inside a Gradio app powered by BGE Embeddings, ChromaDB, and Zephyr 7B Beta LLM.
I am running a model on RTX A1000 GPU but it takes 60 seconds to get an answer
Number of tokens (2662) exceeded maximum context length (512). how to solve this error
hello! python app.py says: ``` python app.py /home/pai/text-generation-webui/Zephyr-7B-beta-RAG-Demo/.conda/lib/python3.10/site-packages/langchain/__init__.py:34: UserWarning: Importing PromptTemplate from langchain root module is no longer supported. Please use langchain.prompts.PromptTemplate instead. warnings.warn( /home/pai/text-generation-webui/Zephyr-7B-beta-RAG-Demo/.conda/lib/python3.10/site-packages/langchain/__init__.py:34: UserWarning: Importing LLMChain from langchain...