PRADEEP T

Results 64 issues of PRADEEP T

I have installed the Tigergraph enterprise edition on a local machine. when I put the ```gadmin status``` it shows like this ![Screenshot from 2022-05-24 10-23-55](https://user-images.githubusercontent.com/41164884/169951943-3d68d62e-0511-43a3-b638-fa11c0b803bf.png) Now I want to launch...

### Describe the bug I asked the to return the response in JSON type in the chatGPT prompt. But it returns always in a raw_text manner inside ```{"message":{"role":"assistant","content":``` How can...

bug

You are claiming that privateGPT not using any openai interface and can work without an internet connection. May I know which LLM model is using inside privateGPT for inference purpose?

enhancement

### Issue you'd like to raise. I am doing the pdf question answering using the below code. Note: I am integrated the ConversationBufferMemory for keeping my chat in the memory...

auto:question

### Feature request how to disable the OpenAI initialization when you're not using an OpenAI model. [Please check this issue](https://github.com/hwchase17/langchain/issues/7189#issuecomment-1621931461) ### Motivation I am trying to build a VectorstoreIndexCreator using...

auto:bug

### System Info langchain==0.0.219 python 3.9 ### Who can help? _No response_ ### Information - [ ] The official example notebooks/scripts - [ ] My own modified scripts ### Related...

auto:bug

### System Info langchain==0.0.219 Python 3.9 ### Who can help? _No response_ ### Information - [ ] The official example notebooks/scripts - [ ] My own modified scripts ### Related...

auto:bug

### Issue you'd like to raise. I am trying to query the documents using the below stack **Langchain+ embedding tool + vectore store + LLM model** There are many tools...

auto:question

### System Info langchain==0.0.219 python 3.9 ### Who can help? _No response_ ### Information - [ ] The official example notebooks/scripts - [ ] My own modified scripts ### Related...

auto:bug

Here is the SFTtrainer method i used for finetuning mistral ``` trainer = SFTTrainer( model=peft_model, train_dataset=data, peft_config=peft_config, dataset_text_field=" column name", max_seq_length=3000, tokenizer=tokenizer, args=training_arguments, packing=packing, ) trainer.train() ``` I found different...