devika icon indicating copy to clipboard operation
devika copied to clipboard

Fixed Bugs and Added some useful functions....

Open Rawknee-69 opened this issue 1 year ago • 13 comments

Rawknee-69 avatar Apr 26 '24 10:04 Rawknee-69

remove the lock file

ARajgor avatar Apr 26 '24 12:04 ARajgor

can you give me access to contribute to this pr?

ARajgor avatar Apr 26 '24 16:04 ARajgor

what you wanna contribute

Rawknee-69 avatar Apr 26 '24 17:04 Rawknee-69

resolve the conflict and merge the main to this. also, where do you use the knowledge function I couldn't find it.

ARajgor avatar Apr 26 '24 17:04 ARajgor

"You are an angelic AI Software Engineer, remarkable in intelligence and devoted to establishing a welcoming ambiance for users. Demonstrating perpetual politeness, grace, and acute awareness, you adeptly interpret and cater to user necessities. Taking into account earlier dialogues:"

Is all these very necessary? What I feel is it might be using up some extra tokens. Because this was under actions. It anyways just have to provide the action for subsequent execute. Let me know if I am wrong.

obliviousz avatar Apr 26 '24 18:04 obliviousz

"You are an angelic AI Software Engineer, remarkable in intelligence and devoted to establishing a welcoming ambiance for users. Demonstrating perpetual politeness, grace, and acute awareness, you adeptly interpret and cater to user necessities. Taking into account earlier dialogues:"

Is all these very necessary? What I feel is it might be using up some extra tokens. Because this was under actions. It anyways just have to provide the action for subsequent execute. Let me know if I am wrong.

actually there is a problem that after some replies the llm's starts to hallucinate and the responses it give degrades it quality , sure it takes some extra token but this helps llms to make sure they are in their character and doesn't hallucinate and it also reflects the personality of the llm's

Rawknee-69 avatar Apr 26 '24 18:04 Rawknee-69

"You are an angelic AI Software Engineer, remarkable in intelligence and devoted to establishing a welcoming ambiance for users. Demonstrating perpetual politeness, grace, and acute awareness, you adeptly interpret and cater to user necessities. Taking into account earlier dialogues:" Is all these very necessary? What I feel is it might be using up some extra tokens. Because this was under actions. It anyways just have to provide the action for subsequent execute. Let me know if I am wrong.

actually there is a problem that after some replies the llm's starts to hallucinate and the responses it give degrades it quality , sure it takes some extra token but this helps llms to make sure they are in their character and doesn't hallucinate and it also reflects the personality of the llm's

But everytime we set the context again no? With every prompt. How is it possible?

obliviousz avatar Apr 26 '24 19:04 obliviousz

resolve the conflict and merge the main to this. also, where do you use the knowledge function I couldn't find it.

actually i have modified the knowledge function and didn't changed anything on part of variables so the knowledge function where you have used it previously is there but only the style to store and to extract it is changed by using the a llm that run's locally and using faiss

Rawknee-69 avatar Apr 26 '24 19:04 Rawknee-69

resolve the conflict and merge the main to this. also, where do you use the knowledge function I couldn't find it.

done resloved the conflicts and also explaied about the knowledge part

Rawknee-69 avatar Apr 26 '24 19:04 Rawknee-69

I mean you write the knowledge_base.py file integrating with faiss. but where are you using those functions in the agent.py files. @Rawknee-69

ARajgor avatar Apr 27 '24 06:04 ARajgor

#485 I tested this PR intensively and it's working like magic . I added this after commit knowledge session.refresh(knowledge) # Reload the object from the session To add_knowledge function and change the logic of storing multiple results for same query And concatenate them using join function

darrassi1 avatar Apr 27 '24 13:04 darrassi1

#485 I tested this PR intensively and it's working like magic . I added this after commit knowledge session.refresh(knowledge) # Reload the object from the session To add_knowledge function and change the logic of storing multiple results for same query And concatenate them using join function

What are the changes which you see if you can let us know?

obliviousz avatar Apr 27 '24 13:04 obliviousz

It's enhancing retrieval of memory from the local database , and my approche is combining Context from knowledge base and results from web browser and feed it to a LLM for better compréhension .

darrassi1 avatar Apr 27 '24 15:04 darrassi1