ChatterBot
ChatterBot copied to clipboard
The bot is really slow with big datasets...
I have successfully import a huge data set of Reddit into the bot. It seems to work fine for common questions and is really hilarious at times. But for some user inputs it take forever. Which is honestly not fun to have a conversation with... I wanted to ask if there was any way to decrease the response time with some tweaks or to just limit it to find fewer responses with a deadline (time limit)?
Yes l have also noticed the same thing. l tried using the cornel movie dataset and it was really slow to respond, l hope a fix is created for that because Chatterbot is very easy to use
@AtomynosAtom @AllanNyasha
I'm interested in furthering the development of ChatterBot! I'm using it in a project currently. I let the database train itself ~200 times as a test, and trying to use it renders ChatterBot essentially bricked. Once the DB file reached ~30mb, slowdown starts. As it reaches ~50-60mb, it becomes a crawl. One of my queries, a simple 5 word question, is still processing and it's been upwards of 30 minutes.
Any recommended fixes or findings from you all in regards to processing speed?
Additionally - @AtomynosAtom - how were you able to get the Reddit dataset into CB?
Additionally - @AtomynosAtom - how were you able to get the Reddit dataset into CB? I went on reddit to r/datasets and got reddit comments https://academictorrents.com/details/85a5bd50e4c365f8df70240ffd4ecc7dec59912b After that I got some scripts to clean the data and save it as a multiple yaml file. But the thing with yaml file is that it can't start with a symbol and all so I had to replace all those kinds of characters with text and then it imported it to CB many of them failed but I got a small chunk of them. I had asked the owner about a better way to do this but he didn't have one. https://github.com/gunthercox/ChatterBot/issues/2037#issuecomment-691721541
Yes i really hope @gunthercox fixes this, its really a pain for having to wait minutes for the chatbot to respond with large datasets.