crawl4ai
crawl4ai copied to clipboard
cannot bypass cache db
Hi !
im currently working with the repo but when i try to webscrap multiple websites this message keeps popping up Error caching URL: database is locked async with AsyncWebCrawler(verbose=False, always_by_pass_cache=True, page_timeout=120000) as crawler: result = await crawler.arun(url=str(url), page_timeout=120000, bypass_cache=True)
maybe is on the crawl4ai\async_webcrawler.py line 48 awarmup that has fixed bypass_cache=False but it not quite sure
@jmontoyavallejo Hello, thanks for using our library. Sorry for the delay. Thanks for reporting this. This issue occurs when crawling multiple files concurrently and has been detected and fixed in the library. Also, issue with setting the 'always_bypass_cache' flag also has resolved. You may pull branch 0.3.73, or wait for the new update and install again from Pypi. Thanks again for reporting this.
hi @unclecode @@i still have this error in version 0.3.73
ERROR:crawl4ai.async_database:Connection error: database is locked ERROR:crawl4ai.async_database:Connection error: database is locked ERROR:crawl4ai.async_database:Operation failed after 3 attempts: database is locked ERROR:crawl4ai.async_database:Error caching URL: database is locked
Are you able to share with me a full script that I can reproduce the same error with it?
do you have a discord to do a little call or email? mine is [email protected]
@jmontoyavallejo I send the invitation link to Crawl4ai channel, there you can also DM me.
Closing this issue due to inactivity. Reopen a bug report if the issue still persists.