scrapydweb
scrapydweb copied to clipboard
[BUG] Unable to fire crawling via redis on Heroku
trafficstars
Describe the bug
I am struggling to make it work as described here https://github.com/my8100/scrapyd-cluster-on-heroku#deploy-and-run-distributed-spiders .
Whenever I try to do this:
r.lpush('mycrawler:start_urls', 'http://books.toscrape.com', 'http://quotes.toscrape.com')
No job is started.
To Reproduce Steps to reproduce the behavior:
- Followed the steps in https://github.com/my8100/scrapyd-cluster-on-heroku#deploy-and-run-distributed-spiders on a new Heroku account ( Deploy Heroku apps in the browser )
- Tried to fire
mycrawler_redisspider from scrapy_redis_demo_project.zip as described in Deploy and run distributed spiders paragraph
Expected behavior
I expect that when I use lpush method on 'mycrawler:start_urls', I can fire a crawl job on Heroku.
Screenshots

Environment (please complete the following information):
- OS: Ubuntu 19.10
- Python: 3.7.5
- ScrapydWeb: installed with heroku button as described in https://github.com/my8100/scrapyd-cluster-on-heroku#deploy-and-run-distributed-spiders
- Scrapyd amount: tried with only one worker and one app
- Browser [e.g. Chrome 71, Firefox 64 or Safari 12]
Additional context I have sent you a mail with the same title of this ticket with a video attached.
- Make sure you have correctly set up REDIS_HOST, REDIS_PASSWORD, and REDIS_PORT environment variables when deploying the Scrapyd app on Heroku.
- Execute
pip install -U redison the PC where you execute the redis commands. - Try the redis commands again.
- If still fail:
(1) Visit https://dashboard.heroku.com/apps/{your-scrapyd-app}/settings,
click the "More" button at the top-right corner, and select "Run console",
then post the result of command
pip list. (2) Post the result of commandpip liston the PC where you execute the redis commands. (3) Post the full log file of your Scrapy job.