scrapy-playwright
scrapy-playwright copied to clipboard
🎭 Playwright integration for Scrapy
Without proxy, cookie applied correctly. But when I use proxy (brightdata), then the cookie is not applied. Did I miss anything? ``` class ScrapyTest(scrapy.Spider): name = 'scrapy test' def start_requests(self):...
Hi! I have a spider that uses playwright with a proxy. **NOTE: the spider works as it should when the proxy is not needed and the proxy works, as the...
I have had a hard time trying to follow links using the Scrapy Playwright to navigate a dynamic website. want to write a crawl spider that will get all available...
Integrated playwright_stealth, and PLAYWRIGHT_STEALTH_ENABLED as an optional config. Attached bot test results. **PLAYWRIGHT_STEALTH_ENABLED = True**  **PLAYWRIGHT_STEALTH_ENABLED = False** 
Hi. I think the results of using playright and scrappy-playright are different in some situations. When i use just playwright, it just propery worked. but same code in scrapy-playwright wasn't...
Hi, I have a strange issue where I am receiving a 400 response from Google after clicking on the "I agree" button on their consent form.  This issue however...
Hi. I crawl a website using scrapy_plawright , I use `wait_for_selector` and when page isn't exist (status = 404) scrapy_playwright wait until `Timeout` and then raise exception. Is there any...
It seems that when using scrapy-playwright Scrapy will not shut down cleanly on SIGINT (`cmd+c`), and you have to force a shutdown with a second `cmd+c`. If you use the...
This function [function](https://playwright.dev/python/docs/api/class-page#page-add-init-script) and either this [function](https://playwright.dev/python/docs/api/class-browsercontext#browser-context-add-init-script) are not implemented in scrapy-playwright. I tried with coroutines (evaluate) but it doesn't give the same results as `add_init_script` as the JS script...
Hi, This issue related to [#18](https://github.com/scrapy-plugins/scrapy-playwright/issues/18) The error still occurred with `scrapy-playwright 0.0.4`. The Scrapy script crawled about 2500 domains in 10k from [majestic](https://majestic.com/reports/majestic-million) and crashed with the last error...