Scrape using Selenium webdriver.
Not well tested. Has probably lots of bugs, unhandled situations, and scrapy integration problems. Use at your own risk.
That being said, it works for me, in all its non-blocking glory.
- If you are using scrapy version 0.18.2 or later, the setting LOG_STDOUT must be False.
- I suspect WebdriverActionRequest and ActionChain are broken. I am not using them.
For now it's not on pypi, but this should work:
pip install https://github.com/zipfworks/scrapy-webdriver/archive/master.zip
Or something like this, in setup.py:
setup(
install_requires=[
'scrapy_webdriver',
...,
],
dependency_links=[
'https://github.com/zipfworks/scrapy-webdriver/archive/master.zip#egg=scrapy_webdriver',
],
...,
)
Add something like this in your scrapy project settings:
DOWNLOAD_HANDLERS = {
'http': 'scrapy_webdriver.download.WebdriverDownloadHandler',
'https': 'scrapy_webdriver.download.WebdriverDownloadHandler',
}
SPIDER_MIDDLEWARES = {
'scrapy_webdriver.middlewares.WebdriverSpiderMiddleware': 543,
}
WEBDRIVER_BROWSER = 'PhantomJS' # Or any other from selenium.webdriver
# or 'your_package.CustomWebdriverClass'
# or an actual class instead of a string.
# Optional passing of parameters to the webdriver
WEBDRIVER_OPTIONS = {
'service_args': ['--debug=true', '--load-images=false', '--webdriver-loglevel=debug']
}
In order to have webdriver handle your downloads, use the provided
class scrapy_webdriver.http.WebdriverRequest
in place of the stock scrapy
Request
, like so:
from scrapy_webdriver.http import WebdriverRequest
yield WebdriverRequest('http://www.example.com')
Parameters not supported (yet?) are: method
, body
, headers
, cookies
.
Pull requests much welcome. Just make sure the tests still pass, and add to them as necessary:
python setup.py test