I am new to all instruments here. My goal is to extract all URLs from a lot of pages which are connected moreless by a "Weiter"/"next" button - that for several URLS. I decided to try that with scrapy. The page is dynamically generated. Then I learned that I need a further instrument and installed Splash for that. The installation is working. I set up the installation according to the tutorials. Then I managed to get the first page by send a "return" in the search-input-field. With a browser that gives me the results I need. My problem is that I try to click the "next" button on the generated page and don't know exactly how. As I ve read on several pages this was not always easy. I tried the suggested solutions without success. I think I am not too far away and would appreciate some help. Thank you.
my settings.py
BOT_NAME = 'gr'
SPIDER_MODULES = ['gr.spiders']
NEWSPIDER_MODULE = 'gr.spiders'
ROBOTSTXT_OBEY = True
DOWNLOADER_MIDDLEWARES = {
'scrapy_splash.SplashCookiesMiddleware': 723,
'scrapy_splash.SplashMiddleware': 725,
'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware': 810,
}
SPLASH_URL = 'http://localhost:8050'
SPIDER_MIDDLEWARES = {
'scrapy_splash.SplashDeduplicateArgsMiddleware': 100,
}
DUPEFILTER_CLASS = 'scrapy_splash.SplashAwareDupeFilter'
HTTPCACHE_STORAGE = 'scrapy_splash.SplashAwareFSCacheStorage'
my spider:
# -*- coding: utf-8 -*-
import scrapy
from scrapy.linkextractors import LinkExtractor
from scrapy_splash import SplashRequest
import json
# import base64
class GrSpider(scrapy.Spider):
name = 'gr_'
allowed_domains = ['lawsearch.gr.ch']
start_urls = ['http://www.lawsearch.gr.ch/le/']
def start_requests(self):
script = """
function main(splash)
assert(splash:go(splash.args.url))
splash:set_viewport_full()
splash:wait(0.3)
splash:send_keys("<Return>")
splash:wait(0.3)
return splash:html()
end
"""
for url in self.start_urls:
yield SplashRequest(url=url,
callback=self.parse,
endpoint='execute',
args={'lua_source': script})
def parse(self, response):
script3 = """
function main(splash)
splash:autoload{url="https://code.jquery.com/jquery-3.2.1.min.js"}
assert(splash:go(splash.args.url))
splash:set_viewport_full()
-- splash:wait(2.8)
-- local element = splash:select('.result-pager-next-active .simplebutton')
-- element:mouse_click()
-- local bounds = element:bounds()
-- assert(element:mouse_click{x=bounds.width, y=bounds.height})
-- naechster VERSCUH
-- https://blog.scrapinghub.com/2015/03/02/handling-javascript-in-scrapy-with-splash/
-- https://mcmap.net/q/1919192/-scrapyjs-splash-click-controller-button
-- assert(splash:runjs("$('#date-controller > a:first-child').click()"))
-- https://github.com/scrapy-plugins/scrapy-splash/issues/27
-- assert(splash:runjs("$('#result-pager-next-active .simplebutton').click()"))
-- https://developer.mozilla.org/de/docs/Web/API/Document/querySelectorAll
-- ANSCHAUEN
-- https://mcmap.net/q/1739260/-splash-lua-script-to-do-multiple-clicks-and-visits
-- elementList = baseElement.querySelectorAll(selectors)
-- var domRect = element.getBoundingClientRect();
-- var rect = obj.getBoundingClientRect();
-- https://mcmap.net/q/117437/-quot-queryselectorall-quot-with-multiple-conditions-in-javascript
local get_dimensions = splash:jsfunc([[
function () {
var doc1 = document.querySelectorAll("result-pager-next-active.simplebutton")[0];
var el = doc1.documentElement;
var rect = el.getClientRects()[0];
return {'x': rect.left, 'y': rect.top}
}
]])
-- splash:set_viewport_full()
splash:wait(0.1)
local dimensions = get_dimensions()
splash:mouse_click(dimensions.x, dimensions.y)
-- splash:runjs("document.querySelectorAll('result-pager-next-active ,simplebutton')[1].click()")
-- assert(splash:runjs("$('.result-pager-next-active .simplebutton')[1].click()"))
-- assert(splash:runjs("$('.simplebutton')[12].click()"))
splash:wait(1.6)
return splash:html()
end
"""
for teil in response.xpath('//div/div/div/div/a'):
yield {
'link': teil.xpath('./@href').extract()
}
next_page = response.xpath('//div[@class="v-label v-widget simplebutton v-label-simplebutton v-label-undef-w"]').extract_first()
# print response.body
print '----------------------'
# print response.xpath('//div[@class="v-slot v-slot-simplebutton"]/div[contains(text(), "Weiter")]').extract_first()
# print response.xpath('//div[@class="v-slot v-slot-simplebutton"]/div[contains(text(), "Weiter")]').extract()
# class="v-slot v-slot-simplebutton"
# nextPage = HtmlXPathSelector(response).select("//div[@class='paginationControl']/a[contains(text(),'Link Text Next')]/@href")
# neue_seite=response.url
# print response.url
if next_page is not None:
# yield SplashRequest(url=neue_seite,
yield SplashRequest(response.url,
callback=self.parse,
dont_filter=True,
endpoint='execute',
args={'lua_source': script3})