I'm quite new to web scraping, and in particular in using Scrapy's spiders, pipelines... I'm getting some 202 status from some spider requests' response, hence the page content is not available yet How to handle these status code properly, like "wait for the page to fully load"? I saw/looked into both Scrapy's downloadermiddleware's get_retry_request and reactor's callLater with a lambda to try to handle retries with delay, but without success so far...
(e.g.
yield scrapy.downloadermiddlewares.retry.get_retry_request(
request=response.request, spider=self, reason='202 Accepted - retrying after delay', max_retry_times=self.max_retries)
or something like:
reactor.callLater(self.retry_delay,
lambda: self.crawler.engine.crawl(
scrapy.Request(
url=response.url,
callback=self.parse,
meta={'location_name': response.meta.get('location_name', ''),'retries': retries + 1})))
)
Thanks in advance for any support!