scrapy 爬虫过滤相同的url,Filtered duplicate request,dont

xiaoxiao2021-02-28  67

yield Request('https://www.zhihu.com', meta={'cookiejar':response.meta['cookiejar']}, headers=self.headers_zhihu, callback=self.parse_index, dont_filter=True )

scrapy默认过滤掉重复的之前爬过的url,在request参数中添加dont_filter=True 设置不过滤url

转载请注明原文地址: https://www.6miu.com/read-81608.html

最新回复(0)