首页
Java
登录
6mi
u
盘
搜
搜 索
Java
scrapy 爬虫过滤相同的url,Filtered duplicate request,dont
scrapy 爬虫过滤相同的url,Filtered duplicate request,dont
xiaoxiao
2021-02-28
67
yield
Request
(
'https://www.zhihu.com'
, meta={
'cookiejar'
:response
.meta[
'cookiejar'
]}, headers=
self
.headers_zhihu, callback=
self
.parse_index, dont_filter=
True
)
scrapy默认过滤掉重复的之前爬过的url,在request参数中添加dont_filter=True 设置不过滤url
转载请注明原文地址: https://www.6miu.com/read-81608.html
技术
最新回复
(
0
)