scrapy 一个项目里同时运行多个爬虫
作者:互联网
在spiders文件同级建立一个commands文件夹,建立一个py文件,我自定义为crawlall.py。
from scrapy.commands import ScrapyCommand class Command(ScrapyCommand): requires_project = True def syntax(self): return '[options]' def short_desc(self): return 'Runs all of the spiders' def run(self, args, opts): spider_list = self.crawler_process.spiders.list() for name in spider_list: self.crawler_process.crawl(name, **opts.__dict__) self.crawler_process.start()
在settings文件里把刚建立的crawlall文件的路径设置好
COMMANDS_MODULE = "ProxyPool.commands"
最后在cmd下 scrapy crawlall 运行
来自:https://blog.csdn.net/u014248032/article/details/83351291
标签:commands,多个,process,self,爬虫,scrapy,crawlall,spiders 来源: https://www.cnblogs.com/yoyowin/p/12403156.html