python scrapy 重复执行
作者:互联网
from twisted.internet import reactor, defer from scrapy.crawler import CrawlerRunner from scrapy.utils.log import configure_logging import time import logging from scrapy.utils.project import get_project_settings #在控制台打印日志 configure_logging() #CrawlerRunner获取settings.py里的设置信息 runner = CrawlerRunner(get_project_settings()) @defer.inlineCallbacks def crawl(): while True: logging.info("new cycle starting") yield runner.crawl("xxxxx") #1s跑一次 time.sleep(1) reactor.stop() crawl() reactor.run()
标签:logging,reactor,python,CrawlerRunner,重复,scrapy,import,crawl 来源: https://www.cnblogs.com/winstonsias/p/12106667.html