scrapy mongo pipeline
作者:互联网
import pymongo
db_configs = {
'type': 'mongo',
'host': '127.0.0.1',
'port': '27017',
"user": "",
"password": "",
'db_name': 'spider'
}
class MongoPipeline():
def __init__(self):
self.db_name = db_configs.get("db_name")
self.host = db_configs.get("host")
self.port = db_configs.get("port")
self.username = db_configs.get("user")
self.password = db_configs.get("passwd")
def open_spider(self, spider):
self.client = pymongo.MongoClient('mongodb://{}:{}'.format(self.host, self.port), connect=False, maxPoolSize=10)
if self.username and self.password:
self.db = self.client[self.db_name].authenticate(self.username, self.password)
self.db = self.client[self.db_name]
def close_spider(self, spider):
self.client.close()
def process_item(self, item, spider):
collection_name = spider.name
self.db[collection_name].update_one({"url": item["url"]}, {'$set': item}, upsert=True)
return item
标签:pipeline,name,get,self,db,spider,scrapy,configs,mongo 来源: https://www.cnblogs.com/c-x-a/p/14411831.html