from queue import Queue
import urllib3
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
class CrawlQueue(object):
def __init__(self, items, args):
self.max_failure = 10
self.num_success = 0
self.num_tol = len(items)
self.args = args
self.infos = self.initInfos(items)
self.job_queue = Queue()
self.job_fail = Queue()
self.job_result = Queue()
for info in self.infos:
self.job_queue.put(info)
#####重写的方法
def parser(self, info: dict):
success = True
res = object
return success, res
def initInfos(self, items: list) -> list:
return [{"id": i + 1, "failure": 0, "item": item, "res": None} for i, item in enumerate(items)]
def handleInfo(self, info: dict):
success, res = self.parser(info)
if s
Python设计爬虫任务队列
最新推荐文章于 2024-07-25 11:26:44 发布