# -*- coding: utf-8 -*-
"""
加载cookies文件,使用requests库爬取数据并动态更新cookies,可以使cookies不失效
"""
import pickle
import time
import requests
import random
class Spider:
def __init__(self,domain='51job.com'):
self.headers_51job={
'Host': 'ehire.51job.com',
'Origin':'http://ehire.51job.com',
'User-Agent':"""Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/56.0.2924.87 Safari/537.36""",
'Accept': """text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8""",
'Accept-Language': """zh-CN,zh;q=0.8,en-US;q=0.5,en;q=0.3""",
'Referer': """http://ehire.51job.com/Jobs/JobSearchPost.aspx?IsHis=N""",
'Upgrade-Insecure-Requests':'1',
'Connection': 'keep-alive'
}
self.s=requests.Session()
self.s.headers=self.headers_51job
self.__domain = domain
self.timeOut=30
self.cookies={}
def SetLoginDomain(self,domain='51job.com'):
"""设置登录域名"""
self.__domain=domain
return self.__domain
def SetTimeOut(self,timeOut=30):
self.__timeOut=timeOut
return self.__timeOut
def set_cookies(self):
"""读取cookie文件 该文件由另外一个登录程序获取"""
with open('./cookies.txt') as f:
cookies = pickle.loads(f.read())
for cookie in cookies:
self.cookies[cookie['name']]=cookie['value']
self.s.cookies.update(self.cookies)
def open_url(self, url,data=None):
"""页面请求方法"""
# 请求页面方法
MaxTryTimes = 20
waite_time = random.uniform(0, 1) # 初始化等待时间
for i in range(MaxTryTimes):
time.sleep(waite_time)
try:
req = self.s.post(url,data=data,headers=self.headers_51job,timeout=self.timeOut)
content=req.text
if req.cookies.get_dict():
self.s.cookies.update(req.cookies)
break
except Exception,e:
print e
content = ''
return content
if __name__ == '__main__':
spider=Spider()
spider.set_cookies()
content=spider.open_url(url='http://ehire.51job.com/Jobs/JobSearchPost.aspx?IsHis=N')
print content
爬虫时保持cookies一直有效的方法
最新推荐文章于 2024-09-16 10:37:31 发布