import requests
from lxml import etree
import lxml
import os
from bs4 import BeautifulSoup
import time
startime = time.time()
#利用异常处理在连接失败时提示连接错误并不断重试直至重新连接
def req(url):
t1 = time.time()
count = 1
headers = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:73.0) Gecko/20100101 Firefox/73.0'}
success = False
while not success:
try:
response = requests.get(url, headers=headers)
success = True
t2 = time.time()
t3 = t2 - t1
t4 = time.strftime('%M分%S秒', time.localtime(t3))
#去掉注释会显示访问网站耗时,方便查看网站是否会对频繁请求进行限制
# print("网站请求耗时:",t4)
return response
except:
if count < 2:
print("连接失败!正在重新连接网站...")
count += 1
#配合pydroid3在安卓系统使用,判断保存文件夹是否存在,不存在则创建文件
单线程python小说爬虫示例,配合pydroid3可在安卓系统使用
于 2023-05-11 10:04:43 首次发布