一个简单的批量检测url存活脚本
思路是先读取文件里面的url,然后通过requests模块去访问它,最后提取访问正常的url下载到指定文件夹
import requests
with open(r'C:\Users\惠普\Desktop\1111\链接.txt','r',encoding = 'utf-8') as f:
u = f.readlines()
length = len(u)
for i in range(0,length):
a = u[i].strip()
try:
headers = {'User-Agent':'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:6.0) Gecko/20100101 Firefox/6.0'}
r = requests.get(u[i].strip(),headers = headers,timeout = 5 ).status_code
if r != 200:
print('无法连接'+':'+a)
else:
with open(r'C:\Users\惠普\Desktop\1111\连接正常url.txt' ,'a') as f:#提取访问正常的文件下载路径
f.write(a+'\n')
except:
print('连接超时'+':'+a)