爬虫——爬取7天天气状况

import requests
from lxml import etree

# 网页地址
    # http://www.weather.com.cn/weather1d/103163100.shtml
    # http://www.weather.com.cn/weather1d/101010100.shtml
    #http://www.weather.com.cn/weather1d/101030100.shtml
    #http://www.weather.com.cn/weather1d/101120101.shtml
    #http://www.weather.com.cn/weather1d/101120201.shtml
    #http://www.weather.com.cn/weather1d/101020100.shtml
    #http://www.weather.com.cn/weather1d/101260101.shtml
url = "http://www.weather.com.cn/weather/101120701.shtml"
headers = {
'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/84.0.4147.135 Safari/537.36'
}

# 准备请求头

response = requests.get(url,headers = headers)

# 获得html文件
html = etree.HTML(response.text)
# print(html)
day_list = html.xpath('//ul[@class="t clearfix"]/li/h1/text()')
wea_list = html.xpath('//ul[@class="t clearfix"]/li/p[@class="wea"]/text()')
tem_list1 = html.xpath('//ul[@class="t clearfix"]/li/p[@class="tem"]/span/text()')
tem_list2 = html.xpath('//ul[@class="t clearfix"]/li/p[@class="tem"]/i/text()')
tem_list = []
for i in range(len(tem_list1)):
    tem_list.append(tem_list1[i]+'/'+tem_list2[i])
list = []

# 显示出来爬取的数据
for i in range(len(tem_list1)):
    str1 = (day_list[i]+':'+wea_list[i]+tem_list[i]).encode("ISO-8859-1").decode('utf-8')
    with open('wea.txt','a',encoding='utf-8') as f:
        f.write(str1)
        f.write('\n')


 

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值