import requests from lxml import etree url = 'https://www.runoob.com/python/python-exercise-example%s.html' #网址的爬取 def get_element(url): headers = { 'cookie': '__gads=Test; Hm_lvt_3eec0b7da6548cf07db3bc477ea905ee=1573454862,1573470948,1573478656,1573713819; Hm_lpvt_3eec0b7da6548cf07db3bc477ea905ee=1573714018; SERVERID=fb669a01438a4693a180d7ad8d474adb|1573713997|1573713863', 'referer': 'https://www.runoob.com/python/python-100-examples.html', 'user-agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/77.0.3865.90 Safari/537.36' } res = requests.get(url, headers=headers) return etree.HTML(res.text) #文件的写入 def write_py(i, text): with open('练习实例%s.py' % i, 'w', encoding='utf-8') as file: file.write(text) def main(): for i in range(1, 101): html = get_element(url % i) content = '题目:' + html.xpath('//div[@id="content"]/p[2]/text()')[0] + '\n' ans = html.xpath('//div[@id="content"]/p[position()>=2]/text()')[0] daima = ''.join(html.xpath('//div[@class="hl-main"]/span/text()')) + '\n' a = '"""\n' + content + ans + daima + '\n"""' write_py(i,a) print(ans) if __name__ == '__main__': main()
学习爬虫第十二天
最新推荐文章于 2024-07-12 16:42:46 发布