import requests
from lxml import etree
import json
from queue import Queue
import threading
class Qiubai:
def __init__(self):
self.headers = {
"User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_12_5) AppleWeb\
Kit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.36"}
self.url_queue = Queue() # 实例化三个队列,用来存放内容
self.html_queue = Queue()
self.content_queue = Queue()
def get_total_url(self):
'''
获取了所有的页面url,并且返回urllist
return :list
'''
url_temp = 'https://www.qiushibaike.com/8hr/page/{}/'
url_list = []
for i in range(1, 36):
# url_list.append(url_temp.format(i))
self.url_queue.put(url_temp.format(i))
def parse_url(self):
'''
一个发送请求,获取响应,同时etree处理html
'''
while self.url_queue.not_empty:
url = self.url_queue.get()
print("parsing url:", url)
response = requests.get(url, headers=self.headers, timeout=10) # 发送请求
html = response.content.decode() # 获取html字符串
html = etree.HTML(html) # 获取element 类型的html
self.html_queue.put(html)
self.url_queue.task_done()
def get_content(self):
'''
:param url:
:return: 一个list,包含一个url对应页面的所有段子的所有内容的列表
'''
while self.html_queue.not_empty:
html = self.html_queue.get()
total_div = html.xpath('//div[@class="article block untagged mb15"]') # 返回divelememtn的一个列表
items = []
for i in total_div: # 遍历div标枪,获取糗事百科每条的内容的全部信息
author_img = i.xpath('./div[@class="author clearfix"]/a[1]/img/@src')
author_img = "https:" + author_img[0] if len(author_img) > 0 else None
author_name = i.xpath('./div[@class="author clearfix"]/a[2]/h2/text()')
author_name = author_name[0] if len(author_name) > 0 else None
author_href = i.xpath('./div[@class="author clearfix"]/a[1]/@href')
author_href = "https://www.qiushibaike.com" + author_href[0] if len(author_href) > 0 else None
author_gender = i.xpath('./div[@class="author clearfix"]//div/@class')
author_gender = author_gender[0].split(" ")[-1].replace("Icon", "") if len(author_gender) > 0 else None
author_age = i.xpath('./div[@class="author clearfix"]//div/text()')
author_age = author_age[0] if len(author_age) > 0 else None
item = dict(
author_name=author_name,
author_img=author_img,
author_href=author_href,
author_gender=author_gender,
author_age=author_age,
)
items.append(item)
self.content_queue.put(items)
self.html_queue.task_done() # task_done的时候,队列计数减一
def save_items(self):
'''
保存items
:param items:列表
'''
while self.content_queue.not_empty:
items = self.content_queue.get()
f = open("qiubai.txt", "a")
for i in items:
json.dump(i, f, ensure_ascii=False, indent=2)
# f.write(json.dumps(i))
f.close()
self.content_queue.task_done()
def run(self):
# 1.获取url list
# url_list = self.get_total_url()
thread_list = []
thread_url = threading.Thread(target=self.get_total_url)
thread_list.append(thread_url)
# 发送网络请求
for i in range(10):
thread_parse = threading.Thread(target=self.parse_url)
thread_list.append(thread_parse)
# 提取数据
thread_get_content = threading.Thread(target=self.get_content)
thread_list.append(thread_get_content)
# 保存
thread_save = threading.Thread(target=self.save_items)
thread_list.append(thread_save)
for t in thread_list:
t.setDaemon(True) # 为每个进程设置为后台进程,效果是主进程退出子进程也会退出
t.start() # 为了解决程序结束无法退出的问题
#
# for t in thread_list:
# t.join()
self.url_queue.join() # 让主线程等待,所有的队列为空的时候才能退出
self.html_queue.join()
self.content_queue.join()
if __name__ == "__main__":
qiubai = Qiubai()
qiubai.run()
自己看的文章
最新推荐文章于 2024-07-28 15:46:11 发布