新的启动事件循环
import time
import redis
import asyncio
import logging
from lxml import etree
from contextlib import closing
from aiohttp import ClientSession
logging.basicConfig(
level = logging.INFO,
format = "%(asctime)s [*] %(message)s"
)
URL = "https://www.fabiaoqing.com/biaoqing/lists/page/{}.html"
def parse(source_, red_):
try:
response = etree.HTML(source_)
tagbqppdiv = response.xpath('//div[@class="tagbqppdiv"]/a')
for index, img in enumerate(tagbqppdiv):
args = (int(index + 1), img.xpath('@title')[0], img.xpath('img/@data-original')[0])
red_.hset("mixintu", args[1], args[2])
logging.info('第 %d 张 标题:%r 链接:%r' % args)
except Exception as e:
print('解析详情页出错!')
pass
async def get_source(url_, red_):
try:
async with ClientSession( ) as session:
async with session.get(url_) as response:
source = await response.read( )
parse(source, red_)
except asyncio.CancelledError:
raise
async def main( ):
now = time.time( )
pool = redis.ConnectionPool(host = 'localhost', port = 6379, db = 1)
red = redis.Redis(connection_pool = pool)
tasks = [asyncio.ensure_future(get_source(URL.format(i), red))
for i in range(1, 11)]
dones, pendings = await asyncio.wait(tasks)
logging.info(f"总用时:{time.time( ) - now}")
if __name__ == "__main__":
with closing(asyncio.get_event_loop( )) as loop:
loop.run_until_complete(main( ))
旧的启动事件循环
import asyncio
import logging
import time
import redis
from aiohttp import ClientSession
from lxml import etree
logging.basicConfig(
level = logging.INFO,
format = "%(asctime)s [*] %(message)s"
)
URL = "https://www.fabiaoqing.com/biaoqing/lists/page/{}.html"
def parse(source_, red_):
try:
response = etree.HTML(source_)
tagbqppdiv = response.xpath('//div[@class="tagbqppdiv"]/a')
for index, img in enumerate(tagbqppdiv):
args = (int(index + 1), img.xpath('@title')[0], img.xpath('img/@data-original')[0])
red_.hset("mixintu", args[1], args[2])
logging.info('第 %d 张 标题:%r 链接:%r' % args)
except Exception as e:
print('解析详情页出错!')
pass
async def get_source(url_, red_):
try:
async with ClientSession( ) as session:
async with session.get(url_) as response:
source = await response.read( )
parse(source, red_)
except asyncio.CancelledError:
raise
async def main( ):
now = time.time( )
pool = redis.ConnectionPool(host = 'localhost', port = 6379, db = 1)
red = redis.Redis(connection_pool = pool)
tasks = [asyncio.ensure_future(get_source(URL.format(i), red))
for i in range(1, 11)]
dones, pendings = await asyncio.wait(tasks)
logging.info(f"总用时:{time.time( ) - now}")
if __name__ == "__main__":
event_loop = asyncio.get_event_loop( )
try:
event_loop.run_until_complete(main( ))
except KeyboardInterrupt:
for task in asyncio.Task.all_tasks( ):
print(task.cancel( ))
event_loop.stop( )
finally:
event_loop.close( )
![在这里插入图片描述](https://img-blog.csdnimg.cn/20191014204926463.JPG?x-oss-process=image/watermark,type_ZmFuZ3poZW5naGVpdGk,shadow_10,text_aHR0cHM6Ly9ibG9nLmNzZG4ubmV0L21peGludHU=,size_16,color_FFFFFF,t_70)