python爬虫数据写入文件,及下载功能实现

from django.http import HttpRequest, HttpResponse, FileResponse
from urllib.request import urlopen, Request
from urllib import parse
from lxml import etree
import ssl


#爬虫爬取写入文件
def worm(request: HttpRequest):
    urls = '/user/worm?urls=https://www.baidu.com/s'
    data = {
        'wd': '娄宇'
    }
    ua = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_0) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/75.0.3770.142 Safari/537.36'
    requestWorm = Request("{}?{}".format(urls, parse.urlencode(data)))
    requestWorm.add_header('User-Agent', ua)
    context = ssl._create_unverified_context()

    with urlopen(requestWorm, context=context) as response:
        returnValue = response.read()
        html = etree.HTML(returnValue)
        returnValue = html.xpath("//h3[@class='t c-gap-bottom-small']//text()")
        
        with open('./data.txt', 'ab+') as file:
            for i in returnValue:
                s=bytes(i,encoding='utf-8')
                file.write(s)
        return HttpResponse(returnValue)

 #下载功能
def downLoad(request: HttpRequest):
    with open('./data.txt') as f:
        file = f.read()
        response = FileResponse(file)
        response['Content-Type'] = 'application/octet-stream'
        response['Content-Disposition']='attachment;filename="models.py"'
        return response
    

 

  • 0
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值