scrapy + selenium实现多层网页爬取和点击获取隐藏内容

1. 安装scrapy部分不细讲了, 要装vs环境的, 尝试了很多方式最后都不行, 最后还是老老实实花几个小时装下吧;

2. 客户端创建scrapy项目

     scrapy startproject chinaDaily

3. pycharm打开项目后创建pychon package, 会自动生成一个scrapy.cfg和__init__.py

4. 创建后项目结构

 

5.修改setting.py

ITEM_PIPELINES = {
   'chinaDaily.pipelines.ChinadailyPipeline': 300,
}

BOT_NAME = 'chinaDaily'

SPIDER_MODULES = ['chinaDaily.spiders']
NEWSPIDER_MODULE = 'chinaDaily.spiders'

DOWNLOADER_MIDDLEWARES = {
   # 'chinaDaily.middlewares.RandomUserAgent': 543,
   'chinaDaily.middlewares.JavaScriptMiddleware': 543,      #添加此行代码

# 该中间件将会收集失败的页面,并在爬虫完成后重新调度。(失败情况可能由于临时的问题,例如连接超时或者HTTP 500错误导致失败的页面)
   'scrapy.downloadermiddlewares.retry.RetryMiddleware': 80,
# 该中间件提供了对request设置HTTP代理的支持。您可以通过在 Request 对象中设置 proxy 元数据来开启代理。
    'scrapy.downloadermiddlewares.httpproxy.HttpProxyMiddleware': 100,
}

6. pipelines.py

# Define your item pipelines here
#
# Don't forget to add your pipeline to the ITEM_PIPELINES setting
# See: https://docs.scrapy.org/en/latest/topics/item-pipeline.html


# useful for handling different item types with a single interface
import json

from itemadapter import ItemAdapter


class ChinadailyPipeline:
    fp = None

    # 重写父类的一个方法:该方法只在开始爬虫的时候被调用一次
    def open_spider(self, spider):
        print('开始爬虫......')
        self.fp = open('./news.txt', 'w', encoding='utf8')

    # process_item专门用来处理item类型对象
    # 该方法item参数可以接收爬虫文件提交的item对象
    # 该方法每接收到一个item就会被调用一次
    def process_item(self, item, spider):
        title = item['title']
        content = item['content']
        self.fp.write(title + ':' + content + '\n')
        return item  # 这个item会传递给下一个即将被执行的管道类

    def close_spider(self, spider):
        print('结束爬虫')
        self.fp.close()

7. middlewares.py

# Define here the models for your spider middleware
#
# See documentation in:
# https://docs.scrapy.org/en/latest/topics/spider-middleware.html
import time
from urllib import request

from scrapy import signals

# useful for handling different item types with a single interface
from itemadapter import is_item, ItemAdapter
from scrapy.http import HtmlResponse
from selenium import webdriver


class ChinadailySpiderMiddleware:
    # Not all methods need to be defined. If a method is not defined,
    # scrapy acts as if the spider middleware does not modify the
    # passed objects.

    @classmethod
    def from_crawler(cls, crawler):
        # This method is used by Scrapy to create your spiders.
        s = cls()
        crawler.signals.connect(s.spider_opened, signal=signals.spider_opened)
        return s

    def process_spider_input(self, response, spider):
        # Called for each response that goes through the spider
        # middleware and into the spider.

        # Should return None or raise an exception.
        return None

    def process_spider_output(self, response, result, spider):
        # Called with the results returned from the Spider, after
        # it has processed the response.

        # Must return an iterable of Request, or item objects.
        for i in result:
            yield i

    def process_spider_exception(self, response, exception, spider):
        # Called when a spider or process_spider_input() method
        # (from other spider middleware) raises an exception.

        # Should return either None or an iterable of Request or item objects.
        pass

    def process_start_requests(self, start_requests, spider):
        # Called with the start requests of the spider, and works
        # similarly to the process_spider_output() method, except
        # that it doesn’t have a response associated.

        # Must return only requests (not items).
        for r in start_requests:
            yield r

    def spider_opened(self, spider):
        spider.logger.info('Spider opened: %s' % spider.name)


class ChinadailyDownloaderMiddleware:
    # Not all methods need to be defined. If a method is not defined,
    # scrapy acts as if the downloader middleware does not modify the
    # passed objects.

    @classmethod
    def from_crawler(cls, crawler):
        # This method is used by Scrapy to create your spiders.
        s = cls()
        crawler.signals.connect(s.spider_opened, signal=signals.spider_opened)
        return s

    def process_request(self, request, spider):
        # Called for each request that goes through the downloader
        # middleware.

        # Must either:
        # - return None: continue processing this request
        # - or return a Response object
        # - or return a Request object
        # - or raise IgnoreRequest: process_exception() methods of
        #   installed downloader middleware will be called
        return None

    def process_response(self, request, response, spider):
        # Called with the response returned from the downloader.

        # Must either;
        # - return a Response object
        # - return a Request object
        # - or raise IgnoreRequest
        return response

    def process_exception(self, request, exception, spider):
        # Called when a download handler or a process_request()
        # (from other downloader middleware) raises an exception.

        # Must either:
        # - return None: continue processing this exception
        # - return a Response object: stops process_exception() chain
        # - return a Request object: stops process_exception() chain
        pass

    def spider_opened(self, spider):
        spider.logger.info('Spider opened: %s' % spider.name)


class JavaScriptMiddleware(object):
    def process_request(self, request, spider):
        # executable_path = '/home/hadoop/crawlcompanyinfo0906/phantomjs'
        print("PhantomJS is starting...")
        # driver = webdriver.PhantomJS(executable_path)  # 指定使用的浏览器
        driver = webdriver.Chrome()
        driver.get(request.url)
        time.sleep(1)
        js = "var q=document.documentElement.scrollTop=10000"
        i = 1
        while (i <= 10):
            driver.find_element_by_xpath('//*[@id="page_bar0"]/div').click()
            i = i + 1
        driver.execute_script(js)  # 可执行js,模仿用户操作。此处为将页面拉至最底端。
        time.sleep(3)
        body = driver.page_source
        print("访问" + request.url)
        return HtmlResponse(driver.current_url, body=body, encoding='utf-8', request=request)

8. 创建main.py

# -*- coding: utf-8 -*-
import os
from scrapy import cmdline

dirpath = os.path.dirname(os.path.abspath(__file__))
# 获取当前路径
os.chdir(dirpath)  # 切换到当前目录
cmdline.execute(['scrapy', 'crawl', 'newsToday22'])


9. 创建spider/severalNewsSpider.py

# -*- coding: utf-8 -*-
import scrapy
import time
import numpy
import re
from chinaDaily.items import ChinadailyItem


class chinaSeveralNewsSpider(scrapy.Spider):
    name = 'newsToday22'   
    #此name就是main中执行的spider名
    allowed_domains = ['www.chinanews.com']
    start_urls = ['http://www.chinanews.com/society/']



    def parse(self, response):
        linkList=[]
        title = response.xpath('//title/text()').extract_first()
        print(title)
        div_list = response.xpath('//*[@id="ent0"]/li')
        print("^" * 20)
        print(len(div_list))
        contentresult = ''
        for div in div_list:
            # titleItem = div.xpath('.//div[@class="news_title"]/em/a/text()').extract_first().strip()
            # content = div.xpath('.//div[@class="news_content"]/a/text()').extract_first().strip()  # 返回字符串
            # print("$" * 20)
            # print(content)
            # item = ChinadailyItem()  # item类型的对象
            # item['title'] = titleItem  # auto为items定义的属性
            # item['content'] = content
            # yield item  # 将item提交给管道
            link = div.xpath('.//div[@class="news_content"]/a/@href').extract_first()
            print(link)
            linkList.append(link)

        links = div.xpath('//*[@id="ent0"]/li//div[@class="news_content"]/a/@href').extract()
        print("$" * 20)
        print(len(links))

        for linkItem in linkList:
            # url = "http://www.51porn.net" + link  # 由于提取出来的url不完整,所以需要拼接为完整的url
            yield scrapy.Request(linkItem, callback=self.parse_s,
                                 dont_filter=True)  # 请求下一层url,方法为第二个parse,dont_filter=True的作用是避免有时候新的url会被作用域过滤掉



    def parse_s(self, response):
        print("*" * 20)
        # print(response.text)
        print("*" * 20)
        title = response.xpath('//title/text()').extract_first()
        print(title)
        div_list = response.xpath('//div[@class="left_zw"]/p')
        print("^" * 20)
        print(len(div_list))
        contentresult = ''
        for div in div_list:
            content = div.xpath('./text()').extract_first().strip()   # 返回字符串
            print("+++++++++++++++++++")
            print(type(content))
            if(type(content) == str):
                print(content)
                contentresult += content
            # 将解析的数据封装存储到item类型的对象
        item = ChinadailyItem()  # item类型的对象
        item['title'] = title  # auto为items定义的属性
        item['content'] = contentresult
        yield item  # 将item提交给管道


 

  • 0
    点赞
  • 15
    收藏
    觉得还不错? 一键收藏
  • 1
    评论
评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值