python中的scrapy爬虫_python 中scrapy爬虫问题

items模块如下:

import scrapy

class DoubanBookItem(scrapy.Item):

# define the fields for your item here like:

# name = scrapy.Field()

name = scrapy.Field()

price = scrapy.Field()

publisher = scrapy.Field()

ratings = scrapy.Field()

edition_year = scrapy.Field()

author = scrapy.Field()

paquspider.py 文件如下:

import scrapy

from douban_book.items import DoubanBookItem

class BookSpider(scrapy.Spider):

"""docstring for BookSpider"""

name = 'douban-book'

allowed_domain = ['douban.com']

start_urls = ['https://book.douban.com/top250']

def parse(self, response):

yield scrapy.Request(response.url, callback = self.parse_page)

for page in response.xpath('//div[@class="paginator"]/a'):

link = page.xpath('@href').extract()[0]

yield scrapy.Request(link, callback = self.parse_page)

def parse_page(self, response):

for item in response.xpath('//tr[@class="item"]'):

book = DoubanBookItem()

book['name'] = item.xpath('td[2]/div[1]/a/@title').extract()[0]

book['ratings'] = item.xpath('td[2]/div[2]/span[@class="rating_nums"]/text()').extract()[0]

# book['ratings'] = item.xpath('td[2]/div[2]/span[2]/text()').extract()[0]

book_info = item.xpath('td[2]/p[1]/text()').extract()[0]

book_info_contents = book_info.strip().split(' / ')

book['author'] = book_info_contents[0]

book['publisher'] = book_info_contents[1]

book['edition_year'] = book_info_contents[2]

book['price'] = book_info_contents[3]

yield book

错误:C:\anacoda\python.exe D:/scrapy/douban_book/douban_book/spiders/bookspider.py.py

Traceback (most recent call last):

File "D:/scrapy/douban_book/douban_book/spiders/bookspider.py.py", line 6, in

from douban_book.items import DoubanBookItem

ModuleNotFoundError: No module named 'douban_book'

这是怎么回事呀?谢谢

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值