![](https://img-blog.csdnimg.cn/20201014180756757.png?x-oss-process=image/resize,m_fixed,h_64,w_64)
爬虫
一个程序员的自我积累
天才在于积累,优秀是一种习惯!
展开
-
不用框架初步应用爬虫
from urllib import request import re def main(): url = 'http://www.baidu.com' response = request.urlopen(url=url) html = response.read()#.decode('utf8') # p = '<title>(.*)<...原创 2018-08-27 00:13:00 · 356 阅读 · 0 评论 -
scrapy入门教程
import scrapy class QuotesSpider(scrapy.Spider): name = "quotes" start_urls = [ 'http://quotes.toscrape.com/tag/humor/', ] def parse(self, response): for quote in r...原创 2018-08-27 23:56:23 · 124 阅读 · 0 评论 -
scrapy爬取quotes
import scrapy class QuotesSpider(scrapy.Spider): name = 'quotes' def start_requests(self): urls = [ 'http://quotes.toscrape.com/page/1/', 'http://quotes.tosc...原创 2018-08-28 20:04:13 · 750 阅读 · 0 评论 -
scrapy 爬取京东商品列表
import scrapy class JdSpider(scrapy.Spider): #scrapy crawl name便是由这个name决定的 name = 'jd' def start_requests(self): url = 'https://www.jd.com/' yield scrapy.Request(url=u...原创 2018-08-28 20:58:32 · 439 阅读 · 0 评论