- 博客(6)
- 收藏
- 关注
原创 requests多进程爬取今日头条街拍--记录
spider.py#!/usr/bin/env python#-*- coding:utf-8 -*-import requestsimport reimport jsonimport pymongoimport osfrom requests.exceptions import RequestExceptionfrom urllib.parse import urlen
2018-01-13 07:03:36
1354
转载 requests爬猫眼电影 -- 记录
#!/usr/bin/env python#-*- coding:utf-8 -*-import requestsimport refrom requests.exceptions import RequestExceptionfrom json import dumpsdef get_one_page(url): try: response = requests.
2018-01-13 02:13:48
294
原创 requests爬小猪租房存入Mongodb--记录
#!/usr/bin/env python#-*- coding:utf-8 -*-import requests,time,pymongofrom bs4 import BeautifulSoupheaders = { "User-Agent":'Mozilla/5.0 (X11; U; Linux x86_64; zh-CN; rv:1.9.2.10) Gecko/20100922
2018-01-03 14:38:58
337
原创 requests爬取图片保存--记录
#!/usr/bin/env python#-*- coding:utf-8 -*-import requests,time,osfrom bs4 import BeautifulSoupheaders = { "User-Agent":'Mozilla/5.0 (X11; U; Linux x86_64; zh-CN; rv:1.9.2.10) Gecko/20100922 Ubunt
2018-01-02 16:39:30
2713
原创 requests爬取小猪租房--记录
#!/usr/bin/env python#-*- coding:utf-8 -*-import requests,timefrom bs4 import BeautifulSoupheaders = { "User-Agent":'Mozilla/5.0 (X11; U; Linux x86_64; zh-CN; rv:1.9.2.10) Gecko/20100922 Ubuntu/1
2018-01-02 12:38:37
395
原创 Scrapy 爬豆瓣--记录
# -*- coding: utf-8 -*-import scrapyimport ConfigParserimport reclass MovieSpider(scrapy.Spider): download_delay = 1 #防止IP被封,设置爬取时间 config = ConfigParser.RawConfigParser() #RawConfigParser
2017-12-31 02:42:39
242
空空如也
空空如也
TA创建的收藏夹 TA关注的收藏夹
TA关注的人