Python爬虫实作篇

本次以Dcard宠物版为例

  1. 程式码撰写逻辑
    1. 先以宠物版首页将所以文章连结爬下来存到list里,那就会考虑到浏览器往下滑换页,所以要送一个GET
    2. 从list将连结一个一个取出来组合成正确的网址,并GET下来,寻找图片的标签
    3. 如果有图片的标签就在本地端新增档案,并将图片存档
    4. 将每一个显示在页面是给使用者看得,都存到一个记事本中当作图片目录

2.先将Python套件import进来requests BeautifulSoup json

import requests
from bs4 import BeautifulSoup
import json

3.一开始需要先GET宠物版首页,并使用BeautifulSoup处理成html样式,还有开一个txt存目录

test = open("spider/pet/test.txt","w",encoding='UTF-8')

p = requests.Session()
url=requests.get("https://www.dcard.tw/f/pet")
soup = BeautifulSoup(url.text,"html.parser")

4.将首页文章的连结网址存入list中

sel = soup.select("div.PostList_wrapper_2BLUM a.PostEntry_root_V6g0r")
a=[]
for s in sel:
    a.append(s["href"])
url = "https://www.dcard.tw"+ a[2]
5.根据把首页往下滑,发现网站会对server送一个GET请求下一个30篇文章
for k in range(0,10):
    post_data={
        "before":a[-1][9:18],
        "limit":"30",
        "popular":"true"
    }
    r = p.get("https://www.dcard.tw/_api/forums/pet/posts",params=post_data, headers = { "Referer": "https://www.dcard.tw/", "User-Agent": "Mozilla/5.0" })
6.发现GET来的档案格式是JSON,那小编比较不会处理JSON怎办?可以把他转成Python就好啦,并组合好网址放进list
data2 = json.loads(r.text)
    for u in range(len(data2)):
        Temporary_url = "/f/pet/p/"+ str(data2[u]["id"]) + "-" + str(data2[u]["title"].replace(" ","-"))
        a.append(Temporary_url)
7.接下来就是将list里的URL,GET网页出来
j=0
q=0
for i in a[2:]:
    url = "https://www.dcard.tw"+i
    j+=1
    print ("第",j,"页的URL为:"+url)
    #file.write("temperature is {} wet is {}%\n".format(temperature, humidity))
    test.write("第 {} 页的URL为: {} \n".format(j,url))
    url=requests.get(url)
    soup = BeautifulSoup(url.text,"html.parser")
8.运用BeautifulSoup查看是否有符合图片的标签,然后运用上次所学,将图片存档
sel_jpg = soup.select("div.Post_content_NKEl9 div div div img.GalleryImage_image_3lGzO")
for c in sel_jpg:
    q+=1
    print("第",q,"张:",c["src"])
    test.write("%\n""第 {} 张: {} \n".format(q,c["src"])) 
    pic=requests.get(c["src"])
    img2 = pic.content
    pic_out = open("spider/pet/"+str(q)+".png",'wb')
    pic_out.write(img2)
    pic_out.close()

9.记得将刚刚的txt目录关档,并写一个爬虫结束提醒自己

test.close()
print("爬重结束")
10.完整的程式码(可改range将文章变多)
#藉由首页取得所有文章的URL
import requests
from bs4 import BeautifulSoup
import json

test = open("spider/pet/test.txt","w",encoding='UTF-8')


p = requests.Session()
url=requests.get("https://www.dcard.tw/f/pet")
soup = BeautifulSoup(url.text,"html.parser")
sel = soup.select("div.PostList_wrapper_2BLUM a.PostEntry_root_V6g0r")
a=[]
for s in sel:
    a.append(s["href"])
url = "https://www.dcard.tw"+ a[2]

for k in range(0,10):
        post_data={
            "before":a[-1][9:18],
            "limit":"30",
            "popular":"true"
        }
        r = p.get("https://www.dcard.tw/_api/forums/pet/posts",params=post_data, headers = { "Referer": "https://www.dcard.tw/", "User-Agent": "Mozilla/5.0" })
        data2 = json.loads(r.text)
        for u in range(len(data2)):
            Temporary_url = "/f/pet/p/"+ str(data2[u]["id"]) + "-" + str(data2[u]["title"].replace(" ","-") )
            a.append(Temporary_url)
j=0 #为了印页数
q=0 #为了印张数
for i in a[2:]:
    url = "https://www.dcard.tw"+i
    j+=1
    print ("第",j,"页的URL为:"+url)
    #file.write("temperature is {} wet is {}%\n".format(temperature, humidity))
    test.write("第 {} 页的URL为: {} \n".format(j,url))
    url=requests.get(url)
    soup = BeautifulSoup(url.text,"html.parser")
    sel_jpg = soup.select("div.Post_content_NKEl9 div div div img.GalleryImage_image_3lGzO")
    for c in sel_jpg:
        q+=1
        print("第",q,"张:",c["src"])
        test.write("%\n""第 {} 张: {} \n".format(q,c["src"])) 
        pic=requests.get(c["src"])
        img2 = pic.content
        pic_out = open("spider/pet/"+str(q)+".png",'wb')
        pic_out.write(img2)
        pic_out.close()

test.close()
print("爬虫结束")




  • 10
    点赞
  • 5
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值