import requests
import csv
import random
import time
from fake_useragent import UserAgent
UA=UserAgent().random
url = "http://www.xinfadi.com.cn/getPriceData.html"
dic = {
"limit": "",
"current": "",
"pubDateStartTime": "",
"pubDateEndTime": "",
"prodPcatid": "",
"prodCatid": "",
"prodName": "",
}
headers={
"User-Agent":UA
}
r1 = requests.post(url, data=dic,headers=headers)
resp = r1.json()
print(resp)
all_count = int(resp["count"])
limit = int(resp["limit"])
all_page_number = int(all_count / limit)
with open("新发地菜价.csv", mode="a+", newline='') as f:
csvwriter = csv.writer(f)
for i in range(1, all_page_number):
# dic1 = {
# "limit": limit,
# "current": i,
# "pubDateStartTime": "",
# "pubDateEndTime": "",
# "prodPcatid": "",
# "prodCatid": "",
# "prodName": "",
# }
r1 = requests.post(url, data=dic,headers=headers)
list = resp["list"]
# count = list
for iter in list:
prodName = iter["prodName"]
avgPrice = iter["avgPrice"]
highPrice = iter["highPrice"]
lowPrice = iter["lowPrice"]
place = iter["place"]
prodCat = iter["prodCat"]
pubDate = iter["pubDate"]
unitInfo = iter["unitInfo"]
csvwriter.writerow([prodName, avgPrice, highPrice, lowPrice, place, prodCat, pubDate, unitInfo])
time.sleep(float(format(random.uniform(0, 1), '.2f')))
print("Over")
爬取新发地菜价信息并保存到csv文件
于 2022-05-06 23:48:24 首次发布