如何将上图一天的csv文件批量下载下来?首先我们点击一天csv文件链接,网页会自动弹出一个方框如下图,我们复制一天数据链接http://tkkankyo.eng.niigata-u.ac.jp/HP/HP/database/japan2/01/010106/tem/1day/(复制的链接不是全部,后面csv文件(如:PTA060211DX.csv)的名字不需要copy。然后直接上代码批量下载。在这里插入代码片)
import requests
import os
from bs4 import BeautifulSoup
archive_url = "http://tkkankyo.eng.niigata-u.ac.jp/HP/HP/database/japan2/01/010106/tem/1day/"#网址链接
def get_video_links():
r = requests.get(archive_url)
soup = BeautifulSoup(r.content, 'lxml')
links = soup.findAll('a')
video_links = [archive_url + link['href'] for link in links if link['href'].endswith('csv')]#如果下载zip文件,将后面的csv改为zip
# print(video.links)
return video_links
def download_video_series(video_links):
for link in video_links:
file_name = link.split('/')[-1]
DIR_ROOT='C:\\Users\\lihuaibao\\Desktop\\li\\'下载到本地自己新建的文件夹
file_name=DIR_ROOT+file_name
print("Downloading file:%s" % file_name)
r = requests.get(link, stream=True)
# download started
with open(file_name, 'wb') as f:
for chunk in r.iter_content(chunk_size=1024 * 1024):
if chunk:
f.write(chunk)
#将读取的第一个CSV文件写入合并后的文件保存
print("%s downloaded!\n" % file_name)
print("All videos downloaded!")
return
if __name__ == "__main__":
video_links = get_video_links()
download_video_series(video_links)