import requests
import pandas as pd
from bs4 import BeautifulSoup
import time
import csv
#构建网页
def get_url(url):
for i in range(2,3):
url=“http://www.boc.cn/sourcedb/whpj/index_"+str(i)+".html”
return url
#爬取网页
def get_html(url):
try:
r=requests.get(url,timeout=3.5)
r.raise_for_status()
r.encoding=r.apparent_encoding
html=r.text
return html
except:
print(‘无法爬取’)
#解析网页
def get_data(name_lsts,html):
soup=BeautifulSoup(html,“html.parser”)
content = soup.prettify()
ths=soup.find_all(‘th’)
#提取表名
name_lst=[]
for th in ths:
name_lst.append(th.string)
name_lsts.append(name_lst)
def get_da(data_lsts,html):
soup=BeautifulSoup(html,“html.parser”)
cont