import requests
import time
from bs4 import BeautifulSoup
import os
defGet_content(url):
urls2=[]
res=requests.get(url).content.decode('gbk')
soup=BeautifulSoup(res,"html.parser")
contents=soup.find_all("div",attrs={
"class":"nav"})for content in contents:
lis=content.find_all("li")for li in lis:
urls2.append(li.a.get("href"))return urls2
defGet_url3(urls):
urls3=[]for url in urls:
res = requests.get(url).content.decode('gbk')
soup = BeautifulSoup(res,"html.parser")#
contents=soup.find_all("div",attrs={
"class":"l"})#热门小说
contents2=soup.find_all("div",attrs={
"class":"r"})#玄幻小说,仙侠小说,都市言情小说
contents3=soup.find_all("div",attrs={
"class":"novelslist"})#更新小说
contents4=soup.find_all("div",attrs={
"id":"newscontent"})for i,content inenumerate(contents):
dts=content.find_all("dt")for dt in dts:try:
urls3.append(dt.a.get("href"))except Exception as e:print(i)for c in contents2:
lis=c.find_all("li")for li in lis:
urls3.append(li.a