豆瓣# coding:utf - 8
from urllib.request import urlopen
from bs4 import BeautifulSoup
html = urlopen("https://movie.douban.com/")
bsObj = BeautifulSoup(html, "lxml") # 将html对象转化为BeautifulSoup对象
liList = bsObj.findAll("li", {"class": "title"}) # 找到所有符合此class属性的li标签
for li in liList:
name = li.a.get_text() # 获取标签中文字
print(name)
简书# -*- coding:utf-8 -*-
from urllib import request
from bs4 import BeautifulSoup
url = r‘http://www.jianshu.com‘
# 模拟真实浏览器进行访问
headers = {‘User-Agent‘:‘Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/55.0.2883.87 Safari/537.36‘}
page = request.Request(url, headers=headers)
page_info = request.urlopen(page).read()
page_info = page_info.decode(‘utf-8‘)
# 将获取到的内容转换成BeautifulSoup格式,并将html.parser作为解析器
soup = BeautifulSoup(page_info, ‘lxml‘)
# 以格式化的形式打印html
# print(soup.prettify())
titles = soup.find_all(‘a‘, ‘title‘) # 查找所有a标签中class=‘title‘的语句
# 打印查找到的每一个a标签的string
for title in titles:
print(title.string)
酷狗def getInfo(self, url):
html = requests.get(url, headers=self.header)
soup = BeautifulSoup(html.text, ‘html.parser‘)
# print(soup.prettify())
ranks = soup.select(‘.pc_temp_num‘)
titles = soup.select(‘.pc_temp_songlist > ul > li > a‘) # 层层标签查找
times = soup.select(‘.pc_temp_time‘)
for rank, title, songTime in zip(ranks, titles, times):
data = {
# rank 全打印就是带HTML标签的
‘rank‘: rank.get_text().strip(),
‘title‘: title.get_text().split(‘-‘)[1].strip(),
‘singer‘: title.get_text().split(‘-‘)[0].strip(),
‘songTime‘: songTime.get_text().strip()
}
s = str(data)
print(‘rank:%2s\t‘ % data[‘rank‘], ‘title:%2s\t‘ % data[‘title‘], ‘singer:%2s\t‘ %data[‘singer‘], ‘songTime:%2s\t‘ % data[‘songTime‘])
with open(‘hhh.txt‘, ‘a‘, encoding=‘utf8‘) as f:
f.writelines(s + ‘\n‘)
【更多参考】