request = urllib2.Request(url,headers = headers)
response = urllib2.urlopen(request)
response = urllib2.urlopen(request)
content = response.read()
soup = BeautifulSoup(content,'html.parser',from_encoding='utf-8')node = soup.find_all('div',class_="article block untagged mb15")
print node
抓取页面的编码是utf-8,但使用find_all()方法输出node后就会产生乱码:
解决办法
这里FindALL是一个list,要输应该循环输出
for _ in FindALL:
print _.encode('utf-8')