#想获取贴吧的名称
import urllib.request
import urllib.parse
from lxml import etree
url=“http://tieba.baidu.com/f?ie=utf-8&”
data={
‘kw’:‘python’,
‘pn’:‘0’,
}
data=urllib.parse.urlencode(data)
url=url+data
headers={‘User-Agent’: ‘Mozilla/5.0(Windows NT 10.0; WOW64; rv:61.0) Gecko/20100101 Firefox/61.0’,}
req = urllib.request.Request(url, headers=headers)
res = urllib.request.urlopen(req,timeout=1)
str=res.read().decode(‘utf-8’)
etr=etree.HTML(str)
bam=etr.xpath(’//div[@class=“card_title”]//a/text()’)
print(bam)
python带haders得不到想要的页面,直接用urlopen()就可以,不懂是什么原因,求大神们指教
最新推荐文章于 2024-04-17 21:31:09 发布