python requests请求失败_用python的requests 请求tmall的详情的一个请求报403错误的问题...

importrequestsdefget_content(url):

s= requests.get(url, headers=headers)

content= s.content.decode('gbk')print(content)returncontentif __name__=="__main__":

headers={"Host":"mdskip.taobao.com","Accept": "*/*","Accept-Encoding": "gzip, deflate, br","Accept-Language": "zh-CN,zh;q=0.8,en-US;q=0.5,en;q=0.3","Cookie":"l=AurqcPuigwQdnQv7WvAfCoR1OlrRQW7h; isg=BHp6mNB79CHqYXpVEiRteXyyyKNcg8YEwjgLqoRvCI3ddxqxbLtOFUBGwwOrZ3ad; thw=cn; cna=VsJQERAypn0CATrXFEIahcz8; t=0eed37629fe7ef5ec0b8ecb6cd3a3577; tracknick=tb830309_22; _cc_=UtASsssmfA%3D%3D; tg=0; ubn=p; ucn=unzbyun; x=e%3D1%26p%3D*%26s%3D0%26c%3D0%26f%3D0%26g%3D0%26t%3D0%26__ll%3D-1%26_ato%3D0; miid=981798063989731689; hng=CN%7Czh-CN%7CCNY%7C156; um=0712F33290AB8A6D01951C8161A2DF2CDC7C5278664EE3E02F8F6195B27229B88A7470FD7B89F7FACD43AD3E795C914CC2A8BEB1FA88729A3A74257D8EE4FBBC; enc=1UeyOeN0l7Fkx0yPu7l6BuiPkT%2BdSxE0EqUM26jcSMdi1LtYaZbjQCMj5dKU3P0qfGwJn8QqYXc6oJugH%2FhFRA%3D%3D; ali_ab=58.215.20.66.1516409089271.6; mt=ci%3D-1_1; cookie2=104f8fc9c13eb24c296768a50cabdd6e; _tb_token_=ee7e1e1e7dbe7; v=0","Referer": "https://detail.tmall.com/item.htm?id=542020037373&rn=ef4dc97f30c05c8ae00e231ca59e5e5d&abbucket=18","User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:47.0) Gecko/20100101 Firefox/47.0"}

url="https://mdskip.taobao.com/core/initItemDetail.htm?isUseInventoryCenter=false&cartEnable=true&service3C=false&isApparel=false&isSecKill=false&tmallBuySupport=true&isAreaSell=false&tryBeforeBuy=false&offlineShop=false&itemId=527909705030&showShopProm=false&cachedTimestamp=1524175497016&isPurchaseMallPage=false&isRegionLevel=false&household=false&sellerPreview=false&queryMemberRight=true&addressLevel=2&isForbidBuyItem=false"get_content(url)

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
好的,以下是一个利用 Python 爬取天猫数据的完整代码示例: ```python import requests from bs4 import BeautifulSoup # 设置请求头,模拟浏览器请求 headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.3'} # 定义爬取天猫商品信息的函数 def get_tmall_goods(keyword): # 构造请求的url链接 url = 'https://list.tmall.com/search_product.htm?q=' + keyword # 发送请求,获取响应内容 response = requests.get(url, headers=headers) html = response.text # 使用BeautifulSoup解析html soup = BeautifulSoup(html, 'html.parser') items = soup.select('.product .product-iWrap') # 遍历每个商品,获取商品名称、价格、链接等信息 for item in items: name = item.select('.productTitle a')[0].text.strip() price = item.select('.productPrice em')[0].text.strip() link = item.select('.productTitle a')[0]['href'] # 打印商品信息 print(name, price, link) # 调用函数,搜索商品信息 get_tmall_goods('手机') ``` 在上面的代码中,我们首先设置了请求头,模拟浏览器请求,避免被反爬虫机制拦截。然后定义了一个 get_tmall_goods 函数,该函数接收一个关键词作为参数,然后构造请求url链接,发送请求,获取响应内容,最后使用 Beautiful Soup 解析 html,遍历每个商品,获取商品名称、价格、链接等信息,然后打印出来。最后调用该函数,输入关键词,即可爬取对应的商品信息。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值