*【艺恩娱数】Python爬虫+数据分析可视化中国影院票房*¶



这个里面的影院名称,省份,城市,票房,场次,人次,平均票价,天数,场均人次这些数据都是我们需要的。

一、记得登入才能看到所有的数据

示例:pandas 是基于NumPy 的一种工具,该工具是为了解决数据分析任务而创建的。

二、使用步骤

进行数据抓包

在这里插入图片描述
点击数据包,找我们需要的数据

在这里插入图片描述
借用一个工具

https://spidertools.cn/#/unQuoteUrl
在这里插入图片描述
在这里插入图片描述
在这里插入图片描述

import requests


headers = {
   
    "Accept": "application/json, text/plain, */*",
    "Accept-Language": "zh-CN,zh;q=0.9,en;q=0.8",
    "Connection": "keep-alive",
    "Content-Type": "application/x-www-form-urlencoded",
    "Origin": "https://ys.endata.cn",
    "Referer": "https://ys.endata.cn/BoxOffice/Org",
    "Sec-Fetch-Dest": "empty",
    "Sec-Fetch-Mode": "cors",
    "Sec-Fetch-Site": "same-origin",
    "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/119.0.0.0 Safari/537.36",
    "sec-ch-ua": "\"Google Chrome\";v=\"119\", \"Chromium\";v=\"119\", \"Not?A_Brand\";v=\"24\"",
    "sec-ch-ua-mobile": "?0",
    "sec-ch-ua-platform": "\"Windows\""
}
cookies = {
   
    "JSESSIONID": "595a72d9-2874-4819-b50e-6ae08914f501",
    "route": "65389440feb63b53ee0576493abca26d",
    "Hm_lvt_82932fc4fc199c08b9a83c4c9d02af11": "1701861215,1702188683",
    "Hm_lpvt_82932fc4fc199c08b9a83c4c9d02af11": "1702191042"
}
url = "https://ys.endata.cn/enlib-api/api/cinema/getcinemaboxoffice_day_list.do"
data = {
   
    "r": "0.2244452227327136",
    "bserviceprice": "0",
    "datetype": "Day",
    "date": "2023-12-09",
    "sdate": "2023-12-09",
    "edate": "2023-12-09",
    "citylevel": "",
    "lineid": "",
    "columnslist": "100,101,102,121,122,103,104,108,123,109",
    "pageindex": "3",
    "pagesize": "20",
    "order": "102",
    "ordertype": "desc"
}
response = requests.post(url, headers=headers, cookies=cookies, data=data)

print(response.text)
print(response)
{
   "status":1,"des":"成功","userstatus":1,"version":0,"data":{
   "table2":[{
   "TotalPage":575,"TotalCounts":11494}],"table1":[{
   "CinemaName":"西安高新万达广场店","ShowCount":102,"ProvinceName":"陕西省","EnbaseID":14865,"AvgBoxOffice":46.88,"CinemaID":12109,"BoxOffice":105954.52,"AudienceCount":2260,"AvgShowAudienceCount":22,"CityName":"西安市","Irank":41,"ShowDay":1},{
   "CinemaName":"呼和浩特万达广场店","ShowCount":53,"ProvinceName":"内蒙古","EnbaseID":2340,"AvgBoxOffice":38.11,"CinemaID":1022,"BoxOffice":105916.87,"AudienceCount":2779,"AvgShowAudienceCount":52,"CityName":"呼和浩特市","Irank":42,"ShowDay":1},{
   "CinemaName":"宁波博纳国际影城(北仑店)","ShowCount":76,"ProvinceName":"浙江省","EnbaseID":7529,"AvgBoxOffice":41.42,"CinemaID":6026,"BoxOffice":104799.56,"AudienceCount":2530,"AvgShowAudienceCount":33,"CityName":"宁波市","Irank":43,"ShowDay":1},{
   "CinemaName":"博纳国际影城秦皇岛茂业IMAX店","ShowCount":60,"ProvinceName":"河北省","EnbaseID":21036,"AvgBoxOffice":41.18,"CinemaID":16711,"BoxOffice":104507.85,"AudienceCount":2538,"AvgShowAudienceCount":42,"CityName":"秦皇岛市","Irank":44,"ShowDay":1},{
   "CinemaName":"星轶STARX影城(上海奉贤龙湖天街店)","ShowCount":43,"ProvinceName":"上海市","EnbaseID":22522,"AvgBoxOffice":44.63,"CinemaID":18208,"BoxOffice":103412.53,"AudienceCount"
当你想要使用Python的requests库抓取网页据时,首先需要安装这个库,如果你还没有安装,可以使用pip命令安装: ```bash pip install requests ``` 以下是一个简单的示例,展示如何编写程序抓取包含特定Python关键词(比如"Python教程")的搜索结果页面的据: ```python import requests from bs4 import BeautifulSoup # 使用BeautifulSoup解析HTML # 定义你要抓取的搜索引擎URL,这里假设是Google搜索 search_url = "https://www.google.com/search?q=Python%E6%95%99%E7%A8%8B" # 发送GET请求获取网页内容 response = requests.get(search_url) # 检查请求是否成功,状态码200表示成功 if response.status_code == 200: # 解析响应内容为HTML soup = BeautifulSoup(response.text, 'html.parser') # 使用BeautifulSoup查找相关的搜索结果元素,例如标题或链接 results = soup.find_all('div', class_='g') # 这里假设搜索结果在class为"g"的div标签内 for result in results: title = result.find('h3', class_='r').text # 找到每个结果的标题 link = result.find('a')['href'] # 找到每个结果的链接 print(f'Title: {title}\nLink: {link}\n') # 打印抓取的信息 else: print("Failed to fetch the page. Status code:", response.status_code) ``` 请注意,实际的HTML结构可能会变化,所以`find_all()`和`find()`函后面的参可能会因网站而异。在抓取过程中,你也可能需要处理反爬虫机制,如验证码、用户代理设置等。
评论 2
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值