源码-西虹市首富python分析 (基于数据麒麟的源码)

源码位置 https://github.com/shujusenlin

有一些调整,安装了wordcloud

遇到的坑

1:地市还是不太对,因此我把 《西虹市首富.xlsx》里面的 县、市 都去掉了,对应的pycharts/datasets/city_coordinates.json

里面也去掉了

2: 错误

Traceback (most recent call last):
  File "E:/lonson/Program Files/JetBrains/PyCharm 2017.2.3/workspace-py/test/src/xihongshi/newxihongshi.py", line 72, in <module>
    geo.add("", attr, value, type="heatmap", visual_range=[0, 200],visual_text_color="#fff", symbol_size=40, is_visualmap=True,is_roam=False)
  File "E:\lonson\Program Files\Python\Python35\lib\site-packages\pyecharts\charts\geo.py", line 50, in add
    self.__add(*args, **kwargs)
  File "E:\lonson\Program Files\Python\Python35\lib\site-packages\pyecharts\charts\geo.py", line 109, in __add
    _coordinate = self.get_coordinate(_name, raise_exception=True)
  File "E:\lonson\Program Files\Python\Python35\lib\site-packages\pyecharts\charts\geo.py", line 43, in get_coordinate
    coordinate = get_coordinate(name)
  File "E:\lonson\Program Files\Python\Python35\lib\site-packages\pyecharts\datasets\coordinates.py", line 45, in get_coordinate
    return _COORDINATE_DATASET.get(name, None)
  File "E:\lonson\Program Files\Python\Python35\lib\site-packages\pyecharts\utils\lazy.py", line 18, in inner
    self._setup()
  File "E:\lonson\Program Files\Python\Python35\lib\site-packages\pyecharts\utils\lazy.py", line 32, in _setup
    self._wrapped = self._setupfunc()
  File "E:\lonson\Program Files\Python\Python35\lib\site-packages\pyecharts\datasets\coordinates.py", line 56, in _load_coordinates
    return json.load(f)
  File "E:\lonson\Program Files\Python\Python35\lib\json\__init__.py", line 268, in load
    parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw)
  File "E:\lonson\Program Files\Python\Python35\lib\json\__init__.py", line 319, in loads
    return _default_decoder.decode(s)
  File "E:\lonson\Program Files\Python\Python35\lib\json\decoder.py", line 339, in decode
    obj, end = self.raw_decode(s, idx=_w(s, 0).end())
  File "E:\lonson\Program Files\Python\Python35\lib\json\decoder.py", line 355, in raw_decode
    obj, end = self.scan_once(s, idx)
json.decoder.JSONDecodeError: Invalid control character at: line 2166 column 4 (char 20145)

在geo.py里面增加打印之后发现 _name= 万宁,调整city_coordinates.json即可。手误!!!

 

调整之后的代码如下

# -*- coding: utf-8 -*-
"""
Created on Sun Jul 29 09:35:03 2018
@author: dell
"""
## 调用要使用的包
import json
import random
import requests
import time
import pandas as pd
import os
from pyecharts import Bar, Geo, Line, Overlap
import jieba
from scipy.misc import imread  # 这是一个处理图像的函数
from wordcloud import WordCloud, ImageColorGenerator
import matplotlib.pyplot as plt
from collections import Counter

os.chdir('E:/lonson/custom/爬虫/西红柿')

## 设置headers和cookie
header = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win32; x32; rv:54.0) Gecko/20100101 Firefox/54.0',
          'Connection': 'keep-alive'}
cookies = 'v=3; iuuid=1A6E888B4A4B29B16FBA1299108DBE9CDCB327A9713C232B36E4DB4FF222CF03; webp=true; ci=1%2C%E5%8C%97%E4%BA%AC; __guid=26581345.3954606544145667000.1530879049181.8303; _lxsdk_cuid=1646f808301c8-0a4e19f5421593-5d4e211f-100200-1646f808302c8; _lxsdk=1A6E888B4A4B29B16FBA1299108DBE9CDCB327A9713C232B36E4DB4FF222CF03; monitor_count=1; _lxsdk_s=16472ee89ec-de2-f91-ed0%7C%7C5; __mta=189118996.1530879050545.1530936763555.1530937843742.18'
cookie = {}
for line in cookies.split(';'):
    name, value = cookies.strip().split('=', 1)
    cookie[name] = value

## 爬取数据,每次理论上可以爬取1.5W调数据,存在大量重复数据,需要多次执行,最后统一去重

    tomato = pd.DataFrame(columns=['date', 'score', 'city', 'comment', 'nick'])
for i in range(0, 1000):
    j = random.randint(1, 1000)
    print(str(i) + ' ' + str(j))
    try:
        time.sleep(2)
        url = 'http://m.maoyan.com/mmdb/comments/movie/1212592.json?_v_=yes&offset=' + str(j)
        html = requests.get(url=url, cookies=cookie, headers=header).content
        data = json.loads(html.decode('utf-8'))['cmts']
        for item in data:
            tomato = tomato.append({'date': item['time'].split(' ')[0], 'city': item['cityName'],
                                    'score': item['score'], 'comment': item['content'],
                                    'nick': item['nick']}, ignore_index=True)

        tomato.to_excel('西虹市首富.xlsx', index=False)
    except:
        continue

## 可以直接读取我们已经爬到的数据进行分析
tomato_com = pd.read_excel('西虹市首富.xlsx')
grouped = tomato_com.groupby(['city'])
grouped_pct = grouped['score']

## 全国热力图
city_com = grouped_pct.agg(['mean', 'count'])
city_com.reset_index(inplace=True)
city_com['mean'] = round(city_com['mean'], 2)
data = [(city_com['city'][i], city_com['count'][i]) for i in range(0, city_com.shape[0])]

for d  in data:
    print("data=",d)

geo = Geo('《西虹市首富》全国热力图', title_color="#fff",
          title_pos="center", width=1200, height=600, background_color='#404a59')
attr, value = geo.cast(data)
geo.add("", attr, value, type="heatmap", visual_range=[0, 200],visual_text_color="#fff", symbol_size=40, is_visualmap=True,is_roam=False)
geo.render('西虹市首富全国热力图.html')

## 主要城市评论数与评分
city_main = city_com.sort_values('count', ascending=False)[0:20]
attr = city_main['city']
v1 = city_main['count']
v2 = city_main['mean']

line = Line("主要城市评分")
line.add("城市", attr, v2, is_stack=True, xaxis_rotate=30, yaxis_min=4.2,
         mark_point=['min', 'max'], xaxis_interval=0, line_color='lightblue',
         line_width=4, mark_point_textcolor='black', mark_point_color='lightblue',
         is_splitline_show=False)

bar = Bar("主要城市评论数")
bar.add("城市", attr, v1, is_stack=True, xaxis_rotate=30, yaxis_min=4.2,
        xaxis_interval=0, is_splitline_show=False)
overlap = Overlap()
# 默认不新增 x y 轴,并且 x y 轴的索引都为 0
overlap.add(bar)
overlap.add(line, yaxis_index=1, is_add_yaxis=True)
overlap.render('主要城市评论数_平均分.html')

## 主要城市评分降序
city_score = city_main.sort_values('mean', ascending=False)[0:20]
attr = city_score['city']
v1 = city_score['mean']
line = Line("主要城市评分")
line.add("城市", attr, v1, is_stack=True, xaxis_rotate=30, yaxis_min=4.2,
         mark_point=['min', 'max'], xaxis_interval=0, line_color='lightblue',
         line_width=4, mark_point_textcolor='black', mark_point_color='lightblue',
         is_splitline_show=False)
line.render('主要城市评分.html')

## 主要城市评分全国分布
city_score_area = city_com.sort_values('count', ascending=False)[0:30]
city_score_area.reset_index(inplace=True)
data = [(city_score_area['city'][i], city_score_area['mean'][i]) for i in range(0,
                                                                                city_score_area.shape[0])]
geo = Geo('《西虹市首富》全国主要城市打分图', title_color="#fff",
          title_pos="center", width=1200, height=600, background_color='#404a59')
attr, value = geo.cast(data)
geo.add("", attr, value, visual_range=[4.4, 4.8],
        visual_text_color="#fff", symbol_size=15, is_visualmap=True,
        is_roam=False)
geo.render('西虹市首富全国主要城市打分图.html')

## 前三天票房对比
piaofang = pd.read_excel('票房.xlsx')
attr1 = piaofang[piaofang['film'] == '西虹市首富']['day']
v1 = piaofang[piaofang['film'] == '西虹市首富']['money']
attr2 = piaofang[piaofang['film'] == '羞羞的铁拳']['day']
v2 = piaofang[piaofang['film'] == '羞羞的铁拳']['money']
line = Line("前三天票房对比")
line.add("西红柿首富", attr1, v1, is_stack=True)
line.add("羞羞的铁拳", attr2, v2, is_stack=True)
line.render('前三天票房对比.html')

## 绘制词云
tomato_str = ' '.join(tomato_com['comment'])
words_list = []
word_generator = jieba.cut_for_search(tomato_str)
for word in word_generator:
    words_list.append(word)
words_list = [k for k in words_list if len(k) > 1]
back_color = imread('西红柿.jpg')  # 解析该图片
wc = WordCloud(background_color='white',  # 背景颜色
               max_words=200,  # 最大词数
               mask=back_color,  # 以该参数值作图绘制词云,这个参数不为空时,width和height会被忽略
               max_font_size=300,  # 显示字体的最大值
               font_path="C:/Windows/Fonts/STFANGSO.ttf",  # 解决显示口字型乱码问题,可进入C:/Windows/Fonts/目录更换字体
               random_state=42,  # 为每个词返回一个PIL颜色
               )
tomato_count = Counter(words_list)
wc.generate_from_frequencies(tomato_count)
# 基于彩色图像生成相应彩色
image_colors = ImageColorGenerator(back_color)
# 绘制结果
plt.figure()
plt.imshow(wc.recolor(color_func=image_colors))
plt.axis('off')

 

  • 0
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
2021年最新火麒麟打赏完整源码/打赏系统/超越冠军/诺诺/多级防封/多支付接口切换/带包天月/多套模板 好友分享的,互站买来的版本,这个买的就不贵,索性就直接分享给大家了,对这一类资源了解不多,直接引用互站介绍吧。 一套小白都能直接上手的打赏系统 市面上大部分人用的都是金牌和冠军两个版本的打赏系统,虽然两套系统已经经过很多人的改良,改良过的系统可以稳定运行,但最近还有很多客户、遇到金额被改、扣量不正常、文件或者密码被恶意篡改等问题。 现本店推出一套全新的打赏系统,一改以往云赏和诺诺的底层概念,让运营者更安全、更稳定、更方便! 声明:此套系统并非云赏、诺诺二次开发产物!!! 新系统特性: 1、功能、限制设定全部在后台搞定,完全不用改代码,就是不懂技术的人都能用; 2、超级防封,内置防封接口,以及域名自动切换; 3、系统自带域名检测功能,检测域名是否被封,自动跳过已红域名; 4、系统自带包天、包月、,不用改代码,直接后台设置; 5、可设定已打赏用户视频观看有效期; 6、系统自带市面上常用的支付接口,一键配置,也可以增加新接口; 7、自带19套盒子模板,19套推广模板,一改以往死板的页面; 8、自带会员注册功能,用户可直接打赏,也可以选择包天、包月; 9、台主可设置每个视频打赏的金额限制范围,保证台子利润; 10、代理可设置每个视频打赏金额范围,一改价格固定不变的局面; 11、一键生成总推广链接,每次生成链接都不一样,防止短链被墙; 12、采用伪文件+参数加密等方式保护系统文件安全不被篡改。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值