import jieba
from collections import defaultdict
import os
from sklearn.feature_extraction.text import TfidfTransformer
from sklearn.feature_extraction.text import CountVectorizer
def readfile(filename):
fh=open(filename,'r',encoding='utf-8')
data=[]
for x in fh.readlines():
if(x.strip()!=''):
data.append(x.strip())
fh.close()
return data
#x=readfile("C:/Users/yyq/Desktop/毕业论文/文档1.txt")
#print(x)
#分词处理
def cut2wd(sentence):
wdlist=jieba.cut(sentence)
wdrst=[]
for w in wdlist:
wdrst.append(w)
stopwds=readfile("C:/Users/yyq/Desktop/毕业论文
python机器学习----利用sklearn进行情感分析
最新推荐文章于 2024-07-08 19:15:00 发布