sklearn


1.GaussianNB

# -*- coding: utf-8 -*-
"""
Created on Wed Jun 20 19:23:06 2018

@author: 12046
"""

from sklearn import datasets,cross_validation  
from sklearn.naive_bayes import GaussianNB  
from sklearn.svm import SVC
from sklearn.ensemble import RandomForestClassifier  
from sklearn import metrics  

x,y=datasets.make_classification(n_samples=1000,n_features=10,n_classes=2)  
kf=cross_validation.KFold(len(x),n_folds=10,shuffle=True)  

  
for train_index,test_index in kf:    
    x_train,y_train=x[train_index],y[train_index]  
    x_test,y_test=x[test_index],y[test_index]  
  
clf=GaussianNB()
clf.fit(x_train,y_train)
pred=clf.predict(x_test)  

acc = metrics.accuracy_score(y_test, pred)  
print("Accuracy_score: "+str(acc))  
f1 = metrics.f1_score(y_test, pred)  
print("f1_score: "+str(f1))  
auc = metrics.roc_auc_score(y_test, pred)  
print("roc_auc_score: "+str(auc))  

结果如下图

2.SVC

# -*- coding: utf-8 -*-
"""
Created on Wed Jun 20 19:23:06 2018

@author: 12046
"""

from sklearn import datasets,cross_validation  
from sklearn.naive_bayes import GaussianNB  
from sklearn.svm import SVC
from sklearn.ensemble import RandomForestClassifier  
from sklearn import metrics  

x,y=datasets.make_classification(n_samples=1000,n_features=10,n_classes=2)  
kf=cross_validation.KFold(len(x),n_folds=10,shuffle=True)  

  
for train_index,test_index in kf:    
    x_train,y_train=x[train_index],y[train_index]  
    x_test,y_test=x[test_index],y[test_index]  
  
Cvalues=[1e-02, 1e-01, 1e00, 1e01, 1e02]
for C in Cvalues:
    clf = SVC(C=1e-01, kernel='rbf', gamma=0.1)
    clf.fit(x_train, y_train)
    pred = clf.predict(x_test)

    print("C:"+str(C))
    acc = metrics.accuracy_score(y_test, pred)  
    print("Accuracy_score: "+str(acc))  
    f1 = metrics.f1_score(y_test, pred)  
    print("f1_score: "+str(f1))  
    auc = metrics.roc_auc_score(y_test, pred)  
    print("roc_auc_score: "+str(auc))  
    print()
结果如下

3.RandomForestClassifier

# -*- coding: utf-8 -*-
"""
Created on Wed Jun 20 19:23:06 2018

@author: 12046
"""

from sklearn import datasets,cross_validation  
from sklearn.naive_bayes import GaussianNB  
from sklearn.svm import SVC
from sklearn.ensemble import RandomForestClassifier  
from sklearn import metrics  

x,y=datasets.make_classification(n_samples=1000,n_features=10,n_classes=2)  
kf=cross_validation.KFold(len(x),n_folds=10,shuffle=True)  

  
for train_index,test_index in kf:    
    x_train,y_train=x[train_index],y[train_index]  
    x_test,y_test=x[test_index],y[test_index]  
  
n_estimators_values=[10,100,1000]
for n_estimators in n_estimators_values:
    clf = RandomForestClassifier(n_estimators)
    clf.fit(x_train, y_train)
    pred = clf.predict(x_test)

    print("n_estimators: "+str(n_estimators))
    acc = metrics.accuracy_score(y_test, pred)  
    print("Accuracy_score: "+str(acc))  
    f1 = metrics.f1_score(y_test, pred)  
    print("f1_score: "+str(f1))  
    auc = metrics.roc_auc_score(y_test, pred)  
    print("roc_auc_score: "+str(auc))  
    print()


分析:三种模型的预测效果都差不多


  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值