机器学习多算法实现对泰坦尼克号幸存者预测(源码及运行结果)

#!pip install pandas
#pip install --upgrade pandas
import pandas as pd
from sklearn.tree import DecisionTreeClassifier#D为S库中的一个分类器,基于决策树算法实现
from sklearn.datasets import make_classification
from sklearn.model_selection import train_test_split
from sklearn.model_selection import GridSearchCV
from sklearn.model_selection import cross_val_score
import matplotlib.pyplot as plt
data = pd.read_csv("taitanic_data.csv",index_col = 0)
# 头信息,部分数据信息
data.head()

# 信息汇总
data.info()

# 删除缺失值过多的列 inplace=True覆盖原表; axis=1 删除列
data.drop(["Cabin"],inplace=True,axis=1)
# 删除和观察判断来说和预测的y没有关系的列
data.drop(["Name","Ticket"],inplace=True,axis=1)
# 处理缺失值,填充均值
data["Age"] = data["Age"].fillna(data["Age"].mean())
# 删除含有缺失值的行
data = data.dropna()
data.info()

#将三分类变量转换为数值型变量
labels = data["Embarked"].unique().tolist()
data["Embarked"] = data["Embarked"].apply(lambda x: labels.index(x))
#性别转换,astype可以将文本类转换为数字,将二分类特征转换为0、1
data["Sex"] = (data["Sex"]== "male").astype("int")

data.head()

X = data.iloc[:,data.columns != "Survived"]
X[:10]

# iloc函数:通过行号来取行数据
y = data.iloc[:,data.columns == "Survived"]
y[:10]

 

from sklearn.model_selection import train_test_split
#划分训练集和测试集
Xtrain, Xtest, Ytrain, Ytest = train_test_split(X,y,test_size=0.3)

#修正测试集和训练集的索引
for i in [Xtrain, Xtest, Ytrain, Ytest]:
    i.index = range(i.shape[0])
    
#查看分好的训练集和测试集
Xtrain.head()

Xtest.head()

# 创建实例
clf = DecisionTreeClassifier(random_state=25)
# 训练模型
clf = clf.fit(Xtrain, Ytrain)
#模型评分
score_ = clf.score(Xtest, Ytest)

score_

# 10折交叉验证
score = cross_val_score(clf,X,y,cv=10).mean()

score

# 选择决策树的深度

tr = []
te = []
#遍历10次,每次修改树的深度
# 在每次分支时,不用全部特征,而是随机选取一部分特征,从中选取不纯度相关指标最优的作为分支用的结点 random_state
for i in range(10):
    
    clf = DecisionTreeClassifier(random_state=20
                                 ,max_depth=i+1
                                 ,criterion="entropy"
                                )
    clf = clf.fit(X, y)
    #
    score_tr = clf.score(Xtrain,Ytrain)
    #10折交叉验证
    score_te = cross_val_score(clf,X,y,cv=10).mean()
    tr.append(score_tr)
    te.append(score_te)
print(tr)
print(max(tr))
print(max(te))

plt.plot(range(1,11),tr,color="red",label="train")
plt.plot(range(1,11),te,color="blue",label="test")
# 坐标轴的单位
plt.xticks(range(1,11))
plt.legend()
plt.show()

# 网格搜索,耗时操作
import numpy as np
# 0到0.5之间  20个顺序的随机数
gini_thresholds = np.linspace(0,0.5,20)

parameters = {'splitter':('best','random')
              ,'criterion':("gini","entropy")
              ,"max_depth":[*range(1,10)]
              ,'min_samples_leaf':[*range(1,50,5)]
              ,'min_impurity_decrease':[*np.linspace(0,0.5,20)]
             }

# 分类器
clf = DecisionTreeClassifier(random_state=25)
# 网格搜索
GS = GridSearchCV(clf, parameters, cv=10)
GS.fit(Xtrain,Ytrain)
# 最优参数
GS.best_params_

#这段代码的作用是使用网格搜索(GridSearchCV)方法来
#找到决策树分类器(DecisionTreeClassifier)在指定参数范围内的最优参数组合

GS.best_score_

逻辑回归 

from sklearn.linear_model import LogisticRegression
from sklearn.model_selection import train_test_split

# 划分特征和标签
X = data[["Age", "Sex", "Fare", "Pclass"]].values
y = data["Survived"]


# 划分训练集和测试集
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)
print(X_train.shape, y_train.shape)
print(X_test.shape, y_test.shape)

# 创建并训练逻辑回归模型
model = LogisticRegression()
model.fit(X_train, y_train)

# 在测试集上进行预测
y_pred = model.predict(X_test)

# 输出模型在测试集上的准确率
accuracy = model.score(X_test, y_test)
print("Accuracy:", accuracy)

支持向量机 


from sklearn.model_selection import train_test_split
from sklearn.svm import SVC
from sklearn.metrics import accuracy_score




# 划分特征和标签
X = data[["Pclass", "Sex", "Age", "Fare"]]
y = data["Survived"]

# 划分训练集和测试集
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)

# 创建 SVM 分类器对象
svm = SVC(kernel='linear', C=1, random_state=42)

# 拟合数据集
svm.fit(X_train, y_train)

# 在测试集上进行预测
y_pred = svm.predict(X_test)

# 计算准确率
accuracy = accuracy_score(y_test, y_pred)
print("准确率:", accuracy)

随机森林 

from sklearn.ensemble import RandomForestClassifier
from sklearn.model_selection import train_test_split
from sklearn.metrics import accuracy_score
import matplotlib.pyplot as plt




# 划分特征和标签
X = data[['Pclass', 'Sex', 'Age', 'Fare']]
y = data['Survived']

# 划分训练集和测试集
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)

# 创建随机森林分类器对象
rf = RandomForestClassifier(n_estimators=100, random_state=42)

# 拟合数据集
rf.fit(X_train, y_train)

# 在测试集上进行预测
y_pred = rf.predict(X_test)

# 计算准确率
accuracy = accuracy_score(y_test, y_pred)
print("准确率:", accuracy)

# 可视化分类结果
x1 = X_test[y_pred == 0]['Age']
y1 = X_test[y_pred == 0]['Fare']
x2 = X_test[y_pred == 1]['Age']
y2 = X_test[y_pred == 1]['Fare']

plt.scatter(x1, y1, c='red', label='Not Survived')
plt.scatter(x2, y2, c='blue', label='Survived')
plt.xlabel('Age')
plt.ylabel('Fare')
plt.legend()
plt.title('Random Forest Classification Result')
plt.show()

KNN 

from sklearn.model_selection import train_test_split
from sklearn.neighbors import KNeighborsClassifier
from sklearn.metrics import accuracy_score


# 划分特征和标签
X = data[["Pclass", "Sex", "Age", "Fare"]]
y = data["Survived"]

# 划分训练集和测试集
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)

# 创建KNN分类器对象
knn = KNeighborsClassifier(n_neighbors=5)

# 拟合数据集
knn.fit(X_train, y_train)

# 在测试集上进行预测
y_pred = knn.predict(X_test)

# 计算准确率
accuracy = accuracy_score(y_test, y_pred)
print("准确率:", accuracy)

  • 3
    点赞
  • 11
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论
好的,下面是使用决策树完成泰坦尼克号幸存者预测的步骤: 1. 导入所需库 ```python import pandas as pd from sklearn.tree import DecisionTreeClassifier from sklearn.model_selection import train_test_split from sklearn.metrics import accuracy_score, confusion_matrix ``` 2. 载入数据集 ```python df = pd.read_csv("titanic.csv") ``` 3. 数据预处理 我们需要将一些非数值型的数据转换成数值型的,比如性别和登船口岸。 ```python df["Sex"] = df["Sex"].map({"male": 0, "female": 1}) df["Embarked"] = df["Embarked"].map({"S": 0, "C": 1, "Q": 2}) df = df.drop(["PassengerId", "Name", "Ticket", "Cabin"], axis=1) # 删除不必要的列 df = df.dropna() # 删除含有缺失值的行 ``` 4. 划分数据集 将数据集分为训练集和测试集,以便我们可以评估模型的性能。 ```python X = df.drop("Survived", axis=1) y = df["Survived"] X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, random_state=42) ``` 5. 训练模型 创建一个决策树分类器,并使用训练集进行训练。 ```python clf = DecisionTreeClassifier(random_state=42) clf.fit(X_train, y_train) ``` 6. 预测并评估模型 使用测试集进行预测,并计算模型的准确率和混淆矩阵。 ```python y_pred = clf.predict(X_test) print("Accuracy:", accuracy_score(y_test, y_pred)) print("Confusion matrix:\n", confusion_matrix(y_test, y_pred)) ``` 以上就是使用决策树完成泰坦尼克号幸存者预测的全部步骤。

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

beluga27

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值