决策树(Decision Tree)是一种简单但是广泛使用的分类器。通过训练数据构建决策树,可以高效的对未知的数据进行分类。决策数有两大优点:1)决策树模型可以读性好,具有描述性,有助于人工分析;2)效率高,决策树只需要一次构建,反复使用,每一次预测的最大计算次数不超过决策树的深度。但是决策树对训练集里不存在数据进行预测时,效果很差。
from sklearn.model_selection._split import train_test_split
dataset = [
["Outlook","Temperature","Humidity","Wind","Play?"],
["sunny","hot","high","false","no"],
["sunny","hot","high","true","no"],
["overcast","hot","high","false","yes"],
["rain","mild","high","false","yes"],
["rain","cool","normal","false","yes"],
["rain","cool","normal",