随机森林 极限随机森林
Random forest is one of the most popular and powerful machine learning algorithms. It is one of the algorithms that can used for both classification and regression tasks and therefore, it is one of the most used algorithms in the machine learning space.
随机森林是最流行和功能最强大的机器学习算法之一。 它是可用于分类和回归任务的算法之一,因此,它是机器学习领域最常用的算法之一。
Random Forest is a supervised learning algorithm. So what exactly is ‘Random Forest’? As the name suggests, this algorithm creates a ‘forest’ with a number of trees. The underlying logic of the algorithm is to have a higher number of trees in the forest to produce high accuracy in results. Simply put, Random Forest builds multiple(ensemble) decision trees and merges them together for an accurate and stable prediction.
随机森林是一种监督学习算法。 那么“随机森林”到底是什么? 顾名思义,该算法创建了具有许多树木的“森林”。 该算法的基本逻辑是在森林中拥有更多的树木,以产生较高的结果准确性。 简而言之,Random Forest会构建多个(集合)决策树并将其合并在一起,以进行准确而稳定的预测。
The decision trees built are not just simple decision trees but rather bagged decision trees that are split on a subset of features on each split. Now, it is getting interesting, isn’t it?
构建的决策树不仅是简单的决策树,而且还是袋装的决策树 ,它们在每个分割的特征子集中进行分割。 现在,它变得越来越有趣了,不是吗?
Let us understand what are bagged trees. But before that we need to understand what is Bootstrapping.
让我们了解什么是袋装树。 但是在此之前,我