机器学习cnn如何改变权值_五个机器学习悖论将改变您对数据的思考方式

机器学习cnn如何改变权值

Paradoxes are one of the marvels of human cognition that are hard to using math and statistics. Conceptually, a paradox is a statement that leads to an apparent self-contradictory conclusion based on the original premises of the problem. Even the best-known and well-documented paradoxes regularly fool domain experts as they fundamentally contradict common sense. As artificial intelligence(AI) looks to recreate human cognition, it’s very common for machine learning models to encounter paradoxical patterns in the training data and arrive to conclusions that seem contradictory at first glance. Today, I would like to explore some of the famous paradoxes that are commonly found in machine learning models.

悖论是人类难以理解的数学和统计数据之一。 从概念上讲,悖论是一种根据问题的原始前提得出明显的自相矛盾结论的陈述。 即使是最著名和有据可查的悖论,也经常愚弄领域专家,因为它们从根本上与常识相矛盾。 随着人工智能(AI)试图重现人类认知,机器学习模型在训练数据中遇到悖论模式并得出乍一看似乎矛盾的结论是非常普遍的。 今天,我想探讨一些机器学习模型中常见的著名悖论。

Paradoxes are typically formulated at the intersection of mathematics and philosophy. A notorious philosophical paradox is known as the Ship of Theseus questions whether an object that has had all of its components replaced remains fundamentally the same object. First, suppose that the famous ship sailed by the hero Theseus in a great battle has been kept in a harbor as a museum piece. As the years go by some of the wooden parts begin to rot and are replaced by new ones. After a century or so, all of the parts have been replaced. Is the “restored” ship still the same object as the original? Alternatively, suppose that each of the removed pieces were stored in a warehouse, and after the century, technology develops to cure their rotting and enable them to be put back together to make a ship. Is this “reconstructed” ship the original ship? And if so, is the restored ship in the harbor still the original ship too?

悖论通常是在数学和哲学的交叉点上提出的。 一个臭名昭著的哲学悖论被称为These修斯之船,它质疑一个已经将其所有组成部分都替换掉的物体是否根本上仍然是同一物体。 首先,假设英雄These修斯(Thusus)在一场激战中航行的那艘著名船已被保留在港口中作为博物馆作品。 随着时间的流逝,一些木制零件开始腐烂,并被新的木制零件取代。 一个世纪左右后,所有零件都被更换了。 “恢复”的船是否仍与原始船相同? 或者,假设每个拆下的零件都存储在仓库中,并且在本世纪之后,技术不断发展,可以治愈它们的腐烂,并使它们重新组合在一起制成一艘船。 这艘“改建”的船是原船吗? 如果是这样,港口中恢复的船舶还是原始船舶吗?

Image for post

The field of mathematics and statistics if full of famous paradoxes. To use a couple of famous examples, legendary mathematician and philosopher Bertrand Russell formulated a paradox that highlighted a contradiction in some of the most powerful ideas in set theory formulated one of the greatest mathematicians of all time: Greg Cantor. In essence, the Russell paradox questions whether a “list of all lists that do not contain themselves”. The paradox arises within native set theory by considering the set of all sets that are not members of themselves. Such a set appears to be a member of itself if and only if it is not a member of itself. Hence the paradox. Some sets, such as the set of all teacups, are not members of themselves. Other sets, such as the set of all non-teacups, are members of themselves. Call the set of all sets that are not members of themselves “R.” If R is a member of itself, then by definition it must not be a member of itself. Similarly, if R is not a member of itself, then by definition it must be a member of itself. What????

数学和统计领域,如果充满着著名的悖论。 举几个著名的例子,传说中的数学家和哲学家贝特朗·罗素提出了一个悖论,突显了集合论中某些最强大的思想中的矛盾,而这是有史以来最伟大的数学家之一: 格雷格·坎托 。 本质上,罗素悖论质疑“一个不包含自己的所有列表的列表”。 在本机集合论中,通过考虑不是自身成员的所有集合的集合而产生了悖论。 当且仅当它不是其自身的成员时,这样的集合才似乎是其自身的成员。 因此自相矛盾。 某些集合(例如所有茶杯的集合)不是其成员。 其他集合,例如所有非茶杯的集合,都是它们自己的成员。 将所有非其自身成员的集合称为“ R” 。 如果R是其自身的成员,那么根据定义,它一定不能是其自身的成员。 类似地,如果R不是其自身的成员,那么根据定义,它必须是其自身的成员。 什么????

Image for post

机器学习模型中的著名悖论 (Famous Paradoxes in Machine Learning Models)

As any form of knowledge building based on data, machine learning models are not exempt of cognitive paradoxes. Quite the opposite, as machine learning try to infer patterns hidden in training datasets and validate their knowledge against a specific environment, they are constantly vulnerable to paradoxical conclusions. Here are a few of the most notorious paradoxes that surface in machine learning solutions.

作为基于数据的任何形式的知识构建,机器学习模型都不能免除认知悖论。 恰恰相反,当机器学习试图推断隐藏在训练数据集中的模式并针对特定环境验证其知识时,它们总是容易受到悖论性结论的攻击。 这是机器学习解决方案中出现的一些最臭名昭著的悖论。

辛普森悖论 (The Simpson’s Paradox)

Named after British mathematician Edward Simpson, the Simpson’s Paradox describes a phenomenon in which a trend that is very apparent several groups of data dissipates as the data within those groups in combined. A real-life case of the paradox happened in 1973. Admission rates were investigated at the University of Berkeley’s graduate schools. The university was sued by women for the gender gap in admissions. The results of the investigation were: When each school was looked at separately (law, medicine, engineering etc.), women were admitted at a higher rate than men! However, the average suggested that men were admitted at a much higher rate than women. How is that possible?

以英国数学家爱德华·辛普森(Edward Simpson)的名字命名的辛普森悖论描述了一种现象,其中一种趋势非常明显,几组数据随着这些组内数据的组合而消失。 1973年发生了一个真实的悖论案例。伯克利大学的研究生院对入学率进行了调查。 妇女因录取中的性别差距而起诉该大学。 调查的结果是:当分别检查每所学校(法律,医学,工科等)时女生的入学率高于男生! 但是,平均数表明,男性的入学率比女性高得多。 那怎么可能?

Image for post

The explanation to the previous use case is that a simple average doesn’t account for the relevance of a specific group within the overall dataset. In this specific example, women applied in large numbers to schools with low admission rates: Like law and medicine. These schools admitted less than 10 percent of students. Therefore the percentage of women accepted was very low. Men, on the other hand, tended to apply in larger numbers to schools with high admission rates: Like engineering, where admission rates are about 50%. Therefore the percentage of men accepted was very high.

对前一个用例的解释是,简单的平均值并不能说明整个数据集中特定组的相关性。 在这个具体例子中,妇女大量申请入学率低的学校:法律和医学一样。 这些学校录取的学生不到10%。 因此,接受女性的比例非常低。 另一方面,男人往往更多地被录取率高的学校使用:像工程学一样,录取率约为50%。 因此,接受男人的比例很高。

In the context of machine learning, many unsupervised learning algorithms infer patterns different training datasets that result in contradictions when combined across the board.

在机器学习的背景下,许多无监督的学习算法会推断出不同的训练数据集的模式,这些数据集在整体组合时会产生矛盾。

布莱斯悖论 (The Braess’s Paradox)

This paradox was proposed in 1968 by German mathematician Dietrich Braes. Using an example of congested traffic networks, Braes explained that, counterintuitively, adding a road to a road network could possibly impede its flow (e.g. the travel time of each driver); equivalently, closing roads could potentially improve travel times. Braess reasoning is based on the fact that, in a Nash equilibrium game, drivers have no incentive to change their routes. In terms of game theory, an individual has nothing to gain from applying new strategies if others stick to the same ones. Here in the case of drivers, a strategy is a route taken. In the case of Braess’s paradox, drivers will continue to switch until they reach Nash equilibrium despite the reduction in overall performance. So, counter-intuitively, closing the roads might ease the congestion.

这种悖论是由德国数学家Dietrich Braes于1968年提出的。 布雷斯以拥挤的交通网络为例,解释说,与直觉相反,在道路网络中增加道路可能会阻碍其流量(例如,每个驾驶员的出行时间); 同样,封闭道路可能会改善出行时间。 Braess推理基于以下事实:在Nash均衡博弈中,驾驶员没有动力改变路线。 就博弈论而言,如果其他人坚持相同的策略,则个人从采用新策略中不会获得任何收益。 在这里,对于驾驶员而言,策略是采取的路线。 在Braess悖论的情况下,尽管整体表现有所下降,但驾驶员将继续转换直到达到Nash平衡。 因此,违反直觉,关闭道路可能会缓解交通拥堵。

Image for post

The Braess’s Paradox is very relevant in autonomous, multi-agent reinforcement learning scenarios in which the models needs to reward agents based on specific decisions in unknown environments.

Braess悖论在自主的多主体强化学习场景中非常相关,在该场景中,模型需要根据未知环境中的特定决策来奖励主体。

莫拉韦克悖论 (The Moravec’s Paradox)

Hans Moravec can be considered one of the greatest AI thinkers of the last few decades. In the 1980s, Moravec formulated a counter-intuitive proposition to the way AI models acquire knowledge. The Moravec Paradox states that, contrary to popular believe, high-level reasoning requires less computation than low-level unconscious cognition. This is an empirical observation that goes against the notion that greater computational capability leads to more intelligent systems.

汉斯·莫拉维克(Hans Moravec)被认为是过去几十年来最伟大的AI思想家之一。 在1980年代,Moravec对AI模型获取知识的方式提出了违反直觉的主张。 莫拉维克悖论指出,与普遍的看法相反,高级推理比低级无意识认知所需的计算更少。 这是一个经验性观察,与以下观点相反:更高的计算能力会导致更智能的系统。

A simpler way to frame the Moravec’s Paradox is that AI models can do incredibly complex statistical and data inference tasks that result impossible for humans. However, many tasks that result trivial for humans like grabbing an object will require expensive AI models. As Moravec writes, “it is comparatively easy to make computers exhibit adult level performance on intelligence tests or playing checkers, and difficult or impossible to give them the skills of a one-year-old when it comes to perception and mobility”.

构造Moravec悖论的一种更简单的方法是,AI模型可以执行极其复杂的统计和数据推断任务,而这对于人类来说是不可能的。 但是,许多对人类来说都是微不足道的任务,例如抓住一个物体,将需要昂贵的AI模型。 正如Moravec所说,“使计算机在智力测验或游戏检查中表现出成年人水平的性能相对容易,而就感知和移动性而言,很难或不可能赋予他们一岁的技能”。

Image for post

From the perspective of machine learning, the Moravec’s Paradox is very applicable in aspect of transfer learning that look to generalize knowledge across different machine learning models. Additionally, the Moravec’s Paradox teaches us that some of the best applications of machine intelligence will come as a combination of humans and algorithms.

从机器学习的角度来看,Moravec的悖论在转移学习方面非常适用,该学习旨在跨不同的机器学习模型概括知识。 此外,Moravec的悖论告诉我们,机器智能的一些最佳应用将是人类和算法的结合。

准确性悖论 (The Accuracy Paradox)

Directly related to machine learning, the Accuracy Paradox states that, counterintuitively, accuracy is not always a good metric to classify the effectiveness of predictive models. How is that for a confusing statement? The Accuracy Para has its roots in imbalanced training datasets. For instance, in a dataset in which the incidence of category A is dominant, being found in 99% of cases, then predicting that every case is category A will have an accuracy of 99% is completely misleading.

与机器学习直接相关的“准确性悖论”指出,与直觉相反,准确性并非总是对预测模型的有效性进行分类的好指标。 对于一个令人困惑的陈述,这是怎么回事? Accuracy Para起源于不平衡的训练数据集。 例如,在一个类别为A的事件占主导地位的数据集中,在99%的案例中均被发现,然后预测每个案例均为类别A的准确性将达到99%,这完全是误导性的。

A simpler way to understand the Accuracy Paradox is to find the balance between precision and recall in machine learning models. In machine learning algorithms, precision is often defined as measuring what fraction of your predictions for the positive class are valid. It is formulated by (True Positives / True Positives + False Positives). Complementary, the recall metric measures how often your predictions actually capture the positive class. It is formulated by (True Positives / True Positives + False Negatives).

理解“准确性悖论”的一种简单方法是在机器学习模型中找到精度和召回率之间的平衡。 在机器学习算法中,精度通常被定义为衡量您对肯定类别的预测中有多少有效。 它由(真肯定/真肯定+假肯定)表示。 作为补充,召回指标用于衡量您的预测多长时间获得一次肯定的评价。 它由(真肯定/真肯定+假否定)表示。

Image for post

In many machine learning models, the balance between precision and recall results a better metric for accuracy. For instance, in case of an algorithm for fraud detection recall is a more important metric. It is obviously important to catch every possible fraud even if it means that the authorities might need to go through some false positives. On the other hand, if the algorithm is created for sentiment analysis and all you need is a high-level idea of emotions indicated in tweets then aiming for precision is the way to go.

在许多机器学习模型中,精度和召回率之间的平衡导致了更好的精度指标。 例如,在欺诈检测算法的情况下,召回是更重要的指标。 即使这意味着当局可能需要进行一些误报,显然也必须抓住一切可能的欺诈行为。 另一方面,如果创建用于情感分析的算法,而您所需要的只是对推文中所指示的情绪的高级理解,那么追求精确性就是路要走。

学习能力-戈德尔悖论 (The Learnability-Godel Paradox)

Saving the most controversial for last, this is a very recent paradox that was published in a research paper earlier this year. The paradox links the ability of a machine learning model to learn to one of the most controversial theories of mathematics: Gödel’s Incompleteness Theorem.

最后,最有争议的是这是今年早些时候发表在研究论文上的一个最新悖论。 悖论将机器学习模型的学习能力与最有争议的数学理论之一: 哥德尔不完全性定理

Kurt Gödel is one of the brightest mathematicians of all time and one that pushed the boundaries of philosophy, physics and mathematics like a few of its predecessors. In 1931, Gödel published his two incompleteness theorems that essentially say some statements cannot be proved either true or false using standard mathematical language. In other words, math is an insufficient language to understand some aspects of the universe. The theorems have come to be known as Gödel ‘s continuum hypothesis.

库尔特·哥德尔(KurtGödel)是有史以来最杰出的数学家之一,并且像其几位前辈一样,推动了哲学,物理学和数学的发展。 1931年,哥德尔发表了他的两个不完全性定理 ,这些定理本质上说,使用标准数学语言无法证明某些陈述是对还是错。 换句话说,数学不足以理解宇宙的某些方面。 这些定理已被称为哥德尔连续统假设。

In a recent work, AI researchers from the Israel Institute of Technology linked Gödel’s continuum hypothesis to the learnability of a machine learning model. In a paradoxical statement that challenges all common wisdom, the researchers define the notion of a learnability limbo. Essentially, the researchers go on to show that if the continuum hypothesis is true, a small sample is sufficient to make the extrapolation. But if it is false, no finite sample can ever be enough. This way they show that the problem of learnability is equivalent to the continuum hypothesis. Therefore, the learnability problem, too, is in a state of limbo that can be resolved only by choosing the axiomatic universe.

在最近的工作中,以色列理工学院的AI研究人员将Gödel的连续体假设与机器学习模型的可学习性联系在一起。 在挑战所有常识的矛盾陈述中,研究人员定义了可学习性的概念。 从本质上讲,研究人员继续表明,如果连续假设是正确的,则仅需少量样本就可以进行推断。 但是,如果它是错误的,则没有任何有限的样本就足够了。 通过这种方式,他们表明可学习性问题等同于连续性假设。 因此,可学习性问题也处于困境,只能通过选择公理宇宙来解决。

In simple terms, the mathematical proofs in the study show that AI problems are subjected to Gödel ‘s continuum hypothesis which means that many problems might be effectively unsolvable by AI. Although this paradox has very little applications to real world AI problems today, it will be paramount to the evolution of the field in the near future.

简而言之,研究中的数学证明表明,人工智能问题服从于哥德尔的连续假设,这意味着许多问题可能是人工智能无法有效解决的。 尽管这种悖论在当今现实世界中的AI问题中几乎没有应用,但在不久的将来对领域的发展至关重要。

Paradoxes are omnipresent in machine learning problems in the real world. You can argue that as, algorithms don’t have a notion of common sense, they might be immune to statistical paradoxes. However, given that most machine learning problems require human analysis and interventions and are based on human-curated datasets, we are going to live in an universe of paradoxes for quite some time.

在现实世界中,机器学习问题中普遍存在悖论。 您可以辩称,由于算法没有常识,它们可能不受统计悖论的影响。 但是,鉴于大多数机器学习问题都需要人工分析和干预,并且都是基于人工编排的数据集,因此我们将在相当长的一段悖论中生存。

翻译自: https://medium.com/dataseries/five-machine-learning-paradoxes-that-will-change-the-way-you-think-about-data-3b82513482b8

机器学习cnn如何改变权值

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值