ai训练系统流程_训练AI系统的西方世界之路

ai训练系统流程

I recently started a new newsletter focus on AI education. TheSequence is a no-BS( meaning no hype, no news etc) AI-focused newsletter that takes 5 minutes to read. The goal is to keep you up to date with machine learning projects, research papers and concepts. Please give it a try by subscribing below:

我最近开始了一份有关AI教育的新时事通讯。 TheSequence是无BS(意味着没有炒作,没有新闻等),它是专注于AI的新闻通讯,需要5分钟的阅读时间。 目标是让您了解机器学习项目,研究论文和概念的最新动态。 请通过以下订阅尝试一下:

West World is one of my favorite TV series of the last few years. The HBO drama combines a stellar group of actors in an engaging plot that touches upon some of the most controversial aspects of the future of artificial intelligence(AI). In almost every episode of the first season of West World, we find humans trying to understand the decisions made by the hosts(robots) on specific circumstances. Every time some human needs an explanation about the host behavior can simply query a system that will proceed the explain the reasoning behind the host decision. Simply saying “Analysis, explain X or Y” and the host will pleasantly proceed to detail the intricacies behind its behaviors or actions. If only things work like that in real artificial intelligence(AI) systems.

西方世界是最近几年我最喜欢的电视连续剧之一。 HBO的电视剧在一个引人入胜的情节中结合了一群杰出的演员,这些情节触及了人工智能(AI)未来最有争议的一些方面。 在《西方世界》第一季的几乎每一集中,我们都发现人们试图了解主机(机器人)在特定情况下所做的决定。 每当有人需要有关宿主行为的解释时,都可以简单地查询系统,该系统将继续进行解释宿主决策背后的原因。 只需说“分析,解释X或Y”,主持人就会很高兴地详细说明其行为或动作背后的复杂性。 如果只有在真正的人工智能(AI)系统中,事情才能像那样工作。

Explaining and interpreting knowledge is one of the hardest problems in modern deep learning systems. In supervised deep learning systems, the processes for training a model and the knowledge built in that model are almost uninterpretable. However interpretation of knowledge is a key element in the way humans learn. Let’s take a classic student-teacher setting in which the teacher is trying to convey a specific concept to the student using a series of examples. Based on the feedback from the student, the teacher will adapt his explanations and try to select the most appropriate examples to improve the knowledge of the student. That pedagogical process works brilliantly for humans but fails miserably for neural networks.

解释和解释知识是现代深度学习系统中最难的问题之一。 在监督式深度学习系统中,用于训练模型的过程和在该模型中建立的知识几乎是无法解释的。 然而,知识的解释是人类学习方式的关键要素。 让我们以经典的学生-教师环境为例,在该环境中,老师尝试使用一系列示例向学生传达特定概念。 根据学生的反馈,老师将调整他的解释,并尝试选择最合适的例子来提高学生的知识。 该教学过程对人类非常有效,但对神经网络却失败了。

Some of the most interesting scenarios in deep learning systems require a seamless collaboration between humans and neural networks. However, in most scenarios, its incredibly difficult to establish that collaboration as both sides speak different protocols. A couple of years ago, OpenAI published one of the most relevant papers in this area, Under the title, Interpretable and Pedagogical Examples, OpenAI proposes a method that tries to address this challenge by proposing a more pedagogical way to tech deep learning systems.

深度学习系统中一些最有趣的场景需要人类与神经网络之间的无缝协作。 但是,在大多数情况下,由于双方使用不同的协议,很难建立这种合作关系。 几年前,OpenAI发表了该领域最相关的论文之一,标题为“ Interpretable and Pedagogical Examples” ,OpenAI提出了一种方法,试图通过为技术深度学习系统提出一种更具教学性的方式来应对这一挑战。

Under the title “Interpretable and Pedagogical Examples” the OpenAI researchers formulate an intriguing thesis about what makes understanding the knowledge of deep learning systems so difficult. In their opinion, part of the challenge is that most deep learning architectures rely on teacher and student neural networks to be train jointly which prevents any feedback loop between the two. Instead of that model, the OpenAI team proposes a structure in which teacher and student networks can be train iteratively which can produce more interpretable teaching strategies.

OpenAI研究人员以可解释和教学实例”为题,提出了一个有趣的论点,即为何很难理解深度学习系统的知识。 他们认为,挑战的一部分在于,大多数深度学习架构都依赖于教师和学生的神经网络进行联合训练,从而阻止了两者之间的任何反馈循环。 代替该模型,OpenAI团队提出了一种结构,在该结构中,可以迭代地训练教师和学生的网络,从而可以产生更具解释性的教学策略。

可解释的机器学习 (Interpretable Machine Learning)

The OpenAI interpretable teaching strategy can be seen as a game dynamic between two neural networks, a student and a teacher. The goal of the game is for the student to guess a particular concept based on examples of that concept and the goal of the teacher is to learn to select the most illustrative examples for the student. Using an image recognition scenario as an analogy, the student should try to guess the concepts in a specific image while the teacher should try to select the most appropriate images to improve the knowledge of the student.

OpenAI可解释的教学策略可以看作是两个神经网络(学生和老师)之间的博弈动态。 游戏的目的是让学生根据该概念的示例来猜测一个特定的概念,而老师的目标是学习为学生选择最具说明性的示例。 使用图像识别场景作为类比,学生应尝试猜测特定图像中的概念,而老师应尝试选择最合适的图像以提高学生的知识水平。

Image for post
Source: https://arxiv.org/abs/1711.00694
资料来源: https : //arxiv.org/abs/1711.00694

The two-stage technique to interpretable teaching works like this: a ‘student’ neural network is given randomly selected input examples of concepts and is trained from those examples using traditional supervised learning methods to guess the correct concept labels. In the second step, the ‘teacher’ network — which has an intended concept to teach and access to labels linking concepts to examples — tests the different examples on the student and see which concept labels the student assigns them, eventually converging on the smallest set of examples it needs to give to let the student guess the intended concept.

可解释性教学的两阶段技术是这样的:“学生”神经网络被赋予随机选择的概念输入示例,并使用传统的监督学习方法从这些示例中进行训练,以猜测正确的概念标签。 在第二步中,“教师”网络-具有预期的概念,可以教书并访问将概念链接到示例的标签-测试学生的不同示例,并查看学生为他们分配了哪些概念标签,最终收敛到最小的一组它需要给出一些示例,以使学生猜出预期的概念。

The key to the OpenAI methods is that the teacher and student networks are being trained iteratively rather than jointly. In the traditional mode, both neural networks will be trained together selecting examples that are hard to interpret by humans. The goal of the OpenAI technique is to produce more interpretable teaching strategies but how do we really quantify interpretable? To evaluate the performance of the mode, the OpenAI team centered in two fundamental metrics:

OpenAI方法的关键是要对教师和学生网络进行迭代式培训,而不是联合培训。 在传统模式下,两个神经网络将一起训练,选择难以被人类解释的示例。 OpenAI技术的目标是产生更多可解释的教学策略,但是我们如何真正量化可解释的内容? 为了评估该模式的性能,OpenAI团队以两个基本指标为中心:

1. Evaluating how similar the selected strategies are to intuitive human-designed strategies in each task.

1.评估每个任务中所选策略与直观的人工设计策略有多相似。

2. Evaluating the effectiveness of the selected strategies at teaching humans.

2.评估所选策略对人类教学的有效性。

The OpenAI researchers applied interpretable strategies across a large variety of scenarios producing remarkable results that vastly improve over traditional techniques. More specifically, interpretable teaching leads the student model to learn an interpretable learning strategy, which then constrains the teacher to learn an interpretable teaching strategy.

OpenAI研究人员在各种各样的场景中应用了可解释的策略,产生了显着的成果,与传统技术相比有了很大的改进。 更具体地说,可解释的教学使学生模型学习可解释的学习策略,然后限制教师学习可解释的教学策略。

翻译自: https://medium.com/swlh/training-ai-systems-the-west-world-way-369438e4fc7b

ai训练系统流程

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值