Deepmind 公司及AI岗位概览

一、公司背景

When we started DeepMind in 2010, there was far less interest in the field of AI than there is today. To accelerate the field, we took an interdisciplinary approach, bringing together new ideas and advances in machine learning, neuroscience, engineering, mathematics, simulation and computing infrastructure, along with new ways of organising scientific endeavour. 2010年成立,关注机器学习,神经科学,工程,数学,模拟仿真等领域。We joined forces with Google in 2014 to accelerate our work, while continuing to set our own research agenda. 2014年并入谷歌。

二、具有影响力的项目

2.1 AlphaGo defeats Lee Sedol in the game of Go

We created AlphaGo, a computer program that combines advanced search tree with deep neural networks. These neural networks take a description of the Go board as an input and process it through a number of different network layers containing millions of neuron-like connections. 将整张棋盘输入深度学习网络中。

One neural network, the “policy network”, selects the next move to play. The other neural network, the “value network”, predicts the winner of the game. We introduced AlphaGo to numerous amateur games to help it develop an understanding of reasonable human play. Then we had it play against different versions of itself thousands of times, each time learning from its mistakes. 关键网络,策略网络和价值网络。

相关论文

1、Mastering the game of Go with deep neural networks and tree search

链接:

https://storage.googleapis.com/deepmind-media/alphago/AlphaGoNaturePaper.pdf

Abstract

A new approach to computer Go that combines Monte-Carlo tree search with deep neural networks that have been trained by supervised learning, from human expert games, and by reinforcement learning, from games of self-play. The first time ever that a computer program has defeated a human professional player.

作者:

David Silver1 *, Aja Huang1 *, Chris J. Maddison1 , Arthur Guez1 , Laurent Sifre1 , George van den Driessche1 , Julian Schrittwieser1 , Ioannis Antonoglou1 , Veda Panneershelvam1 , Marc Lanctot1 , Sander Dieleman1 , Dominik Grewe1 , John Nham2 , Nal Kalchbrenner1 , Ilya Sutskever2 , Timothy Lillicrap1 , Madeleine Leach1 , Koray Kavukcuoglu1 , Thore Graepel1 & Demis Hassabis1

2、Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model

Abstract

MuZero learns a model that, when applied iteratively, predicts the quantities most directly relevant to planning: the reward, the action-selection policy, and the value function. When evaluated on 57 different Atari games - the canonical video game environment for testing AI techniques, in which model-based planning approaches have historically struggled - our new algorithm achieved a new state of the art. When evaluated on Go, chess and shogi, without any knowledge of the game rules, MuZero matched the superhuman performance of the AlphaZero algorithm that was supplied with the game rules.

作者:

Julian SchrittwieserIoannis AntonoglouThomas HubertKaren SimonyanLaurent SifreSimon SchmittArthur GuezEdward LockhartDemis HassabisThore GraepelTimothy Lillicrap & David Silver

番外:作者探究-----------------------------------------

(1)Julian Schrittwieser 经历:

https://uk.linkedin.com/in/julian-schrittwieser-92476516

在deepmind已工作8年。

(2)Arthur Guez 经历:

He holds a M.Sc. in machine learning from McGill University in 2010 and a Ph.D. from Gatsby Computational Neuroscience Unit at University College London in 2015.

读博期间发表四篇论文。链接:Arthur Guez - Chessprogramming wiki

番外:论文探究-----------------------------------------

第一篇论文发表于27 January 2016,Nature.

最后一篇论文发表于23 December 2020,Nature.

2.2 AlphaStar: Grandmaster level in StarCraft II using multi-agent reinforcement learning

AlphaStar is the first AI to reach the top league of a widely popular esport without any game restrictions. This January, a preliminary version of AlphaStar challenged two of the world's top players in StarCraft II, one of the most enduring and popular real-time strategy video games of all time. 在星际争霸游戏中打败了top player。Since then, we have taken on a much greater challenge: playing the full game at a Grandmaster level under professionally approved conditions. 目标:以特技大师级水平完成游戏。

推荐阅读文章:

Neural networks have shown great success in everything from playing Go and Atari games to image recognition and language translation.

Population based training of neural networks | DeepMind

番外:团队探究-----------------------------------------

(1)Ziyu Wang -- Google Research Scientist

加拿大读研+博,牛津大学完成博士学业

https://ca.linkedin.com/in/ziyu-wang-433b4614

(2)Sara Sabour -- Google Research Scientist

多伦多大学研究生毕业,从SWE实习生三年做到Research SWE,两年后加入谷歌

读研期间实习 Architech,多伦多

https://ca.linkedin.com/in/sara-sabour-63019132?trk=public_profile_browsemap

(3)Yuhuai(Tony) Wu -- Google Research Scientist

本地大学毕业后多伦多大学博士,读博前后多次实习

读博前 Mila - Institut Québécois d'Intelligence Artificielle实习, 蒙特利尔

https://ca.linkedin.com/in/yuhuai-tony-wu-02a641b5

三、团队职员

(1)Yazhe Li

Senior Software Engineer at DeepMind

相关岗位:

Research- Scientist Edmonton, Canada

教育经历:

09年读完研究生后 Ecole Nationale des Ponts et Chaussées

14年申请了UCL,英国的研究生后读博士至今,同时Deepmind工作六年。

https://www.linkedin.com/in/yazhe-li-9980b8100/?originalSubdomain=uk

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值