【openai_ros】3 - Exploring the OpenAI Structure: RoboCube. Part 1


Unit 2: How to apply openai_ros to a new robot
在这里插入图片描述

SUMMARY

Estimated time to completion: 10 minutes

In the previous unit, we introduced you the openai_ros package and how it works to train a robot with reinforcement learning.

In the following units, you will learn how to apply the openai_ros package to your own robot.

This unit is just a presentation of the robot we are going to use. We will use as an example the cube robot. However, you will see generic definitions so you can equally apply the method to any other robot.

END OF SUMMARY

在这里插入图片描述

The moving cube robot

This robot was developed by the ETHZurich and it’s a perfect platform for control theory and mechanical physics. You can find more information in this link: Cubli Research

In order to see the video below, select the next cell of the notebook and press the play button.

from IPython.display import YouTubeVideo
# Cubli Robot created By ETHZurich
# Video credit: William Stein.
YouTubeVideo('n_6p-1J551Y')

【注意】
油管上面可以看这个视频,搜’n_6p-1J551Y’即可
另外B站也可,观看链接

The simulation of the robot

We have created a simulation of the Cubli robot, the one you can see on the simulation window. This first simulated version has only One inertia disk, but it s enough to the goal we want to achieve.

The goal we want the robot to learn

We want to make this robot be able to walk around in the direction that we want. And it has to do it by learning by itself, not by using mathematical calculations as was done with the original.

So, the objective is to make the Cubli robot learn how to move forwards in the WORLD Y-AXIS direction.

In the next units of this course, we are going to learn how to build all the software parts to achieve that goal

Structure of the next units:

  1. On unit 3, you are going to create a new Robot Environment for the Cubli that allows you to access the sensors and actuators.
  2. On unit 4, you are going to create the Training Environment that inherits from the Robot Environment that you created. You will use this Environment to define the reward and the conditions to detect when the task is done. You will also use it to provide to the training algorithm the vector of observations, as well as to provide to the Robot Environment the actual commands to move based on the action decided by the training algorithm.
  3. On unit 5, you will create a Training Script for Qlearning and deepQ that uses your created Task Environment.

Switch to the next unit learn how to create a new Robot Environment for a new Robot

  • 1
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值