(1.2)bark-ml

github开源bark-ml:https://github.com/bark-simulator/bark-ml/

bark官网:https://bark-simulator.github.io/

bark-ml指导手册:https://bark-simulator.github.io/tutorials/bark_ml_getting_started/

Youtube视频:https://www.youtube.com/user/fortissTV/search?query=BARK

相关论文:

一、项目下载

# 新建一个Anaconda环境
conda create -n bark-ml python=3.7
# 如果想删除环境采用以下操作
# conda remove -n your_env_name(虚拟环境名称) --all
# 进入环境
conda activate bark-ml
pip install virtualenv==16.7.8
pip install bark-ml
git clone https://github.com/bark-simulator/bark-ml
cd bark-ml

二、配置环境

1、修改install.sh

which python  # 查看自己的python路径:/home/myx/anaconda3/envs/bark-ml/bin/python
sudo vim utils/install.sh  # 修改install.sh文件

python3.7替换为刚刚得到的自己的python路径/home/myx/anaconda3/envs/bark-ml/bin/python

image-20220720193857906

bash utils/install.sh
# 可能需要自己运行一下
# pip install torch==1.9.0+cpu torchvision -f https://download.pytorch.org/whl/torch_stable.html
# 进入虚拟环境
source utils/dev_into.sh

2、修改WORKSPACE

sudo gedit WORKSPACE

将原句

python_configure(name = “local_config_python”)

改为

python_configure(name = “local_config_python”, python_version = “3”)

image-20220720201814346

三、测试

# source utils/dev_into.sh
# deactivate
bazel test //...
bazel test //bark_ml:unit_tests
bazel run //examples:continuous_env
bazel run //examples:blueprint_config
bazel run //examples:tfa
bazel run //examples:tfa_gnn -- --mode=train
bazel run //examples:tfa_counterfactual -- --mode=train
# 如果想训练
# bazel run //examples:tfa -- --mode=train
  • 新建自己的参数文件夹 myx_params
/home/myx/Project/bark-ml/examples/myx_checkpoints/
/home/myx/Project/bark-ml/bark_ml/environments/blueprints/merging/merging.py
# 机器人行为和一大堆参数

/home/myx/anaconda3/envs/bark-ml/lib/python3.7/site-packages/tf_agents/drivers/dynamic_episode_driver.py
# 跑循环神经网络训练_loop_body_fn

/home/myx/Project/bark-ml/bark_ml/environments/blueprints/merging/merging.py
# evaluator 设置

/home/myx/Project/bark-ml/bark_ml/environments/single_agent_runtime.py
# reward, done, info = self._evaluator.Evaluate(

/home/myx/Project/bark-ml/bark_ml/evaluators/general_evaluator.py
  def Evaluate(self, observed_world, action):
    """Returns information about the current world state."""
    eval_results = observed_world.Evaluate()
    reward = 0.
    scheduleTerminate = False

    for _, eval_fn in self._bark_ml_eval_fns.items():
      t, r, i = eval_fn(observed_world, action, eval_results)
      eval_results = {**eval_results, **i} # merge info
      reward += r # accumulate reward
      if t: # if any of the t are True -> terminal
        scheduleTerminate = True

    return reward, scheduleTerminate, eval_results


/home/myx/anaconda3/envs/bark-ml/lib/python3.7/site-packages/tf_agents/agents/ppo/ppo_agent.py
# ppo 训练代码
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值