Transformer模型在强化学习中的应用相关论文

Representation Learning

  • My Body is a Cage: the Role of Morphology in Graph-Based Incompatible Control [PaperApr 2021
    Vitaly Kurin, Maximilian Igl, Tim Rocktäschel, Wendelin Boehmer, Shimon Whiteson
  • The Sensory Neuron as a Transformer: Permutation-Invariant NeuralNetworks for Reinforcement Learning [PaperSept 2021
    Yujin Tang, David Ha
  • CoBERL: Contrastive BERT for Reinforcement Learning [PaperFeb 2022
    Andrea Banino, Adrià Puidomenech Badia, Jacob Walker, Tim Scholtes, Jovana Mitrovic, Charles Blundell
  • Structure-Aware Transformer Policy for Inhomogeneous Multi-Task Reinforcement Learning [PaperMar 2022
    Sunghoon Hong, Deunsol Yoon, Kee-Eung Kim
  • Efficient Spatiotemporal Transformer for Robotic Reinforcement Learning [PaperJune 2022
    Yiming Yang, Dengpeng Xing, Bo Xu
  • Learning to Navigate in Interactive Environments with the Transformer-based Memory [PaperJune 2022
    Weiyuan Li, Ruoxin Hong, Jiwei Shen, Yue Lu
  • CtrlFormer: Learning Transferable State Representation for Visual Control via Transformer [PaperJune 2022
    Yao Mu, Shoufa Chen, Mingyu Ding, Jianyu Chen, Runjian Chen, Ping Luo
  • TVENet: Transformer-Based Visual Exploration Network for Mobile Robot in Unseen Environment [PaperJune 2022
    Tianyao Zhang, Xiaoguang Hu, Jin Xiao, Guofeng Zhang
  • StARformer: Transformer with State-Action-Reward Representations for Visual Reinforcement Learning [PaperJuly 2022
    Jinghuan Shang, Kumara Kahatapitiya, Xiang Li, Michael S. Ryoo
  • Multi-granularity scenarios understanding network for trajectory prediction [PaperAug 2022
    Biao Yang, Jicheng Yang, Rongrong Ni, Changchun Yang, Xiaofeng Liu
  • One for All: One-stage Referring Expression Comprehension with Dynamic Reasoning [PaperOct 2022
    Zhipeng Zhang, Zhimin Wei, Zhongzhen Huang, Rui Niu, Peng Wang
  • Unsupervised Learning of Temporal Abstractions with Slot-based Transformers [PaperNov 2022
    Anand Gopalakrishnan, Kazuki Irie, Jürgen Schmidhuber, Sjoerd van Steenkiste

World Model

  • Improving Sample Efficiency of Value Based Models Using Attention and Vision Transformers [PaperFeb 2022
    Amir Ardalan Kalantari, Mohammad Amini, Sarath Chandar, Doina Precup
  • Dreaming with Transformers [PaperFeb 2022
    Catherine Zeng, Jordan Docter, Alexander Amini, Igor Gilitschenski, Ramin Hasani and Daniela Rus
  • TransDreamer: Reinforcement Learning with Transformer World Models [PaperFeb 2022
    Chang Chen, Yi-Fu Wu, Jaesik Yoon, Sungjin Ahn
  • A model-based approach to meta-Reinforcement Learning: Transformers and tree search [PaperAug 2022
    Brieuc Pinon, Jean-Charles Delvenne, Raphaël Jungers
  • Transformers are Sample Efficient World Models [PaperSep 2022
    Vincent Micheli, Eloi Alonso, François Fleuret
  • Masked World Models for Visual Control [PaperNov 2022
    Younggyo Seo, Danijar Hafner, Hao Liu, Fangchen Liu, Stephen James, Kimin Lee, Pieter Abbeel
  • Transformers-Based World Models Are Happy With 100k Interactions [PaperNov 2022
    Jan Robine, Jan_Robine, Marc Höftmann, Tobias Uelwer, Stefan Harmeling

Reward Learning

  • Comparing BERT-based Reward Functions for Deep Reinforcement Learning in Machine Translation [PaperOct 2022
    Yuki Nakatani, Tomoyuki Kajiwara, Takashi Ninomiya

Policy Leaning

  • UPDeT: Universal Multi-agent Reinforcement Learning via Policy Decoupling with Transformers [PaperFeb 2021
    Siyi Hu, Fengda Zhu, Xiaojun Chang, Xiaodan Liang
  • Offline Reinforcement Learning as One Big Sequence Modeling Problem [PaperJun 2021
    Michael Janner, Qiyang Li, Sergey Levine
  • Decision Transformer: Reinforcement Learning via Sequence Modeling [PaperJun 2021
    Lili Chen, Kevin Lu, Aravind Rajeswaran, Kimin Lee, Aditya Grover, Michael Laskin, Pieter Abbeel, Aravind Srinivas, Igor Mordatch
  • A Novel Deep Reinforcement Learning Based Stock Direction Prediction using Knowledge Graph and Community Aware Sentiments [PaperJuly 2021
    Anil Berk Altuner, Zeynep Hilal Kilimci
  • Pretraining for Language Conditioned Imitation with Transformers [PaperSept 2021
    Aaron L Putterman, Kevin Lu, Igor Mordatch, Pieter Abbeel
  • CubeTR: Learning to Solve the Rubik's Cube using Transformers [PaperSept 2021
    Mustafa Ebrahim Chasmai
  • Transfer learning with causal counterfactual reasoning in Decision Transformers [PaperOct 2021
    Ayman Boustati, Hana Chockler, Daniel C. McNamee
  • Offline Pre-trained Multi-Agent Decision Transformer: One Big Sequence Model Tackles All SMAC Tasks [PaperDec 2021
    Linghui Meng, Muning Wen, Yaodong Yang, Chenyang Le, Xiyun Li, Weinan Zhang, Ying Wen, Haifeng Zhang, Jun Wang, Bo Xu
  • Switch Trajectory Transformer with Distributional Value Approximation for Multi-Task Reinforcement Learning [PaperMar 2022
    Qinjie Lin, Han Liu, Biswa Sengupta
  • FedFormer: Contextual Federation with Attention in Reinforcement Learning [PaperMay 2022
    Liam Hebert, Lukasz Golab, Pascal Poupart, Robin Cohen
  • Transformer with Memory Replay [PaperMay 2022
    Rui Liu, Barzan Mozafari
  • Online Decision Transformer [PaperJuly 2022
    Qinqing Zheng, Amy Zhang, Aditya Grover
  • Multi-Game Decision Transformers [PaperMay 2022
    Kuang-Huei Lee, Ofir Nachum, Mengjiao Yang, Lisa Lee, Daniel Freeman, Winnie Xu, Sergio Guadarrama, Ian Fischer, Eric Jang, Henryk Michalewski, Igor Mordatch
  • Prompting Decision Transformer for Few-Shot Policy Generalization [PaperJune 2022
    Mengdi Xu, Yikang Shen, Shun Zhang, Yuchen Lu, Ding Zhao, Joshua B. Tenenbaum, Chuang Gan
  • Offline Pre-trained Multi-Agent Decision Transformer: One Big Sequence Model Tackles All SMAC Tasks [PaperJun 2022
    Linghui Meng, Muning Wen, Yaodong Yang, Chenyang Le, Xiyun Li, Weinan Zhang, Ying Wen, Haifeng Zhang, Jun Wang, Bo Xu
  • Transformers are Meta-Reinforcement Learners [PaperJune 2022
    Luckeciano C. Melo
  • Transformer-based Value Function Decomposition for Cooperative Multi-agent Reinforcement Learning in StarCraft [PaperAug 2022
    Muhammad Junaid Khan, Syed Hammad Ahmed, Gita Sukthankar
  • Transformer-Based Deep Reinforcement Learning in VizDoom [PaperAug 2022
    Vitalii Sopov, Ilya Makarov
  • Q-learning Decision Transformer: Leveraging Dynamic Programming for Conditional Sequence Modelling in Offline RL [PaperOct 2022
    Taku Yamagata, Ahmed Khalil, Raul Santos-Rodriguez
  • Bootstrapped Transformer for Offline Reinforcement Learning [PaperOct 2022
    Kerong Wang, Hanye Zhao, Xufang Luo, Kan Ren, Weinan Zhang, Dongsheng Li
  • Relational Reasoning via Set Transformers: Provable Efficiency and Applications to MARL [PaperOct 2022
    Fengzhuo Zhang, Boyi Liu, Kaixin Wang, Vincent Y. F. Tan, Zhuoran Yang, Zhaoran Wang
  • Multi-Agent Reinforcement Learning is a Sequence Modeling Problem [PaperOct 2022
    Muning Wen, Jakub Grudzien Kuba, Runji Lin, Weinan Zhang, Ying Wen, Jun Wang, Yaodong Yang
  • You Can't Count on Luck: Why Decision Transformers and RvS Fail in Stochastic Environments [PaperNov 2022
    Keiran Paster, Sheila McIlraith, Jimmy Ba
  • Deep Transformer Q-Networks for Partially Observable Reinforcement Learning [PaperNov 2022
    Kevin Esslinger, Robert Platt, Christopher Amato

Training Strategy

  • Stabilizing Transformers for Reinforcement Learning [PaperOct 2019
    Emilio Parisotto, H. Francis Song, Jack W. Rae, Razvan Pascanu, Caglar Gulcehre, Siddhant M. Jayakumar, Max Jaderberg, Raphael Lopez Kaufman, Aidan Clark, Seb Noury, Matthew M. Botvinick, Nicolas Heess, Raia Hadsell
  • Semi-supervised Offline Reinforcement Learning with Pre-trained Decision Transformers [PaperSept 2021
    Catherine Cang, Kourosh Hakhamaneshi, Ryan Rudes, Igor Mordatch, Aravind Rajeswaran, Pieter Abbeel, Michael Laskin
  • Can Wikipedia Help Offline Reinforcement Learning? [PaperJan 2022
    Machel Reid, Yutaro Yamada, Shixiang Shane Gu
  • Evaluating Vision Transformer Methods for Deep Reinforcement Learning from Pixels [PaperApr 2022
    Tianxin Tao, Daniele Reda, Michiel van de Panne

Explainability

  • ATTEXPLAINER: Explain Transformer via Attention by Reinforcement Learning [PaperJuly 2022
    Runliang Niu, Zhepei Wei, Yan Wang, Qi Wang

 

 

  • 0
    点赞
  • 3
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值