说明:笔记旨在整理我校CS181课程的基本概念(PPT借用了Berkeley CS188)。由于授课及考试语言为英文,故英文出没可能。
目录
1 Markov Decision Processes mechanics
1.1 Markov Decision definitions
1 Markov Decision Processes mechanics
1.1 Markov Decision definitions
A MDP is defined by:
1.2 Markov 涵义
For markov decision processes, "Markov" means action outcomes depend only on the current state:
1.3 最优策略optimal policy
For MDP, we want an <