• MDPs


    Markov Decision Processes  (MDPs)

    Named after Andrey Markov, known at least as early as 1950s.(cf. Bellman 1957)

    Discrete time stochastic control process.

    State:

    Action:

    Reward:

    Markov property: given s and a , it conditionally independen of all previous state and actions.

    Extent of Markov chains: 

    Diff.: addition of actions (allowing choice) and rewards (giving motivation).

    Formally,

    Defination

    a  4-tuple (S,A,P.(.,.),R.(.,.), where

    • S is a finite set of states,
    • A is a finite set of actions (alternatively, A_s is the finite set of actions available from state s),
    • P_a(s, s’) = Pr(s_{t+1} = s’ | s_t = s, a_t = a) is the probability that action a in state s at time t will lead to state s’ at time t+1
    • R_a(s,s’) is the immediate reward (or expected immediate reward) received after transition to state s’ from state s with transition probability P_a(s,s’)

    Problem

    Find a policy for decision maker: a function \pai that specifies that action \pai 9s) that the decision maker will choose when in state s.

    The goal is to choose a policy \pai that will maximize some cumulative function of random rewards:

    \sum_{t=0]{\gama^t R_{a_t}(s_t,s_{t+1})  (where a_t = \pai (s_t})

    where \gama is the discount rate and \in (0,1]. It is typically close to 1.

    Algorithm

     \pi(s) := \arg \max_a \left\{ \sum_{s'} P_a(s,s') \left( R_a(s,s') + \gamma V(s') \right) \right\}

     V(s) := \sum_{s'} P_{\pi(s)} (s,s') \left( R_{\pi(s)} (s,s') + \gamma V(s') \right)

    P : transition function

    R: Reward function

    V: contains real values, V(s) will contain the discounted sum of the rewards to be earned (on average ) bye following that solution from state s.

    Notable variants

    Value iteration

     V(s) := \max_a \left\{ \sum_{s'} P_a(s,s') \left( R_a(s,s') + \gamma V(s') \right) \right\}.

    Polycy iteration \ Modified policy iteration\ Prioritized sweeping

    Extension and generalizations

    Partial observability (POMDP)

    RL:

    Probilities or rewards are unknown.

    Q(s,a) = \sum_{s’}{Pa(s,s])(R_a{s,s]) + \gammaV(s’))}

    I was in state s and I tried doing a and s’ happened.

  • 相关阅读:
    关于ThreadLocal的理解
    常用Linux软件安装
    Spring事务注解@Transactional失效的问题
    使用jackson转换xml格式数据进行响应
    创建简单web程序了解servlet
    JDBC
    StringBuild类
    Canlendar 日期类
    Java Date 时间类的使用
    QWeb
  • 原文地址:https://www.cnblogs.com/justin_s/p/2038228.html
Copyright © 2020-2023  润新知