发明名称 Learning controller with advantage updating algorithm
摘要 A new algorithm for reinforcement learning, advantage updating, is proposed. Advantage updating is a direct learning technique; it does not require a model to be given or learned. It is incremental, requiring only a constant amount of calculation per time step, independent of the number of possible actions, possible outcomes from a given action, or number of states. Analysis and simulation indicate that advantage updating is applicable to reinforcement learning systems working in continuous time (or discrete time with small time steps) for which Q-learning is not applicable. Simulation results are presented indicating that for a simple linear quadratic regulator (LQR) problem with no noise and large time steps, advantage updating learns slightly faster than Q-learning. When there is noise or small time steps, advantage updating learns more quickly than Q-learning by a factor of more than 100,000. Convergence properties and implementation issues are discussed. New convergence results are presented for R-learning and algorithms based upon change in value. It is proved that the learning rule for advantage updating converges to the optimal policy with probability one.
申请公布号 US5608843(A) 申请公布日期 1997.03.04
申请号 US19940283729 申请日期 1994.08.01
申请人 THE UNITED STATES OF AMERICA AS REPRESENTED BY THE SECRETARY OF THE AIR FORCE 发明人 BAIRD, III, LEEMON C.
分类号 G05B13/02;G06F15/18;(IPC1-7):G06E1/00;G06E3/00 主分类号 G05B13/02
代理机构 代理人
主权项
地址