MISC

査読有り
2009年

Robust Approximation in Decomposed Reinforcement Learning

NEURAL INFORMATION PROCESSING, PT 1, PROCEEDINGS
  • Takeshi Mori
  • ,
  • Shin Ishii

5863
開始ページ
590
終了ページ
597
記述言語
英語
掲載種別
DOI
10.1007/978-3-642-10677-4_67
出版者・発行元
SPRINGER-VERLAG BERLIN

Recently, an efficient reinforcement learning method has been proposed, in which the problem of approximating the value function is naturally decomposed into a number of sub-problems, each of which can be solved at small computational cost. While tins method certainly reduces the magnitude of temporal difference error, the value function may be overfitted to sampled data. To overcome this difficulty, we introduce a robust approximation to tins context. Computer experiments show that the value function learning by our method is much more robust than those by the previous methods.

リンク情報
DOI
https://doi.org/10.1007/978-3-642-10677-4_67
Web of Science
https://gateway.webofknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcAuth=JSTA_CEL&SrcApp=J_Gate_JST&DestLinkType=FullRecord&KeyUT=WOS:000279118000067&DestApp=WOS_CPL
ID情報
  • DOI : 10.1007/978-3-642-10677-4_67
  • ISSN : 0302-9743
  • Web of Science ID : WOS:000279118000067

エクスポート
BibTeX RIS