2009年
Robust Approximation in Decomposed Reinforcement Learning
NEURAL INFORMATION PROCESSING, PT 1, PROCEEDINGS
- ,
- 巻
- 5863
- 号
- 開始ページ
- 590
- 終了ページ
- 597
- 記述言語
- 英語
- 掲載種別
- DOI
- 10.1007/978-3-642-10677-4_67
- 出版者・発行元
- SPRINGER-VERLAG BERLIN
Recently, an efficient reinforcement learning method has been proposed, in which the problem of approximating the value function is naturally decomposed into a number of sub-problems, each of which can be solved at small computational cost. While tins method certainly reduces the magnitude of temporal difference error, the value function may be overfitted to sampled data. To overcome this difficulty, we introduce a robust approximation to tins context. Computer experiments show that the value function learning by our method is much more robust than those by the previous methods.
- リンク情報
- ID情報
-
- DOI : 10.1007/978-3-642-10677-4_67
- ISSN : 0302-9743
- Web of Science ID : WOS:000279118000067