論文

査読有り
2005年

Reinforcement learning of optimal supervisor based on language measure

2005 44TH IEEE CONFERENCE ON DECISION AND CONTROL & EUROPEAN CONTROL CONFERENCE, VOLS 1-8
  • Tatsushi Yamasaki
  • ,
  • Kazutaka Taniguchi
  • ,
  • Toshimitsu Ushio

開始ページ
126
終了ページ
131
記述言語
英語
掲載種別
研究論文(国際会議プロシーディングス)
出版者・発行元
IEEE

Recently, Wang and Ray introduced a signed real measure for formal languages, called a language measure, to evaluate performance of strings generated by discrete event systems. They proposed a synthesis method of an optimal supervisor based on the language measure. If exact description of a discrete event system and the specification is not available, a learning-based approach is useful. In this paper, first, we clarify the relationship between the Bellman equation and a performance index of the languages generated by the controlled discrete event systems. Next, using the relationship, we propose a learning method of the optimal supervisor based on reinforcement learning where costs of disabling of events and the evaluation of reaching states are taken into consideration. Finally, by computer simulation, we illustrate an efficiency of the proposed method.

リンク情報
Web of Science
https://gateway.webofknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcAuth=JSTA_CEL&SrcApp=J_Gate_JST&DestLinkType=FullRecord&KeyUT=WOS:000240653700022&DestApp=WOS_CPL
ID情報
  • ISSN : 0191-2216
  • Web of Science ID : WOS:000240653700022

エクスポート
BibTeX RIS