Papers

Peer-reviewed
2017

Communication-Less Cooperative Q-Learning Agents in Maze Problem

INTELLIGENT AND EVOLUTIONARY SYSTEMS, IES 2016
  • Fumito Uwano
  • ,
  • Keiki Takadama

Volume
8
Number
First page
453
Last page
467
Language
English
Publishing type
Research paper (international conference proceedings)
DOI
10.1007/978-3-319-49049-6_33
Publisher
SPRINGER INT PUBLISHING AG

This paper introduces a reinforcement learning technique with an internal reward for a multi-agent cooperation task. The proposed method is an extension of Q-learning which changes the ordinary (external) reward to the internal reward for agent-cooperation under the condition of no communication. To increase the certainty of the proposed methods, we theoretically investigate what values should be set to select the goal for the cooperation among agents. In order to show the effectiveness of the proposed method, we conduct the intensive simulation on the maze problem for the agent-cooperation task, and confirm the following implications: (1) the proposed method successfully enable agents to acquire cooperative behaviors while a conventional method fails to always acquire such behaviors; (2) the cooperation among agents according to their internal rewards is achieved no communication; and (3) the condition for the cooperation among any number of agent is indicated.

Link information
DOI
https://doi.org/10.1007/978-3-319-49049-6_33
Web of Science
https://gateway.webofknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcAuth=JSTA_CEL&SrcApp=J_Gate_JST&DestLinkType=FullRecord&KeyUT=WOS:000402759400033&DestApp=WOS_CPL
ID information
  • DOI : 10.1007/978-3-319-49049-6_33
  • ISSN : 2363-6084
  • Web of Science ID : WOS:000402759400033

Export
BibTeX RIS