Comparison between reinforcement learning methods with different goal selections in multi-agent cooperation

Fumito Uwano, Keiki Takadama

研究成果査読

2 被引用数 (Scopus)

抄録

This study discusses important factors for zero communication, multi-agent cooperation by comparing different modified reinforcement learning methods. The two learning methods used for comparison were assigned different goal selections for multi-agent cooperation tasks. The first method is called Profit Minimizing Reinforcement Learning (PMRL); it forces agents to learn how to reach the farthest goal, and then the agent closest to the goal is directed to the goal. The second method is called Yielding Action Reinforcement Learning (YARL); it forces agents to learn through a Q-learning process, and if the agents have a conflict, the agent that is closest to the goal learns to reach the next closest goal. To compare the two methods, we designed experiments by adjusting the following maze factors: (1) the location of the start point and goal; (2) the number of agents; and (3) the size of maze. The intensive simulations performed on the maze problem for the agent cooperation task revealed that the two methods successfully enabled the agents to exhibit cooperative behavior, even if the size of the maze and the number of agents change. The PMRL mechanism always enables the agents to learn cooperative behavior, whereas the YARL mechanism makes the agents learn cooperative behavior over a small number of learning iterations. In zero communication, multi-agent cooperation, it is important that only agents that have a conflict cooperate with each other.

本文言語English
ページ(範囲)917-929
ページ数13
ジャーナルJournal of Advanced Computational Intelligence and Intelligent Informatics
21
5
DOI
出版ステータスPublished - 9月 2017
外部発表はい

ASJC Scopus subject areas

  • 人間とコンピュータの相互作用
  • コンピュータ ビジョンおよびパターン認識
  • 人工知能

フィンガープリント

「Comparison between reinforcement learning methods with different goal selections in multi-agent cooperation」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル