Conjoined architecture for heterogeneous multi-agent reinforcement learning다기종 멀티에이전트 강화학습을 위한 결합 아키텍처

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 3
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor김종환-
dc.contributor.authorHong, Chansol-
dc.contributor.author홍찬솔-
dc.date.accessioned2024-08-08T19:31:34Z-
dc.date.available2024-08-08T19:31:34Z-
dc.date.issued2024-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1100043&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/322143-
dc.description학위논문(박사) - 한국과학기술원 : 전기및전자공학부, 2024.2,[vi, 59 p. :]-
dc.description.abstractIn this dissertation, we propose the conjoined architecture designed for multi-agent reinforcement learning in environments using global state and heterogeneous agents. Under such environments, semantic mismatches in both input and output layers arise and hinder contemporary multi-agent reinforcement learning algorithms to efficiently train under centralized-training decentralized-execution settings using parameter sharing. On that regard, we propose the conjoined architecture capable of effectively train in environments using global state and heterogeneous agents. The conjoined architecture is a partially parameter sharing architecture where heterogeneous agents are considered as a single team to be trained together using global state as the input to be processed through a team network. Unlike traditional fully-centralized training, the conjoined architecture factorizes the output joint action space into individual agents' action spaces represented with agents' own weights and biases. We exemplify the use of the conjoined architecture through proposing two actor-critic algorithms multi-actor-conjoined-critic and conjoined-actor-conjoined-critic. A conjoined critic evaluates all agents' actions as a single sample. Instead of evaluating joint action-space values for all action combinations of agents, the conjoined critic outputs individual Q-values for each agent to reduce output dimension size. Through value decomposition network, individual Q-values are summed to estimate team Q-values, which is the optimization objective for the critic. For multi-actor-conjoined-critic, individual actors are trained with value estimations from conjoined critic while sharing their internal state among each other through bandwidth-limited communication channel. For conjoined-actor-conjoined-critic, a parameter-efficient conjoined actor is used in addition to the conjoined critic to replace individual actors. We evaluate the proposed algorithms in AI Soccer environment that uses global state and heterogeneous agents and compare the results with existing algorithms to demonstrate conjoined architecture's effectiveness. Finally, we conduct ablation studies to investigate effects of components in the proposed algorithms.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject기계학습▼a인공지능▼a강화학습▼a멀티에이전트 강화학습▼a다기종 에이전트-
dc.subjectMachine learning▼aArtificial intelligence▼aReinforcement learning▼aMulti-agent reinforcement learning▼aHeterogeneous agents-
dc.titleConjoined architecture for heterogeneous multi-agent reinforcement learning-
dc.title.alternative다기종 멀티에이전트 강화학습을 위한 결합 아키텍처-
dc.typeThesis(Ph.D)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :전기및전자공학부,-
dc.contributor.alternativeauthorKim, Jong-Hwan-
Appears in Collection
EE-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0