Understanding the mechanism of human–computer game: a distributed reinforcement learning perspective

Zhinan Peng, Jiangping Hu, Yiyi Zhao, Bijoy K. Ghosh

Research output: Contribution to journalArticlepeer-review

Abstract

In this paper, the mechanism of the human–computer game is investigated with the help of multi-agent systems (MASs) and reinforcement learning (RL). The game is formulated as a bipartite consensus problem while the interactions among humans and computers are modelled as a multi-agent system over a coopetition network. The coopetition network associated with the multi-agent system is represented by a signed graph, where positive/negative edges denote cooperative/competitive interactions. We assume the decision mechanism of the agents are model free and each agent has to make a distributed decision by learning the input and output data from himself/itself and his/its neighbours. The individual decision is developed with the neighbours' state information and a performance index function. A policy iteration (PI) algorithm is proposed to solve the Hamilton-Jacobi-Bellman equation and obtain the optimal decision strategy. Furthermore, an actor-critic neural network is adopted to approximate the performance index and the optimal decision strategy in an online manner. The simulation results are finally given to validate the proposed reinforcement learning approach.

Original languageEnglish
Pages (from-to)2837-2848
Number of pages12
JournalInternational Journal of Systems Science
Volume51
Issue number15
DOIs
StatePublished - Nov 17 2020

Keywords

  • Human–computer game
  • bipartite consensus
  • coopetition network
  • multi-agent system
  • reinforcement learning

Fingerprint

Dive into the research topics of 'Understanding the mechanism of human–computer game: a distributed reinforcement learning perspective'. Together they form a unique fingerprint.

Cite this