Visible to the public 1-recall reinforcement learning leading to an optimal equilibrium in potential games with discrete and continuous actions

Title1-recall reinforcement learning leading to an optimal equilibrium in potential games with discrete and continuous actions
Publication TypeConference Paper
Year of Publication2015
AuthorsTatarenko, T.
Conference Name2015 54th IEEE Conference on Decision and Control (CDC)
Date Publisheddec
Keywords1-recall reinforcement learning, agent dynamics, automata theory, continuous actions, convergence, decision making, discrete actions, Distributed optimization, game theory, Games, learning (artificial intelligence), learning automata, Linear programming, Markov processes, multi-agent systems, multiagent systems, Nash equilibria, Nash equilibrium, optimal equilibrium, Optimization, payoff-based learning, potential games, pubcrawl170110, stochastic approximation, Stochastic processes, strategic decision makers
Abstract

Game theory serves as a powerful tool for distributed optimization in multiagent systems in different applications. In this paper we consider multiagent systems that can be modeled as a potential game whose potential function coincides with a global objective function to be maximized. This approach renders the agents the strategic decision makers and the corresponding optimization problem the problem of learning an optimal equilibruim point in the designed game. In distinction from the existing works on the topic of payoff-based learning, we deal here with the systems where agents have neither memory nor ability for communication, and they base their decision only on the currently played action and the experienced payoff. Because of these restrictions, we use the methods of reinforcement learning, stochastic approximation, and learning automata extensively reviewed and analyzed in [3], [9]. These methods allow us to set up the agent dynamics that moves the game out of inefficient Nash equilibria and leads it close to an optimal one in both cases of discrete and continuous action sets.

DOI10.1109/CDC.2015.7403282
Citation Keytatarenko_1-recall_2015