Multiagent reinforcement learning with spiking and non-spiking agents in the iterated prisoner's dilemma

6Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper investigates Multiagent Reinforcement Learning (MARL) in a general-sum game where the payoffs' structure is such that the agents are required to exploit each other in a way that benefits all agents. The contradictory nature of these games makes their study in multiagent systems quite challenging. In particular, we investigate MARL with spiking and non-spiking agents in the Iterated Prisoner's Dilemma by exploring the conditions required to enhance its cooperative outcome. According to the results, this is enhanced by: (i) a mixture of positive and negative payoff values and a high discount factor in the case of non-spiking agents and (ii) having longer eligibility trace time constant in the case of spiking agents. Moreover, it is shown that spiking and non-spiking agents have similar behaviour and therefore they can equally well be used in any multiagent interaction setting. For training the spiking agents, a novel and necessary modification enhances competition to an existing learning rule based on stochastic synaptic transmission. © 2009 Springer Berlin Heidelberg.

Cite

CITATION STYLE

APA

Vassiliades, V., Cleanthous, A., & Christodoulou, C. (2009). Multiagent reinforcement learning with spiking and non-spiking agents in the iterated prisoner’s dilemma. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5768 LNCS, pp. 737–746). https://doi.org/10.1007/978-3-642-04274-4_76

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free