Learning to negotiate optimally in non-stationary environments

N/ACitations
Citations of this article
22Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We adopt the Markov chain framework to model bilateral negotiaamong agents in dynamic environments and use Bayesian learning to enable them to learn an optimal strategy in incomplete information settings. Specifically, an agent learns the optimal strategy to play against an opponent whose strategy varies with time, assuming no prior information about its negotiation parameters. In so doing, we present a new framework for adaptive negotiation in such non-stationary environments and develop a novel learning algorithm, which is guaranteed to converge, that an agent can use to negotiate optimally over time. We have implemented our algorithm and shown that it converges quickly in a wide range of cases. © Springer-Verlag Berlin Heidelberg 2006.

Cite

CITATION STYLE

APA

Narayanan, V., & Jennings, N. R. (2006). Learning to negotiate optimally in non-stationary environments. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4149 LNAI, pp. 288–300). Springer Verlag. https://doi.org/10.1007/11839354_21

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free