We adopt the Markov chain framework to model bilateral negotiaamong agents in dynamic environments and use Bayesian learning to enable them to learn an optimal strategy in incomplete information settings. Specifically, an agent learns the optimal strategy to play against an opponent whose strategy varies with time, assuming no prior information about its negotiation parameters. In so doing, we present a new framework for adaptive negotiation in such non-stationary environments and develop a novel learning algorithm, which is guaranteed to converge, that an agent can use to negotiate optimally over time. We have implemented our algorithm and shown that it converges quickly in a wide range of cases. © Springer-Verlag Berlin Heidelberg 2006.
CITATION STYLE
Narayanan, V., & Jennings, N. R. (2006). Learning to negotiate optimally in non-stationary environments. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4149 LNAI, pp. 288–300). Springer Verlag. https://doi.org/10.1007/11839354_21
Mendeley helps you to discover research relevant for your work.