Number of the records: 1
Nash Q-learning agents in Hotelling's model: Reestablishing equilibrium
- 1.0542311 - ÚTIA 2022 RIV NL eng J - Journal Article
Vainer, J. - Kukačka, Jiří
Nash Q-learning agents in Hotelling's model: Reestablishing equilibrium.
Communications in Nonlinear Science and Numerical Simulation. Roč. 99, č. 1 (2021), č. článku 105805. ISSN 1007-5704. E-ISSN 1878-7274
Grant - others:Univerzita Karlova(CZ) PRIMUS/19/HUM/17; Univerzita Karlova(CZ) UNCE/HUM/035
Institutional support: RVO:67985556
Keywords : Hotelling’s location model * Agent-based simulation * Reinforcement learning * Nash Q-learning
OECD category: Economic Theory
Impact factor: 4.186, year: 2021 ; AIS: 0.853, rok: 2021
Method of publishing: Limited access
Result website:
http://library.utia.cas.cz/separaty/2021/E/kukacka-0542311.pdf https://www.sciencedirect.com/science/article/pii/S1007570421001167
DOI: https://doi.org/10.1016/j.cnsns.2021.105805
This paper examines adaptive agents’ behavior in a stochastic dynamic version of the Hotelling’s location model. We conduct an agent-based numerical simulation under the Hotelling’s setting with two agents who use the Nash Q-learning mechanism for adaptation. This allows exploring what alternations this technique brings compared to the original analytic solution of the famous static game-theoretic model with strong assumptions imposed on players. We discover that under the Nash Q-learning and quadratic consumer cost function, agents with high enough valuation of future profits learn behavior similar to aggressive market strategy. Both agents make similar products and lead a price war to eliminate their opponent from the market. This behavior closely resembles the Principle of Minimum Differentiation from Hotelling’s original paper with linear consumer costs. However, the quadratic consumer cost function would otherwise result in the maximum differentiation of production in the original model. Thus, the Principle of Minimum Differentiation can be justified based on repeated interactions of the agents and long-run optimization.
Permanent Link: http://hdl.handle.net/11104/0320103
Number of the records: 1