Artificial Intelligence and Spontaneous Collusion (2202.05946v5)
Abstract: We develop a tractable model for studying strategic interactions between learning algorithms. We uncover a mechanism responsible for the emergence of algorithmic collusion. We observe that algorithms periodically coordinate on actions that are more profitable than static Nash equilibria. This novel collusive channel relies on an endogenous statistical linkage in the algorithms' estimates which we call spontaneous coupling. The model's parameters predict whether the statistical linkage will appear, and what market structures facilitate algorithmic collusion. We show that spontaneous coupling can sustain collusion in prices and market shares, complementing experimental findings in the literature. Finally, we apply our results to design algorithmic markets.
- Abada, I. and X. Lambin (2023): “Artificial intelligence: Can seemingly collusive outcomes be avoided?” Management Science.
- Abreu, D. and H. Matsushima (1992): “Virtual implementation in iteratively undominated strategies: complete information,” Econometrica, 993–1008.
- Asker, J., C. Fershtman, and A. Pakes (2022): “Artificial Intelligence, Algorithm Design and Pricing,” American Economic Review, P&P, forthcoming.
- Banchio, M. and A. Skrzypacz (2022): “Artificial Intelligence and Auction Design,” Working paper.
- Benaim, M. (1996): “A Dynamical System Approach to Stochastic Approximations,” SIAM Journal of Control and Optimization, 34, 437–472.
- Bendor, J., D. Mookherjee, and D. Ray (2001a): “Aspiration-Based Reinforcement Learning in Repeated Interaction Games: An Overview,” International Game Theory Review, 03, 159–174.
- ——— (2001b): “Reinforcement Learning in Repeated Interaction Games,” The B.E. Journal of Theoretical Economics, 1.
- Bhandari, J., D. Russo, and R. Singal (2021): “A finite time analysis of temporal difference learning with linear function approximation,” Operations Research, 69, 950–973.
- Borkar, V. S. and S. P. Meyn (2000): “The O.D.E. Method for Convergence of Stochastic Approximation and Reinforcement Learning,” SIAM J. Control. Optim., 38, 447–469.
- Brown, Z. Y. and A. MacKay (2021): “Competition in Pricing Algorithms,” American Economic Journal: Microeconomics, forthcoming.
- Börgers, T. and R. Sarin (1997): “Learning Through Reinforcement and Replicator Dynamics,” Journal of Economic Theory, 77, 1–14.
- Calvano, E., G. Calzolari, V. Denicolo, and S. Pastorello (2020): “Artificial intelligence, algorithmic pricing, and collusion,” American Economic Review, 110, 3267–97.
- Competition & Markets Authority (2021): “Algorithms: How they can reduce competition and harm consumers,” Discussion paper.
- Competition Bureau (2018): “Big Data and Innovation: Implications for Competition Policy in Canada,” Discussion paper.
- Dieci, L. and L. Lopez (2011): “Sliding motion on discontinuity surfaces of high co-dimension. A construction for selecting a Filippov vector field,” Numerische Mathematik, 117, 779–811.
- Erev, I., Y. Bereby-Meyer, and A. E. Roth (1999): “The effect of adding a constant to all payoffs: experimental investigation, and implications for reinforcement learning models,” Journal of Economic Behavior & Organization, 39, 111–128.
- Erev, I. and A. E. Roth (1998): “Predicting How People Play Games: Reinforcement Learning in Experimental Games with Unique, Mixed Strategy Equilibria,” The American Economic Review, 88, 848–881.
- Gomes, E. R. and R. Kowalczyk (2009): “Dynamic Analysis of Multiagent Q-Learning with ϵitalic-ϵ\epsilonitalic_ϵ-Greedy Exploration,” in Proceedings of the 26th Annual International Conference on Machine Learning, 369–376.
- Gonczarowski, Y., O. Heffetz, and C. Thomas (2023): “Strategyproofness-Exposing Mechanism Descriptions,” Working paper.
- Hammond, P. J. (1979): “Straightforward Individual Incentive Compatibility in Large Economies,” The Review of Economic Studies, 46, 263–282.
- Hansen, K., K. Misra, and M. Pai (2021): “Algorithmic Collusion: Supra-Competitive Prices via Independent Algorithms,” Marketing Science, 40, 1–12.
- Harrington, J. E. (2018): “Developing Competition Law for Collusion by Autonomous Artificial Agents,” Journal of Competition Law & Economics, 14, 331–363.
- ——— (2022): “The effect of outsourcing pricing algorithms on market competition,” Management Science, 68, 6889–6906.
- Hopkins, E. and M. Posch (2005): “Attainability of boundary points under reinforcement learning,” Games and Economic Behavior, 53, 110–125.
- Johnson, J., A. Rhodes, and M. Wildenbeest (2023): “Platform Design when Sellers Use Pricing Algorithms,” Econometrica, forthcoming.
- Karandikar, R., D. Mookherjee, D. Ray, and F. Vega-Redondo (1998): “Evolving Aspirations and Cooperation,” Journal of Economic Theory, 80, 292–331.
- Klein, T. (2021): “Autonomous algorithmic collusion: Q-learning under sequential pricing,” The RAND Journal of Economics, 52, 538–558.
- Kurtz, T. G. (1970): “Solutions of ordinary differential equations as limits of pure jump Markov processes,” Journal of Applied Probability, 7, 49–58.
- Lamba, R. and S. Zhuk (2022): “Pricing with Algorithms,” working paper.
- Leisten, M. (2022): “Algorithmic Competition, with Humans,” Working paper.
- Leonardos, S. and G. Piliouras (2022): “Exploration-exploitation in multi-agent learning: Catastrophe theory meets game theory,” Artificial Intelligence, 304.
- Lerer, A. and A. Peysakhovich (2017): ‘‘Maintaining cooperation in complex social dilemmas using deep reinforcement learning,” arXiv preprint arXiv:1707.01068.
- Mertikopoulos, P. and W. H. Sandholm (2016): “Learning in games via reinforcement and regularization,” Mathematics of Operations Research, 41, 1297–1324.
- Musolff, L. A. (2021): “Algorithmic Pricing Facilitates Tacit Collusion: Evidence from E-Commerce,” Working paper.
- OECD (2017): “Algorithms and Collusion: Competition Policy in the Digital Age,” Technical report.
- Parkes, D. C. (2004): “On learnable mechanism design,” in Collectives and the Design of Complex Systems, Springer, 107–131.
- Possnig, C. (2023): “Reinforcement Learning and Collusion,” Working paper.
- Tumer, K. and N. Khani (2009): “Learning from actions not taken in multiagent systems,” Advances in Complex Systems, 12, 455–473.
- Tuyls, K., P. J. Hoen, and B. Vanschoenwinkel (2005): “An Evolutionary Dynamical Analysis of Multi-Agent Learning in Iterated Games,” Autonomous Agents and Multi-Agent Systems, 12, 115–153.
- Watkins, C. J. and P. Dayan (1992): “Q-learning,” Machine learning, 8, 279–292.
- Wunder, M., M. L. Littman, and M. Babes (2010): “Classes of Multiagent Q-learning Dynamics with epsilon-greedy Exploration,” in Proceedings of the 27th Annual International Conference on Machine Learning, 1167–1174.