ESSEC METALAB

RESEARCH

ARTIFICIAL INTELLIGENCE: CAN SEEMINGLY COLLUSIVE OUTCOMES BE AVOIDED?

[ARTICLE] This paper studies how companies using machine learning to compete can unintentionally act like a cartel, even without communicating. It explores ways to regulate this to prevent harm to consumers.

by Xavier Lambin (ESSEC Business School), Ibrahim Abada

Strategic decisions are increasingly delegated to algorithms. We extend previous results of the algorithmic collusion literature to the context of dynamic optimization with imperfect monitoring by analyzing a setting where a limited number of agents use simple and independent machine-learning algorithms to buy and sell a storable good. No specific instruction is given to them, only that their objective is to maximize profits based solely on past market prices and payoffs. With an original application to battery operations, we observe that the algorithms learn quickly to reach seemingly collusive decisions, despite the absence of any formal communication between them. Building on the findings of the existing literature on algorithmic collusion, we show that seeming collusion could originate in imperfect exploration rather than excessive algorithmic sophistication. We then show that a regulator may succeed in disciplining the market to produce socially desirable outcomes by enforcing decentralized learning or with adequate intervention during the learning process.

[Please read the research paper here]

Research list
arrow-right