Privacy Can Arise Endogenously in an Economic System with Learning Agents (2404.10767v1)
Abstract: We study price-discrimination games between buyers and a seller where privacy arises endogenously--that is, utility maximization yields equilibrium strategies where privacy occurs naturally. In this game, buyers with a high valuation for a good have an incentive to keep their valuation private, lest the seller charge them a higher price. This yields an equilibrium where some buyers will send a signal that misrepresents their type with some probability; we refer to this as buyer-induced privacy. When the seller is able to publicly commit to providing a certain privacy level, we find that their equilibrium response is to commit to ignore buyers' signals with some positive probability; we refer to this as seller-induced privacy. We then turn our attention to a repeated interaction setting where the game parameters are unknown and the seller cannot credibly commit to a level of seller-induced privacy. In this setting, players must learn strategies based on information revealed in past rounds. We find that, even without commitment ability, seller-induced privacy arises as a result of reputation building. We characterize the resulting seller-induced privacy and seller's utility under no-regret and no-policy-regret learning algorithms and verify these results through simulations.
- Conditioning prices on purchase history. Marketing Science, 2004.
- The economics of privacy. Journal of Economic Literature, 54(2):442–492, 2016.
- Policy regret in repeated games. Advances in Neural Information Processing Systems, 2020.
- Mechanisms for a no-regret agent: Beyond the common prior. 2020 IEEE 61st Annual Symposium on Foundations of Computer Science (FOCS), 2020.
- Hide and seek: Costly consumer privacy in a market with repeated purchases. Marketing Science, 31(2):277–292, 2012.
- Strategizing against no-regret learners. Advances in Neural Information Processing Systems, 32, 2019.
- Gaussian differential privacy. arXiv preprint arXiv:1905.02383, 2019.
- The algorithmic foundations of differential privacy. Foundations and Trends in Theoretical Computer Science, 9(3–4):211–407, 2014.
- Bad reputation. The Quarterly Journal of Economics, 118(3):785–814, 2003.
- Calibrated learning and correlated equilibrium. Games and Economic Behavior, 21(1-2):40, 1997.
- Behavior-based price discrimination and customer recognition. Handbook on Economics and Information Systems, 1:377–436, 2006.
- Calibrated Stackelberg games: Learning optimal commitments against calibrated agents. Advances in Neural Information Processing Systems, 36, 2024.
- Contract renegotiation and coasian dynamics. The Review of Economic Studies, 55(4):509–540, 1988.
- Johannes Horner. Reputation and competition. American Economic Review, 92(3):644–663, 2002.
- Shota Ichihashi. Online privacy and information disclosure by consumers. American Economic Review, 110(2):569–595, 2020.
- Bandit Algorithms. Cambridge University Press, 2020.
- We need to focus on how our data is used, not just how it is shared. Communications of the ACM, 66(9):32–34, 2023.
- Ilya Mironov. Rényi differential privacy. In 2017 IEEE 30th Computer Security Foundations Symposium (CSF), pages 263–275. IEEE, 2017.
- The value of personal information in markets with endogenous privacy. Center for Economic and International Studies, 13(352), 2015.
- Bridging the gap between computer science and legal approaches to privacy. Harvard Journal of Law & Technology, 31:687, 2017.
- Carl Shapiro. Premiums for high quality products are returns to reputation. Quarterly Journal of Economics, 98(4):659–679, 1983.
- Alicia Solow-Niederman. Information privacy and the inference economy. Northwestern University Law Review, 117:357, 2022.
- Privacy loss in Apple’s implementation of differential privacy on macOS 10.12. arXiv preprint arXiv:1709.02753, 2017.