Data Sharing with Endogenous Choices over Differential Privacy Levels
This research reveals why AI companies struggle to share data—and how to fix it.
A new game theory paper analyzes the strategic tension when agents share data under differential privacy. Each participant chooses their own privacy level, creating a trade-off: more privacy reduces individual risk but degrades the overall dataset's utility for AI training. The study introduces a "robust equilibrium" concept to model stable data-sharing coalitions and analyzes efficiency losses compared to a social optimum, providing a framework for decentralized data markets crucial for AI development.
Why It Matters
This framework is essential for building functional, privacy-preserving data markets needed to train next-generation AI models.