The Role of Learning Algorithms in Collective Action (2405.06582v3)
Abstract: Collective action in machine learning is the study of the control that a coordinated group can have over machine learning algorithms. While previous research has concentrated on assessing the impact of collectives against Bayes (sub-)optimal classifiers, this perspective is limited in that it does not account for the choice of learning algorithm. Since classifiers seldom behave like Bayes classifiers and are influenced by the choice of learning algorithms along with their inherent biases, in this work we initiate the study of how the choice of the learning algorithm plays a role in the success of a collective in practical settings. Specifically, we focus on distributionally robust optimization (DRO), popular for improving a worst group error, and on the ubiquitous stochastic gradient descent (SGD), due to its inductive bias for "simpler" functions. Our empirical results, supported by a theoretical foundation, show that the effective size and success of the collective are highly dependent on properties of the learning algorithm. This highlights the necessity of taking the learning algorithm into account when studying the impact of collective action in machine learning.
- A general class of coefficients of divergence of one distribution from another. Journal of the Royal Statistical Society: Series B (Methodological), 1966.
- Differential privacy has disparate impact on model accuracy. Advances in neural information processing systems, 2019.
- A convex framework for fair regression. arXiv:1706.02409, 2017.
- When users control the algorithms: Values expressed in practices on twitter. In Proceedings of the ACM on Human-Computer Interaction, 2019.
- Julie Yujie Chen. Thrown under the bus and outrunning it! the logic of didi and taxi drivers’ labour and activism in the on-demand economy. New Media & Society, 2018.
- Deep reinforcement learning from human preferences. Advances in neural information processing systems, 2017.
- Imre Csiszár. On information-type measure of difference of probability distributions and indirect observations. Studia Sci. Math. Hungar., 1967.
- Distributionally robust optimization under moment uncertainty with application to data-drivenproblems. Operations Research, 2010.
- Variance-based regularization with convex objectives. Journal of Machine Learning Research, 2019.
- Learning models with uniform performance via distributionally robust optimization. The Annals of Statistics, 2021.
- The like economy: Social buttons and the data-intensive web. New media & society, 2013.
- Algorithmic collective action in machine learning. In International Conference on Machine Learning, volume 2022, 2023.
- Fairness without demographics in repeated loss minimization. In Proceedings of the 35th International Conference on Machine Learning, Proceedings of Machine Learning Research, 2018.
- Benchmarking Neural Network Robustness to Common Corruptions and Perturbations. In International Conference on Learning Representations, 2019.
- The origins and prevalence of texture bias in convolutional neural networks. In Advances in Neural Information Processing Systems, 2020.
- Sgd on neural networks learns functions of increasing complexity. In Advances in Neural Information Processing Systems, 2019.
- Wilds: A benchmark of in-the-wild distribution shifts. In International Conference on Machine Learning, 2021.
- Alex Krizhevsky. Learning multiple layers of features from tiny images, 2009.
- The inductive bias of in-context learning: Rethinking pretraining example design. In International Conference on Learning Representations, 2022.
- Large-scale methods for distributionally robust optimization. In Advances in Neural Information Processing Systems, 2020.
- Just train twice: Improving group robustness without training group information. In Proceedings of the 38th International Conference on Machine Learning, 2021.
- Towards deep learning models resistant to adversarial attacks. In International Conference on Learning Representations, 2018.
- Maximin effects in inhomogeneous large-scale data. The Annals of Statistics, 2015.
- Learning from failure: De-biasing classifier from biased classifier. In Advances in Neural Information Processing Systems, 2020.
- Stochastic gradient methods for distributionally robust optimization with f-divergences. In Advances in Neural Information Processing Systems, 2016.
- Mancur Olson. The logic of collective action. Contemporary Sociological Theory, 1965.
- A law of adversarial risk, interpolation, and label noise. In International Conference on Learning Representations, 2023.
- Toward a better trade-off between performance and fairness with kernel-based distribution matching. arXiv:1910.11779, 2019.
- Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 2024.
- Hatim A. Rahman. The invisible cage: Workers’ reactivity to opaque algorithmic evaluations. Administrative Science Quarterly, 2021.
- Distributionally Robust Neural Networks. In International Conference on Learning Representations, 2020.
- How unfair is private learning? In Uncertainty in Artificial Intelligence, 2022.
- The pitfalls of simplicity bias in neural networks. In Advances in Neural Information Processing Systems, 2020.
- How robust is unsupervised representation learning to distribution shift? In The Eleventh International Conference on Learning Representations, 2023.
- Frapp\\\backslash\’e: A post-processing framework for group fairness regularization. In International Conference on Machine Learning, 2024.
- Examining the inductive bias of neural language models with artificial languages. In Annual Meeting of the Association for Computational Linguistics, 2021.
- Adversarial training for high-stakes reliability. Advances in Neural Information Processing Systems, 2022.