Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
72 tokens/sec
GPT-4o
61 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sociotechnical Implications of Generative Artificial Intelligence for Information Access (2405.11612v2)

Published 19 May 2024 in cs.IR and cs.AI

Abstract: Robust access to trustworthy information is a critical need for society with implications for knowledge production, public health education, and promoting informed citizenry in democratic societies. Generative AI technologies may enable new ways to access information and improve effectiveness of existing information retrieval systems but we are only starting to understand and grapple with their long-term social implications. In this chapter, we present an overview of some of the systemic consequences and risks of employing generative AI in the context of information access. We also provide recommendations for evaluation and mitigation, and discuss challenges for future research.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (240)
  1. Artificial intelligence (ai) standards. URL https://www.iso.org/sectors/it-technologies/ai.
  2. URL https://query.prod.cms.rt.microsoft.com/cms/api/am/binary/RE5cmFl.
  3. URL https://www.ap.org/media-center/press-releases/2023/ap-open-ai-agree-to-share-select-news-content-and-technology-in-new-collaboration/.
  4. Ai used to target kids with disinformation, 2023. URL https://www.bbc.co.uk/newsround/66796495.
  5. Supporting open source and open science in the eu ai act, 2023. URL https://huggingface.co/blog/assets/eu_ai_act_oss/supporting_OS_in_the_AIAct.pdf.
  6. URL https://investor.shutterstock.com/news-releases/news-release-details/shutterstock-expands-partnership-openai-signs-new-six-year.
  7. URL https://foe.org/news/ai-threat-report/.
  8. URL https://www.nist.gov/artificial-intelligence/ai-standards.
  9. URL https://sustainablebrands.com/read/product-service-design-innovation/ai-fueling-climate-change-energy-disinformation.
  10. World economic forum global risks report 2024, 2024. URL https://www.weforum.org/publications/global-risks-report-2024/.
  11. Persistent anti-muslim bias in large language models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, pages 298–306, 2021.
  12. Physiognomy’s new clothes. 2017. URL https://medium.com/@blaisea/physiognomys-new-clothes-f2d4b59fdd6a.
  13. N. Al-Sibai. The top google image for israel kamakawiwo’ole is ai-generated. Futurism, 2023.
  14. Self-consuming generative models go mad. arXiv preprint arXiv:2307.01850, 2023.
  15. M. Altenried. The platform as factory: Crowdwork and the hidden labour behind artificial intelligence. Capital & Class, 44(2):145–158, 2020.
  16. D. ANGUIANO and L. BECKETT. How hollywood writers triumphed over ai–and why it matters. The Guardian, October, 2023.
  17. Generative ai has an intellectual property problem. Harvard Business Review, 7, 2023.
  18. Should we treat data as labor? moving beyond “free”. In aea Papers and Proceedings, volume 108, pages 38–42. American Economic Association 2014 Broadway, Suite 305, Nashville, TN 37203, 2018.
  19. A. Baio. ’most disturbing ai site on internet’ can find every picture of you that exists. Indy100, 2024.
  20. L. Belkhir and A. Elmeligi. Assessing ict global emissions footprint: Trends to 2040 & recommendations. Journal of cleaner production, 177:448–463, 2018.
  21. B. bench authors. Beyond the imitation game: Quantifying and extrapolating the capabilities of language models. Transactions on Machine Learning Research, 2023. ISSN 2835-8856. URL https://openreview.net/forum?id=uyTL5Bvosj.
  22. On the dangers of stochastic parrots: Can language models be too big?. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, 2021.
  23. D. Berreby. As use of a.i. soars, so does the energy and water it requires. Yale Environment 360, 2024.
  24. A. Birhane. Algorithmic colonization of africa. SCRIPTed, 17:389, 2020.
  25. A. Birhane and F. Cummins. Algorithmic injustices: Towards a relational ethics. arXiv preprint arXiv:1912.07376, 2019.
  26. The values encoded in machine learning research. In 2022 ACM Conference on Fairness, Accountability, and Transparency, pages 173–184, 2022.
  27. Demographic dialectal variation in social media: A case study of african-american english. arXiv preprint arXiv:1608.08868, 2016.
  28. Language (technology) is power: A critical survey of" bias" in nlp. arXiv preprint arXiv:2005.14050, 2020.
  29. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Advances in neural information processing systems, 29, 2016.
  30. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258, 2021.
  31. Funding the next generation of content farms: Some of the world’s largest blue chip brands unintentionally support the spread of unreliable ai-generated news websites. NewsGuard, 2023a.
  32. Funding the next generation of content farms: Some of the world’s largest blue chip brands unintentionally support the spread of unreliable ai-generated news websites. NewsGuard, 2023b.
  33. K. Burke. ’biggest act of copyright theft in history’: thousands of australian books allegedly used to train ai model. The Guardian, 2023.
  34. K. Burke. Generative ai is a marvel. is it also built on theft? The Economist, 2024.
  35. M. Burtell and T. Woodside. Artificial influence: An analysis of ai-driven persuasion. arXiv preprint arXiv:2303.08721, 2023.
  36. Semantics derived automatically from language corpora contain human-like biases. Science, 356(6334):183–186, 2017.
  37. T. Cantrell. The true cost of ai innovation. Scientific Computing World.
  38. Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21), pages 2633–2650, 2021.
  39. Characterizing manipulation from ai systems. In Proceedings of the 3rd ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization, pages 1–13, 2023.
  40. K. Chayka. Is a.i. art stealing from artists? The New Yorker, 2023.
  41. S. Chesterman. Good models borrow, great models steal: intellectual property rights and generative ai. Policy and Society, page puae006, 2024.
  42. Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30, 2017.
  43. D. Coffey. Māori are trying to save their language from big tech. Wired UK, 2021.
  44. W. D. Cohan. Ai is learning from stolen intellectual property. it needs to stop. The Washington Post, 2023.
  45. D. Coldewey. Thousands of authors sign letter urging ai makers to stop stealing books. TechCrunch, 2023.
  46. J. Coleman. Ai’s climate impact goes beyond its emissions. Scientific American, 2023.
  47. J. Corbett. Report warns generative ai could turbocharge climate disinformation. Common Dreams, 2024.
  48. A. M. R. Correia. Information literacy for an active and effective citizenship. In White Paper prepared for UNESCO, the US National Commission on Libraries and Information Science, and the National Forum on Information Literacy, for use at the Information Literacy Meeting of Experts, Prague, The Czech Republic, 2002.
  49. N. Couldry and U. A. Mejias. Data colonialism: Rethinking big data’s relation to the contemporary subject. Television & New Media, 20(4):336–349, 2019.
  50. J. Cox. Google news is boosting garbage ai-generated articles. 404 Media, 2024.
  51. J. Coyle. In hollywood writers’ battle against ai, humans win (for now), 2023.
  52. H. Cramer. Practical routes in the ux of ai, or sharing more beaten paths. Interactions, 29(5):89–91, aug 2022. ISSN 1072-5520. doi: 10.1145/3555834. URL https://doi.org/10.1145/3555834.
  53. K. Crawford. The trouble with bias. In Conference on Neural Information Processing Systems, invited speaker, 2017.
  54. C. Criddle and K. Bryan. Ai boom sparks concern over big tech’s water consumption. The Conversation, 2024.
  55. The power of noise: Redefining retrieval for rag systems. arXiv preprint arXiv:2401.14887, 2024.
  56. Are large language models a threat to digital public goods? evidence from activity on stack overflow. arXiv preprint arXiv:2307.07367, 2023.
  57. S. Dhawan. Universities leveraging ai detectors: International students fear they may be wrongly accused of cheating. Financial Express, 2023.
  58. B. Dickson. What is machine learning data poisoning? The Verge, 2020.
  59. C. A. A. Disinformation. Artificial intelligence threats to climate change. 2024.
  60. Measuring the carbon intensity of ai in cloud instances. In Proceedings of the 2022 ACM conference on fairness, accountability, and transparency, pages 1877–1894, 2022.
  61. M. H. Dupré’. Sports illustrated published articles by fake, ai-generated writers. Futurism, 2023a.
  62. M. H. Dupré’. Top google result for "edward hopper" an ai-generated fake. Futurism, 2023b.
  63. G. Duran. The tech baron seeking to “ethnically cleanse” san francisco. The New Republic, 2024.
  64. Not just algorithms: Strategically addressing consumer impacts in information retrieval. In Advances in Information Retrieval: 46th European Conference on Information Retrieval, ECIR 2024, Glasgow, UK, March 24–28, 2024, Proceedings, Part IV, page 314–335, Berlin, Heidelberg, 2024. Springer-Verlag. ISBN 978-3-031-56065-1. doi: 10.1007/978-3-031-56066-8_25. URL https://doi.org/10.1007/978-3-031-56066-8_25.
  65. A mechanism-based approach to mitigating harms from persuasive generative ai. arXiv preprint arXiv:2404.15058, 2024.
  66. M. C. Elish. Moral crumple zones: Cautionary tales in human-robot interaction (pre-print). Engaging Science, Technology, and Society (pre-print), 2019.
  67. E. Ferrara. Should chatgpt be biased? challenges and risks of bias in large language models. arXiv preprint arXiv:2304.03738, 2023.
  68. Cdc museum covid-19 timeline. 2022, 2022.
  69. J. A. Fowler. Turbotax and h&r block now use ai for tax advice. it’s awful. The Washington Post, 2024. URL https://www.washingtonpost.com/technology/2024/03/04/ai-taxes-turbotax-hrblock-chatbot/.
  70. I. Gabriel. Artificial intelligence, values, and alignment. Minds and machines, 30(3):411–437, 2020.
  71. I. Gabriel and V. Ghazavi. The challenge of value alignment: From fairer algorithms to ai safety. arXiv preprint arXiv:2101.06060, 2021.
  72. Ai could create a perfect storm of climate misinformation. arXiv preprint arXiv:2306.12807, 2023.
  73. Red teaming language models to reduce harms: Methods, scaling behaviors, and lessons learned, 2022.
  74. A framework for exploring the consequences of ai-mediated enterprise knowledge access and identifying risks to workers. In Proceedings of the 2024 ACM conference on fairness, accountability, and transparency, 2024.
  75. T. Gebru. Effective altruism is pushing a dangerous brand of ‘ai safety’, 2022.
  76. T. Gebru and É. P. Torres. Eugenics and the promise of utopia through artificial general intelligence. First Monday, 2023.
  77. Statement from the listed authors of stochastic parrots on the “ai pause” letter, 2023. URL https://www.dair-institute.org/blog/letter-statement-March2023/.
  78. Realtoxicityprompts: Evaluating neural toxic degeneration in language models. arXiv preprint arXiv:2009.11462, 2020.
  79. D. Gershgorn. Github’s automatic coding tool rests on untested legal ground. The Verge, 2021.
  80. J. Gertner. Wikipedia’s moment of truth. The New York Times Magazine, 2023.
  81. T. Gillespie. Custodians of the Internet: Platforms, content moderation, and the hidden decisions that shape social media. Yale University Press, 2018.
  82. S. Goldstein. Informed Societies. facet publishing, 2020.
  83. H. Gonen and Y. Goldberg. Lipstick on a pig: Debiasing methods cover up systematic gender biases in word embeddings but do not remove them. In Proc. NAACL, pages 609–614, 2019.
  84. M. González. A better-informed society is a freer society. 2021. URL https://www.unesco.org/en/articles/better-informed-society-freer-society.
  85. Co-audit: tools to help humans double-check ai-generated content. arXiv preprint arXiv:2310.01297, 2023.
  86. C. Gordon. Ai is accelerating the loss of our scarcest natural resource: Water. Forbes, 2024.
  87. Not what you’ve signed up for: Compromising real-world llm-integrated applications with indirect prompt injection. In Proceedings of the 16th ACM Workshop on Artificial Intelligence and Security, pages 79–90, 2023.
  88. F. Guerrini. Ai’s unsustainable water use: How tech giants contribute to global water shortages. Forbes, 2023.
  89. Legalbench: A collaboratively built benchmark for measuring legal reasoning in large language models, 2023.
  90. Ai’s excessive water consumption threatens to drown out its environmental contributions. The Conversation, 2024.
  91. E. Halper. Amid explosive demand, america is running out of power. The Washington Post, 2024.
  92. K. Hao. Artificial intelligence is creating a new colonial world order. MIT Technology Review, 2022.
  93. K. Hao. Ai is taking water from the desert. The Atlantic, 2024.
  94. K. Hao and A. P. Hernández. How the ai industry profits from catastrophe. MIT Technology Review, 2022.
  95. K. Hao and D. Seetharaman. Cleaning up chatgpt takes heavy toll on human workers. The Wall Street Journal, 24, 2023.
  96. G. Hardin. The tragedy of the commons. In Classic Papers in Natural Resource Economics Revisited, pages 145–156. Routledge, 2018.
  97. K. Hays and A. Barr. Ai is killing the grand bargain at the heart of the web. ’we’re in a different world.’. Business Insider, 2024.
  98. M. Heikkila. This new data poisoning tool lets artists fight back against generative ai. MIT Technology Review, 2023.
  99. A. Hern. Techscape: How cheap, outsourced labour in africa is shaping ai english, 2024. URL https://www.theguardian.com/technology/2024/apr/16/techscape-ai-gadgest-humane-ai-pin-chatgpt.
  100. S. Higgins and L. Gregory. Information literacy and social justice: Radical professional praxis. Library Juice Press, 2013.
  101. E. Hoel. Here lies the internet, murdered by generative ai. The Intrinsic Perspective, 2024a. URL https://www.theintrinsicperspective.com/p/here-lies-the-internet-murdered-by.
  102. E. Hoel. A.i.-generated garbage is polluting our culture. The New York Times, 2024b.
  103. Online evaluation for information retrieval. Found. Trends Inf. Retr., 10(1):1–117, jun 2016a. ISSN 1554-0669. doi: 10.1561/1500000051. URL https://doi.org/10.1561/1500000051.
  104. Online evaluation for information retrieval. Foundations and Trends® in Information Retrieval, 10(1):1–117, 2016b. ISSN 1554-0669. doi: 10.1561/1500000051. URL http://dx.doi.org/10.1561/1500000051.
  105. K.-l. Hom. Rage baiting. Westside Seattle, 2015.
  106. C. C. IPCC et al. The physical science basis, the working group i contribution to the un ipcc’s fifth assessment report (wg1 ar5), 2013.
  107. What disease does this patient have? a large-scale open domain question answering dataset from medical exams. Applied Sciences, 11(14), 2021. ISSN 2076-3417. doi: 10.3390/app11146421. URL https://www.mdpi.com/2076-3417/11/14/6421.
  108. The global landscape of ai ethics guidelines. Nature machine intelligence, 1(9):389–399, 2019.
  109. A. Kak and S. M. West. Ai now 2023 landscape: Confronting tech power, 2023. URL https://ainowinstitute.org/2023-landscape.
  110. P. Kalluri et al. Don’t ask if artificial intelligence is good or fair, ask how it shifts power. Nature, 583(7815):169–169, 2020.
  111. A. Kanungo. The green dilemma: Can ai fulfil its potential without harming the environment? Earth.Org, 2023.
  112. On the societal impact of open foundation models. 2024.
  113. A. Kasirzadeh and I. Gabriel. In conversation with artificial intelligence: aligning language models with human values. Philosophy & Technology, 36(2):1–24, 2023.
  114. Why and how is the power of big tech increasing in the policy process? the case of generative ai. Policy and Society, page puae012, 2024.
  115. R. Khattak. The environmental impact of e-waste. Earth.Org, 2023.
  116. H. Khlaaf. Toward comprehensive risk assessments and assurance of ai-based systems. Trail of Bits, 2023.
  117. N. Klein. Ai machines aren’t ‘hallucinating’. but their makers are. The Guardian, 2023.
  118. K. Knibbs. Scammy ai-generated book rewrites are flooding amazon. Wired, 2024a. URL https://www.wired.com/story/scammy-ai-generated-books-flooding-amazon/.
  119. K. Knibbs. Your kid may already be watching ai-generated videos on youtube. Wired, 2024b. URL https://www.wired.com/story/your-kid-may-be-watching-ai-generated-videos-on-youtube/.
  120. Private traits and attributes are predictable from digital records of human behavior. Proceedings of the national academy of sciences, 110(15):5802–5805, 2013.
  121. Gender bias and stereotypes in large language models. In Proceedings of The ACM Collective Intelligence Conference, pages 12–24, 2023.
  122. S. Kugel and S. Hiltner. A new frontier for travel scammers: A.i.-generated guidebooks. The New York Times, 2023. URL https://www.nytimes.com/2023/08/05/travel/amazon-guidebooks-artificial-intelligence.html.
  123. A. LaFrance. The rise of techno-authoritarianism. The Atlantic, 2024.
  124. F. Landymore. Sports illustrated lays off journalists after announcing pivot to ai content. Futurism, 2023. URL https://futurism.com/the-byte/sports-illustrated-lays-off-journalists-ai-content.
  125. How do people change their technology use in protest? understanding. Proceedings of the ACM on Human-Computer Interaction, 3(CSCW):1–22, 2019.
  126. The dimensions of data labor: A road map for researchers, activists, and policymakers to empower data producers. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pages 1151–1161, 2023a.
  127. Making ai less" thirsty": Uncovering and addressing the secret water footprint of ai models. arXiv preprint arXiv:2304.03271, 2023b.
  128. Adversarial example does good: Preventing painting imitation from diffusion models via adversarial examples. arXiv preprint arXiv:2302.04578, 2023a.
  129. Holistic evaluation of language models, 2023b.
  130. Monitoring ai-modified content at scale: A case study on the impact of chatgpt on ai conference peer reviews. arXiv preprint arXiv:2403.07183, 2024.
  131. A. Limbong. Authors push back on the growing number of ai ’scam’ books on amazon. National Public Radio, 2024. URL https://www.npr.org/2024/03/13/1237888126/growing-number-ai-scam-books-amazon.
  132. Automatic and universal prompt injection attacks against large language models. arXiv preprint arXiv:2403.04957, 2024.
  133. Prompt injection attack against llm-integrated applications. arXiv preprint arXiv:2306.05499, 2023.
  134. R. Lonergan. Mr. justice brandeis, great american. Mr. Justice Brandeis, Great American: Press Opinion and Public Appraisal, 1941.
  135. B. Marr. Is generative ai stealing from artists? Forbes, 2023.
  136. Towards understanding the interplay of generative artificial intelligence and the internet. arXiv preprint arXiv:2306.06130, 2023.
  137. T. Mathewson. Ai detection tools falsely accuse international students of cheating. The Markup, 2023.
  138. The substantial interdependence of wikipedia and google: A case study on the relationship between peer production communities and information technologies. In Proceedings of the International AAAI Conference on Web and Social Media, volume 11, pages 142–151, 2017.
  139. Y. Mehdi. Bringing the full power of copilot to more people and businesses, 2024. URL https://blogs.microsoft.com/blog/2024/01/15/bringing-the-full-power-of-copilot-to-more-people-and-businesses/.
  140. Rethinking search: making domain experts out of dilettantes. In Acm sigir forum, volume 55, pages 1–27. ACM New York, NY, USA, 2021.
  141. Studying up machine learning data: Why talk about bias when we mean power? Proceedings of the ACM on Human-Computer Interaction, 6(GROUP):1–14, 2022.
  142. Silo language models: Isolating legal risk in a nonparametric datastore. arXiv preprint arXiv:2308.04430, 2023.
  143. B. Mitra. Search and society: Reimagining information access for radical futures. arXiv preprint arXiv:2403.17901, 2024.
  144. M. Möhlmann. Algorithmic nudges don’t have to be unethical. Harvard Business Review, 22, 2021.
  145. Access to health information on the internet: a public health issue? Revista da Associação Médica Brasileira, 58:650–658, 2012.
  146. K. Mularczyk. Row over deepfake of polish pm in opposition-party broadcast. Brussels Signal, 2023. URL https://brusselssignal.eu/2023/08/row-over-deepfake-of-polish-pm-in-opposition-party-broadcast/.
  147. J. Muldoon and B. A. Wu. Artificial intelligence in the colonial matrix of power. Philosophy & Technology, 36(4):80, 2023.
  148. M. Murgia. Ai academics under pressure to do commercial research. Financial Times, 13, 2019.
  149. S. M. Mutula. Digital divide and economic development: Case study of sub-saharan africa. The Electronic Library, 26(4):468–489, 2008.
  150. J. Naughton. Ai’s craving for data is matched only by a runaway thirst for water and energy. The Guardian, 2024.
  151. Biases in large language models: origins, inventory, and discussion. ACM Journal of Data and Information Quality, 15(2):1–21, 2023.
  152. W. Oremus. He wrote a book on a rare subject. then a chatgpt replica appeared on amazon. The Washington Post, 2023. URL https://www.washingtonpost.com/technology/2023/05/05/ai-spam-websites-books-chatgpt/.
  153. K. Orland. Lazy use of ai leads to amazon products called “i cannot fulfill that request”. Ars Technica, 2024. URL https://arstechnica.com/ai/2024/01/lazy-use-of-ai-leads-to-amazon-products-called-i-cannot-fulfill-that-request/.
  154. M. O’Gorman. At the heart of artificial intelligence is racism and colonialism that we must excise. The Globe and Mail web edition, pages NA–NA, 2023.
  155. On the risk of misinformation pollution with large language models. arXiv preprint arXiv:2305.13661, 2023.
  156. Automatic personality assessment through social media language. Journal of personality and social psychology, 108(6):934, 2015.
  157. Ai deception: A survey of examples, risks, and potential solutions. arXiv preprint arXiv:2308.14752, 2023.
  158. Climate change 2022: Impacts, adaptation and vulnerability, 2022.
  159. Carbon emissions and large neural network training. 2021.
  160. The carbon footprint of machine learning training will plateau, then shrink. Computer, 55(7):18–28, 2022.
  161. J. Pearson. Scientific journal publishes ai-generated rat with gigantic penis in worrying incident. Vice, 2024.
  162. Red teaming language models with language models, 2022.
  163. B. Perrigo. Exclusive: Openai used kenyan workers on less than $2 per hour to make chatgpt less toxic. Last accessed, 19, 2023. URL https://time.com/6247678/openai-chatgpt-kenya-workers/.
  164. D. Pierce. You can now use the dall-e 3 ai image generator inside bing chat. The Verge, 2023.
  165. M.-T. Png. At the tensions of south and north: Critical roles of global south stakeholders in ai governance. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, pages 1434–1445, 2022.
  166. G. Polizzi. Information literacy in the digital age: Why critical digital literacy matters for democracy. Informed Societies: Why information literacy matters for citizenship, participation and democracy, pages 1–23, 2020.
  167. M. Poynting and E. Rivault. 2023 confirmed as world’s hottest year on record, 2024.
  168. A human rights-based approach to responsible ai. arXiv preprint arXiv:2210.02667, 2022.
  169. Our twitter profiles, our selves: Predicting personality with twitter. In 2011 IEEE third international conference on privacy, security, risk and trust and 2011 IEEE third international conference on social computing, pages 180–185. IEEE, 2011.
  170. S. Ren. How much water does ai consume? the public deserves to know it. 2023.
  171. N. Robins-Early. New gpt-4o ai model is faster and free for all users, openai announces. The Guardian, 2024.
  172. Research priorities for robust and beneficial artificial intelligence. AI magazine, 36(4):105–114, 2015.
  173. Ares: An automated evaluation framework for retrieval-augmented generation systems, 2024.
  174. M. Sadeghi and L. Arvanitis. Rise of the newsbots: Ai-generated news websites proliferating online. NewsGuard, 2023.
  175. Re-imagining algorithmic fairness in india and beyond. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, pages 315–328, 2021.
  176. Reduce, reuse, recycle: Green information retrieval research. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2825–2837, 2022.
  177. N. Scientist. Covid-19: the story of a pandemic. New Scientist, 10, 2021.
  178. C. Shah and E. M. Bender. Situating search. In Proceedings of the 2022 Conference on Human Information Interaction and Retrieval, pages 221–232, 2022.
  179. Glaze: Protecting artists from style mimicry by {{\{{Text-to-Image}}\}} models. In 32nd USENIX Security Symposium (USENIX Security 23), pages 2187–2204, 2023a.
  180. Prompt-specific poisoning attacks on text-to-image generative models. arXiv preprint arXiv:2310.13828, 2023b.
  181. Who validates the validators? aligning llm-assisted evaluation of llm outputs with human preferences, 2024.
  182. R. Shrivastava. Openai and microsoft sued by nonfiction writers for alleged ‘rampant theft’ of authors’ works. Forbes, 2023.
  183. The curse of recursion: Training on generated data makes models forget. arXiv preprint arXiv:2305.17493, 2023.
  184. The persuasive effects of political microtargeting in the age of generative artificial intelligence. PNAS Nexus, 3(2):pgae035, 01 2024. ISSN 2752-6542. doi: 10.1093/pnasnexus/pgae035. URL https://doi.org/10.1093/pnasnexus/pgae035.
  185. J. J. Smith and L. Beattie. Recsys fairness metrics: Many to use but which one to choose? arXiv preprint arXiv:2209.04011, 2022.
  186. Scoping fairness objectives and identifying fairness metrics for recommender systems: The practitioners’ perspective. In Proceedings of the ACM Web Conference 2023, WWW ’23, page 3648–3659, New York, NY, USA, 2023. Association for Computing Machinery. ISBN 9781450394161. doi: 10.1145/3543507.3583204. URL https://doi.org/10.1145/3543507.3583204.
  187. I. Solaiman. The gradient of generative ai release: Methods and considerations. In Proceedings of the 2023 ACM conference on fairness, accountability, and transparency, pages 111–122, 2023.
  188. R. Speare-Cole. Generative ai could ‘supercharge’ climate disinformation, report warns. Independent, 2024.
  189. Beyond the imitation game: Quantifying and extrapolating the capabilities of language models, 2023.
  190. B. C. Stahl and D. Eke. The ethics of chatgpt–exploring the ethical issues of an emerging technology. International Journal of Information Management, 74:102700, 2024.
  191. Stanford. Stanford helm, 2024. URL https://crfm.stanford.edu/helm/.
  192. Energy and policy considerations for deep learning in nlp. arXiv preprint arXiv:1906.02243, 2019.
  193. H. Suresh and J. Guttag. A framework for understanding sources of harm throughout the machine learning life cycle. In Proceedings of the 1st ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization, EAAMO ’21, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 9781450385534. doi: 10.1145/3465416.3483305. URL https://doi.org/10.1145/3465416.3483305.
  194. J. Tacheva and S. Ramasubramanian. Ai empire: Unraveling the interlocking systems of oppression in generative ai’s global order. Big Data & Society, 10(2):20539517231219241, 2023.
  195. Understanding the capabilities, limitations, and societal impact of large language models. arXiv preprint arXiv:2102.02503, 2021.
  196. R. Tan and R. Cabato. Behind the ai boom, an army of overseas workers in’digital sweatshops’. The Washington Post, pages NA–NA, 2023.
  197. M. Tani. New york times drops out of ai coalition. Semafor, 2023.
  198. D. Taraborelli. The sum of all human knowledge in the age of machines: a new research agenda for wikimedia. In ICWSM-15 Workshop on Wikipedia, 2015.
  199. A. Tarkowski. How wikipedia can shape the future of ai. Open Future, 2023a.
  200. A. Tarkowski. Stewarding the sum of all knowledge in the age of ai. Open Future, 2023b.
  201. A. Taylor. A historic rise in global conflict deaths suggests a violent new era. 2023. URL https://www.washingtonpost.com/world/2023/06/29/conflict-war-deaths-global-peace-rise-casualty/.
  202. L. Taylor. Covid-19: True global death toll from pandemic is almost 15 million, says who. BMJ: British Medical Journal (Online), 377:o1144, 2022.
  203. Large language models can accurately predict searcher preferences. arXiv preprint arXiv:2309.10621, 2023.
  204. A shocking amount of the web is machine translated: Insights from multi-way parallelism. arXiv preprint arXiv:2401.05749, 2024.
  205. United Nations Meetings Coverage and Press Releases. With highest number of violent conflicts since second world war, united nations must rethink efforts to achieve, sustain peace, speakers tell security council. https://press.un.org/en/2023/sc15184.doc.htm, 2023.
  206. A. Urman and M. Makhortykh. The silence of the llms: Cross-lingual analysis of political bias and false information prevalence in chatgpt, google bard, and bing chat. 2023.
  207. S. Varghese. How a google search could end up endangering a life. iTWire, 2021.
  208. Learn to code sustainably: An empirical study on llm-based green code generation, 2024.
  209. J. Vincent. The invention of ai ‘gaydar’could be the start of something much worse. The Verge, 21, 2017.
  210. J. Vincent. The lawsuit that could rewrite the rules of ai copyright. The Verge, 22, 2022a.
  211. J. Vincent. Shutterstock will start selling ai-generated stock imagery with help from openai. The Verge, 25, 2022b.
  212. J. Vincent. Ai art tools stable diffusion and midjourney targeted with copyright lawsuit. The Verge, 2023a.
  213. N. Vincent. Don’t give openai all the credit for gpt-3: You might have helped create the latest “astonishing” advance in ai too, 2020. URL https://www.psagroup.org/blogposts/62.
  214. N. Vincent. The wga strike is a canary in the coal mine for ai labor concerns, 2023b. URL https://dataleverage.substack.com/p/the-wga-strike-is-a-canary-in-the.
  215. N. Vincent and H. Li. Github copilot and the exploitation of “data labor”: A wake-up call for the tech industry, 2021. URL https://www.psagroup.org/blogposts/62.
  216. N. Vincent and H. Li. Chatgpt stole your work. so what are you going to do?, 2023.
  217. Examining wikipedia with a broader lens: Quantifying the value of wikipedia’s relationships with other large-scale online communities. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, pages 1–13, 2018.
  218. “data strikes”: evaluating the effectiveness of a new form of collective action against technology companies. In The World Wide Web Conference, pages 1931–1943, 2019a.
  219. Mapping the potential and pitfalls of" data dividends" as a means of sharing the profits of artificial intelligence. arXiv preprint arXiv:1912.00757, 2019b.
  220. Data leverage: A framework for empowering the public in its relationship with technology companies. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pages 215–227, 2021.
  221. B. Violino. Ai tools such as chatgpt are generating a mammoth increase in malicious phishing emails. CNBC, 2023. URL https://www.cnbc.com/2023/11/28/ai-like-chatgpt-is-creating-huge-increase-in-malicious-phishing-email.html.
  222. Decodingtrust: A comprehensive assessment of trustworthiness in gpt models, 2024.
  223. T. Warren. Microsoft’s new copilot pro brings ai-powered office features to the rest of us, 2024. URL https://www.theverge.com/2024/1/15/24038711/microsoft-copilot-pro-office-ai-apps.
  224. Long-form factuality in large language models, 2024.
  225. Ethical and social risks of harm from language models. arXiv preprint arXiv:2112.04359, 2021.
  226. Taxonomy of risks posed by language models. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, pages 214–229, 2022.
  227. Challenges in detoxifying language models. arXiv preprint arXiv:2109.07445, 2021.
  228. M. Whittaker. The steep cost of capture. Interactions, 28(6):50–55, nov 2021a. ISSN 1072-5520. doi: 10.1145/3488666. URL https://doi.org/10.1145/3488666.
  229. M. Whittaker. The steep cost of capture. Interactions, 28(6):50–55, 2021b.
  230. The exploited labor behind artificial intelligence. Noema Magazine, 13, 2022. URL https://www.noemamag.com/the-exploited-labor-behind-artificial-intelligence/.
  231. R. Williams. Humans may be more likely to believe disinformation generated by ai. MIT Technology Review, 2023. URL https://www.technologyreview.com/2023/06/28/1075683/humans-may-be-more-likely-to-believe-disinformation-generated-by-ai/.
  232. Sustainable AI: Environmental implications, challenges and opportunities. Proceedings of Machine Learning and Systems, 4:795–813, 2022.
  233. C. Xiang. Openai used kenyan workers making $2 an hour to filter traumatic content from chatgpt. VICE, 2023.
  234. Promoting civil discourse through search engine diversity. Social Science Computer Review, 32(2):145–154, 2014.
  235. Computer-based personality judgments are more accurate than those made by humans. Proceedings of the National Academy of Sciences, 112(4):1036–1040, 2015.
  236. P. K. Yu. Bridging the digital divide: Equality in the information age. Cardozo Arts & Ent. LJ, 20:1, 2002.
  237. Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593, 2019.
  238. S. Zuboff. The age of surveillance capitalism. In Social Theory Re-Wired, pages 203–213. Routledge, 2023.
  239. Surveillance capitalism: an interview with shoshana zuboff. Surveillance & Society, 17(1/2):257–266, 2019.
  240. Beyond co2 emissions: The overlooked impact of water consumption of information retrieval models. In Proceedings of the 2023 ACM SIGIR International Conference on Theory of Information Retrieval, pages 283–289, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Bhaskar Mitra (78 papers)
  2. Henriette Cramer (6 papers)
  3. Olya Gurevich (1 paper)
Citations (2)
Youtube Logo Streamline Icon: https://streamlinehq.com