2000 character limit reached
TASRA: a Taxonomy and Analysis of Societal-Scale Risks from AI (2306.06924v2)
Published 12 Jun 2023 in cs.AI, cs.CR, cs.CY, and cs.LG
Abstract: While several recent works have identified societal-scale and extinction-level risks to humanity arising from artificial intelligence, few have attempted an {\em exhaustive taxonomy} of such risks. Many exhaustive taxonomies are possible, and some are useful -- particularly if they reveal new risks or practical approaches to safety. This paper explores a taxonomy based on accountability: whose actions lead to the risk, are the actors unified, and are they deliberate? We also provide stories to illustrate how the various risk types could each play out, including risks arising from unanticipated interactions of many AI systems, as well as risks from deliberate misuse, for which combined technical and policy solutions are indicated.
- Concrete problems in AI safety. arXiv preprint arXiv:1606.06565.
- Low impact artificial intelligences. arXiv preprint arXiv:1705.10720. (link).
- Barak, B. (2002). Can we obfuscate programs. Retreived From http://www. math. ias/edu/boaz/Papers/obf—informal. html 7. (link).
- Barrat, J. (2013). Artificial intelligence and the end of the human era. New York: Thomas Dunne.
- Bengio, Y. (2023). How rogue AIs may arise. (link).
- Untitled video statement calling for articulation of concrete cases of harm and extinction. (link).
- The complexity of decentralized control of markov decision processes. Mathematics of operations research 27(4), 819–840. INFORMS. (link).
- Indistinguishability obfuscation from functional encryption. Journal of the ACM (JACM) 65(6), 1–37. ACM New York, NY, USA. (link).
- Bostrom, N. (2014). Superintelligence: Paths, Dangers, Strategies. Oxford, UK: Oxford University Press.
- Butler, S. (1863). Darwin among the machines. The Press (Christchurch, New Zealand) June 13.
- Capek, K. (1920). R.U.R. (Rossum’s Universal Robots). Aventinum.
- Carey, R. (2018). Incorrigibility in the cirl framework. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, pp. 30–35. (link).
- Carpenter, D. and D. A. Moss (2013). Preventing regulatory capture: Special interest influence and how to limit it. Cambridge University Press.
- Center for AI Safety (2023). Statement on AI risk. (link).
- Critch, A. (2017). Toward negotiable reinforcement learning: shifting priorities in pareto optimal sequential decision-making. arXiv preprint arXiv:1701.01302. (link).
- Servant of many masters: Shifting priorities in pareto-optimal sequential decision-making. arXiv preprint arXiv:1711.00363. (link).
- Cyberspace Administration of China (2023). Measures for the management of generative artificial intelligence services (draft for comment). (link).
- Dal Bó, E. (2006). Regulatory capture: A review. Oxford review of economic policy 22(2), 203–225. Oxford University Press.
- Dalrymple, D. A. (2022). An open agency architecture for safe transformative AI. AI Alignment Forum. (link).
- Negotiable reinforcement learning for pareto optimal sequential decision-making. In Advances in Neural Information Processing Systems, pp. 4712–4720. (link).
- Ed Yong (2013). Trees trap ants into sweet servitude. National Geographic. (link).
- European Commission (2021). Regulation of the european parliament and of the council: Laying down harmonised rules on artificial intelligence (artificial intelligence act) and amending certain union legislative acts. (link).
- Candidate indistinguishability obfuscation and functional encryption for all circuits. SIAM Journal on Computing 45(3), 882–929. SIAM. (link).
- Hiding secrets in software: A cryptographic approach to program obfuscation. Communications of the ACM 59(5), 113–120. ACM New York, NY, USA. (link).
- Tobacco industry’s elaborate attempts to control a global track and trace system and fundamentally undermine the illicit trade protocol. Tobacco Control 28(2), 127–140. BMJ Publishing Group Ltd. (link).
- Good, I. J. (1966). Speculations concerning the first ultraintelligent machine. In Advances in computers, Volume 6, pp. 31–88. Elsevier.
- Cooperative inverse reinforcement learning. In Advances in neural information processing systems, pp. 3909–3917. (link).
- Hibbard, B. (2001). Super-intelligent machines. ACM SIGGRAPH Computer Graphics 35(1), 11–13. ACM New York, NY, USA.
- Translation: Measures for the management of generative artificial intelligence services (draft for comment). digichina.stanford.edu. (link).
- Learning gentle object manipulation with curiosity-driven deep reinforcement learning. arXiv preprint arXiv:1903.08542. (link).
- Measuring and avoiding side effects using relative reachability. arXiv preprint arXiv:1806.01186. (link).
- Misleading meta-objectives and hidden incen-tives for distributional shift. (link).
- Fault tree analysis, methods, and applications: a review. IEEE transactions on reliability 34(3), 194–203. IEEE.
- Indistinguishability obfuscation from ddh-like assumptions on constant-degree graded encodings. In 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS), pp. 11–20. IEEE. (link).
- Matheny, J. (2023). Artificial intelligence challenges and opportunities for the department of defense. Testimony presented to the U.S. Senate Committee on Armed Services, Subcommittee on Cybersecurity, on April 19, 2023. (link).
- Mearns, A. (1965). Fault tree analysis- the study of unlikely events in complex systems(fault tree analysis as tool to identify component failure as probable cause of undesired event in complex system). In System Safety Symposium, Seattle, Wash, pp. 1965.
- Milli, S. and A. D. Dragan (2019). Literal or pedagogic human? analyzing human model misspecification in objective learning. arXiv preprint arXiv:1903.03877. (link).
- National Institute of Standards and Technology (2022). Ai risk management framework: Initial draft. (link).
- The building blocks of interpretability. Distill 3(3), e10. (link).
- Closing the AI accountability gap: defining an end-to-end framework for internal algorithmic auditing. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, pp. 33–44. (link).
- Identifying the effect of unemployment on crime. The Journal of Law and Economics 44(1), 259–283. The University of Chicago Press. (link).
- Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence 1(5), 206–215. Nature Publishing Group. (link).
- Russell, S. (2014). White paper: Value alignment in autonomous systems.
- Russell, S. (2019). Human compatible: Artificial intelligence and the problem of control. Penguin. (link).
- A study in rashomon curves and volumes: A new perspective on generalization and model simplicity in machine learning. arXiv preprint arXiv:1908.01755. (link).
- Preferences implicit in the state of the world. arXiv preprint arXiv:1902.04198. (link).
- Shapiro, D. and R. Shachter. User-agent value alignment. (link).
- Shapiro, D. G. (2011). The social agency problem. In 2011 AAAI Fall Symposium Series. (link).
- Sheehan, M. (2021). China’s new AI governance initiatives shouldn’t be ignored. The Carnegie Endowment for International Peace. (link).
- Aligning superintelligence with human interests: A technical research agenda. Machine Intelligence Research Institute (MIRI) technical report 8. Citeseer. (link).
- President Biden and U.K. Prime Minister Rishi Sunak hold news conference at White House | full video. CBS News. (link).
- Alignment for advanced machine learning systems. Machine Intelligence Research Institute. (link).
- Can digital computers think? Third Programme. BBC.
- Turing, A. (1951b). Intelligent Machinery, A Heretical Theory (c.1951). Reprinted in The Essential Turing, by B. Jack Copeland., 2004. Oxford University Press. (link).
- Conservative agency via attainable utility preservation. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, pp. 385–391. (link).
- US Senate Judiciary Committee (2023). US Senate Judiciary Committee Hearing on Oversight of A.I. (video footage). (link).
- Compute accounting principles can help reduce AI risks. (link).
- Oversight of A.I.: Rules for Artificial Intelligence. (link).
- Watson, H. A. et al. (1961). Launch control safety study. Bell labs.
- White House (2022). Blueprint for an AI Bill of Rights: Making Automated Systems Work for the American People. (link).
- Wiener, N. (1960). Some moral and technical consequences of automation. Science 131(3410), 1355–1358. JSTOR. (link).
- Yampolskiy, R. V. (2015). Taxonomy of pathways to dangerous ai. arXiv preprint arXiv:1511.03246.
- Yudkowsky, E. et al. (2008). Artificial intelligence as a positive and negative factor in global risk. Global catastrophic risks 1(303), 184.
- Thinking about risks from ai: accidents, misuse and structure. Lawfare. February 11, 2019. (link).
- Andrew Critch (23 papers)
- Stuart Russell (98 papers)