Farsight: Fostering Responsible AI Awareness During AI Application Prototyping (2402.15350v2)
Abstract: Prompt-based interfaces for LLMs have made prototyping and building AI-powered applications easier than ever before. However, identifying potential harms that may arise from AI applications remains a challenge, particularly during prompt-based prototyping. To address this, we present Farsight, a novel in situ interactive tool that helps people identify potential harms from the AI applications they are prototyping. Based on a user's prompt, Farsight highlights news articles about relevant AI incidents and allows users to explore and edit LLM-generated use cases, stakeholders, and harms. We report design insights from a co-design study with 10 AI prototypers and findings from a user study with 42 AI prototypers. After using Farsight, AI prototypers in our user study are better able to independently identify potential harms associated with a prompt and find our tool more useful and usable than existing resources. Their qualitative feedback also highlights that Farsight encourages them to focus on end-users and think beyond immediate harms. We discuss these findings and reflect on their implications for designing AI prototyping experiences that meaningfully engage with AI harms. Farsight is publicly accessible at: https://PAIR-code.github.io/farsight.
- Fatih Kadir Akın. 2022. Awesome ChatGPT Prompts. https://github.com/f/awesome-chatgpt-prompts
- Walking the Walk of AI Ethics: Organizational Challenges and the Individualization of Risk among Ethics Entrepreneurs. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency.
- Software Engineering for Machine Learning: A Case Study. ICSE (2019). https://doi.org/10.1109/ICSE-SEIP.2019.00042
- Guidelines for Human-AI Interaction. CHI (2019). https://doi.org/10.1145/3290605.3300233
- How Polymorphic Warnings Reduce Habituation in the Brain: Insights from an fMRI Study. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. https://doi.org/10.1145/2702123.2702322
- Anthropic. 2023. Core Views on AI Safety: When, Why, What, and How. https://www.anthropic.com/index/core-views-on-ai-safety
- Apple. 2023. Human Interface Guidelines: Machine Learning. https://developer.apple.com/design/human-interface-guidelines/machine-learning
- Ai Ethics Statements: Analysis and Lessons Learnt from Neurips Broader Impact Statements. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency.
- A General Language Assistant as a Laboratory for Alignment. arXiv 2112.00861 (2021). http://arxiv.org/abs/2112.00861
- James Auger. 2013. Speculative Design: Crafting the Speculation. Digital Creativity 24 (2013).
- Judgment Call the Game: Using Value Sensitive Design and Design Fiction to Surface Ethical Concerns Related to Technology. In Proceedings of the 2019 on Designing Interactive Systems Conference. https://doi.org/10.1145/3322276.3323697
- Designing Disaggregated Evaluations of AI Systems: Choices, Considerations, and Tradeoffs. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society (AIES ’21). https://doi.org/10.1145/3461702.3462610
- AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias. arXiv 1810.01943 (2018). http://arxiv.org/abs/1810.01943
- Detecting Discriminatory Risk through Data Annotation Based on Bayesian Inferences. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3442188.3445940
- Browser Interfaces and Extended Validation SSL Certificates: An Empirical Study. In Proceedings of the 2009 ACM Workshop on Cloud Computing Security. https://doi.org/10.1145/1655008.1655012
- Power to the People? Opportunities and Challenges for Participatory AI. In Equity and Access in Algorithms, Mechanisms, and Optimization. https://doi.org/10.1145/3551624.3555290
- Rainer Böhme and Stefan Köpsell. 2010. Trained to Accept?: A Field Experiment on Consent Dialogs. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/1753326.1753689
- On the Opportunities and Risks of Foundation Models. arXiv 2108.07258 (2022). http://arxiv.org/abs/2108.07258
- D33{}^{3}start_FLOATSUPERSCRIPT 3 end_FLOATSUPERSCRIPT Data-Driven Documents. IEEE TVCG 17 (2011). https://doi.org/10.1109/TVCG.2011.185
- Overcoming Failures of Imagination in AI Infused System Development and Deployment. arXiv 2011.13416 (2020). http://arxiv.org/abs/2011.13416
- Virginia Braun and Victoria Clarke. 2006. Using Thematic Analysis in Psychology. Qualitative Research in Psychology 3 (2006). https://doi.org/10.1191/1478088706qp063oa
- Philip AE Brey. 2012. Anticipatory Ethics for Emerging Technologies. Nanoethics 6 (2012).
- José Carlos Brustoloni and Ricardo Villamarín-Salomón. 2007. Improving Security Decisions with Polymorphic and Audited Dialogs. In Proceedings of the 3rd Symposium on Usable Privacy and Security. https://doi.org/10.1145/1280680.1280691
- AHA!: Facilitating AI Impact Assessment by Generating Examples of Harms. arXiv 2306.03280 (2023). http://arxiv.org/abs/2306.03280
- FAIRVIS: Visual Analytics for Discovering Intersectional Bias in Machine Learning. In 2019 IEEE Conference on Visual Analytics Science and Technology (VAST). https://doi.org/10.1109/VAST47406.2019.8986948
- 23 Ways to Nudge: A Review of Technology-Mediated Nudging in Human-Computer Interaction. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3290605.3300733
- Quantifying Memorization Across Neural Language Models. arXiv 2202.07646 (2023). http://arxiv.org/abs/2202.07646
- Extracting Training Data from Large Language Models. arXiv 2012.07805 (2021). http://arxiv.org/abs/2012.07805
- Surveying the Landscape of Ethics-Focused Design Methods. arXiv preprint arXiv:2102.08909 (2021).
- Alexandra Chouldechova. 2017. Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments. Big Data 5 (2017).
- Andrew R Chow and Billy Perrigo. 2023. The AI Arms Race Is Changing Everything. https://time.com/6255952/ai-impact-chatgpt-microsoft-google/
- Jacob Cohen. 1968. Weighted Kappa: Nominal Scale Agreement Provision for Scaled Disagreement or Partial Credit. Psychological Bulletin 70 (1968). https://doi.org/10.1037/h0026256
- Jacob Cohen. 2013. Statistical Power Analysis for the Behavioral Sciences (2nd ed ed.).
- A Systematic Review and Thematic Analysis of Community-Collaborative Approaches to Computing Research. In CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3491102.3517716
- Disentangling Disagreements on Offensiveness: A Cross-Cultural Study. In The 61st Annual Meeting of the Association for Computational Linguistics.
- Stakeholder Participation in AI: Beyond ”Add Diverse Stakeholders and Stir”. arXiv 2111.01122 (2021). http://arxiv.org/abs/2111.01122
- The Participatory Turn in AI Design: Theoretical Foundations and the Current State of Practice. In Equity and Access in Algorithms, Mechanisms, and Optimization. https://doi.org/10.1145/3617694.3623261
- MasterKey: Automated Jailbreak Across Multiple Large Language Model Chatbots. arXiv 2307.08715 (2023). http://arxiv.org/abs/2307.08715
- Whose Ground Truth? Accounting for Individual and Collective Identities Underlying Dataset Annotation. arXiv 2112.04554 (2021). http://arxiv.org/abs/2112.04554
- Deque. 2023. Axe DevTools: Digital Accessibility Testing Tools Dev Teams Love. https://www.deque.com/axe/devtools/
- Erik Derner and Kristina Batistič. 2023. Beyond the Safeguards: Exploring the Security Risks of ChatGPT. arXiv 2305.08005 (2023). http://arxiv.org/abs/2305.08005
- Toxicity in ChatGPT: Analyzing Persona-assigned Language Models. arXiv 2304.05335 (2023). http://arxiv.org/abs/2304.05335
- CrowdWorkSheets: Accounting for Individual and Collective Identities Underlying Crowdsourced Dataset Annotation. In 2022 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3531146.3534647
- Report of the Montréal Declaration for a Responsible Development of Artificial Intelligence. (2018). https://doi.org/1866/27795
- ‘‘That’s Important, but…”: How Computer Science Researchers Anticipate Unintended Consequences of Their Research Innovations. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems.
- Doteveryone. 2019. Consequence Scanning – an Agile Practice for Responsible Innovators. https://doteveryone.org.uk/project/consequence-scanning/
- Dovetail. 2023. Dovetail: All Your Customer Insights in One Place. https://dovetail.com/
- Olive Jean Dunn. 1961. Multiple Comparisons among Means. J. Amer. Statist. Assoc. 56 (1961). https://doi.org/10.1080/01621459.1961.10482090
- Anthony Dunne and Fiona Raby. 2013. Speculative Everything: Design, Fiction, and Social Dreaming.
- You’ve Been Warned: An Empirical Study of the Effectiveness of Web Browser Phishing Warnings. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/1357054.1357219
- Serge Egelman and Stuart Schechter. 2013. The Importance of Being Earnest [in Security Warnings]. In Financial Cryptography and Data Security: 17th International Conference, FC 2013, Okinawa, Japan, April 1-5, 2013, Revised Selected Papers 17.
- Seamful XAI: Operationalizing Seamful Design in Explainable AI. arXiv preprint arXiv:2211.06753 (2022).
- Improving SSL Warnings: Comprehension and Adherence. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. https://doi.org/10.1145/2702123.2702442
- Experimenting at Scale with Google Chrome’s SSL Warning. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/2556288.2557292
- Programming without a Programming Language: Challenges and Opportunities for Designing Developer Tools for Prompt Programming. In Extended Abstracts of the 2023 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3544549.3585737
- What Do We Teach When We Teach Tech Ethics?: A Syllabi Analysis. In Proceedings of the 51st ACM Technical Symposium on Computer Science Education. https://doi.org/10.1145/3328778.3366825
- Batya Friedman. 1996. Value-Sensitive Design. interactions 3 (1996).
- Batya Friedman and David Hendry. 2012. The Envisioning Cards: A Toolkit for Catalyzing Humanistic and Technical Imaginations. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/2207676.2208562
- A Survey of Value Sensitive Design Methods. Foundations and Trends® in Human–Computer Interaction 11 (2017).
- Value Sensitive Design: Theory and Methods. University of Washington technical report 2 (2002).
- Predictability and Surprise in Large Generative Models. In 2022 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3531146.3533229
- More Than ”If Time Allows”: The Role of Ethics in AI Education. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society. https://doi.org/10.1145/3375627.3375868
- Datasheets for Datasets. arXiv:1803.09010 [cs] (2020). http://arxiv.org/abs/1803.09010
- RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models. arXiv 2009.11462 (2020). http://arxiv.org/abs/2009.11462
- Improving Alignment of Dialogue Agents via Targeted Human Judgements. arXiv 2209.14375 (2022). http://arxiv.org/abs/2209.14375
- Stopping Spyware at the Gate: A User Study of Privacy, Notice and Spyware. In Proceedings of the 2005 Symposium on Usable Privacy and Security - SOUPS ’05. https://doi.org/10.1145/1073001.1073006
- Noticing Notice: A Large-Scale Experiment on the Timing of Software License Agreements. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/1240624.1240720
- Google. 2015. Lit: Simple Fast Web Components. https://lit.dev/
- Google. 2016. Lighthouse. https://github.com/GoogleChrome/lighthouse
- Google. 2023a. Google Ai Studio: Prototype with Generative AI. https://aistudio.google.com/app
- Google. 2023b. PaLM API: Safety Guidance. https://developers.generativeai.google/guide/safety_guidance
- Jury Learning: Integrating Dissenting Voices into Machine Learning Models. In CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3491102.3502004
- Grammarly. 2023. Grammarly: Free Writing AI Assistance. https://www.grammarly.com/
- Hans W. A. Hanley and Zakir Durumeric. 2023. Machine-Made Media: Monitoring the Mobilization of Machine-Generated Articles on Misinformation and Mainstream News Websites. arXiv 2305.09820 (2023). http://arxiv.org/abs/2305.09820
- Deconstructing Community-Based Collaborative Design: Towards More Equitable Participatory Design Engagements. Proceedings of the ACM on Human-Computer Interaction 3 (2019). https://doi.org/10.1145/3359318
- It’s Time to Do Something: Mitigating the Negative Impacts of Computing through a Change to the Peer Review Process. arXiv preprint arXiv:2112.09544 (2021).
- SUMMIT: Scaling Deep Learning Interpretability by Visualizing Activation and Attribution Summarizations. IEEE TVCG (2019). https://doi.org/10.1109/TVCG.2019.2934659
- Improving Fairness in Machine Learning Systems: What Do Industry Practitioners Need?. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3290605.3300830
- Planning for Natural Language Failures with the AI Playbook. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3411764.3445735
- Human Factors in Model Interpretability: Industry Practices, Challenges, and Needs. Proceedings of the ACM on Human-Computer Interaction 4 (2020). https://doi.org/10.1145/3392878
- The White House. 2022. Blueprint for an AI Bill of Rights: Making Automated Systems Work for the American People. https://www.whitehouse.gov/ostp/ai-bill-of-rights
- A Call to Alarms: Current State and Future Directions in the Battle against Alarm Fatigue. Journal of Electrocardiology 51 (2018). https://doi.org/10.1016/j.jelectrocard.2018.07.024
- Queer In AI: A Case Study in Community-Led Participatory AI. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. http://arxiv.org/abs/2303.16972
- PromptMaker: Prompt-based Prototyping with Large Language Models. In CHI Conference on Human Factors in Computing Systems Extended Abstracts. https://doi.org/10.1145/3491101.3503564
- Adapting Security Warnings to Counter Online Disinformation. In 30th USENIX Security Symposium (USENIX Security 21). https://www.usenix.org/conference/usenixsecurity21/presentation/kaiser
- Sensible AI: Re-imagining Interpretability and Explainability Using Sensemaking Theory. In 2022 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3531146.3533135
- ProPILE: Probing Privacy Leakage in Large Language Models. arXiv 2307.01881 (2023). http://arxiv.org/abs/2307.01881
- Online Survey on Novel Designs for Supporting Self-Reflection and Emotion Regulation in Online News Commenting. In Proceedings of the 25th International Academic Mindtrek Conference. https://doi.org/10.1145/3569219.3569411
- Shamika Klassen and Casey Fiesler. 2022. ” Run Wild a Little With Your Imagination” Ethical Speculation in Computing Education with Black Mirror. In Proceedings of the 53rd ACM Technical Symposium on Computer Science Education-Volume 1.
- Inherent Trade-Offs in the Fair Determination of Risk Scores. arXiv 1609.05807 (2016). http://arxiv.org/abs/1609.05807
- Jupyter Notebooks-a Publishing Format for Reproducible Computational Workflows. 2016 (2016). https://doi.org/10.3233/978-1-61499-649-1-87
- Learning Is Not a Spectator Sport: Doing Is Better than Watching for Learning from a MOOC. In Proceedings of the Second (2015) ACM Conference on Learning@ Scale.
- GeDi: Generative Discriminator Guided Sequence Generation. In Findings of the Association for Computational Linguistics: EMNLP 2021. https://doi.org/10.18653/v1/2021.findings-emnlp.424
- J. Richard Landis and Gary G. Koch. 1977. The Measurement of Observer Agreement for Categorical Data. Biometrics 33 (1977). https://doi.org/10.2307/2529310
- Michelle Seng Ah Lee and Jatinder Singh. 2020. The Landscape and Gaps in Open Source Fairness Toolkits. SSRN Electronic Journal (2020). https://doi.org/10.2139/ssrn.3695002
- Fei-Fei Li and John Etchemendy. 2022. Annual Report 2022: Stanford Institute for Human-centered Artificial Intelligence. https://hai-annual-report.stanford.edu
- Multi-Step Jailbreaking Privacy Attacks on ChatGPT. arXiv 2304.05197 (2023). http://arxiv.org/abs/2304.05197
- Q. Vera Liao and Jennifer Wortman Vaughan. 2023. AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap. arXiv 2306.01941 (2023). http://arxiv.org/abs/2306.01941
- Examining Responsibility and Deliberation in AI Impact Statements and Ethics Reviews. In Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society. https://doi.org/10.1145/3514094.3534155
- Trustworthy LLMs: A Survey and Guideline for Evaluating Large Language Models’ Alignment. arXiv 2308.05374 (2023). http://arxiv.org/abs/2308.05374
- News from Generative Artificial Intelligence Is Believed Less. In 2022 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3531146.3533077
- Scott M. Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17). https://doi.org/10.48550/arXiv.1705.07874
- Assessing the Fairness of AI Systems: AI Practitioners’ Processes, Challenges, and Needs for Support. Proceedings of the ACM on Human-Computer Interaction 6 (2022). https://doi.org/10.1145/3512899
- Co-Designing Checklists to Understand Organizational Challenges and Opportunities around Fairness in AI. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3313831.3376445
- Tambiama André Madiega. 2021. Artificial Intelligence Act. European Parliament: European Parliamentary Research Service (2021). https://artificialintelligenceact.eu
- On the Educational Impact of ChatGPT: Is Artificial Intelligence Ready to Obtain a University Degree?, In Proceedings of the 2023 Conference on Innovation and Technology in Computer Science Education V. 1. arXiv 2303.11146. https://doi.org/10.1145/3587102.3588827
- H. B. Mann and D. R. Whitney. 1947. On a Test of Whether One of Two Random Variables Is Stochastically Larger than the Other. The Annals of Mathematical Statistics 18 (1947). https://doi.org/10.1214/aoms/1177730491
- The Potential of Generative AI for Personalized Persuasion at Scale. Preprint. PsyArXiv. https://doi.org/10.31234/osf.io/rn97c
- Using Data Type Based Security Alert Dialogs to Raise Online Security Awareness. In Proceedings of the Seventh Symposium on Usable Privacy and Security. https://doi.org/10.1145/2078827.2078830
- Kenneth O. McGraw and S. P. Wong. 1992. A Common Language Effect Size Statistic. Psychological Bulletin 111 (1992). https://doi.org/10.1037/0033-2909.111.2.361
- Sean McGregor. 2020. Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database. arXiv 2011.08512 (2020). http://arxiv.org/abs/2011.08512
- Marry L. McHugh. 2012. Interrater Reliability: The Kappa Statistic. Biochemia Medica (2012). https://doi.org/10.11613/BM.2012.031
- UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv:1802.03426 (2020). http://arxiv.org/abs/1802.03426
- MDN. 2011. WebGL: 2D and 3D Graphics for the Web - Web APIs. https://developer.mozilla.org/en-US/docs/Web/API/WebGL_API
- MDN. 2021. Web Components - Web APIs. https://developer.mozilla.org/en-US/docs/Web/API/Web_components
- Sharan B Merriam et al. 2002. Introduction to Qualitative Research. Qualitative research in practice: Examples for discussion and analysis 1 (2002).
- Meta. 2023. Llama 2: Responsible Use Guide. https://ai.meta.com/llama-project/responsible-use-guide
- Between Subjectivity and Imposition: Power Dynamics in Data Annotation for Computer Vision. Proceedings of the ACM on Human-Computer Interaction 4 (2020). https://doi.org/10.1145/3415186
- Microsoft. 2020. Responsible AI Toolbox. Microsoft. https://github.com/microsoft/responsible-ai-toolbox
- Microsoft. 2022a. Harms Modeling - Azure Application Architecture Guide. https://learn.microsoft.com/en-us/azure/architecture/guide/responsible-innovation/harms-modeling/
- Microsoft. 2022b. Microsoft Responsible AI Impact Assessment Guide. (2022). https://aka.ms/RAIImpactAssessmentGuidePDF
- Model Cards for Model Reporting. In Proceedings of the Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3287560.3287596
- Dealing with Disagreements: Looking beyond the Majority Vote in Subjective Annotations. Transactions of the Association for Computational Linguistics 10 (2022). https://doi.org/10.1162/tacl_a_00449
- Visual Auditor: Interactive Visualization for Detection and Summarization of Model Biases. In VIS. https://doi.org/10.1109/VIS54862.2022.00018
- Delegate the Smartphone User? Security Awareness in Smartphone Platforms. Computers & Security 34 (2013). https://doi.org/10.1016/j.cose.2012.11.004
- Anticipatory Ethics and the Role of Uncertainty. arXiv preprint arXiv:2011.13170 (2020).
- Unpacking the Expressed Consequences of AI Research in Broader Impact Statements. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society.
- InterpretML: A Unified Framework for Machine Learning Interpretability. arXiv (2019). http://arxiv.org/abs/1909.09223
- Donald A. Norman and Stephen W. Draper. 1986. User Centered System Design: New Perspectives on Human-Computer Interaction.
- Bound by the Bounty: Collaboratively Shaping Evaluation Processes for Queer AI Harms. arXiv 2307.10223 (2023). http://arxiv.org/abs/2307.10223
- Cathy O’Neil and Hanna Gunn. 2020. Near-Term Artificial Intelligence and the Ethical Matrix. https://doi.org/10.1093/oso/9780190905033.003.0009
- OpenAI. 2023a. GPT-4 Technical Report. arXiv 2303.08774 (2023). http://arxiv.org/abs/2303.08774
- OpenAI. 2023b. OpenAI Playground. https://platform.openai.com/playground
- Google PAIR. 2019. People + AI Guidebook. https://pair.withgoogle.com/guidebook
- On the Risk of Misinformation Pollution with Large Language Models. arXiv 2305.13661 (2023). http://arxiv.org/abs/2305.13661
- Ellie Pavlick and Tom Kwiatkowski. 2019. Inherent Disagreements in Human Textual Inferences. Transactions of the Association for Computational Linguistics 7 (2019). https://doi.org/10.1162/tacl_a_00293
- Red Teaming Language Models with Language Models. arXiv 2202.03286 (2022). http://arxiv.org/abs/2202.03286
- A Framework to Assess (Dis)Agreement Among Diverse Rater Groups. arXiv 2311.05074 (2023). http://arxiv.org/abs/2311.05074
- Engaging Students with Instructor Solutions in Online Programming Homework. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3313831.3376857
- Institutionalizing Ethics in AI through Broader Impact Requirements. Nature Machine Intelligence 3 (2021).
- You Can’t Sit With Us: Exclusionary Pedagogy in AI Ethics Education. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3442188.3445914
- Closing the AI Accountability Gap: Defining an End-to-End Framework for Internal Algorithmic Auditing. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3351095.3372873
- Where Responsible AI Meets Reality: Practitioner Perspectives on Enablers for Shifting Organizational Practices. Proceedings of the ACM on Human-Computer Interaction 5 (2021). https://doi.org/10.1145/3449081
- Supporting Human-AI Collaboration in Auditing LLMs with LLMs. In Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society. https://doi.org/10.1145/3600211.3604712
- An Experience Sampling Study of User Reactions to Browser Warnings in the Field. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3173574.3174086
- E.M. Reingold and J.S. Tilford. 1981. Tidier Drawings of Trees. IEEE Transactions on Software Engineering SE-7 (1981). https://doi.org/10.1109/TSE.1981.234519
- An Investigation of Phishing Awareness and Education over Time: When and How to Best Remind Users. In Sixteenth Symposium on Usable Privacy and Security (SOUPS 2020). https://www.usenix.org/conference/soups2020/presentation/reinheimer
- Karen Renaud and Marc Dupuis. 2019. Cyber Security Fear Appeals: Unexpectedly Complicated. In Proceedings of the New Security Paradigms Workshop. https://doi.org/10.1145/3368860.3368864
- Marco Tulio Ribeiro and Scott Lundberg. 2022. Adaptive Testing and Debugging of NLP Models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). https://doi.org/10.18653/v1/2022.acl-long.230
- ”Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. https://doi.org/10.1145/2939672.2939778
- Beyond Accuracy: Behavioral Testing of NLP Models with CheckList. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/2020.acl-main.442
- From Plane Crashes to Algorithmic Harm: Applicability of Safety Engineering Frameworks for Responsible ML. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3544548.3581407
- Angler: Helping Machine Translation Practitioners Prioritize Model Improvements. In CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3544548.3580790
- Generating Phishing Attacks Using ChatGPT. arXiv 2305.05133 (2023). http://arxiv.org/abs/2305.05133
- Aequitas: A Bias and Fairness Audit Toolkit. arXiv 1811.05577 (2019). http://arxiv.org/abs/1811.05577
- Juan Pablo Sarmiento and Alyssa Friend Wise. 2022. Participatory and Co-Design of Learning Analytics: An Initial Review of the Literature. In LAK22: 12th International Learning Analytics and Knowledge Conference. https://doi.org/10.1145/3506860.3506910
- Shlomo S. Sawilowsky. 2009. New Effect Size Rules of Thumb. Journal of Modern Applied Statistical Methods 8 (2009). https://doi.org/10.22237/jmasm/1257035100
- Self-Diagnosis and Self-Debiasing: A Proposal for Reducing Corpus-Based Bias in NLP. arXiv 2103.00453 (2021). http://arxiv.org/abs/2103.00453
- Principles to Practices for Responsible AI: Closing the Gap. arXiv 2006.04707 (2020). http://arxiv.org/abs/2006.04707
- Digital Nudging: Guiding Online User Choices through Interface Design. Commun. ACM 61 (2018). https://doi.org/10.1145/3213765
- Howard J Seltman. 2012. Experimental Design and Analysis.
- S. S. Shapiro and M. B. Wilk. 1965. An Analysis of Variance Test for Normality (Complete Samples). Biometrika 52 (1965). https://doi.org/10.1093/biomet/52.3-4.591
- Meaningful Context, a Red Flag, or Both? Preferences for Enhanced Misinformation Warnings Among US Twitter Users. In 2022 European Symposium on Usable Security. https://doi.org/10.1145/3549015.3555671
- Sociotechnical Harms of Algorithmic Systems: Scoping a Taxonomy for Harm Reduction. In Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society. https://doi.org/10.1145/3600211.3604673
- Value Cards: An Educational Toolkit for Teaching Social Impacts of Machine Learning through Deliberation. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3442188.3445971
- Model Evaluation for Extreme Risks. arXiv 2305.15324 (2023). http://arxiv.org/abs/2305.15324
- Thinking through and Writing about Research Ethics beyond” Broader Impact”. arXiv preprint arXiv:2104.08205 (2021).
- Gabriel Simmons. 2023. Moral Mimicry: Large Language Models Produce Moral Rationalizations Tailored to Political Identity. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop). https://doi.org/10.18653/v1/2023.acl-srw.40
- Guy Simon. 2020. OpenWeb Tests the Impact of “Nudges” in Online Discussions. OpenWeb Blog (2020).
- TensorFlow.Js: Machine Learning for the Web and Beyond. arXiv (2019). https://arxiv.org/abs/1901.05350
- REAL ML: Recognizing, Exploring, and Articulating Limitations of Machine Learning Research. In 2022 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3531146.3533122
- Incorporating Ethics in Computing Courses: Barriers, Support, and Perspectives from Educators. In Proceedings of the 54th ACM Technical Symposium on Computer Science Education V. 1. https://doi.org/10.1145/3545945.3569855
- Irene Solaiman and Christy Dennison. 2021. Process for Adapting Language Models to Society (PALMS) with Values-Targeted Datasets. In Advances in Neural Information Processing Systems, Vol. 34. https://proceedings.neurips.cc/paper_files/paper/2021/file/2e855f9489df0712b4bd8ea9e2848c5a-Paper.pdf
- Beyond the Imitation Game: Quantifying and Extrapolating the Capabilities of Language Models. arXiv preprint arXiv:2206.04615 (2022). http://arxiv.org/abs/2206.04615
- Consequences, Schmonsequences! Considering the Future as Part of Publication and Peer Review in Computing Research. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems.
- Harini Suresh and John Guttag. 2021. A Framework for Understanding Sources of Harm throughout the Machine Learning Life Cycle. In Equity and Access in Algorithms, Mechanisms, and Optimization. https://doi.org/10.1145/3465416.3483305
- Elham Tabassi. 2023. AI Risk Management Framework: AI RMF (1.0). Technical Report. National Institute of Standards and Technology. https://doi.org/10.6028/NIST.AI.100-1
- Gemini: A Family of Highly Capable Multimodal Models. arXiv preprint arXiv:2312.11805 (2023). https://arxiv.org/abs/2312.11805
- The Language Interpretability Tool: Extensible, Interactive Visualizations and Analysis for NLP Models. In EMNLP Demo. https://doi.org/10.18653/v1/2020.emnlp-demos.15
- Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv 2307.09288 (2023). https://arxiv.org/abs/2307.09288
- Rama Adithya Varanasi and Nitesh Goyal. 2023. ‘‘It Is Currently Hodgepodge”: Examining AI/ML Practitioners’ Challenges during Co-Production of Responsible AI Values. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems.
- Error Analysis of Statistical Machine Translation Output. In Proceedings of the Fifth International Conference on Language Resources and Evaluation (LREC’06). http://www.lrec-conf.org/proceedings/lrec2006/pdf/413_pdf.pdf
- Designing Responsible AI: Adaptations of UX Practice to Meet Responsible AI Challenges. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3544548.3581278
- Wordflow: Social Prompt Engineering for Large Language Models. arXiv 2401.14447 (2024). http://arxiv.org/abs/2401.14447
- StickyLand: Breaking the Linear Presentation of Computational Notebooks. In CHI Conference on Human Factors in Computing Systems Extended Abstracts. https://doi.org/10.1145/3491101.3519653
- WizMap: Scalable Interactive Visualization for Exploring Large Machine Learning Embeddings. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations). https://aclanthology.org/2023.acl-demo.50
- Interpretability, Then What? Editing Machine Learning Models to Reflect Human Knowledge and Values. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD ’22). https://doi.org/10.1145/3534678.3539074
- NOVA: A Practical Method for Creating Notebook-Ready Visual Analytics. arXiv:2205.03963 (2022). http://arxiv.org/abs/2205.03963
- Fairlearn: Assessing and Improving Fairness of AI Systems. arXiv 2303.16626 (2023). http://arxiv.org/abs/2303.16626
- Ethical and Social Risks of Harm from Language Models. arXiv 2112.04359 (2021). https://doi.org/10.48550/arXiv.2112.04359
- Sociotechnical Safety Evaluation of Generative AI Systems. arXiv 2310.11986 (2023). http://arxiv.org/abs/2310.11986
- Taxonomy of Risks Posed by Language Models. In 2022 ACM Conference on Fairness, Accountability, and Transparency. https://doi.org/10.1145/3531146.3533088
- Benjamin Weiser and Nate Schweber. 2023. The ChatGPT Lawyer Explains Himself. The New York Times (2023). https://www.nytimes.com/2023/06/08/nyregion/lawyer-chatgpt-sanctions.html
- Challenges in Detoxifying Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2021. https://doi.org/10.18653/v1/2021.findings-emnlp.210
- The What-If Tool: Interactive Probing of Machine Learning Models. TVCG 26 (2019). https://doi.org/10.1109/TVCG.2019.2934619
- Meredith Whittaker. 2021. The Steep Cost of Capture. Interactions 28 (2021). https://doi.org/10.1145/3488666
- Richmond Y Wong and Vera Khovanskaya. 2018. Speculative Design in HCI: From Corporate Imaginations to Critical Orientations.
- Seeing Like a Toolkit: How Toolkits Envision the Work of AI Ethics. Proceedings of the ACM on Human-Computer Interaction 7 (2023). https://doi.org/10.1145/3579621
- Richmond Y Wong and Tonya Nguyen. 2021. Timelines: A World-Building Activity for Values Advocacy. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems.
- RECAST: Enabling User Recourse and Interpretability of Toxicity Detection Models with Interactive Visualization. Proceedings of the ACM on Human-Computer Interaction 5 (2021). https://doi.org/10.1145/3449280
- A Comparative Analysis of Industry Human-AI Interaction Guidelines. http://arxiv.org/abs/2010.11761
- Do Security Toolbars Actually Prevent Phishing Attacks?. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/1124772.1124863
- Errudite: Scalable, Reproducible, and Testable Error Analysis. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. https://doi.org/10.18653/v1/P19-1073
- AI Chains: Transparent and Controllable Human-AI Interaction by Chaining Large Language Model Prompts. In CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3491102.3517582
- Detoxifying Language Models Risks Marginalizing Minority Voices. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. https://doi.org/10.18653/v1/2021.naacl-main.190
- Investigating How Practitioners Use Human-AI Guidelines: A Case Study on the People + AI Guidebook. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3544548.3580900
- Why Johnny Can’t Prompt: How Non-AI Experts Try (and Fail) to Design LLM Prompts. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3544548.3581388
- Zijie J. Wang (39 papers)
- Chinmay Kulkarni (15 papers)
- Lauren Wilcox (10 papers)
- Michael Terry (25 papers)
- Michael Madaio (15 papers)