Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security (2401.05459v2)

Published 10 Jan 2024 in cs.HC, cs.AI, and cs.SE
Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security

Abstract: Since the advent of personal computing devices, intelligent personal assistants (IPAs) have been one of the key technologies that researchers and engineers have focused on, aiming to help users efficiently obtain information and execute tasks, and provide users with more intelligent, convenient, and rich interaction experiences. With the development of smartphones and IoT, computing and sensing devices have become ubiquitous, greatly expanding the boundaries of IPAs. However, due to the lack of capabilities such as user intent understanding, task planning, tool using, and personal data management etc., existing IPAs still have limited practicality and scalability. Recently, the emergence of foundation models, represented by LLMs, brings new opportunities for the development of IPAs. With the powerful semantic understanding and reasoning capabilities, LLM can enable intelligent agents to solve complex problems autonomously. In this paper, we focus on Personal LLM Agents, which are LLM-based agents that are deeply integrated with personal data and personal devices and used for personal assistance. We envision that Personal LLM Agents will become a major software paradigm for end-users in the upcoming era. To realize this vision, we take the first step to discuss several important questions about Personal LLM Agents, including their architecture, capability, efficiency and security. We start by summarizing the key components and design choices in the architecture of Personal LLM Agents, followed by an in-depth analysis of the opinions collected from domain experts. Next, we discuss several key challenges to achieve intelligent, efficient and secure Personal LLM Agents, followed by a comprehensive survey of representative solutions to address these challenges.

Introduction to Personal LLM Agents

Intelligent personal assistants (IPAs), commonly recognized as digital aides on various devices, have substantially improved over the years. Yet, they often fall short in understanding complex user intentions and executing tasks beyond simple commands. The advent of foundation models like LLMs holds the promise of transcending these limitations. The paper "Personal LLM Agents: Insights and Survey about the Capability, Efficiency, and Security" contributes to the emerging discourse on the potential role of LLMs in enhancing IPAs. It explores the architecture and intelligence levels of what the authors term Personal LLM Agents—agents embedded with personal data and devices for personalized assistance.

Architecture and Intelligence Levels of Personal LLM Agents

The core component of a Personal LLM Agent is a foundation model that connects various skills, manages local resources, and maintains user-related information. The agent's intelligence is tiered across five levels—ranging from simple step-following to representing users in complex affairs—highlighting the progression from basic command execution to advanced self-evolution and interpersonal interactions.

Challenges to Capability, Efficiency, and Security

Although LLMs can, in principle, significantly enhance IPAs, this integration is not without challenges. The research paper highlights several technical challenges that fall under capability, efficiency, and security. Addressing these challenges is essential for realizing the potential of Personal LLM Agents. A closer inspection reveals that while LLMs are proficient in task execution, context sensing, and memorization, they still require optimizations in model compression, kernel acceleration, and energy efficiency to be truly efficient. Furthermore, the paper points out that guaranteeing security and privacy, ranging from data confidentiality to system integrity, should be an intrinsic part of their design.

Future Directions and Considerations for Personal LLM Agents

The pursuit of integrating LLMs with IPAs demands a multi-faceted approach, encompassing not just the advancement of models but also ensuring their societal acceptability and ethical compliance. Continuous innovation in model design, system architecture, and security protocols is pivotal. Future research should consider the delicate balance between efficiency and capability while ensuring robustness against privacy intrusions and malicious attacks.

In summary, Personal LLM Agents represent a pivotal shift in personal computing paradigms, coupling the advanced cognitive capabilities of LLMs with the intimate personalization requirements of IPAs. As they evolve, they hold the potential to transform mundane tasks into enriched experiences, allowing individuals to focus on what genuinely matters to them.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (385)
  1. Apple. Siri. https://www.apple.com/siri/, 2023a. [Online; accessed December 26, 2023].
  2. Google. Google assistant for android. https://developer.android.com/guide/app-actions/overview, 2023a. [Online; accessed December 24, 2023].
  3. Amazon. Alexa. https://www.alexa.com, 2023. [Online; accessed December 26, 2023].
  4. Mapping natural language instructions to mobile ui action sequences, 2020.
  5. Glider: A reinforcement learning approach to extract ui scripts from websites. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’21, page 1420–1430, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 9781450380379. doi: 10.1145/3404835.3462905.
  6. Reinforcement learning on web interfaces using workflow-guided exploration. ArXiv, abs/1802.08802, 2018.
  7. A survey of large language models, 2023a.
  8. IBM. Ibm shoebox. https://www.ibm.com/ibm/history/exhibits/specialprod1/specialprod1_7.html, 2023. [Online; accessed December 26, 2023].
  9. Bruce Lowerre and R Reddy. The harpy speech recognition system: performance with large vocabularies. The Journal of the Acoustical Society of America, 60(S1):S10–S11, 1976.
  10. 1. 0 TANGORA - a large vocabulary speech recognition system for five languages. In Proc. 2nd European Conference on Speech Communication and Technology (Eurospeech 1991), pages 183–192, 1991. doi: 10.21437/Eurospeech.1991-44.
  11. L. Rabiner and B. Juang. An introduction to hidden markov models. IEEE ASSP Magazine, 3(1):4–16, 1986. doi: 10.1109/MASSP.1986.1165342.
  12. The dragon continuous speech recognition system: A real-time implementation. In Human Language Technology - The Baltic Perspectiv, 1990.
  13. Wikipedia. Speakable items. https://en.wikipedia.org/wiki/Speakable_items, 2023a. [Online; accessed January 5, 2023].
  14. Medspeak: Report creation with continuous speech recognition. In Proceedings of the ACM SIGCHI Conference on Human Factors in Computing Systems, CHI ’97, page 431–438, New York, NY, USA, 1997. Association for Computing Machinery. ISBN 0897918029. doi: 10.1145/258549.258829.
  15. Microsoft. Speech transcript - jim allchin, winhec 2002. https://news.microsoft.com/speeches/speech-transcript-jim-allchin-winhec-2002/, 2002. [Online; accessed January 5, 2023].
  16. John Markoff. Google is taking questions (spoken, via iphone). https://www.nytimes.com/2008/11/14/technology/internet/14voice.html, 2008. [Online; accessed January 5, 2024].
  17. Microsoft. Cortana. https://www.microsoft.com/en-us/cortana, 2023a. [Online; accessed December 26, 2023].
  18. OpenAI. Introduce chatgpt. https://openai.com/blog/chatgpt, 2022. [Online; accessed November 28, 2023].
  19. Microsoft. Announcing microsoft copilot, your everyday ai companion. https://blogs.microsoft.com/blog/2023/09/21/announcing-microsoft-copilot-your-everyday-ai-companion/, 2023b. [Online; accessed December 4, 2023].
  20. Apple. Sirikit: Empower users to interact with their devices through voice, intelligent suggestions, and personalized workflows. https://developer.apple.com/documentation/sirikit/, 2023b. [Online; accessed December 24, 2023].
  21. Apple. Shortcuts user guide. https://support.apple.com/en-hk/guide/shortcuts/welcome/ios, 2023c. [Online; accessed December 24, 2023].
  22. Joaoapps. Tasker: Total automation for android. https://tasker.joaoapps.com, 2023. [Online; accessed December 24, 2023].
  23. Absinthe. Anywhere shortcuts. https://play.google.com/store/apps/details?id=com.absinthe.anywhere_&hl=en_US&pli=1, 2023. [Online; accessed December 24, 2023].
  24. Programming iot devices by demonstration using mobile apps. In End-User Development: 6th International Symposium, IS-EUD 2017, Eindhoven, The Netherlands, June 13-15, 2017, Proceedings 6, pages 3–17. Springer, 2017a.
  25. Ulink: Enabling user-defined deep linking to app content. In Proceedings of the 14th Annual International Conference on Mobile Systems, Applications, and Services, MobiSys ’16, page 305–318, New York, NY, USA, 2016. Association for Computing Machinery. ISBN 9781450342698. doi: 10.1145/2906388.2906416.
  26. "what can i help you with?": Infrequent users’ experiences of intelligent personal assistants. In Proceedings of the 19th International Conference on Human-Computer Interaction with Mobile Devices and Services, MobileHCI ’17, New York, NY, USA, 2017. Association for Computing Machinery. ISBN 9781450350754. doi: 10.1145/3098279.3098539.
  27. A mixed-methods approach to understanding user trust after voice assistant failures. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, CHI ’23, New York, NY, USA, 2023. Association for Computing Machinery. ISBN 9781450394215. doi: 10.1145/3544548.3581152.
  28. "like having a really bad pa": The gulf between user expectation and experience of conversational agents. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, CHI ’16, page 5286–5297, New York, NY, USA, 2016. Association for Computing Machinery. ISBN 9781450333627. doi: 10.1145/2858036.2858288.
  29. Matthew B. Hoy. Alexa, siri, cortana, and more: An introduction to voice assistants. Medical Reference Services Quarterly, 37(1):81–88, 2018. doi: 10.1080/02763869.2018.1404391. PMID: 29327988.
  30. Humanoid: A deep learning-based approach to automated black-box android app testing. In 2019 34th IEEE/ACM International Conference on Automated Software Engineering (ASE), pages 1070–1073. IEEE, 2019.
  31. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, page 6000–6010, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964.
  32. Actionbert: Leveraging user actions for semantic understanding of user interfaces. In AAAI Conference on Artificial Intelligence, 2020.
  33. Vut: Versatile ui transformer for multi-modal multi-task user interface modeling. ArXiv, abs/2112.05692, 2021.
  34. Uibert: Learning generic multimodal representations for ui understanding. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, pages 1705–1712. International Joint Conferences on Artificial Intelligence Organization, 8 2021. doi: 10.24963/ijcai.2021/235. Main Track.
  35. Spotlight: Mobile ui understanding using vision-language models with a focus. ArXiv, abs/2209.14927, 2022.
  36. Lexi: Self-supervised learning of the ui language. ArXiv, abs/2301.10165, 2023.
  37. Uinav: A maker of ui automation agents. arXiv preprint arXiv:2312.10170, 2023a.
  38. World of bits: An open-domain platform for web-based agents. In Proceedings of the 34th International Conference on Machine Learning - Volume 70, ICML’17, page 3135–3144. JMLR.org, 2017.
  39. Learning to navigate the web. ArXiv, abs/1812.09195, 2018.
  40. Dom-q-net: Grounded rl on structured language. ArXiv, abs/1902.07257, 2019.
  41. A data-driven approach for learning to control computers. In International Conference on Machine Learning, pages 9466–9482. PMLR, 2022.
  42. Scaling laws for neural language models, 2020.
  43. Training language models to follow instructions with human feedback, 2022.
  44. Deep reinforcement learning from human preferences, 2023.
  45. Toolformer: Language models can teach themselves to use tools, 2023.
  46. Webgpt: Browser-assisted question-answering with human feedback, 2022.
  47. Multimodal web navigation with instruction-finetuned foundation models. ArXiv, abs/2305.11854, 2023.
  48. Progprompt: Generating situated robot task plans using large language models. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 11523–11530. IEEE, 2023.
  49. Robot task planning based on large language model representing knowledge with directed graph structures, 2023.
  50. Language models as zero-shot planners: Extracting actionable knowledge for embodied agents, 2022a.
  51. Hugginggpt: Solving ai tasks with chatgpt and its friends in hugging face, 2023.
  52. Mathcoder: Seamless code integration in llms for enhanced mathematical reasoning. ArXiv, abs/2310.03731, 2023a.
  53. Code llama: Open foundation models for code. ArXiv, abs/2308.12950, 2023.
  54. Solving challenging math word problems using gpt-4 code interpreter with code-based self-verification, 2023a.
  55. OpenAI. Gpt-4 technical report, 2023.
  56. Microsoft. Reinventing search with a new ai-powered microsoft bing and edge, your copilot for the web. https://blogs.microsoft.com/blog/2023/02/07/reinventing-search-with-a-new-ai-powered-microsoft-bing-and-edge-your-copilot-for-the-web/, 2023c. [Online; accessed December 8, 2023].
  57. Google. Bard: A conversational ai tool by google. https://bard.google.com, 2023b. [Online; accessed December 26, 2023].
  58. Google. Introducing gemini: our largest and most capable ai model. https://blog.google/technology/ai/google-gemini-ai/, 2023c. [Online; accessed December 26, 2023].
  59. Huawei. Reshaping industries with ai: Huawei cloud launches pangu models 3.0 and ascend ai cloud services. https://www.huaweicloud.com/intl/en-us/news/20230707180809498.html, 2023. [Online; accessed November 28, 2023].
  60. XiaoMi. Milm-6b. https://github.com/XiaoMi/MiLM-6B, 2023. [Online; accessed December 24, 2023].
  61. Sayed Naem Bokhari. The linux operating system. Computer, 28(8):74–79, 1995.
  62. Wikipedia. Borda count. https://en.wikipedia.org/wiki/Borda_count, 2023b. [Online; accessed December 13, 2023].
  63. Jinyu Li. Recent advances in end-to-end automatic speech recognition, 2022.
  64. End-to-end speech recognition: A survey, 2023.
  65. A survey on large language model based autonomous agents, 2023b.
  66. The rise and potential of large language model based agents: A survey, 2023.
  67. Igniting language intelligence: The hitchhiker’s guide from chain-of-thought reasoning to language agents, 2023a.
  68. Pomdp-based statistical spoken dialog systems: A review. Proceedings of the IEEE, 101(5):1160–1179, 2013. doi: 10.1109/JPROC.2012.2225812.
  69. Multi-task learning for joint language understanding and dialogue state tracking. In Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue, pages 376–384, Melbourne, Australia, July 2018. Association for Computational Linguistics. doi: 10.18653/v1/W18-5045.
  70. Toby Jia-Jun Li and Oriana Riva. Kite: Building conversational bots from mobile apps. In Proceedings of the 16th Annual International Conference on Mobile Systems, Applications, and Services, MobiSys ’18, page 96–109, New York, NY, USA, 2018. Association for Computing Machinery. ISBN 9781450357203. doi: 10.1145/3210240.3210339.
  71. Sugilite: Creating multimodal smartphone automation by demonstration. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, CHI ’17, page 6038–6049, New York, NY, USA, 2017b. Association for Computing Machinery. ISBN 9781450346559. doi: 10.1145/3025453.3025483.
  72. Can current task-oriented dialogue models automate real-world scenarios in the wild?, 2023a.
  73. Instructtods: Large language models for end-to-end task-oriented dialogue systems, 2023.
  74. Enhancing large language model induced task-oriented dialogue systems through look-forward motivated goals, 2023a.
  75. Are llms all you need for task-oriented dialogue?, 2023.
  76. Unlocking the potential of user feedback: Leveraging large language model as user simulators to enhance dialogue system. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, CIKM ’23, page 3953–3957, New York, NY, USA, 2023b. Association for Computing Machinery. ISBN 9798400701245. doi: 10.1145/3583780.3615220.
  77. Language models are few-shot learners, 2020.
  78. Microsoft. Bing web search api. https://www.microsoft.com/en-us/bing/apis/bing-web-search-api, 2023d.
  79. Gorilla: Large language model connected with massive apis. arXiv preprint arXiv:2305.15334, 2023.
  80. Gpt4tools: Teaching large language model to use tools via self-instruction, 2023.
  81. Toolllm: Facilitating large language models to master 16000+ real-world apis, 2023a.
  82. Chain-of-thought prompting elicits reasoning in large language models. In Advances in Neural Information Processing Systems, volume 35, pages 24824–24837. Curran Associates, Inc., 2022a.
  83. Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601, 2023a.
  84. Mrkl systems: A modular, neuro-symbolic architecture that combines large language models, external knowledge sources and discrete reasoning, 2022.
  85. Camel: Communicative agents for "mind" exploration of large scale language model society, 2023b.
  86. Language models can solve computer tasks. In Thirty-seventh Conference on Neural Information Processing Systems, 2023a.
  87. Chameleon: Plug-and-play compositional reasoning with large language models. In The 37th Conference on Neural Information Processing Systems (NeurIPS), 2023.
  88. ToolkenGPT: Augmenting frozen language models with massive tools via tool embeddings. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  89. Enabling conversational interaction with mobile ui using large language models. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, CHI ’23, New York, NY, USA, 2023c. Association for Computing Machinery. ISBN 9781450394215. doi: 10.1145/3544548.3580895.
  90. Droidbot-gpt: Gpt-powered ui automation for android. arXiv preprint arXiv:2304.07061, 2023a.
  91. Mind2web: Towards a generalist agent for the web, 2023.
  92. Empowering llm to use smartphone for intelligent task automation, 2023b.
  93. Axnav: Replaying accessibility tests from natural language, 2023.
  94. Explore, select, derive, and recall: Augmenting llm with human-like memory for mobile task automation. arXiv preprint arXiv:2312.03003, 2023b.
  95. Meta-gui: Towards multi-modal conversational agents on mobile gui. arXiv preprint arXiv:2205.11029, 2022.
  96. Responsible task automation: Empowering large language models as responsible task automators, 2023b.
  97. You only look at screens: Multimodal chain-of-action agents. arXiv preprint arXiv:2309.11436, 2023.
  98. Reinforced ui instruction grounding: Towards a generic ui task automation api. ArXiv, abs/2310.04716, 2023c.
  99. Gpt-4v in wonderland: Large multimodal models for zero-shot smartphone gui navigation. arXiv preprint arXiv:2311.07562, 2023.
  100. Cogagent: A visual language model for gui agents, 2023a.
  101. Can vision-language models think from a first-person perspective?, 2023.
  102. Lilian Weng. Llm powered autonomous agents. https://lilianweng.github.io/posts/2023-06-23-agent/, 2023.
  103. Autogpt. https://github.com/Significant-Gravitas/AutoGPT, 2023.
  104. Langchain. https://github.com/langchain-ai/langchain, 2023.
  105. Babyagi. https://github.com/yoheinakajima/babyagi, 2023.
  106. Anton Osika. Gpt-engineer. https://github.com/AntonOsika/gpt-engineer, 2023.
  107. Autoagents: The automatic agents generation framework. arXiv preprint, 2023a.
  108. Openagents: An open platform for language agents in the wild, 2023.
  109. KillianLucas. Open interpreter. https://github.com/KillianLucas/open-interpreter, 2023.
  110. Jerry Liu. LlamaIndex, 11 2022. URL https://github.com/jerryjliu/llama_index.
  111. Deshraj Yadav Taranjeet Singh. Embedchain: Data platform for llms - load, index, retrieve, and sync any unstructured data. https://github.com/embedchain/embedchain, 2023.
  112. Agents: An open-source framework for autonomous language agents, 2023b.
  113. Metagpt: Meta programming for a multi-agent collaborative framework, 2023b.
  114. Autogen: Enabling next-gen llm applications via multi-agent conversation framework. arXiv preprint arXiv:2308.08155, 2023.
  115. Automatic macro mining from interaction traces at scale, 2023.
  116. Androidenv: A reinforcement learning platform for android. arXiv preprint arXiv:2105.13231, 2021.
  117. Mobile-Env: An evaluation platform and benchmark for interactive agents in llm era. CoRR, abs/2305.08144, 2023d.
  118. Mapping natural language commands to web elements. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 4970–4976, Brussels, Belgium, October-November 2018. Association for Computational Linguistics. doi: 10.18653/v1/D18-1540.
  119. A dataset for interactive vision language navigation with unknown command feasibility. In European Conference on Computer Vision (ECCV), 2022.
  120. Ugif: Ui grounded instruction following, 2023.
  121. Android in the wild: A large-scale dataset for android device control, 2023.
  122. Webshop: Towards scalable real-world web interaction with grounded language agents. In Advances in Neural Information Processing Systems, volume 35, pages 20744–20757. Curran Associates, Inc., 2022a.
  123. Webarena: A realistic web environment for building autonomous agents. arXiv preprint arXiv:2307.13854, 2023c.
  124. Assistgui: Task-oriented desktop graphical user interface automation, 2023a.
  125. Feverphone: Accessible core-body temperature sensing for fever monitoring using commodity smartphones. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 7(1):1–23, 2023.
  126. Flowsense: Monitoring airflow in building ventilation systems using audio sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 6(1):1–26, 2022.
  127. Microcam: Leveraging smartphone microscope camera for context-aware contact surface sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 7(3):1–28, 2023c.
  128. Muse-fi: Contactless muti-person sensing exploiting near-field wi-fi channel variation. In Proceedings of the 29th Annual International Conference on Mobile Computing and Networking, pages 1–15, 2023d.
  129. Robust inertial motion tracking through deep sensor fusion across smart earbuds and smartphone. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 5(2):1–26, 2021.
  130. Dexar: Deep explainable sensor-based activity recognition in smart-home environments. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 6(1):1–30, 2022.
  131. Towards open-domain twitter user profile inference. In Findings of the Association for Computational Linguistics: ACL 2023, pages 3172–3188, 2023c.
  132. Identifying user habits through data mining on call data records. Engineering Applications of Artificial Intelligence, 54:49–61, 2016.
  133. Fedtherapist: Mental health monitoring with user-generated linguistic expressions on smartphones via federated learning. arXiv preprint arXiv:2310.16538, 2023.
  134. Affective state prediction from smartphone touch and sensor data in the wild. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, pages 1–14, 2022.
  135. Are you killing time? predicting smartphone users’ time-killing moments via fusion of smartphone sensor data and screenshots. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pages 1–19, 2023b.
  136. Remote breathing rate tracking in stationary position using the motion and acoustic sensors of earables. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pages 1–22, 2023.
  137. Samosa: Sensing activities with motion and subsampled audio. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 6(3):1–19, 2022.
  138. A multi-sensor approach to automatically recognize breaks and work activities of knowledge workers in academia. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 4(3):1–20, 2020.
  139. Dancingant: Body-empowered wireless sensing utilizing pervasive radiations from powerline. In Proceedings of the 29th Annual International Conference on Mobile Computing and Networking, pages 1–15, 2023.
  140. Sencom: Integrated sensing and communication with practical wifi. In Proceedings of the 29th Annual International Conference on Mobile Computing and Networking, pages 1–16, 2023.
  141. Sleepmore: Inferring sleep duration at scale via multi-device wifi sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 6(4):1–32, 2023.
  142. Automated mobile sensing strategies generation for human behaviour understanding. arXiv preprint arXiv:2311.05457, 2023b.
  143. M3sense: Affect-agnostic multitask representation learning using multimodal wearable sensors. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 6(2):1–32, 2022.
  144. Cocoa: Cross modality contrastive learning for sensor data. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 6(3):1–28, 2022.
  145. Attend and discriminate: Beyond the state-of-the-art for human activity recognition using wearable sensors. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 5(1):1–22, 2021.
  146. Predicting subjective measures of social anxiety from sparsely collected mobile sensor data. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 4(3):1–24, 2020.
  147. Fall detection based on interpretation of important features with wrist-wearable sensors. In Proceedings of the 28th Annual International Conference on Mobile Computing And Networking, pages 823–825, 2022.
  148. Guoguo: Enabling fine-grained indoor localization via smartphone. In Proceeding of the 11th annual international conference on Mobile systems, applications, and services, pages 235–248, 2013.
  149. Environmental sound recognition with time–frequency audio features. IEEE Transactions on Audio, Speech, and Language Processing, 17(6):1142–1158, 2009.
  150. S Chandrakala and SL Jayalakshmi. Environmental audio scene and sound event recognition for autonomous surveillance: A survey and comparative studies. ACM Computing Surveys (CSUR), 52(3):1–34, 2019.
  151. Complex daily activities, country-level diversity, and smartphone sensing: A study in denmark, italy, mongolia, paraguay, and uk. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pages 1–23, 2023.
  152. Generalization and personalization of mobile sensing-based mood inference models: An analysis of college students in eight countries. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 6(4):1–32, 2023.
  153. Detecting social contexts from mobile sensing indicators in virtual interactions with socially anxious individuals. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 7(3):1–26, 2023d.
  154. Examining the social context of alcohol drinking in young adults with smartphone sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 5(3):1–26, 2021a.
  155. One more bite? inferring food consumption level of college students using smartphone sensing and self-reports. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 5(1):1–28, 2021b.
  156. Exploring large language models for human mobility prediction under public events. arXiv preprint arXiv:2311.17351, 2023.
  157. Penetrative ai: Making llms comprehend the physical world. arXiv preprint arXiv:2310.09605, 2023.
  158. Activity recognition with smartphone sensors. Tsinghua science and technology, 19(3):235–249, 2014.
  159. mteeth: Identifying brushing teeth surfaces using wrist-worn inertial sensors. Proceedings of the ACM on interactive, mobile, wearable and ubiquitous technologies, 5(2):1–25, 2021.
  160. Guard your heart silently: Continuous electrocardiogram waveform monitoring with wrist-worn motion sensor. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 6(3):1–29, 2022.
  161. Healthwalks: Sensing fine-grained individual health condition via mobility data. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 4(4):1–26, 2020.
  162. Moodexplorer: Towards compound emotion detection via smartphone sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 1(4):1–30, 2018.
  163. Identifying mobile sensing indicators of stress-resilience. Proceedings of the ACM on interactive, mobile, wearable and ubiquitous technologies, 5(2):1–32, 2021.
  164. First-gen lens: Assessing mental health of first-generation students across their first year at college using mobile sensing. Proceedings of the ACM on interactive, mobile, wearable and ubiquitous technologies, 6(2):1–32, 2022a.
  165. Smartgpa: how smartphones can assess and predict academic performance of college students. In Proceedings of the 2015 ACM international joint conference on pervasive and ubiquitous computing, pages 295–306, 2015.
  166. Predicting personality traits from physical activity intensity. Computer, 52(7):47–56, 2019.
  167. Detecting job promotion in information workers using mobile sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 4(3):1–28, 2020.
  168. Predictors of life satisfaction based on daily activities from mobile sensor data. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pages 497–500, 2014.
  169. Social sensing: assessing social functioning of patients living with schizophrenia using mobile phone sensing. In Proceedings of the 2020 CHI conference on human factors in computing systems, pages 1–15, 2020a.
  170. Predicting symptom trajectories of schizophrenia using mobile sensing. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 1(3):1–24, 2017a.
  171. Smokingopp: Detecting the smoking’opportunity’context using mobile sensors. Proceedings of the ACM on interactive, mobile, wearable and ubiquitous technologies, 4(1):1–26, 2020.
  172. Zheng Chen. Palr: Personalization aware llms for recommendation. arXiv preprint arXiv:2305.07622, 2023.
  173. Bridging the information gap between domain-specific model and general llm for personalized recommendation. arXiv preprint arXiv:2311.03778, 2023e.
  174. Sentiment analysis through llm negotiations. arXiv preprint arXiv:2311.01876, 2023a.
  175. Conversational health agents: A personalized llm-powered agent framework. arXiv preprint arXiv:2310.02374, 2023.
  176. Lifelogging: Personal big data. Foundations and Trends® in information retrieval, 8(1):1–125, 2014.
  177. ‘outlines of a world coming into existence’: pervasive computing and the ethics of forgetting. Environment and planning B: planning and design, 34(3):431–445, 2007.
  178. Vision-based human activity recognition: a survey. Multimedia Tools and Applications, 79(41-42):30509–30555, 2020.
  179. Predicting personality from patterns of behavior collected with smartphones. Proceedings of the National Academy of Sciences, 117(30):17680–17687, 2020.
  180. Deep learning-based document modeling for personality detection from text. IEEE Intelligent Systems, 32(2):74–79, 2017.
  181. A survey of automatic personality detection from texts. In Proceedings of the 28th international conference on computational linguistics, pages 6284–6295, 2020.
  182. Facial emotion detection using deep learning. In 2020 international conference for emerging technology (INCET), pages 1–5. IEEE, 2020.
  183. Emotion detection of textual data: An interdisciplinary survey. In 2021 IEEE World AI IoT Congress (AIIoT), pages 0255–0261. IEEE, 2021.
  184. Akupm: Attention-enhanced knowledge-aware user preference model for recommendation. In Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pages 1891–1899, 2019.
  185. Automated extraction of personal knowledge from smartphone push notifications. In 2018 IEEE International Conference on Big Data (Big Data), pages 733–742. IEEE, 2018.
  186. An algorithm to transform natural language into sql queries for relational databases. Selforganizology, 3(3):100–116, 2016.
  187. Grammar-based neural text-to-sql generation. arXiv preprint arXiv:1905.13326, 2019.
  188. Privacystreams: Enabling transparency in personal data processing for mobile apps. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 1(3):76, 2017c.
  189. Generative agents: Interactive simulacra of human behavior. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, pages 1–22, 2023.
  190. Mot: Memory-of-thought enables chatgpt to self-improve. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 6354–6374, 2023.
  191. Augmenting language models with long-term memory. arXiv preprint arXiv:2306.07174, 2023e.
  192. Prompt-guided retrieval augmentation for non-knowledge-intensive tasks. arXiv preprint arXiv:2305.17653, 2023.
  193. Show your work: Scratchpads for intermediate computation with language models. arXiv preprint arXiv:2112.00114, 2021.
  194. Cognitive architectures for language agents. arXiv preprint arXiv:2309.02427, 2023.
  195. React: Synergizing reasoning and acting in language models. arXiv preprint arXiv:2210.03629, 2022b.
  196. Check your facts and try again: Improving large language models with external knowledge and automated feedback. arXiv preprint arXiv:2302.12813, 2023.
  197. Multi-stage episodic control for strategic exploration in text games. arXiv preprint arXiv:2201.01251, 2022.
  198. Keep calm and explore: Language models for action generation in text-based games. arXiv preprint arXiv:2010.02903, 2020.
  199. Improving language models by retrieving from trillions of tokens. In International conference on machine learning, pages 2206–2240. PMLR, 2022.
  200. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems, 33:9459–9474, 2020.
  201. Process and content in decisions from memory. Psychological Review, 129(1):73, 2022.
  202. Grounding language to entities and dynamics for generalization in reinforcement learning. In International Conference on Machine Learning, pages 4051–4062. PMLR, 2021.
  203. Human-assisted continual robot learning with foundation models. arXiv preprint arXiv:2309.14321, 2023.
  204. Voyager: An open-ended embodied agent with large language models. arXiv preprint arXiv:2305.16291, 2023f.
  205. Dreamcoder: growing generalizable, interpretable knowledge with wake–sleep bayesian program learning. Philosophical Transactions of the Royal Society A, 381(2251):20220050, 2023.
  206. Bootstrap your own skills: Learning to solve new tasks with large language model guidance. arXiv preprint arXiv:2310.10021, 2023f.
  207. Lifelong pretraining: Continually adapting language models to emerging corpora. arXiv preprint arXiv:2110.08534, 2021.
  208. Continual learning for named entity recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 13570–13577, 2021.
  209. Tool learning with foundation models. arXiv preprint arXiv:2304.08354, 2023b.
  210. Star: Bootstrapping reasoning with reasoning. Advances in Neural Information Processing Systems, 35:15476–15488, 2022.
  211. Large language models can self-improve. arXiv preprint arXiv:2210.11610, 2022b.
  212. Parameter-efficient transfer learning for nlp. In International Conference on Machine Learning, pages 2790–2799. PMLR, 2019a.
  213. Peft: State-of-the-art parameter-efficient fine-tuning methods. https://github.com/huggingface/peft, 2022.
  214. Adamix: Mixture-of-adapter for parameter-efficient tuning of large language models. arXiv preprint arXiv:2205.12410, 1(2):4, 2022b.
  215. Fireact: Toward language agent fine-tuning. arXiv preprint arXiv:2310.05915, 2023c.
  216. Gptq: Accurate post-training quantization for generative pre-trained transformers. arXiv preprint arXiv:2210.17323, 2022.
  217. Awq: Activation-aware weight quantization for llm compression and acceleration. arXiv preprint arXiv:2306.00978, 2023.
  218. Llm-qat: Data-free quantization aware training for large language models. arXiv preprint arXiv:2305.17888, 2023a.
  219. Zeroquant: Efficient and affordable post-training quantization for large-scale transformers. Advances in Neural Information Processing Systems, 35:27168–27183, 2022c.
  220. Smoothquant: Accurate and efficient post-training quantization for large language models. In International Conference on Machine Learning, pages 38087–38099. PMLR, 2023.
  221. Llm-pruner: On the structural pruning of large language models. arXiv preprint arXiv:2305.11627, 2023.
  222. Sparsegpt: Massive language models can be accurately pruned in one-shot. In International Conference on Machine Learning, pages 10323–10337. PMLR, 2023.
  223. A simple and effective pruning approach for large language models. arXiv preprint arXiv:2306.11695, 2023b.
  224. Baby llama: knowledge distillation from an ensemble of teachers trained on a small dataset with no performance penalty. arXiv preprint arXiv:2308.02019, 2023.
  225. Knowledge distillation of large language models. arXiv preprint arXiv:2306.08543, 2023.
  226. Distilling step-by-step! outperforming larger language models with less training data and smaller model sizes. arXiv preprint arXiv:2305.02301, 2023.
  227. Symbolic chain-of-thought distillation: Small models can also" think" step-by-step. arXiv preprint arXiv:2306.14050, 2023c.
  228. Zeroquant-v2: Exploring post-training quantization in llms from comprehensive study to low rank compensation, 2023b.
  229. Losparse: Structured compression of large language models based on low-rank and sparse approximation. arXiv preprint arXiv:2306.11222, 2023d.
  230. Compressing context to enhance inference efficiency of large language models, 2023e.
  231. Llmlingua: Compressing prompts for accelerated inference of large language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP 2023), December 2023a.
  232. Adapting language models to compress contexts. ArXiv, abs/2305.14788, 2023.
  233. Dynamic context pruning for efficient and interpretable autoregressive transformers. arXiv preprint arXiv:2305.15805, 2023.
  234. H2o: Heavy-hitter oracle for efficient generative inference of large language models. arXiv preprint arXiv:2306.14048, 2023g.
  235. Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in Neural Information Processing Systems, 35:16344–16359, 2022.
  236. Tri Dao. Flashattention-2: Faster attention with better parallelism and work partitioning. arXiv preprint arXiv:2307.08691, 2023.
  237. Flashdecoding++: Faster large language model inference on gpus. arXiv preprint arXiv:2311.01282, 2023c.
  238. Accelerating large language model decoding with speculative sampling. arXiv preprint arXiv:2302.01318, 2023d.
  239. Fast inference from transformers via speculative decoding. In International Conference on Machine Learning, pages 19274–19286. PMLR, 2023.
  240. Flexgen: High-throughput generative inference of large language models with a single gpu, 2023.
  241. Qualcomm. Snapdragon 8 gen 3 mobile platform. https://www.qualcomm.com/products/mobile/snapdragon/smartphones/snapdragon-8-series-mobile-platforms/snapdragon-8-gen-3-mobile-platform, 2023.
  242. Efficient deployment of transformer models on edge tpu accelerators: A real system evaluation. In Architecture and System Support for Transformer Models (ASSYST@ ISCA 2023), 2023.
  243. Dfx: A low-latency multi-fpga appliance for accelerating transformer-based text generation. In 2022 55th IEEE/ACM International Symposium on Microarchitecture (MICRO), pages 616–630. IEEE, 2022.
  244. Parameter-efficient transfer learning for NLP. CoRR, abs/1902.00751, 2019b.
  245. Llm-adapters: An adapter family for parameter-efficient fine-tuning of large language models, 2023e.
  246. Lora: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=nZeVKeeFYf9.
  247. Full parameter fine-tuning for large language models with limited resources, 2023.
  248. Sophia: A scalable stochastic second-order optimizer for language model pre-training, 2023b.
  249. Textbooks are all you need, 2023.
  250. Textbooks are all you need ii: phi-1.5 technical report, 2023f.
  251. Phi-2: The surprising power of small language models. https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/, 2023.
  252. Cachegen: Fast context loading for language model applications, 2023c.
  253. Qdrant team. Qdrant. https://github.com/qdrant/qdrant, 2021.
  254. Vespa.ai team. Vespa. https://github.com/vespa-engine/vespa, 2016.
  255. Analyticdb-v: A hybrid analytical engine towards query fusion for structured and unstructured data. Proc. VLDB Endow., 13(12):3152–3165, aug 2020. ISSN 2150-8097. doi: 10.14778/3415478.3415541.
  256. Milvus: A purpose-built vector data management system. In Proceedings of the 2021 International Conference on Management of Data, SIGMOD ’21, page 2614–2627, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 9781450383431. doi: 10.1145/3448016.3457550.
  257. Hqann: Efficient and robust similarity search for hybrid queries with structured and unstructured constraints. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, CIKM ’22, page 4580–4584, New York, NY, USA, 2022a. Association for Computing Machinery. ISBN 9781450392365. doi: 10.1145/3511808.3557610.
  258. Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, 7(3):535–547, 2019.
  259. Quicker adc: Unlocking the hidden potential of product quantization with simd. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(5):1666–1677, May 2021. ISSN 1939-3539. doi: 10.1109/tpami.2019.2952606.
  260. Vald team. Vald. https://github.com/vdaas/vald, 2019.
  261. Locality-sensitive hashing scheme based on p-stable distributions. In Proceedings of the Twentieth Annual Symposium on Computational Geometry, SCG ’04, page 253–262, New York, NY, USA, 2004. Association for Computing Machinery. ISBN 1581138857. doi: 10.1145/997817.997857.
  262. Random projection trees and low dimensional manifolds. In Proceedings of the Fortieth Annual ACM Symposium on Theory of Computing, STOC ’08, page 537–546, New York, NY, USA, 2008. Association for Computing Machinery. ISBN 9781605580470. doi: 10.1145/1374376.1374452.
  263. Randomized partition trees for exact nearest neighbor search, 2013.
  264. SPANN: Highly-efficient billion-scale approximate nearest neighborhood search. In Advances in Neural Information Processing Systems, 2021.
  265. Approximate nearest neighbor algorithm based on navigable small world graphs. Inf. Syst., 45:61–68, 2014.
  266. Efficient and robust approximate nearest neighbor search using hierarchical navigable small world graphs. IEEE Trans. Pattern Anal. Mach. Intell., 42(4):824–836, apr 2020. ISSN 0162-8828. doi: 10.1109/TPAMI.2018.2889473.
  267. Diskann: Fast accurate billion-point nearest neighbor search on a single node. In Advances in Neural Information Processing Systems, volume 32, 2019.
  268. Diskann++: Efficient page-based search over isomorphic mapped graph index using query-sensitivity entry vertex. ArXiv, abs/2310.00402, 2023.
  269. Cxl-anns: Software-hardware collaborative memory disaggregation and computation for billion-scale approximate nearest neighbor search. In USENIX Annual Technical Conference, 2023.
  270. Co-design hardware and algorithm for vector search. Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, 2023b.
  271. Understanding and overcoming the challenges of efficient transformer quantization. arXiv preprint arXiv:2109.12948, 2021.
  272. Outlier suppression: Pushing the limit of low-bit transformer language models. Advances in Neural Information Processing Systems, 35:17402–17414, 2022b.
  273. llama.cpp developers. ggerganov/llama.cpp: Port of facebook’s llama model in c/c++. https://github.com/ggerganov/llama.cpp, 2023.
  274. MLC team. MLC-LLM, 2023. URL https://github.com/mlc-ai/mlc-llm.
  275. Rptq: Reorder-based post-training quantization for large language models. arXiv preprint arXiv:2304.01089, 2023a.
  276. Outlier suppression+: Accurate quantization of large language models by equivalent and optimal shifting and scaling. arXiv preprint arXiv:2304.09145, 2023a.
  277. Qllm: Accurate and efficient low-bitwidth quantization for large language models. arXiv preprint arXiv:2310.08041, 2023d.
  278. Deepspeed-inference: enabling efficient inference of transformer models at unprecedented scale. In SC22: International Conference for High Performance Computing, Networking, Storage and Analysis, pages 1–15. IEEE, 2022.
  279. Efficient memory management for large language model serving with pagedattention. In Proceedings of the 29th Symposium on Operating Systems Principles, pages 611–626, 2023.
  280. Scissorhands: Exploiting the persistence of importance hypothesis for llm kv cache compression at test time. arXiv preprint arXiv:2305.17118, 2023e.
  281. Colt5: Faster long-range transformers with conditional computation. In Conference on Empirical Methods in Natural Language Processing, 2023.
  282. Skipdecode: Autoregressive skip decoding with batching and caching for efficient llm inference. arXiv preprint arXiv:2307.02628, 2023.
  283. Linformer: Self-attention with linear complexity. ArXiv, abs/2006.04768, 2020b.
  284. Specinfer: Accelerating generative llm serving with speculative inference and token tree verification. arXiv preprint arXiv:2305.09781, 2023.
  285. Accelerating llm inference with staged speculative decoding. arXiv preprint arXiv:2308.04623, 2023.
  286. Speculative decoding with big little decoder. In Thirty-seventh Conference on Neural Information Processing Systems, 2023b.
  287. Accelerating attention mechanism on fpgas based on efficient reconfigurable systolic array. ACM Transactions on Embedded Computing Systems, 22(6):1–22, 2023.
  288. Towards a unified view of parameter-efficient transfer learning, 2022.
  289. Prefix-tuning: Optimizing continuous prompts for generation, 2021.
  290. P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks, 2022.
  291. Llama-adapter: Efficient fine-tuning of language models with zero-init attention, 2023h.
  292. Gpt understands, too, 2023f.
  293. Efficient estimation of word representations in vector space, 2013.
  294. Distributed representations of sentences and documents. In Proceedings of the 31st International Conference on Machine Learning, volume 32 of Proceedings of Machine Learning Research, pages 1188–1196, Bejing, China, 22–24 Jun 2014. PMLR.
  295. Reta-llm: A retrieval-augmented large language model toolkit, 2023g.
  296. Eric Melz. Enhancing llm intelligence with arm-rag: Auxiliary rationale memory for retrieval augmented generation, 2023.
  297. Training language models with memory augmentation. ArXiv, abs/2205.12674, 2022. URL https://api.semanticscholar.org/CorpusID:249062699.
  298. A comprehensive survey on vector database: Storage and retrieval technique, challenge. arXiv preprint arXiv:2310.11703, 2023.
  299. Survey of vector database management systems, 2023.
  300. Toni Taipalus. Vector database management systems: Fundamental concepts, use-cases, and current challenges. ArXiv, abs/2309.11322, 2023.
  301. Memorizing transformers. ArXiv, abs/2203.08913, 2022b.
  302. Ret-llm: Towards a general read-write memory for large language models. arXiv preprint arXiv:2305.14322, 2023.
  303. Approximate nearest neighbor search in high dimensional vector databases: Current research and future directions, 2023.
  304. Filtered-diskann: Graph algorithms for approximate nearest neighbor search with filters. In Proceedings of the ACM Web Conference 2023, WWW ’23, page 3406–3416, New York, NY, USA, 2023. Association for Computing Machinery. ISBN 9781450394161. doi: 10.1145/3543507.3583552.
  305. Song: Approximate nearest neighbor search on gpu. 2020 IEEE 36th International Conference on Data Engineering (ICDE), pages 1033–1044, 2020.
  306. Ggnn: Graph-based gpu nearest neighbor search. IEEE Transactions on Big Data, 9:267–279, 2019.
  307. Cagra: Highly parallel graph construction and approximate nearest neighbor search for gpus. ArXiv, abs/2308.15136, 2023.
  308. Llama: Open and efficient foundation language models, 2023.
  309. BlueLM Team. Bluelm: An open multilingual 7b language model. https://github.com/vivo-ai-lab/BlueLM, 2023.
  310. Spqr: A sparse-quantized representation for near-lossless llm weight compression. arXiv preprint arXiv:2306.03078, 2023.
  311. On data banks and privacy homomorphisms. Foundations of secure computation, 4(11):169–180, 1978.
  312. Craig Gentry. Fully homomorphic encryption using ideal lattices. In Proceedings of the forty-first annual ACM symposium on Theory of computing, pages 169–178, 2009.
  313. Cryptonets: Applying neural networks to encrypted data with high throughput and accuracy. In International conference on machine learning, pages 201–210. PMLR, 2016.
  314. The-x: Privacy-preserving transformer inference with homomorphic encryption. arXiv preprint arXiv:2206.00216, 2022.
  315. Cheetah: Optimizing and accelerating homomorphic encryption for private inference. In 2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA), pages 26–39. IEEE, 2021.
  316. A survey on homomorphic encryption schemes: Theory and implementation. ACM Computing Surveys (Csur), 51(4):1–35, 2018.
  317. Slalom: Fast, verifiable and private execution of neural networks in trusted hardware. arXiv preprint arXiv:1806.03287, 2018.
  318. Security vulnerabilities of sgx and countermeasures: A survey. ACM Computing Surveys (CSUR), 54(6):1–36, 2021.
  319. Erika McCallister. Guide to protecting the confidentiality of personally identifiable information, volume 800. Diane Publishing, 2010.
  320. Privacy-preserving neural representations of text. arXiv preprint arXiv:1808.09408, 2018.
  321. Textfusion: Privacy-preserving pre-trained model inference via token fusion. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 8360–8371, 2022.
  322. Textobfuscator: Making pre-trained language model a privacy protector via obfuscating word representations. In Findings of the Association for Computational Linguistics: ACL 2023, pages 5459–5473, 2023d.
  323. Adversarial training for large neural language models. arXiv preprint arXiv:2004.08994, 2020.
  324. User-driven access control: Rethinking permission granting in modern operating systems. In 2012 IEEE Symposium on Security and Privacy, pages 224–238. IEEE, 2012.
  325. Taintdroid: an information-flow tracking system for realtime privacy monitoring on smartphones. ACM Transactions on Computer Systems (TOCS), 32(2):1–29, 2014.
  326. Intriguing properties of neural networks, 2014.
  327. Adversarial attacks and defenses in images, graphs and text: A review. International Journal of Automation and Computing, 17(2):151–178, 2020. doi: 10.1007/s11633-019-1211-x.
  328. Certifying llm safety against adversarial prompting, 2023.
  329. On evaluating adversarial robustness of large vision-language models. arXiv preprint arXiv:2305.16934, 2023b.
  330. Jailbroken: How does llm safety training fail? arXiv preprint arXiv:2307.02483, 2023b.
  331. On the adversarial robustness of multi-modal foundation models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3677–3685, 2023.
  332. Misusing tools in large language models with visual adversarial examples, 2023.
  333. Badnets: Evaluating backdooring attacks on deep neural networks. IEEE Access, 7:47230–47244, 2019. doi: 10.1109/ACCESS.2019.2909068.
  334. Patchbackdoor: Backdoor attack against deep neural networks without model modification. In Proceedings of the 31st ACM International Conference on Multimedia, pages 9134–9142, 2023b.
  335. Backdoor attacks for in-context learning with language models. arXiv preprint arXiv:2307.14692, 2023.
  336. Prompt as triggers for backdoor attack: Examining the vulnerability in language models, 2023c.
  337. Poisonprompt: Backdoor attack on prompt-based large language models, 2023c.
  338. Defending against backdoor attacks in natural language generation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 5257–5265, 2023c.
  339. Not what you’ve signed up for: Compromising real-world llm-integrated applications with indirect prompt injection. In Proceedings of the 16th ACM Workshop on Artificial Intelligence and Security, AISec ’23, page 79–90, New York, NY, USA, 2023. Association for Computing Machinery. ISBN 9798400702600. doi: 10.1145/3605764.3623985.
  340. Ignore previous prompt: Attack techniques for language models, 2022.
  341. Prompt injection attack against llm-integrated applications, 2023h.
  342. Jailbreak in pieces: Compositional adversarial attacks on multi-modal language models, 2023.
  343. Jailbreaking black box large language models in twenty queries, 2023.
  344. Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21), pages 2633–2650. USENIX Association, August 2021. ISBN 978-1-939133-24-3.
  345. Smoothllm: Defending large language models against jailbreaking attacks, 2023.
  346. Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12):1–38, 2023.
  347. A survey of hallucination in large foundation models. arXiv preprint arXiv:2309.05922, 2023.
  348. Dera: enhancing large language model completions with dialog-enabled resolving agents. arXiv preprint arXiv:2303.17071, 2023.
  349. Cumulative reasoning with large language models. arXiv preprint arXiv:2308.04371, 2023i.
  350. Learning from mistakes makes llm better reasoner. arXiv preprint arXiv:2310.20689, 2023.
  351. Large language models can learn rules. arXiv preprint arXiv:2310.07064, 2023.
  352. Don’t stop pretraining: Adapt language models to domains and tasks. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020.
  353. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys, 55(9):1–35, 2023i.
  354. Finetuned language models are zero-shot learners. In International Conference on Learning Representations, 2021.
  355. Rlaif: Scaling reinforcement learning from human feedback with ai feedback. arXiv preprint arXiv:2309.00267, 2023c.
  356. Openchat: Advancing open-source language models with mixed-quality data. arXiv preprint arXiv:2309.11235, 2023g.
  357. Language models (mostly) know what they know. arXiv preprint arXiv:2207.05221, 2022.
  358. Self-refine: Iterative refinement with self-feedback. arXiv preprint arXiv:2303.17651, 2023.
  359. Reflexion: Language agents with verbal reinforcement learning, 2023.
  360. Teaching large language models to self-debug. arXiv preprint arXiv:2304.05128, 2023e.
  361. Selfcheckgpt: Zero-resource black-box hallucination detection for generative large language models. arXiv preprint arXiv:2303.08896, 2023.
  362. Improving factuality and reasoning in language models through multiagent debate. arXiv preprint arXiv:2305.14325, 2023.
  363. Retrieval augmented language model pre-training. In International conference on machine learning, pages 3929–3938. PMLR, 2020.
  364. Knowledge graph embedding: A survey of approaches and applications. IEEE Transactions on Knowledge and Data Engineering, 29(12):2724–2743, 2017b.
  365. Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of naacL-HLT, volume 1, page 2, 2019.
  366. Large language models can be easily distracted by irrelevant context. In International Conference on Machine Learning, pages 31210–31227. PMLR, 2023.
  367. Chain-of-note: Enhancing robustness in retrieval-augmented language models. arXiv preprint arXiv:2311.09210, 2023.
  368. Self-rag: Learning to retrieve, generate, and critique through self-reflection. arXiv preprint arXiv:2310.11511, 2023.
  369. Self-knowledge guided retrieval augmentation for large language models. arXiv preprint arXiv:2310.05002, 2023h.
  370. Learning to filter context for retrieval-augmented generation. arXiv preprint arXiv:2311.08377, 2023i.
  371. Critic: Large language models can self-correct with tool-interactive critiquing. arXiv preprint arXiv:2305.11738, 2023.
  372. Gradient-based constrained sampling from language models. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 2251–2277, 2022.
  373. Cgmh: Constrained sentence generation by metropolis-hastings sampling. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 6834–6842, 2019.
  374. Making language models better reasoners with step-aware verifier. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023g.
  375. Large language models are better reasoners with self-verification. In Findings of the Association for Computational Linguistics: EMNLP 2023, 2023.
  376. A survey of the state of explainable ai for natural language processing. arXiv preprint arXiv:2010.00711, 2020.
  377. Explainability for large language models: A survey. arXiv preprint arXiv:2309.01029, 2023d.
  378. Teach me to explain: A review of datasets for explainable natural language processing. arXiv preprint arXiv:2102.12060, 2021.
  379. What to learn, and how: Toward effective learning from rationales. In Findings of the Association for Computational Linguistics: ACL 2022, 2022.
  380. Rationalization for explainable nlp: A survey. Frontiers in Artificial Intelligence, 6, 2023.
  381. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171, 2022c.
  382. Enhancing chain-of-thoughts prompting with iterative bootstrapping in large language models. arXiv preprint arXiv:2304.11657, 2023d.
  383. Overthinking the truth: Understanding how language models process false demonstrations. arXiv preprint arXiv:2307.09476, 2023.
  384. Inference-time intervention: Eliciting truthful answers from a language model. arXiv preprint arXiv:2306.03341, 2023h.
  385. Mutual information alleviates hallucinations in abstractive summarization. arXiv preprint arXiv:2210.13210, 2022.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (25)
  1. Yuanchun Li (37 papers)
  2. Hao Wen (52 papers)
  3. Weijun Wang (21 papers)
  4. Xiangyu Li (52 papers)
  5. Yizhen Yuan (3 papers)
  6. Guohong Liu (2 papers)
  7. Jiacheng Liu (67 papers)
  8. Wenxing Xu (3 papers)
  9. Xiang Wang (279 papers)
  10. Yi Sun (146 papers)
  11. Rui Kong (9 papers)
  12. Yile Wang (24 papers)
  13. Hanfei Geng (2 papers)
  14. Jian Luan (50 papers)
  15. Xuefeng Jin (2 papers)
  16. Zilong Ye (2 papers)
  17. Guanjing Xiong (3 papers)
  18. Fan Zhang (685 papers)
  19. Xiang Li (1002 papers)
  20. Mengwei Xu (62 papers)
Citations (96)