Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Promptor: A Conversational and Autonomous Prompt Generation Agent for Intelligent Text Entry Techniques (2310.08101v2)

Published 12 Oct 2023 in cs.CL and cs.AI

Abstract: Text entry is an essential task in our day-to-day digital interactions. Numerous intelligent features have been developed to streamline this process, making text entry more effective, efficient, and fluid. These improvements include sentence prediction and user personalization. However, as deep learning-based LLMs become the norm for these advanced features, the necessity for data collection and model fine-tuning increases. These challenges can be mitigated by harnessing the in-context learning capability of LLMs such as GPT-3.5. This unique feature allows the LLM to acquire new skills through prompts, eliminating the need for data collection and fine-tuning. Consequently, LLMs can learn various text prediction techniques. We initially showed that, for a sentence prediction task, merely prompting GPT-3.5 surpassed a GPT-2 backed system and is comparable with a fine-tuned GPT-3.5 model, with the latter two methods requiring costly data collection, fine-tuning and post-processing. However, the task of prompting LLMs to specialize in specific text prediction tasks can be challenging, particularly for designers without expertise in prompt engineering. To address this, we introduce Promptor, a conversational prompt generation agent designed to engage proactively with designers. Promptor can automatically generate complex prompts tailored to meet specific needs, thus offering a solution to this challenge. We conducted a user study involving 24 participants creating prompts for three intelligent text entry tasks, half of the participants used Promptor while the other half designed prompts themselves. The results show that Promptor-designed prompts result in a 35% increase in similarity and 22% in coherence over those by designers.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (61)
  1. Towards a human-like open-domain chatbot. arXiv preprint arXiv:2001.09977, 2020.
  2. T. Baldwin. Online Adaptation for Mobile Device Text Input Personalization. Michigan State University. Computer Science, 2012.
  3. J. R. Bellegarda. Statistical language model adaptation: review and perspectives. Speech communication, 42(1):93–108, 2004.
  4. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  5. Can large language models be an alternative to human evaluations? arXiv preprint arXiv:2305.01937, 2023.
  6. Language model adaptation using mixtures and an exponentially decaying cache. In 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 2, pp. 799–802. IEEE, 1997.
  7. Selection-inference: Exploiting large language models for interpretable logical reasoning. arXiv preprint arXiv:2205.09712, 2022.
  8. Hybrid processing for grammar and style checking. In Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008), pp. 153–160, 2008.
  9. R. DeMori and M. Federico. Language model adaptation. Computational models of speech pattern processing, pp. 280–303, 1999.
  10. Augmentative and alternative communication (aac) advances: A review of configurations for individuals with a speech disability. Sensors, 19(8):1911, 2019.
  11. G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences, 2007.
  12. Grammatical error correction using hybrid systems and type filtering. In Proceedings of the Eighteenth Conference on Computational Natural Language Learning: Shared Task, pp. 15–24, 2014.
  13. Effects of language modeling and its personalization on touchscreen typing performance. In Proceedings of the 33rd annual ACM conference on human factors in computing systems, pp. 649–658, 2015.
  14. A. F. Ganai and F. Khursheed. Predicting next word using rnn and lstm cells: Stastical language modeling. In 2019 Fifth International Conference on Image Information Processing (ICIIP), pp. 469–474. IEEE, 2019.
  15. S. Ghosh and P. O. Kristensson. Neural networks for text correction and completion in keyboard decoding. arXiv preprint arXiv:1709.06429, 2017.
  16. Hybrid model for word prediction using naive bayes and latent information. arXiv preprint arXiv:1803.00985, 2018.
  17. How close is chatgpt to human experts? comparison corpus, evaluation, and detection. arXiv preprint arXiv:2301.07597, 2023.
  18. K. Han. Personalized news recommendation and simulation based on improved collaborative filtering algorithm. Complexity, 2020:1–12, 2020.
  19. J. Hauke and T. Kossowski. Comparison of values of pearson’s and spearman’s correlation coefficients on the same sets of data. Quaestiones geographicae, 30(2):87–93, 2011.
  20. Considering optimization of english grammar error correction based on neural network. Neural Computing and Applications, pp. 1–13, 2022.
  21. T. Kocmi and C. Federmann. Large language models are state-of-the-art evaluators of translation quality. arXiv preprint arXiv:2302.14520, 2023.
  22. Modeling the speed of text entry with a word prediction interface. IEEE transactions on rehabilitation engineering, 2(3):177–187, 1994.
  23. Large language models are zero-shot reasoners. arXiv preprint arXiv:2205.11916, 2022.
  24. P. O. Kristensson. Five challenges for intelligent text entry methods. AI Magazine, 30(4):85–85, 2009.
  25. A Design Engineering Approach for Quantitatively Exploring Context-Aware Sentence Retrieval for Nonspeaking Individuals with Motor Disabilities, p. 1–11. Association for Computing Machinery, New York, NY, USA, 2020.
  26. R. Kuhn and R. De Mori. A cache-based natural language model for speech recognition. IEEE transactions on pattern analysis and machine intelligence, 12(6):570–583, 1990.
  27. Performance of chatgpt on usmle: Potential for ai-assisted medical education using large language models. PLOS Digital Health, 2(2):e0000198, 2023.
  28. T. O. Kvålseth. Note on cohen’s kappa. Psychological reports, 65(1):223–226, 1989.
  29. Deep learning-based context-sensitive spelling typing error correction. IEEE Access, 8:152565–152578, 2020.
  30. Microsoft dialogue challenge: Building end-to-end task-completion dialogue systems. arXiv preprint arXiv:1807.11125, 2018.
  31. V. Liu and L. B. Chilton. Design guidelines for prompt engineering text-to-image generative models. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, pp. 1–23, 2022.
  32. N. Madi and H. S. Al-Khalifa. Grammatical error checking systems: A review of approaches and emerging directions. In 2018 Thirteenth International Conference on Digital Information Management (ICDIM), pp. 142–147. IEEE, 2018.
  33. Personalizing sentences and text. Contemporary Educational Psychology, 16(3):287–292, 1991.
  34. D. Naber et al. A rule-based style and grammar checker. 2003.
  35. OpenAI. Openai evals. https://github.com/openai/evals, 2023.
  36. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pp. 311–318, 2002.
  37. nuqmm: Quantized matmul for efficient inference of large-scale generative language models. arXiv preprint arXiv:2206.09557, 2022.
  38. Improving language understanding by generative pre-training. 2018.
  39. E. Reiter. A structured review of the validity of bleu. Computational Linguistics, 44(3):393–401, 2018.
  40. Performance and user experience of touchscreen and gesture keyboards in a lab setting and in the wild. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, pp. 679–688, 2015.
  41. S. K. Sharma. Rule based grammar checking systems (a survey). International Journal of Computer Applications & Information Technology, 10(1):217–220, 2016.
  42. Kwickchat: A multi-turn dialogue system for aac using context-aware sentence generation by bag-of-keywords. In 27th International Conference on Intelligent User Interfaces, pp. 853–867, 2022.
  43. High-throughput generative inference of large language models with a single gpu. arXiv preprint arXiv:2303.06865, 2023.
  44. Autoprompt: Eliciting knowledge from language models with automatically generated prompts. arXiv preprint arXiv:2010.15980, 2020.
  45. M. Soam and S. Thakur. Next word prediction using deep learning: A comparative study. In 2022 12th International Conference on Cloud Computing, Data Science & Engineering (Confluence), pp. 653–658. IEEE, 2022.
  46. How to fine-tune bert for text classification? In Chinese Computational Linguistics: 18th China National Conference, CCL 2019, Kunming, China, October 18–20, 2019, Proceedings 18, pp. 194–206. Springer, 2019.
  47. Lstm neural networks for language modeling. In Thirteenth annual conference of the international speech communication association, 2012.
  48. Enabling conversational interaction with mobile ui using large language models. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp. 1–17, 2023.
  49. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171, 2022.
  50. A comprehensive survey of grammar error correction. arXiv preprint arXiv:2005.06600, 2020.
  51. Chain of thought prompting elicits reasoning in large language models. arXiv preprint arXiv:2201.11903, 2022.
  52. A prompt pattern catalog to enhance prompt engineering with chatgpt. arXiv preprint arXiv:2302.11382, 2023.
  53. R. F. Woolson. Wilcoxon signed-rank test. Wiley encyclopedia of clinical trials, pp. 1–3, 2007.
  54. B. Yang and J. Shen. Tinkerable aac keyboard. 2023.
  55. Tinkerable aac keyboard. 2023. doi: 10 . 17863/CAM . 91650
  56. Lm-critic: language models for unsupervised grammatical error correction. arXiv preprint arXiv:2109.06822, 2021.
  57. Words prediction based on n-gram model for free-text entry in electronic health records. Health information science and systems, 7:1–7, 2019.
  58. Type, then correct: intelligent text correction techniques for mobile text entry using neural networks. In Proceedings of the 32nd Annual ACM Symposium on User Interface Software and Technology, pp. 843–855, 2019.
  59. Perd: Personalized emoji recommendation with dynamic user preference. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1922–1926, 2022.
  60. Least-to-most prompting enables complex reasoning in large language models. arXiv preprint arXiv:2205.10625, 2022.
  61. Large language models are human-level prompt engineers. arXiv preprint arXiv:2211.01910, 2022.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Junxiao Shen (17 papers)
  2. John J. Dudley (8 papers)
  3. Jingyao Zheng (4 papers)
  4. Bill Byrne (57 papers)
  5. Per Ola Kristensson (45 papers)
Citations (2)