Can LLMs Configure Software Tools (2312.06121v1)
Abstract: In software engineering, the meticulous configuration of software tools is crucial in ensuring optimal performance within intricate systems. However, the complexity inherent in selecting optimal configurations is exacerbated by the high-dimensional search spaces presented in modern applications. Conventional trial-and-error or intuition-driven methods are both inefficient and error-prone, impeding scalability and reproducibility. In this study, we embark on an exploration of leveraging Large-LLMs to streamline the software configuration process. We identify that the task of hyperparameter configuration for machine learning components within intelligent applications is particularly challenging due to the extensive search space and performance-critical nature. Existing methods, including Bayesian optimization, have limitations regarding initial setup, computational cost, and convergence efficiency. Our work presents a novel approach that employs LLMs, such as Chat-GPT, to identify starting conditions and narrow down the search space, improving configuration efficiency. We conducted a series of experiments to investigate the variability of LLM-generated responses, uncovering intriguing findings such as potential response caching and consistent behavior based on domain-specific keywords. Furthermore, our results from hyperparameter optimization experiments reveal the potential of LLMs in expediting initialization processes and optimizing configurations. While our initial insights are promising, they also indicate the need for further in-depth investigations and experiments in this domain.
- Y. Zhang, H. He, O. Legunsen, S. Li, W. Dong, and T. Xu, “An evolutionary study of configuration design and implementation in cloud systems,” in 2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE). IEEE, 2021, pp. 188–200.
- N. Siegmund, N. Ruckel, and J. Siegmund, “Dimensions of software configuration: on the configuration context in modern software development,” in Proceedings of the 28th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, 2020, pp. 338–349.
- Z. Wan, X. Xia, D. Lo, and G. C. Murphy, “How does machine learning change software development practices?” IEEE Transactions on Software Engineering, vol. 47, no. 9, pp. 1857–1871, 2021.
- S. Fahmy, A. Deraman, J. Yahaya, and A. R. Hamdan, “Human competency assessment for software configuration management,” Annals of Emerging Technologies in Computing (AETiC), vol. 5, no. 5, pp. 69–78, 2021.
- J. Kannan, S. Barnett, L. Cruz, A. Simmons, and A. Agarwal, “Mlsmellhound: a context-aware code analysis tool,” in Proceedings of the ACM/IEEE 44th International Conference on Software Engineering: New Ideas and Emerging Results, 2022, pp. 66–70.
- G. Blumenschein, “Monitoring builds in a devops infrastructure/submitted by georg blumenschein,” 2023.
- R. Krishna, M. S. Iqbal, M. A. Javidian, B. Ray, and P. Jamshidi, “Cadet: Debugging and fixing misconfigurations using counterfactual reasoning,” arXiv preprint arXiv:2010.06061, 2020.
- M. S. Iqbal, R. Krishna, M. A. Javidian, B. Ray, and P. Jamshidi, “Unicorn: reasoning about configurable system performance through the lens of causality,” in Proceedings of the Seventeenth European Conference on Computer Systems, 2022, pp. 199–217.
- K. F. Tomasdottir, M. Aniche, and A. Van Deursen, “Why and how JavaScript developers use linters,” in ASE 2017 - Proceedings of the 32nd IEEE/ACM International Conference on Automated Software Engineering. Institute of Electrical and Electronics Engineers Inc., nov 2017, pp. 578–589.
- C. Vassallo, S. Panichella, F. Palomba, S. Proksch, H. C. Gall, and A. Zaidman, “How developers engage with static analysis tools in different contexts,” Empirical Software Engineering, vol. 25, no. 2, pp. 1419–1457, mar 2020. [Online]. Available: https://link.springer.com/article/10.1007/s10664-019-09750-5
- K. F. Tomasdottir, M. Aniche, and A. Van Deursen, “The Adoption of JavaScript Linters in Practice: A Case Study on ESLint,” IEEE Transactions on Software Engineering, vol. 46, no. 8, pp. 863–891, aug 2020.
- K. Wood and E. Pereira, “Impact of misconfiguration in cloud–investigation into security challenges,” International Journal Multimedia and Image Processing, vol. 1, no. 1, pp. 17–25, 2011.
- F. Hutter, M. Lindauer, A. Balint, S. Bayless, H. Hoos, and K. Leyton-Brown, “The configurable sat solver challenge (cssc),” Artificial Intelligence, vol. 243, pp. 1–25, 2017.
- M. Bilal, M. Canini, and R. Rodrigues, “Finding the right cloud configuration for analytics clusters,” in Proceedings of the 11th ACM Symposium on Cloud Computing, 2020, pp. 208–222.
- M. A. Langford, K. H. Chan, J. E. Fleck, P. K. McKinley, and B. H. Cheng, “Modalas: addressing assurance for learning-enabled autonomous systems in the face of uncertainty,” Software and Systems Modeling, pp. 1–21, 2023.
- M. Casimiro, P. Romano, D. Garlan, and L. Rodrigues, “Towards a Framework for Adapting Machine Learning Components,” Proceedings - 2022 IEEE International Conference on Autonomic Computing and Self-Organizing Systems, ACSOS 2022, pp. 131–140, 2022.
- J. Gesi, X. Shen, Y. Geng, Q. Chen, and I. Ahmed, “Leveraging feature bias for scalable misprediction explanation of machine learning models,” in Proceedings of the 45th International Conference on Software Engineering (ICSE), 2023.
- Y. Xiao, I. Beschastnikh, Y. Lin, R. S. Hundal, X. Xie, D. S. Rosenblum, and J. S. Dong, “Self-checking deep neural networks for anomalies and adversaries in deployment,” IEEE Transactions on Dependable and Secure Computing, 2022.
- D. Jin, Z. Jin, Z. Hu, O. Vechtomova, and R. Mihalcea, “Deep learning for text style transfer: A survey,” Computational Linguistics, vol. 48, no. 1, pp. 155–205, 2022.
- M. Shafiq and Z. Gu, “Deep residual learning for image recognition: A survey,” Applied Sciences, vol. 12, no. 18, p. 8972, 2022.
- G. Hinton, L. Deng, D. Yu, G. E. Dahl, A.-r. Mohamed, N. Jaitly, A. Senior, V. Vanhoucke, P. Nguyen, T. N. Sainath et al., “Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups,” IEEE Signal processing magazine, vol. 29, no. 6, pp. 82–97, 2012.
- C. E. Rasmussen, “Gaussian processes in machine learning,” in Summer school on machine learning. Springer, 2003, pp. 63–71.
- S. Wang, T. Tuor, T. Salonidis, K. K. Leung, C. Makaya, T. He, and K. Chan, “When edge meets learning: Adaptive control for resource-constrained distributed machine learning,” in IEEE INFOCOM 2018-IEEE conference on computer communications. IEEE, 2018, pp. 63–71.
- I. Standard, “Green AI : Do deep learning frameworks have different costs,” 2019.
- C.-j. W. Ramya, R. Udit, G. Bilge, A. Newsha, A. Kiwan, G. Chang, F. Aga, B. James, H. Charles, B. Michael, G. Anurag, M. Ott, A. Melnikov, S. Candido, D. Brooks, G. Chauhan, B. Lee, H.-h. S. L. Bugra, A. Max, B. Joe, S. Ravi, J. Mike, and R. Kim, “Sustainable AI: Environmental Implications, Challenges and Opportunities,” 2022.
- J. Kaddour, J. Harris, M. Mozes, H. Bradley, R. Raileanu, and R. McHardy, “Challenges and applications of large language models,” arXiv preprint arXiv:2307.10169, 2023.
- A. Barbu, D. Mayo, J. Alverio, W. Luo, C. Wang, D. Gutfreund, J. Tenenbaum, and B. Katz, “Objectnet: A large-scale bias-controlled dataset for pushing the limits of object recognition models,” Advances in neural information processing systems, vol. 32, 2019.
- P. Malo, A. Sinha, P. Korhonen, J. Wallenius, and P. Takala, “Good debt or bad debt: Detecting semantic orientations in economic texts,” Journal of the Association for Information Science and Technology, vol. 65, 2014.
- A. Hazourli, “Financialbert-a pretrained language model for financial text mining,” Technical report, Tech. Rep., 2022.
- A. Kumar, R. Shen, S. Bubeck, and S. Gunasekar, “How to fine-tune vision models with sgd,” arXiv preprint arXiv:2211.09359, 2022.
- P. Goyal, Q. Duval, I. Seessel, M. Caron, I. Misra, L. Sagun, A. Joulin, and P. Bojanowski, “Vision models are more robust and fair when pretrained on uncurated images without supervision,” arXiv preprint arXiv:2202.08360, 2022.