Human-in-the-Loop Policy Optimization for Preference-Based Multi-Objective Reinforcement Learning (2401.02160v1)
Abstract: Multi-objective reinforcement learning (MORL) aims to find a set of high-performing and diverse policies that address trade-offs between multiple conflicting objectives. However, in practice, decision makers (DMs) often deploy only one or a limited number of trade-off policies. Providing too many diversified trade-off policies to the DM not only significantly increases their workload but also introduces noise in multi-criterion decision-making. With this in mind, we propose a human-in-the-loop policy optimization framework for preference-based MORL that interactively identifies policies of interest. Our method proactively learns the DM's implicit preference information without requiring any a priori knowledge, which is often unavailable in real-world black-box decision scenarios. The learned preference information is used to progressively guide policy optimization towards policies of interest. We evaluate our approach against three conventional MORL algorithms that do not consider preference information and four state-of-the-art preference-based MORL algorithms on two MORL environments for robot control and smart grid management. Experimental results fully demonstrate the effectiveness of our proposed method in comparison to the other peer algorithms.
- Y. Qin, H. Wang, S. Yi, X. Li, and L. Zhai, “An energy-aware scheduling algorithm for budget-constrained scientific workflows based on multi-objective reinforcement learning,” J. Supercomput., vol. 76, no. 1, pp. 455–480, 2020.
- J. Xu, K. Li, and M. Abusara, “Multi-objective reinforcement learning based multi-microgrid system optimisation problem,” in EMO’21: Proc. of 11th International Conference on Evolutionary Multi-Criterion Optimization, vol. 12654. Springer, 2021, pp. 684–696.
- J. Xu, Y. Tian, P. Ma, D. Rus, S. Sueda, and W. Matusik, “Prediction-guided multi-objective reinforcement learning for continuous robot control,” in ICML’20: Proc. of the 37th International Conference on Machine Learning, ser. Proceedings of Machine Learning Research, vol. 119. PMLR, 2020, pp. 10 607–10 616.
- C. F. Hayes, R. Radulescu, E. Bargiacchi, J. Källström, M. Macfarlane, M. Reymond, T. Verstraeten, L. M. Zintgraf, R. Dazeley, F. Heintz, E. Howley, A. A. Irissappane, P. Mannion, A. Nowé, G. de Oliveira Ramos, M. Restelli, P. Vamplew, and D. M. Roijers, “A practical guide to multi-objective reinforcement learning and planning,” Auton. Agents Multi Agent Syst., vol. 36, no. 1, p. 26, 2022.
- Z. Gábor, Z. Kalmár, and C. Szepesvári, “Multi-criteria reinforcement learning,” in ICML’98: Proc. of the 15th International Conference on Machine Learning. Morgan Kaufmann, 1998, pp. 197–205.
- S. Mannor and N. Shimkin, “The steering approach for multi-criteria reinforcement learning,” in NIPS’01: Proc. of 14th Annual Conference on Neural Information Processing Systems. MIT Press, 2001, pp. 1563–1570.
- S. Natarajan and P. Tadepalli, “Dynamic preferences in multi-criteria reinforcement learning,” in ICML’05: Proc. of the 22nd International Conference on Machine Learning, ser. ACM International Conference Proceeding Series, vol. 119. ACM, 2005, pp. 601–608.
- R. Yang, X. Sun, and K. Narasimhan, “A generalized algorithm for multi-objective reinforcement learning and policy adaptation,” in NeurIPS’19: Proc. of the 2019 Annual Conference on Neural Information Processing Systems 2019, 2019, pp. 14 610–14 621.
- A. Ikenaga and S. Arai, “Inverse reinforcement learning approach for elicitation of preferences in multi-objective sequential optimization,” in ICA’18: Proc. of the 2018 IEEE International Conference on Agents. IEEE, 2018, pp. 117–118.
- P. Vamplew, R. Dazeley, A. Berry, R. Issabekov, and E. Dekker, “Empirical evaluation methods for multiobjective reinforcement learning algorithms,” Mach. Learn., vol. 84, no. 1-2, pp. 51–80, 2011.
- D. M. Roijers, L. M. Zintgraf, and A. Nowé, “Interactive thompson sampling for multi-objective multi-armed bandits,” in ADT’17: Proc. of 5th International Conference on Algorithmic Decision Theory, J. Rothe, Ed., vol. 10576. Springer, 2017, pp. 18–34.
- D. M. Roijers, L. M. Zintgraf, P. Libin, and A. Nowé, “Interactive multi-objective reinforcement learning in multi-armed bandits for any utility function,” in ALA workshop at FAIM, vol. 8, 2018.
- M. Peschl, A. Zgonnikov, F. A. Oliehoek, and L. C. Siebert, “MORAL: aligning AI with human norms through multi-objective reinforced active learning,” in AAMAS’22: Proc. of the 21st International Conference on Autonomous Agents and Multiagent Systems, 2022, pp. 1038–1046.
- E. Todorov, T. Erez, and Y. Tassa, “Mujoco: A physics engine for model-based control,” in IROS’12: Proc. 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2012, pp. 5026–5033.
- W. Chiu, H. Sun, and H. V. Poor, “A multiobjective approach to multimicrogrid system design,” IEEE Trans. Smart Grid, vol. 6, no. 5, pp. 2263–2272, 2015.
- D. M. Roijers, P. Vamplew, S. Whiteson, and R. Dazeley, “A survey of multi-objective sequential decision-making,” J. Artif. Intell. Res., vol. 48, pp. 67–113, 2013.
- R. Radulescu, P. Mannion, D. M. Roijers, and A. Nowé, “Multi-objective multi-agent decision making: A utility-based analysis and survey,” in AAMAS’20: Proc. of the 19th International Conference on Autonomous Agents and Multiagent Systems, 2020, pp. 2158–2160.
- M. Rolf, “The need for MORE: need systems as non-linear multi-objective reinforcement learning,” in ICDL-EpiRob’20: Joint IEEE 10th International Conference on Development and Learning and Epigenetic Robotics. IEEE, 2020, pp. 1–8.
- A. Abdolmaleki, S. H. Huang, L. Hasenclever, M. Neunert, H. F. Song, M. Zambelli, M. F. Martins, N. Heess, R. Hadsell, and M. A. Riedmiller, “A distributional view on multi-objective policy optimization,” in ICML’20: Proc. of the 37th International Conference on Machine Learning, ser. Proceedings of Machine Learning Research, vol. 119. PMLR, 2020, pp. 11–22.
- H. Mossalam, Y. M. Assael, D. M. Roijers, and S. Whiteson, “Multi-objective deep reinforcement learning,” CoRR, vol. abs/1610.02707, 2016. [Online]. Available: http://arxiv.org/abs/1610.02707
- M. Reymond and A. Nowé, “Pareto-DQN: Approximating the Pareto front in complex multi-objective decision problems,” in ALA’19: Proc. of the Adaptive and Learning Agents Workshop at AAMAS, 2019.
- S. Parisi, M. Pirotta, N. Smacchia, L. Bascetta, and M. Restelli, “Policy gradient approaches for multi-objective sequential decision making: A comparison,” in ADPRL’14: Proc. of 2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning. IEEE, 2014, pp. 1–8.
- K. Li, T. Zhang, and R. Wang, “Deep reinforcement learning for multiobjective optimization,” IEEE Trans. Cybern., vol. 51, no. 6, pp. 3103–3114, 2021.
- C. Wirth, R. Akrour, G. Neumann, and J. Fürnkranz, “A survey of preference-based reinforcement learning methods,” J. Mach. Learn. Res., vol. 18, pp. 136:1–136:46, 2017.
- D. Amodei, C. Olah, J. Steinhardt, P. F. Christiano, J. Schulman, and D. Mané, “Concrete problems in AI safety,” CoRR, vol. abs/1606.06565, 2016. [Online]. Available: http://arxiv.org/abs/1606.06565
- H. Sugiyama, T. Meguro, and Y. Minami, “Preference-learning based inverse reinforcement learning for dialog control,” in INTERSPEECH’12: Proc. of the 13th Annual Conference of the International Speech Communication Association. ISCA, 2012, pp. 222–225.
- X. Pan and Y. Shen, “Human-interactive subgoal supervision for efficient inverse reinforcement learning,” in AAMAS’18: Proc. of the 17th International Conference on Autonomous Agents and MultiAgent Systems. International Foundation for Autonomous Agents and Multiagent Systems Richland, SC, USA / ACM, 2018, pp. 1380–1387.
- D. S. Brown, W. Goo, P. Nagarajan, and S. Niekum, “Extrapolating beyond suboptimal demonstrations via inverse reinforcement learning from observations,” in ICML’19: Proc. of the 36th International Conference on Machine Learning, ser. Proceedings of Machine Learning Research, vol. 97. PMLR, 2019, pp. 783–792.
- D. S. Brown, W. Goo, and S. Niekum, “Better-than-demonstrator imitation learning via automatically-ranked demonstrations,” in CoRL’19: Proc. of the 3rd Annual Conference on Robot Learning, ser. Proceedings of Machine Learning Research, vol. 100. PMLR, 2019, pp. 330–359.
- M. Kollmitz, T. Koller, J. Boedecker, and W. Burgard, “Learning human-aware robot navigation from physical interaction via inverse reinforcement learning,” in IROS’20: Proc. of 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2020, pp. 11 025–11 031.
- C. Wirth, J. Fürnkranz, and G. Neumann, “Model-free preference-based reinforcement learning,” in AAAI’16: Proc. of the 30th AAAI Conference on Artificial Intelligence. AAAI Press, 2016, pp. 2222–2228.
- P. F. Christiano, J. Leike, T. B. Brown, M. Martic, S. Legg, and D. Amodei, “Deep reinforcement learning from human preferences,” in NIPS’17: Proc. of 2017 Annual Conference on Neural Information Processing Systems, 2017, pp. 4299–4307.
- N. Wanigasekara, Y. Liang, S. T. Goh, Y. Liu, J. J. Williams, and D. S. Rosenblum, “Learning multi-objective rewards and user utility function in contextual bandits for personalized ranking,” in IJCAI’19: Proc. of the 28th International Joint Conference on Artificial Intelligence, 2019, pp. 3835–3841.
- J. Cao, S. Kwong, R. Wang, and K. Li, “AN indicator-based selection multi-objective evolutionary algorithm with preference for multi-class ensemble,” in ICMLC’14: Proc. of the 2014 International Conference on Machine Learning and Cybernetics, 2014, pp. 147–152.
- K. Li, K. Deb, O. T. Altinöz, and X. Yao, “Empirical investigations of reference point based methods when facing a massively large number of objectives: First results,” in EMO’17: Proc. of the 9th International Conference Evolutionary Multi-Criterion Optimization, 2017, pp. 390–405.
- K. Li, R. Chen, G. Min, and X. Yao, “Integration of preferences in decomposition multiobjective optimization,” IEEE Trans. Cybern., vol. 48, no. 12, pp. 3359–3370, 2018.
- J. Zou, C. Ji, S. Yang, Y. Zhang, J. Zheng, and K. Li, “A knee-point-based evolutionary algorithm using weighted subpopulation for many-objective optimization,” Swarm and Evolutionary Computation, vol. 47, pp. 33–43, 2019.
- H. Nie, H. Gao, and K. Li, “Knee point identification based on voronoi diagram,” in SMC’20: Proc. of 2020 IEEE International Conference on Systems, Man, and Cybernetics. IEEE, 2020, pp. 1081–1086.
- R. Chen and K. Li, “Knee point identification based on the geometric characteristic,” in SMC’21: Proc. of 2021 IEEE International Conference on Systems, Man, and Cybernetics. IEEE, 2021, pp. 764–769.
- K. Li, H. Nie, H. Gao, and X. Yao, “Posterior decision making based on decomposition-driven knee point identification,” IEEE Trans. Evol. Comput., vol. 26, no. 6, pp. 1409–1423, 2022.
- K. Li, “Progressive preference learning: Proof-of-principle results in MOEA/D,” in EMO’19: Proc. of the 10th International Conference Evolutionary Multi-Criterion Optimization, 2019, pp. 631–643.
- K. Li, R. Chen, D. A. Savic, and X. Yao, “Interactive decomposition multiobjective optimization via progressively learned value functions,” IEEE Trans. Fuzzy Syst., vol. 27, no. 5, pp. 849–860, 2019.
- K. Li, G. Lai, and X. Yao, “Interactive evolutionary multiobjective optimization via learning to rank,” IEEE Trans. Evol. Comput., vol. 27, no. 4, pp. 749–763, 2023.
- K. Li, M. Liao, K. Deb, G. Min, and X. Yao, “Does preference always help? A holistic study on preference-based evolutionary multiobjective optimization using reference points,” IEEE Trans. Evol. Comput., vol. 24, no. 6, pp. 1078–1096, 2020.
- Q. Zhang and H. Li, “MOEA/D: A multiobjective evolutionary algorithm based on decomposition,” IEEE Trans. Evol. Comput., vol. 11, no. 6, pp. 712–731, 2007.
- K. Li, Á. Fialho, S. Kwong, and Q. Zhang, “Adaptive operator selection with bandits for a multiobjective evolutionary algorithm based on decomposition,” IEEE Trans. Evol. Comput., vol. 18, no. 1, pp. 114–130, 2014.
- K. Li, Q. Zhang, S. Kwong, M. Li, and R. Wang, “Stable matching-based selection in evolutionary multiobjective optimization,” IEEE Trans. Evol. Comput., vol. 18, no. 6, pp. 909–923, 2014.
- K. Li, S. Kwong, Q. Zhang, and K. Deb, “Interrelationship-based selection for decomposition multiobjective optimization,” IEEE Trans. Cybern., vol. 45, no. 10, pp. 2076–2088, 2015.
- M. Wu, S. Kwong, Q. Zhang, K. Li, R. Wang, and B. Liu, “Two-level stable matching-based selection in MOEA/D,” in SMC’15: Proc. of the 2015 IEEE International Conference on Systems, Man, and Cybernetics, 2015, pp. 1720–1725.
- K. Li, K. Deb, Q. Zhang, and S. Kwong, “An evolutionary many-objective optimization algorithm based on dominance and decomposition,” IEEE Trans. Evol. Comput., vol. 19, no. 5, pp. 694–716, 2015.
- K. Li, K. Deb, Q. Zhang, and Q. Zhang, “Efficient nondomination level update method for steady-state evolutionary multiobjective optimization,” IEEE Trans. Cybernetics, vol. 47, no. 9, pp. 2838–2849, 2017.
- M. Wu, K. Li, S. Kwong, Y. Zhou, and Q. Zhang, “Matching-based selection with incomplete lists for decomposition multiobjective optimization,” IEEE Trans. Evol. Comput., vol. 21, no. 4, pp. 554–568, 2017.
- M. Wu, K. Li, S. Kwong, and Q. Zhang, “Evolutionary many-objective optimization based on adversarial decomposition,” IEEE Trans. Cybern., vol. 50, no. 2, pp. 753–764, 2020.
- I. Das and J. E. Dennis, “Normal-boundary intersection: A new method for generating the pareto surface in nonlinear multicriteria optimization problems,” SIAM J. Optim., vol. 8, no. 3, pp. 631–657, 1998.
- M. Wu, S. Kwong, Y. Jia, K. Li, and Q. Zhang, “Adaptive weights generation for decomposition-based multi-objective optimization using gaussian process regression,” in GECCO’17: Proc. of the 2017 Genetic and Evolutionary Computation Conference. ACM, 2017, pp. 641–648.
- M. Wu, K. Li, S. Kwong, Q. Zhang, and J. Zhang, “Learning to decompose: A paradigm for decomposition-based multiobjective optimization,” IEEE Trans. Evol. Comput., vol. 23, no. 3, pp. 376–390, 2019. [Online]. Available: https://doi.org/10.1109/TEVC.2018.2865931
- G. Pruvost, B. Derbel, A. Liefooghe, K. Li, and Q. Zhang, “On the combined impact of population size and sub-problem selection in MOEA/D,” in EvoCOP’20: Proc. of the 20th European Conference on Evolutionary Computation in Combinatorial Optimization, ser. Lecture Notes in Computer Science, vol. 12102. Springer, 2020, pp. 131–147.
- J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” CoRR, vol. abs/1707.06347, 2017. [Online]. Available: http://arxiv.org/abs/1707.06347
- J. Schulman, P. Moritz, S. Levine, M. I. Jordan, and P. Abbeel, “High-dimensional continuous control using generalized advantage estimation,” in ICLR’16: Proc. of the 4th International Conference on Learning Representations, 2016.
- P. Auer, “Using confidence bounds for exploitation-exploration trade-offs,” J. Mach. Learn. Res., vol. 3, pp. 397–422, 2002.
- K. H. Wray, S. Zilberstein, and A. Mouaddib, “Multi-objective mdps with conditional lexicographic reward preferences,” in AAAI’15: Proc. of the 29th AAAI Conference on Artificial Intelligence. AAAI Press, 2015, pp. 3418–3424.
- L. M. Zintgraf, D. M. Roijers, S. Linders, C. M. Jonker, and A. Now, “Ordered preference elicitation strategies for supporting multi-objective decision making,” in AAMAS’18: Proc. of the 17th International Conference on Autonomous Agents and MultiAgent Systems, 2018, pp. 1477–1485.
- W. Chu and Z. Ghahramani, “Preference learning with Gaussian processes,” in ICML’05: Proc. of Proceedings of the Twenty-Second International Conference on Machine Learning, ser. ACM International Conference Proceeding Series, vol. 119. ACM, 2005, pp. 137–144.
- X. Chen, A. Ghadirzadeh, M. Björkman, and P. Jensfelt, “Meta-learning for multi-objective reinforcement learning,” in IROS’19: Proc. of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2019, pp. 977–983.
- R. A. Bradley and M. E. Terry, “Rank analysis of incomplete block designs: I. the method of paired comparisons,” Biometrika, vol. 39, no. 3/4, pp. 324–345, 1952.
- K. Li, K. Deb, and X. Yao, “R-metric: Evaluating the performance of preference-based evolutionary multiobjective optimization using reference points,” IEEE Trans. Evol. Comput., vol. 22, no. 6, pp. 821–835, 2018.
- G. Lai, M. Liao, and K. Li, “Empirical studies on the role of the decision maker in interactive evolutionary multi-objective optimization,” in CEC’21: Proc. of the 2021 IEEE Congress on Evolutionary Computation. IEEE, 2021, pp. 185–192.
- R. Tanabe and K. Li, “Quality indicators for preference-based evolutionary multi-objective optimization using a reference point: A review and analysis,” IEEE Trans. Evol. Comput., 2023, accepted for publication.
- T. Joachims, “Optimizing search engines using clickthrough data,” in SIGKDD’02: Proc. of the 8th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 2002, pp. 133–142.
- R. Chen, K. Li, and X. Yao, “Dynamic multiobjectives optimization with a changing number of objectives,” IEEE Trans. Evol. Comput., vol. 22, no. 1, pp. 157–171, 2018.
- X. Fan, K. Li, and K. C. Tan, “Surrogate assisted evolutionary algorithm based on transfer learning for dynamic expensive multi-objective optimisation problems,” in CEC’20: Proc. of the 2020 IEEE Congress on Evolutionary Computation. IEEE, 2020, pp. 1–8.
- R. Chen and K. Li, “Transfer bayesian optimization for expensive black-box optimization in dynamic environment,” in SMC’21: Proc. of 2021 IEEE International Conference on Systems, Man, and Cybernetics. IEEE, 2021, pp. 1374–1379.
- K. Li, R. Chen, and X. Yao, “A data-driven evolutionary transfer optimization for expensive problems in dynamic environments,” IEEE Trans. Evol. Comput., 2023.
- K. Li, R. Chen, G. Fu, and X. Yao, “Two-archive evolutionary algorithm for constrained multiobjective optimization,” IEEE Trans. Evol. Comput., vol. 23, no. 2, pp. 303–315, 2019.
- X. Shan and K. Li, “An improved two-archive evolutionary algorithm for constrained multi-objective optimization,” in EMO’21: Proc. of the 11th International Conference on Evolutionary Multicriteria Optimization, ser. Lecture Notes in Computer Science, vol. 12654. Springer, 2021, pp. 235–247.
- S. Li, K. Li, and W. Li, “Do we really need to use constraint violation in constrained evolutionary multi-objective optimization?” in PPSN’22: Proc. of the 17th International Conference on Parallel Problem Solving from Nature, ser. Lecture Notes in Computer Science, vol. 13399. Springer, 2022, pp. 124–137.
- S. Wang and K. Li, “Constrained Bayesian optimization under partial observations: Balanced improvements and provable convergence,” in AAAI’24: Proc. of the Thirty-Eighth AAAI Conference on Artificial Intelligence, 2024, accepted for publication. [Online]. Available: https://arxiv.org/abs/2312.03212
- H. Yang and K. Li, “Instoptima: Evolutionary multi-objective instruction optimization via large language model-based instruction operators,” in EMNLP’23: Findings of the Association for Computational Linguistics: EMNLP 2023. Association for Computational Linguistics, 2023, pp. 13 593–13 602.
- R. Chen and K. Li, “Data-driven evolutionary multi-objective optimization based on multiple-gradient descent for disconnected pareto fronts,” in EMO’23: Proc. of the 12th International Conference on Evolutionary Multi-Criterion Optimization, ser. Lecture Notes in Computer Science, vol. 13970. Springer, 2023, pp. 56–70.
- B. Lyu, Y. Yang, S. Wen, T. Huang, and K. Li, “Neural architecture search for portrait parsing,” IEEE Trans. Neural Networks Learn. Syst., vol. 34, no. 3, pp. 1112–1121, 2023.
- B. Lyu, M. Hamdi, Y. Yang, Y. Cao, Z. Yan, K. Li, S. Wen, and T. Huang, “Efficient spectral graph convolutional network deployment on memristive crossbars,” IEEE Trans. Emerg. Top. Comput. Intell., vol. 7, no. 2, pp. 415–425, 2023.
- B. Lyu, L. Lu, M. Hamdi, S. Wen, Y. Yang, and K. Li, “MTLP-JR: multi-task learning-based prediction for joint ranking in neural architecture search,” Comput. Electr. Eng., vol. 105, p. 108474, 2023.
- S. Zhou, K. Li, and G. Min, “Attention-based genetic algorithm for adversarial attack in natural language processing,” in PPSN’22: Proc. of 17th International Conference on Parallel Problem Solving from Nature, ser. Lecture Notes in Computer Science, vol. 13398. Springer, 2022, pp. 341–355.
- ——, “Adversarial example generation via genetic algorithm: a preliminary result,” in GECCO’22: Companion of 2022 Genetic and Evolutionary Computation Conference. ACM, 2022, pp. 469–470.
- P. N. Williams, K. Li, and G. Min, “A surrogate assisted evolutionary strategy for image approximation by density-ratio estimation,” in CEC’23: Proc. of 2023 IEEE Congress on Evolutionary Computation. IEEE, 2023, pp. 1–8.
- ——, “Sparse adversarial attack via bi-objective optimization,” in EMO’23: Proc. of the 12th International Conference on Evolutionary Multi-Criterion Optimization, ser. Lecture Notes in Computer Science, vol. 13970. Springer, 2023, pp. 118–133. [Online]. Available: https://doi.org/10.1007/978-3-031-27250-9_9
- ——, “Black-box adversarial attack via overlapped shapes,” in GECCO ’22: Genetic and Evolutionary Computation Conference, Companion Volume, Boston, Massachusetts, USA, July 9 - 13, 2022. ACM, 2022, pp. 467–468.
- P. N. Williams and K. Li, “Black-box sparse adversarial attack via multi-objective optimisation CVPR proceedings,” in CVPR’23: Proc. of 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE, 2023, pp. 12 291–12 301.
- K. Li, Z. Xiang, and K. C. Tan, “Which surrogate works for empirical performance modelling? A case study with differential evolution,” in CEC’19: Proc. of the 2019 IEEE Congress on Evolutionary Computation, 2019, pp. 1988–1995.
- K. Li, Z. Xiang, T. Chen, S. Wang, and K. C. Tan, “Understanding the automated parameter optimization on transfer learning for cross-project defect prediction: an empirical study,” in ICSE’20: Proc. of the 42nd International Conference on Software Engineering. ACM, 2020, pp. 566–577.
- M. Liu, K. Li, and T. Chen, “DeepSQLi: deep semantic learning for testing SQL injection,” in ISSTA’20: Proc. of the 29th ACM SIGSOFT International Symposium on Software Testing and Analysis. ACM, 2020, pp. 286–297.
- K. Li, Z. Xiang, T. Chen, and K. C. Tan, “BiLO-CPDP: Bi-level programming for automated model discovery in cross-project defect prediction,” in ASE’20: Proc. of the 35th IEEE/ACM International Conference on Automated Software Engineering. IEEE, 2020, pp. 573–584.
- K. Li, H. Yang, and W. Visser, “Danuoyi: Evolutionary multi-task injection testing on web application firewalls,” IEEE Trans. Softw. Eng., 2023, accepted for publication.
- J. Xu, K. Li, M. Abusara, and Y. Zhang, “Admm-based OPF problem against cyber attacks in smart grid,” in SMC’21: Proc. of 2021 IEEE International Conference on Systems, Man, and Cybernetics. IEEE, 2021, pp. 1418–1423.
- J. Xu, K. Li, and M. Abusara, “Preference based multi-objective reinforcement learning for multi-microgrid system optimization problem in smart grid,” Memetic Comput., vol. 14, no. 2, pp. 225–235, 2022.
- J. Billingsley, K. Li, W. Miao, G. Min, and N. Georgalas, “A formal model for multi-objective optimisation of network function virtualisation placement,” in EMO’19: Proc. of 10th International Conference on Evolutionary Multi-Criterion Optimization, ser. Lecture Notes in Computer Science, vol. 11411. Springer, 2019, pp. 529–540.
- ——, “Routing-led placement of vnfs in arbitrary networks,” in CEC’20: Proc. of 2020 IEEE Congress on Evolutionary Computation. IEEE, 2020, pp. 1–8.
- J. Billingsley, W. Miao, K. Li, G. Min, and N. Georgalas, “Performance analysis of SDN and NFV enabled mobile cloud computing,” in GLOBECOM’20: Proc. of 2020 IEEE Global Communications Conference. IEEE, 2020, pp. 1–6.
- J. Billingsley, K. Li, W. Miao, G. Min, and N. Georgalas, “Parallel algorithms for the multiobjective virtual network function placement problem,” in EMO’21: Proc. of 11th International Conference on Evolutionary Multi-Criterion Optimization, ser. Lecture Notes in Computer Science, vol. 12654. Springer, 2021, pp. 708–720.
- J. Cao, S. Kwong, R. Wang, and K. Li, “A weighted voting method using minimum square error based on extreme learning machine,” in ICMLC’12: Proc. of the 2012 International Conference on Machine Learning and Cybernetics, 2012, pp. 411–414.
- K. Li, R. Wang, S. Kwong, and J. Cao, “Evolving extreme learning machine paradigm with adaptive operator selection and parameter control,” International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, vol. supp02, pp. 143–154, 2013.
- K. Li and S. Kwong, “A general framework for evolutionary multiobjective optimization via manifold learning,” Neurocomputing, vol. 146, pp. 65–74, 2014.
- J. Cao, S. Kwong, R. Wang, X. Li, K. Li, and X. Kong, “Class-specific soft voting based multiple extreme learning machines ensemble,” Neurocomputing, vol. 149, pp. 275–284, 2015.
- R. Wang, S. Ye, K. Li, and S. Kwong, “Bayesian network based label correlation analysis for multi-label classifier chain,” Inf. Sci., vol. 554, pp. 256–275, 2021.
- H. Gao, H. Nie, and K. Li, “Visualisation of pareto front approximation: A short survey and empirical comparisons,” in CEC’19: Proc. of the 2019 IEEE Congress on Evolutionary Computation, 2019, pp. 1750–1757.