Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Graph Attention-based Deep Reinforcement Learning for solving the Chinese Postman Problem with Load-dependent costs (2310.15516v4)

Published 24 Oct 2023 in cs.LG

Abstract: Recently, Deep reinforcement learning (DRL) models have shown promising results in solving routing problems. However, most DRL solvers are commonly proposed to solve node routing problems, such as the Traveling Salesman Problem (TSP). Meanwhile, there has been limited research on applying neural methods to arc routing problems, such as the Chinese Postman Problem (CPP), since they often feature irregular and complex solution spaces compared to TSP. To fill these gaps, this paper proposes a novel DRL framework to address the CPP with load-dependent costs (CPP-LC) (Corberan et al., 2018), which is a complex arc routing problem with load constraints. The novelty of our method is two-fold. First, we formulate the CPP-LC as a Markov Decision Process (MDP) sequential model. Subsequently, we introduce an autoregressive model based on DRL, namely Arc-DRL, consisting of an encoder and decoder to address the CPP-LC challenge effectively. Such a framework allows the DRL model to work efficiently and scalably to arc routing problems. Furthermore, we propose a new bio-inspired meta-heuristic solution based on Evolutionary Algorithm (EA) for CPP-LC. Extensive experiments show that Arc-DRL outperforms existing meta-heuristic methods such as Iterative Local Search (ILS) and Variable Neighborhood Search (VNS) proposed by (Corberan et al., 2018) on large benchmark datasets for CPP-LC regarding both solution quality and running time; while the EA gives the best solution quality with much more running time. We release our C++ implementations for metaheuristics such as EA, ILS and VNS along with the code for data generation and our generated data at https://github.com/HySonLab/Chinese_Postman_Problem

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Neural combinatorial optimization with reinforcement learning. arXiv preprint arXiv:1611.09940.
  2. Learning generalizable models for vehicle routing problems via knowledge distillation. Advances in Neural Information Processing Systems, 35:31226–31238.
  3. Combinatorial optimization and reasoning with graph neural networks. In Zhou, Z.-H., editor, Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, pages 4348–4355. International Joint Conferences on Artificial Intelligence Organization. Survey Track.
  4. An algorithm for the rural postman problem. Report IC. OR, 81:81.
  5. Distributed optimization by ant colonies. In Proceedings of the first European conference on artificial life, volume 142, pages 134–142. Paris, France.
  6. Arc routing problems: A review of the past, present, and future. Networks, 77(1):88–115.
  7. The chinese postman problem with load-dependent costs. Transportation Science, 52(2):370–385.
  8. Discriminative embeddings of latent variable models for structured data. In International conference on machine learning, pages 2702–2711. PMLR.
  9. Reinforcement learning with combinatorial actions: An application to vehicle routing. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 609–620. Curran Associates, Inc.
  10. Ant colony optimization. IEEE Computational Intelligence Magazine, 1(4):28–39.
  11. Ant colony system: a cooperative learning approach to the traveling salesman problem. IEEE Transactions on Evolutionary Computation, 1(1):53–66.
  12. Ant system: optimization by a colony of cooperating agents. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 26(1):29–41.
  13. Floyd, R. W. (1962). Algorithm 97: Shortest path. Commun. ACM, 5(6):345.
  14. Improvement procedures for the undirected rural postman problem. INFORMS Journal on computing, 11(1):53–62.
  15. Graph neural network guided local search for the traveling salesperson problem. In International Conference on Learning Representations.
  16. Multi-view graph contrastive learning for solving vehicle routing problems. In Evans, R. J. and Shpitser, I., editors, Proceedings of the Thirty-Ninth Conference on Uncertainty in Artificial Intelligence, volume 216 of Proceedings of Machine Learning Research, pages 984–994. PMLR.
  17. Learning to solve routing problems via distributionally robust optimization. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 9786–9794.
  18. An efficient graph convolutional network technique for the travelling salesman problem. arXiv preprint arXiv:1906.01227.
  19. Learning combinatorial optimization algorithms over graphs. Advances in neural information processing systems, 30.
  20. Learning collaborative policies to solve np-hard routing problems. Advances in Neural Information Processing Systems, 34:10418–10430.
  21. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
  22. Actor-critic algorithms. In Solla, S., Leen, T., and Müller, K., editors, Advances in Neural Information Processing Systems, volume 12. MIT Press.
  23. Attention, learn to solve routing problems! In International Conference on Learning Representations.
  24. Pomo: Policy optimization with multiple optima for reinforcement learning. Advances in Neural Information Processing Systems, 33:21188–21198.
  25. Learning to delegate for large-scale vehicle routing. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W., editors, Advances in Neural Information Processing Systems, volume 34, pages 26198–26211. Curran Associates, Inc.
  26. Attention routing: track-assignment detailed routing using attention-based reinforcement learning. In International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, volume 84003, page V11AT11A002. American Society of Mechanical Engineers.
  27. Evolution algorithms in combinatorial optimization. Parallel computing, 7(1):65–85.
  28. Reinforcement learning for solving the vehicle routing problem. In Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc.
  29. Potvin, J.-Y. (2009). State-of-the art review—evolutionary algorithms for vehicle routing. INFORMS Journal on computing, 21(4):518–548.
  30. Prins, C. (2004). A simple and effective evolutionary algorithm for the vehicle routing problem. Computers & operations research, 31(12):1985–2002.
  31. Attention is all you need. Advances in neural information processing systems, 30.
  32. Pointer networks. Advances in neural information processing systems, 28.
  33. Williams, R. J. (1992). Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine learning, 8:229–256.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Truong Son Hy (28 papers)
  2. Cong Dao Tran (3 papers)
Github Logo Streamline Icon: https://streamlinehq.com