A multi-objective optimization approach for the virtual coupling train set driving strategy
Junting Lin, Maolin Li, Xiaohui Qiu
Railway Engineering Science ›› 2025, Vol. 33 ›› Issue (2) : 169-191.
A multi-objective optimization approach for the virtual coupling train set driving strategy
This paper presents an improved virtual coupling train set (VCTS) operation control framework to deal with the lack of optimization of speed curves in the traditional techniques. The framework takes into account the temporary speed limit on the railway line and the communication delay between trains, and it uses a VCTS consisting of three trains as an experimental object. It creates the virtual coupling train tracking and control process by improving the driving strategy of the leader train and using the leader–follower model. The follower train uses the improved speed curve of the leader train as its speed reference curve through knowledge migration, and this completes the multi-objective optimization of the driving strategy for the VCTS. The experimental results confirm that the deep reinforcement learning algorithm effectively achieves the optimization goal of the train driving strategy. They also reveal that the intrinsic curiosity module prioritized experience replay dueling double deep Q-network (ICM-PER-D3QN) algorithm outperforms the deep Q-network (DQN) algorithm in optimizing the driving strategy of the leader train. The ICM-PER-D3QN algorithm enhances the leader train driving strategy by an average of 57% when compared to the DQN algorithm. Furthermore, the particle swarm optimization (PSO)-based model predictive control (MPC) algorithm has also demonstrated tracking accuracy and further improved safety during VCTS operation, with an average increase of 37.7% in tracking accuracy compared to the traditional MPC algorithm.
High-speed trains / Virtual coupling / Multi-objective optimization / Deep reinforcement learning / Knowledge transfer / Model predictive control
[1.] |
|
[2.] |
|
[3.] |
|
[4.] |
|
[5.] |
|
[6.] |
|
[7.] |
|
[8.] |
|
[9.] |
|
[10.] |
|
[11.] |
|
[12.] |
|
[13.] |
|
[14.] |
|
[15.] |
|
[16.] |
Liu S, Cao F, Xun J et al (2015) Energy-efficient operation of single train based on the control strategy of ATO. In: 2015 IEEE 18th International Conference on Intelligent Transportation Systems (ITSC), Gran Canaria, pp 2580–2586
|
[17.] |
|
[18.] |
|
[19.] |
|
[20.] |
|
[21.] |
|
[22.] |
|
[23.] |
Mehta P, Meyn S (2009) Q-learning and Pontryagin’s minimum principle. In: Proceedings of the 48h IEEE conference on decision and control (CDC) held jointly with 2009 28th Chinese control conference (CCC), Shanghai, pp 3598–3605
|
[24.] |
Liu Y, Halev A, Liu X (2021) Policy learning with constraints in model-free reinforcement learning: A survey. In: The 30th international joint conference on artificial intelligence (IJCAI), Montreal, pp 4508–4515
|
[25.] |
|
[26.] |
|
[27.] |
Fu P, Gao S, Dong H et al (2018) Speed tracking error and rate driven event-triggered PID control design method for automatic train operation system. In: 2018 Chinese Automation Congress (CAC), Xi’an, pp 2889–2894
|
[28.] |
|
[29.] |
|
[30.] |
|
[31.] |
|
[32.] |
|
[33.] |
|
[34.] |
Bujarbaruah M, Zhang X, Rosolia U et al (2018) Adaptive MPC for iterative tasks. In: 2018 IEEE Conference on Decision and Control (CDC). Miami, pp 6322–6327
|
[35.] |
|
[36.] |
Yin L (2022) High speed train modeling and nonlinear speed tracking control based on disturbance observer. Dissertation, Jilin University (in Chinese)
|
[37.] |
Meng X (2021) Energy-efficient train operation control of automatic driving based on Q learning and deep Q learning. Dissertation, Beijing Jiaotong University (in Chinese)
|
[38.] |
Pathak D, Agrawal P, Efros AA et al (2017) Curiosity-driven exploration by self-supervised prediction. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, pp 2778–2787
|
[39.] |
Schaul T, Quan J, Antonoglou I et al (2015) Prioritized experience replay. In: 4th International Conference on Learning Representations, San Juan, May 2–4
|
[40.] |
Wang Z, Schaul T, Hessel M et al (2016) Dueling network architectures for deep reinforcement learning. In: Proceedings of the 33rd International Conference on International Conference on Machine Learning (ICML), New York, pp 1995–2003
|
[41.] |
|
[42.] |
|
[43.] |
|
[44.] |
|
[45.] |
|
[46.] |
|
[47.] |
|
[48.] |
|
[49.] |
|
[50.] |
Ma Y (2022) Research on the integration of high-speed train operation adjustment and energy-saving control under the condition of road network. Dissertation, Lanzhou Jiaotong University (in Chinese)
|
[51.] |
Long SH (2021) Models and algorithms for the integrated optimization of train rescheduling and train control for high-speed railway. Dissertation, Beijing Jiaotong University (in Chinese)
|
[52.] |
|
[53.] |
|
/
〈 |
|
〉 |