Journals
  Publication Years
  Keywords
Search within results Open Search
Please wait a minute...
For Selected: Toggle Thumbnails
Traffic signal control algorithm based on overall state prediction and fair experience replay
Zijun MIAO, Fei LUO, Weichao DING, Wenbo DONG
Journal of Computer Applications    2025, 45 (1): 337-344.   DOI: 10.11772/j.issn.1001-9081.2024010066
Abstract131)   HTML2)    PDF (3199KB)(107)       Save

In order to cope with traffic congestion, efficient traffic signal control algorithms have been designed, which can improve the traffic efficiency of vehicles in the existing transportation network significantly. Although deep reinforcement learning algorithms have shown excellent performance in single intersection traffic signal control problems, their application in multi-intersection environments still faces major challenge — the non-stationarity problem caused by the spatiotemporal partial observability generated by Multi-Agent Reinforcement Learning (MARL) algorithm, resulting in that the deep reinforcement learning algorithms cannot guarantee stable convergence. To this end, a multi-intersection traffic signal control algorithm based on overall state prediction and fair experience replay — IS-DQN was proposed. For one thing, to avoid the problem of non-stationarity caused by spatial observability in algorithm, the state space of IS-DQN was expanded by predicting the overall state of multiple intersections based on historical traffic flow information from different lanes. For another, in order to cope with the time partial observability brought by traditional experience replay strategies, a reservoir sampling algorithm was adopted to ensure the fairness of experience replay pool, so as to avoid non-stationary problems in it. Experimental results on three different traffic pressure simulations in complex multi-intersection environments show that under different traffic pressure conditions, especially in low and medium traffic flow conditions, IS-DQN algorithm has lower average vehicle driving time, better convergence performance and convergence stability compared to independent deep reinforcement learning algorithms.

Table and Figures | Reference | Related Articles | Metrics