WO2011157745A1 - Decentralised autonomic system and method for use in an urban traffic control environment - Google Patents
Decentralised autonomic system and method for use in an urban traffic control environment Download PDFInfo
- Publication number
- WO2011157745A1 WO2011157745A1 PCT/EP2011/059926 EP2011059926W WO2011157745A1 WO 2011157745 A1 WO2011157745 A1 WO 2011157745A1 EP 2011059926 W EP2011059926 W EP 2011059926W WO 2011157745 A1 WO2011157745 A1 WO 2011157745A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- agent
- learning
- action
- agents
- traffic
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/07—Controlling traffic signals
- G08G1/081—Plural intersections under common control
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/07—Controlling traffic signals
- G08G1/08—Controlling traffic signals according to detected number or speed of vehicles
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/07—Controlling traffic signals
- G08G1/087—Override of traffic control, e.g. by signal transmitted by an emergency vehicle
Definitions
- the invention relates to the field of decentralized autonomic systems, and specifically to urban traffic control systems .
- UTC Urban Traffic Control
- TRANSIT TRL, Transit Research Lab in the UK
- MAXBAND US Department of Transport
- Such UTC systems usually consist of several different fixed-time plans designed for morning peak, midday, afternoon peak, and evening/night-time conditions.
- special plans may be produced, for example, for recurring major music or sporting events.
- the major disadvantage of fixed-time plans is that they are rarely kept up to date due to the complexity and duration of the design process for the new plans and that they are not able to deal well with fluctuations in the patterns .
- Widely deployed adaptive traffic control techniques include UTMC (Federal Highway Administration USA) , SCATS (Roads and Traffic Authority of New South Wales, Australia) and SCOOT (Transport and Road Research Laboratory UK, Siemens Traffic Control, Peek Traffic) .
- UTMC initially used fixed-time offline design, but later versions use on-line comparison of traffic data and historical data to re-evaluate the current plan every 15 minutes.
- SCATS and SCOOT also provide online adaptation with signal duration being adjusted at every cycle. Both approaches primarily rely on the use of induction loops to estimate traffic counts and adjust the duration of the signal accordingly, showing significant improvements over fixed- time controllers.
- the main disadvantages of these systems are their centralized or hierarchical control, which limits the amount of data that can be processed in real-time, in turn limiting the accuracy of adaptation, as well as reliance on expensive to install and maintain induction loops.
- These systems also require significant manual pre- configuration (i.e. selecting the traffic phases to be deployed, grouping junctions into subsystems etc.) as well as significant expertise and cost to configure and operate.
- Petri nets are used to model each junction in a simulation of a UTC system, representing vehicle flows entering the junction, leaving the junction, and a traffic light (TL) controller.
- the TL control system consists of a local controller and a priority controller on each junction, and a global supervisor. Each local controller aims to minimize the traffic queues and equalize queue lengths across a junction's approaches.
- an emergency vehicle When an emergency vehicle enters the system, it notifies the global controller, which calculates the shortest path (in terms of waiting time) for an emergency vehicle to take. It then notifies all of the local junction controllers on the path of the time at which an emergency vehicle is estimated to reach them. Based on this information, the local priority controller can either extend the current green signal or shorten future red signals, to ensure an approach with an emergency vehicle receives a green light.
- traffic light controllers act independently and do not cooperate with their neighbours, therefore not accounting for potential agent dependencies.
- Japanese Patent publication number JP 09160605 A assigned to Omron, addresses the issue of traffic lights exchanging their status, however, it does not address how the information exchanged is used in the presence of multiple policies or if and how is it used to address learning the dependencies between multiple policies, multiple junctions, or learning of suitable degrees of cooperation between junctions.
- Single-policy multi-agent approaches as well as multi- policy single-agent approaches, have limited use in real- world application, and in urban traffic control in particular, as the systems consist of multiple, often hundreds of traffic lights (agents) and the traffic lights need to deal with multiple policies, e.g., optimization of global traffic flow, honoring pedestrian requests, prioritizing public transport vehicles, or prioritize emergency vehicles.
- Learning and exchange of status techniques from single-policy multi-agent Q-learning approaches are not easily applicable to multiple policies, and require new approaches and techniques when multiple heterogeneous, potentially conflicting, policies are present.
- W-learning multi-policy single-agent learning
- existing traffic control systems suffer from a number of problems in that they are relatively unsophisticated, and even though they provide a degree of adaptivity, they require a large amount of configuration and manual intervention, and rely on a limited number of, often unreliable sensors, such as induction loops.
- many of the proposed UTC systems focus on optimization of traffic only towards a single traffic policy (e.g., only optimizing general traffic flow without prioritizing public transport) and do not learn the effect that neighbouring junctions have on one another but tend to cooperate and exchange status with a predefined set of neighbouring junctions, regardless of the degree of influence that the junctions might have on one another.
- An object of the present invention is to provide an urban traffic control system and method to overcome at least one of the above mentioned problems and shortcomings of existing prior art systems.
- a system of agents for use in an Urban Traffic Control environment each agent representing a traffic light controller at a traffic junction to control traffic flow, said system comprising: each agent is adapted to collect data local to the junction using one or more sensors and applying a Q learning reinforcement learning model to said collected data, one Q-learning model per each policy that the agent is implementing;
- each agent comprises:
- the exchanged values used to learn preferences of policies implemented by other agents comprises means for using remote policy learning.
- the main operational advantage of the system of the invention is that it utilizes machine learning to learn appropriate behaviours for a variety of traffic conditions, in a fully decentralized distributed self-organizing approach where data collection and analysis is performed by the junctions or intersections locally. It removes the need for extensive pre-configuration, as the agents or nodes can configure themselves based on the observed conditions and learnt behaviours, reducing the configuration, deployment, and operational time and costs. It also enables timely analysis of large amounts of sensor data and determination of the current traffic conditions to deploy learnt optimal signal sequences for that given set of conditions .
- each junction can automatically learn dependencies between neighbouring junctions, i.e., the effect of one junction's traffic light settings on another for a particular set of traffic conditions, removing the need for manual analysis, as well as learning when and how much should junctions take into account neighbouring junctions' preferences when selecting signal settings, by using cooperation coefficient learning.
- the invention provides a system and method for supporting simultaneous optimization on multiple junctions towards multiple performance policies, which enable junctions to learn the dependencies between their own performance and the performance of other junctions, and enable junctions to learn to what degree they should collaborate with other junctions to ensure optimal system performance.
- the system and method makes use of remote policies, which instead of directly using exchanged statuses, enables agents to learn each other's action preferences, and in such a way to enable collaboration.
- each junction uses Q-learning and W- learning, together with remote learning and learning of the cooperation coefficients to provide a distributed W- learning model and to obtain said action preferences. Learning can be utilised to learn the most optimal levels of collaboration between multiple agents .
- the Q-learning reinforcement learning model receives information on current traffic conditions from at least one sensor, and maps that information to one of the available system-state representations.
- the Q-learning reinforcement learning model learns an action to implement the most suitable traffic light sequences in the long-term for the given traffic conditions.
- each agent comprises a set of policies such that each junction is adapted to learn a preferred action to be executed, in parallel with the importance of that action depending on the system state using said distributed W-Learning model.
- each junction combines local Q-learning and W-learning, with remote learning for neighbouring junctions, to provide a distributed W-learning model in order to obtain said action preferences.
- actions are traffic light settings and states can encode any information internally defined as relevant for the decision, e.g., number of cars approaching the junction, or number of buses waiting at approach.
- each local and each remote policy on at least one agent is adapted to decide an action for execution at the next time step, based on the importance of executing that action in the current states of all local and remote policies .
- Q-learning and W- learning data of neighbouring junctions to obtain data necessary for remote learning of the action preferences from the immediate upstream and/or downstream junctions.
- an agent for use in an Urban Traffic Control environment said agent representing a traffic-light controller at a traffic junction to control traffic flow, and adapted to collect data local to the junction using one or more sensors and applying a Q-learning reinforcement learning model to said collected data, one Q-learning model per each policy that the agent is implementing;
- each agent comprises:
- a method for controlling an Urban Traffic Control environment comprising a plurality of agents, each agent representing a traffic light controller at a traffic junction to control traffic flow, the method comprising the steps of:
- a computer program comprising program instructions for causing a computer program to carry out the above method which may be embodied on a recording medium, carrier signal or read-only memory.
- Figure 1 illustrates an overall architecture of the system showing a group of traffic control signals at connected intersections each controlled by a single agent ;
- Figure 2 illustrates agent interaction required for a decision making process on a single distributed W- Learning (DWL) agent for action selection;
- DWL distributed W- Learning
- Figure 3 illustrates an implementation of the DWL action-selection process on a single agent, according to one aspect of the invention
- Figures 4, 5 and 6 illustrate the implementation of the learning process that each DWL agent performs to determine the most suitable cooperation coefficient (C) , according to other aspects of the invention.
- the invention provides a fully self-organizing Urban Traffic Control (UTC) system that uses reinforcement learning (RL) to map the currently-observed traffic conditions (based on the information received from the available road and/or in-car sensor technology) to appropriate traffic light sequences.
- UTC Urban Traffic Control
- RL reinforcement learning
- Such a UTC system is enabled through use of a novel multi-policy multi-agent optimization algorithm, Distributed W-Learning (DWL), which is using RL techniques Q-learning and W-learning, remote learning, and cooperation coefficient learning and described in more detail below.
- DWL Distributed W-Learning
- each junction implements a Q-learning RL process model whereby it receives information on current traffic conditions from available sensors, maps that information to one of the available system state representations, and executes the action (set of traffic light sequences) that is has learnt to be the most suitable in the long-term for the given traffic conditions.
- a DWL agent/ unction learns a preferred action to be executed, as well as the importance of that action in its current system state. For example, for a policy that prioritizes public transport, action importance is high if it has two buses queuing on its approaches with both of them being behind schedule, while it has a very low importance if there are no public transport vehicles approaching.
- Figure 1 shows the overall architecture of the system.
- a group of traffic control signals at connected intersections are each controlled by a single agent implementing DWL .
- the overall system consists of a group of such junctions, each junction hereinafter referred to as an agent.
- Each agent is implementing a DWL process to learn and execute the most suitable actions for its own performance goals/policies as well as the goals/policies of neighbouring (all upstream and all downstream) agents.
- DWL agent Al implements policy pi (e.g., optimize general traffic flow)
- DWL agent A2 implements policies pi (it is also optimizing general traffic flow) and p2 (e.g., prioritize buses)
- DWL agent A3 also implements policy p2 (it is also prioritizing buses) .
- A2 learns the best local traffic-control signal settings to meet its performance policies pi and p2, but also, in collaboration with Al and A3, learns the best local settings to enable Al and A3 to meet their performance policies pi, p2 locally.
- the division into agents can also be physical, i.e., the software that implements DWL behaviours can be running on the traffic-control signals themselves, but all software agents can also be running in the same central location.
- Figure 2 depicts the agent interaction required for a decision making process on a single DWL agent.
- Each agent in the system performs interaction as described in Figure 2 with each of its neighbouring agents, for each of its local policies and for each of its neighbours' policies.
- each agent exchanges with its neighbours a list of its local policies and associated states.
- Each agent initiates its remote policies which perform remote learning, one policy for each corresponding local policy of each of its neighbours.
- each local policy on each agent observes its local environmental conditions, maps them to a state representation, and performs updates on its local processes that learn the optimal actions and the importance of executing the preferred actions in each state.
- Each local policy suggests an action for execution at the next time step, together with the associated W-value (importance) of the policy's current state.
- Each agent also receives from each of its neighbours' state information (a representation of the neighbours' environment conditions) for each of its policies, and based on that information performs updates on its remote processes that learn the optimal actions and the importance of executing the preferred actions in each particular state.
- Each remote policy suggests an action for execution at the next time step, together with the associated W-value (importance) for the policy's current state.
- This decision process is described in more detail in Figure 3 below.
- each DWL agent Before making action decisions, each DWL agent also updates its learning process that learns the optimal value of Cooperation Coefficient (C) for the neighbourhood, i.e., for all its local and remote policies. This step is optional, as agents can either use predefined C, or learn C during the system operation. The process of learning C is described in more detail in Figure 4.
- Figure 3 depicts the implementation of the DWL action- decision process on a single agent, according to one aspect of the invention.
- Figure 3 shows an agent A2, with its local policies pi and p2.
- A2 has 2 neighbouring junctions, Al, which implements policy pi, and A3, which implements policy p2. Therefore, agent A2 learns how to best meet its local policies pi and p2 (lpl, lp2), but also learns how to best help neighbours meet their policies, or so-called remote policies, rpll (policy pi on agent Al ) and rp32 (policy p2 on agent A3).
- each agent makes a decision as to what actions to execute (i.e., what traffic-control signal settings to deploy) based on optimal actions learnt for its local and remote policies.
- Each policy both local and remote, suggests an action, as selected based on the outcome of the ongoing Q-learning processes associated with each policy.
- Each action is also associated with its current importance, expressed as a learnt W-value, learnt as the outcome of the ongoing W- learning processes associated with each policy.
- C is introduced to enable a local agent to give a varying degree of importance to the neighbours' action preferences.
- C can be predefined or can be learnt at runtime, using a learning process which aims to maximize the rewards received on all agent's local and remote policies .
- FIGS 4, 5, and 6 illustrate the implementation of the learning process that each DWL agent performs to learn the most suitable cooperation coefficient (C) .
- C enables agents to take into account action preferences of their neighbours' policies with varying weight, which can be equal to or less than their local action preferences.
- each local and each remote learning policy on a D L agent suggest an action for execution at the next time step, together with an associated W-value, representing the importance of executing that action in the policy's current state.
- An agent takes the suggestions of its local policies with their full weight, but it has the option to scale the weight of remote policy suggestions (i.e., actions suitable for its neighbours) in order to give higher priority to its local preferences.
- each DWL agent is enabled to learn or determine the most suitable value of C, so that resulting behaviour is optimal for the neighbourhood, i.e., for an agent itself and for all the policies that all of its one-hop neighbours are implementing. Determining C is implemented as a learning process on each agent.
- a learning process selects one of the C values and the agent multiplies all W- values received from remote policies with that value, for example as shown in Figure 4.
- Figure 5 shows the action associated with the state with the highest current W-value is executed, and the outcome of that action is observed on all local policies and on all policies on all one-hop neighbours, i.e., the rewards received by all local policies and by all policies on all one-hop neighbours are added up and used to update the value of C used in the last time step using a learning process .
- Figure 6 illustrates how the value of C can be learnt per agent, i.e., an agent uses the same C to scale W-values received for all policies on all one-hop neighbours; can be learnt per agent pair, i.e., an agent has as many Q- learning processes as it has neighbours, and learns a separate C to scale W-values received from each neighbour; or can be learnt per remote policy, i.e., the number of Q- learning processes that are learning C on an agent is the sum of the number of all policies that all of its one-hop neighbours implement, agent learns a separate C to scale W- values received from each policy on each neighbour.
- Figure 4 shows an example where a single C is learnt on an agent
- Figure 6 shows an example where a different C is learnt for each remote policy.
- the DWL technique combines local Q-learning and W-learning, with remote learning for its neighbouring junctions, to obtain action preferences not just from its local policies, but also from the immediate upstream and downstream junctions.
- Each agent can be adapted to obtain action preferences from all of its one-hop neighbours, for example from neighbours multiple hops away or only from a subset of one-hop or multi-hop neighbours . It is envisaged that agents can be adapted to collaborate with only downstream or only upstream neighbours, depending on the application required .
- junctions coordinate with their upstream and downstream neighbours to execute actions that are most appropriate for the traffic conditions not just locally, but for the immediate neighbourhood as well.
- the priority of different policies is easily incorporated into DWL, as if a policy is given a higher reward in the RL process design, it will have a higher importance compared to a lower-priority policy, enabling easy integration of public vehicle and emergency vehicle priority in a UTC system.
- DWL includes the cooperation coefficient C.
- C has a value between 0 and 1, where 0 denotes a non-collaborative junction, i.e., a junction that always executes an action preferred by its local policies, and 1 denotes a fully collaborative junction, i.e., a junction that gives as much weight to neighbours' preferences as to its local preferences when making an action decision.
- C can be predefined (to make particular junctions more/less cooperative based on their importance in the system) or can be learnt by each junction so as to maximize the reinforcement learning reward locally and in its one-hop neighbourhood.
- the system can function as follows, each junction/agent implements a Q-learning reinforcement learning process whereby it receives information on current traffic conditions from available sensors. This information is then mapped to one of the available system state representations, and an action is executed (set of traffic light sequences) that is has learnt to be the most suitable in the long-term for the given traffic conditions.
- a junction/agent learns a preferred action to be executed, as well as the importance of that action in its current system state using W-Learning.
- the overall operation of this embodiment can be summarised in that Distributed W-Learning combines local Q-learning and W-learning, with remote learning on behalf of its neighbouring junctions/agents, and learning of cooperation coefficients, to obtain action preferences not just from its local policies, but also from all its one-hop neighbours, in order to improve traffic flow efficiency.
- the present invention allows increasingly available road-side and in-car sensor technology, as well as car-to-infrastructure and car-to-car communication capabilities to be utilized to enable UTC systems to make more informed, quicker adaptation decisions.
- UTC systems based on distributed w-learning performance is increased and response time to changing traffic conditions is decreased, as well as a reduction in operating costs in traffic control centres, including both human and hardware cost.
- junctions are capable of learning the dependencies between each other's performance and capable of collaboration to help improve not just their own, but each other's performance, and therefore performance of the overall system.
- the embodiments in the invention described with reference to the drawings comprise a computer apparatus and/or processes performed in a computer apparatus to control each agent according to the invention.
- the invention also extends to computer programs, particularly computer programs stored on or in a carrier adapted to bring the invention into practice and for controlling the agent.
- the program may be in the form of source code, object code, or a code intermediate between source and object code, such as in partially compiled form or in any other form suitable for use in the implementation of the method according to the invention.
- the carrier may comprise a storage medium such as ROM, e.g. CD ROM, or magnetic recording medium, e.g. a floppy disk or hard disk.
- the carrier may be an electrical or optical signal which may be transmitted via an electrical or an optical cable or by radio or other means.
- the 'agent' hereinbefore described with respect to the invention can be incorporated in an existing junction node or new junction node in hardware or software or a combination of both.
- the invention is not limited to the embodiments hereinbefore described but may be varied in both construction and detail.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Traffic Control Systems (AREA)
Abstract
The invention provides a system of agents for use in an Urban Traffic Control environment, each agent representing a traffic light controller at a traffic junction to control traffic flow, said system comprising each agent is adapted to collect data local to the junction using one or more sensors and applying a Distributed W-Learning model to said collected data; each agent comprises means for mapping locally collected data to one of the system state representations available to determine action values; each agent is adapted to determine a current importance value of an action value using a Distributed W-Learning model. The main operational advantage of the system of the invention is that it utilizes machine learning to learn appropriate behaviours for a variety of traffic conditions, in a fully decentralized distributed self-organizing approach capable of addressing multiple performance policies on multiple agents, where data collection and analysis is performed by the junctions or intersections locally.
Description
Title
Decentralised Autonomic System and Method for use in an Urban Traffic Control Environment Field of the Invention
The invention relates to the field of decentralized autonomic systems, and specifically to urban traffic control systems . Background to the Invention
Existing Urban Traffic Control (UTC) approaches fall into two main categories: fixed-time and adaptive traffic controllers. In fixed-time systems, selection of traffic light sequences and their duration is designed offline using specialized programs such as TRANSIT (TRL, Transit Research Lab in the UK) and MAXBAND (US Department of Transport) . Such UTC systems usually consist of several different fixed-time plans designed for morning peak, midday, afternoon peak, and evening/night-time conditions. In addition, special plans may be produced, for example, for recurring major music or sporting events. The major disadvantage of fixed-time plans is that they are rarely kept up to date due to the complexity and duration of the design process for the new plans and that they are not able to deal well with fluctuations in the patterns .
Widely deployed adaptive traffic control techniques include UTMC (Federal Highway Administration USA) , SCATS (Roads and Traffic Authority of New South Wales, Australia) and SCOOT (Transport and Road Research Laboratory UK, Siemens Traffic Control, Peek Traffic) . UTMC initially used fixed-time offline design, but later versions use on-line comparison
of traffic data and historical data to re-evaluate the current plan every 15 minutes.
SCATS and SCOOT also provide online adaptation with signal duration being adjusted at every cycle. Both approaches primarily rely on the use of induction loops to estimate traffic counts and adjust the duration of the signal accordingly, showing significant improvements over fixed- time controllers. The main disadvantages of these systems are their centralized or hierarchical control, which limits the amount of data that can be processed in real-time, in turn limiting the accuracy of adaptation, as well as reliance on expensive to install and maintain induction loops. These systems also require significant manual pre- configuration (i.e. selecting the traffic phases to be deployed, grouping junctions into subsystems etc.) as well as significant expertise and cost to configure and operate.
There are also a number of prototype UTC systems currently being tested that aim to provide more reactive real-time adaptivity. The most significant of those are RHODES (University of Arizona), OPAC (University of Massachusetts) and RTACL (University of Pittsburgh/University of Maryland) . OPAC and RHODES attempt to predict future traffic conditions at the intersection and network level, and show some improvements in prototype testing over existing strategies, however there are concerns about their applicability being limited only to main arteries in under- saturated conditions. RTACL uses distributed control where most decision making is done at the local level and by communicating with neighbouring intersections, however, in field tests RTACL shows inconsistencies in performance by
improving travel times along certain routes, while significantly degrading the service on others.
Numerous scientific works also address the issue of UTC using various multi-agent and learning techniques. In a paper published by Bazzan, A. L. (2005). "A distributed approach for coordination of traffic signal agents", Autonomous Agents and Multi-Agent Systems, 10 (1), 131-164, evolutionary game theory is used to model individual traffic controllers as agents that are capable of sensing their local environment and learning optimal parameters for continually changing traffic patterns. Agents receive both local reinforcement from their local detectors, and global reinforcement based on global traffic conditions. For example, if the majority of traffic travels westbound, agents receive higher payoffs for giving longer green signals to that direction. Global matrices of payoffs need to be specified by the designer of the system for each set of traffic conditions and as such require domain knowledge to construct. This work does not address multiple traffic policies but only optimizing global traffic flow, and does not learn the dependencies between agents' performances.
In another published paper by Febbraro et al . (2004) Febbraro, A. D., Giglio, D., & Sacco, N. (2004). "Urban traffic control structure based on hybrid Petri nets", IEEE Transactions on Intelligent Transportation Systems, 5 (4), 224-237, Petri nets are used to model each junction in a simulation of a UTC system, representing vehicle flows entering the junction, leaving the junction, and a traffic light (TL) controller. The TL control system consists of a local controller and a priority controller on each junction, and a global supervisor. Each local controller
aims to minimize the traffic queues and equalize queue lengths across a junction's approaches. When an emergency vehicle enters the system, it notifies the global controller, which calculates the shortest path (in terms of waiting time) for an emergency vehicle to take. It then notifies all of the local junction controllers on the path of the time at which an emergency vehicle is estimated to reach them. Based on this information, the local priority controller can either extend the current green signal or shorten future red signals, to ensure an approach with an emergency vehicle receives a green light. In this publication traffic light controllers act independently and do not cooperate with their neighbours, therefore not accounting for potential agent dependencies.
Richter, S. (2006). "Learning traffic control - towards practical traffic control using policy gradients", Tech. rep., Albert-Ludwigs-Universitat Freiburg discloses using a reinforcement learning (RL) algorithm to optimize average vehicle travel time. Each intersection implements an RL process that receives a reward based on the number of vehicles it allows to proceed. TL agents communicate with their immediate neighbours, in order to use neighbours traffic counts to anticipate their own traffic flows. This work does not address multiple traffic policies but rather is focused on optimizing global traffic flow and does not learn the dependencies between agents' performances.
A number of existing patent publications attempt to address the use of Q-learning in optimizing the traffic flow, for example US 7047224 B, assigned to Siemans, WO 01/86610 A, assigned to Siemans and DE 4436339 A, assigned to IFU GMBH. However, these patent publications do not address the issue
of balancing the multiple policies that traffic systems need to implement, nor the issue of learning the dependencies between the performance of different policies, the dependencies between different junctions, or the degree of agent collaboration required to address these dependencies . Japanese Patent publication number JP 09160605 A, assigned to Omron, addresses the issue of traffic lights exchanging their status, however, it does not address how the information exchanged is used in the presence of multiple policies or if and how is it used to address learning the dependencies between multiple policies, multiple junctions, or learning of suitable degrees of cooperation between junctions. Salkham (2008) in the paper "A collaborative reinforcement learning approach to urban traffic control optimization", IAT 2008, 560-566. also addresses the issue of junctions exchanging their status and Q-learning parameters, however, the exchanged messages are used only for collaborative optimization towards a single system policy.
In the work published by Humphrys in 1995 titled "W- learning: competition among selfish Q-learners", use of W- learning and Q-learning for optimization of multiple policies has been discussed, however the work addresses only a single agent and therefore does not address any of the multi-agent issues (dependencies or degree of collaboration arising) , nor it is applied in a traffic control setting.
Single-policy multi-agent approaches, as well as multi- policy single-agent approaches, have limited use in real- world application, and in urban traffic control in
particular, as the systems consist of multiple, often hundreds of traffic lights (agents) and the traffic lights need to deal with multiple policies, e.g., optimization of global traffic flow, honouring pedestrian requests, prioritizing public transport vehicles, or prioritize emergency vehicles. Learning and exchange of status techniques from single-policy multi-agent Q-learning approaches are not easily applicable to multiple policies, and require new approaches and techniques when multiple heterogeneous, potentially conflicting, policies are present. Due to the nature of Q-learning, the suitability of actions is learnt specific to given state-action pairs; Q-learning processes implementing different policies will not have matching state-action pairs therefore rendering exchange of status useless as receiving agents will not be capable of interpreting that information or using it for collaboration or optimization.
Similarly, techniques from multi-policy single-agent learning (W-learning) are not easily applicable to multiple agents and require new approaches and techniques when multiple heterogeneous agents are present. Due to the nature of W-learning, the importance of an action is learnt specific to a policy state, and different agents implementing different policies will not have matching policy states, therefore rendering exchange of status useless as receiving agents will not be capable of interpreting that information or using it for collaboration or optimization.
In summary, existing traffic control systems suffer from a number of problems in that they are relatively unsophisticated, and even though they provide a degree of
adaptivity, they require a large amount of configuration and manual intervention, and rely on a limited number of, often unreliable sensors, such as induction loops. Furthermore, many of the proposed UTC systems focus on optimization of traffic only towards a single traffic policy (e.g., only optimizing general traffic flow without prioritizing public transport) and do not learn the effect that neighbouring junctions have on one another but tend to cooperate and exchange status with a predefined set of neighbouring junctions, regardless of the degree of influence that the junctions might have on one another.
An object of the present invention is to provide an urban traffic control system and method to overcome at least one of the above mentioned problems and shortcomings of existing prior art systems.
Summary of the Invention
According to the present invention there is provided, as set out in the appended claims, a system of agents for use in an Urban Traffic Control environment, each agent representing a traffic light controller at a traffic junction to control traffic flow, said system comprising: each agent is adapted to collect data local to the junction using one or more sensors and applying a Q learning reinforcement learning model to said collected data, one Q-learning model per each policy that the agent is implementing;
each agent comprises:
means for mapping locally collected data to one or more system-state representations available to determine action values;
means to learn a current importance value of an action relative to actions preferred by other policies on an agent, using a W-Learning model, and determine the dependencies between locally implemented policies; means for sharing Q-learning and W-learning parameters, for example, current system-state representations, environment feedback, action values and associated importance values with neighbouring agents, one-hop from each neighbouring agent;
means to receive and interpret the shared values to remotely learn preferences of policies implemented by other agents, and their relative importance, to learn the dependencies between the policies implemented locally and those implemented by neighbouring agents; and
means to combine local action-effectiveness and importance values with action-effectiveness and importance values of one-hop neighbouring agents in a periodic manner over a number of time steps to optimise the traffic flow at each agent, taking into account neighbour agents preferred actions and importance values, at each time step.
In one embodiment the exchanged values used to learn preferences of policies implemented by other agents comprises means for using remote policy learning.
In one embodiment there is provided means to determine a cooperation coefficient using a learning model and using said cooperation coefficient to scale action preferences of neighbour agents so as to maximize the performance locally and in the immediate agent neighbourhood.
The main operational advantage of the system of the invention is that it utilizes machine learning to learn appropriate behaviours for a variety of traffic conditions, in a fully decentralized distributed self-organizing approach where data collection and analysis is performed by the junctions or intersections locally. It removes the need for extensive pre-configuration, as the agents or nodes can configure themselves based on the observed conditions and learnt behaviours, reducing the configuration, deployment, and operational time and costs. It also enables timely analysis of large amounts of sensor data and determination of the current traffic conditions to deploy learnt optimal signal sequences for that given set of conditions . Using remote learning, each junction can automatically learn dependencies between neighbouring junctions, i.e., the effect of one junction's traffic light settings on another for a particular set of traffic conditions, removing the need for manual analysis, as well as learning when and how much should junctions take into account neighbouring junctions' preferences when selecting signal settings, by using cooperation coefficient learning.
The invention provides a system and method for supporting simultaneous optimization on multiple junctions towards multiple performance policies, which enable junctions to learn the dependencies between their own performance and the performance of other junctions, and enable junctions to learn to what degree they should collaborate with other junctions to ensure optimal system performance. The system and method makes use of remote policies, which instead of directly using exchanged statuses, enables agents to learn each other's action preferences, and in such a way to enable collaboration.
In one embodiment each junction uses Q-learning and W- learning, together with remote learning and learning of the cooperation coefficients to provide a distributed W- learning model and to obtain said action preferences. Learning can be utilised to learn the most optimal levels of collaboration between multiple agents .
In one embodiment the Q-learning reinforcement learning model receives information on current traffic conditions from at least one sensor, and maps that information to one of the available system-state representations.
In one embodiment the Q-learning reinforcement learning model learns an action to implement the most suitable traffic light sequences in the long-term for the given traffic conditions.
In one embodiment each agent comprises a set of policies such that each junction is adapted to learn a preferred action to be executed, in parallel with the importance of that action depending on the system state using said distributed W-Learning model. In one embodiment each junction combines local Q-learning and W-learning, with remote learning for neighbouring junctions, to provide a distributed W-learning model in order to obtain said action preferences. In one embodiment actions are traffic light settings and states can encode any information internally defined as relevant for the decision, e.g., number of cars approaching the junction, or number of buses waiting at approach.
In one embodiment the cooperation coefficient, C, is adapted to enable a local agent to give a varying degree of importance to the neighbours' action preferences, wherein 0 <= C <= 1.
In one embodiment, at each time step, each local and each remote policy on at least one agent is adapted to decide an action for execution at the next time step, based on the importance of executing that action in the current states of all local and remote policies .
In one embodiment there is provided Q-learning and W- learning data of neighbouring junctions, to obtain data necessary for remote learning of the action preferences from the immediate upstream and/or downstream junctions.
In a further embodiment there is provided an agent for use in an Urban Traffic Control environment, said agent representing a traffic-light controller at a traffic junction to control traffic flow, and adapted to collect data local to the junction using one or more sensors and applying a Q-learning reinforcement learning model to said collected data, one Q-learning model per each policy that the agent is implementing;
each agent comprises:
means for mapping locally collected data to one or more system state representations available to determine action values;
means to learn a current importance value of an action relative to actions preferred by other policies on an agent, using a W-Learning model, and determine the dependencies between locally implemented policies;
means for sharing Q-learning and W-learning parameters, for example current system state representations, environment feedback, action values and associated importance values with neighbouring agents, one-hop from each neighbouring agent;
means to receive and interpret the shared values to remotely learn preferences of policies implemented by other agents, and their relative importance, to learn the dependencies between the policies implemented locally and those implemented by neighbouring agents; and
means to combine local action-effectiveness and importance values with action-effectiveness and importance values of one-hop neighbouring agents in a periodic manner over a number of time steps to optimise the traffic flow at each agent, taking into account neighbour agents preferred actions and importance values, at each time step.
In another embodiment there is provided a method for controlling an Urban Traffic Control environment comprising a plurality of agents, each agent representing a traffic light controller at a traffic junction to control traffic flow, the method comprising the steps of:
collecting data local to the junction using one or more sensors and applying a Q-learning reinforcement learning model to said collected data;
mapping locally collected data to one or more system state representations available to determine action values ;
determining a current importance value of an action value using a W-Learning model;
sharing action values, rewards values and associated importance values with neighbouring agents, one-hop from each neighbouring agent;
utilizing shared values using remote learning to determine actions preferred by each neighbouring agent; and combining local action-effectiveness and importance values with action-effectiveness and importance values of one-hop neighbouring agents in a periodic manner over a number of time steps to optimise the traffic flow at each agent, taking into account neighbour agents preferred actions and importance values, at each time step.
There is also provided a computer program comprising program instructions for causing a computer program to carry out the above method which may be embodied on a recording medium, carrier signal or read-only memory.
Brief Description of the Drawings
The invention will be more clearly understood from the following description of an embodiment thereof, given by way of example only, with reference to the accompanying drawings, in which :-
Figure 1 illustrates an overall architecture of the system showing a group of traffic control signals at connected intersections each controlled by a single agent ;
Figure 2 illustrates agent interaction required for a decision making process on a single distributed W- Learning (DWL) agent for action selection;
Figure 3 illustrates an implementation of the DWL action-selection process on a single agent, according to one aspect of the invention; and
Figures 4, 5 and 6 illustrate the implementation of the learning process that each DWL agent performs to determine the most suitable cooperation coefficient (C) , according to other aspects of the invention.
Detailed Description of the Drawings/Operation
The invention provides a fully self-organizing Urban Traffic Control (UTC) system that uses reinforcement learning (RL) to map the currently-observed traffic conditions (based on the information received from the available road and/or in-car sensor technology) to appropriate traffic light sequences. Such a UTC system is enabled through use of a novel multi-policy multi-agent optimization algorithm, Distributed W-Learning (DWL), which is using RL techniques Q-learning and W-learning, remote learning, and cooperation coefficient learning and described in more detail below.
Distributed W-Learning (DWL)
In DWL, each junction implements a Q-learning RL process model whereby it receives information on current traffic conditions from available sensors, maps that information to one of the available system state representations, and executes the action (set of traffic light sequences) that is has learnt to be the most suitable in the long-term for the given traffic conditions. For each of the policies (e.g. prioritizing public transport, optimizing general traffic flow), a DWL agent/ unction learns a preferred action to be executed, as well as the importance of that action in its current system state. For example, for a policy that prioritizes public transport, action importance is high if it has two buses queuing on its approaches with both of them being behind schedule, while it has a very low
importance if there are no public transport vehicles approaching. The importance of the action to a policy is learnt using W-Learning process model. Referring now to the figures and initially Figure 1, Figure 1 shows the overall architecture of the system. A group of traffic control signals at connected intersections are each controlled by a single agent implementing DWL . The overall system consists of a group of such junctions, each junction hereinafter referred to as an agent. Each agent is implementing a DWL process to learn and execute the most suitable actions for its own performance goals/policies as well as the goals/policies of neighbouring (all upstream and all downstream) agents. For example, in Figure 1, DWL agent Al implements policy pi (e.g., optimize general traffic flow), DWL agent A2 implements policies pi (it is also optimizing general traffic flow) and p2 (e.g., prioritize buses) and DWL agent A3 also implements policy p2 (it is also prioritizing buses) . Using DWL, A2 learns the best local traffic-control signal settings to meet its performance policies pi and p2, but also, in collaboration with Al and A3, learns the best local settings to enable Al and A3 to meet their performance policies pi, p2 locally. The division into agents can also be physical, i.e., the software that implements DWL behaviours can be running on the traffic-control signals themselves, but all software agents can also be running in the same central location. Figure 2 depicts the agent interaction required for a decision making process on a single DWL agent. Each agent in the system, at each time step, performs interaction as described in Figure 2 with each of its neighbouring agents, for each of its local policies and for each of its
neighbours' policies. At system initialization, each agent exchanges with its neighbours a list of its local policies and associated states. Each agent initiates its remote policies which perform remote learning, one policy for each corresponding local policy of each of its neighbours. After initialization, at each time step (depicted in Figure 2), each local policy on each agent observes its local environmental conditions, maps them to a state representation, and performs updates on its local processes that learn the optimal actions and the importance of executing the preferred actions in each state. Each local policy suggests an action for execution at the next time step, together with the associated W-value (importance) of the policy's current state. Each agent also receives from each of its neighbours' state information (a representation of the neighbours' environment conditions) for each of its policies, and based on that information performs updates on its remote processes that learn the optimal actions and the importance of executing the preferred actions in each particular state. Each remote policy suggests an action for execution at the next time step, together with the associated W-value (importance) for the policy's current state. This decision process is described in more detail in Figure 3 below. Before making action decisions, each DWL agent also updates its learning process that learns the optimal value of Cooperation Coefficient (C) for the neighbourhood, i.e., for all its local and remote policies. This step is optional, as agents can either use predefined C, or learn C during the system operation. The process of learning C is described in more detail in Figure 4.
Figure 3 depicts the implementation of the DWL action- decision process on a single agent, according to one aspect of the invention. Figure 3 shows an agent A2, with its
local policies pi and p2. A2 has 2 neighbouring junctions, Al, which implements policy pi, and A3, which implements policy p2. Therefore, agent A2 learns how to best meet its local policies pi and p2 (lpl, lp2), but also learns how to best help neighbours meet their policies, or so-called remote policies, rpll (policy pi on agent Al ) and rp32 (policy p2 on agent A3). At each time step, each agent makes a decision as to what actions to execute (i.e., what traffic-control signal settings to deploy) based on optimal actions learnt for its local and remote policies. Each policy, both local and remote, suggests an action, as selected based on the outcome of the ongoing Q-learning processes associated with each policy. Each action is also associated with its current importance, expressed as a learnt W-value, learnt as the outcome of the ongoing W- learning processes associated with each policy. The action that is executed on each agent is the one with the highest current importance, i.e., the highest W-value, after remote policies' W-values have been multiplied by the cooperation coefficient C, where 0 <= C <= 1. C is introduced to enable a local agent to give a varying degree of importance to the neighbours' action preferences. C can range from a fully non-cooperative value, C=0, where an agent does not consider neighbours' action preferences at all, to a fully cooperative, C=l, where neighbours' preferences matter as much as local ones. C can be predefined or can be learnt at runtime, using a learning process which aims to maximize the rewards received on all agent's local and remote policies .
Figures 4, 5, and 6 illustrate the implementation of the learning process that each DWL agent performs to learn the most suitable cooperation coefficient (C) . C enables agents
to take into account action preferences of their neighbours' policies with varying weight, which can be equal to or less than their local action preferences. At each time step, each local and each remote learning policy on a D L agent suggest an action for execution at the next time step, together with an associated W-value, representing the importance of executing that action in the policy's current state. An agent takes the suggestions of its local policies with their full weight, but it has the option to scale the weight of remote policy suggestions (i.e., actions suitable for its neighbours) in order to give higher priority to its local preferences. Using different scaling on different agents in the system can be beneficial, in order to, for example, enable junctions that are more important for the overall system performance to take only their local preferences into account (i.e., use C=0), or to enable less important junctions to execute actions suitable for their more important neighbours. Therefore, each DWL agent is enabled to learn or determine the most suitable value of C, so that resulting behaviour is optimal for the neighbourhood, i.e., for an agent itself and for all the policies that all of its one-hop neighbours are implementing. Determining C is implemented as a learning process on each agent. The set of actions in that learning process consists of various values of C, e.g., {C=0, C=0.1, C=0.2, C=0.3, C=0.4, C=0.5, C=0.6, C=0.7, C=0.8, C=0.9, C=l}. At each time step, a learning process selects one of the C values and the agent multiplies all W- values received from remote policies with that value, for example as shown in Figure 4.
Figure 5 shows the action associated with the state with the highest current W-value is executed, and the outcome of that action is observed on all local policies and on all policies on all one-hop neighbours, i.e., the rewards received by all local policies and by all policies on all one-hop neighbours are added up and used to update the value of C used in the last time step using a learning process . Figure 6 illustrates how the value of C can be learnt per agent, i.e., an agent uses the same C to scale W-values received for all policies on all one-hop neighbours; can be learnt per agent pair, i.e., an agent has as many Q- learning processes as it has neighbours, and learns a separate C to scale W-values received from each neighbour; or can be learnt per remote policy, i.e., the number of Q- learning processes that are learning C on an agent is the sum of the number of all policies that all of its one-hop neighbours implement, agent learns a separate C to scale W- values received from each policy on each neighbour. Figure 4 shows an example where a single C is learnt on an agent, and Figure 6 shows an example where a different C is learnt for each remote policy. It will be appreciated that the DWL technique according to the invention combines local Q-learning and W-learning, with remote learning for its neighbouring junctions, to obtain action preferences not just from its local policies, but also from the immediate upstream and downstream junctions. Each agent can be adapted to obtain action preferences from all of its one-hop neighbours, for example from neighbours multiple hops away or only from a subset of one-hop or multi-hop neighbours . It is envisaged that
agents can be adapted to collaborate with only downstream or only upstream neighbours, depending on the application required . In this way, junctions coordinate with their upstream and downstream neighbours to execute actions that are most appropriate for the traffic conditions not just locally, but for the immediate neighbourhood as well. The priority of different policies is easily incorporated into DWL, as if a policy is given a higher reward in the RL process design, it will have a higher importance compared to a lower-priority policy, enabling easy integration of public vehicle and emergency vehicle priority in a UTC system. In order to enable an agent to decide whether to execute an action preferred by its local policies or by its neighbours' policies, DWL includes the cooperation coefficient C. C has a value between 0 and 1, where 0 denotes a non-collaborative junction, i.e., a junction that always executes an action preferred by its local policies, and 1 denotes a fully collaborative junction, i.e., a junction that gives as much weight to neighbours' preferences as to its local preferences when making an action decision. C can be predefined (to make particular junctions more/less cooperative based on their importance in the system) or can be learnt by each junction so as to maximize the reinforcement learning reward locally and in its one-hop neighbourhood. In one embodiment the system can function as follows, each junction/agent implements a Q-learning reinforcement learning process whereby it receives information on current traffic conditions from available sensors. This information
is then mapped to one of the available system state representations, and an action is executed (set of traffic light sequences) that is has learnt to be the most suitable in the long-term for the given traffic conditions.
For each of a set of policies (e.g. prioritizing public transport, optimizing general traffic flow) , a junction/agent learns a preferred action to be executed, as well as the importance of that action in its current system state using W-Learning.
The overall operation of this embodiment can be summarised in that Distributed W-Learning combines local Q-learning and W-learning, with remote learning on behalf of its neighbouring junctions/agents, and learning of cooperation coefficients, to obtain action preferences not just from its local policies, but also from all its one-hop neighbours, in order to improve traffic flow efficiency. It will be appreciated that the present invention allows increasingly available road-side and in-car sensor technology, as well as car-to-infrastructure and car-to-car communication capabilities to be utilized to enable UTC systems to make more informed, quicker adaptation decisions. In particular, by using UTC systems based on distributed w-learning, performance is increased and response time to changing traffic conditions is decreased, as well as a reduction in operating costs in traffic control centres, including both human and hardware cost. In addition, by using UTC systems based on remote learning and cooperation coefficient learning, junctions are capable of learning the dependencies between each other's performance and capable of collaboration to help improve not just their
own, but each other's performance, and therefore performance of the overall system.
The embodiments in the invention described with reference to the drawings comprise a computer apparatus and/or processes performed in a computer apparatus to control each agent according to the invention. However, the invention also extends to computer programs, particularly computer programs stored on or in a carrier adapted to bring the invention into practice and for controlling the agent. The program may be in the form of source code, object code, or a code intermediate between source and object code, such as in partially compiled form or in any other form suitable for use in the implementation of the method according to the invention. The carrier may comprise a storage medium such as ROM, e.g. CD ROM, or magnetic recording medium, e.g. a floppy disk or hard disk. The carrier may be an electrical or optical signal which may be transmitted via an electrical or an optical cable or by radio or other means.
In the specification the terms "comprise, comprises, comprised and comprising" or any variation thereof and the terms include, includes, included and including" or any variation thereof are considered to be totally interchangeable and they should all be afforded the widest possible interpretation and vice versa. In addition the 'agent' hereinbefore described with respect to the invention can be incorporated in an existing junction node or new junction node in hardware or software or a combination of both. The invention is not limited to the embodiments hereinbefore described but may be varied in both construction and detail.
Claims
Claims
1. A system of agents for use in an Urban Traffic Control environment, each agent representing a traffic light controller at a traffic junction to control traffic flow, said system comprising:
each agent is adapted to collect data local to the junction using one or more sensors and applying a Q- learning reinforcement learning model to said collected data, one Q-learning model per each policy that the agent is implementing;
each agent comprises:
means for mapping locally collected data to one or more system state representations available to determine action values;
means to learn a current importance value of an action relative to actions preferred by other policies on an agent, using a W-Learning model, and determine the dependencies between locally implemented policies; means for sharing Q-learning and W-learning parameters, for example current system state representations, environment feedback, action values and associated importance values with neighbouring agents, one-hop from each neighbouring agent;
means to receive and interpret the shared values to remotely learn preferences of policies implemented by other agents, and their relative importance, to learn the dependencies between the policies implemented locally and those implemented by neighbouring agents; and
means to combine local action-effectiveness and importance values with action-effectiveness and importance values of one-hop neighbouring agents in a periodic manner over a number of time steps to
optimise the traffic flow at each agent, taking into account neighbour agents preferred actions and importance values, at each time step. 2. The system as claimed in claim 1 wherein the exchanged values to learn preferences of policies implemented by other agents comprise means for using remote policy learning . 3. The system as claimed in claims 1 or 2 comprising means to determine a cooperation coefficient using a learning model and using said cooperation coefficient to scale action preferences of neighbour agents so as to maximize the performance locally and in the immediate agent neighbourhood.
The system as claimed in claim 3 wherein the cooperation coefficient, C, is adapted to enable a local agent to give a varying degree of importance to the neighbours' action preferences, wherein 0 <= C <= 1.
The system as claimed in any preceding claim wherein each junction combines local Q-learning and W-learning, with remote learning for neighbouring junctions, to provide a distributed W-learning model in order to obtain said action preferences.
The system as claimed in any preceding claim wherein the Q-learning reinforcement learning model receives information on current traffic conditions from at least one sensor, and adapted to map that information to one of the available system-state representations .
The system as claimed in any preceding claim wherein the learning model executes an action to implement the most suitable traffic-light sequences in the long-term for the given traffic conditions .
The system as claimed in any preceding claim wherein each agent comprises a set of policies such that each junction is adapted to learn a preferred action to be executed, in parallel with the importance of that action in its current system state using said W-Learning model.
The system as claimed in any preceding claim wherein each junction combines local Q-learning and W-learning with remote learning on behalf of its neighbouring junctions/agents, and learning of cooperation coefficients, to provide a distributed W-learning model, to obtain said action preferences.
10. The system as claimed in any preceding claim, wherein actions are traffic light settings and states can encode any information internally defined as relevant for the decision, for example, number of cars approaching the junction ornumber of buses waiting at an approach. 11. The system as claimed in any preceding claim wherein at each time step, each local and each remote policy on at least one agent is adapted to decide an action for execution at the next time step, based on the importance of executing that action in the policy's current state.
12. The system as claimed in any preceding claim wherein Q-learning and W-learning data of neighbouring junctions are exchanged, to obtain action preferences from the immediate upstream and/or downstream junctions.
13. An agent for use in an Urban Traffic Control environment adapted for incorporation into the system as claimed in any preceding claim.
14. A method of controlling a system of agents for use in an Urban Traffic Control environment, each agent representing a traffic light controller at a traffic junction to control traffic flow, said method comprising the steps of:
collecting at each agent data local to the junction using one or more sensors and applying a Q-learning reinforcement learning model to said collected data, one Q-learning model per each policy that the agent is implementing;
mapping at each agent locally collected data to one or more system-state representations available to determine action values;
learning at each agent a current importance value of an action relative to actions preferred by other policies on an agent, using a W-Learning model, and determine the dependencies between locally implemented policies ;
sharing at each agent Q-learning and W-learning parameters, for example current system-state representations, environment feedback, action values and associated importance values with neighbouring agents, one-hop from each neighbouring agent;
receiving and interpreting the shared values to learn preferences of policies implemented by other agents, and their relative importance, to learn the dependencies between the policies implemented locally and those implemented by neighbouring agents; and
combining local action-effectiveness and importance values with action-effectiveness and importance values of one-hop neighbouring agents in a periodic manner over a number of time steps to optimise the traffic flow at each agent, taking into account neighbour agents preferred actions and importance values, at each time step.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US13/703,774 US20130176146A1 (en) | 2010-06-15 | 2011-06-15 | Decentralised Autonomic System and Method for Use in an Urban Traffic Control Environment |
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US35479810P | 2010-06-15 | 2010-06-15 | |
| US61/354,798 | 2010-06-15 | ||
| GBGB1009974.5A GB201009974D0 (en) | 2010-06-15 | 2010-06-15 | Decentralised autonomic system and method for use inan urban traffic control environment |
| GB1009974.5 | 2010-06-15 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2011157745A1 true WO2011157745A1 (en) | 2011-12-22 |
Family
ID=42471657
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/EP2011/059926 Ceased WO2011157745A1 (en) | 2010-06-15 | 2011-06-15 | Decentralised autonomic system and method for use in an urban traffic control environment |
Country Status (2)
| Country | Link |
|---|---|
| GB (1) | GB201009974D0 (en) |
| WO (1) | WO2011157745A1 (en) |
Cited By (19)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN103106801A (en) * | 2013-01-14 | 2013-05-15 | 上海应用技术学院 | Self-organizing traffic signal coordination control method |
| WO2013086629A1 (en) * | 2011-12-16 | 2013-06-20 | El-Tantawy Samah | Multi-agent reinforcement learning for integrated and networked adaptive traffic signal control |
| EP3051516A1 (en) * | 2015-01-30 | 2016-08-03 | Siemens Corporation | Real-time monitoring and diagnostic processing of traffic control data |
| CN106652500A (en) * | 2017-01-05 | 2017-05-10 | 湖南金环交通设施市政建设有限公司 | Guard route traffic signal lamp automatic control system |
| CN106846836A (en) * | 2017-02-28 | 2017-06-13 | 许昌学院 | A kind of Single Intersection signal timing control method and system |
| CN106910351A (en) * | 2017-04-19 | 2017-06-30 | 大连理工大学 | A traffic signal adaptive control method based on deep reinforcement learning |
| CN108335497A (en) * | 2018-02-08 | 2018-07-27 | 南京邮电大学 | A kind of traffic signals adaptive control system and method |
| CN108802769A (en) * | 2018-05-30 | 2018-11-13 | 千寻位置网络有限公司 | Detection method and device of the GNSS terminal on overhead or under overhead |
| CN108806287A (en) * | 2018-06-27 | 2018-11-13 | 沈阳理工大学 | A kind of Traffic Signal Timing method based on collaboration optimization |
| CN109035812A (en) * | 2018-09-05 | 2018-12-18 | 平安科技(深圳)有限公司 | Control method, device, computer equipment and the storage medium of traffic lights |
| CN109389835A (en) * | 2018-11-16 | 2019-02-26 | 北方工业大学 | A Mapping Method of Urban Road Traffic Demand and Control |
| CN109544913A (en) * | 2018-11-07 | 2019-03-29 | 南京邮电大学 | A kind of traffic lights dynamic timing algorithm based on depth Q e-learning |
| CN109785619A (en) * | 2019-01-21 | 2019-05-21 | 南京邮电大学 | Coordinated optimal control system for regional traffic signal and its control method |
| CN110047278A (en) * | 2019-03-30 | 2019-07-23 | 北京交通大学 | A kind of self-adapting traffic signal control system and method based on deeply study |
| CN111696370A (en) * | 2020-06-16 | 2020-09-22 | 西安电子科技大学 | Traffic light control method based on heuristic deep Q network |
| CN112863206A (en) * | 2021-01-07 | 2021-05-28 | 北京大学 | Traffic signal lamp control method and system based on reinforcement learning |
| US11080602B1 (en) | 2020-06-27 | 2021-08-03 | Sas Institute Inc. | Universal attention-based reinforcement learning model for control systems |
| CN113276884A (en) * | 2021-04-28 | 2021-08-20 | 吉林大学 | Intelligent vehicle interactive decision passing method and system with variable game mode |
| CN114973660A (en) * | 2022-05-13 | 2022-08-30 | 黄河科技学院 | Traffic decision method of model linearization iteration updating method |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| DE4436339A1 (en) | 1994-10-11 | 1996-04-18 | Ifu Gmbh | Traffic lights sequence control method |
| JPH09160605A (en) | 1995-12-11 | 1997-06-20 | Omron Corp | Controller used by decentralized control system, and decentralized control method |
| WO2001086610A1 (en) | 2000-05-05 | 2001-11-15 | Siemens Aktiengesellschaft | Method and device for determining an optimized selection of a frame signal diagram from a large number of frame signal diagrams for a traffic system |
| US7047224B1 (en) | 1998-09-23 | 2006-05-16 | Siemens Aktiengesellschaft | Method and configuration for determining a sequence of actions for a system which comprises statuses, whereby a status transition ensues between two statuses as a result of an action |
-
2010
- 2010-06-15 GB GBGB1009974.5A patent/GB201009974D0/en not_active Ceased
-
2011
- 2011-06-15 WO PCT/EP2011/059926 patent/WO2011157745A1/en not_active Ceased
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| DE4436339A1 (en) | 1994-10-11 | 1996-04-18 | Ifu Gmbh | Traffic lights sequence control method |
| JPH09160605A (en) | 1995-12-11 | 1997-06-20 | Omron Corp | Controller used by decentralized control system, and decentralized control method |
| US7047224B1 (en) | 1998-09-23 | 2006-05-16 | Siemens Aktiengesellschaft | Method and configuration for determining a sequence of actions for a system which comprises statuses, whereby a status transition ensues between two statuses as a result of an action |
| WO2001086610A1 (en) | 2000-05-05 | 2001-11-15 | Siemens Aktiengesellschaft | Method and device for determining an optimized selection of a frame signal diagram from a large number of frame signal diagrams for a traffic system |
Non-Patent Citations (8)
| Title |
|---|
| BAZZAN, A. L.: "A distributed approach for coordination of traffic signal agents", AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, vol. 10, no. 1, 2005, pages 131 - 164, XP019227686, DOI: doi:10.1007/s10458-004-6975-9 |
| FEBBRARO, A. D., GIGLIO, D., SACCO, N.: "Urban traffic control structure based on hybrid Petri nets", IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, vol. 5, no. 4, 2004, pages 224 - 237, XP011122957, DOI: doi:10.1109/TITS.2004.838180 |
| HUMPHRYS, W-LEARNING: COMPETITION AMONG SELFISH Q-LEARNERS, 1995 |
| IVANA DUSPARIC ET AL: "Distributed W-Learning: Multi-Policy Optimization in Self-Organizing Systems", SELF-ADAPTIVE AND SELF-ORGANIZING SYSTEMS, 2009. SASO '09. THIRD IEEE INTERNATIONAL CONFERENCE ON, IEEE, PISCATAWAY, NJ, USA, vol. 1, 14 September 2009 (2009-09-14), pages 20 - 29, XP031552273, ISBN: 978-1-4244-4890-6 * |
| IVANA DUSPARIC ET AL: "Multi-policy Optimization in Self-organizing Systems", 14 September 2009, SELF-ORGANIZING ARCHITECTURES, SPRINGER BERLIN HEIDELBERG, BERLIN, HEIDELBERG, PAGE(S) 101 - 126, ISBN: 978-3-642-14411-0, XP019147057 * |
| IVANA DUSPARIC ET AL: "Using Reinforcement Learning for Multi-policy Optimization in Decentralized Autonomic Systems â An Experimental Evaluation", 7 July 2009, AUTONOMIC AND TRUSTED COMPUTING, SPRINGER BERLIN HEIDELBERG, BERLIN, HEIDELBERG, PAGE(S) 105 - 119, ISBN: 978-3-642-02703-1, XP019122660 * |
| RICHTER, S.: "Tech. rep.", 2006, ALBERT-LUDWIGS-UNIVERSITAT FREIBURG, article "Learning traffic control - towards practical traffic control using policy gradients" |
| SALKHAM: "A collaborative reinforcement learning approach to urban traffic control optimization", IAT, 2008, pages 560 - 566, XP058017263, DOI: doi:10.1109/WIIAT.2008.88 |
Cited By (29)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2013086629A1 (en) * | 2011-12-16 | 2013-06-20 | El-Tantawy Samah | Multi-agent reinforcement learning for integrated and networked adaptive traffic signal control |
| CN103106801A (en) * | 2013-01-14 | 2013-05-15 | 上海应用技术学院 | Self-organizing traffic signal coordination control method |
| CN103106801B (en) * | 2013-01-14 | 2015-05-20 | 上海应用技术学院 | Self-organizing traffic signal coordination control method |
| EP3051516A1 (en) * | 2015-01-30 | 2016-08-03 | Siemens Corporation | Real-time monitoring and diagnostic processing of traffic control data |
| US9875654B2 (en) | 2015-01-30 | 2018-01-23 | Siemens Aktiengesellschaft | Real-time monitoring and diagnostic processing of traffic control data |
| US10109196B2 (en) | 2015-01-30 | 2018-10-23 | Siemens Aktiengesellschaft | Real-time monitoring and diagnostic processing of traffic control data |
| CN106652500A (en) * | 2017-01-05 | 2017-05-10 | 湖南金环交通设施市政建设有限公司 | Guard route traffic signal lamp automatic control system |
| CN106846836A (en) * | 2017-02-28 | 2017-06-13 | 许昌学院 | A kind of Single Intersection signal timing control method and system |
| CN106846836B (en) * | 2017-02-28 | 2019-05-24 | 许昌学院 | A kind of Single Intersection signal timing control method and system |
| CN106910351A (en) * | 2017-04-19 | 2017-06-30 | 大连理工大学 | A traffic signal adaptive control method based on deep reinforcement learning |
| CN106910351B (en) * | 2017-04-19 | 2019-10-11 | 大连理工大学 | A traffic signal adaptive control method based on deep reinforcement learning |
| CN108335497A (en) * | 2018-02-08 | 2018-07-27 | 南京邮电大学 | A kind of traffic signals adaptive control system and method |
| CN108802769A (en) * | 2018-05-30 | 2018-11-13 | 千寻位置网络有限公司 | Detection method and device of the GNSS terminal on overhead or under overhead |
| CN108806287A (en) * | 2018-06-27 | 2018-11-13 | 沈阳理工大学 | A kind of Traffic Signal Timing method based on collaboration optimization |
| CN109035812A (en) * | 2018-09-05 | 2018-12-18 | 平安科技(深圳)有限公司 | Control method, device, computer equipment and the storage medium of traffic lights |
| CN109035812B (en) * | 2018-09-05 | 2021-07-27 | 平安科技(深圳)有限公司 | Traffic signal lamp control method and device, computer equipment and storage medium |
| CN109544913A (en) * | 2018-11-07 | 2019-03-29 | 南京邮电大学 | A kind of traffic lights dynamic timing algorithm based on depth Q e-learning |
| CN109389835A (en) * | 2018-11-16 | 2019-02-26 | 北方工业大学 | A Mapping Method of Urban Road Traffic Demand and Control |
| CN109785619A (en) * | 2019-01-21 | 2019-05-21 | 南京邮电大学 | Coordinated optimal control system for regional traffic signal and its control method |
| CN109785619B (en) * | 2019-01-21 | 2021-06-22 | 南京邮电大学 | Coordinated optimal control system for regional traffic signal and its control method |
| CN110047278A (en) * | 2019-03-30 | 2019-07-23 | 北京交通大学 | A kind of self-adapting traffic signal control system and method based on deeply study |
| CN110047278B (en) * | 2019-03-30 | 2021-06-08 | 北京交通大学 | An adaptive traffic signal control system and method based on deep reinforcement learning |
| CN111696370A (en) * | 2020-06-16 | 2020-09-22 | 西安电子科技大学 | Traffic light control method based on heuristic deep Q network |
| US11080602B1 (en) | 2020-06-27 | 2021-08-03 | Sas Institute Inc. | Universal attention-based reinforcement learning model for control systems |
| CN112863206A (en) * | 2021-01-07 | 2021-05-28 | 北京大学 | Traffic signal lamp control method and system based on reinforcement learning |
| CN113276884A (en) * | 2021-04-28 | 2021-08-20 | 吉林大学 | Intelligent vehicle interactive decision passing method and system with variable game mode |
| CN113276884B (en) * | 2021-04-28 | 2022-04-26 | 吉林大学 | Intelligent vehicle interactive decision passing method and system with variable game mode |
| CN114973660A (en) * | 2022-05-13 | 2022-08-30 | 黄河科技学院 | Traffic decision method of model linearization iteration updating method |
| CN114973660B (en) * | 2022-05-13 | 2023-10-24 | 黄河科技学院 | Traffic decision method of model linearization iterative updating method |
Also Published As
| Publication number | Publication date |
|---|---|
| GB201009974D0 (en) | 2010-07-21 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20130176146A1 (en) | Decentralised Autonomic System and Method for Use in an Urban Traffic Control Environment | |
| WO2011157745A1 (en) | Decentralised autonomic system and method for use in an urban traffic control environment | |
| Zhang et al. | Using reinforcement learning with partial vehicle detection for intelligent traffic signal control | |
| Prothmann et al. | Organic control of traffic lights | |
| Keyvan-Ekbatani et al. | Controller design for gating traffic control in presence of time-delay in urban road networks | |
| Faye et al. | A distributed algorithm for adaptive traffic lights control | |
| Prothmann et al. | Organic traffic control | |
| CN113316808A (en) | Spatial control of traffic signals by space-time expansion of traffic conditions | |
| Korecki et al. | Analytically guided reinforcement learning for green it and fluent traffic | |
| Xiang et al. | An adaptive traffic signal coordination optimization method based on vehicle-to-infrastructure communication | |
| JP7296561B2 (en) | Information processing device, traffic signal control system and traffic signal control method | |
| Fu et al. | Modeling and simulation of dynamic lane reversal using a cell transmission model | |
| Chen et al. | Dynamic traffic light optimization and Control System using model-predictive control method | |
| Celtek et al. | Evaluating action durations for adaptive traffic signal control based on deep Q-learning | |
| Hao et al. | Distributed cooperative backpressure‐based traffic light control method | |
| Gunarathna et al. | Real-time lane configuration with coordinated reinforcement learning | |
| Coll et al. | A linear programming approach for adaptive synchronization of traffic signals | |
| Soleimany et al. | Hierarchical federated learning model for traffic light management in future smart | |
| JP7296567B2 (en) | TRAFFIC SIGNAL CONTROL DEVICE, TRAFFIC SIGNAL CONTROL SYSTEM AND TRAFFIC SIGNAL CONTROL METHOD | |
| Hu et al. | Learning model parameters for decentralized schedule-driven traffic control | |
| Ajayi et al. | Optimising Urban Road Transportation Efficiency: AI-Driven Solutions for Reducing Traffic Congestion in Big Cities | |
| Kaths et al. | Traffic signals in connected vehicle environments: Chances, challenges and examples for future traffic signal control | |
| Galvão et al. | Smart traffic signal control in urban networks via deep reinforcement learning and visible light communication | |
| Friesen et al. | Multi-Agent Deep Reinforcement Learning For Real-World Traffic Signal Controls-A Case Study | |
| Korecki et al. | Analytically guided machine learning for green IT and fluent traffic |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11731293 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 13703774 Country of ref document: US |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 11731293 Country of ref document: EP Kind code of ref document: A1 |