WO2024223902A1 - Trajectory prediction for dynamic agents - Google Patents
Trajectory prediction for dynamic agents Download PDFInfo
- Publication number
- WO2024223902A1 WO2024223902A1 PCT/EP2024/061674 EP2024061674W WO2024223902A1 WO 2024223902 A1 WO2024223902 A1 WO 2024223902A1 EP 2024061674 W EP2024061674 W EP 2024061674W WO 2024223902 A1 WO2024223902 A1 WO 2024223902A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- trajectory
- agent
- visual
- features
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/62—Extraction of image or video features relating to a temporal dimension, e.g. time-based feature extraction; Pattern tracking
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W60/00—Drive control systems specially adapted for autonomous road vehicles
- B60W60/001—Planning or execution of driving tasks
- B60W60/0027—Planning or execution of driving tasks using trajectory prediction for other traffic participants
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2420/00—Indexing codes relating to the type of sensors based on the principle of their operation
- B60W2420/40—Photo, light or radio wave sensitive means, e.g. infrared sensors
- B60W2420/403—Image sensing, e.g. optical camera
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2554/00—Input parameters relating to objects
- B60W2554/40—Dynamic objects, e.g. animals, windblown objects
- B60W2554/402—Type
- B60W2554/4029—Pedestrians
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2554/00—Input parameters relating to objects
- B60W2554/40—Dynamic objects, e.g. animals, windblown objects
- B60W2554/404—Characteristics
- B60W2554/4045—Intention, e.g. lane change or imminent movement
Definitions
- the present disclosure pertains to agent trajectory prediction, which has applications, for example, in autonomous driving and robotics.
- predicted agent trajectories may be used as a basis to autonomously plan an ego trajectory for a robot (e.g. autonomous vehicle), referred to as the ego agent.
- a robot e.g. autonomous vehicle
- AV Autonomous Vehicles
- VRUs vulnerable road users
- Pedestrian prediction is hard since they can change direction and start or stop moving, and it is high risk, for example if they enter a road area. Producing conservative estimates of pedestrian motion allows potential actions to be captured and avoided, however can lead to very conservative driving of an AV and prevent progress of the AV.
- trajectory -based predictions techniques are provided herein which incorporate visual appearance information to more accurately identify when changes of motion occur, and thus more accurately predict a future trajectory of a VRU.
- the prediction techniques disclosed herein are particularly well suited to the task of predicting pedestrian trajectories, or trajectories of other vulnerable road users (VRUs), using visual ‘cues’ in combination with historic trajectories.
- VRUs vulnerable road users
- a first aspect herein provides a computer-implemented method of predicting an agent trajectory in a world space, the method comprising: receiving a visual input comprising an image of the agent in an image space; extracting from the visual input using a visual feature extractor a set of visual features; receiving a trajectory history input comprising an observed trajectory of the agent in the world space; extracting from the trajectory history input a set of trajectory features; and generating, based on the set of visual features and the set of trajectory features, a prediction output comprising a predicted trajectory for the agent in the world space.
- Embodiments utilize vision-based classification of motion changes (e.g. motion initiations) to inform trajectory prediction.
- motion changes e.g. motion initiations
- Image features in the image space are thus used to guide trajectory prediction in the world space.
- the world space may, for example, be defined by coordinates within a ground plane (or world plane), which encode a ‘bird’s-eye-view’ (or top-down view) of the agent and its surroundings.
- the world space may be two-dimensional, or three- dimensional (with the addition of a height coordinate above the ground plane).
- the observed and predicted trajectories have coordinates in the ground plane (as possibly height coordinates as well).
- each trajectory may comprise a sequence of historic/predicted agent locations, each defined by a set of ground plane (and, in some cases, height) coordinates. Trajectory features may be extracted in the ground plane (e.g. via temporal convolution).
- the image space may be defined by image coordinates in an image plane lying substantially perpendicular to the ground plane, or more generally non-parallel to it.
- the image space may be two-dimensional, or three-dimensional with the addition of a depth coordinate.
- the image features may be extracted in the image plane (with additional temporal convolution in some embodiments).
- the visual input may comprise a time sequence of images of the agent (appearance frames), such as agent images cropped from a video sequence.
- the set of trajectory features may, for example, be extracted via one or more temporal convolution operations applied to the trajectory history input.
- the visual feature extractor may, for example, have a neural network architecture.
- the visual feature extractor has the form of a 2D convolutional neural network (CNN) feature extractor that extracts visual features in the image space from each appearance frame independently, and a temporal convolution processing component that combines those visual features.
- CNN convolutional neural network
- the 2D CNN feature extractor may, for example, comprise a 2D CNN that processes the appearance frames sequentially, or multiple 2D CNNs that process the appearance frames in parallel.
- a second aspect herein provides a computer-implemented method of training a machine learning visual feature extractor, the method comprising: receiving a training image of an agent in an image space; receiving a known trajectory in a world space associated with the training image; automatically determining a motion state ground truth associated with the training image based on the trajectory in the world space; and extracting using the visual feature extractor a set of visual features from the training image; using the extracted visual features to compute a predicted motion state; and training the visual feature extractor based on the predicted motion state and the ground truth motion state.
- the training may comprise tuning parameters of the visual feature extractor based on a training loss that measures error in the predicted motion state relative to the ground truth motion state.
- the groundtruth/predicted motion states may be, e.g. motion classes (motion classification) or numerical motion value (motion regression).
- Figure 1 shows a schematic block diagram of an autonomous vehicle stack
- Figure 2 illustrates a schematic block diagram of a neural network architecture for interactive prediction
- Figure 3 is an example of cropped pedestrian appearance, showing gait change with motion initiation
- Figure 4 is a schematic block diagram of an appearance-based model for predicting agent trajectories
- Figure 5 shows examples of prediction inputs and outputs
- Figure 6 is a schematic block diagram of a model for motion state prediction.
- Figure 2 shows a neural network-based, multi-modal prediction architecture referred to as DiPA (Diverse and Probabilistically Accurate Interactive Prediction), which generates predicted trajectories, with spatial distribution parameters (e.g. mean and covariance) and mode prediction weights for different prediction modes.
- DiPA Diaverse and Probabilistically Accurate Interactive Prediction
- the present prediction architecture is incorporated in the DiPA architecture in one embodiment.
- FIG. 2 is a schematic block diagram of an example neural network architecture that may be used to perform interaction-based agent trajectory prediction, for example to predict agent trajectories for agents of a driving scenario based on which an autonomous vehicle stack can plan a safe trajectory.
- the network receives a set of trajectory histories 206 comprising a set of past states for each agent of a scenario.
- Each trajectory history may be in the form of a time series of agent states over some past time interval, each state including the position, orientation and velocity of the agent at a corresponding point in time. At least one past agent state is received for each agent.
- Other agent inputs 204 may also be received, such as a set of spatial dimensions for each agent.
- These agent inputs and agent histories may be derived from sensor data by a perception subsystem 102 of an AV stack 100.
- the trajectory histories are processed by one or more temporal convolution layers 208 to generate a feature vector for each agent that is time-independent.
- the set of trajectory histories 206 is an array having shape (agents, time, features), i.e. for each agent, a set of features is input defining the state of the agent at each timestep of the time interval.
- features is used herein to refer to a representative set of values, and different feature representations are used in different parts of the network.
- ‘features’ when used to describe a dimension of an input or output of the network can refer to different numbers and types of features for different inputs and outputs.
- the feature values represent the position, orientation and velocity of the agent at the selected point in time
- the features output from the temporal convolution represent the states of the agent over the entire time interval. Note that the dimensionality shown in Figure 2 excludes the feature dimension.
- the convolved trajectory histories are broadcast over all agents as shown by the 0 symbol, for example by concatenating the feature vector for each agent with each feature vector associated with each other agent of the scenario.
- Each combination of two agents is associated with a respective pairwise feature, which is processed by one or more interaction layers 210 of the neural network. These may be fully connected (FC) layers as shown.
- the output of the interaction layers 210 is a respective feature vector for each pairwise combination of agents, which is subsequently reduced (aggregated) over one agent dimension. Note that each agent is treated as an independent input to the network, similarly to elements of a batch, rather than as components of a single input that the network learns to process according to an assigned role.
- a first agent of each pair is referred to as the ‘reference’ agent while the second agent of the pair is referred to as the ‘comparison’ agent.
- All agents of the scenario act as both a reference agent and a comparison agent.
- the reduction is over the comparison agent dimension, such that a respective interaction representation is output for each agent of the scenario.
- Example reduction operations include max reductions, where for a given reference agent the maximum value for each feature is selected over all comparison agents, and a sum, which gives the sum of each feature over all the comparison agents.
- the reduced interaction representation feature vectors which have dimension (agents, features) is combined with the additional agent inputs, as well as the convolved histories, which also have dimension (agents, features), although, as noted above, the number of features for the interaction representation and the agent inputs need not be the same.
- These are combined in the present example by concatenating the agent input (combined with the convolved agent histories) and the interaction representation vector for each agent as shown by the ⁇ in Figure 2.
- This combined feature representation for each agent is processed by a set of prediction layers 212 and 216.
- a scene context may also be generated by reducing an intermediate output of a first set of prediction layers 212 and processing this in a set of context layers 214, with the output being broadcast to the intermediate output for each agent, for example by concatenating the scene context with each intermediate output as generated for each agent, before processing the combined agent representations in a second set of prediction layers 216 to generate a final predicted output including a trajectory prediction 218.
- the neural network is configured to make predictions for each of a fixed number of prediction ‘modes’, where the number of modes, for example five, is predetermined.
- the output generates a predicted trajectory 218 for each mode, as well as a spatial distribution 220 indicating uncertainty of the trajectory itself in space, as well as a weight 222 (such as a mode probability) which indicates a predicted mode, where the mode with the highest weight is the mode that the network determines is the most likely for the given agent.
- a weight 222 such as a mode probability
- the DiPA architecture provides an effective mechanism for evaluating multimodal predictions in the context of prediction of road users, which considers the various possibilities that should be considered, and evaluating the accuracy of the distribution of predictions.
- this architecture for pedestrian motion prediction, such as changes of motion, is recognized as it provides a general method for evaluating a distribution of future multimodal states.
- DiPA architecture includes the extraction of pairwise interaction features (‘agent x agent’), the use of this is optional.
- An interaction layer 210 is shown as the top layer at the top of Figure 2, but this layer is not necessarily required in the present context. Hence, the interaction layer 210 may be omitted, or deactivated. In this case, agents are considered individually, without explicitly modelling agent interactions.
- Figure 2 also shows a global feature layer at the bottom, which can also be omitted or deactivated in this case. This layer combines the features between the various agents in the scene, and is not required when only handling a single agent at a time.
- the DiPA architecture is capable of handling a varying number of agents, including a single agent at a time, and can therefore be configured to operate with the number of agents set to one).
- the interaction and global layers may be retained, handling multiple agents at a time, making full use of the interactive DiPA architecture.
- a middle agent prediction layer is shown to have three inputs: trajectory history, agent inputs (as in DiPA), and the additional visual features.
- the agent inputs may, for example, comprise agent dimensions.
- the outputs to the network are unchanged.
- Appearance cues such as changes of body pose provide additional information of the actions of pedestrians, such as when gait is changing in order to begin or stop moving. These appearance cues can reliably inform when motion changes are taking place, and provide an early and accurate signal of motion.
- Figure 3 illustrates an example pedestrian appearance, showing gait change with motion initiation.
- a first image 302 shows a pedestrian waiting at a crossing on a pavement at the side of a road. In this image, both feet of the pedestrian are stood on the pavement.
- a second image 304 shows the pedestrian moving closer to the edge of the pavement.
- the pedestrian is shown to be at the edge of the pavement. In this image, the pedestrian is stood such that they appear to intend on walking closer the edge of the pavement, towards the road.
- the pedestrian is seen to leave the pavement and cross the edge such that they have now stepped onto the road.
- the motion cues in the images presented in Figure 3 may be used by a trained neural network to inform trajectory -based motion prediction.
- a visual feature extractor described below with reference to Figure 4 may use these images to predict the motion of a pedestrian.
- Prediction of pedestrian motion is inherently a multimodal task- if a pedestrian is standing beside the road area there are at least two significant possibilities to consider, of whether they remain stationary or begin moving into the road, which would create a hazard for the AV.
- a multimodal predictor can create a predicted trajectory for each mode, and assign a probability estimate for each event.
- An experimental task is evaluated herein for pedestrian prediction, that includes a dataset of cropped images of pedestrians, along with their associated trajectory in world coordinates. This dataset is constructed using data from the nuScenes dataset [5], which combines camera information with trajectories of pedestrians in the world space, to produce an experimental task for prediction of pedestrians including the use of observed appearance.
- This experiment uses a history of images and trajectory positions for a pedestrian, and predicting future positions, evaluated using the multimodal prediction measures. [4],
- This experimental task allows a model to predict modes of behaviour, such as motion initiation and standing still, using the appearance of pedestrians to provide cues when changes of motion take place.
- a network architecture that uses a CNN-based model for interpreting pedestrian appearance is presented herein, which is used along with the DiPA trajectory predictor, and show that this model improves over prediction using trajectory history alone.
- pedestrian appearance such as changes of gait, are informative for estimating the future trajectory modes of pedestrians, using a prediction representation that can be used by an AV planner to control the vehicle while avoiding potential conflicts with pedestrians.
- Multiple prediction models are compared on real data collected from an AV.
- the data includes pedestrian trajectories, and cropped images of appearance for each timestep along the trajectory. Given histories of length Is, trajectory prediction is then evaluated at 1, 2 and 3 s.
- Multi-modal trajectories with spatial distributions are predicted, and evaluated with standard trajectory error measures: minADE/FDE, predRMS, weightRMS and NLL. These measures evaluate closest-mode prediction in addition to probabilistic estimates, which provide complementary evaluations of prediction accuracy, as described in [4], An effective predictor needs to perform well on each measure, indicating the ability to capture distinct modes of behaviour, as well as accurate estimates of the probability that each will occur.
- the use of appearance as an observation can assist with identifying changes of motion, and to focus on this task a dataset selection is created that emphasises instances involving changes of motion, in addition to the full dataset.
- Predictions are produced using 5 modes, which are encoded using a predicted trajectory position for each timestep, as well as a covariance matrix representing the spatial error distribution, and a probability weight for each predicted mode.
- Calculation of the evaluation measures minADE/FDE, predRMS and NLL are described in [4],
- the weightRMS evaluation is used, as this provides an estimate of trajectory prediction error weighted by probability estimates, that improves over the predRMS evaluation as it considers each of the mode prediction weights not just the most probable.
- weightRMS is calculated as follows, for each instance n G N, mode m G M, mode weight W m , ground-truth position x and predicted trajectory position p:
- trajectory-based trajectory predictor which is compared against a number of trajectory-only predictors, including kinematic prediction and a neural -network trajectory predictor (DiPA [4]) that has been demonstrated as effective for prediction of road users including pedestrians.
- the kinematic predictor predicts a single predicted trajectory mode, while the neural network predictor and appearance network produce trajectories with covariances and mode estimates, and are evaluated using multimodal evaluations.
- the first dataset is built from the popular large-scale autonomous driving dataset NuScenes [5], It contains sensor data collected from a fleet of autonomous vehicles operating in various urban environments. It includes 3D trajectory annotations at 2Hz and camera images at a variable rate (10 or 20 Hz). When synchronising trajectory annotations with camera images, trajectories are interpolated at 10 Hz and synchronised with the closest camera images. Pedestrian instances are selected and maintain the original train, validation and test split.
- the second dataset is obtained from data-gathering runs with the Five Al vehicle fleet in urban areas of London and Millbrook (UK). It contains trajectory annotations and camera images at 30 Hz.
- Pedestrian instances are selected and split the datasets into train (70 %), validation (10 %) and test (20 %) based on the pedestrian instance ID.
- These datasets include camera images from different views, e.g. front-left, front-right etc cameras. Since each pedestrian can be visible from multiple views, the datasets contain multiple samples, one for each view.
- cropped regions of images are collected in each available camera view, along with the world trajectory coordinates. The cropped region is expanded to be a loose crop of the pedestrian region, with constant aspect ratio, and rescaled to a common size of 192 x 192.
- the described method processes a sequence of cropped pedestrian images as an input, which is combined with the DiPA [4] trajectory predictor backbone to predict multimodal future trajectories of predictions.
- An overview of the model 400 is shown in Figure 4.
- FIG 4 shows a schematic block diagram of an overview of an appearance-based model 400.
- Visual inputs 402 of pedestrian appearance of objects is encoded per frame by a visual feature extractor 404.
- the visual feature extractor may be a CNN 404.
- the extracted visual features may be interpreted over time using temporal convolutions 406.
- the temporal convolution component 406 may combine the visual features of the visual inputs 402.
- Trajectory history input 408 comprising an observed trajectory is received by a temporal convolution component 410.
- Image and trajectory encodings 412 are combined and decoded by decoding layers 414 to produce predictions of multimodal trajectories 416, covariances 418 and mode probabilities 420 to estimate future motion states of pedestrians.
- the trajectory prediction network (bottom row) uses the DiPA predictor, described with reference to Figure 2, operating with a single agent at a time to focus on the contribution of appearance features for improving multimodal trajectory prediction.
- the proposed model 400 processes a sequence of independent image frames 402 using image features (two-dimensional), without the use of time-based video features (three dimensional including time).
- Image features two-dimensional
- time-based video features three dimensional including time.
- Video-based models of human motion such as C3D [6] or I3D [7] are based on datasets using stationary cameras, and the use on data including cropping errors and camera motion will produce feature responses from background motion and interfere with learning.
- the proposed model 400 uses a sequence of image-based feature processing 404, followed by temporal convolutions 406 to allow inference between frames over time.
- Trajectory histories 408 are provided as input to another temporal convolution component 408.
- Each trajectory history comprises an observed trajectory of an agent in a world space.
- a set of trajectory features are extracted from the trajectory history.
- the encoded features representing appearance are concatenated with the trajectory encoding features 412, and fed into the trajectory decoder 414 to produce predictions of trajectory positions 416 in the world space, covariance estimates 418 and mode probabilities 420.
- the relationship between view orientation and the world space is different in various instances and frames, and it is not feasible to predict trajectory positions or covariances directly from appearance.
- Appearance can inform future motion states, such as initiation of motion, which can be captured in specific predicted behaviour modes.
- the appearance network is trained based on the loss of predicted behaviour modes, and not using the trajectory or covariance losses.
- Each model, appearance- or trajectory -based prediction, is most relevant for a range of cases, and the different models operate based on different information about the scene. Combining the models allows prediction that is better informed than each individual model.
- the output of the model 400 is represented by a set of trajectories 416 with corresponding confidence weights. Each trajectory is described as a series of positions at a number of future time-steps, encoded with a position and covariance matrix. This set of weighted trajectories is an encoding of the predicted probability distribution over space and time.
- the proposed method is a semi-integrated neural -network based approach, where observed trajectory, appearance and/or scene data is provided as input, and a set of predicted trajectories and confidence weights are produced for the agent of interest.
- Trajectory prediction is based on DiPA [4]
- Appearance of objects is encoded as a feature vector extracted from a Convolutional Neural Network (CNN), which learns to predict future motion states from appearance, such as whether a pedestrian moves, or remains stationary. This is trained on motion states automatically estimated from trajectory data.
- Feature responses from the trained network are then fed as input to a joint appearance-trajectory network, with an option of performing integrated training of motion states and trajectory prediction.
- CNN Convolutional Neural Network
- Baselines include a Constant Velocity (CV) and a Decaying Acceleration (DA) model.
- CV Constant Velocity
- DA Decaying Acceleration
- CA Constant Acceleration
- App-net uses a CNN (MobileNetV3 Small [9]) which is trained against the mode prediction loss
- App-pose using pre-calculated pose features [10] which are passed to the temporal convolution layer as a vector of 17 x 2 features of pose positions in the image.
- Figure 5 shows examples of prediction input and output.
- Left images 502a, 502b show cropped appearance data.
- Right images 504a, 504b show ground-truth (past - solid line, future - dotted line) and predicted (dashed line) trajectory data.
- an appearance-based model is introduced that uses the appearance of pedestrians to improve on prediction of motion, compared to prediction using a trajectory only approach. Comparison with a kinematic model shows that the neural -network trajectory predictor improves over the kinematic model, and the appearance-based model improves over the trajectory-only model. This effect is emphasised when a selection of the dataset is used that includes a higher proportion of instances with changes of motion.
- This appearance-based model is useful for prediction of pedestrian motion using camera images from a moving AV, and can support an AV for operating safely when pedestrians are nearby.
- the experiment and described method operates for a single pedestrian at a time. This allows the advantages for prediction from appearance to be demonstrated as an independent task.
- the method can be extended to support the prediction of multiple agents together in a scene, including the use of appearance for each agent.
- multiple pedestrians may be visible together in an instance, and further improvements can provide masking to allow individual agents to be discriminated and to allow for independent prediction of each.
- the first uses classification of a motion state to train the appearance network 600.
- the appearance network 600 may, for example, be separately trained on a motion state prediction task, as shown in Figure 6.
- the training method may be used to train the visual feature extractor 404 described with reference to Figure 4.
- This motion state is based on movement in the world space.
- the second is trained using feedback from the trajectory prediction outputs (mode probabilities, covariances and trajectories).
- the whole network may be jointly trained on trajectory prediction in the world space.
- the motion state is based on movement in the world space.
- the appearance network 600 may produce outputs that classify the motion state of the agent in relation to motion classes, e.g., "moving" vs “not moving”, “turning” vs “not turning” and/or other high-level motion classes. Alternatively or additionally, it may predict a motion value, such as or a real-valued estimate of speed (a form of regression). In this case, the appearance network 600 is trained against the actual future motion of the agent, which is known for the samples in the training set. Motion state ground truth used in training is generated automatically from known trajectory data in the world space, which does not require manual annotation.
- trajectories are defined in the world plane.
- a ground truth may be defined according to whether the pedestrian is moving or not moving in the future (after the observation).
- the appearance network 600 tries to predict if the pedestrian will be moving or not moving. The intention is for the network 600 to identify appearance cues to do this, e.g. it will see that there is a change of pose indicating that they are starting to move, and this will allow classification that the pedestrian will be moving after the observation.
- Training is this context refers to systematic tuning of the parameters of a machine learning (ML) component, such as neural network weights, based on a training loss defined on ground truth and corresponding ML outputs.
- the training loss quantifies error in the ML outputs, by measuring an extend of difference between the ML outputs and the ground truth.
- gradient-based training methods such as stochastic gradient descent/ascent, may be used to tune the parameters based on their gradients with respect of the training gloss,.
- FIG. 1 shows, by way of context, a highly schematic block diagram of an AV runtime stack 100.
- the stack 100 may be fully or semi-autonomous.
- the stack 100 may operate as an Autonomous Driving System (ADS) or Advanced Driver Assist System (ADAS).
- ADS Autonomous Driving System
- ADAS Advanced Driver Assist System
- the run time stack 100 is shown to comprise a perception system 102, a prediction system 104, a planning system (planner) 106 and a control system (controller) 108.
- the prediction system comprises the multi-model prediction system described above, which supports trajectory planning in the presence of pedestrian(s)/VRU(s).
- the perception system 102 receives sensor outputs from an on-board sensor system 110 of the AV, and uses those sensor outputs to detect external agents and measure their physical state, such as their position, velocity, acceleration etc.
- the on-board sensor system 110 can take different forms but generally comprises a variety of sensors such as image capture devices (cameras/optical sensors), lidar and/or radar unit(s), satellitepositioning sensor(s) (GPS etc.), motion/inertial sensor(s) (accelerometers, gyroscopes etc.) etc.
- the onboard sensor system 110 thus provides rich sensor data from which it is possible to extract detailed information about the surrounding environment, and the state of the AV and any external actors (vehicles, pedestrians, cyclists etc.) within that environment.
- the sensor outputs typically comprise sensor data of multiple sensor modalities such as stereo images from one or more stereo optical sensors, lidar, radar etc. Sensor data of multiple sensor modalities may be combined using filters, fusion components etc.
- the perception system 102 typically comprises multiple perception components which cooperate to interpret the sensor outputs and thereby provide perception outputs to the prediction system 104.
- the perception outputs from the perception system 102 are used by the prediction system 104 to predict future behaviour of external actors (agents), such as other vehicles, pedestrians and other VRUs in the vicinity of the AV.
- the stack 100 may be subject to simulationbased testing to verify safety and performance.
- it may or may not be necessary to model the on-board sensor system 100.
- Surrogate model(s) of the perception system 102 can also be used to test planner performance (or planner and controller performance) in the presence of realistic perception errors, without the use of sensor models.
- the perception system 104 may operate on simulated data of various forms.
- Predictions computed by the prediction system 104 are provided to the planner 106, which uses the predictions to make autonomous driving decisions to be executed by the AV in a given driving scenario.
- the inputs received by the planner 106 would typically indicate a drivable area and would also capture predicted movements of any external agents (obstacles, from the AV’s perspective) within the drivable area.
- the driveable area can be determined using perception outputs from the perception system 102 in combination with map information, such as an HD (high definition) map.
- a core function of the planner 106 is the planning of trajectories for the AV (ego trajectories), taking into account predicted agent motion. This may be referred to as trajectory planning.
- a trajectory is planned in order to carry out a desired goal within a scenario. The goal could for example be to enter a roundabout and leave it at a desired exit; to overtake a vehicle in front; or to stay in a current lane at a target speed (lane following).
- the goal may, for example, be determined by an autonomous route planner 120, also referred to as a goal generator 120.
- the controller 108 executes the decisions taken by the planner 106 by providing suitable control signals to an on-board actor system 112 of the AV.
- the planner 106 plans trajectories for the AV and the controller 108 generates control signals to implement the planned trajectories.
- the planner 106 will plan into the future, such that a planned trajectory may only be partially implemented at the control level before a new trajectory is planned by the planner 106.
- the actor system 112 includes “primary” vehicle systems, such as braking, acceleration and steering systems, as well as secondary systems (e.g. signalling, wipers, headlights etc.).
- the planner 106 or controller 108 controls a simulated ego agent in a simulation environment that includes simulated pedestrian(s)/VRU(s).
- a scenario description 116 may be used as a basis for planning and prediction.
- the scenario description 116 is generated using the perception system 102, together with a high-definition (HD) map 114.
- HD high-definition
- the scenario description 116 is, in turn, used as a basis for motion prediction in the prediction system 104, and the resulting motion predictions 118 are used in combination with the scenario description 116 as a basis for planning in the planning system 106.
- a “full” stack typically involves everything from processing and interpretation of low-level sensor data (perception), feeding into primary higher-level functions such as prediction and planning, as well as control logic to generate suitable control signals to implement planninglevel decisions (e.g. to control braking, steering, acceleration etc.).
- level 3 stacks include some logic to implement transition demands and level 4 stacks additionally include some logic for implementing minimum risk maneuvers.
- the stack may also implement secondary control functions e.g. of signalling, headlights, windscreen wipers etc.
- a computer system comprises execution hardware which may be configured to execute the method/algorithmic steps disclosed herein and/or to implement a model trained using the present techniques.
- execution hardware encompasses any form/combination of hardware configured to execute the relevant method/algorithmic steps.
- the execution hardware may take the form of one or more processors, which may be programmable or non-programmable, or a combination of programmable and nonprogrammable hardware may be used. Examples of suitable programmable processors include general purpose processors based on an instruction set architecture, such as CPUs, GPUs/accelerator processors etc.
- Such general-purpose processors typically execute computer readable instructions held in memory coupled to or internal to the processor and carry out the relevant steps in accordance with those instructions.
- Other forms of programmable processors include field programmable gate arrays (FPGAs) having a circuit configuration programmable through circuit description code.
- Examples of nonprogrammable processors include application specific integrated circuits (ASICs). Code, instructions etc. may be stored as appropriate on transitory or non-transitory media (examples of the latter including solid state, magnetic and optical storage device(s) and the like).
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Databases & Information Systems (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Automation & Control Theory (AREA)
- Human Computer Interaction (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Image Analysis (AREA)
Abstract
A computer-implemented method of predicting an agent trajectory in a world space, the method comprising: receiving a visual input comprising an image of the agent in an image space, extracting from the visual input using a visual feature extractor a set of visual features, receiving a trajectory history input comprising an observed trajectory of the agent in the world space, extracting from the trajectory history input a set of trajectory features, and generating, based on the set of visual features and the set of trajectory features, a prediction output comprising a predicted trajectory for the agent in the world space.
Description
Trajectory Prediction for Dynamic Agents
Technical Field
The present disclosure pertains to agent trajectory prediction, which has applications, for example, in autonomous driving and robotics.
Background
Within a robotic system (such as an autonomous driving system), predicted agent trajectories may be used as a basis to autonomously plan an ego trajectory for a robot (e.g. autonomous vehicle), referred to as the ego agent.
Autonomous Vehicles (AV) operate in areas where pedestrians are present, and prediction of pedestrian behaviour is important for avoiding conflict, which is particularly important with vulnerable road users (VRUs). Pedestrian prediction is hard since they can change direction and start or stop moving, and it is high risk, for example if they enter a road area. Producing conservative estimates of pedestrian motion allows potential actions to be captured and avoided, however can lead to very conservative driving of an AV and prevent progress of the AV.
Summary
To address the issue of overly-conservative autonomous driving behaviour in the presence of pedestrians and other vulnerable road users (such as cyclists), trajectory -based predictions techniques are provided herein which incorporate visual appearance information to more accurately identify when changes of motion occur, and thus more accurately predict a future trajectory of a VRU.
Methods which predict future motion based only on an observed history of positions are limited in an important respect. When changes of motion take place, such as initiation of motion to enter a road area from a stationary position, there is a delay before the motion can be accurately observed in the trajectory, and used to make an accurate prediction of motion. Noise is always present in the estimated position, and the greater the noise the later that motion initiation can be reliably observed.
In practice, this means that often motion is only clearly observed in the history of positions after the pedestrian has entered the road, making purely trajectory -based predictors less suitable for VRUs in particular.
This problem is addressed herein using appearance data to inform trajectory prediction. The visual appearance information enables motion changes to be detected earlier than pure trajectory -based methods.
The prediction techniques disclosed herein are particularly well suited to the task of predicting pedestrian trajectories, or trajectories of other vulnerable road users (VRUs), using visual ‘cues’ in combination with historic trajectories.
A first aspect herein provides a computer-implemented method of predicting an agent trajectory in a world space, the method comprising: receiving a visual input comprising an image of the agent in an image space; extracting from the visual input using a visual feature extractor a set of visual features; receiving a trajectory history input comprising an observed trajectory of the agent in the world space; extracting from the trajectory history input a set of trajectory features; and generating, based on the set of visual features and the set of trajectory features, a prediction output comprising a predicted trajectory for the agent in the world space.
Embodiments utilize vision-based classification of motion changes (e.g. motion initiations) to inform trajectory prediction.
Image features in the image space are thus used to guide trajectory prediction in the world space.
The world space (world coordinate system) may, for example, be defined by coordinates within a ground plane (or world plane), which encode a ‘bird’s-eye-view’ (or top-down view) of the agent and its surroundings. The world space may be two-dimensional, or three- dimensional (with the addition of a height coordinate above the ground plane). In this case, the observed and predicted trajectories have coordinates in the ground plane (as possibly height coordinates as well). For example, each trajectory may comprise a sequence of historic/predicted agent locations, each defined by a set of ground plane (and, in some cases, height) coordinates. Trajectory features may be extracted in the ground plane (e.g. via temporal convolution).
The image space may be defined by image coordinates in an image plane lying substantially perpendicular to the ground plane, or more generally non-parallel to it. The image space may be two-dimensional, or three-dimensional with the addition of a depth coordinate.
The image features may be extracted in the image plane (with additional temporal convolution in some embodiments).
The visual input may comprise a time sequence of images of the agent (appearance frames), such as agent images cropped from a video sequence.
The set of trajectory features may, for example, be extracted via one or more temporal convolution operations applied to the trajectory history input.
The visual feature extractor may, for example, have a neural network architecture. For example, in the described embodiments, the visual feature extractor, has the form of a 2D convolutional neural network (CNN) feature extractor that extracts visual features in the image space from each appearance frame independently, and a temporal convolution processing component that combines those visual features.
The 2D CNN feature extractor may, for example, comprise a 2D CNN that processes the appearance frames sequentially, or multiple 2D CNNs that process the appearance frames in parallel.
A second aspect herein provides a computer-implemented method of training a machine learning visual feature extractor, the method comprising: receiving a training image of an agent in an image space; receiving a known trajectory in a world space associated with the training image; automatically determining a motion state ground truth associated with the training image based on the trajectory in the world space; and extracting using the visual feature extractor a set of visual features from the training image; using the extracted visual features to compute a predicted motion state; and training the visual feature extractor based on the predicted motion state and the ground truth motion state.
The training may comprise tuning parameters of the visual feature extractor based on a training loss that measures error in the predicted motion state relative to the ground truth motion state.
The groundtruth/predicted motion states may be, e.g. motion classes (motion classification) or numerical motion value (motion regression).
Brief description of figures
For a better understanding of the present subject matter, illustrative embodiments are described below by way of example only.
Figure 1 shows a schematic block diagram of an autonomous vehicle stack;
Figure 2 illustrates a schematic block diagram of a neural network architecture for interactive prediction;
Figure 3 is an example of cropped pedestrian appearance, showing gait change with motion initiation;
Figure 4 is a schematic block diagram of an appearance-based model for predicting agent trajectories;
Figure 5 shows examples of prediction inputs and outputs; and
Figure 6 is a schematic block diagram of a model for motion state prediction.
Detailed Description
The ability to anticipate pedestrian motion changes is a critical capability for autonomous vehicles. In urban environments, pedestrians may begin to move to legally cross the road, or may walk in the roadway without considering rules and oncoming traffic. Motion is only clearly observed after the pedestrian has entered the road, making purely trajectory -based predictors less suitable. It is recognized herein that appearance data can inform trajectory prediction and classification of motion changes, e.g. motion initiations, earlier. Many prediction methods have been proposed, but they might not consider appearance data or produce diverse probabilistic trajectories. A comparative evaluation of multiple methods is detailed. Methods are presented herein that can predict trajectories and motion states from trajectory and appearance data, showing the benefits of considering multiple input and output modalities. Two trajectory and image datasets from the popular
NuScenes dataset are created and from observations on real-world autonomous vehicle runs, analysing instances with motion changes.
Figure 2 shows a neural network-based, multi-modal prediction architecture referred to as DiPA (Diverse and Probabilistically Accurate Interactive Prediction), which generates predicted trajectories, with spatial distribution parameters (e.g. mean and covariance) and mode prediction weights for different prediction modes.
The present prediction architecture is incorporated in the DiPA architecture in one embodiment.
Figure 2 is a schematic block diagram of an example neural network architecture that may be used to perform interaction-based agent trajectory prediction, for example to predict agent trajectories for agents of a driving scenario based on which an autonomous vehicle stack can plan a safe trajectory. The network receives a set of trajectory histories 206 comprising a set of past states for each agent of a scenario. Each trajectory history may be in the form of a time series of agent states over some past time interval, each state including the position, orientation and velocity of the agent at a corresponding point in time. At least one past agent state is received for each agent. Other agent inputs 204 may also be received, such as a set of spatial dimensions for each agent. These agent inputs and agent histories may be derived from sensor data by a perception subsystem 102 of an AV stack 100. The trajectory histories are processed by one or more temporal convolution layers 208 to generate a feature vector for each agent that is time-independent.
As shown in Figure 2, the set of trajectory histories 206 is an array having shape (agents, time, features), i.e. for each agent, a set of features is input defining the state of the agent at each timestep of the time interval. Note that ‘features’ is used herein to refer to a representative set of values, and different feature representations are used in different parts of the network. In other words, ‘features’ when used to describe a dimension of an input or output of the network can refer to different numbers and types of features for different inputs and outputs. For the trajectory history, for example, the feature values represent the position, orientation and velocity of the agent at the selected point in time, and the features output from the temporal convolution represent the states of the agent over the entire time interval. Note that the dimensionality shown in Figure 2 excludes the feature dimension.
The convolved trajectory histories are broadcast over all agents as shown by the 0 symbol, for example by concatenating the feature vector for each agent with each feature vector
associated with each other agent of the scenario. Each combination of two agents is associated with a respective pairwise feature, which is processed by one or more interaction layers 210 of the neural network. These may be fully connected (FC) layers as shown. The output of the interaction layers 210 is a respective feature vector for each pairwise combination of agents, which is subsequently reduced (aggregated) over one agent dimension. Note that each agent is treated as an independent input to the network, similarly to elements of a batch, rather than as components of a single input that the network learns to process according to an assigned role.
For clarity, herein a first agent of each pair is referred to as the ‘reference’ agent while the second agent of the pair is referred to as the ‘comparison’ agent. All agents of the scenario act as both a reference agent and a comparison agent. The reduction is over the comparison agent dimension, such that a respective interaction representation is output for each agent of the scenario. Example reduction operations include max reductions, where for a given reference agent the maximum value for each feature is selected over all comparison agents, and a sum, which gives the sum of each feature over all the comparison agents. The reduced interaction representation feature vectors, which have dimension (agents, features) is combined with the additional agent inputs, as well as the convolved histories, which also have dimension (agents, features), although, as noted above, the number of features for the interaction representation and the agent inputs need not be the same. These are combined in the present example by concatenating the agent input (combined with the convolved agent histories) and the interaction representation vector for each agent as shown by the © in Figure 2. This combined feature representation for each agent is processed by a set of prediction layers 212 and 216. A scene context may also be generated by reducing an intermediate output of a first set of prediction layers 212 and processing this in a set of context layers 214, with the output being broadcast to the intermediate output for each agent, for example by concatenating the scene context with each intermediate output as generated for each agent, before processing the combined agent representations in a second set of prediction layers 216 to generate a final predicted output including a trajectory prediction 218. In the present example, the neural network is configured to make predictions for each of a fixed number of prediction ‘modes’, where the number of modes, for example five, is predetermined. The output generates a predicted trajectory 218 for each mode, as well as a spatial distribution 220 indicating uncertainty of the trajectory itself in space, as well as a weight 222 (such as a mode
probability) which indicates a predicted mode, where the mode with the highest weight is the mode that the network determines is the most likely for the given agent.
Reference is made to our co-pending United Kingdom Patent Application No. 2208732.4 filed on 14 June 2022, and to [4], each of which is incorporated herein by reference in its entirety. Further details of the DiPA architecture are described therein.
The DiPA architecture provides an effective mechanism for evaluating multimodal predictions in the context of prediction of road users, which considers the various possibilities that should be considered, and evaluating the accuracy of the distribution of predictions. In the present context, the usefulness of this architecture for pedestrian motion prediction, such as changes of motion, is recognized as it provides a general method for evaluating a distribution of future multimodal states.
Note that, whilst the ‘full’ DiPA architecture includes the extraction of pairwise interaction features (‘agent x agent’), the use of this is optional. An interaction layer 210 is shown as the top layer at the top of Figure 2, but this layer is not necessarily required in the present context. Hence, the interaction layer 210 may be omitted, or deactivated. In this case, agents are considered individually, without explicitly modelling agent interactions. Figure 2 also shows a global feature layer at the bottom, which can also be omitted or deactivated in this case. This layer combines the features between the various agents in the scene, and is not required when only handling a single agent at a time.
As an alternative to deactivating the interaction layer, it could be retained, but configured to handle only a single agent at a time (the DiPA architecture is capable of handling a varying number of agents, including a single agent at a time, and can therefore be configured to operate with the number of agents set to one).
Experiments are described below, which have been run on a single agent at a time to highlight the effect of introducing appearance for prediction of individual agents.
However, in a multi-agent setting, the interaction and global layers may be retained, handling multiple agents at a time, making full use of the interactive DiPA architecture.
A middle agent prediction layer is shown to have three inputs: trajectory history, agent inputs (as in DiPA), and the additional visual features. The agent inputs may, for example, comprise agent dimensions. For pedestrian/VRU prediction, these can be set to a reasonably representative and predetermined value (e.g. 1 meter x 1 meter).
The outputs to the network are unchanged.
Appearance cues such as changes of body pose provide additional information of the actions of pedestrians, such as when gait is changing in order to begin or stop moving. These appearance cues can reliably inform when motion changes are taking place, and provide an early and accurate signal of motion. Figure 3 illustrates an example pedestrian appearance, showing gait change with motion initiation.
A first image 302 shows a pedestrian waiting at a crossing on a pavement at the side of a road. In this image, both feet of the pedestrian are stood on the pavement. A second image 304 shows the pedestrian moving closer to the edge of the pavement. In a third image 306, the pedestrian is shown to be at the edge of the pavement. In this image, the pedestrian is stood such that they appear to intend on walking closer the edge of the pavement, towards the road. In a fourth image 308, the pedestrian is seen to leave the pavement and cross the edge such that they have now stepped onto the road.
The motion cues in the images presented in Figure 3 may be used by a trained neural network to inform trajectory -based motion prediction. For example, a visual feature extractor described below with reference to Figure 4 may use these images to predict the motion of a pedestrian.
Existing methods have used pedestrian appearance to estimate whether a pedestrian intends to cross the road, and to inform prediction of the future position of the pedestrian in the camera view. Common datasets for these experiments are PIE [1] and JAAD [2], [3], These methods have demonstrated the ability to classify pedestrian crossing intent, and to improve prediction of future positions within the camera view.
The way that these experiments can contribute to the operation of an autonomous vehicle is not clear, as the way that classification of pedestrian intent and prediction within the camera space can inform the control of an AV has not been shown, and further stages of processing would be needed. Pedestrian intent is judgement-based, and needs to be manually annotated for each instance of the dataset. Prediction within the camera frame would need to be performed before fusion of information between various camera views and modalities, and it is unclear how well it can inform the future position of a pedestrian in the world space.
Prediction of pedestrian motion is inherently a multimodal task- if a pedestrian is standing beside the road area there are at least two significant possibilities to consider, of whether they
remain stationary or begin moving into the road, which would create a hazard for the AV. A multimodal predictor can create a predicted trajectory for each mode, and assign a probability estimate for each event.
An experimental task is evaluated herein for pedestrian prediction, that includes a dataset of cropped images of pedestrians, along with their associated trajectory in world coordinates. This dataset is constructed using data from the nuScenes dataset [5], which combines camera information with trajectories of pedestrians in the world space, to produce an experimental task for prediction of pedestrians including the use of observed appearance. This experiment uses a history of images and trajectory positions for a pedestrian, and predicting future positions, evaluated using the multimodal prediction measures. [4], This experimental task allows a model to predict modes of behaviour, such as motion initiation and standing still, using the appearance of pedestrians to provide cues when changes of motion take place.
To solve this task a network architecture that uses a CNN-based model for interpreting pedestrian appearance is presented herein, which is used along with the DiPA trajectory predictor, and show that this model improves over prediction using trajectory history alone. This shows that pedestrian appearance such as changes of gait, are informative for estimating the future trajectory modes of pedestrians, using a prediction representation that can be used by an AV planner to control the vehicle while avoiding potential conflicts with pedestrians.
Experiments
Multiple prediction models are compared on real data collected from an AV. The data includes pedestrian trajectories, and cropped images of appearance for each timestep along the trajectory. Given histories of length Is, trajectory prediction is then evaluated at 1, 2 and 3 s. Multi-modal trajectories with spatial distributions are predicted, and evaluated with standard trajectory error measures: minADE/FDE, predRMS, weightRMS and NLL. These measures evaluate closest-mode prediction in addition to probabilistic estimates, which provide complementary evaluations of prediction accuracy, as described in [4], An effective predictor needs to perform well on each measure, indicating the ability to capture distinct modes of behaviour, as well as accurate estimates of the probability that each will occur.
The use of appearance as an observation can assist with identifying changes of motion, and to focus on this task a dataset selection is created that emphasises instances involving changes of motion, in addition to the full dataset.
Predictions are produced using 5 modes, which are encoded using a predicted trajectory position for each timestep, as well as a covariance matrix representing the spatial error distribution, and a probability weight for each predicted mode. Calculation of the evaluation measures minADE/FDE, predRMS and NLL are described in [4], In addition, the weightRMS evaluation is used, as this provides an estimate of trajectory prediction error weighted by probability estimates, that improves over the predRMS evaluation as it considers each of the mode prediction weights not just the most probable. weightRMS is calculated as follows, for each instance n G N, mode m G M, mode weight Wm, ground-truth position x and predicted trajectory position p:
Experiments are conducted using the proposed appearance-based trajectory predictor, which is compared against a number of trajectory-only predictors, including kinematic prediction and a neural -network trajectory predictor (DiPA [4]) that has been demonstrated as effective for prediction of road users including pedestrians. The kinematic predictor predicts a single predicted trajectory mode, while the neural network predictor and appearance network produce trajectories with covariances and mode estimates, and are evaluated using multimodal evaluations.
Datasets
Two distinct autonomous driving datasets are analysed. The first dataset is built from the popular large-scale autonomous driving dataset NuScenes [5], It contains sensor data collected from a fleet of autonomous vehicles operating in various urban environments. It includes 3D trajectory annotations at 2Hz and camera images at a variable rate (10 or 20 Hz). When synchronising trajectory annotations with camera images, trajectories are interpolated at 10 Hz and synchronised with the closest camera images. Pedestrian instances are selected and maintain the original train, validation and test split. The second dataset is obtained from data-gathering runs with the Five Al vehicle fleet in urban areas of London and Millbrook (UK). It contains trajectory annotations and camera images at 30 Hz. Pedestrian instances are
selected and split the datasets into train (70 %), validation (10 %) and test (20 %) based on the pedestrian instance ID. These datasets include camera images from different views, e.g. front-left, front-right etc cameras. Since each pedestrian can be visible from multiple views, the datasets contain multiple samples, one for each view. From each pedestrian instance, cropped regions of images are collected in each available camera view, along with the world trajectory coordinates. The cropped region is expanded to be a loose crop of the pedestrian region, with constant aspect ratio, and rescaled to a common size of 192 x 192.
Experimental Method(s)
The described method processes a sequence of cropped pedestrian images as an input, which is combined with the DiPA [4] trajectory predictor backbone to predict multimodal future trajectories of predictions. An overview of the model 400 is shown in Figure 4.
Figure 4 shows a schematic block diagram of an overview of an appearance-based model 400. Visual inputs 402 of pedestrian appearance of objects is encoded per frame by a visual feature extractor 404. The visual feature extractor may be a CNN 404. The extracted visual features may be interpreted over time using temporal convolutions 406. For example, the temporal convolution component 406 may combine the visual features of the visual inputs 402. Trajectory history input 408 comprising an observed trajectory is received by a temporal convolution component 410. Image and trajectory encodings 412 are combined and decoded by decoding layers 414 to produce predictions of multimodal trajectories 416, covariances 418 and mode probabilities 420 to estimate future motion states of pedestrians. The trajectory prediction network (bottom row) uses the DiPA predictor, described with reference to Figure 2, operating with a single agent at a time to focus on the contribution of appearance features for improving multimodal trajectory prediction.
One consideration for observing object appearance from the point of view of an autonomous vehicle, is that the camera is moving with the vehicle, and the detected region of each identified pedestrian will contain errors, resulting in visual effects such as background motion and misalignment between sequential frames, which can interfere with processing of visual features. To compensate for these effects, the proposed model 400 processes a sequence of independent image frames 402 using image features (two-dimensional), without the use of time-based video features (three dimensional including time). Video-based models of human motion such as C3D [6] or I3D [7] are based on datasets using stationary cameras, and the use on data including cropping errors and camera motion will produce feature responses from
background motion and interfere with learning. Instead the proposed model 400 uses a sequence of image-based feature processing 404, followed by temporal convolutions 406 to allow inference between frames over time.
Trajectory histories 408 are provided as input to another temporal convolution component 408. Each trajectory history comprises an observed trajectory of an agent in a world space. A set of trajectory features are extracted from the trajectory history.
The encoded features representing appearance are concatenated with the trajectory encoding features 412, and fed into the trajectory decoder 414 to produce predictions of trajectory positions 416 in the world space, covariance estimates 418 and mode probabilities 420. As a moving camera is used the relationship between view orientation and the world space is different in various instances and frames, and it is not feasible to predict trajectory positions or covariances directly from appearance. Appearance can inform future motion states, such as initiation of motion, which can be captured in specific predicted behaviour modes. To allow this to be captured by the network, the appearance network is trained based on the loss of predicted behaviour modes, and not using the trajectory or covariance losses.
Each model, appearance- or trajectory -based prediction, is most relevant for a range of cases, and the different models operate based on different information about the scene. Combining the models allows prediction that is better informed than each individual model. The output of the model 400 is represented by a set of trajectories 416 with corresponding confidence weights. Each trajectory is described as a series of positions at a number of future time-steps, encoded with a position and covariance matrix. This set of weighted trajectories is an encoding of the predicted probability distribution over space and time. The proposed method is a semi-integrated neural -network based approach, where observed trajectory, appearance and/or scene data is provided as input, and a set of predicted trajectories and confidence weights are produced for the agent of interest. Trajectory prediction is based on DiPA [4], Appearance of objects is encoded as a feature vector extracted from a Convolutional Neural Network (CNN), which learns to predict future motion states from appearance, such as whether a pedestrian moves, or remains stationary. This is trained on motion states automatically estimated from trajectory data. Feature responses from the trained network are then fed as input to a joint appearance-trajectory network, with an option of performing integrated training of motion states and trajectory prediction.
Experimental Results
Multiple methods are compared on the two presented datasets using standard trajectory error metrics. Baselines include a Constant Velocity (CV) and a Decaying Acceleration (DA) model. In past works, , e.g. [8], the DA model was shown to improve over CV and Constant Acceleration (CA) since a constant acceleration model can be more accurate than a constant velocity model in the short term, but it can be very inaccurate in the long term. Considering accelerations can help predicting motion profile changes, e.g. motion initiations, but its estimates might not be accurate. Two implementations are tested, one (App-net) uses a CNN (MobileNetV3 Small [9]) which is trained against the mode prediction loss, and a second implementation (App-pose) using pre-calculated pose features [10] which are passed to the temporal convolution layer as a vector of 17 x 2 features of pose positions in the image.
These appearance-based predictors allow visual cues to influence predicted trajectories, and the estimates of probabilities of each trajectory mode. Results are reported in Tables I and II.
TABLE I: Comparison on the NuScenes-Appearance dataset.
TABLE II: Comparison on the Five Al dataset.
Qualitative examples are included in Figure 5. Figure 5 shows examples of prediction input and output. Left images 502a, 502b show cropped appearance data. Right images 504a, 504b show ground-truth (past - solid line, future - dotted line) and predicted (dashed line) trajectory data.
Conclusions
In order to operate an autonomous vehicle in the vicinity of pedestrians, it is important to be able to estimate their future motion, and particularly to identify significant cases such as changes of motion, which can indicate when they may enter the road area. To focus on this problem a new dataset task is introduced to perform estimation of multimodal trajectories, using pedestrian appearance to inform future motion. This experiment task makes new use of the nuScenes dataset, and provides a useful means of training and evaluating prediction that can support the use of an AV. This task improves over previous datasets such as PIE and JAAD, which are limited to the camera frame, by evaluating prediction of motion in the world space, including evaluation of probabilistic estimates of different modes of motion.
To solve this task an appearance-based model is introduced that uses the appearance of pedestrians to improve on prediction of motion, compared to prediction using a trajectory
only approach. Comparison with a kinematic model shows that the neural -network trajectory predictor improves over the kinematic model, and the appearance-based model improves over the trajectory-only model. This effect is emphasised when a selection of the dataset is used that includes a higher proportion of instances with changes of motion.
This appearance-based model is useful for prediction of pedestrian motion using camera images from a moving AV, and can support an AV for operating safely when pedestrians are nearby.
The experiment and described method operates for a single pedestrian at a time. This allows the advantages for prediction from appearance to be demonstrated as an independent task. The method can be extended to support the prediction of multiple agents together in a scene, including the use of appearance for each agent. In addition, currently multiple pedestrians may be visible together in an instance, and further improvements can provide masking to allow individual agents to be discriminated and to allow for independent prediction of each.
Two variations are considered by way of example. The first uses classification of a motion state to train the appearance network 600. In this case, the appearance network 600 may, for example, be separately trained on a motion state prediction task, as shown in Figure 6. The training method may be used to train the visual feature extractor 404 described with reference to Figure 4.
This motion state is based on movement in the world space.
The second is trained using feedback from the trajectory prediction outputs (mode probabilities, covariances and trajectories). In this case, the whole network may be jointly trained on trajectory prediction in the world space.
In the first case, the motion state is based on movement in the world space. When using motion state classification, the appearance network 600 may produce outputs that classify the motion state of the agent in relation to motion classes, e.g., "moving" vs "not moving", "turning" vs "not turning" and/or other high-level motion classes. Alternatively or additionally, it may predict a motion value, such as or a real-valued estimate of speed (a form of regression). In this case, the appearance network 600 is trained against the actual future motion of the agent, which is known for the samples in the training set. Motion state ground
truth used in training is generated automatically from known trajectory data in the world space, which does not require manual annotation.
In the above examples, trajectories are defined in the world plane. For the purpose of training on motion states, a ground truth may be defined according to whether the pedestrian is moving or not moving in the future (after the observation). The appearance network 600 tries to predict if the pedestrian will be moving or not moving. The intention is for the network 600 to identify appearance cues to do this, e.g. it will see that there is a change of pose indicating that they are starting to move, and this will allow classification that the pedestrian will be moving after the observation.
Training is this context refers to systematic tuning of the parameters of a machine learning (ML) component, such as neural network weights, based on a training loss defined on ground truth and corresponding ML outputs. The training loss quantifies error in the ML outputs, by measuring an extend of difference between the ML outputs and the ground truth. For example, gradient-based training methods, such as stochastic gradient descent/ascent, may be used to tune the parameters based on their gradients with respect of the training gloss,.
Example Application
Figure 1 shows, by way of context, a highly schematic block diagram of an AV runtime stack 100. The stack 100 may be fully or semi-autonomous. For example, the stack 100 may operate as an Autonomous Driving System (ADS) or Advanced Driver Assist System (ADAS).
The run time stack 100 is shown to comprise a perception system 102, a prediction system 104, a planning system (planner) 106 and a control system (controller) 108.
The prediction system comprises the multi-model prediction system described above, which supports trajectory planning in the presence of pedestrian(s)/VRU(s).
In a real -world context, the perception system 102 receives sensor outputs from an on-board sensor system 110 of the AV, and uses those sensor outputs to detect external agents and measure their physical state, such as their position, velocity, acceleration etc. The on-board sensor system 110 can take different forms but generally comprises a variety of sensors such as image capture devices (cameras/optical sensors), lidar and/or radar unit(s), satellitepositioning sensor(s) (GPS etc.), motion/inertial sensor(s) (accelerometers, gyroscopes etc.) etc. The onboard sensor system 110 thus provides rich sensor data from which it is possible
to extract detailed information about the surrounding environment, and the state of the AV and any external actors (vehicles, pedestrians, cyclists etc.) within that environment. The sensor outputs typically comprise sensor data of multiple sensor modalities such as stereo images from one or more stereo optical sensors, lidar, radar etc. Sensor data of multiple sensor modalities may be combined using filters, fusion components etc.
The perception system 102 typically comprises multiple perception components which cooperate to interpret the sensor outputs and thereby provide perception outputs to the prediction system 104. The perception outputs from the perception system 102 are used by the prediction system 104 to predict future behaviour of external actors (agents), such as other vehicles, pedestrians and other VRUs in the vicinity of the AV.
As is increasingly common in the AV space, the stack 100 may be subject to simulationbased testing to verify safety and performance. In a simulation context, depending on the nature of the testing, it may or may not be necessary to model the on-board sensor system 100. For example, when only the planner 106 (or only the planner 106 and controller 108) are tested on a ‘perfect’ representation of the scenario (that is, directly on simulator ground truth), simulated sensor data is not required therefore complex sensor modelling is not required. Surrogate model(s) of the perception system 102 (or part(s) of it) can also be used to test planner performance (or planner and controller performance) in the presence of realistic perception errors, without the use of sensor models.
In a simulation-based testing context, the perception system 104 may operate on simulated data of various forms.
Predictions computed by the prediction system 104 are provided to the planner 106, which uses the predictions to make autonomous driving decisions to be executed by the AV in a given driving scenario. The inputs received by the planner 106 would typically indicate a drivable area and would also capture predicted movements of any external agents (obstacles, from the AV’s perspective) within the drivable area. The driveable area can be determined using perception outputs from the perception system 102 in combination with map information, such as an HD (high definition) map.
A core function of the planner 106 is the planning of trajectories for the AV (ego trajectories), taking into account predicted agent motion. This may be referred to as trajectory planning. A trajectory is planned in order to carry out a desired goal within a scenario. The goal could for example be to enter a roundabout and leave it at a desired exit;
to overtake a vehicle in front; or to stay in a current lane at a target speed (lane following). The goal may, for example, be determined by an autonomous route planner 120, also referred to as a goal generator 120.
The controller 108 executes the decisions taken by the planner 106 by providing suitable control signals to an on-board actor system 112 of the AV. In particular, the planner 106 plans trajectories for the AV and the controller 108 generates control signals to implement the planned trajectories. Typically, the planner 106 will plan into the future, such that a planned trajectory may only be partially implemented at the control level before a new trajectory is planned by the planner 106. The actor system 112 includes “primary” vehicle systems, such as braking, acceleration and steering systems, as well as secondary systems (e.g. signalling, wipers, headlights etc.).
In simulation-based testing, the planner 106 or controller 108 controls a simulated ego agent in a simulation environment that includes simulated pedestrian(s)/VRU(s).
Within the stack 100, a scenario description 116 may be used as a basis for planning and prediction. The scenario description 116 is generated using the perception system 102, together with a high-definition (HD) map 114. By localizing the ego vehicle 114 on the map, it is possible to combine the information extracted in the perception system 104 (including dynamic agent information) with the pre-existing environmental information contained in the HD map 114. The scenario description 116 is, in turn, used as a basis for motion prediction in the prediction system 104, and the resulting motion predictions 118 are used in combination with the scenario description 116 as a basis for planning in the planning system 106.
A “full” stack typically involves everything from processing and interpretation of low-level sensor data (perception), feeding into primary higher-level functions such as prediction and planning, as well as control logic to generate suitable control signals to implement planninglevel decisions (e.g. to control braking, steering, acceleration etc.). For autonomous vehicles, level 3 stacks include some logic to implement transition demands and level 4 stacks additionally include some logic for implementing minimum risk maneuvers. The stack may also implement secondary control functions e.g. of signalling, headlights, windscreen wipers etc.
References herein to components, functions, modules and the like, denote functional components of a computer system which may be implemented at the hardware level in
various ways. A computer system comprises execution hardware which may be configured to execute the method/algorithmic steps disclosed herein and/or to implement a model trained using the present techniques. The term execution hardware encompasses any form/combination of hardware configured to execute the relevant method/algorithmic steps. The execution hardware may take the form of one or more processors, which may be programmable or non-programmable, or a combination of programmable and nonprogrammable hardware may be used. Examples of suitable programmable processors include general purpose processors based on an instruction set architecture, such as CPUs, GPUs/accelerator processors etc. Such general-purpose processors typically execute computer readable instructions held in memory coupled to or internal to the processor and carry out the relevant steps in accordance with those instructions. Other forms of programmable processors include field programmable gate arrays (FPGAs) having a circuit configuration programmable through circuit description code. Examples of nonprogrammable processors include application specific integrated circuits (ASICs). Code, instructions etc. may be stored as appropriate on transitory or non-transitory media (examples of the latter including solid state, magnetic and optical storage device(s) and the like).
References
[1] Amir Rasouli, luliia Kotseruba, Toni Kunic, and John K. Tsotsos. Pie: A large-scale dataset and models for pedestrian intention estimation and trajectory prediction. In The IEEE International Conference on Computer Vision (ICCV), October 2019.
[2] A. Rasouli, I. Kotseruba, and J. K. Tsotsos. Agreeing to cross: How drivers and pedestrians communicate. In 2017 IEEE Intelligent Vehicles Symposium (IV), pages 264-269, 2017.
[3] Amir Rasouli, luliia Kotseruba, and John K Tsotsos. Are they going to cross? a benchmark dataset and baseline for pedestrian crosswalk behavior. In Proceedings of the IEEE International Conference on Computer Vision Workshops, pages 206-213, 2017.
[4] Anthony Knittel, Majd Hawasly, Stefano V Albrecht, John Redford, and Subramanian Ramamoorthy. DiPA: Diverse and probabilistically accurate interactive prediction. arXiv preprint arXiv:2210.06106, 2022.
[5] Holger Caesar, Varun Bankiti, Alex H Lang, Sourabh Vora, Venice Erin Liong, Qiang Xu, Anush Krishnan, Yu Pan, Giancarlo Baldan, and Oscar Beijbom. nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 11621-11631, 2020.
[6] Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE international conference on computer vision, pages 4489-4497, 2015.
[7] Joao Carreira and Andrew Zisserman. Quo vadis, action recognition? a new model and the kinetics dataset. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 6299-6308, 2017.
[8] Morris Antonello, Mihai Dobre, Stefano V Albrecht, John Redford, and Subramanian Ramamoorthy. Flash: Fast and light motion prediction for autonomous driving with bayesian inverse planning and learned motion profiles. In 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 9829- 9836. IEEE, 2022.
[9] Andrew Howard, Mark Sandler, Grace Chu, Liang-Chieh Chen, Bo Chen, Mingxing Tan, Weijun Wang, Yukun Zhu, Ruoming Pang, Vijay Vasudevan, et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF international conference on computer vision, pages 1314-1324, 2019.
[10] Rishabh Bajpai and Deepak Joshi. Movenet: A deep neural network for joint profile prediction across variable walking speeds and slopes. IEEE Transactions on Instrumentation and Measurement, 70: 1-11, 2021.
Claims
1. A computer-implemented method of predicting an agent trajectory in a world space, the method comprising: receiving a visual input comprising an image of the agent in an image space; extracting from the visual input using a visual feature extractor a set of visual features; receiving a trajectory history input comprising an observed trajectory of the agent in the world space; extracting from the trajectory history input a set of trajectory features; and generating, based on the set of visual features and the set of trajectory features, a prediction output comprising a predicted trajectory for the agent in the world space.
2. The method of claim 1, wherein the method comprises encoding the set of visual features with the set of trajectory features to calculate a set of encoded features.
3. The method of claim 2, wherein the method comprises decoding the set of encoded features to generate the prediction output.
4. The method of any preceding claim, wherein the prediction output additionally comprises covariances and mode probabilities.
5. The method of any preceding claim, wherein the visual feature extractor extracts the set of visual features in the image space from each visual input independently; and a temporal convolution processing component combines those visual features.
6. The method of claim 5, wherein the visual feature extractor processes the visual inputs sequentially.
7. The method of any preceding claim, wherein the world space is defined by coordinates within a ground plane.
8. The method of any preceding claim, wherein the world space is two-dimensional.
9. The method of claim 7, wherein the world space is three-dimensional having a height coordinate above the ground plane.
10. The method of claims 7 or 8, wherein the observed trajectory and the predicted trajectory have coordinates in the ground plane.
11. The method of any preceding claim, wherein the predicted trajectory comprises a sequence of historic agent locations, each defined by a set of ground plane coordinates.
12. The method of any of claims 1 to 10, wherein the predicted trajectory comprises a sequence of predicted agent locations, each defined by a set of ground plane coordinates.
13. The method of any of claims 7 to 12, wherein the trajectory features are extracted in the ground plane.
14. The method of any of claims 5 to 13, wherein the image space is defined by image coordinates in an image plane lying substantially perpendicular to the ground plane.
15. The method of any preceding claim, wherein the image space is two-dimensional,
16. The method of any of claims 1 to 14, wherein the image space is three-dimensional.
17. The method of any of claims 14 to 16, wherein the image features are extracted in the image plane.
18. The method of any preceding claim, wherein the visual feature extractor has a neural network architecture.
19. The method of claim 18, wherein the visual feature extractor is a 2D convolutional neural network feature extractor.
20. The method of any preceding claim, wherein the visual input comprises a time sequence of images of the agent.
21. The method of claim 20, wherein the time sequence of images of the agent are agent images cropped from a video sequence.
22. The method of claim 20, wherein the visual feature extractor comprises multiple 2D convolutional neural networks that processes the time sequence of images of the agent in parallel.
23. A computer-implemented method of training a machine learning visual feature extractor, the method comprising: receiving a training image of an agent in an image space; receiving a known trajectory in a world space associated with the training image; automatically determining a motion state ground truth associated with the training image based on the trajectory in the world space; and
extracting using the visual feature extractor a set of visual features from the training image; using the extracted visual features to compute a predicted motion state; and training the visual feature extractor based on the predicted motion state and the ground truth motion state.
24. The method of claim 23, wherein the training comprises tuning parameters of the visual feature extractor based on a training loss that measures error in the predicted motion state relative to the ground truth motion state.
25. The method of claims 23 or 24, wherein the motion state ground truth is a motion class.
26. The method of claims 23 or 24, wherein the motion state ground truth is a numerical motion value.
27. The method of any of claims 23 to 26, wherein the predicted motion state is a motion class.
28. The method of any of claims 23 to 26, wherein the predicted motion state is a numeral motion value.
29. The method of any of claims 1 to 22, wherein the visual feature extractor is trained in accordance with claims 23 to 27.
30. A computer program comprising executable instructions configured, when executed on one or more hardware processors, to implement the method of any preceding claim.
31. A computer system, comprising one or more hardware processors configured to implement the method of any of claims 1 to 29.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| GB2306327.4 | 2023-04-28 | ||
| GBGB2306327.4A GB202306327D0 (en) | 2023-04-28 | 2023-04-28 | Trajectory prediction for dynamic agents |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2024223902A1 true WO2024223902A1 (en) | 2024-10-31 |
Family
ID=86692096
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/EP2024/061674 Pending WO2024223902A1 (en) | 2023-04-28 | 2024-04-26 | Trajectory prediction for dynamic agents |
Country Status (2)
| Country | Link |
|---|---|
| GB (1) | GB202306327D0 (en) |
| WO (1) | WO2024223902A1 (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN120929354A (en) * | 2025-10-14 | 2025-11-11 | 杭州秋果计划科技有限公司 | Code verification method and device |
Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113077489A (en) * | 2021-04-21 | 2021-07-06 | 中国第一汽车股份有限公司 | Pedestrian trajectory prediction method, device, equipment and storage medium |
-
2023
- 2023-04-28 GB GBGB2306327.4A patent/GB202306327D0/en not_active Ceased
-
2024
- 2024-04-26 WO PCT/EP2024/061674 patent/WO2024223902A1/en active Pending
Patent Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113077489A (en) * | 2021-04-21 | 2021-07-06 | 中国第一汽车股份有限公司 | Pedestrian trajectory prediction method, device, equipment and storage medium |
Non-Patent Citations (15)
| Title |
|---|
| A. RASOULII. KOTSERUBAJ. K. TSOTSOS: "Agreeing to cross: How drivers and pedestrians communicate", IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2017, pages 264 - 269 |
| AMIR RASOULIIULIIA KOTSERUBAJOHN K TSOTSOS: "Are they going to cross? a benchmark dataset and baseline for pedestrian crosswalk behavior", PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, 2017, pages 206 - 213, XP033303459, DOI: 10.1109/ICCVW.2017.33 |
| AMIR RASOULIIULIIA KOTSERUBATONI KUNICJOHN K. TSOTSOS: "Pie: A large-scale dataset and models for pedestrian intention estimation and trajectory prediction", THE IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), October 2019 (2019-10-01) |
| ANDREW HOWARDMARK SANDLERGRACE CHULIANG-CHIEH CHENBO CHENMINGXING TANWEIJUN WANGYUKUN ZHURUOMING PANGVIJAY VASUDEVAN ET AL.: "Searching for mobilenetv3", PROCEEDINGS OF THE IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, 2019, pages 1314 - 1324 |
| ANTHONY KNITTEL ET AL: "DiPA: Probabilistic Multi-Modal Interactive Prediction for Autonomous Driving", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 8 March 2023 (2023-03-08), XP091454733 * |
| ANTHONY KNITTELMAJD HAWASLYSTEFANO V ALBRECHTJOHN REDFORDSUBRAMANIAN RAMAMOORTHY: "DiPA: Diverse and probabilistically accurate interactive prediction", ARXIV: 2210. 06106, 2022 |
| DU TRANLUBOMIR BOURDEVROB FERGUSLORENZO TORRESANIMANOHAR PALURI: "Learning spatiotemporal features with 3d convolutional networks", PROCEEDINGS OF THE IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION, 2015, pages 4489 - 4497 |
| HOLGER CAESAR, VARUN BANKITI, ALEX H LANG, SOURABH VORA, VENICE ERIN LIONG, QIANG XU, ANUSH KRISHNAN, YU PAN, GIANCARLO BALDAN, OS: "nuscenes: A multimodal dataset for autonomous driving", PROCEEDINGS OF THE IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, 2020, pages 11621 - 11631 |
| JOAO CARREIRAANDREW ZISSERMAN: "Quo vadis, action recognition? a new model and the kinetics dataset", PROCEEDINGS OF THE IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, 2017, pages 6299 - 6308 |
| KNITTEL ANTHONY ET AL: "Comparison of Pedestrian Prediction Models from Trajectory and Appearance Data for Autonomous Driving", 25 May 2023 (2023-05-25), XP093171008, Retrieved from the Internet <URL:https://arxiv.org/pdf/2305.15942v1> [retrieved on 20240606] * |
| KNITTEL ANTHONY ET AL: "DiPA: Diverse and Probabilistically Accurate Interactive Prediction", 12 October 2022 (2022-10-12), XP093171009, Retrieved from the Internet <URL:https://export.arxiv.org/pdf/2210.06106v1> [retrieved on 20240606] * |
| MORRIS ANTONELLOMIHAI DOBRESTEFANO V ALBRECHTJOHN REDFORDSUBRAMANIAN RAMAMOORTHY: "2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)", 2022, IEEE, article "Flash: Fast and light motion prediction for autonomous driving with bayesian inverse planning and learned motion profiles", pages: 9829 - 9836 |
| RISHABH BAJPAIDEEPAK JOSHI: "Movenet: A deep neural network for joint profile prediction across variable walking speeds and slopes", IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, vol. 70, 2021, pages 1 - 11, XP011853560, DOI: 10.1109/TIM.2021.3073720 |
| SADEGHIAN AMIR ET AL: "SoPhie: An Attentive GAN for Predicting Paths Compliant to Social and Physical Constraints", 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), IEEE, 15 June 2019 (2019-06-15), pages 1349 - 1358, XP033686463, DOI: 10.1109/CVPR.2019.00144 * |
| XUE HAO ET AL: "SS-LSTM: A Hierarchical LSTM Model for Pedestrian Trajectory Prediction", 2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), IEEE, 12 March 2018 (2018-03-12), pages 1186 - 1194, XP033337787, DOI: 10.1109/WACV.2018.00135 * |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN120929354A (en) * | 2025-10-14 | 2025-11-11 | 杭州秋果计划科技有限公司 | Code verification method and device |
Also Published As
| Publication number | Publication date |
|---|---|
| GB202306327D0 (en) | 2023-06-14 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Laugier et al. | Probabilistic analysis of dynamic scenes and collision risks assessment to improve driving safety | |
| EP3881226B1 (en) | Object classification using extra-regional context | |
| US11361557B2 (en) | Attention-based recurrent convolutional network for vehicle taillight recognition | |
| CN111670468B (en) | Moving body behavior prediction device and moving body behavior prediction method | |
| CN112203916B (en) | Method and apparatus for determining a vehicle comfort metric for an autonomous vehicle | |
| EP4181091B1 (en) | Pedestrian behavior prediction with 3d human keypoints | |
| JP7625679B2 (en) | Prediction device, prediction method, program, and vehicle control system | |
| Vatavu et al. | Stereovision-based multiple object tracking in traffic scenarios using free-form obstacle delimiters and particle filters | |
| Leon et al. | A review of tracking, prediction and decision making methods for autonomous driving | |
| CN118953402B (en) | A method, system and storage medium for predicting risk situation of autonomous driving vehicles based on multimodal information fusion and large model deduction | |
| CN111771207A (en) | Enhanced vehicle tracking | |
| EP4137845A1 (en) | Methods and systems for predicting properties of a plurality of objects in a vicinity of a vehicle | |
| US12091023B2 (en) | Information processing system, information processing method, computer program product, and vehicle control system | |
| US20240351592A1 (en) | Support tools for av testing | |
| US20250368208A1 (en) | Motion prediction for mobile agents | |
| Chavez-Garcia | Multiple sensor fusion for detection, classification and tracking of moving objects in driving environments | |
| Bhaggiaraj et al. | Deep learning based self driving cars using computer vision | |
| CN116868239A (en) | Static occupancy tracking | |
| EP4517696A1 (en) | Road geometry estimation for vehicles | |
| WO2024223902A1 (en) | Trajectory prediction for dynamic agents | |
| Nejad et al. | Vehicle trajectory prediction in top-view image sequences based on deep learning method | |
| CN117795566A (en) | Perception of three-dimensional objects in sensor data | |
| Qiao et al. | Human driver behavior prediction based on urbanflow | |
| CN119964109A (en) | Vehicle intention recognition method, device, vehicle and medium | |
| Wang et al. | Multimodal data trajectory prediction: A review |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 24722598 Country of ref document: EP Kind code of ref document: A1 |