RedMotion: Intelligent Motion Prediction for Autonomous Vehicles
Principal Investigator:
Dr. Andreas Lintermann
Affiliation:
Forschungszentrum Jülich GmbH, Jülich Supercomputing Centre, Jülich, Germany
Local Project ID:
genai-ad
Date published:
RedMotion: Intelligent Motion Prediction for Autonomous Vehicles
Autonomous vehicles must predict the motion of other road users within fractions of a second in complex urban environments with hundreds of lanes, traffic lights, and vehicles. RedMotion addresses this challenge through a novel transformer architecture that learns augmentation-invariant and redundancy-reduced descriptors of road environments. By compressing up to 1,200 local environmental features into exactly 100 compact tokens through self-supervised learning, RedMotion achieves efficient and accurate motion prediction. Training on millions of traffic scenes from the Waymo and Argoverse datasets required extensive parallel computations on the GPU nodes of JUWELS Booster.
Autonomous vehicles must predict within fractions of a second where other road users will move. This prediction is particularly challenging in complex urban environments with many vehicles, pedestrians, and changing traffic rules. Traditional approaches struggle with information overload: How do you process data about hundreds of lanes, traffic lights, and other vehicles without losing important details?
FZI has developed RedMotion, a novel method that learns to extract only the truly important information from the road environment. The system uses two innovative "redundancy reduction" mechanisms:
1. Redundancy Reduction: RedMotion distills up to 1,200 local information elements (e.g., overlapping lanes, closely clustered vehicles) into exactly 100 compact "RED tokens" -- similar to how a human reduces a complex traffic situation to its essential aspects.
2. Self-Supervised Learning: The model automatically learns which features are important by viewing the same traffic scene from slightly different perspectives and developing robust, invariant representations through a Barlow Twins loss function.
The Architecture
RedMotion consists of three principal components (Figure 1):

Figure 1: RedMotion architecture showing the trajectory encoder, road environment encoder with local and global (RED) tokens, and the two redundancy reduction stages.
Developing RedMotion required massive computational power:
The ability to run multiple experiments in parallel on JUWELS was crucial for exploring different hyperparameters, training strategies, and architectural variants.
RedMotion achieves competitive prediction accuracy while being significantly more efficient than existing methods. By processing the road environment in a structured, compact form with fixed-size representations, it enables:
The research results were published in the prestigious "Transactions on Machine Learning Research" (2024) and provide important foundations for safer autonomous driving systems. The learned representations are reusable for various downstream tasks beyond motion prediction, making RedMotion a versatile building block for autonomous vehicle perception systems.
Royden Wagner, Omer Sahin Tas, Marvin Klemp, Carlos Fernandez, and Christoph Stiller, RedMotion: Motion Prediction via Redundancy Reduction, Transactions on Machine Learning Research, 2024.