Anticipating environmental impacts on robotic systems is crucial for agile control, mirroring how a human driver predicts road conditions to inform their actions. Paul Brunzema of Aachen University, alongside Thomas Lew, Ray Zhang, Takeru Shirasawa, John Subosits, and Marcus Greiff, all from Toyota Research Institute, address this challenge with a novel approach to predictive modelling. Their research introduces a vision-conditioned variational Bayesian model that anticipates changes in system dynamics by leveraging visual context. This proactive adaptation distinguishes itself from reactive methods, offering a significant advantage in rapidly changing environments and proving particularly effective in high-performance applications like autonomous vehicle racing. Validated on a Lexus LC500 navigating wet conditions, the team’s model successfully completed twelve laps, while baseline systems consistently failed, highlighting the importance of anticipating dynamic shifts.
Traditional modelling approaches frequently struggle to accurately represent abrupt variations in system behaviour, whereas adaptive methods are inherently reactive and may adapt too late to guarantee safety. This research proposes a vision-conditioned variational Bayesian last-layer dynamics model, designed to leverage visual context in order to anticipate environmental changes. The model initially learns nominal vehicle dynamics and is subsequently fine-tuned using feature-wise affine transformations of latent features, facilitating context-aware dynamics prediction.
The resulting model was integrated into an optimal controller specifically for vehicle racing applications, allowing for proactive adjustments based on anticipated conditions. Validation of the method was performed using a Lexus LC500 racing vehicle navigating through water puddles, a scenario demanding precise dynamic prediction. This approach demonstrates the potential to improve autonomous vehicle performance and safety in dynamic and unpredictable environments by anticipating rather than simply reacting to changes. The use of a variational Bayesian framework allows for principled uncertainty quantification, further enhancing robustness.
Autonomous Racing via Learning and Adaptation
This research details a system for autonomous car racing, focusing on learning and adaptation. The core problem addressed is building robust and adaptable autonomous racing systems, which often struggle with real-world variability, the sim-to-real gap, data efficiency, and extreme maneuvers. The proposed system is a learning-based Model Predictive Control (MPC) framework, incorporating a neural network dynamics model that predicts the car’s future state given its current state and control inputs. Bayesian Optimization (BO) efficiently tunes the parameters of the dynamics model and MPC controller, enabling adaptation to new conditions with limited data.
Visual conditioning, achieved using FiLM (Feature-wise Linear Modulation), improves the accuracy of the dynamics model by incorporating camera images. Uncertainty quantification, using Bayesian Last Layer networks, further enhances robustness. An active learning strategy selects the most informative data points for training, improving data efficiency.
Vision-Based Friction Prediction Improves Autonomous Racing Performance
Scientists achieved a breakthrough in autonomous vehicle control by developing a vision-conditioned variational Bayesian last-layer dynamics model (VcVBLL). This innovative model anticipates environmental changes, specifically variations in road friction, allowing for proactive adaptation of vehicle dynamics. The research team validated their approach using a Lexus LC500 racing car, demonstrating a significant advancement in autonomous racing capabilities. Experiments revealed that the system successfully completed all 12 attempted laps under varying conditions, including navigating water puddles, a scenario that consistently caused failures in baseline systems.
The core of this achievement lies in the model’s ability to learn nominal vehicle dynamics and then refine these predictions using visual context through feature-wise affine transformations of latent features. Measurements confirm that this conditioning on visual information is crucial for maintaining control in dynamic environments. The team integrated the VcVBLL model into a model predictive control (MPC) framework, enabling high-speed autonomous racing where conventional, proprioception-only baselines consistently spun out.
Vision Anticipates Track Conditions for Racing
This work demonstrates a vision-conditioned variational Bayesian last-layer (VcVBLL) model capable of proactive control for autonomous vehicle racing on surfaces with varying conditions. By integrating visual context into a learned dynamics model via FiLM, the system anticipates changes in road surface properties, enabling robust performance where traditional methods fail. The researchers achieved this through a two-stage training process, initially establishing a nominal model on dry surfaces before fine-tuning it with limited data from wet conditions. Hardware experiments conducted with a Lexus LC500 racing through water puddles confirmed the efficacy of the approach.
The vision-conditioned model successfully completed all twelve attempted laps under varying conditions, a significant achievement given that baseline methods consistently lost control. The authors acknowledge the limited availability of data representing water interactions as a constraint, and note that the current model focuses specifically on the transition between dry and wet surfaces. Future research will extend this work to more complex scenarios, including autonomous drifting and off-road driving, with the aim of incorporating a richer visual context for improved predictive control. Anticipating environmental impacts on robotic systems is crucial for agile control, mirroring how a human driver predicts road conditions to inform their actions.
👉 More information
🗞 Vision-Conditioned Variational Bayesian Last Layer Dynamics Models
🧠 ArXiv: https://arxiv.org/abs/2601.09178
