Highway Lane Change
This example shows how to perceive surround-view information and use it to design an automated lane change maneuver system for highway driving scenarios.
An automated lane change maneuver (LCM) system enables the ego vehicle to automatically move from one lane to another lane. An LCM system models the longitudinal and lateral control dynamics for an automated lane change. LCM systems scan the environment for most important objects (MIOs) using onboard sensors, identify an optimal trajectory that avoids these objects, and steer the ego vehicle along the identified trajectory.
This example shows how to create a test bench model to test the sensor fusion, planner, and controller components of an LCM system. This example uses five vision sensors and one radar sensor to detect other vehicles from the surrounding view of the ego vehicle. It uses a joint probabilistic data association (JPDA) based tracker to track the fused detections from these multiple sensors. The lane change planner then generates a feasible trajectory for the tracks to negotiate a lane change that is executed by the lane change controller. In this example, you:
Partition the algorithm and test bench — The model is partitioned into lane change algorithm models and a test bench model. The algorithm models implement the individual components of the LCM system. The test bench includes the integration of the algorithm models and testing framework.
Explore the test bench model — The test bench model contains the testing framework, which includes the sensors and environment, ego vehicle dynamics model, and metrics assessment using ground truth.
Explore the algorithm models — Algorithm models are reference models that implement the sensor fusion, planner, and controller components to build the lane change application.
Simulate and visualize system behavior — Simulate the test bench model to test the integration of sensor fusion and tracking with planning and controls to perform lane change maneuvers on a curved road with multiple vehicles.
Explore other scenarios — These scenarios test the system under additional conditions.
You can apply the modeling patterns used in this example to test your own LCM system.
Partition Algorithm and Test Bench
The model is partitioned into separate algorithm models and a test bench model.
Algorithm models — Algorithm models are reference models that implement the functionality of individual components.
Test bench model — The Highway Lane Change Test Bench specifies the stimulus and environment for testing the algorithm models.
Explore Test Bench Model
In this example, you use a system-level simulation test bench model to explore the behavior of a probabilistic sensor-based LCM system.
To explore the test bench model, load the highway lane change project.
Open the system-level simulation test bench model.
Opening this model runs the
helperSLHighwayLaneChangeSetup function, which initializes the road scenario using the
drivingScenario object in the base workspace. It also configures the sensor configuration parameters, tracker design parameters, planner configuration parameters, controller design parameters, vehicle model parameters, and the Simulink® bus signals required for defining the inputs and outputs for the
The test bench model contains these subsystems:
Scenario and Environment— Subsystem that specifies the scene, vehicles, sensors, and map data used for simulation. This example uses five vision sensors, one radar sensor, and an INS sensor.
Surround Vehicle Sensor Fusion— Subsystem that fuses the detections from multiple sensors to produce tracks.
Planner Configuration Parameters— Subsystem that specifies the configuration parameters required for the planner algorithm.
Highway Lane Change Planner— Subsystem that implements the lane change planner algorithm for highway driving.
Lane Change Controller— Subsystem that specifies the path-following controller that generates control commands to steer the ego vehicle along the generated trajectory.
Vehicle Dynamics— Subsystem that specifies the dynamic model for the ego vehicle.
Metrics Assessment— Subsystem that specifies metrics to assess system-level behavior.
Highway Lane Change Planner,
Lane Change Controller, and
Metrics Assessment subsystems are the same as those in the Highway Lane Change Planner and Controller example. However, whereas the lane change planner in the Highway Lane Change Planner and Controller example, uses ground truth information from the scenario to detect MIOs, the lane change planner in this example uses tracks from surround vehicle sensor fusion to detect the MIOs. The
Vehicle Dynamics subsystem models the ego vehicle using a Bicycle Model block, and updates its state using commands received from the
Lane Change Controller subsystem.
Scenario and Environment subsystem uses the Scenario Reader block to provide road network and vehicle ground truth positions. This block also outputs map data required for the highway lane change planner algorithm. This subsystem outputs the detections from the vision sensors, clusters from the radar sensor, and ego-estimated position from the INS sensor required for the sensor fusion and tracking algorithm. Open the
Scenario and Environment subsystem.
open_system("HighwayLaneChangeTestBench/Scenario and Environment")
The Scenario Reader block configures the driving scenario and outputs actor poses, which control the positions of the target vehicles.
The Vehicle To World block converts actor poses from the coordinates of the ego vehicle to the world coordinates.
The Vision Detection Generator block simulates object detections using a camera sensor model.
The Driving Radar Data Generator block simulates object detections based on a statistical model. It also outputs clustered object detections for further processing.
The INS block models the measurements from the inertial navigation system and global navigation satellite system and outputs the fused measurements. It outputs the noise-corrupted position, velocity, and orientation of the ego vehicle.
The subsystem configures five vision sensors and a radar sensor to capture the surround view of the vehicle. These sensors are mounted on different locations on the ego vehicle to capture a 360-degree view.
The Bird's-Eye Scope displays sensor coverage using a cuboid representation. The radar coverage area and detections are in red. The vision coverage area and detections are in blue.
Vehicle Dynamics subsystem uses a Bicycle Model block to model the ego vehicle. For more details on the
Vehicle Dynamics subsystem, see the Highway Lane Following example. Open the
Vehicle Dynamics subsystem.
The Bicycle Model block implements a rigid two-axle single-track vehicle body model to calculate longitudinal, lateral, and yaw motion. The block accounts for body mass, aerodynamic drag, and weight distribution between the axles due to acceleration and steering. For more details, see Bicycle Model (Automated Driving Toolbox).
Metric Assessment subsystem enables system-level metric evaluations using the ground truth information from the scenario. Open the
Metrics Assessment subsystem.
Collision Detectionsubsystem detects the collision of the ego vehicle with other vehicles and halts the simulation if it detects a collision. The subsystem also computes the TimeGap parameter using the distance to the lead vehicle (headway) and the longitudinal velocity of the ego vehicle. This parameter is evaluated against prescribed limits.
Jerk Metricssubsystem computes the LongitudinalJerk and LateralJerk parameters using longitudinal velocity and lateral velocity, respectively. These parameters are evaluated against prescribed limits.
For more details on how to validate the metrics automatically using Simulink Test, see the Automate Testing for Highway Lane Change example.
Explore Algorithm Models
The lane change system is developed by integrating the surround vehicle sensor fusion, lane-change planner, and lane-following controller components.
The surround vehicle sensor fusion algorithm model fuses vehicle detections from cameras and radar sensors and tracks the detected vehicles using the central-level tracking method. Open the
Surround Vehicle Sensor Fusion algorithm model.
The surround vehicle sensor fusion model takes the vehicle detections from vision sensors and clusters from the radar sensor as inputs.
Vision Detection Concatenationblock concatenates the vision detections.
Delete Velocity From Visionblock is a MATLAB Function block that deletes velocity information from vision detections.
Vision and Radar Detection Concatenationblock concatenates the vision and radar detections.
Add Localization Informationblock is a MATLAB Function block that adds localization information for the ego vehicle to the concatenated detections using an estimated ego vehicle pose from the INS sensor. This enables the tracker to track in the global frame, and minimizes the effect on the tracks of lane change maneuvers by the ego vehicle.
helperJPDATrackerblock performs fusion and manages the tracks of stationary and moving objects. The tracker fuses the information contained in the concatenated detections and tracks the objects around the ego vehicle. It estimates tracks in the Frenet coordinate system. It uses mapInfo from the scenario to estimate the tracks in Frenet coordinate system. The tracker then outputs a list of confirmed tracks. These tracks are updated at a prediction time driven by a digital clock in the
Scenario and Environmentsubsystem.
For more details on the algorithm, see the Object Tracking and Motion Planning Using Frenet Reference Path example.
The highway lane change planner is a fundamental component of a highway lane change system. This component is expected to handle different driving behaviors to safely navigate the ego vehicle from one point to another point. The
Highway Lane Change Planner algorithm model contains a terminal state sampler, motion planner, and motion prediction module. The terminal state sampler samples terminal states based on the planner parameters and the current state of both the ego vehicle and other vehicles in the scenario. The motion prediction module predicts the future motion of MIOs. The motion planner samples trajectories and outputs an optimal trajectory. Open the
Highway Lane Change Planner algorithm model.
The algorithm model implements the main algorithm for the highway lane change system. The reference model reads map data, actor poses (in world coordinates), and planner parameters from the
Scenario and Environment subsystem to perform trajectory planning. The model uses the Frenet coordinate system to find the MIOs surrounding the ego vehicle. Then, the model samples terminal states for different behaviors, predicts the motion of target actors, and generates multiple trajectories. Finally, the model evaluates the costs of generated trajectories and checks for the possibility of collision and kinematic feasibility to estimate the optimal trajectory. For more details, see the Generate Code for Highway Lane Change Planner example.
Lane Change Controller reference model simulates a path-following control mechanism that keeps the ego vehicle traveling along the generated trajectory while tracking a set velocity. Open the
Lane Change Controller reference model.
The controller adjusts both the longitudinal acceleration and front steering angle of the ego vehicle to ensure that the ego vehicle travels along the generated trajectory. The controller computes optimal control actions while satisfying velocity, acceleration, and steering angle constraints using adaptive model predictive control (MPC). For more details on the integration of the highway lane change planner and controller, see the Highway Lane Change Planner and Controller example.
Simulate and Visualize System Behavior
Set up and run the
HighwayLaneChangeTestBench simulation model to visualize the behavior of the system during a lane change. The
Visualization block in the model creates a MATLAB figure that shows the chase view and top view of the scenario and plots the ego vehicle, tracks, sampled trajectories, capsule list, and other vehicles in the scenario.
Disable the MPC update messages.
HighwayLaneChangeTestBench model to use the
During the simulation, the model logs signals to the base workspace as
logsout. You can analyze the simulation results and debug any failures in the system behavior using the
HelperAnalyzeLCSimulationResults object. The
visualizeSimulationData function of the method creates a MATLAB figure and plots a chase view of the scenario along with detections and tracks. For more details on this figure, see the Generate Code for Highway Lane Change Planner example. Run the function and explore the plot.
visualizatonObj = HelperAnalyzeLCSimulationResults(logsout); visualizatonObj.visualizeSimulationData()
Explore Other Scenarios
In this example, you have explored the system behavior for the
scenario_LC_15_StopnGo_Curved scenario, but you can use the same test bench model to explore other scenarios. This is a list of scenarios that are compatible with the
Each of these scenarios have been created using the Driving Scenario Designer and exported to a scenario file. Examine the comments in each file for more details on the road and vehicles in each scenario. You can configure the
HighwayLaneChangeTestBench model and workspace to simulate these scenarios using the
helperSLHighwayLaneChangeSetup function. For example, you can configure the simulation for a curved road scenario using this command.
In this example, you designed and simulated a highway lane change maneuver system using information perceived from surround view. This example showed how to integrate sensor fusion, planner, and controller components to simulate a highway lane change system in a closed-loop environment. The example also demonstrated various evaluation metrics to validate the performance of the designed system. If you have a Simulink Coder™ license and Embedded Coder™ license, you can generate ready-to-deploy code of the algorithm models for an embedded real-time target (ERT).
Enable the MPC update messages again.
trajectoryOptimalFrenet (Navigation Toolbox)