Autonomous vehicles accurately find their way with the aid of sensors. To ensure this in every possible situation, vehicles, controllers, and sensors are efficiently validated at early development stages: they are tested in virtual test drives with an immense variety of test cases. dSPACE precisely addresses this requirement by offering “Sensor Simulation” a unique and complete solution based on simulating a sensor’s physical phenomena and properties.
More realistic sensor simulation and higher performance.
Presenting a new test system for validating an autonomous, AI-based vehicle using realistic sensor simulation.
In this webinar recording dSPACE shows the different dSPACE solutions for sensor simulations: starting from ground truth sensor models to raw data stimulation via GPU-based physical sensor models.
Sensor Simulation contains models that create a virtual 3-D world to represent real objects in the surroundings of the vehicle. It also provides models of camera, radar, and lidar sensors to simulate the perception of this world. For this purpose, the model types simulate two essential aspects:
The sensor models provide sensor-realistic simulation data that can be used to validate functions for autonomous driving or a subset of them, such as data fusion or object detection. For best performance, the sensor models can be executed on platforms, such as the Sensor Simulation PC, which is equipped with powerful graphics processor units (GPU).
Sensor Simulation supports the reuse of models and test scenarios on different platforms. Therefore, tests created and used on the developer’s PC can be performed on a simulator or in the cloud and vice versa. This enables easy and fast scaling of parallelized tests. The entire validation process from software-in-the-loop (SIL) simulation to hardware-in-the-loop (HIL) simulation is supported. This enables versatile test methods with real and virtual ECUs or sensors.
Sensor Simulation provides flexibility via an application programming interface (API) to support customized solutions. The online postprocessing API is an efficient means of implementing sensor specific extensions directly in the product. It can be used for adapting the output format or integrating a sensor model of the sensor supplier, for example. The resulting custom code is executed on the GPU during the overall process. Therefore, you can use the same code in SIL and HIL domains.
Realistic simulation of camera and fish-eye camera sensor.Lidar Module
Realistic simulation of lidar signals.Radar Module
Realistic simulation of radar signals.
Sensor simulation supports entirely software-based testing by means of the SIL method.Hardware-in-the-Loop (HIL) Simulation
Sensor Simulation offers the option to test real sensors in combination with ADAS/AD control units in a closed loop.