An Auto-Calibrating System for Sensors in Autonomous Vehicles

Radars, Cameras, Lidars, Ultrasonic sensors help an autonomous vehicle make decisions and their accuracy is paramount to safe autonomous driving.

Technology creates new solutions but with increasing complexity. For instance, in the automotive industry, technology has come a long way from a simple speed control feature in the early 1900s to fully Autonomous Vehicles (AV) today. Vehicle manufacturers (OEMs) constantly aim to make driving as comfortable as possible for their customers.

Autonomous Driving (AD), which was once science fiction, is made possible using multiple sensors like radars, cameras, lidars, etc., which continuously feed information and help complex applications along with hardware and control systems to make critical driving decisions.

AVs can depend on more than 20 sensors mounted at various locations. Each sensor provides different kinds of data. For example, cameras provide images, radars and lidars provide point clouds, etc. The Autonomous Driving (AD) software uses information/data from each of these sensors individually and collectively from multiple sensors (Radar + Camera) to make precise driving decisions.

Autonomous Vehicle

Figure 1: The Autonomous Vehicle depends on multiple sensors to make a driving decision

Since the AV depends on such a wide range of sensors, calibration of these sensors becomes very important. Mis-calibrated sensors can have a catastrophic impact on the safety and functionality of the AV. Calibration of individual sensors is a long-drawn-out process, and different sensors i.e., LiDAR, radar or cameras etc. require different calibration methodologies. This calibration is typically carried out at the end of the production line.

The general understanding in the industry is that intrinsic calibrations, i.e., geometric distortion, focal length, etc., do not change over time. They are typically done only once. However, extrinsic calibration, i.e., position and orientation, which can change based on various reasons, from poor road conditions to sensors mounted on movable car parts such as the side view mirror, is a challenge. For example, a rear-facing camera mounted on the auto-folding side- view mirror, is used for lane change functionality or to detect the presence of a vehicle in the adjacent lane, a yaw angle miscalibration of 0.5 degrees can cause a lateral error of almost one meter depending on the distance to the vehicle. A lateral error of one meter could result in a vehicle present in the adjacent lane going undetected.

Challenges in sensor calibration

The typical challenges or concerns about sensor calibration include:

  • Calibration of individual sensors takes anywhere between 20 – 60 mins, and for 20 sensors, it could be 8-10 hours
  • Calibration of sensors depends on external rigs
  • Calibration frequency is influenced by factors like wear and tear of the vehicle, accidents, servicing cycle, etc.
  • Liability for safety violations arising from mis-calibrated sensors
  • Reliability and availability of features that depend on sensors
  • Impact on finances of poorly functioning features

typical rig to calibrate sensors

Figure 2: A typical rig to calibrate sensors

Sensor calibration solutions

The challenges of sensor calibration for AVs can be resolved as follows.

  • A robust multi-modality sensor calibration algorithm can help vehicle manufacturers (OEMs) to auto-calibrate the sensor’s extrinsic parameters.
  • Also, an AV that can auto-calibrate its sensors will be able to provide information/data more reliably over an extended period with minimum visits to the service center.
Requirements of a multi-modality sensor calibration system

A multi-modality sensor calibration system requires the following features:

  • Self-diagnosis and determination of error required for correction.
  • Auto calibration without the need for an external rig.
  • Determination of an ideal scenario to auto-calibrate.
  • Ability to function in both highways and urban scenarios.
  • Duration to calibrate must not exceed a few seconds for each sensor
  • Absence of external human involvement
  • Auto-correction of safety-related errors to meet functional safety requirements

Additionally, it is important to calibrate individual sensors and multiple sensors with respect to each other. Multi-modality sensor calibration refers to the calibration of different types of sensors such as Camera with respect to LiDAR or RADAR about LiDAR.

Let us look at an example to understand multi-modality sensor calibration.

In an AV, features such as Lane Departure Warning (LDW), Forward Collision Warning (FCW), Traffic Sign Recognition (TSR) rely on the front-facing camera. A well-calibrated camera can position the vehicle in the center of the lane and maintain an appropriate distance from the vehicle in the front. But a minor error in the pitch angle of the camera mounting can result in the vehicle in front appearing further away than it is. It could result in delayed system response to avoid a collision with a vehicle in the front.

Autonomous Vehicle uses the Front Camera

Figure 3: An Autonomous Vehicle uses the Front Camera to assess the road for various features like FCW, TSR and LDW

Alternatively, a LiDAR or a RADAR sensor can reliably estimate the distance to vehicles even if there is a variation in its pitch angle. Thus, either one of these sensors can be used to determine and compensate for a camera’s pitch angle mounting errors.

For a camera or LiDAR to effectively predict the yaw angle mounting error, an Inertial Measurement Unit (IMU) sensor in the vehicle can compare vehicle movement with the tracking and movement of other stationary objects, such as road barriers, traffic signs or lane markings.

Multiple such solutions can be created where a single point of truth must be used to calibrate all sensors. The single point of truth can be another sensor, such as the LiDAR or camera. But it is also important to note that any calibration error in the main sensor will magnify the errors for other sensors. An auto calibrating system is important.

With the increase in competition between OEMs to develop quicker and more reliable and advanced Autonomous Driving features, the effective use of perfectly calibrated sensors is key to their success.

KPIT has worked on developing multiple models for sensor misalignment. These models could be extended to every sensor and help in autocalibration. We have the knowledge, skills and experience to develop state-of-the-art technologies and features. We require access to millions of kilometers of vehicle data, test vehicle setup and customer-specific requirements from our customers.

At KPIT, over the past decade, we have developed a strong Autonomous Driving Practice, and we are strategic software development and integration partners to Global OEMs and Tier 1s for Autonomous Driving.

Our teams help clients across all domains with AD development and deploy key technologies like AI, ML, DL, Sensor Fusion, Computer vision, Virtual Simulation and more.

Related Articles

How to achieve the best Sensor Fusion for ADAS/AV vehicles?

How to achieve the best Sensor Fusion for ADAS/AV vehicles?

Software Industrialization is the next step in a Commercial Vehicle’s Autonomous Driving Journey

Software Industrialization is the next step in a Commercial Vehicle’s Autonomous Driving Journey

System Engineering Plays A Key Role In Autonomous Vehicle Journey

System Engineering Plays A Key Role In Autonomous Vehicle Journey

Your feedback form has been submitted successfully!