How to realize the real-time online calibration design of autonomous driving system sensors

Publisher:和谐相伴Latest update time:2024-04-19 Source: elecfans Reading articles on mobile phones Scan QR code
Read articles on your mobile phone anytime, anywhere

2. Detect contour points by color segmentation

This module detects the contour points of the red inner octagon from the stop sign (see Figure b above). In the image preprocessing stage, the red part with threshold can be extracted using the color space conversion from RGB to HSV.

Then, on the HSV image, a conventional edge detection algorithm (such as the Canny/Devernay algorithm) is used to extract the edge, so as to obtain the contour points of the octagon with sub-pixel accuracy. Since the edge detection algorithm is very important in this article, the technical principles of the Canny algorithm and Devernay can be briefly explained here.

A. Canny algorithm

The Canny algorithm actually applies the first-order derivative of the Gaussian filter to image detection, so that the image gradient can be obtained, where the edge point is the local maximum value along the normal direction n.

The direction n orthogonal to the edge can be approximated by the image gradient:

wKgZomUBDRWARhJpAAAOzMsLlDU285.jpg

In the above formula, G and I are Gaussian filter and image respectively, and * represents the convolution operator.

B. Devernay sub-pixel correction

Although the basic edge detection process can be achieved through the Canny algorithm, the edge points obtained are at the whole pixel accuracy level. For our system, the processing process requires sub-pixel accuracy, which can be improved by using the Devernay correction algorithm to improve the Canny algorithm. Assume that B is the pixel point obtained by the Canny algorithm (corresponding to the local gradient maximum in the normal direction n). The gradient of two adjacent points of B along n can be approximated by linear interpolation. Then, the sub-pixel position of the edge point is refined to the maximum value of the one-dimensional quadratic interpolation of all three points (along n).

Next, we need to link the contour points of the processed results. To achieve this goal, we need to group the above independent edge detection points into several chains. In order to ensure the uniqueness of the contour detection, we need to use the dense point detection method, and the one containing the most points will be selected as the contour of the inner octagon.

3. Edge line fitting

Although the overall contour of the octagon is detected as above, it is still necessary to rely on a certain edge line fitting algorithm to fit these contour points into a real octagon.

The operation process is to fit eight lines to a set of two-dimensional points X ∈ RN×2 on the stop sign contour above, and the estimation of each line will generate an edge of the octagon. One edge is estimated at a time by finding the line with the most "supporting points". When the distance from the point to the line is within a threshold (in this paper, the threshold is fixed to 0.5 pixel distance to ensure that sub-pixel is accurately estimated), the contour point is considered to "support" the estimation of the corresponding line.

The selection of contour lines is done by running the RANSAC algorithm on the original set of contour points. Note that the number of iterations is limited by N2 by going through all pairs. Once an edge is fitted, all supporting points are removed from the original set so that the same algorithm can be repeated to find different edges.

Combined with the probability p ∈ [0, 1), the probability that the final least squares line is the most supported line, the number of iterations K can be calculated according to the following formula:

wKgaomUBDRWALHIKAAAUBRkMnak362.jpg

where w = 1(8−i)2 , i ∈ {0, ..., 7} is the chance of selecting a point from the support set, assuming that there are few outliers outside the contour in X. Determining p involves considering the trade-off between estimation accuracy and running time. If accuracy is more important than computational efficiency, then p should be set as close to 1 as possible (e.g., 0.999).

4. Vertical search line refinement

The size of the gradient of the RGB/HSV type image can actually indicate the location of the octagon outline, where the color changes most abruptly in the neighborhood. Considering that smart cars may be driving at night or in scenes with insufficient visibility, the original point may deviate from the outline due to insufficient lighting, which may lead to misleading color segmentation in the HSV space. Therefore, line thinning is necessary, where the moved line can be adjusted back to the true octagon edge through local search.

Starting from the estimated line, we search for the local maximum gradient along the perpendicular direction of the line and rotate the line to align with it. In the case of underexposure/overexposure, the gradient peak at the edge may disappear. Therefore, introducing a refinement boundary can solve this problem well. Assuming that the refinement should be subtle, a line will be updated only when more than half of the S refined points are within the boundary.

The refined border can be parameterized by a percentage B ∈ [0, 1], which means that the refined point is at most 100% of the white border width away from the original position in the vertical direction. To estimate the width of the white border of the stop sign, the length of the estimated line segment can be multiplied by the width-to-length ratio of the inner polygon edge recorded in the traffic sign rule.

5. Intersection point estimation

The eight estimated lines have at most 28 intersections, of which only eight are valid octagon corners. A simple, effective and commonly used approach is to select intersections close to the line segment endpoints (for example, within a distance of 10 pixels). Then, any estimate with a corner count not equal to 8 is discarded.

For example, sorting the corner points in clockwise order (relative to the geometric center of the points) helps to obtain 2D-3D correspondences. In addition, the first corner can always be set to the top left corner of the character "S" because stop signs are usually upright.

In order to further verify the validity of the corner points, the Hausdorff distance method is used to ensure that the final set of eight points forms an octagon under affine transformation.

6. Planar Object Calibration

Usually such standard models are calibrated using the pinhole camera model. That is, a 3D point (X, Y, Z) in world coordinates is projected into the image as a pixel (u, v). (u, v) can be related to (X, Y, Z) by the following equation:

wKgaomUBDRWAH3vPAAAX6CrjjC0812.jpg

Where s is an unknown scalar, A ∈ R 3×3 is the camera intrinsic matrix (assuming no skew constraints). r1, r2, r3 are the columns of the rotation matrix, and t is the translation vector. They are often called camera extrinsic parameters because they have the function of transforming world coordinate points into the camera coordinate system.

During calibration, if the 3D coordinate system is fixed to a flat object such as a chessboard or a stop sign, the Z component in Equation 3 is always zero, resulting in a simplified equation:

wKgaomUBDRaAcwOTAAASS_Td-VU070.jpg

We use M = (X, Y, 1)T and m = (u, v, 1)T to represent the position of a point on a plane object and its position in the image, respectively. Their relationship is defined by the homography matrix H as follows:

wKgaomUBDRWAFUGZAAAQoXublck720.jpg

Obviously, H is a 3 × 3 matrix up to a scaling factor. The homography matrix H can be computed using N ≥ 4 points.

In our system, after detecting the 8 corner points of the stop sign, H can be represented by H = [h1 h2 h3]. From formula 5, we can get:

wKgZomUBDRWAIcWCAAASEqgncAI186.jpg

Based on the knowledge that r1 and r2 are orthogonal, we obtain the two constraints required to calculate the intrinsic parameter matrix A as follows:

wKgZomUBDRWANzVEAAAaf9rypz0151.jpg

In this way, the two constraints required to calculate the intrinsic parameter matrix A are obtained.

In our system, the principal point is assumed to be fixed at the center of the image plane and only the focal lengths fx and fy are calculated and updated iteratively, so for the chessboard calibration we also need to enforce this assumption for comparison purposes.

The closed-form solution can be improved by maximum likelihood inference: the original calibration needs to be used as a nonlinear optimization problem for initial estimation, which can also be solved using the Levenberg-Marquardt algorithm. Currently, this part is not included in our system.

7. Time update of Kalman filter calibration

The optimization update target here is that some other noise may be generated over time, so it is necessary to integrate the camera intrinsic parameter matrix through continuous optimization. We generally assume that the inference process will introduce some standard Gaussian noise, so we can use Kalman filters with different noise covariance matrices for optimization and update.

The standard Kalman filter prediction formula is as follows:

wKgZomUBDRWAHbQKAAAM-2nxF74595.jpg

where xt, xt+1 are the camera intrinsic parameter vectors estimated at different times. A basic assumption of the study is that if the intrinsic parameters are assumed to be constant in the short-term estimation, then F can become the identity matrix.

At this point, the entire update step is performed by the following formula:

wKgaomUBDRWAV5dDAAAIlagaCoc962.jpg

The process noise and measurement noise δ are Gaussian random vectors with zero mean. Their covariance is represented by the (diagonal) matrices Q = E( T ) and R = E(δδT ), respectively. Q specifies how much we believe the intrinsic parameters to vary over time and should be lower for shorter estimation periods. R determines how much noise is in one estimate. Intuitively, this measurement noise should decrease from a high value as the process progresses, because the more images used, the more accurate the calibration will be.

03

in conclusion

Advanced driver assistance systems have very strict requirements on the robust performance of their sensors, because this is the key to whether the system can achieve accurate environmental perception. From the current development trend, the most promising sensing mode is the imaging mode, because low-cost and rich environmental information can be calculated from the image. Since camera systems are often used for distance perception, speed reasoning and target detection. For such applications, camera parameters are usually crucial for reconstructing part of the scene and affecting the measurement accuracy. The accuracy of the parameters often comes from the precise calibration and adaptive optimization process of the original camera. However, when the camera is installed on the vehicle, the inherent parameters of the camera due to mechanical vibration caused by road bumps or temperature fluctuations caused by weather will change. Therefore, it will also be a challenge to repeatedly calibrate the camera for long-term use.

[1] [2] [3]
Reference address:How to realize the real-time online calibration design of autonomous driving system sensors

Previous article:Silergy Automotive Grade Electronic Fuse SA21816
Next article:How to significantly improve the thermal efficiency of automobile engines?

Recommended ReadingLatest update time:2024-11-16 12:54

Typical applications of RGB sensors
Since the physical structure design is very important for many of these applications, the RGB ambient light sensor needs to be small and accurate. If the device side length is less than 2 mm, it can be easily packaged near the image sensor or mounted in an inconspicuous position in the display glass panel fram
[Embedded]
Typical applications of RGB sensors
51 MCU-Temperature Sensor Code Analysis Ⅱ
1. Conversion less than 0 degrees We know that when the actual temperature measured is below 0 degrees, the situation when temp is not converted is as shown in the figure below We already know that the upper 5 bits being 1 indicates a negative number, so why is it 1111 1111 1111 1000 when it is -0.5 degrees? The
[Microcontroller]
51 MCU-Temperature Sensor Code Analysis Ⅱ
Sensata Technologies-EMB Braking Force Sensor丨Confirmed to apply for 2023 Golden Edition Award
Application technology | EMB brake force sensor Application field: Intelligent chassis Product Description: ① Sensata Technologies' EMB braking force sensor is a further exploration and application promotion of new packaging and fields based on existing technologies.
[Automotive Electronics]
Sensata Technologies-EMB Braking Force Sensor丨Confirmed to apply for 2023 Golden Edition Award
SmartSens Launches First Linear CMOS Image Sensor to Enable Industrial Linear Camera Applications
SmartSens Launches First Linear CMOS Image Sensor to Enable Industrial Linear Camera Applications December 8, 2022, Shanghai, China— SmartSens (Shanghai) Electronic Technology Co., Ltd. (stock abbreviation: SmartSens) has launched the first LA (Linear) series 4K resolution high-speed industrial
[sensor]
SmartSens Launches First Linear CMOS Image Sensor to Enable Industrial Linear Camera Applications
Huawei to develop radar and build in-vehicle sensor ecosystem for self-driving cars
According to Reuters, Huawei's rotating chairman Xu Zhijun said that in the future Huawei will build millimeter-wave radar based on 5G technology and develop lidar to solve the cost and performance problems of its current smart cars. Xu Zhijun said that Huawei will build an "ecosystem" of in-vehicle sensors. "Hua
[Automotive Electronics]
Huawei to develop radar and build in-vehicle sensor ecosystem for self-driving cars
Hardcore | OmniVision Automotive Image Sensor Series Special Topic Launched
In 2022, the intelligentization of automobiles will accelerate. The in-vehicle cameras, known as the "eyes of intelligent driving", need to increase both "quantity" and "quality". Image sensors have become an important growth point in the automotive parts industry chain. According to the lates
[sensor]
Hardcore | OmniVision Automotive Image Sensor Series Special Topic Launched
Apple's under-screen optical fingerprint sensor makes fingerprint reading more reliable
Recently, Apple analysts said that Apple is considering adding an in-display Touch ID feature to the new iPhone. In cases where Face ID is not in optimal condition, the auxiliary Touch ID option will be very useful. Apple is using a redesigned optical sensor for in-screen fingerprint reading, which is more reliable th
[Mobile phone portable]
Apple's under-screen optical fingerprint sensor makes fingerprint reading more reliable
Image sensors are experiencing a "species explosion" and emerging applications are flourishing everywhere
For a long time, the evolution of CMOS image sensors used in camera modules of digital cameras and smartphones has been centered on enhancing imaging performance, increasing the number of pixels, and improving sensitivity. Now, CMOS image sensors have entered an evolutionary path centered on multifunctionality. In add
[Analog Electronics]
Image sensors are experiencing a
Latest Embedded Articles
Change More Related Popular Components

EEWorld
subscription
account

EEWorld
service
account

Automotive
development
circle

About Us Customer Service Contact Information Datasheet Sitemap LatestNews


Room 1530, 15th Floor, Building B, No.18 Zhongguancun Street, Haidian District, Beijing, Postal Code: 100190 China Telephone: 008610 8235 0740

Copyright © 2005-2024 EEWORLD.com.cn, Inc. All rights reserved 京ICP证060456号 京ICP备10001474号-1 电信业务审批[2006]字第258号函 京公网安备 11010802033920号