Recently, Andrej Karpathy, senior director of Tesla AI, revealed at a machine learning conference that he is also doing some research on pseudo-LiDAR technology. He believes that the gap between virtual LiDAR and real LiDAR is getting smaller and smaller.
pseudo-LiDAR
The earliest paper proposing the concept of "pseudo-LiDAR" data can be traced back to a technical paper from Cornell University in 2018. The authors Yan Wang and Wei-Lun Chao are both Chinese. The paper proposed a new method to shorten the performance gap between pure vision technology architecture and LiDAR.
This paper changes the 3D information presentation form of the stereo camera target detection system, converts the image-based stereo vision data into a 3D point cloud similar to that generated by LiDAR, and switches it to the final view format through data conversion. Although the visual effect is not comparable to LiDAR in terms of experimental results, this new low-cost method provides a way of thinking for the visual solution.
Two years ago, Cornell University and others published papers on visual depth estimation, target recognition, 3D Packing, etc. based on this method. Some researchers found that after adopting their new method, the camera's performance in target detection is close to that of LiDAR, and its cost is only a fraction of the latter. Analyzing the images captured by the camera with a bird's-eye view instead of a front view can increase the accuracy of target detection by 2 times, making stereo cameras a viable alternative to LiDAR, and its cost is much lower than the latter.
Tesla is also doing similar research
Musk's attitude towards LiDAR is self-evident. He believes that relying on LiDAR is like relying on crutches for walking. Recently, Tesla AI Senior Director Andrej Karpathy revealed at the Machine Learning Conference held this year that Tesla is also doing some research on "pseudo LiDAR".
Along the way, he shared a specific example of how Tesla can achieve the accuracy of traditional LiDAR with only a few cameras. The secret in Tesla's evolving solution is not the camera itself, but the advanced processing and neural network built by the camera to fully understand the range and quality of the input. Through the stitching of cameras in different directions, visual depth estimation is projected into a bird's-eye view and used as a local navigation map. At the same time, the depth of each pixel in the picture is estimated, just like the LiDAR point cloud, to form a 3D object detection.
Tesla's camera-based approach is much cheaper and easier to implement in terms of hardware, but requires extremely complex algorithms to convert raw camera input and vehicle telematics into useful information.
Fundamentally, computers can identify lane markings, signs, and other vehicles from a series of continuous static images (also known as video). Tesla takes computer vision to an unprecedented level, analyzing not only the image, but the individual pixels within the image.
“We take a pseudo-lidar approach where we basically predict the depth of each pixel, and we project that out,” Karpathy said. Doing this over time can replicate many of the capabilities of traditional lidar systems, but it requires a lot of real-time processing power to make the image deconstruction capabilities work.
Vehicles are driven in real time, so having a system that can make determinations or predictions based on images does no good if the results aren’t instantaneous. Thankfully, Tesla built its own hardware for the third major version of its Autopilot computer, and it’s designed specifically to run Tesla’s code.
The gap with LiDAR is narrowing
Tesla’s so-called pseudo-lidar solution is getting better. Karpathy showed off a series of lidar-like 3D maps of the world that looked an awful lot like the results from a cutting-edge lidar solution. Of course, the visualizations are more for humans than computers, so they don’t really convey how impactful Tesla’s advances in computer vision are. “Using only vision techniques and pseudo-lidar approaches, the gap will close very quickly,” Karpathy said.
However, some experts have raised doubts that the vision-based method is highly dependent on image clarity, camera pixels and fiber strength, which may be a difficult problem to solve at present. Therefore, it depends on whether the price of lidar drops faster or the visual algorithm improves faster.
Previous article:Sunny Automotive Optics joins Leddar ecosystem to create best-in-class LiDAR solutions
Next article:DARPA develops 3D infrared sensor for autonomous driving navigation
- Popular Resources
- Popular amplifiers
- GigaDevice's full range of automotive-grade SPI NOR Flash GD25/55 wins ISO 26262 ASIL D functional safety certification
- Breaking through the intelligent competition, Changan Automobile opens the "God's perspective"
- The world's first fully digital chassis, looking forward to the debut of the U7 PHEV and EV versions
- Design of automotive LIN communication simulator based on Renesas MCU
- When will solid-state batteries become popular?
- Adding solid-state batteries, CATL wants to continue to be the "King of Ning"
- The agency predicts that my country's public electric vehicle charging piles will reach 3.6 million this year, accounting for nearly 70% of the world
- U.S. senators urge NHTSA to issue new vehicle safety rules
- Giants step up investment, accelerating the application of solid-state batteries
- Intel promotes AI with multi-dimensional efforts in technology, application, and ecology
- ChinaJoy Qualcomm Snapdragon Theme Pavilion takes you to experience the new changes in digital entertainment in the 5G era
- Infineon's latest generation IGBT technology platform enables precise control of speed and position
- Two test methods for LED lighting life
- Don't Let Lightning Induced Surges Scare You
- Application of brushless motor controller ML4425/4426
- Easy identification of LED power supply quality
- World's first integrated photovoltaic solar system completed in Israel
- Sliding window mean filter for avr microcontroller AD conversion
- What does call mean in the detailed explanation of ABB robot programming instructions?
- USB Type-C® and USB Power Delivery: Designed for extended power range and battery-powered systems
- RAQ #223: How to measure and determine soft-start timing without a soft-start equation?
- RAQ #223: How to measure and determine soft-start timing without a soft-start equation?
- GigaDevice's full range of automotive-grade SPI NOR Flash GD25/55 wins ISO 26262 ASIL D functional safety certification
- GigaDevice's full range of automotive-grade SPI NOR Flash GD25/55 wins ISO 26262 ASIL D functional safety certification
- New IsoVu™ Isolated Current Probes: Bringing a New Dimension to Current Measurements
- New IsoVu™ Isolated Current Probes: Bringing a New Dimension to Current Measurements
- Infineon Technologies Launches ModusToolbox™ Motor Kit to Simplify Motor Control Development
- Infineon Technologies Launches ModusToolbox™ Motor Kit to Simplify Motor Control Development
- STMicroelectronics IO-Link Actuator Board Brings Turnkey Reference Design to Industrial Monitoring and Equipment Manufacturers
- On-chip memory resources of RSL10
- Comprehensive analysis of IPTV
- UART Protocol Timing Summary
- NXP Smart Car Northeast Competition First Prize, we will open source after the competition
- Non-inverting proportional amplifier feedback resistor setting
- I encountered a problem about Keil today
- Cumulative error analysis and clock frequency optimization for UART communication in MSP430FR2311
- [Shanghai Hangxin ACM32F070 development board + touch function evaluation board] Voltage meter implementation and ADC performance rough test
- Xianji official engineer dry goods: Based on HPM6750 CAN2.0 and CAN-FD operation guide
- Saber simulation problem