DJI Automotive Autonomous Driving Business Analysis Report: Leading the NOA Market by Virtue of Unique Technology Route - ResearchAndMarkets.com

The "Analysis on DJI Automotive's Autonomous Driving Business, 2024" report has been added to ResearchAndMarkets.com's offering.

Research on DJI Automotive: lead the NOA market by virtue of unique technology route

In 2016, DJI Automotive's internal technicians installed a set of stereo sensors + vision fusion positioning system into a car and made it run successfully. DJI Automotive's technologies such as perception, positioning, decision and planning accumulated in the drone field have been successfully transferred to intelligent driving field.

Almost all founding and management team members of DJI Automotive came from DJI's drone projects. DJI Automotive had only about 10 members at the beginning, mainly composed of representatives temporarily transferred from the Flight Control Department and Vision Department of DJI at that time.

DJI claims that it is a company specializing in the research of intelligent robots, and drones and autonomous vehicles are different forms of intelligent robots. Relying on its unique technology route, DJI holds lead in the mass production and application of NOA. By DJI Automotive's estimates, around 2 million passenger cars taking to road will be equipped with DJI Automotive's intelligent driving systems in 2025.

Continuously optimize stereo vision sensors

One of the core technologies of DJI Automotive is stereo vision. Even when other sensors like GPS fail, based on visual perception of the stereo camera, drones can still enable hovering, obstacle avoidance, and speed measurement among others.

After applying stereo vision technology to autonomous vehicles, DJI Automotive continues to optimize stereo vision sensors according to requirements of different autonomous driving levels.

In 2023, to meet the needs of NOA, DJI Automotive launched the second-generation inertial navigation stereo vision system, which eliminates the overall lens hood by adding a customized optical polarizer and cancels the rigid connecting rod using a better self-calibration algorithm. This makes it easier to install the sensor, and the distance between two cameras can be flexibly configured from 180 mm to 400 mm. Elimination of the rigid connecting rod is a huge progress in stereo vision sensors, allowing stereo cameras to be applied in much more scenarios.

Based on the needs of L3 autonomous driving, in 2024 DJI Automotive introduced a LiDAR-vision system, which combines LiDAR, stereo sensor, long-focus mono camera and inertial navigation. Compared with the currently common 'LiDAR + front camera' solution on the market, the system can reduce the costs by 30% to 40%, while enabling 100% performance and replacing all the functions. Thanks to the integrated design, the 'LiDAR-vision' solution can also be built into the cabin as a whole, reducing the overall installation costs.

The 'LiDAR-vision' solution can further enhance safety in vehicle longitudinal control. Thanks to LiDAR's precise ranging capabilities and robustness to illumination, the 'LiDAR-vision' solution can further improve safety and comfort of intelligent driving system in such scenarios as cut-in at close range, complex traffic flow in urban areas, response to vulnerable road users (VRU), arbitrary obstacle avoidance, detour, and VRU at night.

Use drone technologies for data acquisition and simulation

Among the three autonomous driving data acquisition methods, acquisition by vehicles is the most common, but the proportion of effective data is low, and it is easy to interfere with real behaviors of surrounding vehicles, and it is unable to record data in blind spots of sensors. Another method is acquisition in field, with low flexibility and insufficient reliability, a result of angle skew and low image accuracy.

According to the in-depth research by fka, the automotive technology research institute of RWTH Aachen University, and DJI Automotive's own practices in the past two years, aerial survey data acquisition by drones has obvious advantages. Drones can collect richer and more complete scenario data, and can directly collect aerial objective shots of all vehicles in blind spots of the target vehicle without obstruction, reflecting more realistic and interference-free human driving behaviors, and more efficiently collecting data in specific road sections and special driving scenarios, for example, on/off-ramps and frequent cut-ins.

Why does the implementation of vision-only autonomous driving suddenly accelerate?

Why has the pace of implementing vision-only technology solutions suddenly quicken since 2024? The answer is foundation models. The research shows that a truly autonomous driving system needs at least about 17 billion kilometers of road verification before being production-ready. The reason is that even if the existing technology can handle more than 95% of common driving scenarios, problems may still occur in the remaining 5% corner cases.

Generally, learning a new corner case requires collecting more than 10,000 samples, and the entire cycle is more than 2 weeks. Even if a team has 100 autonomous vehicles conducting road tests 24 hours a day, the time required to accumulate data is measured in 'hundred years' - which is obviously unrealistic.

Foundation models are used to quickly restore real scenarios and generate corner cases in various complex scenarios for model training. Foundation models (such as Pangu model) can shorten the closed-loop cycle of autonomous driving corner cases from more than two weeks to two days.

Currently, DJI Automotive, Baidu, PhiGent Robotics, GAC, Tesla and Megvii among others have launched their vision-only autonomous driving solutions. This weekly report summarizes and analyzes vision-only autonomous driving routes.

Key Topics Covered:

1 Overview

  • Profile
  • Team
  • Development History
  • All-scenario Intelligent Driving User Journey Map
  • Products Layout
  • DJI's Judgment on Evolution Trends of High-level Intelligent Driving

2 Core Technologies

  • Inertial Navigation Stereo Vision: Developed to the 2nd Generation
  • Stereo Vision Perception Technology
  • Parameter Comparison between Vision Sensors
  • LiDAR-vision System and Specifications
  • Intelligent Driving Domain Controller
  • Domain Controller Middleware
  • Lingxi Intelligent Driving System 2.0
  • BEV Perception Technology

3 Intelligent Driving Solutions

  • Intelligent Driving Solutions
  • Solution Comparison: Parking Functions and Sensor Configurations
  • Solution Comparison: Driving Functions and Sensor Configurations
  • Lingxi Intelligent Driving Technology Route

4 Mass Production and Application of DJI Automotive's Intelligent Driving Solutions

  • Application Case
  • Lingxi Intelligent Driving 2.0 for Baojun Yunduo

5 How Does DJI Automotive Develop NOA System?

  • Introduction to DJI Automotive's NOA Solution
  • How to Choose Perception Routes for Global Vision
  • How to Establish Environment Perception and Prediction Capabilities
  • How to Establish High-precision Local Pose Estimation Capabilities
  • DJI Automotive's Algorithms and Models to Enable NOA
  • How DJI Automotive Realizes NOA
  • How to Ensure Reliability

For more information about this report visit https://www.researchandmarkets.com/r/nvv0p0

About ResearchAndMarkets.com

ResearchAndMarkets.com is the world's leading source for international market research reports and market data. We provide you with the latest data on international and regional markets, key industries, the top companies, new products and the latest trends.

Contacts

ResearchAndMarkets.com

Laura Wood, Senior Press Manager

press@researchandmarkets.com



For E.S.T Office Hours Call 1-917-300-0470

For U.S./ CAN Toll Free Call 1-800-526-8630

For GMT Office Hours Call +353-1-416-8900

Data & News supplied by www.cloudquote.io
Stock quotes supplied by Barchart
Quotes delayed at least 20 minutes.
By accessing this page, you agree to the following
Privacy Policy and Terms and Conditions.