市场调查报告书
商品编码
1482386
中国自动驾驶模拟产业(2024年)Autonomous Driving Simulation Industry Report, 2024 |
2023年11月17日,工业及资讯化部等三部会发布了《关于智慧网联汽车试驾及道路交通的通知》。截至目前,多家整车厂已获得高速公路或城市L3级自动驾驶测试牌照,包括BYD、BMW、IM、Mercedes-Benz、Deepal、Avatr、ARCFOX、AITO、Jiyue、GAC Aion等。目前,以城市NOA为代表的高级智慧驾驶功能的使用正在加速,L3级及以上的自动驾驶系统设计得足够安全且鲁棒,能够应对城市地区无数的边缘/长尾情况。 。
L3智慧驾驶系统的实际落地需要超过10亿公里的试驾,实际驾驶测试成本高、耗时长、用例覆盖率低。然而,模拟测试可以在短时间内、低成本地解决这个问题。对于Xpeng来说,除了车主每天提供的道路数据外,Xpeng还致力于在虚拟空间中建立极端场景并结合模拟,供智慧驾驶系统学习和理解。截至2023年底,Xpeng模拟里程已达1.22亿公里。
在智慧驾驶的 "三大支柱" 测试方法中,模拟测试透过虚拟环境模拟各种交通场景、路况、天气照度、异常情况,对自动驾驶系统在各种情况下的功能、反应和判断进行评估。的能力。
上图中,自动驾驶仿真平台包含交通场景模拟(静态场景重建与动态场景模拟)、环境感知感测器模拟(摄影机、光达、雷达、GPS/IMU等感测器的建模与模拟)、需支援仿真等,验证从感知到控制的模拟测试。根据测试对象的不同,自动驾驶模拟平台可以用作MIL(Model in the Loop)、SIL(Software in the Loop)、HIL(Hardware in the Loop)、DIL(Driver in the Loop)、VIL(Vehicle in the Loop)启用循环测试。目前,仿真测试公司的能力参差不齐。
趋势一:自动驾驶模拟测试进入高保真、高还原精度模拟阶段。
在整个感知-预测-判断-规划-控制关係中,感知对应的是收集车辆外部环境资讯的各类感测器,如车流、路况、天气、光照、异常等。主要包括摄影机仿真、光达仿真、雷达仿真、定位仿真(GPS、IMU)。
目前,许多公司都在从事细粒度的模拟工程实践,例如忠实地模拟真实的道路环境、动态交通场景、车辆和行人的行为,透过详细的物理现象和动态感测器性能来快速验证自动驾驶系统的性能。
对于PilotD Automotive,基于PilotD PlenRay物理射线技术的完整物理感测器模型提供了电磁波的多路径反射、折射和干扰等详细的物理现象,以及检测丢失率、目标分辨率、不确定性检测和 "重影" " 可以模拟物理现像等动态感测器性能,以实现感测器模型所需的高保真度。迄今为止的模拟还原率接近 95%。
本报告针对中国自动驾驶模拟产业进行调查分析,提供模拟技术、国内外解决方案供应商、模拟测试趋势等资讯。
Autonomous Driving Simulation Research: Three Trends of Simulation Favoring the Implementation of High-level Intelligent Driving.
On November 17, 2023, the Ministry of Industry and Information Technology and other three ministries issued the Notice on Piloting Access and On-road Passage of Intelligent Connected Vehicles. Up to now, many OEMs including BYD, BMW, IM, Mercedes-Benz, Deepal, Avatr, ARCFOX, AITO, Jiyue and GAC Aion have obtained highway or urban L3 autonomous driving test licenses. At present, the application of high-level intelligent driving functions represented by urban NOA is being accelerated, and L3 and above autonomous driving systems should be safe and robust enough to deal with countless edge/long tail cases in urban areas.
The commercialization of L3 intelligent driving systems needs more than one billion kilometers of test mileage, and actual road tests are costly and time-consuming, with low use case coverage. However, simulation tests can quickly solve this problem in a short time and at low cost. In Xpeng's case, in addition to the roadside data provided by car owners every day, Xpeng is working to build extreme scenarios in virtual space combining simulation for the intelligent driving system to learn and understand. By the end of 2023, the simulation mileage of Xpeng had reached 122 million kilometers.
In the "three-pillar" test approach for intelligent driving, in simulation tests, different traffic scenes, road conditions, weather illumination and abnormalities are simulated through the virtual environment to evaluate the functions, response and decision capabilities of the autonomous driving systems in various circumstances.
In the above figure, the autonomous driving simulation platform should support traffic scene simulation (static scene restoration and dynamic scene simulation), environment-aware sensor simulation (modeling and simulation of sensors such as camera, LiDAR, radar and GPS/IMU), vehicle dynamics simulation, etc., so as to verify simulation tests ranging from perception to control. According to tested objects, the autonomous driving simulation platform enables in-the-loop tests such as: model in the loop (MIL), software in the loop (SIL), hardware in the loop (HIL), driver in the loop (DIL) and vehicle in the loop (VIL). At present, simulation test companies vary in capabilities, as shown in the table below.
Trend 1: Autonomous driving simulation tests have entered the precise simulation stage with high fidelity and high reduction.
In the whole link of "perception-prediction-decision-planning-control", perception corresponds to all kinds of sensors which collect the external environment information of the vehicle, such as traffic flow, road conditions, weather, illumination and abnormalities. It mainly involves camera simulation, LiDAR simulation, radar simulation, and positioning simulation (GPS, IMU).
At present, many companies are working on fine-grained simulation engineering practices, for example, high fidelity simulation of real road environment, dynamic traffic scene and vehicle/pedestrian behavior, and accurate restoration of detailed physical phenomena and dynamic sensor performance, so as to quickly verify the performance of autonomous driving systems and provide comprehensive test and verification reports.
In PilotD Automotive's case, the full physical sensor model based on PilotD PlenRay physical ray technology can simulate detailed physical phenomena such as multi-path reflection, refraction and interference of electromagnetic waves, or dynamic sensor performance such as detection loss rate, target resolution, uncertainty detection and "ghost" physical phenomena, so as to obtain the high fidelity required by the sensor model. Up to now, the simulation reduction rate is close to 95%.
NVIDIA DRIVE Sim(TM) supported by NVIDIA Omniverse is an end-to-end simulation platform, physically based multi-sensor simulation with high fidelity. It can generate numerous real-world digital twin scenes. At present, the LiDAR models of HESAI and RoboSense have been integrated into NVIDIA DRIVE Sim to simulate the performance of LiDAR in various aspects such as beam control, user-defined scanning modes and resolution, and generate synthetic data sets. Users such as OEMs or autonomous driving solution providers can directly call the LiDAR models for R&D or testing through DRIVE Sim.
Similar to NVIDIA, the 3D scene and high-precision physical sensor simulation of dSPACE AURELION generates highly realistic raw data for radar, LiDAR and cameras in real time. AURELION optimizes the impact of materials on radar echoes and multi-path ray tracing technology to ensure real-world-like measurement effects (e.g. ghost spot effects). For radar, dSPACE offers DARTS, that is, over-the-air simulation of radar echoes in real time.
In the vehicle's perception of environmental information, it is easy to ignore the simulation of the interaction between vehicles and pedestrians. For example, the Qianxing simulation platform of RisenLighten has added rich and realistic pedestrian models to support the functions of self-defining the micro-trajectory of pedestrians and generating pedestrians in batches. When editing scenes, users can simulate the crowded and sparse pedestrian distribution in reality as needed, and can also build complex long-tail scenes such as pedestrians walking randomly, sudden pedestrians in front, courtesy to people and cars, and dispute on right of way to test the comprehensive performance of the autonomous driving system. The platform also provides different pedestrian behavior style models, covering scenarios such as human-vehicle interaction, crossing the road and crossing the intersection obliquely, to simulate an intelligent pedestrian traffic flow. In addition, with the diverse driver behaviors, the platform models three driving styles of drivers (conservative drivers, conventional drivers and radical drivers), and refines each parameter through a certain probability distribution, so that the driving behaviors of vehicles in the environment are diversified and randomized.
In addition, multi-sensor concurrent simulation tests greatly improve the R&D and testing efficiency of perception algorithms. In terms of engineering practice, in May 2023 51Sim and VCARSYSTEM cooperated to successfully fulfill the closed loop of domain controllers from SIL to HIL in China's autonomous driving tests, and realized the thorough localization of domain controller in-the-loop simulation tests. In this domestic domain controller in-the-loop solution based on Journey 5, the intelligent driving data reinjection system independently developed by VCARSYSTEM supports simultaneous injection of sensor data from multiple high-definition cameras, LiDAR, radar, ultrasonic radar, GNSS&IMU and so on, and easily reproduces specific scenes and environments via 51Sim-One, an autonomous driving simulation test platform, thereby greatly improving the R&D and testing efficiency of perception algorithms.
Trend 2: Automatic generation and scene generalization are essential.
At present, how to build a corner case scene is a big challenge for the industry. It is the significance of simulation tests to reproduce scenes such as high-risk working conditions, extreme weather conditions, complex traffic environment and edge events, which are difficult to cover in actual road tests. Especially for large-scale tests of safety-critical scenes, automated simulation technology based on AI technology is needed to cover more scenarios.
Coverage-based quality is a more detailed and comprehensive autonomous driving safety test method. It focuses on the quality of test coverage, that is, whether the system has experienced various possible situations and scenarios. By defining a range of test cases and test scenes, this method can ensure that autonomous driving systems can be tested in various road conditions, traffic conditions and abnormalities. The quality of coverage can involve changes in road conditions, traffic behaviors, special weather conditions, emergencies and more.
Moveover, AI technology and large language models are gradually integrated into simulation tests, playing an increasingly important role in automatic scene generation, automatic annotation, accelerating the construction of scene libraries, reducing the cost of simulation tests, lowering the threshold of simulation test technology and shortening the vehicle development cycle.
In the case of natural language interaction, 51WORLD's AIGC-Scenario Copilot supports fully natural language interaction. Without tedious manual editing and code, it only needs scene descriptions, for example, "add an action, first change the lane to the right, and then slow down to 0". By using an AI large language model, an autonomous driving simulation test scene conforming to the OpenSCENARIO standard can be generated, and an unknown dangerous scene can also be generated to expand the boundary of the simulation test.
In addition, Huawei Pangu Models feature the following in generation of autonomous driving scenes:
Autonomous driving scene understanding replaces manual annotation and classification, and 10,000 video clips are processed in minutes.
Autonomous driving scene generation, vehicle type change, lane change, scene combination rendering and other applications are realized through NeRF technology.
Autonomous driving pre-annotation replaces manual annotation, and supports 2D, 3D and 4D automatic annotation, with an accuracy rate of over 90%.
Autonomous driving multimodal retrieval supports multi-dimensional retrieval capabilities such as searching for pictures by text and searching for pictures by pictures, and realizes minute-level retrieval of millions of pictures.
In addition to simulation platforms and scene library generalization capabilities mentioned above, simulation evaluation systems for autonomous driving testing are also essential in autonomous driving technology commercialization. Simulation evaluation refers to the evaluation and optimization of all aspects of an autonomous driving system by means of simulation testing to ensure its safe, reliable and efficient operation on actual roads. Simulation evaluation mainly includes autonomous driving system evaluation and simulation test system evaluation, of which simulation test system evaluation includes the evaluation of scenario coverage, scene realness, scene effectiveness and simulation efficiency.
Trend 3: Capitalization and sharing of scene library data help to drag down costs and improve efficiency of high-level autonomous driving training and testing.
In simulation tests, in addition to automatic scene generation based on road data (dSPACE Autera, NI data collection solution, VI-Grade AutoHawk, etc.), an all-scenario synthetic data simulation material library can help developers keep training, testing and verifying autonomous driving systems in massive driving scenes, especially safety-critical scenes, to improve algorithm iteration efficiency and closed-loop test efficiency and depth.
For example, the all-scenario synthetic data simulation library of OASIS DATA, Oasis' autonomous driving data platform, embraces common traffic participants, obstacles, road facilities and other traffic environment elements. Combined with physical sensor simulation models, it further generates multimodal, high-fidelity, and accurately annotated simulation materials on a large scale. In terms of synthetic data generation efficiency, it can generate up to 100,000 frames per day, saving more than 90% of data collection and annotation costs.
In addition, 51Sim DataOne has powerful data-driven capabilities, including Dataverse and Synthverse. Wherein, Dataverse, a data platform, is capable of data cleaning, data calculation, data management, data visualization, data statistics, etc. and enables a data-driven simulation closed loop; Synthverse, a synthetic data platform, can automatically generate 3D scenes based on HD maps, restore specific scenes with high fidelity through 3D reconstruction technology, and generalize dynamic and static elements of 3D reconstructed scenes via scene editing generalization tools.
In November 2023, 51Sim, together with Volcengine, TZTEK and MXNAVI, the ecosystem partners of Horizon Robotics, launched the industry's first full-chain data-driven closed-loop ecosystem solution to accelerate the mass production and application of autonomous vehicles. This solution provides chips, domain controllers, data acquisition, data processing, algorithm training, refeed testing, and simulation software and hardware integrated testing, aiming to solve common problems in the intelligent driving industry such as low data utilization and difficult data-driven closed loop construction, and promote the mass production and application of high-level autonomous driving. 51Sim provides a full set of virtual simulation test capabilities for the data-driven closed-loop ecosystem.
Amid surging demand for training and test data of autonomous driving systems, it is difficult to collect diverse and high-quality long-tail scenes on a large scale, and filter the required scenes. In view of this, some simulation solution providers such as SYNKROTRON, IAE and 51Sim have begun to work on the data assetization of simulation scene libraries (including standard regulatory scenes, accident scenes, natural driving scenes, dangerous/extreme scenes, and reconstruction scenes). They have also made a positive response to the "Three-Year Action Plan for 'Data Elements x" (2024-2026), and assisted the intelligent connection industry with assetization of data, which have been traded on Shenzhen Data Exchange, Shanghai Data Exchange, Suzhou Big Data Exchange, and Northern Big Data Trading Center.
In the aspect of simulation data sharing, it is necessary to handle vehicle-cloud cooperation. For example, in the conventional stand-alone R&D environment, the data silo within a team has been a serious problem. How to achieve an efficient working mode among XIL test engineers, tool chain R&D engineers, algorithm training engineers and algorithm test engineers has been tackled. For example, in April 2024, 51SimOne officially released the cooperative version of the "cloud+terminal" integrated product. Through centralized storage and integrated design, it seamlessly connects the client and the cloud, and supports multi-person cooperation, so that data is fully shared among a team. The client supports local integration and debugging of R&D tasks, and the cloud bolsters phased large-scale automated testing in algorithm development, so that a platform can meet various needs and greatly accelerate the iteration and optimization of autonomous driving algorithms.