Open Access Paper
28 December 2022 Research on construction technology of real digital twin traffic scene based on edge-cloud collaboration
Huibin Duan, Yun Shen, Peng Ding
Author Affiliations +
Proceedings Volume 12506, Third International Conference on Computer Science and Communication Technology (ICCSCT 2022); 125064H (2022) https://doi.org/10.1117/12.2662510
Event: International Conference on Computer Science and Communication Technology (ICCSCT 2022), 2022, Beijing, China
Abstract
Based on the needs of building real traffic twin scenes, this paper proposes a traffic twin scene scheme architecture based on edge-cloud collaboration, and introduces the key technologies involved. This paper explores the data processing scheme of multi-node cooperation when the road traffic flow is large, and combines the detection and identification technology of traffic participants to complete the accurate mapping of traffic participants and twin models, and establish the driving and exit mechanism of the real traffic participant model to achieve High-fidelity digital traffic twin scenarios.

1.

INTRODUCTION

The digitalization of transportation facilities and participants is one of the most important development directions of future intelligent transportation systems. The digital twin system can record, reproduce and even interpret the physical system, which helps to promote the digital evolution of the transportation system and the continuous improvement of autonomous driving. It is currently being widely used in intelligent transportation and smart cities1-4. The currently constructed twin traffic systems are more for simulation, and lack the twin practice of rendering real-time models of vehicles and distinguishing vehicle information. When the traffic flow is large, the construction of a real traffic digital twin system puts forward higher requirements for data acquisition, transmission, processing, model calling, rendering, and driving. Currently, there are few high-fidelity real-time traffic twin systems.

This paper explores a simple collaborative scheduling method of multi-way side edge computing node resources and the accurate invocation and exit mechanism of the traffic participant model. In order to ensure that a large number of traffic flows and people suddenly appear in some road sections, the amount of data increases, and the fast processing and cloud twin rendering are realized when the roadside nodes have insufficient computing power. Based on the real-time data collected by the large-scale deployment of roadside perception devices (cameras, radars, etc.) on digital roads, combined with AI recognition, a dynamic model library is established. It provides a solution for the rapid twin presentation of mixed traffic and the real-time mapping of traffic participants, traffic events, traffic environment and other elements in the metaverse in the future real world. It can precipitate data and provide scenarios and data foundations for virtual training of autonomous vehicles in the metaverse.

2.

A TRAFFIC DIGITAL TWIN SCENE CONSTRUCTION SOLUTION BASED ON EDGE-CLOUD COLLABORATION

2.1

Technical solution architecture

With the continuous maturity of the digital twin system, comprehensive digital modeling of road pavement, surrounding environment, roadside equipment, etc. can be realized by means of point cloud scanning and aerial photography. And use the sensors distributed on the road to fully perceive and dynamically monitor the traffic conditions. Data is processed by roadside computing nodes and rendered in the cloud, combined with digital maps, to achieve accurate mapping and information expression of physical traffic in the twin world, so as to achieve complete connection between physical space and virtual space5,6.

This solution mainly includes three parts: roadside collection terminal, edge node and regional cloud, as shown in Figure 1.

Figure 1.

Architecture of traffic twin technology solution based on edge-cloud collaboration.

00157_PSISDG12506_125064H_page_2_1.jpg

2.1.1

Terminal.

The roadside collection terminal mainly includes sensors such as cameras and radars, which are responsible for the collection and return of real-time traffic information.

2.1.2

Edge Node.

The roadside edge node is characterized by miniaturization. It is mainly responsible for the aggregation and processing of a section of road and intersection information, including data access, data analysis, data forwarding, resource monitoring, communication and other modules to achieve collaborative analysis and forwarding of traffic information.

  • Data access module: It is responsible for accessing roadside ipc, radar, signal and other equipment data and data forwarded by other roadside nodes, and sending multi-source data into the data analysis or forwarding module.

  • Data AI analysis module: It is responsible for analyzing and detecting multi-source data, and sending the analyzed structured data to the data forwarding module.

  • Data forwarding module: It is responsible for forwarding the data after detection and analysis to the regional cloud data access module; forwarding the original data that has not been analyzed due to insufficient node resources to other edge nodes.

  • Resource monitoring module: It monitors the resource usage of roadside nodes, and informs the surrounding nodes of the resource usage through the communication module. If the resource usage rate is higher than the set threshold, the data access module will be notified to send the roadside data directly to other edge nodes for processing through the forwarding module, and the data processing tasks sent by other nodes will be rejected at the same time.

2.1.3

Regional Cloud.

It mainly includes modules such as roadside node data access, roadside edge node management, various model databases, data processing and model calling, digital twin engine, etc., to realize the construction of real traffic twin scenarios.

  • Edge node management module: It is responsible for managing roadside edge nodes in the area and helping them establish communication between nodes.

  • Data processing and model calling module: It receives traffic data sent by the data access module, selects the traffic participant model and related data to send to the engine, and realizes the real-time entry and exit of traffic participants in the virtual world.

  • Digital twin engine module: It is responsible for rendering various models, as well as rendering simulations of traffic environments, such as wind, frost, rain, snow, etc., and using real-time data to drive models to move in the virtual world.

  • Environment and traffic model library: It stores road environment models and traffic models, mainly vehicles and roadside equipment.

  • Dynamic traffic participant model library: It stores current traffic participant models and data.

  • Hybrid traffic database: It stores uploaded structured data such as roadside real-time radar video, as a data source for future traffic simulation training.

2.2

Key technologies for the solution

2.2.1.

Collaborative Data Processing of Multi-way Side Nodes Based on Resource Monitoring.

When the traffic flow peaks, the amount of roadside data increases, and the single edge node may have a processing bottleneck, resulting in the inability to upload valid data in real time and affect the construction of twin traffic scenarios. This paper proposes a scheme to efficiently detect the resource usage of edge nodes and allocate and use the resources of surrounding roadside nodes to ensure the efficient processing and uploading of traffic data. This scheme combines active monitoring of node resources with dynamic monitoring that dynamically adjusts the monitoring frequency according to resource usage. Reducing the monitoring frequency that is too low or too high will result in inaccurate monitoring, waste of resources, and the generation of massive monitoring data to run on system resources. The details are as follows:

  • The regional cloud center selects N adjacent nodes for each roadside edge node according to the stored geographical traffic environment information and the connection of each node, and helps to establish communication connections between nodes. Nodes report resource usage to each other, so that load balancing and coordination can be performed on edge servers close to the client, improving resource utilization, reducing data processing delays, and ensuring fast upload of processed traffic data.

  • Every period T, the edge node will report its current resource usage to its connected N edge nodes. Among them, T is dynamically variable. When the continuous change range of the resources of an edge node is small, T is increased, otherwise, T is decreased. Each edge node resource sets a threshold, and when the resource reaches the threshold, it also actively reports its resource usage to the edge server connected to it. Monitoring period T setting method: When the resource utilization rate of the node received by the resource monitoring module for three consecutive times changes little, it means that the resource utilization rate is in a relatively stable state, and the monitoring period of the resource becomes longer, that is, increase T, T=α*T (α>1). When the resource utilization rate of a certain resource received by the monitoring module for three consecutive times changes greatly, that is, the monitoring period T of the resource is reduced, T=β*T (0<β<1).

  • In this solution, the threshold of edge computing resources is set to 80%. When the resource utilization rate of the edge node reaches 80%, the node’s access data is forwarded to other nodes for processing, and the computing tasks sent by the adjacent edge nodes are no longer received. The method is highly portable, and the node resource monitoring service is lightweight, which is convenient for large-scale and small nodes on the roadside to realize resource coordination of processing data and ensure the timeliness of traffic twin data.

2.2.2

Detection and Identification Technology Based on Multi-source Data Fusion.

Aiming at the problem of traffic information acquisition and identification error in bad light at night or in weather such as wind, snow, rain and fog. Through the fusion of radar and video information, the influence of multiple factors can be effectively reduced, and the target detection and recognition of traffic participants can be accurately realized7. Firstly, the radar and video data are preprocessed, and their detection results are tracked respectively, and then the spatiotemporal coordinates of the multi-source data are unified through coordinate transformation and time alignment. Then, through the deep learning model, the information and data from multiple sensors are automatically analyzed and synthesized under certain criteria, and the required decision-making and estimation are finally completed.

The main solution used by the deep learning model is the neural network. A large number of simple processing units are fused into a complex model, a large amount of data is imported into the neural network for training, and data features of different dimensions are learned, thereby eliminating the interference of non-target noise, thereby eliminating the noise problem caused by too many sensors. Deep learning models have strong fault tolerance and adaptive capabilities, and can simulate complex nonlinear mapping. For example, for vehicle target recognition, the data information contained in the vehicle can be processed to obtain target feature information. Due to the noise caused by multi-sensor data, the fusion process of these uncertain information is equivalent to the uncertainty inference process. The neural network classifies and deduces the received information on the weight distribution of the network through the acquired information of each sensor, and uses a specific learning algorithm to obtain the learned knowledge and obtain the inference mechanism of uncertainty, thereby realizing multi-sensor data fusion recognition ability. Fusion analysis is mainly divided into three types of solutions: data-level fusion, feature-level fusion and decision-level fusion. The data to be processed in data-level fusion are collected under the same type of sensors, so data fusion cannot handle heterogeneous data. The feature-level fusion is used to reflect the attributes of the monitored physical quantities by extracting the feature vectors contained in the collected data, and is a fusion of the characteristics of the monitored objects. Decision-level fusion performs certain discrimination, classification, and simple logical operations according to the data features obtained by feature-level fusion, and makes higher-level decisions according to application requirements, resulting in application-oriented fusion.

We deployed an AI analysis module at the roadside edge node using a feature-level fusion identification scheme as shown in Figure 2. FPN network is used for feature extraction and prediction, and AVOD framework is selected as the basic framework of multi-source data fusion algorithm to detect and identify traffic targets. Traffic participant data is obtained, including participant ID, time stamp, target type (car, bus, bicycle, person, etc.), target feature (license plate number, color, vehicle brand, vehicle type etc.), space location, speed, acceleration, angular speed, etc. This data will be used to drive the traffic actor participant of the twin world.

Figure 2.

Detection and identification technology based on multi-source data fusion.

00157_PSISDG12506_125064H_page_4_1.jpg

2.2.3

Construction Method of Real Traffic Twin Scene.

The system imports road environment data and traffic model data, and relies on the digital twin engine for rendering8. All dynamic elements such as the presentation of the environment, the change of the road state, the loading of the vehicle model, and the change of the driving path are driven by data. Using meteorological information and time information, the rendering engine is driven to simulate weather such as day and night, wind, snow, rain and fog under various lighting conditions to achieve a more realistic virtual traffic environment.

For traffic participants, target detection and recognition technology is mainly used to achieve the calibration of the same traffic participant (mainly cars, but also people) through the comparison of ID, license plate or other features, and then the participant model is calibrated and assigned. This technology can realize accurate mapping between real traffic participants and the traffic participant model in the twin system, so as to complete the driving and exit of virtual traffic participants, and maintain the consistency of virtual and real scenes. This paper takes a real-time vehicle in a certain road section as an example to introduce its twin presentation and exit. The following traffic participants are mainly vehicles. The specific steps are as follows, also refer to Figure 1.

  • Traffic data is collected by roadside sensors and enters its directly connected roadside node. If the node has sufficient resources, AI analysis and processing is performed. If the node has insufficient computing resources, its forwarding module sends the data to adjacent nodes for processing.

  • The roadside edge node data AI analysis module processes the roadside multi-source data to obtain the traffic participant data Dt1 at the current sampling time point. The data includes road section information (calibrated by the location of the collection device), ID of each traffic participant (the ID of the same traffic participant is determined to remain unchanged in multiple sampling), traffic participant type, location, three-dimensional size, speed, angle, Acceleration, license plate number, vehicle type and other data. The data will be sent to the data access module of the regional cloud center through the data forwarding module.

  • The data access module of the regional cloud center stores Dt1 in the hybrid traffic database and sends it to the data processing and model calling module. Based on the vehicle type information, a vehicle model is selected from the dynamic traffic participant model library, and model parameters (road segments, IDs, license plates, etc.) are added. The model data is put into the twin engine. If there is no vehicle model of this type in the dynamic library, the relevant model is selected from the traffic model library to store in the dynamic library, the model parameters are set, and fed into the engine.

  • The twin engine loads and renders the input model data in real time, maps it with the location of the high-precision map, and displays it in the virtual world. The model is driven by the corresponding traffic data (speed, angle, etc.) to realize the dynamic presentation of the real traffic situation in the meta universe. The movement track and state change of the model can keep consistent with the physical world.

  • The fusion perception data Dt2 at the next sampling time point enters the data processing and model calling module. This module parses the Dt2 road segment information, related ID and license plate information, and matches the vehicle model parameter information of the road segment from the dynamic model library (at least one of ID and license plate needs to be matched). The matched vehicle information is uploaded to the twin engine and continue to drive the virtual vehicle model with the relevant ID of the road section using the latest data. At the same time, the model parameter lines are deleted that do not match DT2 data in the relevant models under the road section in the dynamic library, which means deleting a certain vehicle information (the vehicle has driven out of the road section and is not captured by the current roadside equipment), and the twin engine is informed to delete the corresponding vehicle in the virtual world. The newly added traffic participants in Dt2 can perform the process in the third step above, a model is selected, model parameters are added, and it is sent to the twin engine for rendering and driving.

The model parameters are a parameter matrix, and each row is related information of a traffic participant ID, which is convenient for the same model to correspond to different traffic participants, such as vehicles of the same type with different license plates. Therefore, deleting a row is equivalent to deleting a traffic participant corresponding to the model. This mode can realize the correspondence between virtual participants and actual traffic participants in the traffic twin, and the entry and exit of its model in the twin traffic scene. Based on accurate target detection and recognition technology, this method can be extended to the realization of virtual presentation and driving of traffic participants including pedestrians, tricycles, etc., and promote the realization of more realistic virtual and real traffic scenes.

3.

APPLICATION OUTLOOK

Digital twins enable transportation systems to migrate from the physical to the digital world. Combined with the rendering engine, the traffic environment is constructed, and through the fusion and driving of various sensor data, the digital expression of physical entities and processes such as motor vehicles, non-motor vehicles, people, pavement elements and roadside elements on the road is realized. More realistic virtual traffic can be widely used in scenarios such as unmanned virtual testing in various environments, analysis and experimentation of traffic problems, remote operation and maintenance of transportation facilities, and development of new transportation vehicles. In various scenarios, the virtual system can be used to set various conditions and parameters that cannot be set in reality for testing and verification, avoiding possible problems and conducting more innovative practices9,10. With the development of related technologies such as network, interaction, AI, simulation, and real-time rendering, digital twins will affect the real world more, and integrate with the real world, and develop towards the development of virtual and real. The digital twin model can generate certain strategies, dynamically control traffic objects, and realize the interaction between virtual and real traffic. The results of the effect on real traffic will be fed back in the form of data, so as to further optimize the model, realize the autonomous evolution of the model, and drive the transportation development and improvement in the physical world.

4.

CONCLUSION

This paper mainly studies the solution of constructing real-time traffic twin scenes of roads. Based on the collaborative data processing technology of multi-way side edge nodes, it reduces the risk that the data volume of the road section is too large to be quickly processed and uploaded during the traffic peak, and ensures the timeliness of the twin presentation of real-time traffic scenes in the virtual world. A method is proposed to use the AI detection and recognition technology of roadside multi-source data to quickly select the traffic participant model to present, drive and exit in the metaverse. This solution can be widely used to realize real digital twin scenarios of mixed road traffic conditions, and build a digital twin analysis and experimental environment for real traffic operation conditions around the pain points and technological bottlenecks in the transportation industry. In the future, high-quality rendering engines will make virtual scenes more realistic, and AI big data will drive virtual and real traffic to form a closed loop, realize autonomous optimization of the traffic network, and promote the formation of a traffic metaverse where the virtual and the real are merged.

REFERENCES

[1] 

Wu, Z. H., Wu, X. B. and Wang, L., “Prospect of development trend of smart transportation under the background of building China into a country with strong transportation network,” Transportation Research, 5 (4), 26 –36 (2019). Google Scholar

[2] 

Tao, F., Liu, W. R. and Liu, J. H., “Digital twin and its application exploration,” Computer Integrated Manufacturing System, 24 (1), 4 –21 (2018). Google Scholar

[3] 

He, Y., Guo, J. and Zheng, X., “From surveillance to digital twin: challenges and recent advances of signal processing for industrial Internet of Things,” IEEE Signal Processing Mag, 35 (5), 120 –129 (2018). https://doi.org/10.1109/MSP.79 Google Scholar

[4] 

Tao, F., et al., “Digital twin in industry: State-of-the-art,” IEEE Trans. Industrial Informatics, 15 (4), 2405 –2415 (2019). https://doi.org/10.1109/TII.9424 Google Scholar

[5] 

Wang, F., Qin, R., Li, J., Yuan, Y. and Wang, X., “Parallel societies: A computing perspective of social digital twins and virtual-real interactions,” IEEE Trans. Comput. Soc. Syst, 7 (1), 2 –7 (2020). https://doi.org/10.1109/TCSS.6570650 Google Scholar

[6] 

Wang, Z., Liao, X., Zhao, X., Han, K., et al., “A digital twin paradigm: Vehicle-to-cloud based advanced driver assistance systems,” in 2020 IEEE 91st Vehicular Technology Conference, (2020). https://doi.org/10.1109/VTC2020-Spring48590.2020 Google Scholar

[7] 

Chen, X., Ma, H. and Wan, J., “Multi-view 3D object detection network for autonomous driving,” in 2017 IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), (2017). https://doi.org/10.1109/CVPR.2017.691 Google Scholar

[8] 

Javaid, M. and Haleem, A., “Virtual reality applications toward the medical field,” Clinical Epidemiology and Global Health, 8 (2), 600 –605 (2020). https://doi.org/10.1016/j.cegh.2019.12.010 Google Scholar

[9] 

Almeaibed, S., AL-Rubaye, S., Tsourdos, A., et al., “Digital twin analysis to promote safety and security in autonomous vehicles,” IEEE Communications Standards Magazine, 5 (1), 40 –46 (2021). https://doi.org/10.1109/MCOMSTD.011.2100004 Google Scholar

[10] 

Wang, Z., Han, K. and Tiwari, P., “Augmented reality-based advanced driver-assistance system for connected vehicles,” in 2020 IEEE International Conference on Systems, Man, and Cybernetics, (2020). https://doi.org/10.1109/SMC42975.2020 Google Scholar
© (2022) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Huibin Duan, Yun Shen, and Peng Ding "Research on construction technology of real digital twin traffic scene based on edge-cloud collaboration", Proc. SPIE 12506, Third International Conference on Computer Science and Communication Technology (ICCSCT 2022), 125064H (28 December 2022); https://doi.org/10.1117/12.2662510
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Data modeling

Roads

Data fusion

Data processing

Data communications

Artificial intelligence

Clouds

Back to Top