Volume 2025, Issue 1 5353470
Research Article
Open Access

Lidar-Binocular Camera-Integrated Navigation System for Underground Parking

Wei He

Wei He

Academy for Engineering & Technology , School of Computer Science , Fudan University , Shanghai , 200433 , China , fudan.edu.cn

Search for more papers by this author
Rui Li

Corresponding Author

Rui Li

Key Laboratory of Maritime Intelligent Cyberspace Technology of Ministry of Education , Hohai University , Changzhou , 213200 , China , hhu.edu.cn

Search for more papers by this author
Wenjie Liao

Wenjie Liao

Artificial Intelligence Industry Academy School of Computer Engineering , Jiangsu University of Technology , Changzhou , 213001 , China , just.edu.cn

Shanghai Huace Navigation Technology Co., Ltd , Shanghai , 201702 , China

Search for more papers by this author
First published: 19 May 2025
Academic Editor: Roberta Di Pace

Abstract

It is well known that vehicles highly rely on satellite navigation in an open intelligent traffic environment. However, satellite navigation cannot obtain accurate positioning information for vehicles in the interior of underground garage, as they comprise a semienclosed navigation space, worse light than outdoors in a special traffic environment. To address this problem in this research, the Lidar-binocular camera-integrated navigation system (LBCINS) is established for underground parking indoor environment. The obtained Lidar data from the simulation experiment are preprocessed, and the matching results of the inertial navigation system (INS) under the normal distributions transform (NDT) algorithm and the iterative closest point (ICP) algorithm are compared. The simulation experiment results show that in the complex underground parking environment, the INS under Lidar-NDT algorithm with binocular camera achieves a better performance. Then, in the field experiment, the 3D cloud point data were collected by the test vehicle that equipped with the proposed navigation system from an underground parking and obtained 199 pairs of feature points’ distances. Finally, four different statistical methods were used to analyze the calculated distance errors. Results show that under different error statistical methods, the distance error values of the proposed navigation system are 0.00901, 0.059, 0.00766, and 0.087 m, respectively which present a much higher precision than 5.0 m in the specification requested for inertial-integrated navigation terminal.

1. Introduction

With the development of cities, a sharp increase of private cars ownership makes parking difficulties increasing prominent. Research data show that car ownership in China had reached 217 million. However, the ratio of private cars to parking lots is about 1:0.8 in large cities, only about 1:0.5 in small and medium cities [1]. That means there is still an insufficiency of more than 50 million car parking lots in China. Limited by the shortage of ground space resources, therefore accelerating the construction of underground parking lots has also become a significant part of the plan in urban areas. However, due to the poor signal penetration capability of satellites, worse light than outdoors makes it difficult to get a high precision positioning and navigation at indoor environments. Also, indoor positioning data cannot be acquired correctly, which make underground parking position and navigation become a special case.

In recent years, the intelligent transportation has got a vigorous development, for large cities have a lot of researchers paying their attention on shared autonomous vehicles to relieve the pressure of urban traffic and parking problems [2]. Also, the impact of automated vehicles on traffic assignment, mode split, traffic risk evaluation and parking behavior also get more and more attentions [3]. Meanwhile, artificial intelligence (AI) becomes a clear development trend in a lot of industries. AI technology can help make many fields achieve “unmanned” operation, reducing labor costs and improving system operation safety and efficiency [46]. Underground parking is a vertical field of the international union (ITU) industry that combined computer vision (CV), automatic number plate recognition (ANPR), and intelligent camera technology and gradually evolving into an “unmanned” operation form [7, 8]. Jiang et al. based on ultra-wideband (UWB) technique proposed a three-stage practical and economical layout planning approach for UWB base stations, considering regional differentiation accuracy requirements for automated valet parking in an indoor environment [9]. Vision-Zenit released a series of binocular smart cameras, which is an innovative system that employs a high-accuracy recognition and localization approach, utilizing a heterogeneous dual lens to establish binocular stereo vision, extract depth information from the scene, and facilitate three-dimensional (3D) vehicle modeling [10].

According to a study on light detection and ranging (Lidar) in navigation systems, the National Aeronautics and Space Administration (NASA) has successfully developed airborne marine Lidar with scanning and high-velocity data recording capabilities. Bathymetric surveys were performed in Chesapeake Bay and the Atlantic Ocean, and the seafloor topography was drawn where the water depth was less than 10 m [11]. Therefore, the enormous application potential contained in airborne Lidar began to attract attention and was soon applied to land terrain measurement [12]. The big progress got in Global Positioning System (GPS) and inertial navigation systems (INSs) recent years has made it possible to locate and determine attitudes accurately and instantaneously during laser scanning [13]. Regarding research on integrated navigation, Wu et al. studied the map mapping (MM), dead reckoning (DR), INSs, and GPS-integrated navigation methods. They also found that the purpose of integrated navigation was to reduce costs, improve navigation accuracy, enhance reliability, and develop digital navigation [14].

Li et al. based on the integrated navigation of the GPS and the Micro–Electro–Mechanical System (MEMS) inertial navigation system to solved the problem of switching navigation algorithms in the entrance and exit of underground parking lots and proposed a low-cost vehicular seamless navigation technology based on the reduced inertial sensor system (RISS)/GPS between the outdoors and an underground garage [15]. Li et al. used Lidar and INS for indoor combination to realize unmanned aerial vehicle (UAV) positioning technology under technology fusion [16]. Fu et al. detected vehicle motion through the INS and sensed the surrounding environment through Lidar and millimeter wave radar, both of which were combined to achieve the automatic localization of driverless cars on the road [17]. Azurmendi proposed a simultaneous object-detection and distance-estimation algorithm based on YOLOv5 for obstacle detection in indoor autonomous vehicles [18]. Liu et al. explored the performance of INS and global navigation satellite system (GNSS) and found that the processor computing performance of the integration of INS/GNSS is lower, while the processor performance of the INS model of the graph optimization fusion algorithm is better [19]. For the Lidar application field, Kim et al. investigated different Lidar methods in terms of light source, ranging, and imaging. They found that a 905 nm light source had a low cost and was relatively mature [20]. Liu et al. used a 16-channel Lidar to help the visual system build an accurate semantic map and proposed an AVP-SLAM solution for underground parking lot [21].

The development of autonomous cameras and image processing technology has greatly promoted visual navigation technology. Visual navigation is an emerging navigation method that uses charge-coupled device (CCD) cameras to capture road images and adopts machine vision and other related technologies to identify paths and realize automatic navigation. Visual navigation is currently a new technology. It captures images of the surrounding environment through cameras, filters and calculates the images, completes its attitude determination and path recognition, and makes navigation decisions [22]. Because visual navigation usually adopts a passive working mode, simple equipment, low cost, good economy, wide application range, and theoretically the best guidance flexibility, it has developed very rapidly in recent years. It does not rely on any external equipment and only needs to calculate the information in the storage system and the environment to obtain navigation information. Luque-Vega realized autonomous intelligent identification and navigation of intelligent vehicles based on an intelligent vehicle system and utilized visual navigation technology [23].

However, the studies on Lidar sensor navigation systems about underground parking are most focus on autonomous driving and intelligent robots. With the development of Internet of Things technology, intelligent transportation and intelligent travel have attracted increasing attention from the public. Through independent innovation and technology research and development, “electronic parking fees and management, real-time intelligent update of parking information, and maximum utilization of parking space resources” will be the research hot point focus on parking problems.

In summary, the current research explores the Lidar imaging method in different ranging light sources and provides a theoretical basis for Lidar research in INSs but still faces the following challenges:
  • 1.

    Compared with ground traffic scenarios, underground parking lots are more challenging to localize due to enclosed environments and GNSS failure. Existing localization methods for underground parking lots require additional devices such as WIFI or UWB with unstable localization. A robust vision-based localization method to all carpark signs in underground parking lots is necessary.

  • 2.

    A high-precision localization based on feature matching is critical. However, current methods still suffer from feature extraction problems due to low efficiency and instability. Also, high repetition and complex conditions of underground parking lots exacerbate the difficulty of feature matching, the affect, the accuracy, and robustness of localization. A new visual feature for indoor scenarios must be proposed.

The contributions of this research are as follows:
  • 1.

    An Lidar-binocular camera-integrated navigation system (LBCINS) for underground parking is designed by integrating Lidar-binocular camera. We introduced the characteristics and functions of the Intelligent Sensor Networks (ISNs) and analyzed the application of light detection and ranging (Lidar) in automatic navigation. The combination of Lidar and binocular cameras enables vehicle navigation and obstacle recognition in underground parking to address the high costs of current methods that require additional auxiliary equipment and markers. Low-cost, high-precision localization is realized by using Binocular Camera-based sensors.

  • 2.

    A theoretical basis is proposed for exploring Lidar-binocular camera on INS. The LBCINS is formed by fusing Lidar and a binocular camera. Finally, the LBCINS performance under the light detection and ranging-normal distributions transform (NDT) algorithm is analyzed to address the instability of current features. The method has good matching stability and strong robustness to cope with the changing of illumination in underground parking scenario.

The rest of this paper is organized as follows. Section 2 describes the LBCINS. Section 3 describes experimental environment and settings. The experimental results of the proposed methods are shown in Section 4, followed by discussion and conclusions in Section 5.

2. Materials

2.1. Application of ISN Technology in the Underground Parking

Sensors are a significant part of modern information technology. Traditionally, sensor output is mostly an analog signal without signal processing and networking functions. They should be connected to specific measuring instruments to perform transmission functions and signal processing. Intelligent sensors can process raw data internally, exchange data with the outside world through standard interfaces, and change the way the sensor works through software control according to actual demands, achieving intelligence and networking. Because the standard bus interface is adopted, the intelligent sensor has good openness and scalability, which brings much expansion space for the system [24].

The ISN is composed of sensors, microprocessors, and related electrical circuits shown as Figure 1.

Details are in the caption following the image
The composition networks of ISN.

The sensor converts the measured chemical or physical quantities into corresponding electrical signals and then sends them to signal modulation circuit, which is filtered, amplified, and analog-to-digital (A/D) converter, and the microprocessor. The microprocessor calculates, stores, and analyzes the received signal through the feedback loop; it adjusts the sensor and signal conditioning circuit to regulate and control the measurement process, meanwhile transmits the processing results to the output interface. After being processed by the interface circuit, the digitized measurement results are output based on the output format and interface customization. The microprocessor is the core of the smart sensor because it fully utilizes various software functions, makes the sensor intelligent, and greatly improves the corresponding performance [25]. The ISN features are exhibited in Table 1.

Table 1. The characteristics of this sensor network.
Characteristics Channel of realization
High precision Automatic zeroing to eliminate accidental errors
High reliability and high stability Automatic compensation for drift of system characteristics for emergency handling of abnormal situations
High signal-to-noise ratio (SNR) and high resolution Correlation analysis and processing such as digital filtering eliminates cross-sensitivity in multiparameter state
Good adaptability Operated in an optimal low-power mode to adapt to different environments and improve transmission efficiency
Higher cost performance Microprocessor/microcomputer combination, implemented using inexpensive integrated circuit technology and chips and powerful software

The function of the intelligent sensor is designed by simulating the coordinated actions of the human brain and senses, combined with long-term research of testing technology and practical experience. It is an independent intelligent unit, which reduces the stringent requirements on the original hardware performance. Due to the software, the sensor performance is greatly improved [26]. The main functions of the intelligent sensor are demonstrated in Figure 2.

Details are in the caption following the image
The main intelligent sensor functions.

In Figure 2, the intelligent sensor functions include seven aspects. Various physical and chemical quantities can be measured by the composite sensitive function and information that more fully reflects the laws of motion of matter can be provided. The adaptive function extends the life of the device. It also automatically adapts to various environmental conditions, thus expanding the field of work. Adaptive technology improves sensor repeatability and accuracy. The self-diagnostic function performs a self-test at power-on, and this is a diagnostic test to determine if a component is faulty. Moreover, online calibration can be performed through the usage time, and the microprocessor adopts the measured characteristic data stored in electrically erasable programmable read-only memory (E2PROM) for comparison and calibration. Intelligent sensors store a large amount of information that users can inquire about at any moment. The data processing function amplifies and digitizes the signal and then implements signal conditioning in the software. The configuration function can choose the requirements at will according to the needs and can be used for different tasks or applications depending on the scenario. The intelligent sensor has a microcontroller, so digital serial communication can be configured naturally via an external connection.

To apply ISN to underground parking navigation systems, a large number of intelligent sensor nodes are randomly scattered in the target monitoring area. Network topology planning is carried out on the ISN. During the data collection process, the effective data collected by each smart sensor need to be sent back to each monitoring node and base station in a timely and accurate manner. The intelligent sensor node collects a large quantity of real-time data for compression processing. The system provides total number of occupied parking spaces in parking lot and more specific area information. Intelligent sensors are placed at the entrances, exits, and zone transitions of each underground parking area. Sensors at the entrances and exits wirelessly transmit vehicle data from these two locations to a central base station at the entrances and exits. Sensor nodes monitoring zone transition points detect traffic flow and direction and determine if vehicles are moving between zones and send data to a central base station, which analyzes all incoming data and provides real-time total available parking spaces in the underground parking spaces and the number of parking spaces in each area.

With the descriptions above, the ISN applied to underground parking is mainly able to implement three functions. The intelligent sensor nodes installed in the parking lot can sense whether there is a car parked in the corresponding parking space and send the data back to the central system. Through ISN, it is convenient for the entire underground parking area to automatically count the number and location of free parking spaces. The wireless transmission nodes distributed in the underground parking guide the vehicle to enter the free parking space and display the running status of the nearby vehicles at any time, avoiding the occurrence of accidents such as scratches. Integrating the ISN and radio frequency identification (RFID) technology can intelligently sense the stay time of vehicles in the parking lot through sensor nodes and only set charging nodes at the exit to realize nonstop charging.

2.2. The Composition and Working Principle of Lidar

Based on the ISN function, Lidar has been widely used in automatic navigation. Lidar is an all-in-one system of lasers, GPS, and IMU to gain point cloud data and produce precise digital 3D models. The combination of these three techniques can acquire the surrounding 3D reality with consistent absolute measurement point positions [27]. The working principle of Lidar is indicated in Figure 3.

Details are in the caption following the image
The operating principle of Lidar.

Figure 3 shows that Lidar uses a pulsed laser to measure distance. Lidar fires a laser beam and calculates the distance from the Lidar to the target point. Data points are formed by measuring how long it takes for light to hit an object or surface and bounce back. Millions of data points may be acquired in the process, which is called a “point cloud.” By processing these data, accurate 3D stereoscopic images are presented in seconds. Since the speed of light is known, the distance between the laser and the measured object can be calculated by the difference between the time when the laser beam is emitted and the backscattered light is received, which similar to radar, transmitting system sends out signals, which are collected by the receiving system after being reflected by the target. The distance to the target is defined by measuring the propagation time of the reflected light [28].

Integrated navigation uses multiple sensors to navigate and locate. Its essence is the information fusion of multiple sensors. Due to the difference in reference datum, the information of each sensor needs to complete the coordinate transformation on the basis of time synchronization and space alignment. The INS in the system performs state estimation to obtain the final navigation parameters. Lidar calculates the 3D coordinates of the target point by measuring the position and attitude of the target point relative to the laser scanning center and then colorizes and texturizes the point cloud according to the mapping relationship between the photo and the object. The reconstructed 3D surface shape of the object is attached to the inherent color texture of the object to make it realistic [29].

2.3. Application of the Binocular Camera in the Navigation System

During binocular positioning process, the two cameras are on the same plane, and the optical axes are parallel to each other, similar to a person’s two eyes. For a feature point on the object, two cameras fixed at different positions are used to take images of the object, and the coordinates of the point on the image planes of the two cameras are obtained. As long as the precise relative positions of the two cameras are known, the coordinates of the feature points in the fixed coordinate system of one camera can be obtained by geometric methods; that is, the feature point positions can be determined [30]. Figure 4 depicts the binocular-positioning process.

Details are in the caption following the image
The binocular-positioning process.
The binocular camera-positioning process includes four stages.
  • Step 1: obtain and calibrate the internal and external parameters of the left and right cameras.

  • Step 2: calibrate and align the left and right images through the stereo coordinates.

  • Step 3: determine the relative positional relationship between the two cameras.

  • Step 4: calculate coordinates.

The positioning system of the binocular vision image is developed by Micro vision (3D vision image), which is used to identify and locate the chip position during the chip-linking process to better cure the chip in the desired position. The two cameras are positioned at the same point S on the image to match the points of the left and right pictures, which involves the binocular correction action. If the point feature on one image is used to match the corresponding point in another 2D image space, this process can be very time consuming [31]. To reduce the computational complexity of the matching search, limit constraints are employed to decrease the matching of corresponding points from a 2D search space to a one-dimensional search space. Figure 4 denotes the limit constraint principle.

In Figure 5, the binocular correction strictly aligns the two images after the distortion is removed in the horizontal direction, and the epipolar lines of the two images are completely on the same horizontal line so that any point on one image matches a point on the other image. The binocular correction must have the same row number, and one-dimensional search of the row will be satisfied to match the corresponding point. The mathematic expression of its binocular correction is shown in the following equation:
()
where xlxf represents the positional disparity of the imaging points on the left and right binocular cameras, T indicates the center distance between the two cameras, and Z portrays the vertical distance from point P to the camera.
Details are in the caption following the image
The limit constraint principle.

3. Methodology

The point cloud data were obtained from testing site under the normal working state of the system and extract the coordinate values of feature points from point cloud data. The system detection environment of this experiment is in Windows 10. Also, the control network is arranged in the selected control field by using GPS static observation to avoid the binocular camera obtaining the high-precision 3D coordinate value of the feature point as the real value of the feature point in the test. This process was completed before the experiment.

3.1. Mechanism Design for LBCINS

Integrated navigation needs multiple sensors to navigate and locate objects, and the core of the system is the information fusion of multiple sensors on the carrier. However, according to the different reference benchmarks of underground parking, the sensor needs to complete the information interaction and coordinate change synchronously in time and space. The LBCINS estimates the situation at the same time and space, then obtains the parameters required for the final navigation. The integrated navigation process of the Lidar-binocular camera is shown in Figure 6.

Details are in the caption following the image
The integrated navigation process of the Lidar-binocular camera.
The information data input by the Lidar will be fused dependent on their feature by the proposed LBCINS shown in Figure 6. The advanced concept of the Lidar stereo matching channel includes extracting 2D features from obstacle-target stereo pairs, obtaining pixel correspondences and, finally, performing disparity calculations. By exploiting the complementarity of Lidar and binocular modalities, the integrated navigation model can produce highly accurate disparity estimates. The fusion features of the binocular camera combine the geometric information of the sparse Lidar depth with red, green, and blue (RGB) images as input and learn a joint feature representation in the cost calculation stage. By using these two proposed techniques, the disparity computation stage produces highly accurate disparity estimates. At each pixel point, according to the discrete disparity value of the corresponding Lidar data, the scan matching algorithm is used to iteratively calculate the adjacent points of the obstacle target. When performing point cloud matching, the normal distributions transform (NDT) algorithm is used to match the optimal point cloud. The concept of this algorithm is to construct a normal distribution of multidimensional variables from reference data. If the transformation parameters make the two-laser data match well, the probability density of the transformed point in the frame of reference is high. The NDT algorithm first divides the surrounding environment of the obstacle target and calculates and analyzes the pixels containing three feature units. Calculating the mean value and probability density of the pixel set is demonstrated in equations (2) to (4):
()
()
()
In equation (2), p(x) shows the set of scanning points from pixels that are extracted, q refers to the mean of the pixel set, and C denotes the covariance matrix of the pixel set and represents the probability density of the reference frame in each pixel set. Polar coordinate transformation is performed on all frame data points to obtain the transformed scan point set of the reference frame. Equation (5) displays the conversion expressions:
()
where (tx, ty) is the translation distance of (x, y); the conversion parameter is , ϕ signifies the rotation angle and denotes the probability distribution sum of all pixels; using p represents a transformation parameter. Equation (6) illustrates the probabilities of all transformed pixels:
()

3.2. Parameters of LBCINS

The input fusion and conditional cost volume normalization (CCVNorm) of Lidar information are tightly integrated with the CCV components commonly used in stereo matching neural networks for binocular cameras. The stereo matching channel of the binocular camera extracts two-dimensional features, obtains the pixel correspondence, and performs the final disparity calculation. In the cost regularization stage of stereo matching, CCVNorm replaces the batch normalization layer and modulates the characteristic quantity of CCV conditioned on the Lidar data. By using these two proposed techniques, the disparity computation stage produces highly accurate disparity estimates.

The positioning system of the binocular camera adopts two Microvision MV-808H cameras, a VS-M1024 continuous zoom lens, and an MV-8002 two-way high-definition image acquisition card to capture images simultaneously. The spot welding position of the chip is for accurate positioning when installing. The binocular vision inspection system accurately obtains the coordinate information of the installation or processing position on the circuit board through image analysis, processing, and image measurement, calculates the position coordinates, and provides the operation control of the manipulator. The hardware configuration of the binocular camera positioning is as follows. Camera: VS-808HC color high-definition camera, 520 lines, 1/3″ Sony CCD; resolution: 752∗582. HD frame grabber: MV-8002, which supports secondary development and multiple operating systems. Continuous magnifier: VS-M1024; high brightness ring light emitting diode (LED) for machine vision light source: VS-RL100R, and its advantages are adjustable brightness, low temperature, balanced, no flicker, no shadow, and long service life. IMU: WHEELTEC N300, with redundant sensor technology, three gyroscopes with zero-bias stability of 2°/h are fused to achieve higher 3D angle measurement accuracy and contains two independent sets of three-axis gyroscope, three-axis accelerometer, three-axis magnetometer and thermometer, belonging to the 12-axis attitude sensor series.

Binocular vision alignment system software includes Micro vision, one video cable, one trigger plug, and one camera support rod. The image extraction process of the LBCINS is outlined in Figure 7.

Details are in the caption following the image
The image extraction process of the LBCINS.

4. Experiment

4.1. Simulation Experiment

4.1.1. Analysis of the Preprocessing Results of Lidar Data

Before deploying the field experiment in true environment, we design the simulation environment to preliminary explore the performance of the proposed LBCINS. A great feature of the robot-operating system (ROS) is that it has a large number of useful components and tools. In this research, we introduced the Gazebo, a 3D dynamic simulator with the ability to accurately and efficiently simulate populations of robots in complex indoor and outdoor environments. Similar to game engines, Gazebo offers physics simulation at a much higher degree of fidelity, a suite of sensors, and interfaces for both users and programs.

The simulation environment of integrated Lidar-binocular camera navigation is constructed as shown in Figure 8, which is a closed rectangular annular corridor space with three cubic obstacles distributed. The constructed moving object has dynamic characteristics, and the Lidar carries can output obstacle information in the simulated environment in real time. Figure 8 presents a simulation environment for LBCINS.

Details are in the caption following the image
Simulation environment for LBCINS.

In this experiment, the order p of the filtering polynomial is set as 3, M = 30, and the collected data are preprocessed by Lidar data. Figure 9 exhibits the Lidar data preprocessing effect.

Details are in the caption following the image
Comparison of targeted obstacle trajectory Lidar data before and after preprocessing.

Figure 9 illustrates the Lidar data outcome of the targeted obstacle moving trajectory distance error before and after preprocessing. Compared with the processed data, the raw data show a bigger scatter than processed ones, which cannot show the complete real state of the obstacle target. After preprocessing, the degree of data scatter changes, the data noise is reduced, and closer to a straight line, which can better reflect the real state of the complete targeted obstacle. Therefore, the Lidar data-preprocessing process proposed by this research is effective.

4.1.2. Performance Analysis of LBCINS Under the Lidar-NDT Algorithm

To explore the performance of LBCINS under the Lidar-NDT algorithm, the distance error matching results as shown in Figure 10, and the pixel conversion heading angle of the obstacle target matched under the NDT algorithm as shown in Figure 11.

Details are in the caption following the image
Matching results of LBCINS under the Lidar-NDT algorithm.
Details are in the caption following the image
The scan-matching results of the NDT algorithm in the simulation environment.

From Figure 10, we can see that the scan matching results of the NDT algorithm is overall in line with the previous scan data and the last scan data. Besides, along with the targeted obstacle moving the scan matching result of the algorithm and the previous scan data are getting closer, which indicate that the matching result of the LBCINS under the NDT algorithm is feasible. In Figure 11, we can see that the actual value and calculated value of pixel heading angle is 12.43 and 12.51, respectively, and the actual value and calculated value of distance are 0.307 and 0.144. The absolute value of error is 0.08 and 0.163, respectively, which are smaller than the integrated navigation accuracy requirements in the specification for BDS/MEMS inertial-integrated navigation terminal, that is, the minimum distance error is no more than 5 m and the minimum heading error is no more than 1.5°, which illustrate that the matching calculation result of the NDT algorithm can meet the requirements of the combined calculation in the LBCINS.

To further verify the LBCINS performance under the Lidar-NDT algorithm, the NDT algorithm is being compared with the iterative closest point (ICP) algorithm in terms of navigation performance in different environmental conditions. The LBCINS performance of underground parking is analyzed in three scenes. The LBCINS performance under the NDT algorithm is shown in Figure 12. The LBCINS performance under the ICP algorithm is displayed in Figure 13.

Details are in the caption following the image
The LBCINS performance under the NDT algorithm.
Details are in the caption following the image
The performance of the LBCINS under the ICP algorithm.

As shown in the Figures 12 and 13, both the LBCINS under the NDT and ICP algorithms present largest errors in the navigation position and attitude when the line and surface features are missing and the errors are smaller when in the scenes of rich and sparse features. This means that in the case of the existence of targeted obstacle characteristics in underground parking, the navigation system performance can achieve a more accurate navigation direction in the navigation position and attitude. However, the integrated navigation under the ICP algorithm presents bigger errors in scenes where line and surface features are sparse or missing, especially in the scenes where line and surface features are missing. The LBCINS under the NDT algorithm performs better than LBCINS under the ICP algorithm.

4.2. Field Experiment

In the field experiment, cloud point data were collected from an underground parking divided into south and north areas. The test vehicle was designed to enter this underground parking from west entrance following the green arrow route to collect the data at south area and then enter from east entrance following the blue arrow route to collect the data at north area. The yellow boxes represent the no-parking signs. The test vehicle data collection route is shown in Figure 14.

Details are in the caption following the image
Test vehicle data collection route.

4.2.1. Cloud Point Data Collection

The test vehicle, mainly data collection equipment, is shown in Figure 15. These equipment include power supply unit, miniPC with AMD R9 5900HX and 32 GB RAM, WHEELTEC N300 Inertial measurement unit (IMU), an Intel RealSense Depth Camera D455 camera, three imx355 industrial vision cameras, Robosense Helios-16 lidar, and G200mini integrated navigation system. The view field of D455 camera is 86° in horizontal and 57° in vertical. The field view of Robosense Helios 16 is 360° in horizontal and 30° (−15°∼+15°) in vertical.

Details are in the caption following the image
Integrated binocular camera and radar equipment of test vehicle. (a) Integrated binocular camera and radar equipment. (b) Test vehicle.
Details are in the caption following the image
Integrated binocular camera and radar equipment of test vehicle. (a) Integrated binocular camera and radar equipment. (b) Test vehicle.

In the field experiment, we get the 3D point cloud data of vehicles at underground parking environment; the working flow of multisensor data interaction system is as shown in Figure 16.

Details are in the caption following the image
The multisensor data interaction system.

4.2.2. Data Processing

We stored the point cloud data from north and south areas in two bags, respectively, and calibrate the camera internal parameters and camera radar external parameter information as shown in Figure 17.

Details are in the caption following the image
Collected data bags.

In order to verify the accuracy of multisensor data, camera internal parameters, and camera radar external parameters, the R3live is used for multimodal data fusion to generate RGB point cloud scene model. The point cloud scenes’ model maps of the two areas are shown in Figure 18.

Details are in the caption following the image
Point cloud scenes’ model maps. (a) South area. (b) North area.
Details are in the caption following the image
Point cloud scenes’ model maps. (a) South area. (b) North area.

4.2.3. Data Analysis

In the colored point cloud model established through R3live, we randomly marked 398 feature points containing special objects such as ground signs, support columns, signs, and fire cabinets and then combined them in pairs to obtain 199 feature point distances as shown in Figure 19.

Details are in the caption following the image
Feature point distance mark.

We express the actual measured distance by Si, the model calculated distance by , and the distance error by () and calculate the distance error of the 199 feature points as shown in Figure 20. It can be seen that the point distance is between 0 and 8.4 m and the point distance error is between −0.37 and +0.27 m as shown in Figure 20.

Details are in the caption following the image
Point distance error.
We use manual measurement for actual measured distance Si, and the modeling distance is obtained by the (x, y, z) three-axis spatial information in the point cloud information, calculated by using the following equation:
()
For the error analysis between actual measured distance and modeling distance, we used statistical methods, respectively, and calculated mean error (ME), mean absolute error (MAE), mean squared error (MSE), and standard deviation (SD). And the calculated distance error by different methods is shown in Table 2. We can see that all the distance error calculated by different error is in acceptable range. The Lidar-binocular camera navigation system proposed in this research has a higher precision of distance position than 5.0 m in the specification requested for inertial integrated navigation terminal.
()
()
()
()
Table 2. Calculated distance error by different methods.
Index Value (m)
ME −0.00901
MAE 0.059
MSE 0.00766
SD 0.087

5. Discussion and Conclusions

In this paper, we developed a LBCINS for underground parking. First, we introduced the Lidar technology in the context of ISN and proposed the integrated navigation system that combined the INS and Lidar-binocular camera. Then, to preliminary match the results and performance of the LBCINS under the Lidar-NDT algorithm, simulation experiment in Gazebo is designed. The simulation results show that in the complex environment of underground parking, the INS under the NDT algorithm and the binocular camera achieves better performance. Finally, field experiment results show the INS and Lidar-binocular camera-combined navigation system proposed in this research have a high precision of distance position compared with actual distance. The significance of this research is to establish an INS based on a Lidar-binocular camera so that the system can be applied to an underground parking navigation environment and provide a good research basis for exploring Lidar and binocular cameras on INS.

In the research of INS nondestructive detection under the NDT algorithm, there is a lack of data processing and optimization process for the pixel data of target obstacles, which leads to a long pixel data processing time for obstacle targets. Also, there is no specific experimental analysis of the specific underground parking environment. Thus, a performance analysis of the LBCINS carried out based on the recognition of different obstacles encountered in underground parking is needed. In the future research, the performance analysis of the LBCINS should be combined with the specific underground parking environments, making the integrated navigation system combined with real maps to apply the proposed system to the navigation system in more practical scenarios.

Conflicts of Interest

The authors declare no conflicts of interest.

Funding

This work was jointly supported by the Fundamental Research Funds for the Central Universities under Grant B240201136, the Jiangsu Funding Program for Excellent Postdoctoral Talent under Grant 2024ZB335, and the Qingpu District Industry University Research Cooperation Development Foundation under Grant 202314.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

    The full text of this article hosted at iucr.org is unavailable due to technical difficulties.