iBet uBet web content aggregator. Adding the entire web to your favor.
iBet uBet web content aggregator. Adding the entire web to your favor.



Link to original content: https://doi.org/10.3390/s150923286
INS/GPS/LiDAR Integrated Navigation System for Urban and Indoor Environments Using Hybrid Scan Matching Algorithm
Next Article in Journal
A Novel Phonology- and Radical-Coded Chinese Sign Language Recognition Framework Using Accelerometer and Surface Electromyography Sensors
Next Article in Special Issue
Integrated WiFi/PDR/Smartphone Using an Unscented Kalman Filter Algorithm for 3D Indoor Localization
Previous Article in Journal
Workload Model Based Dynamic Adaptation of Social Internet of Vehicles
Previous Article in Special Issue
An Improved WiFi Indoor Positioning Algorithm by Weighted Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

INS/GPS/LiDAR Integrated Navigation System for Urban and Indoor Environments Using Hybrid Scan Matching Algorithm

1
College of Automation, Harbin Engineering University, 145 Nantong Street, Nangang District, Harbin 150001, China
2
Department of Electrical and Computer Engineering, Royal Military College of Canada, P.O. Box 17000, Station Forces, Kingston, ON K7K 7B4, Canada
*
Author to whom correspondence should be addressed.
Sensors 2015, 15(9), 23286-23302; https://doi.org/10.3390/s150923286
Submission received: 10 August 2015 / Revised: 29 August 2015 / Accepted: 9 September 2015 / Published: 15 September 2015
(This article belongs to the Special Issue Sensors for Indoor Mapping and Navigation)

Abstract

:
This paper takes advantage of the complementary characteristics of Global Positioning System (GPS) and Light Detection and Ranging (LiDAR) to provide periodic corrections to Inertial Navigation System (INS) alternatively in different environmental conditions. In open sky, where GPS signals are available and LiDAR measurements are sparse, GPS is integrated with INS. Meanwhile, in confined outdoor environments and indoors, where GPS is unreliable or unavailable and LiDAR measurements are rich, LiDAR replaces GPS to integrate with INS. This paper also proposes an innovative hybrid scan matching algorithm that combines the feature-based scan matching method and Iterative Closest Point (ICP) based scan matching method. The algorithm can work and transit between two modes depending on the number of matched line features over two scans, thus achieving efficiency and robustness concurrently. Two integration schemes of INS and LiDAR with hybrid scan matching algorithm are implemented and compared. Real experiments are performed on an Unmanned Ground Vehicle (UGV) for both outdoor and indoor environments. Experimental results show that the multi-sensor integrated system can remain sub-meter navigation accuracy during the whole trajectory.

1. Introduction

Navigation can be simply concluded to solve the problem of determining the time varying position and attitude of a moving object based on the equipped proprioceptive sensors (inertial sensors, odometer, etc.) and exteroceptive sensors (laser scanner, camera, etc.). The techniques of positioning can be roughly categorized into relative positioning and absolute positioning. Relative positioning is the process of calculating the current position by using a known initial position, and the estimated or known speed over elapsed time. Inertial Navigation System (INS), which is a self-contained system and is immune to interference, belongs to this category. Due to the advancements in microelectronics and micromachining technologies, inertial sensors have evolved from bulky and expensive mechanical design to compact and affordable grade. However, whether it be the traditional mechanical or the Micro-Electro-Mechanical System (MEMS)-based inertial sensors, they are subject to different level of systematic and stochastic errors. These errors will accumulate in the integration process and grow without bound in the long run. Absolute positioning is to make measurements with respect to given reference systems, including Global Positioning System (GPS), active beacons (like wireless local area network, Bluetooth, ultra-wideband, active Radio-frequency Identification (RFID), etc.), passive beacons (like passive RFID, Near Field Communication (NFC) [1,2]), and manmade or natural landmarks, etc. GPS can provide accurate position, velocity and time information all-day in all weather conditions anywhere on the earth when line of sight to at least four satellites are available. However, GPS is susceptible to signal jamming and outage in challenging environments like urban and indoor areas, where satellite signals are unreliable or unavailable. Therefore, both relative positioning and absolute positioning possess advantages and disadvantages, which makes the standalone positioning technique fail to achieve sustainable accuracy. This limitation can be solved by integrating relative positioning and absolute positioning techniques, like the integration of INS and GPS. Fruitful achievements have been gained in the different integration schemes of INS and GPS. However, urban and indoor environments still remain as challenges due to the unreliability and unavailability of GPS.
Another technique that is commonly used in positioning and mapping applications is Light Detection and Ranging (LiDAR). LiDAR can provide accurate bearing and range measurements about the objects in the environments at high frequency. These measurements at each sampling epoch is a sequence of scanned points representing the general contour of the environment observed at the location of the LiDAR, which is defined as a scan. Comparing the working environments of GPS and LiDAR, it is revealed that they work well under complementary environments [3]. Specifically, in open sky, there is less potential of satellites signals blockage for GPS, while LiDAR tends to have less measurements due to the limited scanning range. In this case, it is either difficult to extract features from LiDAR measurements or to successfully match scanned points to estimate pose. In urban area, where there are tall buildings around, GPS suffers from signal jamming and multipath, while LiDAR can detect the building walls which can be extracted as features in the LiDAR measurements. In indoor environments, GPS loses line of sight to satellites, while LiDAR measurements are rich and it is easy to extract features from them. Therefore, integrating GPS and LiDAR with INS can provide continuous corrections to INS. In this paper, this multi-sensor navigation system for both outdoor and indoor environments is introduced.
The organization of the paper is as follows: Section 2 provides a review of literatures related to LiDAR-based integrated navigation systems and scan matching methods. The quaternion-based INS mechanization is given in Section 3. Section 4 introduces the hybrid scan matching algorithm. Section 5 is the filter design. Section 6 presents the experimental results and analysis. Section 7 is the conclusion.

2. Literature Review

In the literature, LiDAR has been successfully applied on platforms like land vehicles, aerial vehicles and pedestrians to implement autonomous navigation, localization and mapping. The integration of LiDAR with other sensors has been intensively studied, like the integration with inertial sensors in [4,5,6], the integration with INS and camera in [7,8], the integration with GPS in [9,10], and integration with INS and GPS in [3,11]. In most of the INS/LiDAR integration works, LiDAR is loosely coupled with INS, where the pose estimation from LiDAR and INS respectively are fused through Kalman Filter (KF) or Partial Filter (PF). While in [12,13], the tight integration of LiDAR and INS are reported, where extracted feature information is used in the filter as update.
Scan matching is a common technique that has been adopted to estimate pose change from LiDAR measurements. The scan matching method can be broadly classified into three different categories: feature-based scan matching, point-based scan matching and mathematical property-based scan matching [14]. Feature-based scan matching method transforms the raw LiDAR scan points into a more efficient representation. By tracking the feature parameters change or pose change oven two scans, the pose change of the vehicle can be derived. Some commonly used features are line segments [15,16,17,18], corners and jump edges [19,20], lane markers [21] and curvature functions [22]. The point-based scan matching method directly deals with the raw LiDAR scan points by searching and matching corresponding points between two scans. The Iterative Closest Point (ICP) algorithm [23,24] and its variants [25] are the most popular methods to solve the point-based scan matching problem, due to their simplicity and effectiveness. For the mathematical property-based scan matching method, it can be based on histogram [26], Hough transformation [27], normal distribution transform [28] or cross-correlation [29].
In our earlier works [13] and [30], we respectively implemented tightly coupled and loosely coupled INS and LiDAR integration with feature-based scan matching method for 2D indoor navigation. The feature-based scan matching method is efficient and accurate. However, it relies on the features in the environments. If the system is extended to outdoor or indoor unconstructed environments, it may fail to work properly. Besides, a loosely coupled INS and LiDAR integration with ICP-based scan matching method is presented in [31]. ICP-based scan matching method is accurate and independent of environment features. Efforts are made to accelerate this scan matching method. However, iterations to optimally align the scans are inevitable.
Different from existing works, a multi-sensor integrated navigation system for both outdoor and indoor environments with an innovative scan matching algorithm is proposed and implemented on an Unmanned Ground Vehicle (UGV). The main contributions of this paper are summarized as below:
  • GPS and LiDAR are used as aiding systems to alternatively provide periodic corrections to INS in different environments. A quaternion-based error model is used to fuse multi-sensor information.
  • An innovative hybrid scan matching algorithm that combines feature-based scan matching method with ICP-based scan matching method is proposed due to their complementary characteristics.
  • Based on the proposed hybrid scan matching algorithm, both loosely coupled and tightly coupled INS and LiDAR integration are implemented and compared using real experimental data.

3. Quaternion-Based INS Mechanization

The INS mechanization involves different frames and their transformation. The navigation solutions are represented in navigation frame while inertial sensors measure angular velocity and acceleration of the vehicle directly in the body frame. The definitions of the frames used in the paper are listed in Table 1.
Table 1. Frames definitions.
Table 1. Frames definitions.
FramesDefinition
Body frameOrigin: Vehicle center of mass.
Y: Longitudinal (forward) direction.
X: Transversal (lateral) direction.
Z: Up vertical direction.
Navigation frameOrigin: Vehicle center of mass.
Y: True north direction.
X: East direction.
Z: Up direction.
There are many formulations to represent the orientation of the vehicle such as Direct Cosine Matrix (DCM), Eular angles, and quaternion. In this work, the quaternion formulation is used due to its computational efficiency and lack of singularities. The quaternion vector is denoted as:
Q = [ q 0 q 1 q 2 q 3 ] T
The quaternion-based rotation matrix from body frame to navigation frame C b n is given in Equation (2), where the superscript n means the navigation frame and the subscript b denotes the body frame.
C b n = [ q 0 2 + q 1 2 q 2 2 q 3 2 2 ( q 1 q 2 q 0 q 3 ) 2 ( q 1 q 3 + q 0 q 2 ) 2 ( q 1 q 2 + q 0 q 3 ) q 0 2 q 1 2 + q 2 2 q 3 2 2 ( q 2 q 3 q 0 q 1 ) 2 ( q 1 q 3 q 0 q 2 ) 2 ( q 2 q 3 + q 0 q 1 ) q 0 2 q 1 2 q 2 2 + q 3 2 ]

3.1. Position Mechanization Equations

The position vector p is expressed in geodetic coordinate in the Earth-centered Earth-fixed (ECEF) frame. The position vector is given as:
p = [ φ λ h ]
where φ is the latitude, λ is the longitude, h is the altitude. As the vehicle moves, the rate of change in position can be expressed as below:
φ ˙ = v n R M + h
λ ˙ = v e ( R N + h ) c o s φ
h ˙ = v u
where v e , v n and v u are the velocity components in the east, north and up directions respectively. R M is the meridian radius of the ellipsoid while R N is the normal radius of the ellipsoid.

3.2. Velocity Mechanization Equations

The measurements of the three-axis accelerometer are the vehicle specific force in the body frame. It should be transformed to navigation frame first, and then integrates to derive vehicle velocity in the navigation frame. To transform the specific force from body frame to navigation frame, the quaternion-based rotation matrix is used. The transformation is given by Equation (7):
f n = C b n f b
where f b = [ f x f y f z ] T is specific force vector in body frame, and f n = [ f e f n f u ] T is acceleration components in the navigation frame. However, the acceleration derived from Equation (7) cannot be used directly due to the following effects:
(1) Earth rotation rate ω i e . The earth rotation rate in the navigation frame is:
ω i e n = [ 0 ω i e cos φ ω i e sin φ ] T
(2) Angular velocity caused by the change of orientation of the navigation frame with respect to the earth ω e n n , which is also called transportation rate. It can be expresses as:
ω e n n = [ v n R M + h v e R N + h v e tan φ R N + h ] T
(3) The earth’s gravity field g . The gravity field vector in the navigation frame is:
g = [ 0 0 g ] T
These factors need to be compensated for in f n . The rate of change in velocity is given as:
v ˙ = f n ( 2 ω i e n + ω e n n ) × v + g
where v = [ v e v n v u ] T is velocity vector in the navigation frame, ( 2 ω i e n + ω e n n ) × is the skew-symmetric matrix of ( 2 ω i e n + ω e n n ) . Due to the high noise level of the MEMS grade inertial sensors, the earth rotation rate cannot be detected. Furthermore, when the vehicle moves in low speed, the transportation rate is negligible. Therefore, the simplified rate of change in velocity can be rewritten as:
v ˙ = f n + g = C b n f b + g

3.3. Attitude Mechanization Equations

The differential equation of the quaternion is [32]:
Q ˙ = 1 2 [ 0 ω n b b ω n b b ( ω n b b ) × ] Q
where ω n b b is the angular velocity of the body frame with respect to the navigation frame represented in the body frame. It is given in Equation (14). ω i b b = [ ω x ω y ω z ] T is the gyroscope measurement in the body frame, representing the angular velocity of the body frame with respect to the inertial frame.
ω n b b = ω i b b C n b ( ω i e n + ω e n n )
However, since the earth rotation rate and the transportation rate are negligible, Equation (14) can be simplified as:
ω n b b ω i b b
By implementing matrix multiplication in Equation (13), it can be rewritten as:
q ˙ 0 = ( q 1 ω x q 2 ω y q 3 ω z ) q ˙ 1 = ( q 0 ω x q 3 ω y + q 2 ω z ) q ˙ 2 = ( q 3 ω x + q 0 ω y q 1 ω z ) q ˙ 3 = ( q 3 ω x + q 1 ω y + q 0 ω z )
After updating the quaternion using Equation (16), the quaternion-based rotation matrix is also updated. The attitude of the UGV can be computed based on the relationship between attitude angles and the quaternion-based rotation matrix. The formulations of pitch, roll and azimuth are given as:
p = sin 1 ( C b n ( 3 , 2 ) ) r = tan 1 ( C b n ( 3 , 1 ) C b n ( 3 , 3 ) ) A = tan 1 ( C b n ( 1 , 2 ) C b n ( 2 , 2 ) )

4. Hybrid Scan Matching Algorithm

4.1. Feature-Based Scan Matching Method

In this paper, the line feature is used due to its richness in indoor environments. The line feature is characterized by two parameters: the perpendicular distance from the origin of the vehicle body frame to the extracted line ρ , and the angle between the x-axis of the body frame and the norm of the extracted line α as shown in Figure 1. If the estimated vehicle relative position and azimuth changes Δ x , Δ y , and Δ A are used as LiDAR updates in the filter, it is defined as loosely coupled. In this case, at least two non-collinear line features have to be matched over two scans to accurately estimate relative position change. In the other integration scheme, change of the perpendicular distance from LiDAR to the line feature over two scans Δ ρ , and azimuth change Δ A are used as LiDAR updates, which is defined as tightly coupled. In this case, at least one line feature is required to be matched over two scans. Therefore, INS/LiDAR integration with line feature-based scan matching method cannot work when no line features are successfully matched over two scans.
Figure 1. The vehicle pose change over two consecutive scans.
Figure 1. The vehicle pose change over two consecutive scans.
Sensors 15 23286 g001
Figure 1 demonstrates the same line feature in the body frame at epoch i and i + 1 . The relative position changes along both axes are defined as Δ x and Δ y , while the azimuth change is defined as Δ A . The change of the perpendicular distance from LiDAR to the line feature over two scans Δ ρ is connected with the vehicle relative position changes Δ x and Δ y in the following format:
Δ ρ = ρ i ρ i + 1 = Δ x cos ( α i ) + Δ y sin ( α i )
The relation between vehicle azimuth change and line feature angle change is straightforward, which can be represented as:
Δ A = α i + 1 α i

4.2. Iterative Closest Point (ICP) Scan Matching Method

Given the current scan and the reference scan, the ICP-based scan matching method is described as follows:
(1)
Inertial sensors provide initial rotation and translation;
(2)
Transform the current scan using the current rotation and translation;
(3)
For each point in the transformed current scan, find its two closest corresponding points in the reference scan;
(4)
Minimize the sum of the square distance from point in the transformed current scan to the line segment containing the two closest corresponding points.
(5)
Check whether the convergence is reached. If so, the algorithm will continue to process the next new scan. Otherwise, it will return to step two to search new correspondences again and repeat the procedures.
Although the point-to-line distance metric has quadratic convergence with a good initial rotation and translation guess, the correspondences search and the iteration are still time-consuming, which limits its ability for real-time application.

4.3. Hybrid Scan Matching Algorithm

From the description above, the featured-based scan matching method works accurately and efficiently in feature rich environments. However, it heavily relies on the features in the environments and has limited ability to cope with unstructured or cluttered environments. The ICP-based scan matching method can perform accurately and robustly in more general environments. However, it requires time-consuming correspondences search and iterations to optimally align the scans. It is easy to find out that the advantages and disadvantages of the two scan matching methods are complementary. Therefore, the hybrid scan matching algorithm is proposed to combine the two scan matching methods to benefit from the advantages and avoid the disadvantages. Specifically, the hybrid scan matching method will work in the default feature-based scan matching mode. However, when less than two non-collinear line features (for loosely coupled system) or no lines (for tightly coupled system) are matched over two scans, the ICP-based scan matching will be activated, and the hybrid scan matching algorithm will transit to ICP-based scan matching mode. In this way, the hybrid scan matching algorithm can achieve efficiency and robustness concurrently.

5. Filter Design

Information from multi-sensor are fused through an Extended Kalman Filter (EKF). The system model and measurement model are described in the following subsections. The block diagram of the multi-sensor integrated navigation system is shown in Figure 2.
Figure 2. The block diagram of the multi-sensor integrated navigation system.
Figure 2. The block diagram of the multi-sensor integrated navigation system.
Sensors 15 23286 g002

5.1. System Model

In the proposed system, the error state vector for the EKF is defined as:
δ x = [ δ p δ v δ q δ b ω δ b f δ Δ x δ Δ y δ Δ A ] T
where the elements in δ x are error in position δ p , error in velocity δ v , error in quaternion δ q , error in gyroscope bias δ b ω , error in accelerometer biases δ b f , error in relative position changes δ Δ x , δ Δ y , and error in azimuth change δ Δ A respectively. δ b ω and δ b f are modelled as first-order Gauss-Markov process, which is given in the following general form:
x ˙ ( t ) = β x ( t ) + 2 σ 2 β w ( t )
where σ is the standard deviation, β is the reciprocal of the time constant of the autocorrelation function of x ( t ) , and w ( t ) is zero-mean Gaussian noise. By applying Taylor expansion to INS position and velocity mechanization equations, quaternion differential equations, relative position changes and azimuth change equations, the linearized system error model is given as:
δ x ˙ = F δ x + G w
where the transition matrix F can be written as [33]:
F = [ F p p F p v 0 3 × 4 0 3 × 3 0 3 × 3 0 3 × 1 0 3 × 1 0 3 × 1 0 3 × 3 0 3 × 3 F v q 0 3 × 3 C n b 0 3 × 1 0 3 × 1 0 3 × 1 0 4 × 3 0 4 × 3 1 2 F q q F q b ω 0 4 × 3 0 4 × 1 0 4 × 1 0 4 × 1 0 3 × 3 0 3 × 3 0 3 × 4 F b ω 0 3 × 3 0 3 × 1 0 3 × 1 0 3 × 1 0 3 × 3 0 3 × 3 0 3 × 4 0 3 × 3 F b f 0 3 × 1 0 3 × 1 0 3 × 1 0 1 × 3 F Δ x v 0 1 × 4 0 1 × 3 0 1 × 3 0 0 F Δ x Δ A 0 1 × 3 F Δ y v 0 1 × 4 0 1 × 3 0 1 × 3 0 0 F Δ y Δ A 0 1 × 3 0 1 × 3 0 1 × 4 F Δ A b ω 0 1 × 3 0 0 0 ]
In the derivation of F matrix, the terms that are divided by the square of the earth radius are ignored. The elements in F matrix are given as below:
F p p = [ 0 0 0 v e tan φ / ( ( R N + h ) cos φ ) 0 0 0 0 0 ] , F p v = [ 0 1 / ( R M + h ) 0 1 / ( ( R N + h ) cos φ ) 0 0 0 0 1 ] , F v q = [ F v q 1 F v q 2 F v q 3 F v q 4 F v q 4 F v q 3 F v q 2 F v q 1 F v q 3 F v q 4 F v q 1 F v q 2 ] ,
where F v q 1 = 2 ( q 0 ( f x b f x ) q 3 ( f y b f y ) + q 2 ( f z b f z ) ) ,
F v q 2 = 2 ( q 1 ( f x b f x ) + q 2 ( f y b f y ) + q 3 ( f z b f z ) ) , F v q 3 = 2 ( q 2 ( f x b f x ) + q 1 ( f y b f y ) + q 0 ( f z b f z ) ) , F v q 4 = 2 ( q 3 ( f x b f x ) q 0 ( f y b f y ) + q 1 ( f z b f z ) ) , F q q = [ 0 ( ω x b ω x ) ( ω y b ω y ) ( ω z b ω z ) ω x b ω x 0 ω z b ω z ( ω y b ω y ) ω y b ω y ( ω z b ω z ) 0 ω x b ω x ω z b ω z ω y b ω y ( ω x b ω x ) 0 ] F q b ω = 1 2 [ q 1 q 2 q 3 q 0 q 3 q 2 q 3 q 0 q 1 q 2 q 1 q 0 ] ,
F b ω = d i a g ( β ω x , β ω y , β ω z ) , F b f = d i a g ( β f x , β f y , β f z ) , where β ω x , β ω y , β ω z and β f x , β f y , β f z are the reciprocal of the time constant of the random process associated with the gyroscope bias and accelerometer bias respectively.
F Δ x v = cos p sin Δ A v [ v e v n v u ] F Δ y v = cos p cos Δ A v [ v e v n v u ] F Δ x Δ A = v cos p cos Δ A , F Δ y Δ A = v cos p sin Δ A F Δ A b ω = [ cos p sin r sin p cos p cos r ] .
where v = v e 2 + v n 2 + v u 2 is vehicle velocity. The process noise vector contains the noises in gyroscope measurements, noises in accelerometer measurements, and the Gaussian noises associated in the first order Gauss-Markov process of gyroscope bias and accelerometer bias. The non-zeros elements in the noise coupling matrix G are:
G ( 4 : 6 , 4 : 6 ) = C n b G ( 7 : 10 , 1 : 3 ) = F q b ω G ( 19 , 1 : 3 ) = F Δ A b ω , G ( 11 : 13 , 7 : 9 ) = d i a g ( 2 σ ω x 2 β ω x , 2 σ ω y 2 β ω y , 2 σ ω z 2 β ω z ) ,
G ( 14 : 16 , 10 : 12 ) = d i a g ( 2 σ f x 2 β f x , 2 σ f y 2 β f y , 2 σ f z 2 β f z ) , where σ ω x , σ ω y , σ ω z and σ f x , σ f y , σ f z are the standard deviation of the gyroscope bias and accelerometer bias respectively.

5.2. Measurement Model

5.2.1. GPS Measurements

In outdoor environments where GPS signals are available, position and velocity solutions from GPS are integrated with INS at the rate of 1 Hz. The measurement model for INS/GPS loosely coupled scheme is:
[ p g p s p i n s v g p s v i n s ] = I 6 × 6 [ δ p δ v ]
where the subscript means the corresponding system measurements.

5.2.2. LiDAR Measurements

The relative position and azimuth changes from LiDAR and INS respectively are integrated at the rate of 5 Hz. If the estimated vehicle relative position and azimuth changes Δ x , Δ y , and Δ A are used as LiDAR measurements, the measurement model can be given in Equation (24). This measurement model is used when INS and LiDAR are loosely coupled with feature-based scan matching method or ICP-based scan matching method.
[ Δ x l i d a r Δ x i n s Δ y l i d a r Δ y i n s Δ A l i d a r Δ A i n s ] = I 3 × 3 [ δ Δ x δ Δ y δ Δ A ]
where Δ x i n s , Δ y i n s and Δ A i n s are given as below:
Δ x i n s = v cos p sin Δ A · T
Δ y i n s = v cos p cos Δ A · T
Δ A i n s = ω z l T
where T is sampling interval, and ω z l is the angular velocity along the vertical axis projected in the horizontal plane, where all LiDAR scanned points are projected into before line feature extraction and scan matching procedures. It can be written as:
ω z l = ω x cos p sin r + ω y sin p + ω z cos p cos r
If change of the perpendicular distance from LiDAR to the line feature over two scans Δ ρ , and azimuth change Δ A are used as LiDAR measurements, the measurement model can be given in Equation (29). This measurement model is used when INS and LiDAR are tightly coupled with feature-based scan matching method.
[ Δ ρ l i d a r , 1 Δ ρ i n s , 1 Δ ρ l i d a r , n Δ ρ i n s , n Δ A l i d a r , 1 Δ A i n s , 1 Δ A l i d a r , n Δ A i n s , n ] = [ cos α 1 sin α 1 cos α n sin α n 0 n × 1 0 n × 2 I n × 1 ] [ δ Δ x δ Δ y δ Δ A ]
where n is the number of line features that has been matched over two scans, and n 1 should holds. Δ ρ i n s , j represents the perpendicular distance change of the j th matched line feature pair over two scans. It has the following format:
Δ ρ i n s , j = Δ x i n s cos ( α j ) + Δ y i n s sin ( α j )

5.2.3. Odometer and Barometer Measurements

The velocity from odometer and height information from barometer are fused with INS velocity and altitude estimation at the rate of 10 Hz. The measurement model is written as below:
[ v o d o v i n s h b a r h i n s ] = I 4 × 4 [ δ v δ h ]

6. Experimental Results and Analysis

The UGV used in the experiment is called Husky A200 from Clearpath Robotics Inc. (Kitchener, ON, Canada) as shown in Figure 3. This UGV is wirelessly controlled and the platform is equipped with a quadrature encoder, a SICK laser scanner LMS111 [34] and a GPS/INS data logger that includes a GNSS receiver (u-blox LEA-6T), an IMU from VTI Technologies Inc., and a barometer from Measurements Specialties.
Figure 3. Experiment platform: Husky A200.
Figure 3. Experiment platform: Husky A200.
Sensors 15 23286 g003
The experiment starts outdoors on a square without building or foliage blockage. Then the UGV travels from outdoor into the building through four ramps. For the rest of the experiment, the UGV moves along the corridor clockwise for one loop and stops in front of the elevator where it starts the motion indoors. While moving in the corridor, the UGV enters a copy room, a classroom and a lab room respectively. The length of the trajectory that the UGV travels is around 440 m and the travelling time is around 16 min. GPS is integrated with INS/odometer/barometer in the beginning of the experiment outdoor for 234 s before the GPS solutions become unreliable. Then, LiDAR replaces GPS to integrate with INS/odometer/barometer in the ramps part and indoor environments.
Two integration scheme of INS and LiDAR with hybrid scan matching algorithm are implemented and compared. The generated trajectories are very close and only the trajectory from tightly coupled navigation system is presented in Figure 4.
Figure 4. The generated trajectory from tightly coupled navigation system.
Figure 4. The generated trajectory from tightly coupled navigation system.
Sensors 15 23286 g004
For better presentation, the trajectory after GPS ends has been rotated counter clockwise by an angle of 17 degrees. For the indoor part, the trajectory is plotted on the map of the building. While traveling indoor, the UGV stops at seven points for at least 10 s to generate reference locations, which are defined as waypoints. The LiDAR-aided part trajectory, the rooms that UGV enters, and seven waypoints are illustrated in Figure 5.
Figure 5. Indoor part trajectory.
Figure 5. Indoor part trajectory.
Sensors 15 23286 g005
During the LiDAR-aided navigation period, the filter implements 3527 times LiDAR measurement updates. For loosely coupled INS/LiDAR system with hybrid scan matching algorithm, the hybrid scan matching algorithm works in line feature-based scan matching mode for most of the time, and ICP-based scan matching mode is activated for 135 times. Among them, 105 times happen outside on the ramps and 30 times occur mostly in the copy room and classroom. This is reasonable since the railings of the ramps are made of vertical metal bars, and the laser beams can go through the space between the bars and does not reflect back. This will cause discontinuous short line segments. The line segments shorter than a threshold will be discarded in the line extraction process. Therefore, in unstructured environments like ramps and rooms with cluttered objects, it is difficult to extract and match at least two non-collinear line features.
However, for tightly coupled INS/LiDAR system with hybrid scan matching algorithm, ICP-based scan matching mode is activated for only 6 times and all of them happen on the ramps. This means for 99.83% of the trajectory, at least one line feature can be matched over two scans. The comparison of the two integration schemes of INS and LiDAR with hybrid scan matching algorithm is shown in Table 2. It can be clearly seen from the table, while performing the hybrid scan matching algorithm, the line feature-based scan matching scheme plays a dominant role, which maintains the computational efficiency. The ICP-based scan matching mode fills the gaps when less than two non-collinear line features (for loosely couple) or no line features (for tightly coupled) are matched over two scans. This guarantees the accuracy and robustness of the LiDAR-derived measurement updates.
Table 2. Comparison of loosely coupled and tightly coupled INS/LiDAR systems.
Table 2. Comparison of loosely coupled and tightly coupled INS/LiDAR systems.
Integration SchemesFeature-Based Scan Matching Activated Times (Percentage)ICP-Based Scan Matching Activated Times (Percentage)
OutdoorIndoor
Loosely coupled system3392 (96.17%)105 (2.98%)30 (0.85%)
Tightly coupled system3521 (99.83%)6 (0.17%)0
Position errors are calculated at the waypoints to assess the accuracy of the integrated solutions after GPS ends for INS, INS/LiDAR loosely coupled and tightly coupled systems. The results are demonstrated in Table 3. For INS, the position solutions are derived based on dead reckoning. As UGV moves, the position error of INS standalone grows gradually and reaches 13.53 m in the end. This is unsatisfactory for accurate indoor navigation. In contrast, the integrated navigation systems (loosely coupled and tightly coupled) can remain an average sub-meter level accuracy.
Table 3. Position errors.
Table 3. Position errors.
Localization Errors(m)1234567Average
INS1.434.6610.968.736.226.4813.537.43
Loosely coupled system0.180.690.620.600.460.730.600.55
Tightly coupled system0.120.450.270.630.320.510.800.44
If we compare the loosely coupled system and tightly coupled system, it can be found that the position accuracy of the tightly coupled system is slightly higher than the loosely coupled system with shorter running time and less limitation on the environment structure. This makes the tightly coupled system preferable over the loosely coupled system.
The attitude angles are shown in Figure 6. As can be seen from the figure, pitch and roll are relatively smaller when the UGV is stationary or in indoor environments. During the period of 300 s to 400 s, pitch grows from around zeros degree to around five degrees and then decreases to zeros degree again for four times. This corresponds to the four ramps with an inclination angle of five degrees.
Figure 6. Attitude angles.
Figure 6. Attitude angles.
Sensors 15 23286 g006

7. Conclusions

In this paper, a multi-sensor integrated navigation system for both urban and indoor environments has been proposed. GPS and LiDAR are used as aiding systems to alternatively provide periodic corrections to INS in different environments since they work in complementary environments. Besides, due to the complementary properties of feature-based scan matching method and ICP-based scan matching method, the hybrid scan matching algorithm is proposed to combine the two scan matching methods to benefit from the advantages of each method. Two integration schemes of INS and LiDAR are implemented and compared. The tightly coupled system is preferable over the loosely coupled system when considering accuracy, running time and limitations on environments. Experimental results show that sub-meter navigation accuracy and accurate attitude angles estimation can be accomplished during the whole trajectory including outdoor and indoor environments.

Author Contributions

Yanbin Gao provided insights in formulating the ideas and experimental design. Shifei Liu and Mohamed Maher Atia performed the experiments and analyzed the data. Shifei Liu and Mohamed Maher Atia wrote the paper. Aboelmagd Noureldin offered plenty of comments, experimental equipment support and useful discussion.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cassimon, D.; Engelen, P.-J.; Yordanov, V. Compound real option valuation with phase-specific volatility: A multi-phase mobile payments case study. Technovation 2011, 31, 240–255. [Google Scholar] [CrossRef] [Green Version]
  2. Ozdenizci, B.; Coskun, V.; Ok, K. NFC internal: An indoor navigation system. Sensors 2015, 15, 7571–7595. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Soloviev, A. Tight coupling of GPS, laser scanner, and inertial measurements for navigation in urban environments. In Proceedings of the 2008 IEEE/ION Position, Location and Navigation Symposium (PLANS), Monterey, CA, USA, 6–8 May 2008; pp. 511–525.
  4. Kohlbrecher, S.; von Stryk, O.; Meyer, J.; Klingauf, U. A flexible and scalable slam system with full 3D motion estimation. In Proceedings of the 2011 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), Kyoto, Japan, 1–5 November 2011; pp. 155–160.
  5. Hornung, A.; Wurm, K.M.; Bennewitz, M. Humanoid robot localization in complex indoor environments. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Taipei, Taiwan, 18–22 October 2010; pp. 1690–1695.
  6. Bry, A.; Bachrach, A.; Roy, N. State estimation for aggressive flight in GPS-denied environments using onboard sensing. In Proceedings of the 2012 IEEE International Conference on Robotics and Automation (ICRA), Saint Paul, MN, USA, 14–18 May 2012.
  7. Shen, S.; Michael, N.; Kumar, V. Autonomous multi-floor indoor navigation with a computationally constrained MAV. In Proceedings of the 2011 IEEE international conference on Robotics and automation (ICRA), Shanghai, China, 9–13 May 2011; pp. 20–25.
  8. Fallon, M.F.; Johannsson, H.; Brookshire, J.; Teller, S.; Leonard, J.J. Sensor fusion for flexible human-portable building-scale mapping. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vilamoura, Portugal, 7–12 October 2012; pp. 4405–4412.
  9. Joerger, M.; Pervan, B. Range-domain integration of GPS and laser scanner measurements for outdoor navigation. In Proceedings of the ION GNSS 19th International Technical Meeting, Fort Worth, TX, USA, 26–29 September 2006; pp. 1115–1123.
  10. Hentschel, M.; Wulf, O.; Wagner, B. A GPS and laser-based localization for urban and non-urban outdoor environments. In Proceedings of the 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2008, Nice, France, 22–26 September 2008; pp. 149–154.
  11. Jabbour, M.; Bonnifait, P. Backing up GPS in urban areas using a scanning laser. In Proceedings of the 2008 IEEE/ION Position, Location and Navigation Symposium, Monterey, CA, USA, 5–8 May 2008; pp. 505–510.
  12. Soloviev, A.; Bates, D.; van Graas, F. Tight coupling of laser scanner and inertial measurements for a fully autonomous relative navigation solution. Navigation 2007, 54, 189–205. [Google Scholar] [CrossRef]
  13. Liu, S.; Atia, M.M.; Karamat, T.B.; Noureldin, A. A LiDAR-aided indoor navigation system for UGVs. J. Navig. 2015, 68, 253–273. [Google Scholar] [CrossRef]
  14. Martínez, J.L.; González, J.; Morales, J.; Mandow, A.; García‐Cerezo, A.J. Mobile robot motion estimation by 2D scan matching with genetic and iterative closest point algorithms. J. Field Robot. 2006, 23, 21–34. [Google Scholar] [CrossRef]
  15. Garulli, A.; Giannitrapani, A.; Rossi, A.; Vicino, A. Mobile robot slam for line-based environment representation. In Proceedings of the 44th IEEE Conference on Decision and Control, 2005 and 2005 European Control Conference. CDC-ECC’05, Seville, Spain, 12–15 December 2005; pp. 2041–2046.
  16. Arras, K.O.; Siegwart, R.Y. Feature extraction and scene interpretation for map-based navigation and map building. In Proceedings of the SPIE 3210, Mobile Robots XII, Pittsburgh, PA, USA, 25 January 1998; pp. 42–53.
  17. Pfister, S.T.; Roumeliotis, S.I.; Burdick, J.W. Weighted line fitting algorithms for mobile robot map building and efficient data representation. In Proceedings of the IEEE International Conference on Robotics and Automation, ICRA’03, 14–19 September 2003; pp. 1304–1311.
  18. Nguyen, V.; Martinelli, A.; Tomatis, N.; Siegwart, R. A comparison of line extraction algorithms using 2D laser rangefinder for indoor mobile robotics. In Proceedings of the 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2005, Edmonton, AB, Canada, 2–6 August 2005; pp. 1929–1934.
  19. Lingemann, K.; Nüchter, A.; Hertzberg, J.; Surmann, H. High-speed laser localization for mobile robots. Robot. Auton. Syst. 2005, 51, 275–296. [Google Scholar] [CrossRef]
  20. Aghamohammadi, A.A.; Taghirad, H.D.; Tamjidi, A.H.; Mihankhah, E. Feature-based range scan matching for accurate and high speed mobile robot localization. In Proceedings of the 3rd European Conference on Mobile Robots, Freiburg, Germany, 19–21 September 2007.
  21. Kammel, S.; Pitzer, B. LiDAR-based lane marker detection and mapping. In Proceedings of the 2008 IEEE Intelligent Vehicles Symposium, Eindhoven, The Netherlands, 4–6 June 2008; pp. 1137–1142.
  22. Núñez, P.; Vazquez-Martin, R.; Bandera, A.; Sandoval, F. Fast laser scan matching approach based on adaptive curvature estimation for mobile robots. Robotica 2009, 27, 469–479. [Google Scholar] [CrossRef]
  23. Segal, A.V.; Haehnel, D.; Thrun, S. Generalized-ICP. In Proceedings of the Robotics: Science and Systems, Seattle, WA, USA, 28 June–1 July 2009.
  24. Besl, P.J.; Mckay, N.D. A method for registration of 3-D shapes. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 239–256. [Google Scholar] [CrossRef]
  25. Censi, A. An ICP variant using a point-to-line metric. In Proceedings of the 2008 IEEE International Conference on Robotics and Automation, ICRA 2008, Pasadena, CA, USA, 19–23 May 2008; pp. 19–25.
  26. Weiß, G.; von Puttkamer, E. A map based on laserscans without geometric interpretation. In Intelligent Autonomous Systems; IOS Press: Amsterdam, The Netherlands, 1995; pp. 403–407. [Google Scholar]
  27. Censi, A.; Iocchi, L.; Grisetti, G. Scan matching in the Hough domain. In Proceedings of the 2005 IEEE International Conference on Robotics and Automation, ICRA 2005, Barcelona, Spain, 18–22 April 2005; pp. 2739–2744.
  28. Burguera, A.; González, Y.; Oliver, G. On the use of likelihood fields to perform sonar scan matching localization. Auton. Robots 2009, 26, 203–222. [Google Scholar] [CrossRef]
  29. Olson, E.B. Real-time correlative scan matching. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, ICRA’09, Kobe, Japan, 12–17 May 2009; pp. 4387–4393.
  30. Liu, S.; Atia, M.M.; Gao, Y.; Noureldin, A. Adaptive covariance estimation method for LiDAR-aided multi-sensor integrated navigation systems. Micromachines 2015, 6, 196–215. [Google Scholar] [CrossRef]
  31. Liu, S.; Atia, M.M.; Gao, Y.; Noureldin, A. An inertial-aided lidar scan matching algorithm for multisensor land-based navigation. In Proceedings of the 27th International Technical Meeting of the Satellite Division of the Institute of Navigation (ION GNSS+ 2014), Tampa, FL, USA, 8–12 September 2014; pp. 2089–2096.
  32. Farrell, J. Aided Navigation: GPS with High Rate Sensors; McGraw-Hill, Inc.: New York, NY, USA, 2008. [Google Scholar]
  33. Zhou, J. Low-cost MEMS-INS/GPS Integration Using Nonlinear Filtering Approaches. Ph.D. Thesis, Universität Siegen, Siegen, Germany, 18 April 2013. [Google Scholar]
  34. SICK. Operating Instructions: Lms100/111/120 Laser Measurement Systems; SICK AG: Waldkirch, Germany, 2008. [Google Scholar]

Share and Cite

MDPI and ACS Style

Gao, Y.; Liu, S.; Atia, M.M.; Noureldin, A. INS/GPS/LiDAR Integrated Navigation System for Urban and Indoor Environments Using Hybrid Scan Matching Algorithm. Sensors 2015, 15, 23286-23302. https://doi.org/10.3390/s150923286

AMA Style

Gao Y, Liu S, Atia MM, Noureldin A. INS/GPS/LiDAR Integrated Navigation System for Urban and Indoor Environments Using Hybrid Scan Matching Algorithm. Sensors. 2015; 15(9):23286-23302. https://doi.org/10.3390/s150923286

Chicago/Turabian Style

Gao, Yanbin, Shifei Liu, Mohamed M. Atia, and Aboelmagd Noureldin. 2015. "INS/GPS/LiDAR Integrated Navigation System for Urban and Indoor Environments Using Hybrid Scan Matching Algorithm" Sensors 15, no. 9: 23286-23302. https://doi.org/10.3390/s150923286

APA Style

Gao, Y., Liu, S., Atia, M. M., & Noureldin, A. (2015). INS/GPS/LiDAR Integrated Navigation System for Urban and Indoor Environments Using Hybrid Scan Matching Algorithm. Sensors, 15(9), 23286-23302. https://doi.org/10.3390/s150923286

Article Metrics

Back to TopTop