Skip to main content

ORIGINAL RESEARCH article

Front. Robot. AI, 03 January 2022
Sec. Robot Vision and Artificial Perception

Calib-Net: Calibrating the Low-Cost IMU via Deep Convolutional Neural Network

Ruihao Li&#x;Ruihao Li 1Chunlian FuChunlian Fu2Wei Yi&#x;
Wei Yi 1*Xiaodong Yi&#x;Xiaodong Yi 1
  • 1Artificial Intelligence Research Center (AIRC), Defense Innovation Institute, Beijing, China
  • 2Tianjin Artificial Intelligence Innovation Center (TAIIC), Tianjin, China

The low-cost Inertial Measurement Unit (IMU) can provide orientation information and is widely used in our daily life. However, IMUs with bad calibration will provide inaccurate angular velocity and lead to rapid drift of integral orientation in a short time. In this paper, we present the Calib-Net which can achieve the accurate calibration of low-cost IMU via a simple deep convolutional neural network. Following a carefully designed mathematical calibration model, Calib-Net can output compensation components for gyroscope measurements dynamically. Dilation convolution is adopted in Calib-Net for spatio-temporal feature extraction of IMU measurements. We evaluate our proposed system on public datasets quantitively and qualitatively. The experimental results demonstrate that our Calib-Net achieves better calibration performance than other methods, what is more, and the estimated orientation with our Calib-Net is even comparable with the results from visual inertial odometry (VIO) systems.

1 Introduction

Low-cost inertial measurement units (IMU) are widely used in our daily life (Nebot and Durrant-Whyte, 1999; Ahmad et al., 2013; Rehder and Siegwart, 2017). Cheap IMU can provide the attitude and position information by integrating three-axis angular velocity measurements and three-axis linear acceleration measurements. Many devices (such as smartphones, autonomous robots, AR/VR devices, etc) are equipped with low-cost IMU sensors to obtain inertial measurements in order to achieve different features. However, the accuracy of IMU is easily affected by calibration parameters include scaling factors, axes misalignments, etc. Inaccurate calibration will produce imprecise angular velocity and linear acceleration, and the integration operation will lead to rapid error accumulation and estimation drift.

In order to limit the drift as much as possible, precise IMU calibration must be performed for both inertial odometry (IO) and visual-inertial odometry (VIO). A lot of research about IMU calibration or camera/IMU calibration are studied in the past several years. Most methods (Furgale et al., 2013; Rohac et al., 2015; Rehder et al., 2016) prefer to design a mathematical model and take calibration model parameters as constant values. External devices are usually needed for the pre-calibration procedure. Online calibration methods (Qin and Shen, 2018; Yang et al., 2020) which optimize the spatio-temporal camera/IMU model parameters all together are proved effective for the tracking performance of VIO systems. As deep learning technology has achieved great success in the past decade, many researchers also introduce deep learning technology into tasks like orientation estimation (Esfahani et al., 2019b), inertial odometry (Chen et al., 2018; Esfahani et al., 2019a; Brossard et al., 2020a), visual-inertial odometry (Clark et al., 2017; Chen et al., 2019b), and most of them adopt the recurrent neural network which is memory consuming and computing consuming.

In this paper, we propose a lightweight and efficient deep convolutional neural network for low-cost IMU calibration. It is designed based on a mathematical model and is trained driven by historical data. The proposed method can calibrate the IMU measurements dynamically. By using the calibrated IMU measurements, accurate orientation estimation can be obtained and the results can even be comparable with VIO methods.

Our contribution can be summarized as below:

• We present a deep convolutional neural network called Calib-Net for low-cost IMU calibration. The Calib-Net adopts dilation convolution for spatio-temporal feature extraction, and learns to produce the compensation for gyroscope measurements dynamically.

• We introduce a mathematical calibration model to construct the training and calibration framework. Both constant calibration matrix (3 × 3) and Calib-Net hyper-parameters can be trained and optimized driven by a carefully designed loss function.

• We implement the proposed framework and validate it quantitatively and qualitatively on public datasets. The experimental evaluations show that our Calib-Net achieves satisfactory performance in IMU calibration, and is also fruitful for VIO systems.

The rest of this paper is organized as follows. Section 2 introduces the related works of IMU calibration and orientation estimation. Section 3 introduces the overview architecture of the proposed calibration framework. Section 4 describes the details of the used mathematical model, neural network architecture, and the loss function. Section 5 presents the results of experimental evaluations. Finally, the conclusion is drawn in Section 6.

2 Related Work

In the late 1990s, Nebot and Durrant-Whyte (1999) presented an inertial error model for gyros and accelerometers aiming at the IMU calibration for vehicle applications. Tedaldi et al. (2014) proposed an IMU calibration method without using external equipments, they need to move the IMU device by hand and place it in multiple static positions. Based on a specific sensor error model, Rohac et al. (2015) performed the low-cost inertial sensor calibration to obtain the model parameters using the nonlinear optimization under static conditions. Furgale et al. (2013) present the open-source Kalibr toolbox for the spatial and temporal calibration of multiple sensors (cameras/IMUs). Afterward, Rehder et al. (2016) extended the Kalibr toolbox and enabled precise IMU intrinsic calibration following an inertial noise model. They Rehder and Siegwart (2017) then further improved the calibration method by introducing the displacement of individual accelerometer axes into the mathematical noise model. Barrau and Bonnabel (2020) recently adopted Lie group for IMU error propagation. These calibration methods usually adopt different inertial model variants and estimate the parameters of these sensor error models. For the calibration methods above, the parameters (scale factors, misalignment errors, offsets, etc.) are usually taken as constant values. Nevertheless, these parameters are always varying as time goes on and the environment changes.

Aiming at the visual-inertial navigation system, Qin and Shen (2018) presented an online calibration method to optimize model parameters dynamically, and proposed VINS-mono system (Qin et al., 2018) which is one of the state-of-the-art visual-inertial systems. Yang et al. (2020) studied and discussed the necessity and importance of online IMU intrinsic calibration for visual-inertial navigation systems, and proved that online calibration is helpful for improving fusion performance. They also demonstrated an open-source visual-inertial odometry system called OpenVINS (Geneva et al., 2020). Campos et al. took the IMU measurement uncertainty into account (Campos et al., 2020), and proposed to use maximum a posteriori (MAP) estimation during IMU initialization in the ORB-SLAM3 system (Campos et al., 2021). Most of these methods rely on a mathematical model and need to calibrate every device before use.

As to the usage of deep learning in IMU calibration and propagation, Yan et al. (2018) proposed to use the machine learning technology to regress velocity vector with linear accelerations and angular velocities as inputs. Nobre and Heckman (2019) proposed to model the IMU calibration as a Markov Decision Process (MDP) and use reinforcement learning to achieve the regression of calibration parameters. Clark et al. (2017) presented VINet which takes the visual-inertial odometry problem as a sequence-to-sequence learning problem to solve and avoids manual camera/IMU calibration operation. Chen et al. (2018) introduced IONet for inertial odometry using the recurrent neural network. Ionet formulates the odometry as an optimization problem based on the neural network and estimates the trajectories with raw measurements as network input. Afterward, they Chen et al. (2019a) further extended the network and also used it to predict model uncertainty. However, the above learning methods usually adopt neural networks with large weights which will consume lots of computation resources.

Esfahani et al. (2019b) presented a recurrent neural network called OriNet for orientation estimation. 3D orientation can be obtained by OriNet with only low-cost IMU measurements as the network input. They Esfahani et al. (2019a) also proposed AbolDeepIO which simulates the noise model during training and achieves robust inertial odometry with a novel deep neural network. Instead of using recurrent neural networks, Brossard et al. (2020b) adopted a convolutional neural network to estimate the orientation with an IMU device. They Brossard et al. (2020a) further achieved accurate and robust dead-reckoning with only a commercial IMU. The Kalman filter is introduced and a deep neural network is utilized to predict the dynamic parameters of the filter.

3 Methods

3.1 Mathematical Model for Low-Cost IMU Calibration

3.2 Architecture Overview

The overview of our proposed framework is shown in Figure 1. Calib-Net takes sequential gyroscope measurements and acceleration measurements as the input, and outputs the compensation for raw IMU measurements. Dilation convolution is utilized for spatio-temporal feature extraction instead of recurrent neural networks. By designing a mathematical calibration model, the measurements can be dynamically calibrated and corrected with the network output. In this way, accurate orientation can be directly obtained through simple integration. Based on only hundred seconds of labeled data, the hyperparameters of the Calib-Net and the constant model parameters (matrix includes) can be optimized at the same time driven by a carefully designed loss function.

FIGURE 1
www.frontiersin.org

FIGURE 1. The overview of the proposed framework. By selecting a mathematical model, we introduce a simple but effective convolutional neural network for dynamic IMU calibration.

For a typical low-cost Inertial Measurement Unit (IMU), it usually consists of a three-axis gyroscope, a three-axis accelerometer, and sometimes a magnetometer. The three-axis gyroscope can provide angular velocity information, the three-axis accelerometer can provide the linear acceleration information. Considering the noise and bias of the sensor measurements (Rehder et al., 2016), the IMU sensor model can be represented as below:

ω̂tat̂=Cωtat+btωbta+ntωnta(1)

Where angular velocity ωtR3 and linear acceleration atR3 are the measurements of gyroscope and accelerometer, btωR3 and btaR3 are the gyroscope and accelerometer bias, ntωR3 and ntaR3 are additive zero-mean Gaussian noises for gyroscope and accelerometer, angular velocity ω̂tR3 and linear acceleration at̂R3 are the calibrated measurements which would be used for integration. The C is the intrinsic calibration matrix (approximate equals to I6) for the IMU model (Rehder et al., 2016), and can be modeled as:

C=CωC×03×3Ca(2)

Where Cω contains scale factor and axis misalignment for gyroscope measurements, Ca contains scale factor and axis misalignment for accelerometer measurements, both Cω and Ca are approximately equal to identity matrix I3. C× is the coefficient matrix. It indicates the effect that the linear accelerometer has on the gyroscope, and is approximately equal to 03×3. So the gyroscope measurement (angular velocity) model can be represented as:

ω̂t=Cωωt+C×at+btω+ntω=Cωωt+Δωt(3)

In the formula above, we abbreviate the compensation component for gyroscope measurements as:

Δωt=C×at+btω+ntω(4)

Where the compensation component Δωt is related to both gyroscope measurements and accelerometer measurements. By integrating the corrected angular velocity ω̂t, we can achieve the estimation of 3D rigid rotation follows the equation below:

Rt+ΔT=RtExpω̂tΔT(5)

Where RtSO(3) and RtTSO(3) are the rotation matrices. t is the timestamp that IMU outputs measurements, and ΔT is the minimal time interval between two consecutive IMU readings. In our case, the IMU runs at 200 Hz, and the ΔT is 5 ms. Exp (⋅) is the exponential map for SO(3). As indicated in the above formula, incorrect ω̂t will lead to continuous error accumulation as more integrations propagate.

3.3 Calib-Net for IMU Measurement Correction

As shown in Eq. 3 and Figure 1, Cω and Δωt play an important role in IMU calibration and orientation estimation. In most cases, Cω will be taken as a constant matrix, and Δωt is a small time-varying vector affected by many factors. In our proposed calibration system, we use the Caib-Net to estimate the compensation part Δωt for angular velocity measurements. The detailed structure of the Calib-Net is shown in Figure 2. It takes the long temporal sequential gyroscope measurements and accelerometer measurements as the network input, and network can be represented as below:

Δωt=FωtkΔT,atkΔT,,ωt,at(6)

Where k is the number of IMU readings used as the input of the proposed Calib-Net. F (⋅) represents the nonlinear function that the Calib-Net stands for. Dilation convolution is adopted to extract spatio-temporal features of IMU measurements. As shown in Figure 3, by introducing the dilation convolution, the historically temporal measurements are fed into the network and used for feature extraction. Dilation size in dilation convolution indicates the spacing between the convolution kernel points, so with different kernel sizes and dilation sizes, different temporal IMU readings will be taken accordingly. To be specific, the length of the input depends on the product of kernel size and dilation size. In the presented network shown in Figure 2, the maximum product value is 448 which means the network will take k = 448 IMU readings as the network input.

FIGURE 2
www.frontiersin.org

FIGURE 2. The details of the Calib-Net structure. The proposed Calib-Net takes temporal gyroscope measurements and accelerometer measurements as inputs, and outputs the compensation part for angular velocity measurements.

FIGURE 3
www.frontiersin.org

FIGURE 3. The illustration of 1D Dilation Convolution. Dilation size indicates the spacing between the convolution kernel points.

The Calib-Net is composed of two dilation convolutional layers, one residual block, and two fully connected layers. The residual block includes two dilation convolutional layers and one shortcut. The detailed configuration of each layer is given in Figure 2. As to the matrix Cω, we set it to the trainable variable. In this way, following the mathematical model shown in Figure 1, the parameters of Calib-Net and Δωt can be trained and optimized driven by the carefully designed loss function.

3.4 Loss Function for Regression

There are many losses that are widely used for solving the regression problem (Jadon, 2020). The losses include L1 Loss using Mean Square Error (MSE), L2 Loss using Mean Absolute Error (MAE), Huber Loss using Smooth Mean Absolute Error, etc. L1 Loss is robust to outliers. however, its gradient is always the same during the training and is still large even faced with small loss values. In this way, it is hard and inefficient for L1 Loss to find the minima at the end of training. On the contrary, L2 Loss is sensitive to outliers, but it is easier for L2 Loss to find a stable solution. Huber loss combines the advantages of L1 Loss and L2 Loss, it is more robust to outliers than L2 Loss and can decrease the gradient around the minima.

In our proposed Calib-Net, we choose to use the log hyperbolic cosine (log-cosh) loss for gyroscope measurement correction regression. Compared with the Huber Loss, it approximately equals (yy′)2/2 for the small loss and to yylog(2) for the large loss, so it has all the advantages of Huber Loss and is twice differentiable for both small loss and large loss. The loss function is defined as below:

Ly,y=i=1nlogcoshyy(7)

Where y is the label (ground truth) of orientation in our case, and y′ is the estimated orientation using the calibrated gyroscope value (namely angular velocity).

We also use multi-scale orientation loss to achieve better calibration performance. The orientation is estimated through the integration of calibrated angular velocity. So we adopt different time intervals to compute the loss and perform the back-propagation driven by the loss. The loss used for network training is shown below:

L=i=1nLΔt+i=1n/2L2Δt+i=1n/4L4Δt++LnΔt(8)

Where n ⋅Δt is the maximum time interval used for computing the loss, and its value is n = 2x. By carefully selecting different Δt and n using the trial and error method, we can adopt the suitable rotation transformation combination to train the calibration parameters.

4 Experimental Evaluation

In this section, we evaluate the performance of our proposed Calib-Net using the public EuROC dataset Burri et al. (2016). Both quantitive and qualitative experiments are performed for comparison. We first compare the orientation estimation performance of our Calib-Net with different deep learning methods. Then we replace the gyroscope measurements used for the OpenVINS system with the calibrated angular velocity from our Calib-Net, and compare the performance with state-of-the-art VIO methods.

For the training of our Calib-Net, most computers with a normally configured GPU are enough. In our case, we use a desktop equipped with an Intel i7-8700K CPU with 3.7-GHz and an Nvidia GeForce GTX 1060 GPU with 6 GB memory. The framework of Calib-Net is implemented based on PyTorch. We use the ADAM optimizer for the network training, the learning rate is set to 0.01, and the weight decay is 0.1. The consine anneal warm restart scheduler is adopted for network learning. The weight parameter for log-cosh loss is 1e6. The IMU runs at 200 Hz, so the ΔT is 5 ms in our case. For loss computing, we set Δt to 80 ms and n to 2 (as illustrated in Eq. 8) which means we use 16 and 32 IMU readings for integration. We training the network for 1,200 epochs and it only takes about 8.5min with the Nvidia GTX 1060 GPU which can be reached easily. For the test procedure, our proposed lightweight Calib-Net can easily reach real-time performance for calibration and orientation estimation (the IMU reading is 200 Hz).

4.1 Calibration Performance Evaluation Among Learning-Based Methods

We first evaluate the proposed Calib-Net by comparing the orientation estimation performance with other learning-based methods. Absolute Orientation Error (AOE) is adopted as the evaluation metric. We choose the evaluation tool (Zhang and Scaramuzza, 2018) to compute the metric. Except for AOE, we also adopt the yaw error as the metric as achieving the accurate yaw angle estimation is most difficult for IMU.

The public EuROC dataset (Burri et al., 2016) is used for the evaluation, and the uncalibrated ADIS16448 IMU is adopted in the dataset. We take MH_01_easy, MH_03_medium, MH_05_difficult, V1_02_medium, V2_01_easy, V2_03_difficult as the training sequences, and take MH_02_easy, MH_04_difficult, V1_01_easy, V1_03_difficult, V2_02_medium as the testing sequences. As shown in Table 1, our proposed Calib-Net outperforms the raw IMU data in terms of the orientation estimation and yaw estimation. When comparing with the state-of-the-art learning-based calibration methods, our framework defeats both OriNet (Esfahani et al., 2019b) and GyroNet (Brossard et al., 2020b) on most sequences. What is more, OriNet (Esfahani et al., 2019b) adopts the Long Short-Term Memory (LSTM) component for spatio-temporal feature extraction and needs larger GPU memory. Our framework adopts dilation convolutions and proper loss function which construct an easy (less hyperparameters) but efficient (better feature learning) neural network.

TABLE 1
www.frontiersin.org

TABLE 1. Orientation estimation results of different learning-based methods. The networks shown in the table are all trained using same sequences of the EuROC dataset (Burri et al., 2016). Both 3D orientation and yaw estimation results are given in the table. The best results are made in bold.

We also plot the estimated orientation in Figure 4 to give a more compact demonstration of our system’s performance. The blue line indicates the ground truth of the orientation (roll, yaw, pitch), the red line indicates the estimated orientation with our Calib-Net, the green line indicates the estimated orientation with the GyroNet, and the yellow line indicates estimated orientation from the raw IMU data. All orientations here are got only through the integration of the angular velocity. As shown in the figure, when using the raw IMU data to perform orientation estimation, the errors will be accumulated and the integrated orientation will drift in a short time due to inaccurate calibration. Our Calib-Net (red line) is most close to the ground truth and achieves the best performance among these methods.

FIGURE 4
www.frontiersin.org

FIGURE 4. Plots of estimated orientations with different methods. (A) Estimated orientation for MH_04_difficult with different methods. (B) Estimated orientation for V1_03_difficult with different learning-based methods.

4.2 Calibration Performance Evaluation Using VIO Methods

In order to further prove the effectiveness of our proposed framework, we also introduce the calibrated angular velocity from our Calib-Net into a well-known VIO system (Open-VINS) which achieves state-of-the-art performance recently. The new VIO system combined with our Calib-Net is called OpenVINS*. We present the quantitative orientation estimation results in Table 2. As shown in the table, OpenVINS* performs better than OpenVINS in most cases, and achieves the best mean results in terms of orientation and yaw estimation. We also plot the trajectories with these methods in Figure 5. The OpenVINS* with carefully calibrated angular velocity achieves better performance in trajectory estimation. This proves that our proposed Calib-Net can accurately correct the angular velocity measurements and the calibrated angular velocity is fruitful for VIO systems.

TABLE 2
www.frontiersin.org

TABLE 2. Orientation estimation results of different VIO methods. The Open-VINS* method takes the calibrated gyroscope data (produced by our proposed Calib-Net) as the input. Both 3D orientation and yaw estimation results are given in the table. The best results are made in bold.

FIGURE 5
www.frontiersin.org

FIGURE 5. Plots of estimated trajectories with different VIO methods. (A) Estimated trajectories for sequence V1_01_easy. (B) Estimated trajectories for sequence V2_02_medium. Open-VINS* takes the corrected IMU measurements from our proposed Calib-Net.

When comparing the estimated orientation of Calib-Net (shown in table I) with that of VIO methods (shown in table II), we can find that our Calib-Net can even compete with VIO methods in terms of orientation estimation. What is more, the average yaw error of our Calib-Net is 0.87°, and it performs better than all VIO methods which include Open-VINS* with 1.11°, Open-VINS with 1.37°, VINS-Mono with 2.14°. This further indicates that a carefully calibrated low-cost IMU can achieve similar (even better in some aspects) performance when comparing with visual-inertial methods.

5 Conclusion and Future Work

In this paper, we present a light-weight deep convolutional neural network for low-cost IMU calibration which is called Calib-Net. A mathematical calibration model is introduced to design the training and calibration framework. Dilation convolution is adopted for spatio-temporal feature extraction of IMU measurements. Driven by a carefully designed loss function, the Calib-Net can be optimized to output the compensation for raw gyroscope measurements. Corresponding experimental evaluations are performed to prove the effectiveness of our proposed Calib-Net. The results show that our framework achieves quite good calibration performance and the orientation estimation performance of our Calib-Net can even compete with state-of-the-art VIO methods. However, the generalization ability of the proposed network to different types of IMU is still challenging, and we would like to try more datasets and study more about it. What is more, we plan to use the deep neural network to perform odometry to dead-reckon both translation and rotation. We also consider introducing deep learning technology for visual-inertial odometry.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://projects.asl.ethz.ch/datasets/.

Author Contributions

RL and WY propose the idea, RL and CF design and implement the experiment. All authors contribute to the writing and data analysis. All authors agree to be accountable for the content of the work.

Funding

This work was supported by Science and Technology Innovation 2030 Major Project under Grant No. 2020AAA0104802. It was also supported by the National Natural Science Foundation of China under Grant No. 61903377 and Grant No. 91948303.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

The authors would like to thank Martin Brossard for his kindly help and insightful discussion.

References

Ahmad, N., Ghazilla, R. A. R., Khairi, N. M., and Kasi, V. (2013). Reviews on Various Inertial Measurement Unit (IMU) Sensor Applications. Ijsps 1, 256–262. doi:10.12720/ijsps.1.2.256-262

CrossRef Full Text | Google Scholar

Barrau, A., and Bonnabel, S. (2020). “A Mathematical Framework for IMU Error Propagation with Applications to Preintegration,” in 2020 IEEE International Conference on Robotics and Automation (ICRA) (IEEE), Paris, France, May 31–August 31, 2020, 5732–5738. doi:10.1109/icra40945.2020.9197492

CrossRef Full Text | Google Scholar

Brossard, M., Barrau, A., and Bonnabel, S. (2020a). AI-IMU Dead-Reckoning. IEEE Trans. Intell. Veh. 5, 585–595. doi:10.1109/tiv.2020.2980758

CrossRef Full Text | Google Scholar

Brossard, M., Bonnabel, S., and Barrau, A. (2020b). Denoising IMU Gyroscopes with Deep Learning for Open-Loop Attitude Estimation. IEEE Robotics Automation Lett. 5, 4796–4803. doi:10.1109/lra.2020.3003256

CrossRef Full Text | Google Scholar

Burri, M., Nikolic, J., Gohl, P., Schneider, T., Rehder, J., Omari, S., et al. (2016). The EuRoC Micro Aerial Vehicle Datasets. Int. J. Robotics Res. 35, 1157–1163. doi:10.1177/0278364915620033

CrossRef Full Text | Google Scholar

Campos, C., Elvira, R., Rodríguez, J. J. G., Montiel, J. M., and Tardós, J. D. (2021). ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM. IEEE Trans. Robotics.

Google Scholar

Campos, C., Montiel, J. M., and Tardós, J. D. (2020). “Inertial-only Optimization for Visual-Inertial Initialization,” in 2020 IEEE International Conference on Robotics and Automation (ICRA) (IEEE), Paris, France, May 31–August 31, 2020, 51–57. doi:10.1109/icra40945.2020.9197334

CrossRef Full Text | Google Scholar

Chen, C., Lu, X., Markham, A., and Trigoni, N. (2018). “IONet: Learning to Cure the Curse of Drift in Inertial Odometry,” in Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, Unites States, February 2–7, 2018.

Google Scholar

Chen, C., Lu, X., Wahlstrom, J., Markham, A., and Trigoni, N. (2019a). Deep Neural Network Based Inertial Odometry Using Low-Cost Inertial Measurement Units. IEEE Trans. Mobile Comput.

Google Scholar

Chen, C., Rosa, S., Miao, Y., Lu, C. X., Wu, W., Markham, A., et al. (2019b). “Selective Sensor Fusion for Neural Visual-Inertial Odometry,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, United States, June 16–20, 2019, 10542–10551. doi:10.1109/cvpr.2019.01079

CrossRef Full Text | Google Scholar

Clark, R., Wang, S., Wen, H., Markham, A., and Trigoni, N. (2017). “VINet: Visual-Inertial Odometry as a Sequence-To-Sequence Learning Problem,” in Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, United States, February 4–9, 2017.

Google Scholar

Esfahani, M. A., Wang, H., Wu, K., and Yuan, S. (2019a). AbolDeepIO: A Novel Deep Inertial Odometry Network for Autonomous Vehicles. IEEE Trans. Intell. Transportation Syst. 21, 1941–1950. doi:10.1109/LRA.2019.2959507

CrossRef Full Text | Google Scholar

Esfahani, M. A., Wang, H., Wu, K., and Yuan, S. (2019b). OriNet: Robust 3-D Orientation Estimation with a Single Particular IMU. IEEE Robotics Automation Lett. 5, 399–406.

Google Scholar

Furgale, P., Rehder, J., and Siegwart, R. (2013). “Unified Temporal and Spatial Calibration for Multi-Sensor Systems,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, November 3–7, 2013 (IEEE), 1280. doi:10.1109/iros.2013.6696514

CrossRef Full Text | Google Scholar

Geneva, P., Eckenhoff, K., Lee, W., Yang, Y., and Huang, G. (2020). “OpenVINS: A Research Platform for Visual-Inertial Estimation,” in 2020 IEEE International Conference on Robotics and Automation (ICRA) (IEEE), Paris, France, May 31–August 31, 2020, 4666–4672. doi:10.1109/icra40945.2020.9196524

CrossRef Full Text | Google Scholar

Jadon, S. (2020). “A Survey of Loss Functions for Semantic Segmentation,” in 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB) (IEEE), Via del Mar, Chile, October 27–29, 2020. doi:10.1109/cibcb48159.2020.9277638

CrossRef Full Text | Google Scholar

Nebot, E., and Durrant-Whyte, H. (1999). Initial Calibration and Alignment of Low-Cost Inertial Navigation Units for Land Vehicle Applications. J. Robotic Syst. 16, 81–92. doi:10.1002/(sici)1097-4563(199902)16:2<81:aid-rob2>3.0.co;2-9

CrossRef Full Text | Google Scholar

Nobre, F., and Heckman, C. (2019). Learning to Calibrate: Reinforcement Learning for Guided Calibration of Visual-Inertial Rigs. Int. J. Robotics Res. 38, 1388–1402. doi:10.1177/0278364919844824

CrossRef Full Text | Google Scholar

Qin, T., Li, P., and Shen, S. (2018). VINS-mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 34, 1004–1020. doi:10.1109/tro.2018.2853729

CrossRef Full Text | Google Scholar

Qin, T., and Shen, S. (2018). “Online Temporal Calibration for Monocular Visual-Inertial Systems,” in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE), Madrid, Spain, October 1–5, 2018, 3662–3669. doi:10.1109/iros.2018.8593603

CrossRef Full Text | Google Scholar

Rehder, J., Nikolic, J., Schneider, T., Hinzmann, T., and Siegwart, R. (2016). “Extending Kalibr: Calibrating the Extrinsics of Multiple IMUs and of Individual Axes,” in IEEE International Conference on Robotics and Automation (ICRA) (IEEE), Stockholm, Sweden, May 16–21, 2016, 4304–4311. doi:10.1109/icra.2016.7487628

CrossRef Full Text | Google Scholar

Rehder, J., and Siegwart, R. (2017). Camera/IMU Calibration Revisited. IEEE Sensors J. 17, 3257–3268. doi:10.1109/jsen.2017.2674307

CrossRef Full Text | Google Scholar

Rohac, J., Sipos, M., and Simanek, J. (2015). Calibration of Low-Cost Triaxial Inertial Sensors. IEEE Instrum. Meas. Mag. 18, 32–38. doi:10.1109/mim.2015.7335836

CrossRef Full Text | Google Scholar

Tedaldi, D., Pretto, A., and Menegatti, E. (2014). “A Robust and Easy to Implement Method for IMU Calibration without External Equipments,” in IEEE International Conference on Robotics and Automation (ICRA) (IEEE), Hong Kong, China, May 31–June 7, 2014, 3042–3049. doi:10.1109/icra.2014.6907297

CrossRef Full Text | Google Scholar

Yan, H., Shan, Q., and Furukawa, Y. (2018). “RIDI: Robust IMU Double Integration,” in Proceedings of the European Conference on Computer Vision, Munich, Germany, September 8–14, 2018 (ECCV)), 621–636. doi:10.1007/978-3-030-01261-8_38

CrossRef Full Text | Google Scholar

Yang, Y., Geneva, P., Zuo, X., and Huang, G. (2020). Online Imu Intrinsic Calibration: Is it Necessary. in” Proc. Of the Robotics: Science and Systems. Corvallis, Oregon, 716–720.

CrossRef Full Text | Google Scholar

Zhang, Z., and Scaramuzza, D. (20182018). “A Tutorial on Quantitative Trajectory Evaluation for Visual (-inertial) Odometry,” in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, October 1–5, 2018 (IEEE), 7244–7251. doi:10.1109/iros.2018.8593941

CrossRef Full Text | Google Scholar

Keywords: IMU calibration, deep neural network, orientation estimation, spatio-temporal, visual inertial odometry

Citation: Li  R, Fu C, Yi  W and Yi  X (2022) Calib-Net: Calibrating the Low-Cost IMU via Deep Convolutional Neural Network. Front. Robot. AI 8:772583. doi: 10.3389/frobt.2021.772583

Received: 08 September 2021; Accepted: 15 November 2021;
Published: 03 January 2022.

Edited by:

Xuebo Zhang, Nankai University, China

Reviewed by:

Zuyuan Zhu, University of Essex, United Kingdom
Yan Zhuang, Dalian University of Technology, China
Xunyu Zhong, Xiamen University, China

Copyright © 2022 Li , Fu, Yi  and Yi . This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Wei Yi , yi_wei_cs@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.