Preview

Giroskopiya i Navigatsiya

Advanced search

x-RIO: Radar Inertial Odometry with Multiple Radar Sensors and Yaw Aiding

https://doi.org/10.17285/0869-7035.0080

Abstract

A robust and accurate real-time navigation system is crucial for autonomous robotics. In particular, GNSS denied and poor visual conditions are still very challenging as vision based approaches tend to fail in darkness, direct sunlight, fog or smoke. Therefore, we are taking advantage of inertial data and FMCW radar sensors as both are not affected by such conditions. In this work, we propose a framework, which uses several 4D mm Wave radar sensors simultaneously. The extrinsic calibration of each radar sensor is estimated online. Based on a single radar scan, the 3D ego velocity and optionally yaw measurements based on Manhattan world assumptions are fused. An extensive evaluation with real world datasets is presented. We achieve even better accuracies than state of the art stereo Visual Inertial Odometry (VIO) while being able to cope with degraded visual conditions and requiring only very little computational resources.

About the Authors

Ch. Doer
Institute of Control Systems, Karlsruhe Institute of Technology (KIT), Karlsruhe, Germany
Germany


G.F. Trommer
Institute of Control Systems, Karlsruhe Institute of Technology (KIT), Karlsruhe, Germany
Germany


References

1. Bloesch, M., Omari, S., Hutter, M., and Siegwart, R., Robust visual inertial odometry using a direct EKF-based approach, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2015, pp. 298–304.

2. Sun, K., Mohta, K., Pfrommer, B., Watterson, M., Liu, S., Mulgaonkar, Y., Taylor, C.J., and Kumar, V., Robust stereo visual inertial odometry for fast autonomous flight, IEEE Robotics and Automation Letters, 2018, vol. 3, no. 2, pp. 965–972.

3. Qin, T., Li, P., and Shen, S., VINS-Mono: A robust and versatile monocular visual-inertial state estimator, IEEE Transactions on Robotics, 2018, vol. 34, no. 4, pp. 1004–1020.

4. Khattak, S., Papachristos, C., and Alexis, K., Keyframe-based thermal-inertial odometry, Journal of Field Robotics, 2020, vol. 37, no. 4, pp. 552–579.

5. Saputra, M.R.U., de Gusmao, P.P.B., Lu, C.X., Almalioglu, Y., Rosa, S., Chen, C., Wahlstroem, J., Wang, W., Markham, A., and Trigoni, N., DeepTIO: A deep thermal-inertial odometry with visual hallucination, IEEE Robotics and Automation Letters, 2020, vol. 5, no. 2.

6. Zhao, S., Wang, P., Zhang, H., Fang, Z., and Scherer, S., TP-TIO: A robust thermal-inertial odometry with deep thermalpoint, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

7. Dickmann, J., Klappstein, J., Hahn, M., Appenrodt, N., Bloecher, H., Werber, K., and Sailer, A., Automotive radar – The key technology for autonomous driving: From detection and ranging to environmental understanding, 2016 IEEE Radar Conference (RadarConf).

8. Doer, C. and Trommer, G.F., Radar inertial odometry with online calibration, 2020 European Navigation Conference (ENC), 2020, pp. 1–10.

9. Doer, C. and Trommer, G.F., Yaw aided radar inertial odometry uisng Manhattan world assumptions, 28th St. Petersburg International Conference on Integrated Navigation Systems (ICINS), 2021.

10. Kramer, A., Stahoviak, C., Santamaria-Navarro, A., Aghamohammadi, A.-A., and Heckman, C., Radar-inertial ego-velocity estimation for visually degraded environments, IEEE International Conference on Robotics and Automation, 2020.

11. Doer, C. and Trommer, G.F., An EKF-based approach to radar inertial odometry, IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI), 2020, pp. 152–159.

12. Lu, C.X., Saputra, M.R.U., Zhao, P., Almalioglu, Y., de Gusmao, P.P.B., Chen, C., Sun, K., Trigoni, N., and Markham, A., Milliego: Single-chip mmWave radar aided egomotion estimation via deep sensor fusion, Proceedings of the 18th Conference on Embedded Networked Sensor Systems, ser. SenSys ’20, 2020.

13. Quist, E.B. and Beard, R.W., Radar odometry on fixed-wing small unmanned aircraft, IEEE Transactions on Aerospace and Electronic Systems, 2016, vol. 52, no. 1, pp. 396–410.

14. Cen, S.H. and Newman, P., Precise ego-motion estimation with millimeter-wave radar under diverse and challenging conditions, IEEE International Conference on Robotics and Automation (ICRA), IEEE, 2018, pp. 1–8. 15. Aldera, R., De Martini, D., Gadd, M., and Newman, P., Fast radar motion estimation with a learnt focus of attention using weak supervision, 2019 International Conference on Robotics and Automation (ICRA), 2019, pp. 1190–1196.

15. Park, Y. S., Shin, Y. S., and Kim, A., Pharao: Direct radar odometry using phase correlation, 2020 IEEE International Conference on Robotics and Automation (ICRA), 2020, pp. 2617–2623.

16. Almalioglu, Y., Turan, M., Lu, C.X., Trigoni, N., and Markham, A., Milli-RIO: Ego-motion estimation with low-cost millimetre-wave radar, IEEE Sensors Journal, 2021, vol. 21, no. 3, pp. 3314–3323. 18. Rapp, M., Barjenbruch, M., Dietmayer, K., Hahn, M., and Dickmann, J., A fast probabilistic ego-motion estimation framework for radar, 2015 European Conference on Mobile Robots (ECMR), 2015, pp. 1–6. 19. Kellner, D., Barjenbruch, M., Klappstein, J., Dickmann, J., and Dietmayer, K., Instantaneous ego-motion estimation using Doppler radar, Conference on Intelligent Transportation Systems (ITSC 2013), IEEE, 2013, pp. 869–874.

17. Doer, C. and Trommer, G.F., Radar visual inertial odometry and radar thermal inertial odometry: Robust navigation even in challenging visual conditions, 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2021.

18. Sola, J., Quaternion kinematics for the error-state KF, 2017.

19. Roumeliotis, S.I. and Burdick, J.W., Stochastic cloning: A generalized framework for processing relative state measurements, Proc. 2002 IEEE International Conference on Robotics and Automation, 2002.

20. Fischler, M.A. and Bolles, R.C., Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography, Communications of the ACM, 1981.

21. Ghilani, C.D., Adjustment Computations: Spatial Data Analysis, Hoboken, New Jersey: John Wiley & Sons, 2017.

22. Delmerico, J. and Scaramuzza, D., A benchmark comparison of monocular visual-inertial odometry algorithms for flying robots, 2018 IEEE International Conference on Robotics and Automation (ICRA), 2018, pp. 2502–2509.

23. Umeyama, S., Least-squares estimation of transformation parameters between two point patterns, IEEE Transactions on Pattern Analysis and Machine Intelligence, 1991, vol. 13, no. 4, pp. 376–380.

24. Zhang, Z. and Scaramuzza, D., A tutorial on quantitative trajectory evaluation for visual(-inertial) odometry, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2018.

25. Guennebaud, G., Jacob, B. et al., Eigen v3, http://eigen.tuxfamily.org, 2010.


Review

For citations:


Doer Ch., Trommer G. x-RIO: Radar Inertial Odometry with Multiple Radar Sensors and Yaw Aiding. Giroskopiya i Navigatsiya. 2021;29(4):78-96. (In Russ.) https://doi.org/10.17285/0869-7035.0080

Views: 3


Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.


ISSN 0869-7035 (Print)
ISSN 2075-0927 (Online)