[1] G. K. Tam, Z.-Q. Cheng, Y.-K. Lai, F. C. Langbein, Y. Liu, D. Marshall, R. R. Martin, X.-F. Sun, and P. L. Rosin, “Registration of 3d point clouds and meshes: A survey from rigid to nonrigid,” IEEE transactions on visualization and computer graphics, vol. 19, no. 7, pp. 1199–1217, 2012.
[2] R. Mur-Artal, J. M. M. Montiel, and J. D. Tardos, “Orb-slam: a versatile and accurate monocular slam system,” IEEE transactions on robotics, vol. 31, no. 5, pp. 1147–1163, 2015.
[3] R. Mur-Artal and J. D. Tard ´os, “Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras,” IEEE Transactions on Robotics, vol. 33, no. 5, pp. 1255–1262, 2017.
[4] D. Scaramuzza and F. Fraundorfer, “Visual odometry [tutorial],” IEEE robotics & automation magazine, vol. 18, no. 4, pp. 80–92, 2011.
[5] K. R. Konda and R. Memisevic, “Learning visual odometry with a convolutional network.” in VISAPP (1), 2015, pp. 486–490.
[6] S. Wang, R. Clark, H. Wen, and N. Trigoni, “Deepvo: Towards end-to-end visual odometry with deep recurrent convolutional neural networks,” in 2017 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2017, pp. 2043–2050.
[7] R. Clark, S. Wang, H. Wen, A. Markham, and N. Trigoni, “Vinet: Visual-inertial odometry as a sequence-to-sequence learning problem,” in Thirty-First AAAI Conference on Artificial Intelligence, 2017.
[8] N. Yang, R. Wang, J. Stuckler, and D. Cremers, “Deep virtual stereo odometry: Leveraging deep depth prediction for monocular direct sparse odometry,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 817–833.
[9] A. Geiger, P. Lenz, C. Stiller, and R. Urtasun, “Vision meets robotics: The kitti dataset,” The International Journal of Robotics Research, vol. 32, no. 11, pp. 1231–1237, 2013.
[10] A. Segal, D. Haehnel, and S. Thrun, “Generalized-icp.” in Robotics: science and systems, vol. 2, no. 4. Seattle, WA, 2009, p. 435.
[11] E. Mendes, P. Koch, and S. Lacroix, “Icp-based pose-graph slam,” 2016 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), pp. 195–200, 2016.
[12] K. Yoneda, H. T. Niknejad, T. Ogawa, N. Hukuyama, and S. Mita, “Lidar scan feature for localization with highly precise 3-d map,” 2014 IEEE Intelligent Vehicles Symposium Proceedings, pp. 1345– 1350, 2014.
[13] A. Nicolai, R. Skeele, C. Eriksen, and G. A. Hollinger, “Deep learning for laser based odometry estimation.”
[14] M. Magnusson, A. Lilienthal, and T. Duckett, “Scan registration for autonomous mining vehicles using 3d-ndt,” Journal of Field Robotics, vol. 24, no. 10, pp. 803–827, 2007.
[15] B. Zhou, Z. Tang, K. Qian, F. Fang, and X. Ma, “A lidar odometry for outdoor mobile robots using ndt based scan matching in gps-denied environments,” in 2017 IEEE 7th Annual International Conference on CYBER Technology in Automation, Control, and Intelligent Systems (CYBER). IEEE, 2017, pp. 1230–1235.
[16] P. Egger, P. V. Borges, G. Catt, A. Pfrunder, R. Siegwart, and R. Dub´e, “Posemap: Lifelong, multi-environment 3d lidar localization,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2018, pp. 3430–3437.
[17] W. Wang, M. R. U. Saputra, P. Zhao, P. Gusmao, B. Yang, C. Chen, A. Markham, and N. Trigoni, “Deeppco: End-to-end point cloud odometry through deep parallel neural network,” arXiv preprint arXiv:1910.11088, 2019.
[18] Y. Cho, G. Kim, and A. Kim, “Deeplo: Geometry-aware deep lidar odometry,” arXiv preprint arXiv:1902.10562, 2019.
[19] K. Ji, H. Chen, H. Di, J. Gong, G. Xiong, J. Qi, and T. Yi, “Cpfgslam:a robust simultaneous localization and mapping based on lidar in off-road environment,” in 2018 IEEE Intelligent Vehicles Symposium (IV), June 2018, pp. 650–655.
[20] J. Deschaud, “IMLS-SLAM: scan-to-model matching based on 3d data,” CoRR, vol. abs/1802.08633, 2018. [Online]. Available: http://arxiv.org/abs/1802.08633
[21] K. Pathak, A. Birk, N. Vaskevicius, M. Pfingsthorn, S. Schwertfeger, and J. Poppinga, “Online three-dimensional slam by registration of large planar surface segments and closed-form pose-graph relaxation,” Journal of Field Robotics, vol. 27, no. 1, pp. 52–84, 2010.
[22] W. S. Grant, R. C. Voorhies, and L. Itti, “Finding planes in lidar point clouds for real-time registration,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2013, pp. 4347– 4354. [23] H. Yin, L. Tang, X. Ding, Y. Wang, and R. Xiong, “Locnet: Global localization in 3d point clouds for mobile vehicles,” in 2018 IEEE Intelligent Vehicles Symposium (IV). IEEE, 2018, pp. 728–733.
[24] J. Behley and C. Stachniss, “Efficient surfel-based slam using 3d laser range data in urban environments.”
[25] J. Zhang and S. Singh, “Loam: Lidar odometry and mapping in realtime.”
[26] T. Shan and B. Englot, “Lego-loam: Lightweight and groundoptimized lidar odometry and mapping on variable terrain,” in 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2018, pp. 4758–4765.
[27] X. Ji, L. Zuo, C. Zhang, and Y. Liu, “Lloam: Lidar odometry and mapping with loop-closure detection based correction,” in 2019 IEEE International Conference on Mechatronics and Automation (ICMA), Aug 2019, pp. 2475–2480.
[28] J. Lin and F. Zhang, “A fast, complete, point cloud based loop closure for lidar odometry and mapping,” 09 2019.
[29] G. Elbaz, T. Avraham, and A. Fischer, “3d point cloud registration for localization using a deep neural network auto-encoder,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017, pp. 2472–2481.
[30] X. Chen, A. M. E. Palazzolo, P. Gigu`ere, J. Behley, and C. Stachniss, “Suma + + : Efficient lidar-based semantic slam,” 2019.
[31] A. Cramariuc, R. Dub´e, H. Sommer, R. Siegwart, and I. Gilitschenski, “Learning 3d segment descriptors for place recognition,” arXiv preprint arXiv:1804.09270, 2018.
[32] R. Dube, D. Dugas, E. Stumm, J. Nieto, R. Siegwart, and C. Cadena, “Segmatch: Segment based place recognition in 3d point clouds,” 05 2017, pp. 5266–5272.
[33] R. Dub´e, M. G. Gollub, H. Sommer, I. Gilitschenski, R. Siegwart, C. Cadena, and J. Nieto, “Incremental-segment-based localization in 3-d point clouds,” IEEE Robotics and Automation Letters, vol. 3, no. 3, pp. 1832–1839, 2018.
[34] R. Dub, A. Gawel, H. Sommer, J. Nieto, R. Siegwart, and C. Cadena, “An online multi-robot slam system for 3d lidars,” in 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Sep. 2017, pp. 1004–1011.
[35] R. Dub´e, A. Cramariuc, D. Dugas, J. Nieto, R. Siegwart, and C. Cadena, “Segmap: 3d segment mapping using data-driven descriptors,” arXiv preprint arXiv:1804.09557, 2018.
[36] C. Park, S. Kim, P. Moghadam, C. Fookes, and S. Sridharan, “Probabilistic surfel fusion for dense lidar mapping,” 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), pp. 2418–2426, 2017.
[37] Q. Li, S. Chen, C. Wang, X. Li, C. Wen, M. Cheng, and J. Li, “Lo-net: Deep real-time lidar odometry,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 8473–8482. [38] M. Velas, M. Spanel, and A. Herout, “Collar line segments for fast odometry estimation from velodyne point clouds,” 2016 IEEE International Conference on Robotics and Automation (ICRA), pp. 4486–4495, 2016.
[39] T. Tang, D. Yoon, F. Pomerleau, and T. D. Barfoot, “Learning a bias correction for lidar-only motion estimation,” in 2018 15th Conference on Computer and Robot Vision (CRV). IEEE, 2018, pp. 166–173.
[40] L. Sun, J. Zhao, X. He, and C. Ye, “Dlo: Direct lidar odometry for 2.5 d outdoor environment,” in 2018 IEEE Intelligent Vehicles Symposium (IV). IEEE, 2018, pp. 1–5.