Augmented Reality Pose Estimation: A Systematic Review of Visual Localization and SLAM

Zhiqian Zhang, Hai Huang, Tong Wu, Wenpeng Huang, Zhenghan Zhong

Article ID: 8394
Vol 7, Issue 1, 2026
DOI: https://doi.org/10.54517/m8394

Download PDF

Abstract

Augmented Reality (AR) has attracted increasing attention as it enhances user perception and interaction by overlaying virtual content onto the physical world. Accurate and real-time six degrees of freedom (6-DoF) pose estimation is a core requirement for reliable spatial registration. To help researchers understand recent advances and select appropriate methods for AR applications, this survey provides a systematic overview of AR pose estimation algorithms from two complementary perspectives: visual localization with prior scene knowledge and simultaneous localization and mapping (SLAM), which performs online pose estimation and mapping in unknow environments. For visual localization, pose estimation methods can be grouped into three major lines, including feature-matching-based localization, scene coordinate regression, and pose regression; their evolution toward improved robustness and scalability is also discussed. For SLAM-based pose estimation, representative approaches are organized into traditional visual SLAM (VSLAM), deep learning–enhanced SLAM, rendering-based SLAM, highlighting key design choices as well as the strengths and limitations of each category under AR constraints. In addition, evaluation metrics and commonly used benchmarks are reviewed, and reported performance of representative algorithms on selected datasets is summarized. Finally, the review summarizes the current state of AR pose estimation and outlines future research directions.


Keywords

Augmented Reality; Pose Estimation; Visual Localization; Visual Simultaneous Localization and Mapping; Absolute Pose Error; Relative Pose Error


References

1. Li G, Luo H, Chen D, Wang P, Yin X, Zhang J. Augmented Reality in Higher Education: A Systematic Review and Meta-Analysis of the Literature from 2000 to 2023. Education Sciences. 2025; 15(6):678. https://doi.org/10.3390/educsci15060678

2. Wei, W., Gao, J., Luo, X. et al. A scoping review of the applications of augmented reality in nursing. BMC Nurs 24, 1105 (2025). https://doi.org/10.1186/s12912-025-03750-1

3. Lin, L., Wu, Z., Lu, Y., Chen, Z., & Guo, W. Lightweight deep learning with multi-scale feature fusion for high-precision and low-latency eye tracking[J]. Displays, 2026, 91: 103260. https://doi.org/10.1016/j.displa.2025.103260

4. Syed TA, Siddiqui MS, Abdullah HB, Jan S, Namoun A, Alzahrani A, Nadeem A, Alkhodre AB. In-Depth Review of Augmented Reality: Tracking Technologies, Development Tools, AR Displays, Collaborative AR, and Security Concerns. Sensors. 2023; 23(1):146. https://doi.org/10.3390/s23010146

5. Sheng X , Mao S , Yan Y ,et al.Review on SLAM algorithms for Augmented Reality[J].Displays, 2024, 84:102806. https://doi.org/10.1016/j.displa.2024.102806

6. Xu M , Wang Y ,Bin XuJun ZhangJian RenZhao HuangStefan PosladPengfei Xu.A critical analysis of image-based camera pose estimation techniques[J].Neurocomputing, 2024, 570(Feb.14):127125.1-127125.27. https://doi.org/10.1016/j.neucom.2023.127125

7. Chandio Y , Selialia K , Degol J ,et al.Lost in Tracking Translation: A Comprehensive Analysis of Visual SLAM in Human-Centered XR and IoT Ecosystems[J]. 2024. https://doi.org/10.48550/arXiv.2411.07146

8. Torsten Sattler, Bastian Leibe, and Leif Kobbelt. Efficient &effective prioritized matching for large-scale image-based localization. IEEE transactions on pattern analysis and ma chine intelligence, 39(9):1744–1756, 2016. https://doi.org/10.1109/TPAMI.2016.2611662

9. Hajime Taira, Masatoshi Okutomi, Torsten Sattler, Mircea Cimpoi, Marc Pollefeys, Josef Sivic, Tomas Pajdla, and Ak ihiko Torii. Inloc: Indoor visual localization with dense matching and view synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 7199–7209, 2018. https://doi.org/10.1109/CVPR.2018.00752

10. R. Arandjelović, P. Gronat, A. Torii, T. Pajdla and J. Sivic, "NetVLAD: CNN Architecture for Weakly Supervised Place Recognition," in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 6, pp. 1437-1451, 1 June 2018, doi: 10.1109/TPAMI.2017.2711011. https://doi.org/10.1109/TPAMI.2017.2711011

11. Sarlin, Paul-Edouard et al. “From Coarse to Fine: Robust Hierarchical Localization at Large Scale.” 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2018): 12708-12717. https://doi.org/10.1109/CVPR.2019.01300  

12. Paul-Edouard Sarlin, Daniel DeTone, Tomasz Malisiewicz, and Andrew Rabinovich. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4938–4947, 2020. https://doi.org/10.1109/CVPR42600.2020.00499

13. Paul-Edouard Sarlin, Ajaykumar Unagar, M˚ans Larsson, Hugo Germain, Carl Toft, Victor Larsson, Marc Pollefeys, Vincent Lepetit, Lars Hammarstrand, Fredrik Kahl, and Torsten Sattler. Back to the Feature: Learning Robust Cam era Localization from Pixels to Pose. In CVPR, 2021. https://doi.org/10.48550/arXiv.2103.09213

14. Fei Xue and Ignas Budvytis and Roberto Cipolla. PRAM: Place Recognition Anywhere Model for Efficient Visual Localization. arXiv preprint arXiv:2404.0778, 2024. https://doi.org/10.48550/arXiv.2404.07785

15. Siyan Dong, Shaohui Liu, Hengkai Guo, Baoquan Chen, and Marc Pollefeys. Lazy visual localization via motion averaging. arXiv preprint arXiv:2307.09981, 2023. https://doi.org/10.48550/arXiv.2307.09981

16. E. Brachmann et al., "DSAC — Differentiable RANSAC for Camera Localization," 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 2017, pp. 2492-2500. https://doi.org/10.1109/CVPR.2017.267

17. E. Brachmann and C. Rother, "Learning Less is More - 6D Camera Localization via 3D Surface Regression," 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 2018, pp. 4654-4662. https://doi.org/10.1109/CVPR.2018.00489

18. E. Brachmann and C. Rother, “Visual camera re-localization from rgb and rgb-d images using dsac,” TPAMI, vol. 44, no. 9, pp. 5847–5865, 2022. https://doi.org/10.1109/TPAMI.2021.3070754

19. X. Li, S. Wang, Y. Zhao, J. Verbeek, and J. Kannala, “Hierarchical scene coordinate classification and regression for visual localization,” in CVPR, 2020. https://doi.org/10.1109/CVPR42600.2020.01200

20. Shuzhe Wang, Zakaria Laskar, Iaroslav Melekhov, Xiaotian Li, Yi Zhao, Giorgos Tolias, and Juho Kannala. Hscnet++: Hierarchical scene coordinate classification and regression for visual localization with transformer. International Jour nal of Computer Vision, pages 1–21, 2024. https://doi.org/10.48550/arXiv.2305.03595

21. Eric Brachmann, Tommaso Cavallari, and Victor Adrian Prisacariu. Accelerated coordinate encoding: Learning to relocalize in minutes using rgb and poses. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pat tern Recognition, pages 5044–5053, 2023. https://doi.org/10.48550/arXiv.2305.14059

22. F. Wang, X. Jiang, S. Galliani, C. Vogel and M. Pollefeys, "GLACE: Global Local Accelerated Coordinate Encoding," 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 2024, pp. 5819-5828, doi: 10.1109/CVPR52733.2024.02037. https://doi.org/10.1109/CVPR52733.2024.02037

23. Jiang, Xudong et al. “R-SCoRe: Revisiting Scene Coordinate Regression for Robust Large-Scale Visual Localization.” 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2025): 11536-11546. https://doi.org/10.1109/CVPR52734.2025.01077

24. K. Xu, Z. Jiang, H. Cao, S. Yuan, C. Wang and L. Xie, "Enhancing Scene Coordinate Regression With Efficient Keypoint Detection and Sequential Information," in IEEE Robotics and Automation Letters, vol. 10, no. 10, pp. 9932-9939, Oct. 2025, doi: 10.1109/LRA.2025.3598670. https://doi.org/10.1109/LRA.2025.3598670

25. S. Tang, S. Tang, A. Tagliasacchi, P. Tan and Y. Furukawa, "NeuMap: Neural Coordinate Mapping by Auto-Transdecoder for Camera Localization," 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 2023, pp. 929-939, doi: 10.1109/CVPR52729.2023.00096. https://doi.org/10.1109/cvpr52729.2023.00096

26. J. Liu, Q. Nie, Y. Liu, and C. Wang, “Nerf-loc: Visual localization with conditional neural radiance field,” in ICRA, 2023. https://doi.org/10.48550/arXiv.2304.07979

27. B. Mildenhall, P.P. Srinivasan, M. Tancik, J.T. Barron, R. Ramamoorthi, R. Ng, Nerf: Representing scenes as neural radiance fields for view synthesis, Commun. ACM 65 (1) (2021) 99106. https://doi.org/10.48550/arXiv.2003.08934

28. A. Kendall, M. Grimes, R. Cipoll. “PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization.” 2015 IEEE International Conference on Computer Vision (ICCV) (2015): 2938-2946. https://doi.org/10.1109/ICCV.2015.336

29. Kendall, Alex and Roberto Cipolla. “Geometric Loss Functions for Camera Pose Regression with Deep Learning.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017): 6555-6564.  https://doi.org/10.1109/CVPR.2017.694

30. Wang, Bing et al. “AtLoc: Attention Guided Camera Localization.” ArXiv abs/1909.03557 (2019): n. pag. https://doi.org/10.1609/aaai.v34i06.6608

31. Balntas, Vassileios et al. “RelocNet: Continuous Metric Learning Relocalisation Using Neural Nets.” European Conference on Computer Vision (2018). https://doi.org/10.1007/978-3-030-01264-9_46

32. Soham Saha, Girish Varma, and CV Jawahar. Improved visual relocalization by discovering anchor points. arXiv preprint arXiv:1811.04370, 2018. https://doi.org/10.48550/arXiv.1811.04370

33. Ding, Mingyu et al. “CamNet: Coarse-to-Fine Retrieval for Camera Re-Localization.” 2019 IEEE/CVF International Conference on Computer Vision (ICCV) (2019): 2871-2880. https://doi.org/10.1109/ICCV.2019.00296

34. Dong, Siyan et al. “Reloc3r: Large-Scale Training of Relative Camera Pose Regression for Generalizable, Fast, and Accurate Visual Localization.” 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024): 16739-16752. https://doi.org/10.1109/CVPR52734.2025.01560

35. Sattler, Torsten et al. “Understanding the Limitations of CNN-Based Absolute Camera Pose Regression.” 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019): 3297-3307. https://doi.org/10.1109/CVPR.2019.00342

36. Chen, Changhao et al. “Deep Learning for Visual Localization and Mapping: A Survey.” IEEE Transactions on Neural Networks and Learning Systems 35 (2023): 17000-17020. https://doi.org/10.1109/tnnls.2023.3309809/mm1

37. Moreau, Arthur et al. “LENS: Localization enhanced by NeRF synthesis.” Conference on Robot Learning (2021). https://doi.org/10.48550/arXiv.2110.06558

38. Chen, Shuai et al. “DFNet: Enhance Absolute Pose Regression with Direct Feature Matching.” European Conference on Computer Vision (2022). https://doi.org/10.48550/arXiv.2204.00559

39. S. Chen, Y. Bhalgat, X. Li, J.-W. Bian, K. Li, Z. Wang, and V. A. Prisacariu, “Neural refinement for absolute pose regression with feature synthesis,” in CVPR, 2024, pp. 20987–20996. https://doi.org/10.1109/CVPR52733.2024.01983

40. Davison, Andrew J. et al. “MonoSLAM: Real-Time Single Camera SLAM.” IEEE Transactions on Pattern Analysis and Machine Intelligence 29 (2007): 1052-1067. https://doi.org/10.1109/TPAMI.2007.1049

41. Klein, Georg S. W. and David William Murray. “Parallel Tracking and Mapping for Small AR Workspaces.” 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality (2007): 225-234. https://doi.org/10.1109/ISMAR.2007.4538852

42. R. Mur-Artal, J.M.M. Montiel, J.D. Tardos. “ORB-SLAM: A Versatile and Accurate Monocular SLAM System.” IEEE Transactions on Robotics 31 (2015): 1147-1163. https://doi.org/10.1109/TRO.2015.2463671

43. Kesorn K, Poslad S. An enhanced bag of visual word vector space model to represent visual content in athletics images[J].IEEE Trans on Multimedia, 2011,14(1): 211-222. https://doi.org/10.1109/TMM.2011.2170665

44. Mur-Artal, Raul and Juan D. Tardós. “ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras.” IEEE Transactions on Robotics 33 (2016): 1255-1262. https://doi.org/10.1109/TRO.2017.2705103

45. C. Campos, R. Elvira, J.J.G. Rodríguez, J.M. Montiel, J.D. Tardós. “ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM.” IEEE Transactions on Robotics 37 (2020): 1874-1890. https://doi.org/10.1109/TRO.2021.3075644

46. Newcombe, Richard A. et al. “DTAM: Dense tracking and mapping in real-time.” 2011 International Conference on Computer Vision (2011): 2320-2327. https://doi.org/10.1109/ICCV.2011.6126513

47. J. Engel, T. Schöps, D. Cremers. “LSD-SLAM: Large-Scale Direct Monocular SLAM.” European Conference on Computer Vision (2014). https://doi.org/10.1007/978-3-319-10605-2_54

48. J. Engel, V. Koltun, D. Cremers. “Direct Sparse Odometry.” IEEE Transactions on Pattern Analysis and Machine Intelligence 40 (2016): 611-625. https://doi.org/10.1109/TPAMI.2017.2658577

49. Schöps, Thomas et al. “BAD SLAM: Bundle Adjusted Direct RGB-D SLAM.” 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019): 134-144. https://doi.org/10.1109/CVPR.2019.00022

50. Yang, Nan et al. “D3VO: Deep Depth, Deep Pose and Deep Uncertainty for Monocular Visual Odometry.” 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020): 1278-1289. https://doi.org/10.1109/CVPR42600.2020.00136

51. K. Tateno, F. Tombari, I. Laina, N. Navab. “CNN-SLAM: Real-Time Dense Monocular SLAM with Learned Depth Prediction.” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017): 6565-6574. https://doi.org/10.1109/CVPR.2017.695

52. Zhou, Huizhong et al. “DeepTAM: Deep Tracking and Mapping.” European Conference on Computer Vision (2018). https://doi.org/10.48550/arXiv.1808.01900

53. Yu, Chao et al. “DS-SLAM: A Semantic Visual SLAM towards Dynamic Environments.” 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (2018): 1168-1174. https://doi.org/10.1109/IROS.2018.8593691

54. Yang, Shichao and Sebastian A. Scherer. “CubeSLAM: Monocular 3-D Object SLAM.” IEEE Transactions on Robotics 35 (2018): 925-938. https://doi.org/10.1109/TRO.2019.2909168

55. Zins, Matthieu et al. “OA-SLAM: Leveraging Objects for Camera Relocalization in Visual SLAM.” 2022 IEEE International Symposium on Mixed and Augmented Reality (ISMAR) (2022): 720-728. https://doi.org/10.48550/arXiv.2209.08338

56. Kerbl, Bernhard et al. “3D Gaussian Splatting for Real-Time Radiance Field Rendering.” ACM Transactions on Graphics (TOG) 42 (2023): 1 - 14. https://doi.org/10.1145/3592433

57. E. Sucar, S. Liu, J. Ortiz, A.J. Davison. “iMAP: Implicit Mapping and Positioning in Real-Time.” 2021 IEEE/CVF International Conference on Computer Vision (ICCV) (2021): 6209-6218. https://doi.org/10.48550/arXiv.2103.12352

58. Z. Zhu, S. Peng, V. Larsson, W. Xu, H. Bao, Z. Cui, M.R. Oswald, M. Pollefeys. “NICE-SLAM: Neural Implicit Scalable Encoding for SLAM.” 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2021): 12776-12786. https://doi.org/10.48550/arXiv.2112.12130

59. Zou, Danping and Ping Tan. “CoSLAM: Collaborative Visual SLAM in Dynamic Environments.” IEEE Transactions on Pattern Analysis and Machine Intelligence 35 (2013): 354-366. https://doi.org/10.1109/TPAMI.2012.104

60. Yang, Xingrui et al. “Vox-Fusion: Dense Tracking and Mapping with Voxel-based Neural Implicit Representation.” 2022 IEEE International Symposium on Mixed and Augmented Reality (ISMAR) (2022): 499-507. https://doi.org/10.48550/arXiv.2210.15858

61. Johari, M. M., Carta, C., & Fleuret, F. “ESLAM: Efficient Dense SLAM System Based on Hybrid Representation of Signed Distance Fields.” 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2022): 17408-17419. https://doi.org/10.1109/cvpr52729.2023.01670

62. C. Yan, D. Qu, D. Xu, B. Zhao, Z. Wang, D. Wang, X. Li. “GS-SLAM: Dense Visual SLAM with 3D Gaussian Splatting.” 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2023): 19595-19604. https://doi.org/10.1109/CVPR52733.2024.01853

63. Hidenobu Matsuki, Riku Murai, Paul H.J. Kelly, and An drew J. Davison. Gaussian splatting slam. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18039–18048, 2024. https://doi.org/10.1109/CVPR52733.2024.01708

64. Huajian Huang, Longwei Li, Hui Cheng, and Sai-Kit Yeung. Photo-slam: Real-time simultaneous localization and photo realistic mapping for monocular stereo and rgb-d cameras. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 21584–21593, 2024. https://doi.org/ 10.1109/CVPR52733.2024.02039

65. N. Keetha, J. Karhade, K.M. Jatavallabhula, G. Yang, S. Scherer, D. Ramanan, J. Luiten. “SplaTAM: Splat, Track & Map 3D Gaussians for Dense RGB-D SLAM.” 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2023): 21357-21366. https://doi.org/10.48550/arXiv.2312.02126

66. Chaoyang Guo, Chunyan Gao,Yiyang Bai. “RD-SLAM: Real-Time Dense SLAM Using Gaussian Splatting.” Applied Sciences (2024): n. pag. https://doi.org/10.3390/app14177767

67. Xinggang Hu, Chenyangguang Zhang, Mingyuan Zhao, et al. “DyGS-SLAM: Realistic Map Reconstruction in Dynamic Scenes Based on Double-Constrained Visual SLAM.” Remote Sensing (2025): n. pag. https://doi.org/10.3390/rs17040625

68. Mingrui Li,Yiming Zhou,Hongxing Zhou, et al. “Dy3DGS-SLAM: Monocular 3D Gaussian Splatting SLAM for Dynamic Environments.” 2025 IEEE International Conference on Robotics and Automation (ICRA) (2025): 14572-14578. https://doi.org/10.1109/ICRA55743.2025.11127324

69. Tianci Wen, Zhiang Liu, Yongchun Fang. “SEGS-SLAM: Structure-enhanced 3D Gaussian Splatting SLAM with Appearance Embedding.” (2025). https://doi.org/10.48550/arXiv.2501.05242

70. Zhexi Peng, Tianjia Shao, Yong Liu, Jingke Zhou, Yin Yang, Jingdong Wang, and Kun Zhou. Rtg-slam: Real-time 3d re construction at scale using gaussian splatting. In ACM SIG GRAPH, 2024. https://doi.org/10.48550/arXiv.2404.19706 

Supporting Agencies



Copyright (c) 2026 Authors

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 International License.


This site is licensed under a Creative Commons Attribution 4.0 International License (CC BY 4.0).