[an error occurred while processing this directive]

ZTE Communications ›› 2023, Vol. 21 ›› Issue (1): 81-88.DOI: 10.12142/ZTECOM.202301010

• • 上一篇    下一篇

  

  • 收稿日期:2022-11-01 出版日期:2023-03-25 发布日期:2024-03-15

Scene Visual Perception and AR Navigation Applications

LU Ping1,2, SHENG Bin2(), SHI Wenzhe1,2   

  1. 1.State Key Laboratory of Mobile Network and Mobile Multimedia Technology, Shenzhen 518055, China
    2.ZTE Corporation, Shenzhen 518057, China
    3.Shanghai Jiao Tong University, Shanghai 200240, China
  • Received:2022-11-01 Online:2023-03-25 Published:2024-03-15
  • About author:LU Ping is the Vice President and general manager of the Industrial Digitalization Solution Department of ZTE Corporation, and Executive Deputy Director of the National Key Laboratory of Mobile Network and Mobile Multimedia Technology. His research directions include cloud computing, big data, augmented reality, and multimedia service-based technologies. He has supported and participated in major national science and technology projects and national science and technology support projects. He has published multiple papers, and authored two books.
    SHENG Bin ( shengbin@cs.sjtu.edu.cn) is a professor of computer science and engineering from Shanghai Jiao Tong University, China. His research directions include virtual reality and computer graphics. He has presided over two projects on the National Natural Science Foundation of China, one youth project of the National Natural Science Foundation of China, and participates in one high-technology research and development plan (the “863” plan) and one key project of the National Natural Science Foundation of China. He has published 121 papers in different journals.
    SHI Wenzhe is a strategy planning engineer with ZTE Corporation, a member of the National Key Laboratory for Mobile Network and Mobile Multimedia Technology, and an engineer of XRExplore Platform Product Planning. His research interests include indoor visual AR navigation, SFM 3D reconstruction, visual SLAM, real-time cloud rendering, VR, and spatial perception.
  • Supported by:
    ZTE Industry‐University‐Institute Cooperation Funds;HC‐CN‐20210707004

Abstract:

With the rapid popularization of mobile devices and the wide application of various sensors, scene perception methods applied to mobile devices occupy an important position in location-based services such as navigation and augmented reality (AR). The development of deep learning technologies has greatly improved the visual perception ability of machines to scenes. The basic framework of scene visual perception, related technologies and the specific process applied to AR navigation are introduced, and future technology development is proposed. An application (APP) is designed to improve the application effect of AR navigation. The APP includes three modules: navigation map generation, cloud navigation algorithm, and client design. The navigation map generation tool works offline. The cloud saves the navigation map and provides navigation algorithms for the terminal. The terminal realizes local real-time positioning and AR path rendering.

Key words: 3D reconstruction, image matching, visual localization, AR navigation, deep learning