{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T06:22:11Z","timestamp":1769840531385,"version":"3.49.0"},"reference-count":99,"publisher":"Association for Computing Machinery (ACM)","issue":"1","license":[{"start":{"date-parts":[[2022,3,29]],"date-time":"2022-03-29T00:00:00Z","timestamp":1648512000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-sa\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CSR-1903136, CNS-1908051, CAREER-2046072"],"award-info":[{"award-number":["CSR-1903136, CNS-1908051, CAREER-2046072"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["326346, 319710, 332307, 338854"],"award-info":[{"award-number":["326346, 319710, 332307, 338854"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Interact. Mob. Wearable Ubiquitous Technol."],"published-print":{"date-parts":[[2022,3,29]]},"abstract":"<jats:p>Mobile Augmented Reality (AR) demands realistic rendering of virtual content that seamlessly blends into the physical environment. For this reason, AR headsets and recent smartphones are increasingly equipped with Time-of-Flight (ToF) cameras to acquire depth maps of a scene in real-time. ToF cameras are cheap and fast, however, they suffer from several issues that affect the quality of depth data, ultimately hampering their use for mobile AR. Among them, scale errors of virtual objects - appearing much bigger or smaller than what they should be - are particularly noticeable and unpleasant. This article specifically addresses these challenges by proposing InDepth, a real-time depth inpainting system based on edge computing. InDepth employs a novel deep neural network (DNN) architecture to improve the accuracy of depth maps obtained from ToF cameras. The DNN fills holes and corrects artifacts in the depth maps with high accuracy and eight times lower inference time than the state of the art. An extensive performance evaluation in real settings shows that InDepth reduces the mean absolute error by a factor of four with respect to ARCore DepthLab. Finally, a user study reveals that InDepth is effective in rendering correctly-scaled virtual objects, outperforming DepthLab.<\/jats:p>","DOI":"10.1145\/3517260","type":"journal-article","created":{"date-parts":[[2022,3,29]],"date-time":"2022-03-29T13:42:46Z","timestamp":1648561366000},"page":"1-25","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":25,"title":["InDepth"],"prefix":"10.1145","volume":"6","author":[{"given":"Yunfan","family":"Zhang","sequence":"first","affiliation":[{"name":"Duke University, Durham, NC, USA"}]},{"given":"Tim","family":"Scargill","sequence":"additional","affiliation":[{"name":"Duke University, Durham, NC, USA"}]},{"given":"Ashutosh","family":"Vaishnav","sequence":"additional","affiliation":[{"name":"Aalto University, Espoo, Finland"}]},{"given":"Gopika","family":"Premsankar","sequence":"additional","affiliation":[{"name":"University of Helsinki, Helsinki, Finland"}]},{"given":"Mario","family":"Di Francesco","sequence":"additional","affiliation":[{"name":"Aalto University, Espoo, Finland"}]},{"given":"Maria","family":"Gorlatova","sequence":"additional","affiliation":[{"name":"Duke University, Durham, NC, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,3,29]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2012.120"},{"key":"e_1_2_1_2_1","unstructured":"Amazon. 2021. Amazon AR view. https:\/\/www.amazon.com\/adlp\/arview."},{"key":"e_1_2_1_3_1","unstructured":"Apple. 2021. Augmented Reality - Apple. https:\/\/www.apple.com\/augmented-reality\/."},{"key":"e_1_2_1_4_1","unstructured":"Apple. 2022. ARKit overview. https:\/\/developer.apple.com\/augmented-reality\/arkit\/."},{"key":"e_1_2_1_5_1","doi-asserted-by":"crossref","unstructured":"Jonathan T Barron and Ben Poole. 2016. The fast bilateral solver. In ECCV.","DOI":"10.1007\/978-3-319-46487-9_38"},{"key":"e_1_2_1_6_1","volume-title":"Zakieh Sadat Hashemifar, and Karthik Dantu","author":"Ben Ali Ali J.","year":"2020","unstructured":"Ali J. Ben Ali, Zakieh Sadat Hashemifar, and Karthik Dantu. 2020. Edge-SLAM: Edge-Assisted Visual Simultaneous Localization and Mapping. In ACM MobiSys."},{"key":"e_1_2_1_7_1","unstructured":"Mary Branscombe. 2018. How Microsoft is making its most sensitive HoloLens depth sensor yet. https:\/\/www.zdnet.com\/article\/howmicrosoft-is-making-its-most-sensitive-hololens-depth-sensor-yet."},{"key":"e_1_2_1_8_1","first-page":"855","article-title":"Close-range camera calibration","volume":"37","author":"Brown Duane C.","year":"1971","unstructured":"Duane C. Brown. 1971. Close-range camera calibration. Photogrammetric Engineering 37, 8 (1971), 855--866.","journal-title":"Photogrammetric Engineering"},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.1986.4767851"},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/3DV.2017.00081"},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2017.2699184"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2947374"},{"key":"e_1_2_1_13_1","volume-title":"Mask-ToF: Learning Microlens Masks for Flying Pixel Correction in Time-of-Flight Imaging","author":"Chugunov Ilya","unstructured":"Ilya Chugunov, Seung-Hwan Baek, Qiang Fu, Wolfgang Heidrich, and Felix Heide. 2021. Mask-ToF: Learning Microlens Masks for Flying Pixel Correction in Time-of-Flight Imaging. In IEEE CVPR."},{"key":"e_1_2_1_14_1","doi-asserted-by":"crossref","unstructured":"J. Deng W. Dong R. Socher L. Li K. Li and F. Li. 2009. ImageNet: A large-scale hierarchical image database. In IEEE CVPR.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_2_1_15_1","unstructured":"Ruofei Du Eric Turner Maksym Dzitsiuk Luca Prasso Ivo Duarte Jason Dourgarian Joao Afonso Jose Pascoal Josh Gladstone Nuno Cruces Shahram Izadi Adarsh Kowdle Konstantine Tsotsos and David Kim. 2020. DepthLab: Real-Time 3D Interaction With Depth Maps for Mobile Augmented Reality. In ACM UIST."},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00138-017-0831-9"},{"key":"e_1_2_1_17_1","unstructured":"Elsevier. [n. d.]. 3D4Medical - Augmented Reality. https:\/\/3d4medical.com\/support\/complete-anatomy\/ar."},{"key":"e_1_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/JSEN.2010.2101060"},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCI.2015.2510506"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2015.09.023"},{"key":"e_1_2_1_21_1","volume-title":"Design preferences on Industrial Augmented Reality: a survey with potential technical writers","author":"Gattullo Michele","year":"2020","unstructured":"Michele Gattullo, Lucilla Dammacco, Francesca Ruospo, Alessandro Evangelista, Michele Fiorentino, Jan Schmitt, and Antonio E Uva. 2020. Design preferences on Industrial Augmented Reality: a survey with potential technical writers. In IEEE ISMAR Adjunct (2020)."},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.measurement.2015.03.042"},{"key":"e_1_2_1_23_1","unstructured":"Google. 2021. Experience 3D & augmented reality in search. https:\/\/support.google.com\/websearch\/answer\/9817187?co=GENIE.Platform%3DAndroid&hl=en&oco=0."},{"key":"e_1_2_1_24_1","unstructured":"Google. 2021. Introduction to Depth on Unity targeting Android. https:\/\/developers.google.com\/ar\/develop\/unity\/depth\/introduction."},{"key":"e_1_2_1_25_1","unstructured":"Google. 2022. Use Raw Depth in your Android app. https:\/\/developers.google.com\/ar\/develop\/java\/depth\/raw-depth."},{"key":"e_1_2_1_26_1","volume-title":"Literature survey on stereo vision disparity map algorithms. Journal of Sensors 2016","author":"Hamzah Rostam Affendi","year":"2016","unstructured":"Rostam Affendi Hamzah and Haidi Ibrahim. 2016. Literature survey on stereo vision disparity map algorithms. Journal of Sensors 2016 (2016)."},{"key":"e_1_2_1_27_1","volume-title":"Time-of-flight cameras: Principles, methods and applications","author":"Hansard Miles","unstructured":"Miles Hansard, Seungkyu Lee, Ouk Choi, and Radu Horaud. 2012. Time-of-flight cameras: Principles, methods and applications. Springer."},{"key":"e_1_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Alastair Harrison and Paul Newman. 2010. Image and Sparse Laser Fusion for Dense Scene Reconstruction. In Field and Service Robotics.","DOI":"10.1007\/978-3-642-13408-1_20"},{"key":"e_1_2_1_29_1","unstructured":"Kaiming He Xiangyu Zhang Shaoqing Ren and Jian Sun. 2015. Deep Residual Learning for Image Recognition. arXiv:1512.03385 [cs.CV]"},{"key":"e_1_2_1_30_1","volume-title":"An overview of depth cameras and range scanners based on time-of-flight technologies. Machine vision and applications 27, 7","author":"Horaud Radu","year":"2016","unstructured":"Radu Horaud, Miles Hansard, Georgios Evangelidis, and Cl\u00e9ment M\u00e9nier. 2016. An overview of depth cameras and range scanners based on time-of-flight technologies. Machine vision and applications 27, 7 (2016), 1005--1020."},{"key":"e_1_2_1_31_1","volume-title":"PENet: Towards Precise and Efficient Image Guided Depth Completion","author":"Hu Mu","unstructured":"Mu Hu, Shuling Wang, Bin Li, Shiyu Ning, Li Fan, and Xiaojin Gong. 2021. PENet: Towards Precise and Efficient Image Guided Depth Completion. In IEEE ICRA."},{"key":"e_1_2_1_32_1","volume-title":"Indoor Depth Completion with Boundary Consistency and Self-Attention. In IEEE ICCV Workshops.","author":"Huang Yu-Kai","unstructured":"Yu-Kai Huang, Tsung-Han Wu, Yueh-Cheng Liu, and Winston H. Hsu. 2019. Indoor Depth Completion with Boundary Consistency and Self-Attention. In IEEE ICCV Workshops."},{"key":"e_1_2_1_33_1","unstructured":"Huawei. 2021. HUAWEI P40 Pro. https:\/\/consumer.huawei.com\/en\/phones\/p40-pro\/specs\/."},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.3390\/s20041021"},{"key":"e_1_2_1_35_1","volume-title":"Chen Change Loy, and Xiaoou Tang","author":"Hui Tak-Wai","year":"2016","unstructured":"Tak-Wai Hui, Chen Change Loy, and Xiaoou Tang. 2016. Depth map super-resolution by deep multi-scale guidance. In ECCV."},{"key":"e_1_2_1_36_1","unstructured":"IKEA. 2021. IKEA Apps. https:\/\/www.ikea.com\/us\/en\/customer-service\/mobile-apps\/."},{"key":"e_1_2_1_37_1","unstructured":"International Telecommunication Union (ITU) Radiocommunication Sector. 2005. BT.470: Conventional analogue television systems. Available at https:\/\/www.itu.int\/rec\/R-REC-BT.470-6-199811-S\/en."},{"key":"e_1_2_1_38_1","unstructured":"International Telecommunications Union (ITU) Telecommunication Standardization Sector. 2018. P.808: Subjective evaluation of speech quality with a crowdsourcing approach. Available at https:\/\/www.itu.int\/rec\/T-REC-P.808-201806-I\/en."},{"key":"e_1_2_1_39_1","unstructured":"International Telecommunications Union (ITU) Telecommunication Standardization Sector. 2018. P.809: Subjective evaluation methods for gaming quality. Available at https:\/\/www.itu.int\/rec\/T-REC-P.809-201806-I\/en."},{"key":"e_1_2_1_40_1","volume-title":"Temporal RVL: A Depth Stream Compression Method. In 2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW). IEEE, 664--665","author":"Jun Hanseul","year":"2020","unstructured":"Hanseul Jun and Jeremy Bailenson. 2020. Temporal RVL: A Depth Stream Compression Method. In 2020 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW). IEEE, 664--665."},{"key":"e_1_2_1_41_1","volume-title":"Towards augmented reality user interfaces in 3D media production","author":"Krichenbauer Max","year":"2014","unstructured":"Max Krichenbauer, Goshiro Yamamoto, Takafumi Taketomi, Christian Sandor, and Hirokazu Kato. 2014. Towards augmented reality user interfaces in 3D media production. In IEEE ISMAR (2014)."},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/3DV.2016.32"},{"key":"e_1_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-03731-8_38"},{"key":"e_1_2_1_44_1","unstructured":"Luyang Liu Hongyu Li and Marco Gruteser. 2019. Edge Assisted Real-Time Object Detection for Mobile Augmented Reality. In ACM MobiCom."},{"key":"e_1_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICNP.2018.00011"},{"key":"e_1_2_1_46_1","doi-asserted-by":"crossref","unstructured":"Zida Liu Guohao Lan Jovan Stojkovic Yunfan Zhang Carlee Joe-Wong and Maria Gorlatova. 2020. CollabAR: Edge-assisted Collaborative Image Recognition for Mobile Augmented Reality. In ACM\/IEEE IPSN.","DOI":"10.1109\/IPSN48710.2020.00-26"},{"key":"e_1_2_1_47_1","doi-asserted-by":"crossref","unstructured":"Zifan Liu Hongzi Zhu Junchi Chen Shan Chang and Lili Qiu. 2019. HyperSight: Boosting distant 3D vision on a single dual-camera smartphone. In ACM SenSys.","DOI":"10.1145\/3356250.3360029"},{"key":"e_1_2_1_48_1","volume-title":"Sparse-to-dense: Depth prediction from sparse depth samples and a single image","author":"Ma Fangchang","year":"2018","unstructured":"Fangchang Ma and Sertac Karaman. 2018. Sparse-to-dense: Depth prediction from sparse depth samples and a single image. In IEEE ICRA."},{"key":"e_1_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/MMSP.2011.6093803"},{"key":"e_1_2_1_50_1","volume-title":"Robust Monocular Visual-Inertial Depth Completion for Embedded Systems","author":"Merrill Nate","unstructured":"Nate Merrill, Patrick Geneva, and Guoquan Huang. 2021. Robust Monocular Visual-Inertial Depth Completion for Embedded Systems. In IEEE ICRA."},{"key":"e_1_2_1_51_1","unstructured":"Microsoft. 2021. Microsoft Mixed Reality - Manufacturing. https:\/\/www.microsoft.com\/en-us\/hololens\/industry-manufacturing."},{"key":"e_1_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2020.12.089"},{"key":"e_1_2_1_53_1","volume-title":"Real-time scattering compensation for time-of-flight camera","author":"Mure-Dubois James","unstructured":"James Mure-Dubois and Heinz H\u00fcgli. 2007. Real-time scattering compensation for time-of-flight camera. In IEEE ICCV."},{"key":"e_1_2_1_54_1","unstructured":"Niantic. 2021. Catching Pok\u00e9mon in AR+ mode. https:\/\/niantic.helpshift.com\/a\/pokemon-go\/?s=accessories&f=catchingpokemon-in-ar-mode&amp;p=web."},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/ISMAR.2011.6092372"},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2015.2459902"},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3463498"},{"key":"e_1_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10278-019-00272-6"},{"key":"e_1_2_1_59_1","volume-title":"DeepLiDAR: Deep Surface Normal Guided Depth Prediction for Outdoor Scene From Sparse LiDAR Data and Single Color Image","author":"Qiu Jiaxiong","unstructured":"Jiaxiong Qiu, Zhaopeng Cui, Yinda Zhang, Xingdi Zhang, Shuaicheng Liu, Bing Zeng, and Marc Pollefeys. 2019. DeepLiDAR: Deep Surface Normal Guided Depth Prediction for Outdoor Scene From Sparse LiDAR Data and Single Color Image. In IEEE CVPR."},{"key":"e_1_2_1_60_1","unstructured":"Qualtrics. 2021. Qualtrics. https:\/\/www.qualtrics.com."},{"key":"e_1_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3097895.3097903"},{"key":"e_1_2_1_62_1","doi-asserted-by":"crossref","unstructured":"Xukan Ran Carter Slocum Maria Gorlatova and Jiasi Chen. 2019. ShareAR: Communication-efficient multi-user mobile augmented reality. In ACM HotNets.","DOI":"10.1145\/3365609.3365867"},{"key":"e_1_2_1_63_1","doi-asserted-by":"crossref","unstructured":"Gernot Riegler Matthias R\u00fcther and Horst Bischof. 2016. ATGV-Net: Accurate depth super-resolution. In ECCV.","DOI":"10.1007\/978-3-319-46487-9_17"},{"key":"e_1_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"e_1_2_1_65_1","unstructured":"Samsung. 2020. What is ToF camera technology on Galaxy and how does it work? https:\/\/www.samsung.com\/global\/galaxy\/whatis\/tof-camera\/."},{"key":"e_1_2_1_66_1","volume-title":"MobileNetV2: Inverted residuals and linear bottlenecks","author":"Sandler Mark","unstructured":"Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, and Liang-Chieh Chen. 2018. MobileNetV2: Inverted residuals and linear bottlenecks. In IEEE CVPR."},{"key":"e_1_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1109\/MPRV.2009.82"},{"key":"e_1_2_1_68_1","volume-title":"Here to stay: Measuring hologram stability in markerless smartphone augmented reality. arXiv preprint arXiv:2109.14757","author":"Scargill Tim","year":"2021","unstructured":"Tim Scargill, Jiasi Chen, and Maria Gorlatova. 2021. Here to stay: Measuring hologram stability in markerless smartphone augmented reality. arXiv preprint arXiv:2109.14757 (2021)."},{"key":"e_1_2_1_69_1","volume-title":"Saman Zadtootaghaj, and Sebastian M\u00f6ller.","author":"Schmidt Steven","year":"2020","unstructured":"Steven Schmidt, Babak Naderi, Saeed Shafiee Sabet, Saman Zadtootaghaj, and Sebastian M\u00f6ller. 2020. Assessing Interactive Gaming Quality of Experience Using a Crowdsourcing Approach. In IEEE QoMEX."},{"key":"e_1_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2016.09.007"},{"key":"e_1_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1109\/LWC.2021.3057114"},{"key":"e_1_2_1_72_1","doi-asserted-by":"crossref","unstructured":"Yawar Siddiqui Julien Valentin and Matthias NieBner. 2020. ViewAL: Active Learning With Viewpoint Entropy for Semantic Segmentation. In IEEE CVPR.","DOI":"10.1109\/CVPR42600.2020.00945"},{"key":"e_1_2_1_73_1","volume-title":"Very Deep Convolutional Networks for Large-Scale Image Recognition. In International Conference on Learning Representations.","author":"Simonyan Karen","year":"2015","unstructured":"Karen Simonyan and Andrew Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. In International Conference on Learning Representations."},{"key":"e_1_2_1_74_1","volume-title":"SUN RGB-D: A RGB-D Scene Understanding Benchmark Suite","author":"Song Shuran","unstructured":"Shuran Song, Samuel P. Lichtenberg, and Jianxiong Xiao. 2015. SUN RGB-D: A RGB-D Scene Understanding Benchmark Suite. In IEEE CVPR."},{"key":"e_1_2_1_75_1","volume-title":"Asian Conference on Computer Vision.","author":"Song Xibin","year":"2016","unstructured":"Xibin Song, Yuchao Dai, and Xueying Qin. 2016. Deep depth super-resolution: Learning depth super-resolution using deep convolutional neural network. In Asian Conference on Computer Vision."},{"key":"e_1_2_1_76_1","volume-title":"EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In International Conference on Machine Learning.","author":"Tan Mingxing","year":"2019","unstructured":"Mingxing Tan and Quoc Le. 2019. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In International Conference on Machine Learning."},{"key":"e_1_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2020.3040528"},{"key":"e_1_2_1_78_1","unstructured":"TechInsights. 2020. Sony d-ToF sensor identified in Apple's New LiDAR Camera. https:\/\/www.techinsights.com\/blog\/sony-d-tofsensor-found-apples-new-lidar-camera."},{"key":"e_1_2_1_79_1","unstructured":"Texas Instruments. 2014. Time-of-Flight Camera - An Introduction (Rev. B). Technical white paper. https:\/\/www.ti.com\/lit\/wp\/sloa190b\/sloa190b.pdf."},{"key":"e_1_2_1_80_1","unstructured":"Texas Instruments. 2019. Introduction to Time-of-Flight Long Range Proximity and Distance Sensor System Design (Rev. B). https:\/\/www.ti.com\/lit\/pdf\/sbau305."},{"key":"e_1_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1145\/3272127.3275041"},{"key":"e_1_2_1_82_1","volume-title":"DIODE: A dense indoor and outdoor depth dataset. CoRR abs\/1908.00463","author":"Vasiljevic Igor","year":"2019","unstructured":"Igor Vasiljevic, Nick Kolkin, Shanyi Zhang, Ruotian Luo, Haochen Wang, Falcon Z. Dai, Andrea F. Daniele, Mohammadreza Mostajabi, Steven Basart, Matthew R. Walter, and Gregory Shakhnarovich. 2019. DIODE: A dense indoor and outdoor depth dataset. CoRR abs\/1908.00463 (2019). http:\/\/arxiv.org\/abs\/1908.00463"},{"key":"e_1_2_1_83_1","volume-title":"Peters","author":"Vassallo Reid","year":"2017","unstructured":"Reid Vassallo, Adam Rankin, Elvis C. S. Chen, and Terry M. Peters. 2017. Hologram stability evaluation for Microsoft HoloLens. In SPIE Medical Imaging 2017: Image Perception, Observer Performance, and Technology Assessment."},{"key":"e_1_2_1_84_1","volume-title":"International Conference on Neural Information Processing Systems (NeurIPS).","author":"Veit Andreas","year":"2016","unstructured":"Andreas Veit, Michael Wilber, and Serge Belongie. 2016. Residual Networks Behave like Ensembles of Relatively Shallow Networks. In International Conference on Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_2_1_85_1","volume-title":"BiFuse: Monocular 360 Depth Estimation via Bi-Projection Fusion","author":"Yeh Yu-Hsuan","unstructured":"Fu-EnWang, Yu-Hsuan Yeh, Min Sun,Wei-Chen Chiu, and Yi-Hsuan Tsai. 2020. BiFuse: Monocular 360 Depth Estimation via Bi-Projection Fusion. In IEEE CVPR."},{"key":"e_1_2_1_86_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2933987"},{"key":"e_1_2_1_87_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107274"},{"key":"e_1_2_1_88_1","unstructured":"Wayfair. 2020. Augmented Reality with a purpose. https:\/\/www.aboutwayfair.com\/augmented-reality-with-a-purpose."},{"key":"e_1_2_1_89_1","doi-asserted-by":"publisher","DOI":"10.1145\/3132272.3134144"},{"key":"e_1_2_1_90_1","unstructured":"Zhiyuan Xie Xiaomin Ouyang and Guoliang Xing. 2021. UltraDepth: Exposing High-Resolution Texture from Depth Cameras. In ACM SenSys."},{"key":"e_1_2_1_91_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2017.2669347"},{"key":"e_1_2_1_92_1","volume-title":"Spatial-Depth Super Resolution for Range Images","author":"Yang Qingxiong","unstructured":"Qingxiong Yang, Ruigang Yang, James Davis, and David Nister. 2007. Spatial-Depth Super Resolution for Range Images. In IEEE CVPR."},{"key":"e_1_2_1_93_1","doi-asserted-by":"crossref","unstructured":"W. Yin Y. Liu C. Shen and Y. Yan. 2019. Enforcing Geometric Constraints of Virtual Normal for Depth Prediction. In IEEE ICCV.","DOI":"10.1109\/ICCV.2019.00578"},{"key":"e_1_2_1_94_1","volume-title":"International Conference on Learning Representations.","author":"Yu Fisher","year":"2016","unstructured":"Fisher Yu and Vladlen Koltun. 2016. Multi-Scale Context Aggregation by Dilated Convolutions. In International Conference on Learning Representations."},{"key":"e_1_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.1145\/3384419.3430721"},{"key":"e_1_2_1_96_1","volume-title":"Deep Depth Completion of a Single RGB-D Image","author":"Zhang Yinda","unstructured":"Yinda Zhang and Thomas Funkhouser. 2018. Deep Depth Completion of a Single RGB-D Image. In IEEE CVPR."},{"key":"e_1_2_1_97_1","doi-asserted-by":"crossref","unstructured":"Yiqin Zhao and Tian Guo. 2020. PointAR: Efficient Lighting Estimation for Mobile Augmented Reality. In ECCV.","DOI":"10.1007\/978-3-030-58592-1_40"},{"key":"e_1_2_1_98_1","volume-title":"RGB-D Image Analysis and Processing","author":"Zollh\u00f6fer Michael","unstructured":"Michael Zollh\u00f6fer. 2019. Commodity RGB-D sensors: Data acquisition. In RGB-D Image Analysis and Processing. Springer, 3--13."},{"key":"e_1_2_1_99_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2016.11.001"}],"container-title":["Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3517260","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3517260","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3517260","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T04:25:40Z","timestamp":1752467140000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3517260"}},"subtitle":["Real-time Depth Inpainting for Mobile Augmented Reality"],"short-title":[],"issued":{"date-parts":[[2022,3,29]]},"references-count":99,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2022,3,29]]}},"alternative-id":["10.1145\/3517260"],"URL":"https:\/\/doi.org\/10.1145\/3517260","relation":{},"ISSN":["2474-9567"],"issn-type":[{"value":"2474-9567","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,3,29]]},"assertion":[{"value":"2022-03-29","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}