Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article
Public Access

LITAR: Visually Coherent Lighting for Mobile Augmented Reality

Published: 07 September 2022 Publication History

Abstract

An accurate understanding of omnidirectional environment lighting is crucial for high-quality virtual object rendering in mobile augmented reality (AR). In particular, to support reflective rendering, existing methods have leveraged deep learning models to estimate or have used physical light probes to capture physical lighting, typically represented in the form of an environment map. However, these methods often fail to provide visually coherent details or require additional setups. For example, the commercial framework ARKit uses a convolutional neural network that can generate realistic environment maps; however the corresponding reflective rendering might not match the physical environments. In this work, we present the design and implementation of a lighting reconstruction framework called LITAR that enables realistic and visually-coherent rendering. LITAR addresses several challenges of supporting lighting information for mobile AR.
First, to address the spatial variance problem, LITAR uses two-field lighting reconstruction to divide the lighting reconstruction task into the spatial variance-aware near-field reconstruction and the directional-aware far-field reconstruction. The corresponding environment map allows reflective rendering with correct color tones. Second, LITAR uses two noise-tolerant data capturing policies to ensure data quality, namely guided bootstrapped movement and motion-based automatic capturing. Third, to handle the mismatch between the mobile computation capability and the high computation requirement of lighting reconstruction, LITAR employs two novel real-time environment map rendering techniques called multi-resolution projection and anchor extrapolation. These two techniques effectively remove the need of time-consuming mesh reconstruction while maintaining visual quality. Lastly, LITAR provides several knobs to facilitate mobile AR application developers making quality and performance trade-offs in lighting reconstruction. We evaluated the performance of LITAR using a small-scale testbed experiment and a controlled simulation. Our testbed-based evaluation shows that LITAR achieves more visually coherent rendering effects than ARKit. Our design of multi-resolution projection significantly reduces the time of point cloud projection from about 3 seconds to 14.6 milliseconds. Our simulation shows that LITAR, on average, achieves up to 44.1% higher PSNR value than a recent work Xihe on two complex objects with physically-based materials.

Supplementary Material

zhao (zhao.zip)
Supplemental movie, appendix, image and software files for, LITAR: Visually Coherent Lighting for Mobile Augmented Reality

References

[1]
Ibraheem Alhashim and Peter Wonka. 2018. High Quality Monocular Depth Estimation via Transfer Learning. arXiv e-prints abs/1812.11941, Article arXiv:1812.11941 (2018). arXiv:1812.11941 https://arxiv.org/abs/1812.11941
[2]
Amazon. 2020. Amazon AR View. https://www.amazon.com/adlp/arview. Accessed: 2020-7-2.
[3]
Christopher Andrews, Michael K Southworth, Jennifer N A Silva, and Jonathan R Silva. 2019. Extended Reality in Medical Practice. Curr. Treat. Options Cardiovasc. Med. 21, 4 (March 2019), 18.
[4]
Apple. 2022. ARCoachingOverlayView. https://developer.apple.com/documentation/arkit/arcoachingoverlayview.
[5]
Apple. 2022. iPhone 13 Pro Tech Specs. https://www.apple.com/iphone-13-pro/specs/.
[6]
Dejan Azinovic, Tzu-Mao Li, Anton Kaplanyan, and Matthias Nießner. 2019. Inverse path tracing for joint material and lighting estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE, Long Beach, CA, 2447--2456.
[7]
Ali J Ben Ali, Zakieh Sadat Hashemifar, and Karthik Dantu. 2020. Edge-SLAM: edge-assisted visual simultaneous localization and mapping. In Proceedings of the 18th International Conference on Mobile Systems, Applications, and Services (MobiSys '20). 325--337.
[8]
F. Bernardini, J. Mittleman, H. Rushmeier, C. Silva, and G. Taubin. 1999. The ball-pivoting algorithm for surface reconstruction. IEEE Transactions on Visualization and Computer Graphics 5, 4 (1999), 349--359. https://doi.org/10.1109/2945.817351
[9]
Paul J Besl and Neil D McKay. 1992. Method for registration of 3-D shapes. In Sensor fusion IV: control paradigms and data structures, Vol. 1611. Spie, 586--606.
[10]
Dan Cernea. 2020. OpenMVS: Multi-View Stereo Reconstruction Library. (2020). https://cdcseacave.github.io/openMVS
[11]
Angel Chang, Angela Dai, Thomas Funkhouser, Maciej Halber, Matthias Niebner, Manolis Savva, Shuran Song, Andy Zeng, and Yinda Zhang. 2017. Matterport3D: Learning from RGB-D Data in Indoor Environments. In 2017 International Conference on 3D Vision (3DV). IEEE Computer Society, 667--676.
[12]
Dachuan Cheng, Jian Shi, Yanyun Chen, Xiaoming Deng, and Xiaopeng Zhang. 2018. Learning Scene Illumination by Pairwise Photos from Rear and Front Mobile Cameras. Comput. Graph. Forum 37, 7 (2018), 213--221. http://dblp.uni-trier.de/db/journals/cgf/cgf37.html#ChengSCDZ18
[13]
RidgeRun Embedded Linux Developer Connection. 2022. NVIDIA CUDA Memory Management. https://developer.ridgerun.com/wiki/index.php?title=NVIDIA_CUDA_Memory_Management.
[14]
Massimiliano Corsini, Marco Callieri, and Paolo Cignoni. 2008. Stereo light probe. In Computer Graphics Forum, Vol. 27. Wiley Online Library, 291--300.
[15]
Paul Debevec. 2006. Image-based lighting. In ACM SIGGRAPH 2006 Courses. 4-es.
[16]
Paul Debevec. 2008. Rendering synthetic objects into real scenes: Bridging traditional and image-based graphics with global illumination and high dynamic range photography. In ACM SIGGRAPH 2008 classes. 1--10.
[17]
Devin Larson. 2014. Standard Proportions of the Human Body. https://www.makingcomics.com/2014/01/19/standard-proportions-human-body/. Accessed: 2021-11-5.
[18]
Ufuk Dilek and Mustafa Erol. 2018. Detecting position using ARKit II: generating position-time graphs in real-time and further information on limitations of ARKit. Physics Education 53, 3 (2018), 035020.
[19]
Ruofei Du, Eric Turner, Maksym Dzitsiuk, Luca Prasso, Ivo Duarte, Jason Dourgarian, Joao Afonso, Jose Pascoal, Josh Gladstone, Nuno Cruces, Shahram Izadi, Adarsh Kowdle, Konstantine Tsotsos, and David Kim. 2020. DepthLab: Real-Time 3D Interaction With Depth Maps for Mobile Augmented Reality. In Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology (UIST). ACM, 15 pages. https://doi.org/10.1145/3379337.3415881
[20]
Farshad Einabadi, Jean-Yves Guillemaut, and Adrian Hilton. 2021. Deep neural models for illumination estimation and relighting: A survey. Comput. Graph. Forum 40, 6 (Sept. 2021), 315--331.
[21]
Epic Games. 2021. Unreal Engine - Real-Time 3D Creation Tool. https://www.unrealengine.com. Accessed: 2021-11-5.
[22]
Marc-André Gardner, Kalyan Sunkavalli, Ersin Yumer, Xiaohui Shen, Emiliano Gambaretto, Christian Gagné, and Jean-François Lalonde. 2017. Learning to Predict Indoor Illumination from a Single Image. ACM Transactions on Graphics (2017).
[23]
Marc-Andre Gardner, Yannick Hold-Geoffroy, Kalyan Sunkavalli, Christian Gagne, and Jean-Francois Lalonde. 2019. Deep Parametric Indoor Lighting Estimation.
[24]
Mathieu Garon, Kalyan Sunkavalli, Sunil Hadap, Nathan Carr, and Jean-François Lalonde. 2019. Fast Spatially-Varying Indoor Lighting Estimation. CVPR (2019).
[25]
Google. 2020. ARCore. https://developers.google.com/ar.
[26]
Google. 2022. Pixel 6 Tech Specs. https://store.google.com/product/pixel_6_specs?hl=en-US.
[27]
Google for Education. 2022. Bringing virtual and augmented reality to school | Google for Education. https://edu.google.com/products/vrar/?modal_active=none. Accessed: 2020-7-24.
[28]
Thorsten Grosch, Tobias Eble, and Stefan Mueller. 2007. Consistent interactive augmentation of live camera images with correct near-field illumination. In Proceedings of the 2007 ACM symposium on Virtual reality software and technology. 125--132.
[29]
Charles R. Harris, K. Jarrod Millman, Stéfan J. van der Walt, Ralf Gommers, Pauli Virtanen, David Cournapeau, Eric Wieser, Julian Taylor, Sebastian Berg, Nathaniel J. Smith, Robert Kern, Matti Picus, Stephan Hoyer, Marten H. van Kerkwijk, Matthew Brett, Allan Haldane, Jaime Fernández del Río, Mark Wiebe, Pearu Peterson, Pierre Gérard-Marchant, Kevin Sheppard, Tyler Reddy, Warren Weckesser, Hameer Abbasi, Christoph Gohlke, and Travis E. Oliphant. 2020. Array programming with NumPy. Nature 585, 7825 (Sept. 2020), 357--362. https://doi.org/10.1038/s41586-020-2649-2
[30]
Vlastimil Havran, Miloslaw Smyk, Grzegorz Krawczyk, Karol Myszkowski, and Hans-Peter Seidel. 2005. Interactive System for Dynamic Scene Lighting using Captured Video Environment Maps. In Rendering Techniques. 31--42.
[31]
HUAWEI. 2021. HUAWEI Mate 30 Pro Specifications | HUAWEI Global. https://consumer.huawei.com/en/phones/mate30-pro/specs/. Accessed: 2020-7-8.
[32]
Apple Inc. 2020. iPad Pro 2020. https://www.apple.com/ipad-pro/specs/.
[33]
Apple Inc. 2022. Adding Realistic Reflections to an AR Experience. https://developer.apple.com/documentation/arkit/camera_lighting_and_effects/adding_realistic_reflections_to_an_ar_experience.
[34]
Apple Inc. 2022. Introducing ARKit 5. https://developer.apple.com/augmented-reality/arkit/.
[35]
Inter IKEA Systems B. V. 2017. IKEA Place. https://apps.apple.com/us/app/ikea-place/id1279244498. Accessed: 2020-7-2.
[36]
Brian Karis and Epic Games. 2013. Real shading in unreal engine 4. Proc. Physically Based Shading Theory Practice 4, 3 (2013).
[37]
Siu Kwan Lam, Antoine Pitrou, and Stanley Seibert. 2015. Numba: A LLVM-Based Python JIT Compiler. In Proceedings of the Second Workshop on the LLVM Compiler Infrastructure in HPC (Austin, Texas) (LLVM '15). Association for Computing Machinery, New York, NY, USA, Article 7, 6 pages. https://doi.org/10.1145/2833157.2833162
[38]
Junxuan Li, Hongdong Li, and Yasuyuki Matsushita. 2021. Lighting, Reflectance and Geometry Estimation From 360deg Panoramic Stereo. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 10591--10600.
[39]
Zhengqin Li, Mohammad Shafiei, Ravi Ramamoorthi, Kalyan Sunkavalli, and Manmohan Chandraker. 2020. Inverse rendering for complex indoor scenes: Shape, spatially-varying lighting and svbrdf from a single image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2475--2484.
[40]
Z Liu, G Lan, J Stojkovic, Y Zhang, C Joe-Wong, and M Gorlatova. 2020. CollabAR: Edge-assisted Collaborative Image Recognition for Mobile Augmented Reality. In 2020 19th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN). 301--312.
[41]
Robert Maier, Kihwan Kim, Daniel Cremers, Jan Kautz, and Matthias Nießner. 2017. Intrinsic3D: High-Quality 3D Reconstruction by Joint Appearance and Geometry Optimization with Spatially-Varying Lighting. (Aug. 2017). arXiv:1708.01670 [cs.CV]
[42]
Morgan McGuire and Michael Mara. 2014. Efficient GPU screen-space ray tracing. Journal of Computer Graphics Techniques (JCGT) 3, 4 (2014), 73--85.
[43]
Pierre Moulon, Pascal Monasse, Romuald Perrot, and Renaud Marlet. 2016. OpenMVG: Open multiple view geometry. In International Workshop on Reproducible Research in Pattern Recognition. Springer, 60--74.
[44]
Raúl Mur-Artal and Juan D. Tardós. 2017. ORB-SLAM2: an Open-Source SLAM System for Monocular, Stereo and RGB-D Cameras. IEEE Transactions on Robotics 33, 5 (2017), 1255--1262. https://doi.org/10.1109/TRO.2017.2705103
[45]
Nvidia. 2022. Jetson AGX Xavier Developer Kit. https://developer.nvidia.com/embedded/jetson-agx-xavier-developer-kit.
[46]
Rohit Pandey, Sergio Orts Escolano, Chloe Legendre, Christian Häne, Sofien Bouaziz, Christoph Rhemann, Paul Debevec, and Sean Fanello. 2021. Total relighting: learning to relight portraits for background replacement. ACM Trans. Graph. 40, 4 (July 2021), 1--21.
[47]
Siddhant Prakash, Alireza Bahremand, Linda D Nguyen, and Robert LiKamWa. 2019. Gleam: An illumination estimation framework for real-time photorealistic augmented reality on mobile devices. In Proceedings of the 17th Annual International Conference on Mobile Systems, Applications, and Services. 142--154.
[48]
René Ranftl, Katrin Lasinger, David Hafner, Konrad Schindler, and Vladlen Koltun. 2020. Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) (2020).
[49]
Science Buddies. 2013. Stepping Science: Estimating Someone's Height from Their Walk. https://www.scientificamerican.com/article/bring-science-home-estimating-height-walk/. Accessed: 2021-11-5.
[50]
Scott Stein. 2021. Lidar is one of the iPhone and iPad's coolest tricks. Here's what else it can do. https://www.cnet.com/tech/mobile/lidar-is-one-of-the-iphone-ipad-coolest-tricks-its-only-getting-better/. Accessed: 2021-11-5.
[51]
Gowri Somanath and Daniel Kurz. 2021. HDR Environment Map Estimation for Real-Time Augmented Reality. CVPR (2021).
[52]
Shuran Song and Thomas Funkhouser. 2019. Neural Illumination: Lighting Prediction for Indoor Environments. CVPR (2019).
[53]
Pratul P. Srinivasan, Ben Mildenhall, Matthew Tancik, Jonathan T. Barron, Richard Tucker, and Noah Snavely. 2020. Lighthouse: Predicting Lighting Volumes for Spatially-Coherent Illumination. In CVPR. 8077--8086. https://doi.org/10.1109/CVPR42600.2020.00810
[54]
Tiancheng Sun, Jonathan T Barron, Yun-Ta Tsai, Zexiang Xu, Xueming Yu, Graham Fyffe, Christoph Rhemann, Jay Busch, Paul Debevec, and Ravi Ramamoorthi. 2019. Single image portrait relighting. ACM Trans. Graph. 38, 4 (July 2019), 1--12.
[55]
Three.js Organization. 2021. Three.js - JavaScript 3D library. https://threejs.org. Accessed: 2021-11-5.
[56]
TornadoWeb. 2022. Tornado Web Server. https://www.tornadoweb.org/en/stable/.
[57]
Mihran Tuceryan et al. 2019. AR360: dynamic illumination for augmented reality with real-time interaction. In 2019 IEEE 2nd International Conference on Information and Computer Technologies (ICICT). IEEE, 170--174.
[58]
Jonas Unger, Joel Kronander, Per Larsson, Stefan Gustavson, and Anders Ynnerman. 2013. Temporally and spatially varying image based lighting using HDR-video. In 21st European Signal Processing Conference (EUSIPCO 2013). IEEE, Marrakech, Morocco, 1--5.
[59]
Unity. 2020. AR Foundation 4.2.0-preview.5. https://docs.unity3d.com/Packages/[email protected]/manual/index.html.
[60]
Unity3D. 2022. Unity3D. https://docs.unity3d.com/. Accessed: 2022-5-14.
[61]
Jingao Xu, Guoxuan Chi, Zheng Yang, Danyang Li, Qian Zhang, Qiang Ma, and Xin Miao. 2021. FollowUpAR: enabling follow-up effects in mobile AR applications. In Proceedings of the 19th Annual International Conference on Mobile Systems, Applications, and Services (Virtual Event, Wisconsin) (MobiSys '21). Association for Computing Machinery, New York, NY, USA, 1--13.
[62]
Juheon Yi and Youngki Lee. 2020. Heimdall: mobile GPU coordination platform for augmented reality applications. In Proceedings of the 26th Annual International Conference on Mobile Computing and Networking (London, United Kingdom) (MobiCom '20, Article 35). Association for Computing Machinery, New York, NY, USA, 1--14.
[63]
Yunfan Zhang, Tim Scargill, Ashutosh Vaishnav, Gopika Premsankar, Mario Di Francesco, and Maria Gorlatova. 2022. InDepth: Real-Time Depth Inpainting for Mobile Augmented Reality. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 6, 1, Article 37 (mar 2022), 25 pages. https://doi.org/10.1145/3517260
[64]
Yiqin Zhao and Tian Guo. 2020. PointAR: Efficient Lighting Estimation for Mobile Augmented Reality. In Computer Vision - ECCV 2020, Andrea Vedaldi, Horst Bischof, Thomas Brox, and Jan-Michael Frahm (Eds.). Springer International Publishing, Cham, 678--693.
[65]
Yiqin Zhao and Tian Guo. 2021. Xihe: A 3D Vision-Based Lighting Estimation Framework for Mobile Augmented Reality. In Proceedings of the 19th Annual International Conference on Mobile Systems, Applications, and Services (MobiSys '21). 28--40.

Cited By

View all
  • (2024)Acceptance of augmented reality for laboratory safety training: methodology and an evaluation studyFrontiers in Virtual Reality10.3389/frvir.2024.13225435Online publication date: 26-Apr-2024
  • (2024)Can Large Language Models Be Good Companions?Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/36596008:2(1-41)Online publication date: 15-May-2024
  • (2024)Systematic Review of Extended Reality for Smart Built Environments Lighting Design SimulationsIEEE Access10.1109/ACCESS.2024.335916712(17058-17089)Online publication date: 2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies
Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies  Volume 6, Issue 3
September 2022
1612 pages
EISSN:2474-9567
DOI:10.1145/3563014
Issue’s Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 07 September 2022
Published in IMWUT Volume 6, Issue 3

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. 3D vision
  2. lighting estimation
  3. mobile augmented reality

Qualifiers

  • Research-article
  • Research
  • Refereed

Funding Sources

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)219
  • Downloads (Last 6 weeks)27
Reflects downloads up to 12 Sep 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Acceptance of augmented reality for laboratory safety training: methodology and an evaluation studyFrontiers in Virtual Reality10.3389/frvir.2024.13225435Online publication date: 26-Apr-2024
  • (2024)Can Large Language Models Be Good Companions?Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/36596008:2(1-41)Online publication date: 15-May-2024
  • (2024)Systematic Review of Extended Reality for Smart Built Environments Lighting Design SimulationsIEEE Access10.1109/ACCESS.2024.335916712(17058-17089)Online publication date: 2024
  • (2023)A Socioenactive Perspective on Emotion Contagion and Senses: Analyzing the Phenomenon via EEG SignalsProceedings of the XXII Brazilian Symposium on Human Factors in Computing Systems10.1145/3638067.3638132(1-11)Online publication date: 16-Oct-2023
  • (2023)Toward Scalable and Controllable AR ExperimentationProceedings of the 1st ACM Workshop on Mobile Immersive Computing, Networking, and Systems10.1145/3615452.3617941(237-246)Online publication date: 6-Oct-2023
  • (2023)CASESProceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies10.1145/36109107:3(1-31)Online publication date: 27-Sep-2023
  • (2023)Audio Augmented Reality OutdoorsProceedings of the 2023 ACM International Conference on Interactive Media Experiences10.1145/3573381.3597028(199-204)Online publication date: 12-Jun-2023
  • (2023)Multi-Camera Lighting Estimation for Photorealistic Front-Facing Mobile Augmented RealityProceedings of the 24th International Workshop on Mobile Computing Systems and Applications10.1145/3572864.3580337(68-73)Online publication date: 22-Feb-2023
  • (2023)Evaluating Light Probe Estimation Techniques for Mobile Augmented Reality2023 International Conference on Cyberworlds (CW)10.1109/CW58918.2023.00029(141-148)Online publication date: 3-Oct-2023

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Get Access

Login options

Full Access

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media