skip to main content
research-article

Video Extrapolation Using Neighboring Frames

Authors Info & Claims
Published:08 April 2019Publication History
Skip Abstract Section

Abstract

With the popularity of immersive display systems that fill the viewer’s field of view (FOV) entirely, demand for wide FOV content has increased. A video extrapolation technique based on reuse of existing videos is one of the most efficient ways to produce wide FOV content. Extrapolating a video poses a great challenge, however, due to the insufficient amount of cues and information that can be leveraged for the estimation of the extended region. This article introduces a novel framework that allows the extrapolation of an input video and consequently converts a conventional content into one with wide FOV. The key idea of the proposed approach is to integrate the information from all frames in the input video into each frame. Utilizing the information from all frames is crucial because it is very difficult to achieve the goal with a two-dimensional transformation based approach when parallax caused by camera motion is apparent. Warping guided by three-dimensnional scene points matches the viewpoints between the different frames. The matched frames are blended to create extended views. Various experiments demonstrate that the results of the proposed method are more visually plausible than those produced using state-of-the-art techniques.

References

  1. Aseem Agarwala, Ke Colin Zheng, Chris Pal, Maneesh Agrawala, Michael Cohen, Brian Curless, David Salesin, and Richard Szeliski. 2005. Panoramic video textures. In ACM Transactions on Graphics, Vol. 24. ACM, 821--827. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Amit Aides, Tamar Avraham, and Yoav Y. Schechner. 2011. Multiscale ultrawide foveated video extrapolation. In Proceedings of the 2011 IEEE International Conference on Computational Photography (ICCP’11). IEEE, 1--8.Google ScholarGoogle Scholar
  3. Tamar Avraham and Yoav Y. Schechner. 2011. Ultrawide foveated video extrapolation. IEEE J. Select. Top. Sign. Process. 5, 2 (2011), 321--334.Google ScholarGoogle ScholarCross RefCross Ref
  4. Barco. 2014. Barco Escape. Retrieved from http://www.ready2escape.com.Google ScholarGoogle Scholar
  5. Matthew Brown and David G. Lowe. 2007. Automatic panoramic image stitching using invariant features. Int. J. Comput. Vis. 74, 1 (2007), 59--73. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Che-Han Chang, Yuuki Sato, and Yung-Yu Chuang. 2014. Shape-preserving half-projective warps for image stitching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3254--3261. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Sunghyun Cho, Jue Wang, and Seungyong Lee. 2012. Video deblurring for hand-held cameras using patch-based synthesis. ACM Trans. Graph. 31, 4 (2012), 64. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Antonio Criminisi, Patrick Pérez, and Kentaro Toyama. 2004. Region filling and object removal by exemplar-based image inpainting. IEEE Trans. Image Process. 13, 9 (2004), 1200--1212. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Carolina Cruz-Neira, Daniel J. Sandin, and Thomas A. DeFanti. 1993. Surround-screen projection-based virtual reality: The design and implementation of the CAVE. In Proceedings of the 20th Annual Conference on Computer Graphics and Interactive Techniques. ACM, 135--142. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Andrew J. Davison, Ian D. Reid, Nicholas D. Molton, and Olivier Stasse. 2007. MonoSLAM: Real-time single camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 29, 6 (2007). Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Abhijeet Ghosh, Matthew Trentacoste, Helge Seetzen, and Wolfgang Heidrich. 2005. Real illumination from virtual environments. In ACM SIGGRAPH 2005 Sketches. ACM, 41. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Miguel Granados, Kwang Kim, James Tompkin, Jan Kautz, and Christian Theobalt. 2012. Background inpainting for videos with dynamic objects and a free-moving camera. In Proceedings of the Computer Vision (ECCV’12). 682--695. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Jan Herling and Wolfgang Broll. 2014. High-quality real-time video inpaintingwith PixMix. IEEE Trans. Vis. Comput. Graph. 20, 6 (2014), 866--879. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Takeo Igarashi, Tomer Moscovich, and John F. Hughes. 2005. As-rigid-as-possible shape manipulation. ACM Trans. Graph. 24, 3 (2005), 1134--1141. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Wei Jiang and Jinwei Gu. 2015. Video stitching with spatial-temporal content-preserving warping. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. 42--48.Google ScholarGoogle ScholarCross RefCross Ref
  16. Brett Jones, Rajinder Sodhi, Michael Murdock, Ravish Mehra, Hrvoje Benko, Andrew Wilson, Eyal Ofek, Blair MacIntyre, Nikunj Raghuvanshi, and Lior Shapira. 2014. Roomalive: Magical experiences enabled by scalable, adaptive projector-camera units. In Proceedings of the 27th Annual ACM Symposium on User Interface Software and Technology. ACM, 637--644. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Brett R. Jones, Hrvoje Benko, Eyal Ofek, and Andrew D. Wilson. 2013. IllumiRoom: Peripheral projected illusions for interactive experiences. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, 869--878. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Georg Klein and David Murray. 2009. Parallel tracking and mapping on a camera phone. In Proceedings of the 8th IEEE International Symposium on Mixed and Augmented Reality (ISMAR’09). IEEE, 83--86. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Felix Klose, Oliver Wang, Jean-Charles Bazin, Marcus Magnor, and Alexander Sorkine-Hornung. 2015. Sampling based scene-space video processing. ACM Trans. Graph. 34, 4 (2015), 67. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Johannes Kopf, Michael F. Cohen, and Richard Szeliski. 2014. First-person hyper-lapse videos. ACM Trans. Graph. 33, 4 (2014), 78. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Jungjin Lee, Bumki Kim, Kyehyun Kim, Younghui Kim, and Junyong Noh. 2016. Rich360: Optimized spherical representation from structured panoramic camera arrays. ACM Trans. Graph. 35, 4 (2016), 63. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Jungjin Lee, Sangwoo Lee, Younghui Kim, and Junyong Noh. 2017. ScreenX: Public immersive theatres with uniform movie viewing experiences. IEEE Trans. Vis. Comput. Graph. 23, 2 (2017), 1124--1138. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Shiwei Li, Lu Yuan, Jian Sun, and Long Quan. 2015. Dual-feature warping-based motion model estimation. In Proceedings of the IEEE International Conference on Computer Vision. 4283--4291. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Jing Liao, Mark Finch, and Hugues Hoppe. 2015. Fast computation of seamless video loops. ACM Trans. Graph. 34, 6 (2015), 197. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Chung-Ching Lin, Sharathchandra U. Pankanti, Karthikeyan Natesan Ramamurthy, and Aleksandr Y. Aravkin. 2015. Adaptive as-natural-as-possible image stitching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1155--1163.Google ScholarGoogle Scholar
  26. Kaimo Lin, Shuaicheng Liu, Loong-Fah Cheong, and Bing Zeng. 2016. Seamless video stitching from hand-held camera inputs. In Computer Graphics Forum, Vol. 35. Wiley Online Library, 479--487.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Wen-Yan Lin, Siying Liu, Yasuyuki Matsushita, Tian-Tsong Ng, and Loong-Fah Cheong. 2011. Smoothly varying affine stitching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 345--352. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Feng Liu, Michael Gleicher, Hailin Jin, and Aseem Agarwala. 2009. Content-preserving warps for 3D video stabilization. ACM Trans. Graph. 28, 3 (2009), 44. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. Abhimitra Meka, Michael Zollhöfer, Christian Richardt, and Christian Theobalt. 2016. Live intrinsic video. ACM Trans. Graph. 35, 4 (2016), 109. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Peter Mills, Alia Sheikh, Graham Thomas, and Paul Debenham. 2011. BBC surround video. Retrieved from http://www.bbc.com/news/magazine-22315685.Google ScholarGoogle Scholar
  31. Alasdair Newson, Andrés Almansa, Matthieu Fradet, Yann Gousseau, and Patrick Pérez. 2014. Video inpainting of complex scenes. SIAM J. Imag. Sci. 7, 4 (2014), 1993--2019.Google ScholarGoogle ScholarCross RefCross Ref
  32. Daniel Novy and Bove V. Michael. 2012. Infinity by nine. Retrieved from http://video.mit.edu/watch/infinity-by-nine-11565/.Google ScholarGoogle Scholar
  33. F. Perazzi, A. Sorkine-Hornung, H. Zimmer, P Kaufmann, O. Wang, S. Watson, and M. Gross. 2015. Panoramic video from unstructured camera arrays. Comput. Graph. Forum 34, 2 (2015), 57--68. Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Wen Qu, Yifei Zhang, Daling Wang, Shi Feng, and Ge Yu. 2015. Semantic movie summarization based on string of IE-RoleNets. Comput. Vis. Media 1, 2 (2015), 129--141.Google ScholarGoogle ScholarCross RefCross Ref
  35. Arno Schödl, Richard Szeliski, David H. Salesin, and Irfan Essa. 2000. Video textures. In Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques. ACM Press/Addison-Wesley Publishing Co., 489--498. Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. Heung-Yeung Shum and Richard Szeliski. 2000. Systems and experiment paper: Construction of panoramic image mosaics with global and local alignment. Int. J. Comput. Vis. 36, 2 (2000), 101--130. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Jaewon Song. 2017. Personal communication with jaewon song, CTO, Digital Idea Corp. (2017).Google ScholarGoogle Scholar
  38. Richard Szeliski. 2006. Image alignment and stitching: A tutorial. Found. Trends Comput. Graph. Vis. 2, 1 (2006), 1--104. Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. Laura Turban, Fabrice Urban, and Philippe Guillotel. 2017. Extrafoveal video extension for an immersive viewing experience. IEEE Trans. Visu. Comput. Graph. 23, 5 (2017), 1520--1533. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Miao Wang, Yukun Lai, Yuan Liang, Ralph Robert Martin, and Shi-Min Hu. 2014. BiggerPicture: Data-driven image extrapolation using graph matching. ACM Trans. Graph. 33, 6 (2014). Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. A Weffers-Albu, S. de Waele, W. Hoogenstraaten, and C. Kwisthout. 2011. Immersive TV viewing with advanced ambilight. In Proceedings of the 2011 IEEE International Conference on Consumer Electronics (ICCE’11). IEEE, 753--754.Google ScholarGoogle Scholar
  42. Jordi Zaragoza, Tat-Jun Chin, Quoc-Huy Tran, Michael S. Brown, and David Suter. 2014. As-projective-as-possible image stitching with moving DLT. IEEE Trans. Pattern Anal. Mach. Intell. 36, 7 (2014), 1285--1298. Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. Fan Zhang and Feng Liu. 2014. Parallax-tolerant image stitching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3262--3269. Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. Fang-Lue Zhang, Jue Wang, Han Zhao, Ralph R. Martin, and Shi-Min Hu. 2015. Simultaneous camera path optimization and distraction removal for improving amateur video. IEEE Trans. Image Process. 24, 12 (2015), 5982--5994.Google ScholarGoogle ScholarCross RefCross Ref
  45. Fang-Lue Zhang, Xian Wu, Hao-Tian Zhang, Jue Wang, and Shi-Min Hu. 2016. Robust background identification for dynamic video editing. ACM Trans. Graph. 35, 6 (2016), 197. Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Yinda Zhang, Jianxiong Xiao, James Hays, and Ping Tan. 2013. Framebreak: Dramatic image extrapolation by guided shift-maps. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 1171--1178. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Video Extrapolation Using Neighboring Frames

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Graphics
        ACM Transactions on Graphics  Volume 38, Issue 3
        June 2019
        125 pages
        ISSN:0730-0301
        EISSN:1557-7368
        DOI:10.1145/3322934
        Issue’s Table of Contents

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 8 April 2019
        • Revised: 1 January 2019
        • Accepted: 1 January 2019
        • Received: 1 November 2017
        Published in tog Volume 38, Issue 3

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format