ABSTRACT
We propose a system, called ExtVision, to augment visual experiences by generating and projecting context-images onto the periphery of the television or computer screen. A peripheral projection of the context-image is one of the most effective techniques to enhance visual experiences. However, the projection is not commonly used at present, because of the difficulty in preparing the context-image. In this paper, we propose a deep neural network-based method to generate context-images for peripheral projection. A user study was performed to investigate the manner in which the proposed system augments traditional visual experiences. In addition, we present applications and future prospects of the developed system.
References
- Amit Aides, Tamar Avraham, and Yoav Y. Schechner. 2011. Multiscale ultrawide foveated video extrapolation. In 2011 IEEE International Conference on Computational Photography (ICCP), 1--8.Google Scholar
- Tamar Avraham and Yoav Y. Schechner. 2011. Ultrawide Foveated Video Extrapolation. IEEE Journal of Selected Topics in Signal Processing 5, 2: 321--334.Google Scholar
Cross Ref
- Patrick Baudisch, Nathaniel Good, and Paul Stewart. 2001. Focus plus context screens. In Proceedings of the 14th annual ACM symposium on User interface software and technology - UIST '01, 31.Google Scholar
Digital Library
- Carolina Cruz-Neira, Daniel J. Sandin, and Thomas A. DeFanti. 1993. Surround-screen projection-based virtual reality. In Proceedings of the 20th annual conference on Computer graphics and interactive techniques - SIGGRAPH '93, 135--142. Google Scholar
Digital Library
- Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Figure 11. Percentage of each answer to some questions of "Drama". Courville, and Yoshua Bengio. 2014. Generative Adversarial Nets. Advances in Neural Information Processing Systems 27: 2672--2680. Google Scholar
Digital Library
- Yixiang Huang, Xuan F. Zha, Jay Lee, and Chengliang Liu. 2013. Discriminant diffusion maps analysis: A robust manifold learner for dimensionality reduction and its applications in machine condition monitoring and fault diagnosis. Mechanical Systems and Signal Processing 34, 1--2: 277--297.Google Scholar
Cross Ref
- Satoshi Iizuka, Edgar Simo-Serra, and Hiroshi Ishikawa. 2017. Globally and Locally Consistent Image Completion. ACM Transactions on Graphics ACM Trans. Graph. Article 36, 13. Google Scholar
Digital Library
- Satoshi Iizuka, Edgar Simo-Serra, and Hiroshi Ishikawa. 2016. Let there be Color!: Joint End-to-end Learning of Global and Local Image Priors for Automatic Image Colorization with Simultaneous Classification. ACM Transactions on Graphics (Proc. of SIGGRAPH 2016) 35, 4. Google Scholar
Digital Library
- Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A. Efros. 2017. Image-to-Image Translation with Conditional Adversarial Networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 5967--5976.Google Scholar
- Brett R. Jones, Hrvoje Benko, Eyal Ofek, and Andrew D. Wilson. 2013. IllumiRoom: Peripheral Projected Illusions for Interactive Experiences. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems - CHI '13: 869. Google Scholar
Digital Library
- Mehdi Mirza and Simon Osindero. 2014. Conditional Generative Adversarial Nets. Retrieved January 9, 2018 from http://arxiv.org/abs/1411.1784Google Scholar
- Daniel E Novy. 2013. COMPUTATIONAL IMMERSIVE DISPLAYS. Retrieved January 9, 2018 from http://excedrin.media.mit.edu/wpcontent/uploads/sites/10/2013/07/novyms.pdfGoogle Scholar
- Deepak Pathak, Philipp Krahenbuhl, Jeff Donahue, Trevor Darrell, and Alexei A. Efros. 2016. Context Encoders: Feature Learning by Inpainting.Google Scholar
- Alec Radford, Luke Metz, and Soumith Chintala. 2015. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks.Google Scholar
- Recommendation Itu-. Methodology for the subjective assessment of video quality in multimedia applications. Retrieved January 9, 2018 from https://www.itu.int/dms_pubrec/itu-r/rec/bt/R-RECBT.1788-0--200701-I!!PDF-E.pdfGoogle Scholar
- Laura Turban, Fabrice Urban, and Philippe Guillotel. 2017. Extrafoveal Video Extension for an Immersive Viewing Experience. IEEE Transactions on Visualization and Computer Graphics 23, 5: 1520-- 1533. Google Scholar
Digital Library
- Saiwen Wang, Jie Song, Jaime Lien, Ivan Poupyrev, and Otmar Hilliges. 2016. Interacting with Soli: Exploring Fine-Grained Dynamic Gesture Recognition in the Radio-Frequency Spectrum. Google Scholar
Digital Library
- Kentaro Yoshida, Seki Inoue, Yasutoshi Makino, and Hiroyuki Shinoda. 2017. VibVid: VIBration Estimation from VIDeo by using Neural Network. 37--44.Google Scholar
- Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. 2017. Places: A 10 million Image Database for Scene Recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence: 1--1.Google Scholar
- SCREENX. Retrieved January 9, 2018 from https://screenx.co.kr/Google Scholar
- Ready 2 Escape -- The ultimate immersive cinema experience. Retrieved January 9, 2018 from https://ready2escape.com/Google Scholar
- Philips TV. Experience Ambilight | Philips. Retrieved January 9, 2018 from https://www.philips.co.uk/c-mso/televisions/p/ambilightGoogle Scholar
- Phillipi. 2017. phillipi/pix2pix. (December 2017). Retrieved January 8, 2018 from https://github.com/phillipi/pix2pixGoogle Scholar
Supplemental Material
Index Terms
ExtVision


Jun Rekimoto


Comments