skip to main content
research-article

Towards efficient context-specific video coding based on gaze-tracking analysis

Published:12 December 2007Publication History
Skip Abstract Section

Abstract

This article discusses a framework for model-based, context-dependent video coding based on exploitation of characteristics of the human visual system. The system utilizes variable-quality coding based on priority maps which are created using mostly context-dependent rules. The technique is demonstrated through two case studies of specific video context, namely open signed content and football sequences. Eye-tracking analysis is employed for identifying the characteristics of each context, which are subsequently exploited for coding purposes, either directly or through a gaze prediction model. The framework is shown to achieve a considerable improvement in coding efficiency.

References

  1. Agrafiotis, D., Canagarajah, N., Bull, D. R., Dye, M., Twyford, H., Kyle, J., and Chung-How, J. 2003. Optimised sign language video coding based on eye-tracking analysis. In Proceedings of the International Conference on Visual Communications and Image Processing, Lugano, Switzerland.Google ScholarGoogle Scholar
  2. Agrafiotis, D., Canagarajah, N., Bull D. R., Kyle, J., Seers, H., and Dye, M. 2006. A perceptually optimised video coding system for sign language communication at low bit rates. Signal Proc. Image Commun. 21, 531--549Google ScholarGoogle ScholarCross RefCross Ref
  3. Appleby, S., Crabtree, B., Jeffery, R., Mulroy, P., and Nilsson, M. 2006. Video coding and delivery challenges for next generation IPTV. BT Technol. J. 24, 174--179. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Chen, M. J., Chi, M. C., Hsu, C. T., and Chen, J. W. 2003. ROI video coding based on H.263+ with robust skin-color detection technique. IEEE Trans. Consum. Electron. 49, 3. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Cheng, W., Chu, W., and Wu, J. 2005. A visual attention based region-of-interest determination framework for video sequences. IEICE Trans. Inf. Syst. E88-D, 1578--1586. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Crabtree, B. 2006. Video compression using focus of attention. In Proceedings of the Picture Coding Symposium, China.Google ScholarGoogle Scholar
  7. Daly, S., Matthews, K., and Ribas-Corbera, J. 1998. Face-based visually-optimized image sequence coding. In Proceedings of the International Conference on Image Processing, Chicago.Google ScholarGoogle Scholar
  8. Geisler, W. S. and Perrys, J. S. 1998. A real-time foveated multiresolution system for low-bandwidth video communication. In Proceedings of the SPIE Conference on Human Vision and Electronic Imaging, 3299.Google ScholarGoogle Scholar
  9. ISO/IEC. 2003. ISO/IEC 14496--10 and ITU-T recommendation H.264 2003. Coding of audiovisual objects---Part 10.Google ScholarGoogle Scholar
  10. ISO/IEC. 2000. ISO/IEC 14496--2 2000. Information technology-Coding of audiovisual objects-Part 2:Visual.Google ScholarGoogle Scholar
  11. Itti, L. 2004. Automatic foveation for video compression using a neurobiological model of visual attention. IEEE Trans. Image Proc. 13, 1304--1318. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Itti, L. and Baldi, P. 2005a. Bayesian surprise attracts human attention. Adv. Neural Inf. Proc. Syst. 19, 1--8.Google ScholarGoogle Scholar
  13. Itti, L. and Baldi, P. 2005b. A principled approach to detecting surprising events in video. In Proceedings of the International Conference on Computer Vision and Pattern Recognition, San Diego. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Itti, L., Koch, C., and Niebur, E. 1998. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Analy. Mach. Intell. 20, 1254--1259. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Itu 1995. R BT.500-11 2002. Methodology for the subjective assessment of the quality of television pictures.Google ScholarGoogle Scholar
  16. Lee, S. and Bovik, A. C. 2003. Fast algorithms for foveated video processing. IEEE Trans. Circ. Syst. Video Technol. 13, 149--162. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Lin, C. W., Chang, Y. Y., and Chen, Y. C. 2000. Low-Complexity face-assisted video coding. In Proceedings of the International Conference on Image Processing, Vancouver, Canada.Google ScholarGoogle Scholar
  18. Liu, Y., Li, Z. G., Soh, Y. C., and Loke, M. H. 2006. Conversational video communication of H.264/AVC with region-of-interest concern. In Proceedings of the International Conference on Image Processing, Atlanta, GA.Google ScholarGoogle Scholar
  19. Muir, L., Richardson, I., and Leaper, S. 2003. Gaze tracking and its application to video coding for sign language. In Proceedings of the Picture Coding Symposium, Saint Malo, France.Google ScholarGoogle Scholar
  20. Nadenau, M. J., Reichel, J., and Kunt, M. 2002. Performance comparison of masking models based on a new psychovisual test method with natural scenery stimuli. Signal Proc. Image Commun. 17, 807--823.Google ScholarGoogle ScholarCross RefCross Ref
  21. Privitera, C. and Stark, L. W. 1997. Algorithms for defining visual region-of-interest: Comparison with eye fixations. Tech. Rep. UCB/ERL M97/72, Electrical Engineering and Computer Science, Department, University of California, Berkeley.Google ScholarGoogle Scholar
  22. Tang, C. W., Chen, C. H., Yu, Y. H., and Tsai, C. J. 2006. Visual sensitivity guided bit allocation for video coding. IEEE Trans. Multimedia 8, 11--18. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Towards efficient context-specific video coding based on gaze-tracking analysis

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader
    About Cookies On This Site

    We use cookies to ensure that we give you the best experience on our website.

    Learn more

    Got it!