skip to main content
research-article

The Replate

Published:25 July 2018Publication History
Skip Abstract Section

Abstract

Replates are a new way to experience replays of sporting events. We use computer vision tools, such as camera tracking and background subtraction, to align all frames in the original video with regard to each other. Using this alignment, we create a single plate as it would be seen from a wide-angle static camera. Key players and events are then reinserted into this arena using an interactive tool, recovering the big picture that is typically lost due to camera panning and zoom. Replates are played in continuous loop. We allow replate authors to include multiple instances of each player, offset by regular intervals, so viewers can quickly and repeatedly inspect their favorite moments. Authors can also choose from a variety of effects that help them construct a narrative of their favourite plays. The interactive system gives immediate feedback allowing authors to adjust the rendered replates interactively. Replates take a fraction of the size of the original input video, and produce an effect reminiscent of video game reenactments of famous sporting events. The footage, however, is real.

References

  1. Aseem Agarwala, Ke Colin Zheng, Chris Pal, Maneesh Agrawala, Michael Cohen, Brian Curless, David Salesin, and Richard Szeliski. 2005. Panoramic Video Textures. ACM Transactions on Graphics (Proceedings of ACM SIGGRAPH 2005) (2005), 821--827. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Jackie Assa, Yaron Caspi, and Daniel Cohen-Or. 2006. Action Synopsis: Pose Selection and Illustration. ACM Transactions on Graphics (Proceedings of ACM SIGGRAPH 2005) (2006), 667--676. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Jiamin Bai, Aseem Agarwala, Maneesh Agrawala, and Ravi Ramamoorthi. 2012. Selectively De-animating Video. ACM Transactions on Graphics (Proceedings of ACM SIGGRAPH 2012) (2012), 66:1--66:10. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. O. Barnich and M. Van Droogenbroeck. 2011. ViBe: A Universal Background Subtraction Algorithm for Video Sequences. Trans. Img. Proc. 20, 6 (June 2011), 1709--1724. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Herbert Bay, Andreas Ess, Tinne Tuytelaars, and Luc Van Gool. 2008. Speeded-Up Robust Features (SURF). Comput. Vis. Image Underst. 110, 3 (June 2008), 346--359. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Matthew Brown and David G. Lowe. 2007. Automatic Panoramic Image Stitching Using Invariant Features. Int. J. Comput. Vision 74, 1 (Aug. 2007), 59--73. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Yaron Caspi, Anat Axelrod, Yasuyuki Matsushita, and Alon Gamliel. 2006. Dynamic Stills and Clip Trailers. The Visual Computer 22, 9 (2006), 642--652. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. J. P. Collomosse, D. Rowntree, and P. M. Hall. 2005. Rendering Cartoon-style Motion Cues in Post-production Video. Graph. Models 67, 6 (Nov. 2005), 549--564. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Carlos D. Correa and Kwan-Liu Ma. 2010. Dynamic Video Narratives. ACM Transactions on Graphics (Proceedings of ACM SIGGRAPH 2010) (2010), 88:1--88:9. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. M. Irani, P. Anandan, J. Bergen, R. Kumar, and S. Hsu. 1995. Mosaic Based Representations of Video Sequences and Their Applications. Signal Processing: Image Communication, special issue on Coding Techniques for Low Bit-rate Video (1995).Google ScholarGoogle Scholar
  11. M. Irani, S. Hsu, and P. Anandan. 1996. Efficient Representations of Video Sequences and Their Applications. Signal Processing: Image Communication, special issue on Image and Video Semantics: Processing, Analysis, and Application (1996).Google ScholarGoogle Scholar
  12. Neel Joshi, Sisil Mehta, Steven Drucker, Eric Stollnitz, Hugues Hoppe, Matt Uyttendaele, and Michael Cohen. 2012. Cliplets: Juxtaposing Still and Dynamic Imagery. In Proceedings of the 25th Annual ACM Symposium on User Interface Software and Technology (UIST '12). 251--260. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. B. Kim and I. Essa. 2005. Video-based Nonphotorealistic and Expressive Illustration of Motion. In Proceedings of the Computer Graphics International 2005 (CGI '05). 32--35. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Zicheng Liao, Neel Joshi, and Hugues Hoppe. 2013. Automated Video Looping with Progressive Dynamism. ACM Transactions on Graphics (Proceedings of ACM SIGGRAPH 2013) (2013), 77:1--77:10. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Manao, Fizick, Tsp, TSchniede, and SEt. 2004--2011. MVTools. http://avisynth.org.ru/mvtools/mvtools2.html. (2004-2011).Google ScholarGoogle Scholar
  16. Cuong Nguyen, Yuzhen Niu, and Feng Liu. 2012. Video Summagator: An Interface for Video Summarization and Navigation. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '12). 647--650. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Yael Pritch, Alex Rav-Acha, and Shmuel Peleg. 2008. Nonchronological Video Synopsis and Indexing. IEEE Trans. Pattern Anal. Mach. Intell. 30, 11 (Nov. 2008). Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Arno Schödl, Richard Szeliski, David H. Salesin, and Irfan Essa. 2013. Video Textures. In Proceedings of ACM SIGGRAPH 2013. 489--498.Google ScholarGoogle Scholar
  19. SportsVision. 2014. http://www.sportvision.com. (2014).Google ScholarGoogle Scholar
  20. Alexandru Telea. 2004. An Image Inpainting Technique Based on the Fast Marching Method. J. Graphics, GPU, 8 Game Tools 9, 1 (2004), 23--34.Google ScholarGoogle Scholar
  21. Laura Teodosio and Walter Bender. 2005. Salient Stills. ACM Trans. Multimedia Comput. Commun. Appl. 1, 1 (Feb. 2005), 16--36. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Graham Thomas. 2012. Sports TV Applications of Computer Vision. Technical Report. British Broadcasting Corporation.Google ScholarGoogle Scholar
  23. James Tompkin, Fabrizio Pece, Kartic Subr, and Jan Kautz. 2011. Towards Moment Imagery: Automatic Cinemagraphs. In Proceedings of the 2011 Conference for Visual Media Production (CVMP '11). 87--93. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Vizrt. 2014. http://www.vizrt.com/products/viz_libero/. (2014).Google ScholarGoogle Scholar

Index Terms

  1. The Replate

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader
      About Cookies On This Site

      We use cookies to ensure that we give you the best experience on our website.

      Learn more

      Got it!