Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Perception of blending in stereo motion panoramas

Published: 02 August 2012 Publication History

Abstract

Most methods for synthesizing panoramas assume that the scene is static. A few methods have been proposed for synthesizing stereo or motion panoramas, but there has been little attempt to synthesize panoramas that have both stereo and motion. One faces several challenges in synthesizing stereo motion panoramas, for example, to ensure temporal synchronization between left and right views in each frame, to avoid spatial distortion of moving objects, and to continuously loop the video in time. We have recently developed a stereo motion panorama method that tries to address some of these challenges. The method blends space-time regions of a video XYT volume, such that the blending regions are distinct and translate over time. This article presents a perception experiment that evaluates certain aspects of the method, namely how well observers can detect such blending regions. We measure detection time thresholds for different blending widths and for different scenes, and for monoscopic versus stereoscopic videos. Our results suggest that blending may be more effective in image regions that do not contain coherent moving objects that can be tracked over time. For example, we found moving water and partly transparent smoke were more effectively blended than swaying branches. We also found that performance in the task was roughly the same for mono versus stereo videos.

References

[1]
Agarwala, A., Zheng, K., Pal, C., Agrawala, M., Cohen, M., Curless, B., Salesin, D., and Szeliski, R. 2005. Panoramic video textures. ACM Trans. Graph. 24, 3, 821--827.
[2]
Burt, P. J. and Adelson, E. H. 1983. A multiresolution spline with application to image mosaics. ACM Trans. Graph. 2, 4, 217--236.
[3]
Couture, V., Langer, M. S., and Roy, S. 2010. Capturing non-periodic omnistereo motions. In Proceedings of the 10th Workshop on Omnidirectional Vision, Camera Networks and Non-classical Cameras (OMNIVIS).
[4]
Couture, V., Langer, M. S., and Roy, S. 2011. Panoramic stereo video textures. In Proceedings of the IEEE International Conference on Computer Vision (ICCV). 1251--1258.
[5]
Daly, S., Held, R., and Hoffman, D. 2011. Perceptual issues in stereoscopic signal processing. IEEE Trans. Broadcas. 57, 2, 347--361.
[6]
Derpanis, K. G. and Wildes, R. P. 2012. Spacetime texture representation and recognition based on a spatiotemporal orientation analysis. IEEE Trans. Pattern Anal. Mach. Intell. 34, 6, 1193--1205.
[7]
Dixon, W. and Mood, A. M. 1948. A method for obtaining and analyzing sensitivity data. J. Ameri. Statist. Assoc. 43, 109--126.
[8]
Held, R. T. and Banks, M. S. 2008. Misperceptions in stereoscopic displays: A vision science perspective. In Proceedings of the 5th Symposium on Applied Perception in Graphics and Visualization (APGV). ACM, New York, 23--32.
[9]
Huang, H.-C. and Hung, Y.-P. 1998. Panoramic stereo imaging system with automatic disparity warping and seaming. Graph. Models . Image Proces. 60, 3, 196--208.
[10]
Ishiguro, H., Yamamoto, M., and Tsuji, S. 1992. Omni-Directional stereo. IEEE Trans. Pattern Anal. Mach. Intell. 14, 2, 257--262.
[11]
Kingdom, F. A. and Prins, N. 2010. Psychophysics: A Practical Introduction. Academic Press.
[12]
Kwatra, V., Schdl, A., Essa, I., Turk, G., and Bobick, A. 2003. Graphcut textures: Image and video synthesis using graph cuts. ACM Trans. Graph. 22, 3, 277--286.
[13]
Mendiburu, B. 2009. 3D Movie Making: Stereoscopic Digital Cinema from Script to Screen. Focal Press.
[14]
Naemura, T., Kaneko, M., and Harashima, H. 1998. Multi-User immersive stereo. In Proceedings of the IEEE International Conference on Image Processing 1. 903.
[15]
Peleg, S., Ben-Ezra, M., and Pritch, Y. 2001. Omnistereo: Panoramic stereo imaging. IEEE Trans. Pattern Anal. Mach. Intell. 23, 3, 279--290.
[16]
Perez, P., Gangnet, M., and Blake, A. 2003. Poisson image editing. In SIGGRAPH Conference Proceedings. ACM Press, New York, 313--318.
[17]
Rav-Acha, A., Pritch, Y., Lischinski, D., and Peleg, S. 2007. Dynamosaicing: Mosaicing of dynamic scenes. IEEE Trans. Pattern Anal. Mach. Intell. 29, 10, 1789--1801.
[18]
Schödl, A., Szeliski, R., Salesin, D., and Essa, I. 2000. Video textures. In SIGGRAPH Conference Proceedings. ACM Press/Addison-Wesley Publishing, New York, 489--498.
[19]
Seitz, S., Kalai, A., and Shum, H. 2002. Omnivergent stereo. Int. J. Comput. Vis. 48, 3, 159--172.
[20]
Szeliski, R. 2010. Computer Vision: Algorithms and Applications. Springer, Chapter 9.3 (Compositing).
[21]
Tardif, J.-P., Roy, S., and Trudeau, M. 2003. Multi-Projectors for arbitrary surfaces without explicit calibration nor reconstruction. In Proceedings of the International Conference on 3-D Digital Imaging and Modeling. 217--224.
[22]
Vangorp, P., Chaurasia, G., Laffont, P.-Y., Fleming, R., and Drettakis, G. 2011. Perception of visual artifacts in image-based rendering of faades. Comput. Graph. Forum 30, 4, 1241--1250.
[23]
Vision3D. 2011. http://vision3d.iro.umontreal.ca/en/projects/omnistereo/.
[24]
Wetherill, G. B. and Levitt, H. 1965. Sequential estimation of points on a psychometric function. Brit. J. Math. Statis. Psycho. 18, 1--10.
[25]
Woeste, H. 2009. Mastering Digital Panoramic Photography. Rocky Nook.

Cited By

View all
  • (2017)Gigapixel Panorama Video LoopsACM Transactions on Graphics10.1145/314445537:1(1-15)Online publication date: 16-Nov-2017
  • (2013)Omnistereo Video Textures without GhostingProceedings of the 2013 International Conference on 3D Vision10.1109/3DV.2013.17(64-70)Online publication date: 29-Jun-2013

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Transactions on Applied Perception
ACM Transactions on Applied Perception  Volume 9, Issue 3
July 2012
74 pages
ISSN:1544-3558
EISSN:1544-3965
DOI:10.1145/2325722
Issue’s Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 02 August 2012
Accepted: 01 June 2012
Received: 01 June 2012
Published in TAP Volume 9, Issue 3

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Stereo
  2. blending
  3. motion
  4. omnistereo
  5. perception
  6. visual perception

Qualifiers

  • Research-article
  • Research
  • Refereed

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)2
  • Downloads (Last 6 weeks)1
Reflects downloads up to 28 Sep 2024

Other Metrics

Citations

Cited By

View all
  • (2017)Gigapixel Panorama Video LoopsACM Transactions on Graphics10.1145/314445537:1(1-15)Online publication date: 16-Nov-2017
  • (2013)Omnistereo Video Textures without GhostingProceedings of the 2013 International Conference on 3D Vision10.1109/3DV.2013.17(64-70)Online publication date: 29-Jun-2013

View Options

Get Access

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media