Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/2072298.2072347acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

Prediction of the inter-observer visual congruency (IOVC) and application to image ranking

Published: 28 November 2011 Publication History
  • Get Citation Alerts
  • Abstract

    This paper proposes an automatic method for predicting the inter-observer visual congruency (IOVC). The IOVC reflects the congruence or the variability among different subjects looking at the same image. Predicting this congruence is of interest for image processing applications where the visual perception of a picture matters such as website design, advertisement, etc. This paper makes several new contributions. First, a computational model of the IOVC is proposed. This new model is a mixture of low-level visual features extracted from the input picture where model's parameters are learned by using a large eye-tracking database. Once the parameters have been learned, it can be used for any new picture. Second, regarding low-level visual feature extraction, we propose a new scheme to compute the depth of field of a picture. Finally, once the training and the feature extraction have been carried out, a score ranging from 0 (minimal congruency) to 1 (maximal congruency) is computed. A value of 1 indicates that observers would focus on the same locations and suggests that the picture presents strong locations of interest. A second database of eye movements is used to assess the performance of the proposed model. Results show that our IOVC criterion outperforms the Feature Congestion measure \cite{Rosenholtz2007}. To illustrate the interest of the proposed model, we have used it to automatically rank personalized photograph.

    References

    [1]
    R. Althoff and N. Cohen. Eye-movement-based memory effect: a reprocessing effect in face perception. Jounral Of Experimental Psychology-Learning Memory and Cognition, 25(4):997--1010, 1999.
    [2]
    R. Baddeley and B. Tatler. High frequency edges (but not contrast predict where we fixate: A bayesian system identification analysis. Vision Research, 46:2824--2833, 2006.
    [3]
    S. Bhattacharya, R. Sukthankar, and M. Shah. A coherent framework for photo-quality assessment and enhancement based on visual aesthetics. In in ACM Multimedia International conference, 2010.
    [4]
    C. Christoudias, B. Georgescu, and P. Meer. Synergism in low-level vision. In 16th International Conference on Pattern Recognition, volume IV, pages 105--155, 2002.
    [5]
    H. Chua, J. Boland, and R. Nisbett. Cultural variation in eye movements during scene perception. In Proceedings of the National Academy of Sciences, volume 102, pages 12629--12633, 2005.
    [6]
    D. Cohen-Or, O. Sorkine, R. Gal, T. Leyvand, and Y. Xu. Color harmonization. In ACM Transactions on Graphics (Proceedings of ACM SIGGRAPH), volume 56, pages 624--630, 2006.
    [7]
    D. Comaniciu and P. Meer. Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Analysis and Machine Intelligence, 24:603--619, 2002.
    [8]
    L. Cowen, L. Ball, and J. Delin. An eye-movement analysis of web-page usability. In L. S. V. Ltd, editor, People and Computers XVI-Memorable yet invisible: Proceedings of HCI 2002, pages 317--335, 2002.
    [9]
    K. Ehinger, B. Hidalgo-Sotelo, A. Torralba, and A. Oliva. Modeling search for people in 900 scenes. Visual Cognition, 17:945--978, 2009.
    [10]
    H. Einhorn. Accepting erro to make less error. Journal of Personality Assessment, 50(3):387--395, 1986.
    [11]
    H. Frey, C. Honey, and P. Konig. What's color got to do with it? the influence of color on visual attention in different categories. Journal of Vision, 8(14), October 2008.
    [12]
    Gershnfel. The nature of mathematical modelling. Cambridge, Univ. Press, 1999.
    [13]
    H. Golberg and X. Kotval. Computer interface evaluation using eye movements: methods and constructs. International Journal of Industrial Ergonomics, 24:631--645, 1999.
    [14]
    R. Gordon. Attentional allocation during the perception of scenes. Journal of Experimental Psychology: Human Perception and Performance, 30:760--777, 2004.
    [15]
    T. Hastie, R. Tibshirani, and J. Friedman. The elements of statistical learning. Springer Series in Statistics, 2001.
    [16]
    J. Henderson. Regarding scenes. Current Directions in Psychological Science, 16:219--222, 2007.
    [17]
    J. Henderson, M. Chanceaux, and T. Smith. The influence of clutter on real-world scene search: Evidence from search efficiency and eye movements. Journal of Vision, 9(1), January 2009.
    [18]
    M. Jordan and R. Jacobs. Hierarchical mixtures of experts and the em algorihtm. Neural Computation, 6:181--214, 1994.
    [19]
    T. Judd, F. Durand, and A. Torralba. Fixations on low-resolution images. Journal of Vision, 11(4), 2011.
    [20]
    T. Judd, K. Ehinger, F. Durand, and A. Torralba. Learning to predict where people look. In ICCV, 2009.
    [21]
    O. Le Meur, P. Le Callet, D. Barba, and D. Thoreau. A coherent computational approach to model the bottom-up visual attention. IEEE Trans. On PAMI, 28(5):802--817, May 2006.
    [22]
    A. Levin. Blind motion deblurring using image statistics. In NIPS, 2006.
    [23]
    R. Lienhart and J. Maydt. An extended set of haar-like features for rapid object detection. In ICIP, volume 1, pages 900--903, 2002.
    [24]
    G. Loftus and N. Mackworth. Cognitive determinants of fixation location during picture viewing. Journal of Experimental Psychology: Human Perception and Performances, 4:565--572, 1978.
    [25]
    Y. Luo and X. Tang. Photo and video quality evaluation: focussing on the subject. In ECCV, pages 386--399, 2008.
    [26]
    Y. Matsuda. Coor design. In Asakura Shoten, 1995.
    [27]
    R. Nisbett. The geography of thought: how Asians and Westerners think differently... and why. New York: Free Press, 2003.
    [28]
    A. Oliva, M. Mack, M. Shrestha, and A. Peeper. Identifying the perceptual dimensions of visual complexity of scenes. In 26th annual meeting of the Cognitive Science Society Meeting, 2004.
    [29]
    D. Parkhurst, K. Law, and E. Niebur. Modelling the role of salience in the allocation of overt visual attention. Vision Research, 42:107--123, 2002.
    [30]
    W. Press, S. Teukolsky, W. Vetterling, and B. Flannery. Numerical Recipes in C: the art of Scientific Computing. Cambridge University Press, New York, NY, USA, 1992.
    [31]
    K. Rayner. Eye movements in reading and information processing: 20 years of research. Psychological Bulletin, 124(3):372--422, 1998.
    [32]
    K. Rayner, M. Catelhano, and J. Yang. Eye movements when looking at unusual-weird scenes: are there cultural differences? Journal of Experimental psychology: learning, Memory and cognition, 35(1):154--259, 2009.
    [33]
    R. Rosenholtz, Y. Li, and L. Nakano. Measuring visual clutter. Journal of Vision, 7(2), March 2007.
    [34]
    M. Ross and A. Oliva. Estimating perception of scene layout properties from global image features. Journal Of Vision, 10(1), Januray 2010.
    [35]
    C. Rother, L. Bordeaux, Y. Hamadi, and A. Black. Autocollage. In in ACM Transactions on Graphics (SIGGRAPH), 2006.
    [36]
    G. A. Rousselet, M. J.-M. Macé, and M. Fabre-Thorpe. Is it an animal? is it a human face? fast processing in upright and inverted natural scenes. Journal of Vision, 3:440--455, 2003.
    [37]
    X. Sun, H. Yao, R. Ji, and S. Liu. Photo assessment based on computatinal visual attention model. In ACM Multimedia, pages 541--544, 2009.
    [38]
    B. W. Tatler, R. J. Baddeley, and I. D. Gilchrist. Visual correlates of fixation selection: effects of scale and time. Vision Research, 45:643--659, 2005.
    [39]
    M. Tokumaru, N. Muranaka, and S. Imanishi. Color design support system considering coor harmony. In IEEE International Conference on Fuzzy Systems, pages 378--383, 2002.
    [40]
    A. Torralba and A. Oliva. Depth estimation from image structure. IEEE Pattern Analysis and Machine Intelligence, 24(9):1226--1238, 2002.
    [41]
    A. Torralba and A. Oliva. Statistics of natural image catagories. network, 14:391--421, 2003.
    [42]
    A. Torralba, A. Oliva, M. Castelhano, and J. Henderson. Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search. Psychological review, 113(4):766--786, 2006.
    [43]
    G. Underwood and T. Foulsham. Visual saliency and semantic incongruency influence eye movements when inspecting pictures. The Quarterly journal of experimental psychology, 59(11):1931--1949, 2006.
    [44]
    P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. In CVPR, 2001.
    [45]
    C.-G. Yeh, Y. Ho, B. Barsky, and M. Ouhyoung. Personalized photograph ranking and selection system. In ACM Multimedia, 2010.
    [46]
    Q. Zhao and C. Koch. Learning a saliency map using fixated locations in natural scenes. Journal of Vision, 11(3):1--15, 2011.

    Cited By

    View all
    • (2024)The Use of Machine Learning in Eye Tracking Studies in Medical Imaging: A ReviewIEEE Journal of Biomedical and Health Informatics10.1109/JBHI.2024.337189328:6(3597-3612)Online publication date: Jun-2024
    • (2024)Revisiting the Heider and Simmel experiment for social meaning attribution in virtual realityScientific Reports10.1038/s41598-024-65532-014:1Online publication date: 24-Jul-2024
    • (2023)High-level cinematic knowledge to predict inter-observer visual congruencyProceedings of the 2023 ACM International Conference on Interactive Media Experiences Workshops10.1145/3604321.3604331(103-108)Online publication date: 12-Jun-2023
    • Show More Cited By

    Index Terms

    1. Prediction of the inter-observer visual congruency (IOVC) and application to image ranking

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      MM '11: Proceedings of the 19th ACM international conference on Multimedia
      November 2011
      944 pages
      ISBN:9781450306164
      DOI:10.1145/2072298
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 28 November 2011

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. congruency
      2. eye tracking
      3. images ranking
      4. visual dispersion

      Qualifiers

      • Research-article

      Conference

      MM '11
      Sponsor:
      MM '11: ACM Multimedia Conference
      November 28 - December 1, 2011
      Arizona, Scottsdale, USA

      Acceptance Rates

      Overall Acceptance Rate 995 of 4,171 submissions, 24%

      Upcoming Conference

      MM '24
      The 32nd ACM International Conference on Multimedia
      October 28 - November 1, 2024
      Melbourne , VIC , Australia

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)24
      • Downloads (Last 6 weeks)2
      Reflects downloads up to 09 Aug 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)The Use of Machine Learning in Eye Tracking Studies in Medical Imaging: A ReviewIEEE Journal of Biomedical and Health Informatics10.1109/JBHI.2024.337189328:6(3597-3612)Online publication date: Jun-2024
      • (2024)Revisiting the Heider and Simmel experiment for social meaning attribution in virtual realityScientific Reports10.1038/s41598-024-65532-014:1Online publication date: 24-Jul-2024
      • (2023)High-level cinematic knowledge to predict inter-observer visual congruencyProceedings of the 2023 ACM International Conference on Interactive Media Experiences Workshops10.1145/3604321.3604331(103-108)Online publication date: 12-Jun-2023
      • (2023)Selective Visual Attention Revealed by Integrating FMRI and Eye-Tracking2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI)10.1109/ISBI53787.2023.10230382(1-5)Online publication date: 18-Apr-2023
      • (2022)A domain adaptive deep learning solution for scanpath prediction of paintingsProceedings of the 19th International Conference on Content-based Multimedia Indexing10.1145/3549555.3549597(57-63)Online publication date: 14-Sep-2022
      • (2022)Single Image Reflection Removal Based on Reflection Region2022 5th International Conference on Artificial Intelligence and Big Data (ICAIBD)10.1109/ICAIBD55127.2022.9820344(254-260)Online publication date: 27-May-2022
      • (2022)Self Supervised Scanpath Prediction Framework for Painting Images2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)10.1109/CVPRW56347.2022.00160(1538-1547)Online publication date: Jun-2022
      • (2022)RETRACTED ARTICLE: FPGA-based reflection image removal using cognitive neural networksApplied Nanoscience10.1007/s13204-022-02352-613:3(2539-2553)Online publication date: 7-Feb-2022
      • (2021)Inter-Observer Visual Congruency in Video-Viewing2021 International Conference on Visual Communications and Image Processing (VCIP)10.1109/VCIP53242.2021.9675428(1-5)Online publication date: 5-Dec-2021
      • (2021)Viewport-Dependent Saliency Prediction in 360° VideoIEEE Transactions on Multimedia10.1109/TMM.2020.298768223(748-760)Online publication date: 2021
      • Show More Cited By

      View Options

      Get Access

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media