Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/2502081.2502282acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

Large-scale visual sentiment ontology and detectors using adjective noun pairs

Published: 21 October 2013 Publication History

Abstract

We address the challenge of sentiment analysis from visual content. In contrast to existing methods which infer sentiment or emotion directly from visual low-level features, we propose a novel approach based on understanding of the visual concepts that are strongly related to sentiments. Our key contribution is two-fold: first, we present a method built upon psychological theories and web mining to automatically construct a large-scale Visual Sentiment Ontology (VSO) consisting of more than 3,000 Adjective Noun Pairs (ANP). Second, we propose SentiBank, a novel visual concept detector library that can be used to detect the presence of 1,200 ANPs in an image. The VSO and SentiBank are distinct from existing work and will open a gate towards various applications enabled by automatic sentiment analysis. Experiments on detecting sentiment of image tweets demonstrate significant improvement in detection accuracy when comparing the proposed SentiBank based predictors with the text-based approaches. The effort also leads to a large publicly available resource consisting of a visual sentiment ontology, a large detector library, and the training/testing benchmark for visual sentiment analysis.

References

[1]
H. Aradhye, G. Toderici, and J. Yagnik. Video2Text: Learning to Annotate Video Content. Internet Multimedia Mining, 2009.
[2]
D. Borth, A. Ulges, and T.M. Breuel. Lookapp - Interactive Construction of web-based Concept Detectors. ICMR, 2011.
[3]
D. Borth and S-F. Chang. Constructing Structures and Relations in SentiBank Visual Sentiment Ontology. Technical Report#CUCS-020--13, Columbia University, Computer Science Dep., 2013.
[4]
E. Dan-Glauser et al. The Geneva Affective Picture Database (GAPED): a new 730-picture database focusing on valence and normative significance. Behavior Research Methods, 2011.
[5]
Charles Darwin. The Expression of the Emotions in Man and Animals. Oxford University Press, USA, 1872 / 1998.
[6]
R. Datta, D. Joshi, J. Li, and J. Wang. Studying Aesthetics in Photographic Images using a Computational Approach. ECCV, 2006.
[7]
J. Deng et al. ImageNet: A Large-Scale Hierarchical Image Database. CVPR, 2009.
[8]
P. Ekman et al. Facial Expression and Emotion. American Psychologist, 48:384--384, 1993.
[9]
A. Esuli and F. Sebastiani. SentiWordnet: A publicly available Lexical Resource for Opinion Mining. LREC, 2006.
[10]
M. Everingham, et al. The Pascal Visual Object Classes (VOC) Challenge. Int. J. of Computer Vision, 88(2):303--338, 2010.
[11]
A. Hanjalic, C. Kofler, and M. Larson. Intent and its Discontents: the User at the Wheel of the Online Video Search Engine. ACM MM, 2012.
[12]
P. Isola, J. Xiao, A. Torralba, and A. Oliva. What makes an Image Memorable? CVPR, 2011.
[13]
J. Jia, S. Wu, X. Wang, P. Hu, L. Cai, and J. Tang. Can we understand van Gogh's Mood?: Learning to infer Affects from Images in Social Networks. ACM MM, 2012.
[14]
Y.-G. Jiang, G. Ye, S.-F. Chang, D. Ellis, and A. Loui. Consumer Video Understand.: Benchmark Database and an Eval. of Human and Machine Performance. ICMR, 2011.
[15]
D. Joshi, R. Datta, E. Fedorovskaya, Q. Luong, J. Wang, J. Li, and J. Luo. Aesthetics and Emotions in Images. Signal Processing Magazine, 28(5):94--115, 2011.
[16]
L. Kennedy, S.-F. Chang, and I. Kozintsev. To Search or to Label?: Predicting the Performance of Search-based Automatic Image Classifiers. MIR Workshop, 2006.
[17]
P. Lang, M. Bradley, and B. Cuthbert. International Affective Picture System (IAPS): Technical Manual and Affective Ratings, 1999.
[18]
B. Li, et al. Scaring or Pleasing: Exploit Emotional Impact of an Image. ACM MM, 2012.
[19]
X. Li, C. Snoek, M. Worring, and A. Smeulders. Harvesting Social Images for Bi-Concept Search. IEEE Transactions on Multimedia, 14(4):1091--1104, 2012.
[20]
N. Codella et al. IBM Research and Columbia University TRECVID-2011 Multimedia Event Detection (med) System. NIST TRECVID Workshop, 2011.
[21]
J. Machajdik and A. Hanbury. Affective Image Classification using Features inspired by Psychology and Art Theory. ACM MM, 2010.
[22]
L. Marchesotti, F. Perronnin, D. Larlus, and G. Csurka. Assessing the Aesthetic Quality of Photographs using Generic Image Descriptors. ICCV, 2011.
[23]
M. Naphade, J. Smith, J. Tesic, S. Chang, W. Hsu, L. Kennedy, A. Hauptmann, and J. Curtis. Large-Scale Concept Ontology for Multimedia. IEEE MultiMedia, 13(3):86--91, 2006.
[24]
C. Osgood, G. Suci, and P. Tannenbaum. The Measurement of Meaning, volume 47. University of Illinois Press, 1957.
[25]
P. Over et al. Trecvid 2012 -- An Overview of the Goals, Tasks, Data, Evaluation Mechanisms and Metrics. TRECVID Workshop, 2012.
[26]
B. Pang and L. Lee. Opinion Mining and Sentiment Analysis. Information Retrieval, 2(1--2):1--135, 2008.
[27]
Robert Plutchik. Emotion: A Psychoevolutionary Synthesis. Harper & Row, Publishers, 1980.
[28]
C. Snoek and M. Worring. Concept-based Video Retrieval. Foundations and Trends in Inf. Retrieval, 4(2), 2009.
[29]
S. Strassel et al. Creating HAVIC: Heterogeneous Audio Visual Internet Collection. LREC, 2012.
[30]
M. Thelwall et al. Sentiment Strength Detection in Short Informal Text. J. of the American Soc. for Information Science and Tech., 61(12):2544--2558, 2010.
[31]
A. Ulges, C. Schulze, M. Koch, and T. Breuel. Learning Automatic Concept Detectors from Online Video. Journal on Comp. Vis. Img. Underst., 114(4):429--438, 2010.
[32]
V. Vonikakis and S. Winkler. Emotion-based Sequence of Family Photos. ACM MM, 2012.
[33]
W. Wang and Q. He. A Survey on Emotional Semantic Image Retrieval. IEEE ICIP, 2008.
[34]
X. Wang, J. Jia, P. Hu, S. Wu, J. Tang, and L. Cai. Understanding the Emotional Impact of Images. ACM MM, 2012.
[35]
T. Wilson et al. Recognizing Contextual Polarity in phrase-level Sentiment Analysis. HLT/EMNLP, 2005.
[36]
V. Yanulevskaya et al. In the Eye of the Beholder: Employing Statistical Analysis and Eye Tracking for Analyzing Abstract Paintings. ACM MM, 2012.
[37]
V. Yanulevskaya et al. Emotional Valence Categorization using Holistic Image Features. IEEE ICIP, 2008.
[38]
Li et al. ObjectBank: A high-level Image Rep. for Scene Classification and Semantic Feature Sparsification. NIPS, 2010.
[39]
Torresani et al. Efficient Object Category Recognition using Classemes. ECCV, 2010.
[40]
A. Olivai and A. Torralba. Modeling the Shape of the Scene: a Holistic Representation of the Spatial Envelope. Int. J. of Computer Vision, 42(3):145--175, 2001.
[41]
S. Bhattacharya and R. Sukthankar and M. Shah. A holistic Approach to Aesthetic Enhancement of Photographs. TOMCCAP, 7(1), 2011.

Cited By

View all
  • (2024)Enhancing Multimodal Tourism Review Sentiment Analysis Through Advanced Feature Association TechniquesInternational Journal of Information Systems in the Service Sector10.4018/IJISSS.34956415:1(1-21)Online publication date: 17-Jul-2024
  • (2024)Multimodal Emotion Cognition Method Based on Multi-Channel Graphic InteractionInternational Journal of Cognitive Informatics and Natural Intelligence10.4018/IJCINI.34996918:1(1-17)Online publication date: 17-Sep-2024
  • (2024)Multi-Modal Sentiment Analysis Based on Image and Text Fusion Based on Cross-Attention MechanismElectronics10.3390/electronics1311206913:11(2069)Online publication date: 27-May-2024
  • Show More Cited By

Index Terms

  1. Large-scale visual sentiment ontology and detectors using adjective noun pairs

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    MM '13: Proceedings of the 21st ACM international conference on Multimedia
    October 2013
    1166 pages
    ISBN:9781450324045
    DOI:10.1145/2502081
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 21 October 2013

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. concept detection
    2. ontology
    3. sentiment prediction
    4. social multimedia

    Qualifiers

    • Research-article

    Conference

    MM '13
    Sponsor:
    MM '13: ACM Multimedia Conference
    October 21 - 25, 2013
    Barcelona, Spain

    Acceptance Rates

    MM '13 Paper Acceptance Rate 47 of 235 submissions, 20%;
    Overall Acceptance Rate 2,145 of 8,556 submissions, 25%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)241
    • Downloads (Last 6 weeks)24
    Reflects downloads up to 10 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Enhancing Multimodal Tourism Review Sentiment Analysis Through Advanced Feature Association TechniquesInternational Journal of Information Systems in the Service Sector10.4018/IJISSS.34956415:1(1-21)Online publication date: 17-Jul-2024
    • (2024)Multimodal Emotion Cognition Method Based on Multi-Channel Graphic InteractionInternational Journal of Cognitive Informatics and Natural Intelligence10.4018/IJCINI.34996918:1(1-17)Online publication date: 17-Sep-2024
    • (2024)Multi-Modal Sentiment Analysis Based on Image and Text Fusion Based on Cross-Attention MechanismElectronics10.3390/electronics1311206913:11(2069)Online publication date: 27-May-2024
    • (2024)A cross-model hierarchical interactive fusion network for end-to-end multimodal aspect-based sentiment analysisIntelligent Data Analysis10.3233/IDA-23030528:5(1293-1308)Online publication date: 19-Sep-2024
    • (2024)A multifactor model using large language models and investor sentiment from photos and news: new evidence from ChinaSSRN Electronic Journal10.2139/ssrn.4708979Online publication date: 2024
    • (2024)Bridging Visual Affective Gap: Borrowing Textual Knowledge by Learning from Noisy Image-Text PairsProceedings of the 32nd ACM International Conference on Multimedia10.1145/3664647.3680875(602-611)Online publication date: 28-Oct-2024
    • (2024)TGCA-PVT: Topic-Guided Context-Aware Pyramid Vision Transformer for Sticker Emotion RecognitionProceedings of the 32nd ACM International Conference on Multimedia10.1145/3664647.3680781(9709-9718)Online publication date: 28-Oct-2024
    • (2024)CH-Mits: A Cross-Modal Dataset for User Sentiment Analysis on Chinese Social MediaProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3679125(5390-5394)Online publication date: 21-Oct-2024
    • (2024)A Novel Dual-Pipeline based Attention Mechanism for Multimodal Social Sentiment AnalysisCompanion Proceedings of the ACM Web Conference 202410.1145/3589335.3651967(1816-1822)Online publication date: 13-May-2024
    • (2024)Does image sentiment of major public emergency affect the stock market performance? New insight from deep learning techniquesAccounting & Finance10.1111/acfi.13313Online publication date: 14-Aug-2024
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media