Abstract
Metaphors are a common language tool used in communication that widely exist in natural language. Metaphor detection is beneficial to improving the performance of natural language processing tasks such as machine translation. The current metaphor detection method captures the semantic incoherence of the metaphorical word and its surrounding text. However, it ignores the different contributions of phrases and words. In addition, word information and its contextual information should be deeply integrated. We propose a learning framework that combines word representations, contextual representations and combined representations. Specifically, we establish word-level and phrase-level attention mechanisms to learn enhanced feature representations. For the word-level attention, we extract word embedding, part-of-speech (POS) and distance features as multilevel representations and use multi-attention to obtain weight information. For the phrase-level attention, the IndRNN and self-attention are employed to obtain the deep semantic representation of the sentence. By using this strategy, our model has the ability to mine richer semantic representations. Experiments on the VUA metaphor corpus show that our method achieves an F-score of 69.7% on the test dataset, thereby surpassing the scores of all the existing models by a considerable margin.
Similar content being viewed by others
References
Assaf D, Neuman Y, Cohen Y et al (2013) Why “dark thoughts” aren’t really dark: A novel algorithm for metaphor identification[C]. 2013 IEEE Symposium on Computational Intelligence, Cognitive Algorithms, Mind, and Brain (CCMB). IEEE, 60-65
Cameron L (2003) Metaphor in educational discourse[M]. A&C Black
Coltheart M (1981) The MRC psycholinguistic database[J]. Q J Exp Psychol A 33(4):497–505
Del Tredici M, Bel N (2016) Assessing the potential of metaphoricity of verbs using corpus data[C]. Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), 4573-4577
Do Dinh EL, Gurevych I (2016)Token-level metaphor detection using neural networks[C]. Proceedings of the Fourth Workshop on Metaphor in NLP, 28-33
Fan GF, Peng LL, Hong WC et al (2016) Electric load forecasting by the SVR model with differential empirical mode decomposition and auto regression[J]. Neurocomputing 173:958–970
Gao G, Choi E, Choi Y et al (2018) Conference on Empirical Methods in Natural Language Processing, 607-613
Hanks P (2004) The syntagmatics of metaphor and idiom[J]. Int J Lexicogr 17(3):245–274
Hinton GE, Srivastava N, Krizhevsky A et al (2012) Improving neural networks by preventing co-adaptation of feature detectors[J]. arXiv preprint arXiv:1207.0580,
Hong WC, Fan GF (2019) Hybrid empirical mode decomposition with support vector regression model for short term load forecasting[J]. Energies 12(6):1093
Huang TH (2014) Social metaphor detection via topical analysis[C]. InternationalJournal of Computational Linguistics & Chinese Language Processing, vol 19, no 2
Iwendi C, Srivastava G, Khan S et al (2020) Cyberbullying detection solutions based on deep learning architectures[J]. Multimed Syst: 1–14
Iwendi C, Moqurrab SA, Anjum A et al (2020) N-Sanitization: A semantic privacy-preserving framework for unstructured medical datasets[J]. Comput Commun 161:160–171
Klebanov BB, Leong CW, Gutierrez ED et al (2016) Semantic classifications for detection of verb metaphors[C]. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, vol 2, Short Papers, 101-106
Lakoff G, Johnson M (2008) Metaphors we live by[M]. University of Chicago press
Li S et al (2018) Independently recurrent neural network (indrnn): Building a longer and deeper rnn. Proceedings of the IEEE conference on computer vision and pattern recognition
Mohler M, Bracewell D, Tomlinson M et al (2013) Semantic signatures for example-based linguistic metaphor detection[C]. Proceedings of the First Workshop on Metaphor in NLP, 27-35
Mykowiecka A, Wawer A, Marciniak M (2018) Detecting figurative word occurrences using recurrent neural networks[C]. Proceedings of the Workshop on Figurative Language Processing, 24-127
Neuman Y, Assaf D, Cohen Y et al (2013) Metaphor identification in large texts corpora[J]. PLoS ONE 8(4):e62343
Pennington J, Socher R, Manning CD, Glove (2014) Global vectors for word representation[C]. Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP),1532-1543
Peters M, Neumann M, Iyyer M et al (2018) Deep contextualized word representations[C]. Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume1 (Long Papers), 2227-2237
Pramanick M, Gupta A, Mitra P (2018) An lstm-crf based approach to token-level metaphor detection[C]. Proceedings of the Workshop on Figurative Language Processing, 67-75
Rai S, Chakraverty S, Tayal DK (2016) Supervised metaphor detection using conditional random fields[C]. Proceedings of the Fourth Workshop on Metaphor in NLP, 18-27
Rei M, Bulat L, Kiela D et al (2017) Grasping the finer point: a supervised similarity network for metaphor detection[C]. Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 1537-1546
Semino E (2007) MIP: a method for identifying metaphorically used words in discourse[J]. Metaphor Symb 22(1):1–39
Shutova E, Teufel S (2010) Metaphor corpus annotated for source-target domain mappings[C]. LREC. 2:222
Shutova E, Kiela D, Maillard J (2016) Black holes and white rabbits: Metaphor identification with visual features[C]. Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 160-170
Shutova E, Sun L, Gutiérrez ED et al (2017) Multilingual metaphor processing: Experiments with semi-supervised and unsupervised learning[J]. Comput Linguist 43(1):71–123
Steen GJ, Dorst AG, Herrmann JB, et al. (2010) A Method for Linguistic Metaphor Identification: From MIP to MIPVU[M]. John Benjamins Publishing
Stowe K, Moeller S, Michaelis L, Palmer M (2019) Linguistic analysis improves neural metaphor detection. In: Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pp 362-371
Swarnkar K, Singh AK (2018)Di-LSTM contrast: A deep neural network for metaphor detection[C]. Proceedings of the Workshop on Figurative Language Processing, 115-120
Tsvetkov Y, Boytsov L, Gershman A et al (2014) Metaphor detection with cross-lingual model transfer[C]. Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, vol 1, Long Papers, 248-258
Turney P, Neuman Y, Assaf D et al (2011) Literal and metaphorical sense identification through concrete and abstract context[C]. Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, 680-690
Wu C, Wu F, Chen Y et al (2018) Neural metaphor detecting with CNN-LSTM model[C]. Proceedings of the Workshop on Figurative Language Processing, 110-114
Zhang Z, Hong WC (2019) Electric load forecasting by complete ensemble empirical mode decomposition adaptive noise and support vector regression with quantum-based dragonfly algorithm[J]. Nonlinear Dyn 98(2):1107–1136
Acknowledgements
We thank the anonymous reviewers for their insightful comments. This work was supported by the National Natural Science Foundation of China (61962057), the Key Program of the National Natural Science Foundation of China (U2003208), and the Major Science and Technology Projects in the Autonomous Region (2020A03004-4).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Yang, Q., Yu, L., Tian, S. et al. Word-level and phrase-level strategies for figurative text identification. Multimed Tools Appl 81, 14339–14353 (2022). https://doi.org/10.1007/s11042-022-12233-3
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-022-12233-3