Abstract
Commonsense reasoning is one of the abilities necessary for artificial intelligence to be as intelligent as humans. However, how to make AI understand commonsense has been a problem that has plagued artificial intelligence for more than 60 years. Existing efforts focus more on the means of knowledge acquisition and strive to enrich the capacity of commonsense knowledge (CSK) bases and dimensions of CSK through advanced methods. Unfortunately, this exuberance has obscured a general consideration of CSK, such as how to follow human habits to obtain the most representative knowledge we need to understand the world. In this paper, this representative knowledge is referred to as core CSK. The influence of core CSK is extensive, and it constitutes almost the fundamental element of human life and the most fundamental cognition of the world. Harnessing human curiosity to find solutions to the above problems is an effective and straightforward route. Specifically, we focus on a special corpus to mine core CSK, namely, why-questions. For example, we can harvest “the sky is blue” from “why is the sky blue?”. To this end, we propose a novel method to extract CSK from why-questions, which mainly consist of two modules. The first is a question classification module used to determine whether a question contains CSK. In this module, we propose a classifier based on a one-sided bootstrapping method and design several informative features for the classifier. The second is a crowdsourcing module used to improve the quality of the extracted commonsense. We conduct extensive experiments, and the experimental results show that our method effectively mines CSK from question corpora. Furthermore, statistical analysis demonstrates the feasibility of this curiosity-driven approach, implying that we provide a basic idea for collecting core CSK. Remarkably, today’s outstanding large language models do not have such simple knowledge summarization capabilities, demonstrating the barrier between the excellence of language models and the universality of CSK.
Similar content being viewed by others
Data Availibility Statement
Data will be available upon request.
Code Availibility Statement
Code will be available upon request.
Notes
Another class of curiosity is called perceptual curiosity.
References
Wang C, Liu J, Liu J, Wang W. Inference of ISA commonsense knowledge with lexical taxonomy. Appl Intell. 2022;1–14.
Amodei D, Ananthanarayanan S, Anubhai R, Bai J, Battenberg E, Case C, Casper J, Catanzaro B, Cheng Q, Chen G et al. Deep speech 2: End-to-end speech recognition in English and mandarin. In: International Conference on Machine Learning, 2016;173–182.
Hassan H, Aue A, Chen C, Chowdhary V, Clark J, Federmann C, Huang X, Junczys-Dowmunt M, Lewis W, Li M, et al. Achieving human parity on automatic Chinese to English news translation. arXiv preprint arXiv:1803.05567 2018.
Levesque H, Davis E, Morgenstern L. The winograd schema challenge. In: Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning 2012. Citeseer
Lenat DB, Guha RV, Pittman K, Pratt D, Shepherd M. Cyc: toward programs with common sense. Commun ACM. 1990;33(8):30–49.
Singh P, Lin T, Mueller ET, Lim G, Perkins T, Zhu WL. Open mind common sense: knowledge acquisition from the general public. In: OTM Confederated International Conferences" On the Move to Meaningful Internet Systems", 2002:1223–37. Springer
Pasca M, Van Durme B. What you seek is what you get: extraction of class attributes from query logs. IJCAI. 2007;7:2832–7.
Miller GA. Wordnet: a lexical database for English. Commun ACM. 1995;38(11):39–41.
Fabian M, Gjergji K, Gerhard W et al. Yago: A core of semantic knowledge unifying wordnet and Wikipedia. In: 16th International World Wide Web Conference, WWW, 2007:697–706
Bollacker K, Evans C, Paritosh P, Sturge T, Taylor J. Freebase: a collaboratively created graph database for structuring human knowledge. In: Proceedings of the 2008 ACM SIGMOD International Conference on Management of Data, 2008:1247–50.
Hoffart J, Suchanek FM, Berberich K, Weikum G. Yago2: A spatially and temporally enhanced knowledge base from Wikipedia. Artif Intell. 2013;194:28–61.
Auer S, Bizer C, Kobilarov G, Lehmann J, Cyganiak R, Ives Z. Dbpedia: a nucleus for a web of open data. In: The Semantic Web: 6th International Semantic Web Conference, 2nd Asian Semantic Web Conference, ISWC 2007+ ASWC 2007, Busan, Korea, November 11-15, 2007. Proceedings, 2007:722–35. Springer
Tandon N, Melo G, Weikum G. Acquiring comparative commonsense knowledge from the web. In: Proceedings of the AAAI Conference on Artificial Intelligence, 2014:28.
Litman JA. Interest and deprivation factors of epistemic curiosity. Personality Individ Differ. 2008;44(7):1585–95.
Collins RP, Litman JA, Spielberger CD. The measurement of perceptual curiosity. Personality Individ Differ. 2004;36(5):1127–41.
Vracheva VP, Moussetis R, Abu-Rahma A. The mediational role of engagement in the relationship between curiosity and student development: a preliminary study. J Happiness Stud. 2020;21(4):1529–47.
Berlyne DE. A theory of human curiosity. Br J Psychol. 1954;45(3):180.
Darlington K. Common sense knowledge, crucial for the success of AI systems. OpenMind BBVA 2020.
Bougie N, Ichise R. Hierarchical learning from human preferences and curiosity. Appl Intell. 2022;52(7):7459–79.
Wang C, Zhu T, Li Z, Liu J. Slr: A million-scale comprehensive crossword dataset for simultaneous learning and reasoning. Neurocomputing 2023;126591.
Lecler A, Duron L, Soyer P. Revolutionizing radiology with gpt-based models: current applications, future possibilities and limitations of chatgpt. Diagn Interv Imaging. 2023;104(6):269–74.
Zang L-J, Cao C, Cao Y-N, Wu Y-M, Cun-Gen C. A survey of commonsense knowledge acquisition. J Comput Sci Technol. 2013;28(4):689–719.
Xu FF, Lin BY, Zhu K. Automatic extraction of commonsense locatednear knowledge. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2018:96–101 .
Tandon N, De Melo G, Suchanek F, Weikum G. Webchild: Harvesting and organizing commonsense knowledge from the web. In: Proceedings of the 7th ACM International Conference on Web Search and Data Mining, 2014:523–32.
Wang C, Liu J, Liu J, Wang W. Inference of ISA commonsense knowledge with lexical taxonomy. Appl Intell. 2023;53(5):5290–303.
Sherin B. Common sense clarified: the role of intuitive knowledge in physics problem solving. J Res Sci Teach. 2006;43(6):535–55.
Liu H, Singh P. Conceptnet-a practical commonsense reasoning tool-kit. BT Technol J. 2004;22(4):211–26.
Davis E. Representations of commonsense knowledge. Morgan Kaufmann; 2014.
Paul D. Social commonsense reasoning with structured knowledge in text. PhD thesis 2024.
Wang C, Liu J, Liu J, Jiang S, Li Z, Xiao Y. Sweet apple, company? or food? adjective-centric commonsense knowledge acquisition with taxonomy-guided induction. Knowl-Based Syst. 2023;280:110988.
Liu J, Chen T, Wang C, Liang J, Chen L, Xiao Y, Chen Y, Jin K. Vocsk: Verb-oriented commonsense knowledge mining with taxonomy-guided induction. Artif Intell. 2022;310:103744.
He M, Fang T, Wang W, Song Y. Acquiring and modeling abstract commonsense knowledge via conceptualization. Artif Intell. 2024:104149.
Tandon N, Varde AS, de Melo G. Commonsense knowledge in machine intelligence. ACM SIGMOD Rec. 2018;46(4):49–52.
Tandon N, De Melo G, Weikum G. Webchild 2.0: Fine-grained commonsense knowledge distillation. In: Proceedings of ACL 2017, System Demonstrations, 2017:115–20.
Hwang JD, Bhagavatula C, Le Bras R, Da J, Sakaguchi K, Bosselut A, Choi Y. (comet-) atomic 2020: On symbolic and neural commonsense knowledge graphs. In: Proceedings of the AAAI Conference on Artificial Intelligence, 2021:35;6384–92.
Zhao Z, Lee WS, Hsu D. Large language models as commonsense knowledge for large-scale task planning. Adv Neural Inf Process Syst. 2024;36.
Liu C, Wang C, Peng Y, Li Z. Zvqaf: Zero-shot visual question answering with feedback from large language models. Neurocomputing. 2024;580:127505.
Tang X, Zheng Z, Li J, Meng F, Zhu S-C, Liang Y, Zhang M. Large language models are in-context semantic reasoners rather than symbolic reasoners. arXiv preprint arXiv:2305.14825 2023.
Wei J, Wang X, Schuurmans D, Bosma M, Xia F, Chi E, Le QV, Zhou D, et al. Chain-of-thought prompting elicits reasoning in large language models. Adv Neural Inf Process Syst. 2022;35:24824–37.
West P, Bhagavatula C, Hessel J, Hwang JD, Jiang L, Bras RL, Lu X, Welleck S, Choi Y. Symbolic knowledge distillation: from general language models to commonsense models. arXiv preprint arXiv:2110.07178 2021.
Gu Y, Mishra BD, Clark P. Do language models have coherent mental models of everyday things? arXiv preprint arXiv:2212.10029 2022.
Bian N, Han X, Sun L, Lin H, Lu Y, He B, Jiang S, Dong B. Chatgpt is a knowledgeable but inexperienced solver: an investigation of commonsense problem in large language models. arXiv preprint arXiv:2303.16421 2023.
Chen J, Shi W, Fu Z, Cheng S, Li L, Xiao Y. Say what you mean! large language models speak too positively about negative commonsense knowledge. arXiv preprint arXiv:2305.05976 2023.
Cheng Q, Sun T, Liu X, Zhang W, Yin Z, Li S, Li L, Chen K, Qiu X. Can AI assistants know what they don’t know? arXiv preprint arXiv:2401.13275 2024.
Wu W, Li H, Wang H, Zhu KQ. Probase: A probabilistic taxonomy for text understanding. In: Proceedings of the 2012 ACM SIGMOD International Conference on Management of Data, 2012:481–92.
Wang Z, Wang H, Wen J-R, Xiao Y. An inference approach to basic level of categorization. In: Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, 2015:653–62.
Han D. On indexical. Journal of Renmin University of China 2015.
Zhou L-Z, He Y-K, Wang J-Y. Survey on research of sentiment analysis. J Comput Appl. 2008;28(11):2725–8.
Wang X-D, Wang J, Zhang Z. Computation on orientation for subjective sentence based on sentiment words ontology. J Comput Appl. 2012;32(6):1678–81.
Jones KS. A statistical interpretation of term specificity and its application in retrieval. J Document. 1972.
Reimers N, Gurevych I. Making monolingual sentence embeddings multilingual using knowledge distillation. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020:4512–25.
Devlin J, Chang M-W, Lee K, Toutanova K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 2018.
Fleiss JL. Measuring nominal scale agreement among many raters. Psychol Bull. 1971;76(5):378.
McHugh ML. The chi-square test of independence. Biochemia medica. 2013;23(2):143–9.
Speer R, Chin J, Havasi C. Conceptnet 5.5: An open multilingual graph of general knowledge. In: Thirty-First AAAI Conference on Artificial Intelligence 2017.
Tandon N. Commonsense knowledge acquisition and applications. PhD thesis, Saarland University, Saarbrücken, Germany 2016.
Wang C, Liu J, Zhuang T, Li J, Liu J, Xiao Y, Wang W, Xie R. A sequence-to-sequence model for large-scale chinese abbreviation database construction. In: Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, 2022:1063–71.
Song R, Liu Z, Chen X, An H, Zhang Z, Wang X, Xu H. Label prompt for multi-label text classification. Appl Intell. 2022:1–15.
Faal F, Schmitt K, Yu JY. Reward modeling for mitigating toxicity in transformer-based language models. Appl Intell. 2022:1–15.
Zhang Z, Pang J, Xie X, Zhou Y. Research on crowdsourcing quality control strategies and evaluation algorithm. Chin J Comput. 2013;36(8):1636–49.
Liu X, Zheng Y, Du Z, Ding M, Qian Y, Yang Z, Tang J. Gpt understands, too. arXiv preprint arXiv:2103.10385 2021.
Sunday-Grève S, Williamson T. Philosophy and common sense 2: cultivating curiosity. Philosophers’ Magazine. 2022;96:24–30.
Acknowledgements
We thank all the anonymous reviewers for their constructive suggestions for the manuscript. This work was supported by the Natural Science Foundation of Shanghai (No. 23ZR1422800).
Author information
Authors and Affiliations
Contributions
All authors contributed to the main ideas present in this manuscript. The first draft of the manuscript were written by Chao Wang, Tao Chen and Jingping Liu. All authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Conflict of Interest
The authors have no relevant financial or non-financial interests to disclose.
Consent for Publication
The authors all agree to publish the article in Machine Learning.
Consent to Participate
Not Applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, C., Chen, T. & Liu, J. Don’t Ignore the Drive of Curiosity: Rethinking Subtleties Between Universality of Commonsense Knowledge and Excellence of Large Language Models. SN COMPUT. SCI. 5, 798 (2024). https://doi.org/10.1007/s42979-024-03165-w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s42979-024-03165-w