Nothing Special   »   [go: up one dir, main page]

Skip to main content
Log in

A relation-aware representation approach for the question matching system

  • Published:
World Wide Web Aims and scope Submit manuscript

Abstract

Online question matching is the process of comparing user queries with system questions to find appropriate answers. This task has become increasingly important with the popularity of knowledge sharing social networks in product search and intelligent Q &A in customer service. Many previous studies have focused on designing complex semantic structures through the questions themselves. In fact, the online user’s queries accumulate a large number of similar sentences, which have been grouped by semantics in the retrieval system. However, how to use these sentences to enhance the understanding of system questions is rarely studied. In this paper, we propose a novel Relation-aware Semantic Enhancement Network (RSEN) model. Specifically, we leverage the labels of the history records to identify different semantically related sentences. Then, we construct an expanded relation network to integrate the representation of different semantic relations. Furthermore, we interact we integrate the features of the system question with the semantically related sentences to augment the semantic information. Finally, we evaluate our proposed RSEN on two publicly available datasets. The results demonstrate the effectiveness of our proposed RSEN method compared to the advanced baselines.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Availability of data and materials

The two datasets used in this paper are public datasets. They are available from the following two URLs. The Quora dataset is available from this URL (https://www.kaggle.com/quora/question-pairs-dataset). And the BQ dataset is available from this URL (http://icrc.hitsz.edu.cn/Article/show/175.html).

Notes

  1. https://www.kaggle.com/quora/question-pairs-dataset

  2. http://icrc.hitsz.edu.cn/Article/show/175.html

References

  1. Li, H., Lu, Z.: Deep learning for information retrieval. In: Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1203–1206 (2016)

  2. Lan, W., Xu, W.: Neural network models for paraphrase identification, semantic textual similarity, natural language inference, and question answering. In: Proceedings of the 27th International Conference on Computational Linguistics, pp. 3890–3902 (2018)

  3. Karan, M., Šnajder, J.: Paraphrase-focused learning to rank for domain-specific frequently asked questions retrieval. Expert Syst. Appl. 91, 418–433 (2018)

    Article  Google Scholar 

  4. Guo, J., Fan, Y., Pang, L., Yang, L., Ai, Q., Zamani, H., Wu, C., Croft, W.B., Cheng, X.: A deep look into neural ranking models for information retrieval. Inf. Process. Manage. 57(6), 102067 (2020)

    Article  Google Scholar 

  5. Yang, R., Zhang, J., Gao, X., Ji, F., Chen, H.: Simple and effective text matching with richer alignment features. arXiv:1908.00300 (2019)

  6. Pang, L., Lan, Y., Guo, J., Xu, J., Wan, S., Cheng, X.: Text matching as image recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30 (2016)

  7. Mueller, J., Thyagarajan, A.: Siamese recurrent architectures for learning sentence similarity. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30 (2016)

  8. Severyn, A., Moschitti, A.: Learning to rank short text pairs with convolutional deep neural networks. In: Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 373–382 (2015)

  9. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. arXiv preprint arXiv:1706.03762 (2017)

  10. Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805 (2018)

  11. Shen, Y., Deng, Y., Yang, M., Li, Y., Du, N., Fan, W., Lei, K.: Knowledge-aware attentive neural network for ranking question answer pairs. In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 901–904 (2018)

  12. Yang, M., Chen, L., Chen, X., Wu, Q., Zhou, W., Shen, Y.: Knowledge-enhanced hierarchical attention for community question answering with multi-task and adaptive learning. In: IJCAI, pp. 5349–5355 (2019)

  13. Santoro, A., Raposo, D., Barrett, D.G.T., Malinowski, M., Pascanu, R., Battaglia, P., Lillicrap, T.: A simple neural network module for relational reasoning. (2017)

  14. Ghasemi, N., Momtazi, S.: Neural text similarity of user reviews for improving collaborative filtering recommender systems. Electron. Commer. Res. Appl. 45, 101019 (2020)

    Article  Google Scholar 

  15. Art, Y.I., Sanchoy, D.: Measuring design-level information quality in online reviews. Electron. Commer. Res. Appl. 30, 102–110 (2018)

    Article  Google Scholar 

  16. Ngai, E.W., Lee, M.C., Luo, M., Chan, P.S., Liang, T.: An intelligent knowledge-based chatbot for customer service: Electron. Commer. Res. Appl. 50, 101098 (2021)

    Article  Google Scholar 

  17. Fu, R., Guo, J., Qin, B., Che, W., Wang, H., Liu, T.: Learning semantic hierarchies: A continuous vector space approach. IEEE/ACM Trans. Audio Speech Lang. Process. 23(3), 461–471 (2015)

    Article  Google Scholar 

  18. Li, S., Luo, H., Zhao, G., Tang, M., Liu, X.: bi-directional bayesian probabilistic model based hybrid grained semantic matchmaking for web service discovery. World Wide Web 25(2), 445–470 (2022)

    Article  Google Scholar 

  19. Yin, W., Schütze, H., Xiang, B., Zhou, B.: Abcnn: Attention-based convolutional neural network for modeling sentence pairs. Trans. Assoc. Comput. Linguist. 4, 259–272 (2016)

    Article  Google Scholar 

  20. Lu, Z., Li, H.: A deep architecture for matching short texts. Adv Neural Inf Process Syst 26, 1367–1375 (2013)

    Google Scholar 

  21. Wang, Z., Hamza, W., Florian, R.: Bilateral multi-perspective matching for natural language sentences. arXiv:1702.03814 (2017)

  22. Cho, K., Courville, A., Bengio, Y.: Describing multimedia content using attention-based encoder-decoder networks. IEEE Trans. Multimedia 17(11), 1875–1886 (2015)

    Article  Google Scholar 

  23. Liu, Y., Sun, C., Lin, L., Wang, X.: Learning natural language inference using bidirectional lstm model and inner-attention. arXiv:1605.09090 (2016)

  24. Zhao, P., Lu, W., Wang, S., Peng, X., Jian, P., Wu, H., Zhang, W.: Multi-granularity interaction model based on pinyins and radicals for chinese semantic matching. World Wide Web 25(4), 1703–1723 (2022)

    Article  Google Scholar 

  25. Gu, Y., Gu, M., Long, Y., Xu, G., Yang, Z., Zhou, J., Qu, W.: An enhanced short text categorization model with deep abundant representation. World Wide Web 21, 1705–1719 (2018)

    Article  Google Scholar 

  26. Kim, S., Kang, I., Kwak, N.: Semantic sentence matching with densely-connected recurrent and co-attentive information. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 6586–6593 (2019)

  27. Parikh, A.P., Täckström, O., Das, D., Uszkoreit, J.: A decomposable attention model for natural language inference. arXiv:1606.01933 (2016)

  28. Wu, S., Dou, C., Wang, D., Li, J., Zhang, X., Feng, Z., Wang, K., Yitagesu, S.: Phrase-level attention network for few-shot inverse relation classification in knowledge graph. World Wide Web, 1–26 (2023)

  29. Chen, Q., Zhu, X., Ling, Z., Wei, S., Jiang, H., Inkpen, D.: Enhanced lstm for natural language inference. arXiv:1609.06038 (2016)

  30. Sakata, W., Shibata, T., Tanaka, R., Kurohashi, S.: Faq retrieval using query-question similarity and bert-based query-answer relevance. In: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1113–1116 (2019)

  31. Wang, B., Kuo, C.-C.J.: Sbert-wk: A sentence embedding method by dissecting bert-based word models. IEEE/ACM Trans. Audio Speech Lang. Process. 28, 2146–2157 (2020)

    Article  Google Scholar 

  32. Li, L., Zhang, M., Chao, Z., Xiang, J.: Using context information to enhance simple question answering. World Wide Web 24, 249–277 (2021)

    Article  Google Scholar 

  33. Li, L., Kong, M., Li, D., Zhou, D.: A multi-granularity semantic space learning approach for cross-lingual open domain question answering. World Wide Web 24(4), 1065–1088 (2021)

    Article  Google Scholar 

  34. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al.: Language models are unsupervised multitask learners. OpenAI Blog 1(8), 9 (2019)

    Google Scholar 

  35. Hu, N., Wu, Y., Qi, G., Min, D., Chen, J., Pan, J.Z., Ali, Z.: An empirical study of pre-trained language models in simple knowledge graph question answering. World Wide Web, 1–32 (2023)

  36. Pennington, J., Socher, R., Manning, C.D.: Glove: Global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1532–1543 (2014)

  37. Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, R., Le, Q.V.: Xlnet: Generalized autoregressive pretraining for language understanding. arXiv:1906.08237 (2019)

  38. Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., Stoyanov, V.: Roberta: A robustly optimized bert pretraining approach. arXiv:1907.11692 (2019)

  39. Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: Albert: A lite bert for self-supervised learning of language representations. arXiv:1909.11942 (2019)

  40. Sun, R., Li, Z., Liu, Q., Wang, Z., Duan, X., Huai, B., Yuan, N.J.: Multi-sentence matching via exploiting list-level semantics expansion. In: 2022 IEEE International Conference on Knowledge Graph (ICKG), pp. 249–256 (2022). IEEE

  41. Zou, Y., Liu, H., Gui, T., Wang, J., Zhang, Q., Tang, M., Li, H., Wang, D.: Divide and conquer: Text semantic matching with disentangled keywords and intents. arXiv:2203.02898 (2022)

  42. Wang, J., Chen, Z., Zhang, Y., He, D., Lin, F.: Preciser comparison: Augmented multi-layer dynamic contrastive strategy for text2text question classification. Neurocomputing 544, 126299 (2023)

    Article  Google Scholar 

  43. Maguitman, A.G., Menczer, F., Erdinc, F., Roinestad, H., Vespignani, A.: Algorithmic computation and approximation of semantic similarity. World Wide Web 9, 431–456 (2006)

    Article  Google Scholar 

  44. Gu, Y., Yang, Z., Xu, G., Nakano, M., Toyoda, M., Kitsuregawa, M.: Exploration on efficient similar sentences extraction. World Wide Web 17, 595–626 (2014)

    Article  Google Scholar 

  45. Zhang, H., Xiao, L., Chen, W., Wang, Y., Jin, Y.: Multi-task label embedding for text classification. arXiv:1710.07210 (2017)

  46. Pappas, N., Henderson, J.: Gile: A generalized input-label embedding for text classification. Trans. Assoc. Comput. Linguist. 7, 139–155 (2019)

    Article  Google Scholar 

  47. Wang, G., Li, C., Wang, W., Zhang, Y., Shen, D., Zhang, X., Henao, R., Carin, L.: Joint embedding of words and labels for text classification. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (vol. 1: Long Papers) (2018)

  48. Zhong, Q., Cui, M., Mai, H., Zhang, Q., Xu, S., Liu, X., Du, Y.: The short text matching model enhanced with knowledge via contrastive learning. arXiv:2304.03898 (2023)

  49. Shan, H., Zhang, Q., Liu, Z., Zhang, G., Li, C.: Beyond two-tower: Attribute guided representation learning for candidate retrieval. In: Proceedings of the ACM Web Conference 2023, pp. 3173–3181 (2023)

  50. Schlichtkrull, M., Kipf, T.N., Bloem, P., Van Den Berg, R., Titov, I., Welling, M.: Modeling relational data with graph convolutional networks. In: European Semantic Web Conference, pp. 593–607 (2018). Springer

  51. Mou, L., Men, R., Li, G., Xu, Y., Jin, Z.: Natural language inference by tree-based convolution and heuristic matching. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (ACL) (2016)

  52. Weeds, J., Clarke, D., Reffin, J., Weir, D., Keller, B.: Learning to distinguish hypernyms and co-hyponyms. In: Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics (2014)

  53. Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv:1412.6980 (2014)

  54. Reimers, N., Gurevych, I.: Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv:1908.10084 (2019)

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China under Grants (U20A20229), in part by the Natural Science Foundation of Xinjiang Uygur Autonomous Region of China (2022D01A227).

Funding

This work was supported by the National Natural Science Foundation of China under Grants (U20A20229), in part by the Natural Science Foundation of Xinjiang Uygur Autonomous Region of China (2022D01A227).

Author information

Authors and Affiliations

Authors

Contributions

Y.C designed the theoretical foundation and logical framework of the study, and was responsible for the experimental operations and data processing and analysis. She also conducted in-depth research on the experimental results and wrote the first draft of the paper. E.C was responsible for the development and implementation of the research methodology, and conducted in-depth theoretical discussions and practical analysis of the research questions. In addition, he carefully reviewed and revised the article. K.Z improved and deepened the theoretical foundation and logical framework of the article, and provided detailed guidance and suggestions on data presentation and processing. In addition, he conducted an in-depth study of the experimental results, and carefully reviewed and revised the whole article. Q.L refined and deepened the research methodology, and provided guidance on the theoretical foundation and logical framework of the study. In addition, he carefully reviewed and revised the full text. R.S. worked extensively on the data presentation and processing of the article, and conducted in-depth research and discussion of the experimental results. In addition, he meticulously revised and embellished the structure and language of the article.

Corresponding author

Correspondence to Enhong Chen.

Ethics declarations

Ethical approval

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chen, Y., Chen, E., Zhang, K. et al. A relation-aware representation approach for the question matching system. World Wide Web 27, 17 (2024). https://doi.org/10.1007/s11280-024-01255-6

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s11280-024-01255-6

Keywords

Navigation