ClueReader: Heterogeneous Graph Attention Network for Multi-Hop Machine Reading Comprehension
<p>Our proposed <span class="html-italic">ClueReader</span>: a heterogeneous graph attention network for multi-hop MRC. The detailed explanations of <span class="html-italic">S</span>, <span class="html-italic">C</span>, and <span class="html-italic">q</span> are in task formalization (<a href="#sec3dot1-electronics-12-03183" class="html-sec">Section 3.1</a>). <span class="html-italic">S</span>, <span class="html-italic">C</span>, and <span class="html-italic">q</span> are encoded in three independent <span class="html-italic">Bi-LSTMs</span> (<a href="#sec3dot2-electronics-12-03183" class="html-sec">Section 3.2</a>). Following the graph construction strategies in <a href="#sec3dot3-electronics-12-03183" class="html-sec">Section 3.3</a>, the outputs of three encoders are applied to <span class="html-italic">Co-attention</span> and <span class="html-italic">Self-attention</span> to initialize the reasoning graph features, which is explained in <a href="#sec3dot4-electronics-12-03183" class="html-sec">Section 3.4</a>. Then the topology information and node features are passed into the GAT layer. A much larger network computation behind <span class="html-italic">grandmother cells</span> is performed in the GAT layer, and n-hops message passing is calculated in n parameter shared layers, which are represented in <a href="#sec3dot4dot2-electronics-12-03183" class="html-sec">Section 3.4.2</a>. Finally, <span class="html-italic">grandmother cell</span> selectivity is combined in <a href="#sec3dot5-electronics-12-03183" class="html-sec">Section 3.5</a>, outputting the final predicted answer.</p> "> Figure 2
<p>Heterogeneous reasoning graph in <span class="html-italic">ClueReader</span>. Different nodes are filled in different colors, and the edges are distinguished by the types of lines. Subject nodes are gray, reasoning nodes are orange, mention nodes are green, support nodes are red, and candidate nodes are blue. The nodes in the light yellow square are all selected to input to the two MLP obtaining the prediction score distribution.</p> "> Figure 3
<p>Samples of <span class="html-small-caps">WikiHop</span> and <span class="html-small-caps">MedHop</span>. Subject entities, reasoning entities, mention entities, and candidate entities are shown in gray, orange, green, and blue colors, respectively. The occurrence of the correct answer is shown by a square frame outside. (<b>a</b>) A sample from the <span class="html-small-caps">WikiHop</span>. (<b>b</b>) A sample from the <span class="html-small-caps">MedHop</span>.</p> "> Figure 4
<p>Statistics of the model performance with different numbers of support documents on the <span class="html-small-caps">WikiHop</span> development set.</p> "> Figure 5
<p>Statistics of the model performance with different numbers of support documents on the <span class="html-small-caps">MedHop</span> development set.</p> "> Figure 6
<p>Visualizations of reasoning graphs on the <span class="html-small-caps">WikiHop</span> development set that are correctly answered. A thicker edge corresponds to a higher attention weight, and darker green nodes or darker blue nodes represent higher output values among the same type of nodes. (<b>a</b>–<b>f</b>) Visualized samples from the <span class="html-small-caps">WikiHop</span> development set.</p> "> Figure 7
<p>Visualizations of reasoning graphs on the <span class="html-small-caps">MedHop</span> development set that are correctly answered. A thicker edge corresponds to a higher attention weight, and darker green nodes or darker blue nodes represent higher output values among the same type of nodes. (<b>a</b>–<b>f</b>) Visualized samples from the <span class="html-small-caps">MedHop</span> development set.</p> "> Figure 8
<p>Generated HTML file of sample # 543 in <span class="html-small-caps">WikiHop</span> development set. The mark <b>MENMAX</b> means the final output of <math display="inline"><semantics><mrow><msub><mi mathvariant="bold">MLP</mi><mrow><mi>m</mi><mi>e</mi><mi>n</mi></mrow></msub></mrow></semantics></math>. For more details, please refer to <a href="https://cluereader.github.io/WH_dev_543.html" target="_blank">https://cluereader.github.io/WH_dev_543.html</a> (accessed on 21 June 2023).</p> ">
Abstract
:1. Introduction
- Selectivity. The grandmother cells concept organizes the neurons in a hierarchical “sparse” coding scheme. It activates some specific neurons to respond to stimulation, similar to the manner in which we store reasoning evidence maps (neurons) in our minds during reading and recall-related evidence maps to reason the answer with a question (stimulation) constrained.
- Specificity. The concept implies that brains contain grandmother neurons that are so specialized and dedicated to a specific object, which is similar to a particular MRC question resulting in a specific answer among multiple reading passages and their complex reasoning evidence.
- Class character. Amazing selectivity is captured in grandmother cells. However, it results from computation by much larger networks and the collective operations of many functionally different low-level cells, similar to human multi-hop reading in which evidence is usually gathered from different levels as much as possible and the final answer is decided in some candidate endpoints.
- In order to construct a more reasonable graph, ClueReader draws inspiration from the concept of grandmother cells in the brain during information cognition, in which cells in the brain only output specific entities. This leads to the creation of heterogeneous graph attention networks with multiple types of nodes.
- By taking the subject of queries as the starting point, potential reasoning entities in multiple documents as bridge points, and mention entities consistent with candidate answers as end points, the proposed ClueReader is a heuristic way of constructing MRC chains.
- Before outputting predicted answers, ClueReader innovatively visualizes the internal state of the heterogeneous graph attention network, providing intuitive quantitative data displays for analyzing the effectiveness, rationality, and explainability.
2. Related Work
2.1. Sequential Reading Models for Multi-Hop MRC
2.2. Graph Neural Networks for Multi-Hop MRC
3. Methodology
3.1. Task Formalization
3.2. Encoding Layer
3.3. Heterogeneous Reasoning Graph
- The query (or the question) locates the related neurons at a low level, which then stimulates higher-level neurons to trigger computation;
- The higher-level neurons begin to respond to increasingly broader portions of other neurons for reasoning, and to avoid a broadcast storm, informative selectivity takes place in this step;
- At the top-level, some independent neurons are responsible for the computations that occurred in step 2. We refer to these neurons as grandmother cells and expect them to provide the appropriate results that correspond to the query.
3.3.1. Nodes Definition
- Subject Nodes—As the form of query q, the subject entity s is given in . For example, the subject entity of the query sequence context Where is the basketball team that Mike DiNunno plays for based? is certainly Mike DiNuuno. We extract all the named entities that match with s from documents, and regard them as the subject nodes to open up the reading clues triggering the further computations. The subject nodes are denoted as and colored in gray in Figure 2.
- Reasoning Nodes—In light of the requirements of the multi-hop MRC, there are some gaps between the subject entities and candidates. To build bridges between the two and make the reasoning clues as complete as possible, we replenish those clues with the named recognition entities and nominal phrases from the documents containing the question subjects and answer candidates. The reasoning nodes are marked as and colored in orange in Figure 2.
- Mention Nodes—A series of candidate entities are given in , they may occur in multiple times within the document set . As a result, we traverse the documents and extract the named entities corresponding to each candidate as mention nodes, serving as the soft endpoint of the reasoning chain. It should be noted that mention nodes will participate in the semi-supervised learning process and will be involved in the final answer prediction. The mention nodes are presented as and colored in green in Figure 2.
- Candidate Nodes—To imitate grandmother cells, we consider candidate nodes as hard endpoints of the reasoning chain to gather relevant information from the heterogeneous reasoning graph. For the mention nodes of a candidate answer , when , candidate nodes are established as grandmother cells to provide the final prediction. The candidate nodes are denoted as and colored in blue in Figure 2.
3.3.2. Edges Definition
3.3.3. Graph Construction
3.4. Heterogeneous Graph Attention Network for Multi-Hop Reading
3.4.1. Query-Aware Contextual Information
3.4.2. Message Passing in the Heterogeneous Graph Attention Network
3.4.3. Gating Mechanism
3.5. Output Layer
4. Experiments
4.1. Dataset for Experiments
- Whether they knew the fact before;
- Whether the fact follows from the texts (with options follows, likely, and not follows);
- Whether multiple documents are required to answer the question.
4.2. Experiments Settings
4.3. Results and Analyses
4.4. Ablation Study
4.5. Visualization
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Wang, Y.; Liu, K.; Liu, J.; He, W.; Lyu, Y.; Wu, H.; Li, S.; Wang, H. Multi-Passage Machine Reading Comprehension with Cross-Passage Answer Verification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 1918–1927. [Google Scholar]
- Dai, Y.; Fu, Y.; Yang, L. A multiple-choice machine reading comprehension model with multi-granularity semantic reasoning. Appl. Sci. 2021, 11, 7945. [Google Scholar] [CrossRef]
- Hassabis, D.; Kumaran, D.; Summerfield, C.; Botvinick, M. Neuroscience-inspired artificial intelligence. Neuron 2017, 95, 245–258. [Google Scholar] [CrossRef] [Green Version]
- Page, M. Connectionist modelling in psychology: A localist manifesto. Behav. Brain Sci. 2000, 23, 443–467. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Dehaene, S. Reading in the Brain: The New Science of How We Read; Penguin: London, UK, 2010. [Google Scholar]
- Quiroga, R.Q.; Reddy, L.; Kreiman, G.; Koch, C.; Fried, I. Invariant visual representation by single neurons in the human brain. Nature 2005, 435, 1102–1107. [Google Scholar] [CrossRef] [Green Version]
- Battaglia, P.W.; Hamrick, J.B.; Bapst, V.; Sanchez-Gonzalez, A.; Zambaldi, V.; Malinowski, M.; Tacchetti, A.; Raposo, D.; Santoro, A.; Faulkner, R.; et al. Relational inductive biases, deep learning, and graph networks. arXiv 2018, arXiv:1806.01261. [Google Scholar]
- Yu, W.; Chang, T.; Guo, X.; Wang, M.; Wang, X. An interaction-modeling mechanism for context-dependent Text-to-SQL translation based on heterogeneous graph aggregation. Neural Netw. 2021, 142, 573–582. [Google Scholar] [CrossRef] [PubMed]
- Seo, M.; Kembhavi, A.; Farhadi, A.; Hajishirzi, H. Bidirectional Attention Flow for Machine Comprehension. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
- Cui, Y.; Chen, Z.; Wei, S.; Wang, S.; Liu, T.; Hu, G. Attention-over-Attention Neural Networks for Reading Comprehension. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017; pp. 593–602. [Google Scholar]
- Zhao, Y.; Wang, L.; Wang, C.; Du, H.; Wei, S.; Feng, H.; Yu, Z.; Li, Q. Multi-granularity heterogeneous graph attention networks for extractive document summarization. Neural Netw. 2022, 155, 340–347. [Google Scholar] [CrossRef] [PubMed]
- Li, S.; Sun, C.; Liu, B.; Liu, Y.; Ji, Z. Modeling Extractive Question Answering Using Encoder-Decoder Models with Constrained Decoding and Evaluation-Based Reinforcement Learning. Mathematics 2023, 11, 1624. [Google Scholar] [CrossRef]
- Welbl, J.; Stenetorp, P.; Riedel, S. Constructing datasets for multi-hop reading comprehension across documents. Trans. Assoc. Comput. Linguist. 2018, 6, 287–302. [Google Scholar] [CrossRef] [Green Version]
- Kenton, J.D.M.W.C.; Toutanova, L.K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the NAACL-HLT, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
- Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A robustly optimized bert pretraining approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
- Beltagy, I.; Peters, M.E.; Cohan, A. Longformer: The long-document transformer. arXiv 2020, arXiv:2004.05150. [Google Scholar]
- Razeghi, Y.; Logan, R.L., IV; Gardner, M.; Singh, S. Impact of pretraining term frequencies on few-shot reasoning. arXiv 2022, arXiv:2202.07206. [Google Scholar]
- Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
- Zhang, C.; Zha, D.; Wang, L.; Mu, N.; Yang, C.; Wang, B.; Xu, F. Graph Convolution Network over Dependency Structure Improve Knowledge Base Question Answering. Electronics 2023, 12, 2675. [Google Scholar] [CrossRef]
- Ding, M.; Zhou, C.; Chen, Q.; Yang, H.; Tang, J. Cognitive Graph for Multi-Hop Reading Comprehension at Scale. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 2694–2703. [Google Scholar]
- Evans, J.S.B. Heuristic and analytic processes in reasoning. Br. J. Psychol. 1984, 75, 451–468. [Google Scholar] [CrossRef]
- Sloman, S.A. The empirical case for two systems of reasoning. Psychol. Bull. 1996, 119, 3. [Google Scholar] [CrossRef]
- De Cao, N.; Aziz, W.; Titov, I. Question Answering by Reasoning Across Documents with Graph Convolutional Networks. In Proceedings of the 2019 Annual Conference of the North American Chapter of the Association for Computational Linguistics. Association for Computational Linguistics, Minneapolis, MN, USA, 3–7 June 2019; pp. 2306–2317. [Google Scholar]
- Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep Contextualized Word Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers). Association for Computational Linguistics, New Orleans, LA, USA, 1–6 June 2018; pp. 2227–2237. [Google Scholar]
- Cao, Y.; Fang, M.; Tao, D. BAG: Bi-directional Attention Entity Graph Convolutional Network for Multi-hop Reasoning Question Answering. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 3–7 June 2019; pp. 357–362. [Google Scholar]
- Tang, Z.; Shen, Y.; Ma, X.; Xu, W.; Yu, J.; Lu, W. Multi-hop reading comprehension across documents with path-based graph convolutional network. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 3905–3911. [Google Scholar]
- Tu, M.; Wang, G.; Huang, J.; Tang, Y.; He, X.; Zhou, B. Multi-hop Reading Comprehension across Multiple Documents by Reasoning over Heterogeneous Graphs. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 2704–2713. [Google Scholar]
- Jia, M.; Liao, L.; Wang, W.; Li, F.; Chen, Z.; Li, J.; Huang, H. Keywords-aware dynamic graph neural network for multi-hop reading comprehension. Neurocomputing 2022, 501, 25–40. [Google Scholar] [CrossRef]
- Zhang, Y.; Meng, F.; Zhang, J.; Chen, Y.; Xu, J.; Zhou, J. MKGN: A Multi-Dimensional Knowledge Enhanced Graph Network for Multi-Hop Question and Answering. IEICE Trans. Inf. Syst. 2022, 105, 807–819. [Google Scholar] [CrossRef]
- Song, L.; Wang, Z.; Yu, M.; Zhang, Y.; Florian, R.; Gildea, D. Evidence integration for multi-hop reading comprehension with graph neural networks. IEEE Trans. Knowl. Data Eng. 2020, 34, 631–639. [Google Scholar] [CrossRef]
- Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
- Zhou, P.; Shi, W.; Tian, J.; Qi, Z.; Li, B.; Hao, H.; Xu, B. Attention-based bidirectional long short-term memory networks for relation classification. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Berlin, Germany, 7–12 August 2016; pp. 207–212. [Google Scholar]
- Tang, H.; Li, H.; Liu, J.; Hong, Y.; Wu, H.; Wang, H. DuReader_robust: A Chinese Dataset Towards Evaluating Robustness and Generalization of Machine Reading Comprehension in Real-World Applications. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), Virtual, 1–6 August 2021; pp. 955–963. [Google Scholar]
- Zhong, V.; Xiong, C.; Keskar, N.S.; Socher, R. Coarse-grain Fine-grain Coattention Network for Multi-evidence Question Answering. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April– 3 May 2018. [Google Scholar]
- Gilmer, J.; Schoenholz, S.S.; Riley, P.F.; Vinyals, O.; Dahl, G.E. Neural message passing for quantum chemistry. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1263–1272. [Google Scholar]
- Bird, S. NLTK: The natural language toolkit. In Proceedings of the COLING/ACL 2006 Interactive Presentation Sessions, Sydney, Australia, 17–18 July 2006; pp. 69–72. [Google Scholar]
- Hashimoto, K.; Xiong, C.; Tsuruoka, Y.; Socher, R. A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 9–11 September 2017; pp. 1923–1933. [Google Scholar]
- Fey, M.; Lenssen, J.E. Fast graph representation learning with PyTorch Geometric. arXiv 2019, arXiv:1903.02428. [Google Scholar]
- Hagberg, A.; Swart, P.; Chult, D.S. Exploring Network Structure, Dynamics, and Function Using NetworkX; Technical Report; Los Alamos National Laboratory: Los Alamos, NM, USA, 2008. [Google Scholar]
- Dhingra, B.; Jin, Q.; Yang, Z.; Cohen, W.; Salakhutdinov, R. Neural Models for Reasoning over Multiple Mentions Using Coreference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), New Orleans, LA, USA, 1–6 June 2018; pp. 42–48. [Google Scholar]
Edges | Definition |
---|---|
If the support document contains the j-th subject node , an undirected edge denoted as is established to connect the support node of and the subject node . | |
If the support document contains the j-th candidate node , an undirected edge denoted as is established to connect the support node of and the candidate node . | |
If the support document contains the j-th mention node , an undirected edge denoted as is established to connect the support node of and the mention node . | |
If the j-th mention node and the i-th candidate node represent the same entity, an undirected edge denoted as is established to connect the two nodes. | |
If the i-th subject node and the j-th reasoning node extracted from the same document, an undirected edge denoted as is established to connect the two nodes. | |
If the i-th reasoning node and the j-th mention node extracted from the same document, an undirected edge denoted as is established to connect the two nodes. | |
All the mention nodes are fully connected using undirected edge . | |
If two mention nodes and are extracted from the same document, the two nodes will be connected as . | |
If two mention nodes and are extracted from different documents represent the same entity, the two nodes will be connected as . | |
If two reasoning nodes and are extracted from the same document or represent the same entity, the two nodes will be connected as . |
Training | Development | Test | Total | |
---|---|---|---|---|
WikiHop | 43,738 | 5129 | 2451 | 51,318 |
MedHop | 1620 | 342 | 546 | 2508 |
Single Models | WikiHop Accuracy (%) | MedHop Accuracy (%) | ||
---|---|---|---|---|
Dev | Test | Dev | Test | |
Coref-GRU [43] | 56.0 | 59.3 | - | - |
MHQA-GRN [31] | 62.8 | 65.4 | - | - |
Entity-GCN [24] | 64.8 | 67.6 | - | - |
HDE [28] | 68.1 | 70.9 | - | - |
BAG [26] | 66.5 | 69.0 | - | - |
Path-based GCN [27] | 64.5 | - | - | - |
Document-cue [13] | - | 36.7 | - | 44.9 |
FastQA [13] | - | 25.7 | - | 23.1 |
TF-IDF [13] | - | 25.6 | - | 9.0 |
BiDAF [13] | - | 42.9 | - | 47.8 |
ClueReader | 66.5 | 72.0 | 48.2 | 46.0 |
Annotation | Accuracy (%) | |
---|---|---|
follows fact | requires multiple documents | 74.9 |
requires single document | 74.0 | |
likely follows fact | requires multiple documents | 71.4 |
requires single document | 71.4 | |
not follows is not given | 71.5 |
Model | Accuracy (%) | |||
---|---|---|---|---|
WikiHop | MedHop | |||
Full Model | 71.45 | - | 48.25 | - |
52.69 | 18.76 | 37.72 | 10.53 | |
70.95 | 0.5 | 47.37 | 0.88 | |
63.34 | 8.11 | 4.97 | 43.28 | |
70.77 | 0.68 | 47.37 | 0.88 | |
62.02 | 9.43 | 48.54 | −0.29 | |
65.87 | 5.58 | 44.77 | 3.48 |
Hyperparameters | Value | Acc. of WikiHop | Acc. of MedHop |
---|---|---|---|
l | 3 | 57.8 | 42.4 |
4 | 58.5 | 43.3 | |
5 | 66.5 | 48.2 | |
6 | 64.2 | 45.0 | |
0 | 59.7 | 42.7 | |
0.5 | 66.1 | 44.2 | |
1.0 | 66.5 | 48.2 | |
1.5 | 59.1 | 43.3 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gao, P.; Gao, F.; Wang, P.; Ni, J.-C.; Wang, F.; Fujita, H. ClueReader: Heterogeneous Graph Attention Network for Multi-Hop Machine Reading Comprehension. Electronics 2023, 12, 3183. https://doi.org/10.3390/electronics12143183
Gao P, Gao F, Wang P, Ni J-C, Wang F, Fujita H. ClueReader: Heterogeneous Graph Attention Network for Multi-Hop Machine Reading Comprehension. Electronics. 2023; 12(14):3183. https://doi.org/10.3390/electronics12143183
Chicago/Turabian StyleGao, Peng, Feng Gao, Peng Wang, Jian-Cheng Ni, Fei Wang, and Hamido Fujita. 2023. "ClueReader: Heterogeneous Graph Attention Network for Multi-Hop Machine Reading Comprehension" Electronics 12, no. 14: 3183. https://doi.org/10.3390/electronics12143183
APA StyleGao, P., Gao, F., Wang, P., Ni, J. -C., Wang, F., & Fujita, H. (2023). ClueReader: Heterogeneous Graph Attention Network for Multi-Hop Machine Reading Comprehension. Electronics, 12(14), 3183. https://doi.org/10.3390/electronics12143183