Information Retrieval 27th China Conference CCIR 2021 Dalian China October 29 31 2021 Proceedings 1st Edition Hongfei Lin Min Zhang Liang Pang
Information Retrieval 27th China Conference CCIR 2021 Dalian China October 29 31 2021 Proceedings 1st Edition Hongfei Lin Min Zhang Liang Pang
Information Retrieval 27th China Conference CCIR 2021 Dalian China October 29 31 2021 Proceedings 1st Edition Hongfei Lin Min Zhang Liang Pang
com
https://ebookmeta.com/product/information-
retrieval-27th-china-conference-ccir-2021-dalian-
china-october-29-31-2021-proceedings-1st-edition-
hongfei-lin-min-zhang-liang-pang-2/
OR CLICK BUTTON
DOWLOAD EBOOK
https://ebookmeta.com/product/information-retrieval-27th-china-
conference-ccir-2021-dalian-china-
october-29-31-2021-proceedings-1st-edition-hongfei-lin-min-zhang-
liang-pang-2/
https://ebookmeta.com/product/logic-and-argumentation-4th-
international-conference-clar-2021-hangzhou-china-
october-20-22-2021-proceedings-1st-edition-pietro-baroni/
https://ebookmeta.com/product/evolutionary-multi-criterion-
optimization-11th-international-conference-emo-2021-shenzhen-
china-march-28-31-2021-proceedings-1st-edition-hisao-ishibuchi/
Pattern Recognition and Computer Vision 4th Chinese
Conference PRCV 2021 Beijing China October 29 November
1 2021 Proceedings Part IV Lecture Notes in Computer
Science Huimin Ma (Editor)
https://ebookmeta.com/product/pattern-recognition-and-computer-
vision-4th-chinese-conference-prcv-2021-beijing-china-
october-29-november-1-2021-proceedings-part-iv-lecture-notes-in-
computer-science-huimin-ma-editor/
https://ebookmeta.com/product/machine-translation-17th-china-
conference-ccmt-2021-xining-china-october-8-10-2021-revised-
selected-papers-1st-edition-jinsong-su-rico-sennrich/
https://ebookmeta.com/product/machine-translation-17th-china-
conference-ccmt-2021-xining-china-october-8-10-2021-revised-
selected-papers-1st-edition-jinsong-su-rico-sennrich-2/
https://ebookmeta.com/product/natural-language-processing-and-
chinese-computing-10th-ccf-international-conference-
nlpcc-2021-qingdao-china-october-13-17-2021-proceedings-part-i-
lu-wang/
https://ebookmeta.com/product/health-information-science-10th-
international-conference-his-2021-melbourne-vic-australia-
october-25-28-2021-proceedings-siuly-siuly-editor/
Hongfei Lin
Min Zhang
Liang Pang (Eds.)
LNCS 13026
Information Retrieval
27th China Conference, CCIR 2021
Dalian, China, October 29–31, 2021
Proceedings
Lecture Notes in Computer Science 13026
Founding Editors
Gerhard Goos
Karlsruhe Institute of Technology, Karlsruhe, Germany
Juris Hartmanis
Cornell University, Ithaca, NY, USA
Information Retrieval
27th China Conference, CCIR 2021
Dalian, China, October 29–31, 2021
Proceedings
123
Editors
Hongfei Lin Min Zhang
Dalian University of Technology Department of Computer Science
Dalian, China Tsinghua University
Beijing, China
Liang Pang
Institute of Computing Technology
Chinese Academy of Sciences
Beijing, China
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
General Chairs
Hong Liu Dalian University of Foreign Languages, China
Jiafeng Guo Institute of Computing Technology, Chinese Academy
of Sciences, China
Proceedings Chairs
Ruihua Qi Dalian University of Foreign Languages, China
Liang Yang Dalian University of Technology, China
Publicity Chair
Zhumin Chen Shandong University, China
Publication Chair
Liang Pang Institute of Computing Technology, Chinese Academy
of Sciences, China
Webmaster
Yuan Lin Dalian University of Technology, China
Sponsorship Chairs
Qi Zhang Fudan University, China
Zhongyuan Wang Kwai Inc., China
Treasurers
Kan Xu Dalian University of Technology, China
Song Yang Dalian University of Foreign Languages, China
Award Chair
Ru Li Shanxi University, China
Program Committee
Ting Bai Beijing University of Posts and Telecommunications,
China
Fei Cai National University of Defense Technology, China
Jiawei Chen University of Science and Technology of China, China
Xiaoliang Chen Xihua University, China
Yubo Chen Institute of Automation, Chinese Academy of Sciences,
China
Zhumin Chen Shandong University, China
Zhicheng Dou Renmin University of China, China
Yajun Du Xihua University, China
Yixing Fan Institute of Computing Technology, Chinese Academy
of Sciences, China
Shengxiang Gao Kunming University of Science and Technology, China
Zhongyuan Han Foshan University, China
Yanbin Hao University of Science and Technology of China, China
Ben He University of Chinese Academy of Sciences, China
Xiangnan He University of Science and Technology of China, China
Yu Hong Soochow University, China
Zhuoren Jiang Zhejiang University, China
Ting Jin Hainan University, China
Yanyan Lan Tsinghua University, China
Chenliang Li Wuhan University, China
Lishuang Li Dalian University of Technology, China
Ru Li Shanxi University, China
Shangsong Liang Sun Yat-sen University, China
Xiangwen Liao Fuzhou University, China
Hongfei Lin Dalian University of Technology, China
Yuan Lin Dalian University of Technology, China
Chang Liu Peking University, China
Peiyu Liu Shandong Normal University, China
Organization ix
NLP for IR
IR in Education
IR in Biomedicine
1 Introduction
Users tend to issue short queries in search engines. These short queries are usu-
ally ambiguous or vague [12,16,26,27]. Taking the query “apple” as an example,
the actual user intents behind the query can be either the fruit “apple” or “Apple
Company”. Besides, a user intent can also cover multiple aspects (such as “how
to learn JAVA” or “download JAVA IDE” for the intent “JAVA programming
language”). In order to satisfy those diversified user intents, the technology of
search result diversification is necessary for search engines. The ranking models
of search result diversification aims at re-ranking the result documents to satisfy
diversified user intents at former ranking positions. Depending on whether to
model the user intent coverage explicitly, previous studies can be categorized
into implicit and explicit diversification methods. The implicit diverse ranking
c Springer Nature Switzerland AG 2021
H. Lin et al. (Eds.): CCIR 2021, LNCS 13026, pp. 3–15, 2021.
https://doi.org/10.1007/978-3-030-88189-4_1
4 X. Qin et al.
2 Related Work
We briefly review some related work about search result diversification and neu-
ral matching in different tasks.
Search Result Diversification. The earliest typical diversification models
is Max Margin Relevance (MMR) [4]. It compares each candidate document
with selected document sequence, greedily selects the document with the best
ranking score, and appends it to the selected sequence. The ranking scores of
documents are computed based on their relevance to the query and the nov-
elty compared with the selected documents. The “novelty” here is measured
by the dissimilarity between documents. The original MMR uses handcrafted
features and scoring functions to calculate the similarity, which limits its appli-
cation. Many approaches extend MMR by using supervised learning methods to
learn the features and functions automatically (e.g., SVM-DIV [33], R-LTR [36],
PAMM [30], and PAMM-NTN [31]). These methods are called implicit diversi-
fication approaches. On the contrary, explicit diversification approaches model
the coverage of different user intents (represented as subtopics) of each docu-
ment. A novel document is expected to cover new user intents which have not
been covered by selected sequence. Several unsupervised and supervised explicit
diversification approaches have been proposed, e.g., xQuAD [24], PM2 [9],
HxQuAD, HPM2 [14], and DSSA [17]. Recently, a group of new approaches
(e.g., DVGAN [21] and DESA [22]) have been proposed as ensemble approaches.
They use both implicit inter-document features and explicit subtopic coverage
features.
Note that most of these approaches use document representations pre-trained
by unsupervised tools, such as doc2vec [20] and LDA [3]. The document similar-
ity is computed by cosine similarity of two document embeddings. Different from
these methods, we represent documents at term-level, based on which an inter-
action is conducted. Therefore, our method can capture fine-grained matching
signals which are more accurate in computing document similarity.
Neural Matching in Different Tasks. In recent years, researchers have pro-
posed a group of deep-learning based relevance matching models for multiple IR
tasks. Compared with traditional approaches, these neural matching methods
can better measure the semantic similarity between queries and documents. In
general, these methods can be divided into two categories: representation-based
methods [15,25] and interaction-based methods [8,32]. The representation-based
methods use neural networks to generate the dense vector representations of
the queries and documents and compute their similarity based on the repre-
sentations. In contrast, the interaction-based methods first capture term-level
interaction signals between queries and documents and then aggregate them to
compute the similarity. In the view of neural matching, the previous diversifi-
cation approaches can be seen as representation-based approaches. In addition
to the ranking task, there are also a group of studies [28,35] leveraging neu-
ral models to measure the similarity between dialogue context and response
6 X. Qin et al.
Ed Relevance
v1 Features
Candidate
Document Ranking
xd
Cross-Attention
MLP
d Score
s
BERT
RNN
RNN
vi vd
Selected
Document
Idi
List
vm
[d1 , · · · , dm ]
Edi
3 Methodology
In this section, we first give the formulation of search result diversification prob-
lem. Then, we introduce the overall structure of our framework and describe
the details of each component. Finally, we describe the training and inference
process.
3.2 MatchingDIV
Ed = Linear(Norm(BERT([D]))), (2)
where Ed ∈ Rld ×h , and ld is the length of the document. [D] denotes the word-
pieces of documents after tokenization, and Norm(·) denotes the operation of
normalization. “BERT” denotes a BERT-like encoder which can be replaced by
other pretrained models, such as DistilBERT [23] or ELECTRA [6]. Following
the previous work [18], we apply a linear projection layer to compress the term
representation into h-dimension for reducing the storage cost.
Interaction via Cross-Attention. To capture the fine-grained matching sig-
nals, we use cross-attention as the interaction function to let each document in
the selected document sequence interact with the candidate document. Similar
to the self-attention which is widely used in Transformer-based models [11,29],
the cross-attention operation is also based on multi-head attention (MHA):
qK
Attn(q, K, V) = Softmax( √ )V, (3)
d
MHA(q, K, V) = [a1 ; . . . ; ah ], (4)
ai = Attn(qWQ K V
i , KWi , VWi ), i ∈ [1, h]. (5)
Due to space limitation, we omit the details of multi-head attention, which can
be referred to at [29]. For the selected document di in the sequence C and the
candidate document d, the interacted representations of di can be defined as:
The last hidden state hi,ldi is used as the integrated representation of the docu-
ment di . To simplify the notation, we use vi = hi,ldi , and it contains matching
signals between the selected document di and the candidate document d.
Afterwards, when obtaining integrated representations of all selected doc-
uments, we employ another RNN to aggregate the information of the whole
selected documents sequence as:
We use the last hidden state hd,|C| to represent the selected document sequence.
This vector contains the matching information between the candidate document
and the selected document sequence, and it is denoted as vd for simplification.
Note that in practice, we use GRU cells for all RNNs.
Ranking Score. Inheriting the spirit of MMR [4], the final ranking score is cal-
culated based on both the relevance and the novelty. For a candidate document
d, its ranking score s is calculated as:
Loss Function. In the process of training, we use the sum of all the documents’
ranking score si as the score sr of a given ranking sequence r. Following previous
work [17,22], we apply a list-pairwise sampling approach to generate training
samples in limited datasets. With the positive and negative ranking pair (r1 , r2 ),
MatchingDIV 9
the loss function for list-pairwise samples is defined as a binary classification log-
loss formation:
L= |ΔM|[ys log(P (r1 , r2 )) + (1 − ys ) log(1 − P (r1 , r2 ))]. (10)
q∈Q s∈Sq
Here |ΔM| = |M(r1 ) − M(r2 )|, and P (r1 , r2 ) = σ(sr1 − sr2 ), where σ(·) is
the sigmoid function. Due to space limitation, we omit the detailed introduc-
tion of list-pairwise sampling method, and more details can be found in [17].
MatchingDIV is optimized in an end-to-end manner, where the BERT encoder
is fine-tuned, and the other components are trained from scratch.
Reducing Online Inference Latency. As MatchingDIV leverages BERT,
which is a large model, to encode the documents, for online ranking tasks, we
propose two strategies to reduce the ranking latency.
4 Experiment Settings
4.1 Datasets and Metrics
We use the Web Track dataset from TREC 2009 to 2012 with 198 queries in total.
The query #95 and #100 without diversity judgements are not used. Each query
includes 3 to 8 user intent annotations, and the relevance rating is marked as
10 X. Qin et al.
In the training phase, we use 5-fold cross validation to tune the parameters in
all experiments with the widely used α-nDCG@20. In each fold, there are 160
queries for training and 40 queries for testing. In our experiment, the hidden
size of GRU is 128, and the BERT-based embeddings are compressed into 128
dimension. The batch size is 32. We use Adam [19] optimizer. The learning rate
of the BERT encoder is 3e−5, while that of other network components is 1e−3.
We compare MatchingDIV with baselines including:
1
https://github.com/jzbjyb/DSSA.
MatchingDIV 11
5 Experimental Results
5.1 Overall Results and Discussion
Table 1 shows the results of all models. We can observe: (1) MatchingDIV out-
performs all the implicit and explicit baseline models, and the improvement is
statistically significant (with p-value <0.05) on all the metrics except for Pre-IA.
These results clearly demonstrate the effectiveness of our proposed RMEDiv.
(2) Compared with those approaches based on pre-trained document embed-
dings, our framework can capture and aggregate the fine-grained matching sig-
nals between selected and candidate documents, thus improving the performance
of search result diversification. (3) Intriguingly, MatchingDIV, as an implicit
method without using subtopic coverage, can perform slightly better than the
ensemble approaches (DVGAN and DESA). This reflects that our proposed
interaction-based document matching is very effective. Besides, this result also
implies the advantage of enhancing the relevance matching component for diver-
sification. (4) Pre-trained language models (such as BERT) are reported to have
great capability of representation. By integrating it into the baseline DSSA, we
see a slight performance improvement. However, there is still a large gap between
Table 1. Performance of all approaches. The baselines include: (1) non-diversed meth-
ods; (2) explicit methods; (3) implicit methods; and (4) ensemble methods. The best
results are in bold. † indicates that our model significantly outperforms all implicit and
explicit approaches (p-value <0.05 in two-tailed paired t-test).
the performance of DSSA (BERT) and that of our proposed MatchingDIV. This
indicates that the better performance we obtained is not merely due to BERT
embeddings but also to the proposed interaction-based document matching.
As we introduced in Sect. 3.3, the inference latency is very important when apply-
ing a diversification model in practice. On the one hand, MatchingDIV employs
2
https://github.com/huggingface/transformers.
MatchingDIV 13
References
1. Agrawal, R., Gollapudi, S., Halverson, A., Ieong, S.: Diversifying search results.
In: WSDM (2009)
14 X. Qin et al.
2. Baeza-Yates, R., Hurtado, C., Mendoza, M.: Query recommendation using query
logs in search engines. In: Lindner, W., Mesiti, M., Türker, C., Tzitzikas, Y., Vakali,
A.I. (eds.) EDBT 2004. LNCS, vol. 3268, pp. 588–596. Springer, Heidelberg (2004).
https://doi.org/10.1007/978-3-540-30192-9 58
3. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet allocation. J. Mach. Learn.
Res. 3, 993–1022 (2003)
4. Carbonell, J.G., Goldstein, J.: The use of MMR, diversity-based reranking for
reordering documents and producing summaries. In: SIGIR (1998)
5. Chapelle, O., Metlzer, D., Zhang, Y., Grinspan, P.: Expected reciprocal rank for
graded relevance. In: CIKM. ACM (2009)
6. Clark, K., Luong, M., Le, Q.V., Manning, C.D.: ELECTRA: pre-training text
encoders as discriminators rather than generators. In: ICLR. OpenReview.net
(2020)
7. Clarke, C.L.A., et al.: Novelty and diversity in information retrieval evaluation. In:
SIGIR. ACM (2008)
8. Dai, Z., Xiong, C., Callan, J., Liu, Z.: Convolutional neural networks for soft-
matching n-grams in ad-hoc search. In: WSDM. ACM (2018)
9. Dang, V., Croft, W.B.: Diversity by proportionality: an election-based approach
to search result diversification. In: SIGIR (2012)
10. Dang, V., Croft, W.B.: Term level search result diversification. In: SIGIR (2013)
11. Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidi-
rectional transformers for language understanding. In: NAACL-HLT. Association
for Computational Linguistics (2019)
12. Dou, Z., Song, R., Wen, J.: A large-scale evaluation and analysis of personalized
search strategies. In: WWW. ACM (2007)
13. Guo, J., et al.: A deep look into neural ranking models for information retrieval.
Inf. Process. Manag. 57(6), 102067 (2020)
14. Hu, S., Dou, Z., Wang, X., Sakai, T., Wen, J.: Search result diversification based
on hierarchical intents. In: CIKM (2015)
15. Huang, P., He, X., Gao, J., Deng, L., Acero, A., Heck, L.P.: Learning deep struc-
tured semantic models for web search using clickthrough data. In: CIKM. ACM
(2013)
16. Jansen, B.J., Spink, A., Saracevic, T.: Real life, real users, and real needs: a study
and analysis of user queries on the web. Inf. Process. Manag. 36(2), 207–227 (2000)
17. Jiang, Z., Wen, J., Dou, Z., Zhao, W.X., Nie, J., Yue, M.: Learning to diversify
search results via subtopic attention. In: SIGIR (2017)
18. Khattab, O., Zaharia, M.: ColBERT: efficient and effective passage search via
contextualized late interaction over BERT. In: SIGIR. ACM (2020)
19. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Bengio, Y.,
LeCun, Y. (eds.) ICLR (2015)
20. Le, Q., Mikolov, T.: Distributed representations of sentences and documents. In:
ICML 2014 (2014)
21. Liu, J., Dou, Z., Wang, X., Lu, S., Wen, J.: DVGAN: a minimax game for search
result diversification combining explicit and implicit features. In: SIGIR (2020)
22. Qin, X., Dou, Z., Wen, J.: Diversifying search results using self-attention network.
In: CIKM. ACM (2020)
23. Sanh, V., Debut, L., Chaumond, J., Wolf, T.: Distilbert, a distilled version of
BERT: smaller, faster, cheaper and lighter. CoRR abs/1910.01108 (2019)
24. Santos, R.L.T., Macdonald, C., Ounis, I.: Exploiting query reformulations for web
search result diversification. In: WWW (2010)
MatchingDIV 15
25. Shen, Y., He, X., Gao, J., Deng, L., Mesnil, G.: Learning semantic representations
using convolutional neural networks for web search. In: WWW. ACM (2014)
26. Silverstein, C., Henzinger, M.R., Marais, H., Moricz, M.: Analysis of a very large
web search engine query log. In: SIGIR Forum, vol. 33, no. 1 (1999)
27. Song, R., Luo, Z., Wen, J., Yu, Y., Hon, H.: Identifying ambiguous queries in web
search. In: WWW. ACM (2007)
28. Tao, C., Wu, W., Xu, C., Hu, W., Zhao, D., Yan, R.: Multi-representation fusion
network for multi-turn response selection in retrieval-based chatbots. In: WSDM
(2019)
29. Vaswani, A., et al.: Attention is all you need. In: NeurIPS (2017)
30. Xia, L., Xu, J., Lan, Y., Guo, J., Cheng, X.: Learning maximal marginal relevance
model via directly optimizing diversity evaluation measures. In: SIGIR (2015)
31. Xia, L., Xu, J., Lan, Y., Guo, J., Cheng, X.: Modeling document novelty with
neural tensor network for search result diversification. In: SIGIR (2016)
32. Xiong, C., Dai, Z., Callan, J., Liu, Z., Power, R.: End-to-end neural ad-hoc ranking
with kernel pooling. In: SIGIR. ACM (2017)
33. Yue, Y., Joachims, T.: Predicting diverse subsets using structural SVMs. In: ICML.
ACM International Conference Proceeding Series, vol. 307 (2008)
34. Zhai, C., Cohen, W.W., Lafferty, J.D.: Beyond independent relevance: methods
and evaluation metrics for subtopic retrieval. In: SIGIR (2003)
35. Zhou, X., et al.: Multi-turn response selection for chatbots with deep attention
matching network. In: ACL (2018)
36. Zhu, Y., Lan, Y., Guo, J., Cheng, X., Niu, S.: Learning for search result diversifi-
cation. In: SIGIR (2014)
Various Legal Factors Extraction Based
on Machine Reading Comprehension
Abstract. With the rapid growth of legal cases, professionals are under
pressure to go through lengthy documents and grasp informative pieces
of text in the limited time. Most of the existing techniques focus on
simple legal information retrieval task, such as name or address of the
prosecutor or the defendant, which can be easily accomplished with the
help of handcrafted patterns or sequence labeling methods. Yet compli-
cated texts always challenge such pattern-based methods and sequence
labeling approaches. These texts state the same facts or describe the
same events, but they do not share common or similar patterns. In
this paper, we design a unified framework to extract legal information
in various formats, including directly extracted information (a piece of
span) and information that needs to be deduced. The framework follows
the methodology to answer questions in machine reading comprehension
(MRC) tasks. We treat the extraction fact labels as the counterpart of
questions in MRC task and propose several strategies to represent them.
We construct several datasets regarding different cases for training and
testing. Our best strategy achieves up to 4% enhancement in F1 score
on each dataset compared to the MRC baseline.
1 Introduction
from the document. However, different forms of legal factors cannot be easily
extracted by a single model simultaneously using existing techniques. Thus, we
hope to build a unified framework to extract all factors regarding the case type.
Generally, information extraction task [6] extracts entities (person, organi-
zation, etc.) and facts (relations, events, etc.) from given texts [16], helping to
acquire the desired information and reconstruct massive contents. This is usually
done by sequence labeling models, such as Lattice LSTM [18] and transformer-
based models [9,17]. They conduct experiments on Chinese named entity recog-
nition (NER) task, and achieve high performance in benchmarks, such as Chinese
NER MSRA [8], OntoNotes 4.0, and Resume NER [18]. However, they fail to
maintain competitive results in our task. Our extraction targets are of multiple
forms, including spans and deductions. Spans are often entity names and event
descriptions, while deductions include answers to predefined legal conditions,
for example, whether the document agrees (answers “Yes”) or opposes (answers
“No”) to a given condition. These challenge the sequence labeling models since
the answers are not spans from the context. Inspired by MRC tasks [7,12,13],
whose question types are similar to ours, for example, asking for a named entity
(span) and deducing a piece of opinion (“Yes” or “No”), we can employ such
ideas to solve the legal factor extraction task. The difference lies in that the
questions of MRC tasks are arbitrary, but the extraction fact labels (the coun-
terpart to questions in MRC) are fixed along with the case type. We pre-define
these fixed legal factors to a given case as a set of extraction fact labels.
In this paper, we design a unified framework, particularly for multi-formed
legal factor extraction. It follows the MRC methodology to encode and interact
with the document and the extraction fact labels. Some researchers proved that
more specified and informative queries could improve the extraction precision
[10]. As our extraction fact labels are fixed, to achieve better performance, we
expand the labels to obtain more specified expressions with query expansion
components before the extraction. Our framework solves the problem discussed
above and our best strategy achieves 4% improvement compared to the baseline.
Our contributions are as follows: i. We achieve automatic extraction of various
forms of legal factor with a single model. ii. We proposed a strategy to represent
legal factors that outperform the others.
2 Related Works
Some existing works investigates the feasibility of using MRC approaches to
solving information extraction tasks. Li et al. [10] use the MRC model to
solve the named entity recognition task and achieved good results in nested
entity recognition. For each more nested entity, one more question requires to
be answered. And the reading comprehension model is designed to handle the
question-answering task. They suppose that the reading comprehension model is
a natural solution to the nested entity problem. Similarly, in our task, informa-
tion of factors is often nested or overlapped, we could exploit the MRC framework
for legal factor extraction.
Query expansion is comprehensively used because of its simplicity and prac-
ticality. There are several ways to reach query expansion. The IBM algorithm in
18 B. Wang et al.
the machine translation model is migrated to directly rewrite queries [5]. People
find that it is too rough to directly rewrite queries, and instead use the Seq2Seq
model to incorporate richer semantic information, and reinforcement learning is
used to fine-tune the rewrite model [1]. Of course, there are other ideas, such
as using a large amount of query click data, various concept words are mined
out. Further, associating the concept words with the knowledge graph in order
to replace the simple query clustering scheme [11].
Primarily, we introduce two important concepts, legal factors and the legal fac-
tor extraction task. As mentioned in Sect. 1, legal factor is a jurisprudence
concept, referring to the key information that affects the final sentencing, which
includes the descriptions of a judicial event, such as persons, actions, causes,
consequences and etc., and the conclusions to predefined judicial conditions. We
define the legal factor extraction task as the procedure to find out the text stat-
ing such information from original paragraphs. Our framework consists of three
parts: input documents and extraction fact labels embeddings, documents and
labels interactive encoding, and result answers prediction. The overall structure
is shown in Fig. 1 and Fig. 2 is an example shows how our framework actually
works.
Fig. 1. Architecture of the framework. All the four forms of extraction fact labels can
be fed into the encoder independently for extraction. The dotted arrows refer to the
semantic enhancement.
Various Legal Factors Extraction Based on Machine Reading Comprehension 19
In the embedding part, the inputs are the case descriptions and extrac-
tion targets. Note that the case type is an inherent attribute of a case
description and the targets of extraction are determined along with the case
type. We design different representations of the extraction fact labels and
gradually enriched their semantic. First, we employ special tokens ([Token1],
[Token2], etc.) as the targets. These tokens function as signals prompting the
model to extracting different legal factors. Then, we represent the labels by
text of legal factors, such as “ ” (the injury level) and “ ”
(course/means). Further, the legal factors are expanded into more specific ques-
tions, for instance, “ ” (What is the victim’s injury level?) and
“ ” (If the defendant use tools?). Finally, the questions are
reinforced to form longer questions via query expansion methods. We will discuss
different extraction targets in the following subsection. The rest parts interac-
tively encode documents and extraction fact labels with multi-head self attention
[15], and predict the results according to the types of extraction fact labels. If it
is unanswerable, an empty string returned. Otherwise, it answers “Yes” or “No”
to a yes/no question, or returns the result of the extraction through the start
and end indexes in the input passage.
Fig. 2. Examples of the workflow (Left: Answer as span. Right: Answer as “YES”).
We do not translate sentences after data augmentation (DA) because DA is usually a
paraphrase of the original question, and the English version could remain unchanged.
The Special Tokens. To verify our hypothesis that the different forms of legal
factor can be extracted by a unified framework, we represent each extraction
fact label as different [Token] rather than using the actual text of factor. They
20 B. Wang et al.
act as signals guiding the model to extract designated content. Specific, we set a
series of special tokens according to the legal factors that need to be extracted,
where the factors are one-to-one mapped to the tokens. It is equivalent to cipher
or semaphore. When the model encounters a special token, it learns to extract
a specific kind of information. Commencing from the special tokens, we enrich
the semantic gradually in the rest extraction fact labels representations.
From Factor to Question. Our produced datasets are dedicated to the type
of cases, and the information that needs to be extracted from each dataset is
summarized according to the experiences of the judiciary (legal factors). Tak-
ing the dataset of traffic accidents as an example, the factors are “ ”
(time of the accident), “ ” (site of the accident), “ ” (division of
responsibilities), “ ” (the loss), etc. These factors are the counterpart of
questions in typical question-answering datasets such as SQuAD [13], SQuAD2.0
[12], and CJRC [4]. Although the factors are not in the form of interrogative sen-
tences, they contain all the core information needed for extraction. Hence, we
use it the factors as another form of extraction fact labels.
Nevertheless, the factors are not written in general question format, they
are just short key phrases indicating what content to look into. We move for-
ward to gain question-like text to see if expanding a factor into question format
benefits the extraction. We convert the factors into question format according
to linguistic rules such as the wh-question patterns. For instance, “ ”
(time of the accident) is a factor in the dataset of traffic accidents. We apply
linguistic rules to switch it to its corresponding question, which results in
“ ” (What was the time of the accident?). Similarly,
“ ” (site of the accident) is switched to “ ”
(Where did the accident take place?), and “ ” (the loss) turns to
“ ” (What losses does the plaintiff suffer?). Some fac-
tors are expanded into more than one question because they refer to multiple
concrete facets. For example, “ ” (division of responsibilities) becomes
“ ” (What are the responsibilities of the defendant?) and
“ ” (What are the responsibilities of the plaintiff?). Such
expansion literally enriches the semantic compared to the vanilla factors, and is
highly likely to enhance the performance in the question-answering experiment.
word order. We design two strategies to find the synonyms of each word,
“M1 ” and “M2 ”. “M1 ” looks up the existing dictionary for synonyms, such
as BigCilin [19]. While “M2 ” builds a special dictionary concerning both con-
texts and questions in each dataset. The embeddings of words in the whole
created dictionary are generated by BERT [3]. We use the embedding vectors
for calculating cosine similarities of the input word and the rest words from
the dictionary. We rank the candidates and resume top K words as retrieved
synonyms. Then, the synonyms are directly piled up to form a new sentence.
For example, in the dataset of recourse for labor remuneration when sup-
posing K = 3, the question “ ” (Whether the
party signs the labor contract?) will be expanded as “
”.
– The second is sentence-wise synonym expansion and we name it data aug-
mentation (DA). The sentence is treated as a whole when searching for syn-
onyms. To augmentate a whole sentence, we need to gain its the general
semantic representation. Hence, we rely on the corpus of Baidu zhidao, which
is a knowledge encyclopedia written in Chinese. This corpus is used to pre-
train language model that better understand the correlation between Chinese
words and phrases. We refer readers to SimBERT [14] for training details. We
use this model to gain a ranked list of K similar sentences. After acquiring
the sentence-wise synonyms, we propose several strategies to join them. In
addition to direct concatenation, we utilize bi-LSTM and bi-GRU to encode
semantic of these synonyms. Compared to the direct concatenation, this strat-
egy extract hidden features and control dimension of the output in order to
satisfy BERT’s input length limit. The direct concatenation easily exceeds
this limitation and the rest part of text is ignored. We report selected aug-
mentation results of K = 5 in Table 1 and Table 5 (in the appendix).
4 Experiments
4.1 Datasets and Experimental Settings
We composed three datasets regarding the following case types, intentional injury
(II), recourse for labor remuneration (RLR), and refusal to execute judgments
or rulings (REJR). The statistics are shown in Table 2. We select 200 question-
answer pairs from each dataset as the testset. Note that the number of questions
answered by “No” is markedly low. This is a common feature of judicial data,
that officers tend to ask what they’ve already had evidence by hands, leading to
few negative answers. We do not manually disturb this, because we believe that
it is better to restore the actual dataset distribution.
As for experimental settings, we employ BERT-base encode the documents
and extraction fact labels and refer readers to BERT [3] for detailed model
descriptions and hyper-parameter settings. Since the input length of Bert is
limited by 512, the length (LF ) allocated to the extraction fact label should also
be constrained and well-designed. Overall, the input extraction fact label tokens
of our model is no longer than 60, and the rest are preserved for document input.
Specific, LF = 1 if using [Token] as the representation; LF ≤ 10 for factors and
questions; while LF ≤ 60 for expansions. In practice, most of the expanded
questions are longer than 10, so any K > 5 results in input length suppressing.
We will seek the best effect by varying K.
– “Qori ”: To Directly generate questions from the factors. The result of the
question-answering experiment in “Qori ” will be the baseline of the whole
experiments.
– “modif y”: We manually modify some questions, mainly to fix some mis-
takes caused by the generation process and to make them closer to
the contexts. The modification includes correction of typos and examina-
tion of punctuation characters. In the II dataset, we modify the ques-
tion “ ” (What happened to the victim?) to
“ ”. Although the English translation for both
“ ” and “ ” is victim, only “ ” appears in the contexts of
the dataset rather than “ ”. Therefore, we modify this word to be con-
sistent with the context.
– “+F ”: To directly concatenate the factors with original questions as a whole.
– “QEsyn -M 1”: Word-wise expansion using method M 1 described in Sect. 3.2.
– “QEsyn -M2 ”: Word-wise expansion using method M 2 described in Sect. 3.2.
– “QEda -random”: To replace the original question with a randomly question
obtained by data augmentation.
– “QEda -lstm”: To add a bi-LSTM layer after data augmentation to extract
semantic information from augmented questions.
– “QEda -gru”: To add a bi-GRU layer after data augmentation to extract
semantic information from augmented questions.
– “QEda -topK”: To directly concatenate the first K augmented sentences with
the original question. We experiment with K = 1, 3, 5.
– “QEda -mix”: To only expand questions whose answers are spans, but use the
original question for those answered by “Yes” or “No”, and those with no
answers (NG).
– “QEda -yn”: Do not expand question those answered by “Yes” or “No”, but
perform expansion on other kinds of questions.
– “QEda -last5”: To concatenate the last five out of ten augmented sentences
with the original question. We hypothesize that taking the last five sentences
increases the diversity of semantic because they are less similar to the original
question than the first five sentences.
Among, the above strategies, QEda -topK, QEda -mix, QEda -yn and QEda -
last5 are four ablation studies that reveal what indeed leads to the enhancement
of extraction results.
4.4 Results
The results of different methods to extract legal information and the perfor-
mances of different forms of extraction fact labels are reported in Table 3. Not as
24 B. Wang et al.
expected, the sequence labeling methods much underperform the rest methods.
The F1 score of Lattice only reaches 11.17%. Similarly, Seq2SeqBERT does not
perform much better, with F1 score 22.8%. We can conclude that the sequence
labeling methods are incapable of handling our task. Among the forms of extrac-
tion fact labels, the one with query expansion module achieves best F1 results
for RLR (88.51), II (86.11) and REJE (81.47) datasets. This method also con-
tributes to the highest EM scores on RLR (77.07) and REJR (67.34) datasets.
While in the II dataset, representing fact labels as questions peaks by the EM
score of 74.88. Table 3 shows that along with the increasing of semantic infor-
mation from tokens to expanded questions, the results are broadly improved,
reaching up to 20% in F1 and 15% in EM. These prove the idea that richer
semantic features boosts the model’s performance.
Table 4 and Table 6 (in the appendix) report the results of query expansion
strategies and the ablation experiments. Taking RLR dataset as an example, we
apply all the methods and strategies described in Sect. 4.3 to test the effective-
ness. Among these strategies, the data augmentation-based strategies outper-
form the others. The synonym-based experiments show similar pattern as the
RLR for the other two datasets. So we omit the synonym-based results due to the
page limitation and focus on the data augmentation-based experiments. In RLR
dataset, “QEda -top5” has a positive effect and its EM and F1 enhance about
1.5% and 2%, respectively compared to “QEda -lstm” and “QEda -gru”. After
these experiments, we realize that query expansion methods work unequally for
all kinds of questions. Some work well on questions whose answers are spans,
but poorly on “Yes”, “No”, or “NG” questions. Hence, apart from expanding
question of all answer types (QEda -topK), we design another experiments, QEda -
mix, QEda -yn, and QEda -last5, as ablation experiments to test the influence on
different types of extractions and to find out the best strategy. The QEda -mix
achieves an overall good result and is the recorded strategy in Table 3, but it
affects datasets differently.
Table 3. Exact Match(%) and F1(%) of different forms of extraction fact labels
(token, factor, question and expanded question) and sequence labeling method. Specific,
F1Seq1 : the F1 of Lattice [18]; F1Seq2 : the F1 of Seq2SeqBERT ; Expansion: QEda -mix.
5 Analyses
5.1 Extract Legal Factor as a Human Judge
Table 3 examines the feasibility of sequence labeling approaches in solving the
proposed legal factor extraction task and confirms that they are unsuitable to our
task. The “entities” in our datasets are nearly ten times as long as the entities
Various Legal Factors Extraction Based on Machine Reading Comprehension 25
in the standard NER datasets, such as Resume NER [18]. Our “entities” are
essentially the answer spans to pre-defined questions, and are uneven in length.
Whilst in the other datasets, most of the entities are of just a few tokens, and
the number of entity tags are far less than ours. Therefore, the sequence labeling
method is unqualified for information extraction task of this magnitude. These
urges us to introduce the MRC-based means as better solutions.
In general MRC task, the model knows what answers to searching for because
the questions are written in plain text. But in the “Token” experiment, we
only feed the token symbols to the model without the specific target of the
extraction descriptions. By doing so, we imitate the procedure of a judge reading
through the documents. As we discussed in Sect. 1, these legal factors are fixed
regarding the case type, and a judge can fast decide the what to looking for when
26 B. Wang et al.
given a case description. This experiment proves that the model can learn the
semantic meanings through this special token without been told the exact target.
To a certain extent, it completes the work of expert knowledge and artificial
settings by itself. This differs our methods from general MRC approaches, and
is instructive for follow-up experiments.
Table 4 and Appendix Table 6). In the ablation experiments, some questions are
preserved and the others are expanded. We first expand all the questions to find
out what types of questions end in poor results, and leave these questions not
expanded while expand the rest questions who have outstanding performance.
The final strategy is QEda -mix, whose experimental results on all three datasets
get improved by up to 4 points in both F1 and EM.
For different datasets, the same type of questions does not necessarily show
similar results before and after query expansion. For example, in the RLR and
REJR datasets, query expansion works very well on questions whose answers are
spans, and poorly on questions whose answers are “YES”, “NO” and “NG”. But
in the II dataset, the results are exactly the opposite. Practically, it is difficult
for us to fully judge the results of the experiments through pre-speculation. For
instance, we intuitively thought that it is unnecessary to expand yes/no questions
because their answers are simple and naive, and the question format is uniform.
However, as far as the experimental results are concerned, our speculation can
only be incompletely wrong. The performance of different types of questions
before and after query expansion could be inconsistent, and cannot be predicted
despite of the datasets they belong to.
6 Conclusion
In this paper, we design a unified framework to extract different types of legal fac-
tors. It is a MRC-based framework but with pre-determined extraction fact labels
according to the datasets and achieves automatic extraction without manually
feeding the questions. To verify the effectiveness of our approach, we constructed
three datasets, including intentional injury (II), recourse for labor remuneration
(RLR), and refusal to execute judgments or rulings (REJR) cases. Experiments
show that such model suits well to our information extraction task compared
to the sequence labeling models. To improve the model performance, we design
different strategies and conduct plentiful experiments that discuss the impact
of adding semantic information to the extraction fact labels. The strategies fall
into two main categories, synonym-based and data augmentation-based. Among
all the strategies of these two categories, the QEda -mix performs generally well,
improving the performance of up to 4% on each dataset.
7 Future Works
Our proposed framework aim at extracting different types of legal factors. So far,
it is able to extract span-like factors, deduce yes/no factors and judge whether a
factor is included in the document or not. However, there are still several types of
legal factors that are not supported, such as multiple-choice factors, numerical
derivation factors, summarisation factors, etc. multiple-choice factors refer to
the fact that involves a fixed number of parties. For instance, some disputes over
rental contracts often involve three parties: the landlord, the tenant, and the
intermediary. The numerical derivation factors usually ask for the total amount
of money, weight, and the number of criminal activities. These information are
28 B. Wang et al.
not directly and clearly written, nor can they be derived during straightforward
extractions. In the future, we will manage to expand the framework’s capability
to let it handle more types of factors.
Experiments show that although query expansion has a positive effect on
each dataset, yet different datasets have different responses to query expansion.
For the RLR and REJR datasets, query expansion works well on questions whose
answers are spans. While query expansion gives a good result on the other types
of instance for II dataset. We believe this is due to differences in the structures
and content of different case types, but we have not quantified these differences
or looked for common patterns. In the future, we will try to find a universal
query expansion strategy to streamline the current solution, as well as adding
new factor types into consideration.
Acknowledgments. The authors would like to thank all the reviewers for their
insight reviews. This paper is funded by National Key R&D Program of China
(No.2018YFC0807701).
Appendix
References
1. Buck, C., et al.: Ask the right questions: active question reformulation with rein-
forcement learning. In: 6th International Conference on Learning Representations,
ICLR 2018, Vancouver, BC, Canada, 30 April–3 May 2018, Conference Track Pro-
ceedings. OpenReview.net (2018). https://openreview.net/forum?id=S1CChZ-CZ
Another random document with
no related content on Scribd:
talk, but his uncle, Halil Pasha, the former Commander of the
Mesopotamian and Caucasian fronts, caused much merriment by
fighting a sham duel with wooden swords. His opponent was a
singer dressed as a medieval knight.
Besides the actresses several of the old aristocracy were gracious to
Enver and even the pretty wife of a Commissar attached to the
Foreign Office wrote him what would have been considered an
intriguing note, in another time and place. She offered to teach him
Russian if he would teach her French and he replied curtly that he
was not a “professor.”
But Enver was not always over-serious and unbending. In a certain
small circle of friends he was quite otherwise. And circumstances,
which so largely decide our destinies, gave me an opportunity to
know him as well as one alien to his religion could know him. He was
allotted by the Foreign Office quarters in the little palace where I was
living. For half a year I saw him every day, sat next to him at table
and occasionally we went to the theatres and the Turkish Embassy.
During that time Enver told me a great deal about his life and his
ambitions.
He had known my husband in Constantinople, and was away from
Moscow at the time of my husband’s death. As soon as he returned
he called on me, bringing Halil and the Turkish Ambassador, Ali
Fued Pasha, with him. All three were extremely kind and
sympathetic. From that moment until I left Moscow the Turks did
everything possible to make life less tragic for me, and I gained an
insight into the Turkish character which I had never imagined. The
Turks have a peculiar capacity for friendship. And friendship, once
given, has no bournes; a friend is a friend through everything—
sorrow, dishonor, poverty, as well as wealth and success. An enemy,
on the other hand, is beyond all consideration; he is spared nothing,
forgiven nothing.
Enver has the personal vanity of the enthusiast and he imagines that
everything he does he does well. The only way to cope with his
conceit is to be brutally frank. I discovered early in our acquaintance
that frankness by no means offends him. One of Enver’s nicest
qualities is that he likes to discover his faults as well as his virtues;
he is eager to improve himself.
He has a passion for making pencil sketches of people he meets and
always goes about with a pencil and a pad of paper in his pocket. In
the house where we lived he made sketches of all the guests and all
the servants. He made, in all, six very uninteresting portraits of me.
One morning when the tea was more tasteless than ever and the
bread especially sour and muddy so that I felt I could not manage to
eat a single bite, I could not help feeling unpleasantly resentful to
see Enver busy with his sketches and full of enthusiasm. And while I
sat staring at the terrible meal, he proceeded to make a life-size
portrait of me which was incredibly bad. I remember that I wondered
where on earth he ever got such a large piece of paper in Russia.
When he had finished the sketch he signed his name with a flourish
and presented it to me. I took it but said nothing. Enver has the
curiosity of a child, and, after a long silence, he asked me if it was
possible that I did not really like it. I said that I thought he had no
talent for drawing. He became suddenly quite angry and said in a
low voice, “But do you realize I have signed my name to it?”
“Your name doesn’t mean anything on a picture,” I explained. “If it
was an order for an execution or an advance it would be another
matter. You can’t make a good drawing just by signing your name to
it”
He frowned and then grew cordial as suddenly as he had grown
angry. He rose and bowed to me in a most courtly way. “You can’t
imagine,” he said, “how pleasing arrogance is to me.” His three
dominating characteristics being bravery, hauteur and recklessness,
he imagined that these motives also guided the actions of his
friends.
Enver never seemed to be able to loaf in the easy manner of most
Orientals. His mental and physical vitality is more like that of an
enthusiastic and healthy American. Every morning he rose early to
go for a long walk, he read a great deal, took at least three lessons
in some foreign language every week and was constantly writing
articles for Turkish papers which he printed on a hand press in his
own room, and he held almost daily conferences either with the
Russians or the Mohammedans. He does not drink or smoke and is
devoutly religious.
He likes any discussion which reveals another person’s deepest
emotions. If he cannot rouse one any other way he does so by some
antagonistic remark which often he does not mean at all. For
example, he is extremely liberal in his opinions about women and
does not think they should be excluded from political life.
Nevertheless, he said to a young actress at tea one afternoon in my
apartment, when they were talking about woman suffrage, that she
would be better off in a harem. Being an ardent feminist, she rose
and fairly shouted at him: “Enver Pasha, you may be a great man in
the East, but just listen to me! I am one of the first actresses of my
profession. In my world it is every bit as great and important for me
to remain an actress as it is for you in your world to remain a warrior
or a diplomat.”
Enver took his scolding in very good humor. Afterwards he told me
that he had never liked this actress before. “Independence is a great
thing in women. Our women lack it and many of them are just
puppets on this account.”
He was always extremely interested in American ideas and
American opinions. He said he could never understand why
Americans were so sentimental about Armenians. “Do they imagine
that Armenians never kill Turks? That is indeed irony.”
At the table he used to ask Mr. Vanderlip questions about his
proposed Kamchatka concessions. Vanderlip, like many
Californians, is rather violently anti-Japanese. His idea of having a
naval base at Kamchatka amused Enver. He said Vanderlip was
killing two birds with one stone, that he wished to manœuver the
American Government into a war with Japan, prove himself a patriot,
and at the same time protect his own interests and grow rich. “So
that,” said Enver, “if it really came about—the next war would be for
Vanderlip and should be known as ‘Vanderlip’s War.’”
When I asked, “Would you be sorry to see America and Japan at
war?” he replied, “Not if England was involved. Anything which tends
to draw England’s attention away from us or which weakens the
great powers, naturally gives Turkey a better chance for
reconstruction. You understand that I’m not saying I want to see
another war; I am simply saying that if those nations interested in
destroying Turkey are occupied elsewhere it relieves us of war
burdens and gives us a chance to carry out our own destinies.”
He tried to get Mr. Vanderlip’s reaction on women by the same
tactics he employed with the actress. One day he said, “I have three
wives and I’m looking for another.” This was not true, but Mr.
Vanderlip proved entirely gullible. “Good heavens,” he said,
regarding Enver in shocked surprise, “we Anglo-Saxons consider
one wife enough tyranny....”
“Naturally,” Enver conceded, suavely, “with one there must be
tyranny but with three or four or a hundred.... Ah, you must agree
that is quite a different matter.”
His sudden appearance in Moscow during the blackest days of the
blockade as well as the blackest days for the Central Powers proves
him an incomparable soldier-of-fortune. With two suits of clothes, a
pair of boots, a good revolver and a young German mechanician
whom he could trust, he started by aeroplane from Berlin to Moscow.
The story of how they had to land because of engine trouble near
Riga, of how he was captured and spent two months in the Riga jail
just at the moment when the whole Allied world was calling loudest
for his blood, will remain a story which will have scanty
advertisement from those British Secret Service men who like so
well to turn journalists and write their own brave autobiographies.
Enver sat in the Riga jail as plain “Mr. Altman” who could not speak
anything but German. He was scrutinized by every Secret Service
man in the vicinity and pronounced unanimously a Jewish German
Communist of no importance. By appearing humble, inoffensive and
pleasant, he soon worked his way into the confidence of the warden,
was released and escaped to Moscow. He arrived just in time to rush
off for the dramatic Baku conference.
The Communists understood perfectly well that Enver Pasha was
not at the Oriental Conference as a sudden and sincere convert to
Internationalism, and he knew that they knew. Both Zinoviev and
Enver were actors taking the leading rôles in a significant historical
pageant. The results are really all that matter, since the motives will
soon be forgotten.
When Enver turned to Moscow he had no other place to turn to and
when Zinoviev took him to Baku, Zinoviev knew no other means of
effectively threatening the English in order to change their attitude on
the blockade. Zinoviev could not complain about Enver’s shallow
attitude towards Socialism since there was hardly anything
Socialistic about Zinoviev’s appeal for a “holy war.” Enver summed
up his feelings about the new alliance thus: “For the future of Turkey
and the future of the East a friendship with Russia is worth more to
the Turks than any number of military victories. And we have to build
that friendship while we have the opportunity.”
His way of living without any regrets and as if there were no to-
morrows is rather startling at times. I remember when Talaat Pasha,
his lifelong friend, was murdered by an Armenian in Berlin, he read
the message with no show of emotion and his only comment was:
“His time had come!” But against an excessive temptation on the
part of fate to record Enver’s death prematurely, in his own words, he
“sleeps with one eye open,” carries a dagger and a loaded
automatic. Once when we talked about the possibility of his being
assassinated he said, “I have been near death so many times that
these days I live now seem to be a sort of present to me.”
Enver is no open sesame to those who do not know him well. He
really has the traditional Oriental inscrutability. The first two or three
times I talked with him, we stumbled along rather lamely in French.
Someone suggested to me that he probably spoke several
languages which, for some unknown reason, he would not admit. So
one day I said abruptly, “Oh, let’s speak English.” He looked at me
with one of his sudden, rare smiles and answered in my own
language, “Very well, if you prefer it.”
When I asked him how he learned English he told me he had
learned it from an English spy. “He came to me as a valet and
professed deep love for Turkey. For several months we studied
diligently. One day I thought I would test his love for Turkey so I
ordered him to the front. He was killed. Later, we found his papers.”
“Were you surprised?” I asked him.
“Why, not at all,” said Enver. “He really showed a great deal of pluck.
The only thing I had against him was that he taught me a lot of
expressions not used at court.”
“Like what?”
“Like ‘don’t mention it,’” said Enver, laughing. “And the terrible thing
about learning such an expression,” he said, “is that it is so sharp
and so definite and often fits an occasion so aptly that it flashes in
one’s mind and can’t be forgotten. American slang is extremely
picturesque and expressive, but it is not dignified enough to be used
by diplomats.”
Everyone is familiar with Enver’s “direct action” method of playing
politics. One of the ways he was wont to remove troublesome rivals
in the days of the Young Turk Revolution was to go out and shoot
them with his own hand. This “impulsiveness” got him into grave
trouble with the Soviets in spite of all his sensible utterances to the
contrary. When he was “shifted” to Bokhara so that he would not be
in the way of either Kemal or the Russians, he got bored and started
a war of his own. One night he fled into the hills of Afghanistan and
soon began to gather recruits around him. A few nights after that one
of the principal officials of the Bokharan Republic also fled to join
Enver. This performance was repeated until over half the Bokharan
cabinet had fled. Then fighting began and we got vague rumors of
battles, but only through the Soviet press.
In August, 1922, we heard that Enver had been killed, that his body
had been found on the field of battle. There was even romance
surrounding him and his supposed death. Stories were circulated
that when the body was picked up and examined, the letters of an
American girl were found next his heart. I went to see Jacob Peters,
who has charge now of all the Eastern Territory and who was then in
Moscow. He laughed heartily and said he would show me the
“information.” It consisted of three very hazy telegrams which had
been three weeks on the way. The men who sent the telegrams and
discovered the body had never seen Enver. There was no mention of
letters. And Peter’s opinion of the whole affair was that there was
nothing at all authentic in the story or else it was “a trick of Enver’s to
sham being dead.” Peters’ theory proved true. Within a few days
fighting began again and Enver began to win.
He had conceived the notion of uniting all Turkestan and Bokhara
and Keiva to the Angora government. It places the Soviets in a
strange position. They may have to give in to him, though he will not
actually be an “enemy,” because neither the Turks nor the Russians
can afford to break their treaty. Therefore, his private war in the
south embarrasses the Soviets much more than it does Kemal, who
needs only to disavow any connection with it, as do all the Turkish
officials. If Enver wins he will add a nice slice to Turkish territory; if
he loses, Turkey will be in the same position as before.
Enver, while he will always maintain a great prestige in the
Mussulman world, will never oust Kemal. Mustapha Kemal Pasha is
the great popular hero of a victorious Turkey, which but for him might
never have even survived. There were times in the past when Enver
was more important than Kemal, but that can never happen again in
Kemal’s life. Both men rose from the ranks and both are the sons of
peasants. And Kemal at this moment is more important than the
Sultan. Greater than that no man can be under the banners of
Mohammed.
One can hardly over-estimate the importance of the new
Mohammedan unity, that new patriotic energy which has taken the
place of the former lethargy and which already reaches out far
beyond the borderland of the Faithful. The Mussulman world is
reviving after a long sleep. And not only Mohammedans are uniting
but the entire East and Middle East. Aside from Japan, a significant
harmony is rapidly taking place, a harmony which evolves itself into
a tremendous power. This power may decide the world’s destiny
before another generation.
Enver and Kemal Pasha, being aware of the purport of beginning
that great concord by interwoven treaties with Russia, read the stars
well. There must come a day also when that great sleeping giant,
China, will be part of this alliance. And the seeds of that friendship
have also been planted. The Chinese official delegations which
came to Moscow were not only well received by the Russians, but
they hob-nobbed with the Mohammedans like brothers.
TIKON AND THE RUSSIAN CHURCH
GREGORY VASSILIEVITCH
TCHICHERIN
My first interview with Tchicherin was at midnight and my last
interview was at five in the morning. This happens to cover a fairly
complete rotation of the official hours of the Soviet Foreign Office.
One evening at a box party in the Bolshoi Theatre, Enver Pasha
remarked: “I have to kill time somehow for three hours after the play.
Halil Pasha and I have an appointment with Mr. Tchicherin at two
o’clock.” In spite of his smiling Oriental inscrutability and a palpable
diplomatic duty to conform to everything Russian, one could feel an
amused disapproval of such official unconventionality.
This eccentric habit of turning night into day, with every floor of the
Foreign Office blazing like a lighthouse in a city which by municipal
decree is put to bed before midnight in order to save fuel, naturally
creates an almost fantastic air of whimsicality. Mr. Tchicherin makes
no excuse for this “vice,” as one of his secretaries very cleverly
phrased it; he simply finds night more harmonious for his tasks than
day and with that lack of consideration which dreamers always
consciously or unconsciously assume, he forces his whole staff to
follow his example. The result is that his clerks make a mad
scramble to get transferred into another government department.
Everything about Tchicherin is as consistently contrary to an ordered
life as his inversion of working hours. Born an aristocrat, trained
under the Tsar for the diplomatic service, delicate, cultured, aloof,
with a fine gesture of Quixotic generosity, he has thrown his life and
his fortunes in with the cause of the proletariat with all the abandon
of religious fervor.
His aloofness is so evident that one can hardly find any concordance
about the astounding decision of such an obvious æsthete to
become an active part of revolution—which is sweat and blood and
violence. Perhaps that explains why he wraps his vision round him
like a cloak and shuts out the sun in order not to be disturbed and
disillusioned by reality. We were all brought up on stories about kings
who were gay-fellows-well-met and could outdance and outdrink
their soldiers; on nobles who turned out to be Robin Hoods. But,
alas, who can imagine Tchicherin rollicking at a workers’ picnic or
smoking a friendly pipe with a Red soldier?
No simple person will ever feel intimate or at home with his super-
class indifference to material surroundings. A scrubwoman is just as
uncomfortable in his presence as was the intrepid Mrs. Sheridan,
who was able to rub such gay elbows with the other commissars. Mr.
Tchicherin’s way of arching an eyebrow at life upsets the best brand
of poise.
Living alone in a barren room on the top floor of the Foreign Office,
he is as far removed socially and physically from the lower as the
upper crust. Perhaps only an aristocrat is able to attain this dizzy
height of indifference to human contact with one’s fellows. And I
can’t help feeling that there is something rather splendid about such
complete isolation.
Outside of politics, the telephone and the cable, all up-to-dateness
offends him. He abhors new clothes, does not like to ride in
automobiles, refuses to have modern office paraphernalia about him,
does every little task for himself, like sharpening his own pencils and
running all over the building on office-boy errands. This attitude
produces the same effect as if he distrusted all his subordinates. His
secretaries stand helpless and ill at ease while he searches for a lost
telegram or answers the telephone.
Last winter they told an amusing story of how Karakhan, who is
Commissar of Eastern Affairs, lured Tchicherin into donning a new
suit. Tchicherin’s one suit was literally in rags when the Turkish treaty
and the Afghan treaty and the Persian treaty and all the other
Oriental treaties were about to be signed. These affairs had to be
arranged with more or less bourgeois pomp, since the Orientals are
rather keen on ceremony. So Mr. Karakhan, taking a long chance,
went ahead and ordered a new suit for Mr. Tchicherin from a Soviet
tailor, then one morning while Tchicherin slept, he changed the suits.
In a few minutes he came rushing back again and exclaimed with
emotion, “There’s a new note from Lord Curzon!” Tchicherin was up
in one bound and struggling into the new trousers. Whatever he
thought privately of Mr. Karakhan’s presumption, they continued in
an apparently pleasant relationship.
In appearance Mr. Tchicherin is tall, with the bent shoulders of the
man who stoops to go through doors. His eyes, not through any
evasiveness, but because of an extreme shyness, continually seek
other places than the face of his interviewer. Yet when one meets his
quick, occasional glance, one is startled by the intelligence and
gentleness of his expression.
Diplomacy is an inseparable part of Mr. Tchicherin’s existence. He
eats, drinks and sleeps with the affairs of state, looks at life as a
chess game and is continually checkmating, even in ordinary
conversation. Lenin approves of him and feels for him a warm
personal affection in spite of the fact that the Premier so dislikes
eccentricities. He knows that Tchicherin can be trusted, that he has
an invaluable knowledge of international affairs and more important
than all that, that he will never make any real decision without
consulting Lenin.
Mr. Bullitt told me that during his negotiations he found Tchicherin so
brilliant that it was difficult to get anywhere. The Foreign Minister was
always quite justified from the Soviet angle but the Soviets were
being forced to make hard concessions. Invariably when they came
to a deadlock, he telephoned Lenin and Lenin gave in.
During our first talk, when we discussed the campaign of lies about
Russia which has so long flooded English, French and American
papers, I said that I thought it was partly due to the fact that no
reporters were permitted at that time to go in and investigate actual
conditions. It was characteristic of Tchicherin to interrupt very
suddenly and ask, “Will you tell me why American reporters come
over here and claim they are impartial observers, even profess
friendliness towards us, and then go home and write such
astounding lies?”
I thought it wasn’t fair to generalize. The most unfair stories have
always been manufactured at Riga and Reval or at Paris by
interested political groups or by disappointed reporters who never
got inside. As for the reporters who actually witnessed the revolution,
certainly the majority remained fair and sympathetic, in spite of the
fact that it grew particularly difficult, especially in America, even to
maintain one’s equilibrium about Russia after Brest-Litovsk. To my
mind came back unhappy recollections of Overman and Lusk
investigations, raids, deportations and general war-hysteria. Perhaps
some such thought came also to Tchicherin because he said, “Yes,
yes, I suppose in the main, you are right, but how do you account for
a man like——?”
Tchicherin is full of old-fashioned honor. The idea that foreign papers
sanctioned false reports in order to justify intervention or the
blockade seemed so outrageous to him that he could never realize
that this sort of propaganda has become as much a part of modern
warfare as liquid fire or submarines.
Very late one night I saw Tchicherin running up the stairs to his office
in a high state of excitement because a New York evening
newspaper carried on its front page a fake interview with Lenin in
which he discussed everything from the Irish situation to the Russian
Ballet. Tchicherin saw no humor in this. His comment was, “How can
a reputable American paper allow such a thing? After all, Comrade
Lenin is the Premier of a great country.”
Men who give themselves completely to an ideal quite naturally
become supersensitive and unreasonable. At least that is the rule,
and Tchicherin is no exception. The deliberate misinterpretation
abroad, during long hard years, of every effort of the Soviet
Government at peace or reconstruction or defense or negotiations,
has got under his skin. So while he insisted on the strictest
adherence to the truth in all reports sent over the government wire,
at the same time he permitted himself a mild dissipation in
extravagant adjectives by way of retaliation, in his too long and too
complicated “notes.” He allowed even more unrestrained language in
Vestnik. Vestnik is the official bulletin of the Soviet Government—
very much like the bulletin issued by the Bureau of Public
Information during the war. The young man who edited this sheet
was a talented and educated Russian but his idea of an unemotional
government report was very much like that of our own George Creel.
I used to tease him about his passion for such words as “scurrilous”
in reference to capitalists or White Guards. But it never made any
impression. He confessed that he found my cables flat and
uninteresting.
Besides my radios to American papers, which were transmitted by
way of Berlin, and the government bulletin which was sent out to the
whole world and rarely used by anybody, there was also a wire to
London for the Daily Herald. Every one of these telegrams had to be