Abstract
Language models are at the heart of numerous works, notably in the text mining and information retrieval communities. These statistical models aim at extracting word distributions, from simple unigram models to recurrent approaches with latent variables that capture subtle dependencies in texts. However, those models are learned from word sequences only, and authors’ identities, as well as publication dates, are seldom considered. We propose a neural model, based on recurrent language modeling (e.g., LSTM), which aims at capturing language diffusion tendencies in author communities through time. By conditioning language models with author and dynamic temporal vector states, we are able to leverage the latent dependencies between the text contexts. The model captures language evolution of authors via a shared temporal prediction function in a latent space, which allows to handle a variety of modeling tasks, including completion and prediction of language models through time. Experiments show the performances of the approach, compared to several temporal and non-temporal language baselines on two real-world corpora.
Similar content being viewed by others
Notes
Code available at https://github.com/edouardelasalles/dar.
Please note that this dataset contains papers with multi-authors. In the experiments reported below, we consider each text as written individually by each of its authors (i.e., it is contained in the individual text set of each of its authors). Combinations of author representations to model multi-authorship could correspond to an interesting perspective as future work.
We did not use WordPiece in the NYT corpus since we noticed in experiments that it led to dramatic overfitting.
References
Ammar W, Groeneveld D, Bhagavatula C, Beltagy I, Crawford M, Downey D, Dunkelberger J, Elgohary A, Feldman S, Ha V, Kinney R, Kohlmeier S, Lo K, Murray T, Ooi H-H, Peters M, Power J, Skjonsberg S, Wang LL, Wilhelm C, Yuan Z, van Zuylen M, Etzioni O (2018) Construction of the literature graph in semantic scholar. In: NAACL
Bai S, Kolter JZ, Koltun V (2018) An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. In: CoRR
Bamler R, Mandt S (2017) Dynamic word embeddings. In: ICML
Bengio Y, Ducharme R, Vincent P, Jauvin C (2003) A neural probabilistic language model. JMLR 3:1137–1155
Bishop CM, Hinton GE, Strachan IG (1997) GTM through time. In: ICANN
Blei DM, Lafferty JD (2006) Dynamic topic models. In: ICML
Blei DM, Ng AY, Jordan MI (2002) Latent dirichlet allocation. JMLR 3:993–1022
Chiu C-C, Sainath TN, Wu Y, Prabhavalkar R, Nguyen P, Chen Z, Kannan A, Weiss RJ, Rao K, Gonina K et al (2018) State-of-the-art speech recognition with sequence-to-sequence models. In: ICASSP
Chung J, Kastner K, Dinh L, Goel K, Courville AC, Bengio Y (2015) A recurrent latent variable model for sequential data. In: Cortes C, Lawrence ND, Lee DD, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates Inc, Red Hook, pp 2980–2988
Devlin J, Chang M-W, Lee K, Toutanova K (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. In: ACL
Ding SH, Fung BC, Iqbal F, Cheung WK (2017) Learning stylometric representations for authorship analysis. IEEE Trans Cybern 49:107–121
Eger S, Mehler A (2017) On the linearity of semantic change: Investigating meaning variation via dynamic graph models. In: ACL
Fedus W, Goodfellow I, Dai AM (2018) Maskgan: better text generation via filling in the\_. In: ICLR
Fraccaro M, Sønderby SK, Paquet U, Winther O (2016) Sequential neural models with stochastic layers. In: NeurIPS
Gerrish SM, Blei DM (2010) A language-based approach to measuring scholarly impact. In: ICML
Hall D, Jurafsky D, Manning CD (2008) Studying the history of ideas using topic models. In: EMNLP
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Iwata T, Yamada T, Sakurai Y, Ueda N (2012) Sequential modeling of topic dynamics with multiple timescales. In: TKDD
Kabán A, Girolami MA (2002) A dynamic probabilistic model to visualise topic evolution in text streams. JIIS 18:107–125
Kenter T, Wevers M, Huijnen P, de Rijke M (2015) Ad hoc monitoring of vocabulary shifts over time. In: ICIKM
Kingma DP, Welling M (2014) Auto-encoding variational bayes. In: ICLR
Krishnan RG, Shalit U, Sontag D (2017) Structured inference networks for nonlinear state space models. AAAI, Menlo Park
Lample G, Ott M, Conneau A, Denoyer L, Ranzato M (2018) Phrase-based and neural unsupervised machine translation. In: EMNLP
Le QV, Mikolov T (2014) Distributed representations of sentences and documents. In: ICML
Loper E, Bird S (2002) Nltk: the natural language toolkit. In: ACL workshop on effective tools and methodologies for teaching natural language processing and computational linguistics
Melis G, Dyer C, Blunsom P (2018) On the state of the art of evaluation in neural language models. In: ICLR
Merity S, Keskar NS, Socher R (2018a) An analysis of neural language modeling at multiple scales. In: CoRR
Merity S, Keskar NS, Socher R (2018b) Regularizing and optimizing LSTM language models. In: ICLR
Mikolov T, Karafiát M, Burget L, Černockỳ J, Khudanpur S (2010) Recurrent neural network based language model. In: ISCA
Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J (2014) Distributed representations of words and phrases and their compositionality. In: NeurIPS
Peters M, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, Zettlemoyer L (2018) Deep contextualized word representations. In: NAACL
Radford A, Wu J, Child R, Luan D, Amodei D, Sutskever I (2019) Language models are unsupervised multitask learners. In: OpenAI Blog
Rezende DJ, Mohamed S, Wierstra D (2014) Stochastic backpropagation and approximate inference in deep generative models. In: ICML
Rudolph MR, Blei DM (2017) Dynamic bernoulli embeddings for language evolution. In: CoRR
Rudolph MR, Ruiz FJR, Athey S, Blei DM (2017) Structured embedding models for grouped data. In: NeurIPS
Saito K, Kimura M, Ohara K, Motoda H (2009) Learning continuous-time information diffusion model for social behavioral data analysis. ACML, Ahmedabad
Song F, Croft WB (1999) A general language model for information retrieval. In: ICIKM
Subramanian S, Lample G, Smith EM, Denoyer L, Ranzato M, Boureau Y-L (2018) Multiple-attribute text style transfer. In: ICLR
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. In: NeurIPS
Vinyals O, Toshev A, Bengio S, Erhan D (2017) Show and tell: lessons learned from the 2015 mscoco image captioning challenge. In: PAMI
Wang C, Blei D, Heckerman D (2008) Continuous time dynamic topic models. In: UAI
Wang E, Silva J, Willett R, Carin L (2011) Dynamic relational topic model for social network analysis with noisy links. In: SSP
Wang W, Gan Z, Xu H, Zhang R, Wang G, Shen D, Chen C, Carin L (2019) Topic-guided variational auto-encoder for text generation. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers). Association for Computational Linguistics, Minneapolis, pp 166–177
Wang X, Mccallum A (2006) Topics over time: A non-Markov continuous-time model of topical trends. In: SIGKDD
Yang Z, Hu Z, Dyer C, Xing EP, Berg-Kirkpatrick T (2018) Unsupervised text style transfer using language models as discriminators. In: NeurIPS
Yao Z, Sun Y, Ding W, Rao N, Xiong H (2017) Discovery of evolving semantics through dynamic word embedding learning. In: CoRR
Yao Z, Sun Y, Ding W, Rao N, Xiong H (2018) Dynamic word embeddings for evolving semantic discovery. In: WSDM
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Delasalles, E., Lamprier, S. & Denoyer, L. Deep dynamic neural networks for temporal language modeling in author communities. Knowl Inf Syst 63, 733–757 (2021). https://doi.org/10.1007/s10115-020-01539-z
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10115-020-01539-z