Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3539618.3591695acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
research-article

Unsupervised Readability Assessment via Learning from Weak Readability Signals

Published: 18 July 2023 Publication History

Abstract

Unsupervised readability assessment aims to evaluate the reading difficulty of text without any manually-labeled data for model training. This is a challenging task because the absence of labeled data makes it difficult for the model to understand what readability is. In this paper, we propose a novel framework to Learn a neural model from Weak Readability Signals (LWRS). Instead of relying on labeled data, LWRS utilizes a set of heuristic signals that specialize in describing text readability from different aspects to guide the model in outputting readability scores for ranking. Specifically, to effectively use multiple heuristic weak signals for model training, we build a multi-signal learning model that ranks the unlabeled texts from multiple readability-related aspects based on intra- and inter-signal learning. We also adopt the pairwise ranking paradigm to reduce the cascade coupling among partial-order pairs. Furthermore, we propose identifying the most representative signal based on the batch-level consensus distribution of all signals. This strategy helps identify the predicted signal that is most correlated with readability in the absence of ground-truth labels. We conduct experiments on three public readability assessment datasets. The experimental results demonstrate that our LWRS outperforms each heuristic signal and their combinations significantly, and can even perform comparably with some supervised methods. Additionally, our LWRS trained on one dataset can be effectively transferred to other datasets, including those in other languages, which indicates its good generalization and potential for wide application.

Supplemental Material

MP4 File
Presentation video for full paper Unsupervised Readability Assessment via Learning from Weak Readability Signals. This is my first time recording this kind of video, I am a bit nervous and not very good at speaking, so please feel free to contact me if you have any questions about the content of the paper

References

[1]
Hélder Antunes and Carla Teixeira Lopes. 2020. Proposal and Comparison of Health Specific Features for the Automatic Assessment of Readability. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval (Virtual Event, China) (SIGIR '20). Association for Computing Machinery, New York, NY, USA, 1973--1976. https: //doi.org/10.1145/3397271.3401187
[2]
Oghenemaro Anuyah, Ion Madrazo Azpiazu, David McNeill, and Maria Soledad Pera. 2017. Can readability enhance recommendations on community question answering sites? CEUR Workshop Proceedings 1905 (2017). 2017 Poster Track of the 11th ACM Conference on Recommender Systems, Poster-Recsys 2017; Conference date: 28-08-2017 Through 28-08-2017.
[3]
Ion Madrazo Azpiazu and Maria Soledad Pera. 2016. Is Readability a Valuable Signal for Hashtag Recommendations?. In Proceedings of the Poster Track of the 10th ACM Conference on Recommender Systems (RecSys 2016), Boston, USA, September 17, 2016 (CEUR Workshop Proceedings, Vol. 1688), Ido Guy and Amit Sharma (Eds.). CEUR-WS.org. http://ceur-ws.org/Vol-1688/paper-21.pdf
[4]
Richard Bamberger and Annette T. Rabin. 1984. New Approaches to Readability: Austrian Research. The Reading Teacher, vol. 37, no. 6 (1984), 512--519.
[5]
Kepa Bengoetxea and Itziar Gonzalez-Dios. 2021. MultiAzterTest: A Multilingual Analyzer on Multiple Levels of Language for Readability Assessment. arXiv:2109.04870 [cs]
[6]
Tianyuan Cai, Ho Hung Lim, John S. Y. Lee, and Meichun Liu. 2022. Enhancing Automatic Readability Assessment with Verb Frame Features. In International Conference on Asian Language Processing, IALP 2022, Singapore, October 27-28, 2022, Rong Tong, Yanfeng Lu, Minghui Dong, Wengao Gong, and Haizhou Li (Eds.). IEEE, 413--418. https://doi.org/10.1109/IALP57159.2022.9961289
[7]
Meri; Coleman and T. L Liau. 1975. A computer readability formula designed for machine scoring. Journal of Applied Psychology, Vol. 60, pp (1975), 283--284.
[8]
E. Dale and J.S Chall. 1948. A Formula for Predicting Readability. Educational Research Bulletin (1948), 37--54.
[9]
Tovly Deutsch, Masoud Jasbi, and Stuart M. Shieber. 2020. Linguistic Features for Readability Assessment. In Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications, BEA@ACL 2020, Online, July 10, 2020, Jill Burstein, Ekaterina Kochmar, Claudia Leacock, Nitin Madnani, Ildikó Pilán, Helen Yannakoudakis, and Torsten Zesch (Eds.). Association for Computational Linguistics, 1--17. https://doi.org/10.18653/v1/2020.bea-1.1
[10]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805 [cs]
[11]
Yo Ehara. 2021. LURAT: A Lightweight Unsupervised Automatic Readability Assessment Toolkit for Second Language Learners. In 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI). 806--814. https://doi.org/ 10.1109/ICTAI52525.2021.00129
[12]
Yo Ehara. 2022. Uncertainty-aware Personalized Readability Assessment Framework for Second Language Learners. J. Inf. Process. 30 (2022), 352--360. https: //doi.org/10.2197/ipsjjip.30.352
[13]
Mohamed El-Madkouri and Beatriz Soto Aranda. 2022. Readability and Communication in Machine Translation of Arabic Phraseologisms into Spanish. In Computational and Corpus-Based Phraseology - 4th International Conference, Europhras 2022, Malaga, Spain, 28-30 September, 2022, Proceedings (Lecture Notes in Computer Science, Vol. 13528), Gloria Corpas Pastor and Ruslan Mitkov (Eds.). Springer, 78--89. https://doi.org/10.1007/978-3-031-15925-1_6
[14]
R Flesch. 1948. A new readability yardstick. Journal of Applied Psychology 32(3 (1948), 221--233.
[15]
Robert Gunning et al. 1952. Technique of clear writing. (1952).
[16]
Julia Hancke, Sowmya Vajjala, and Detmar Meurers. 2012. Readability Classification for German using Lexical, Syntactic, and Morphological Features. In Proceedings of the 24th International Conference on Computational Linguistics. 1063--1080.
[17]
Michael Ibañez, Lloyd Lois Antonie Reyes, Ranz Sapinit, Mohammed Hussien, and Joseph Marvin Imperial. 2022. On Applicability of Neural Language Models for Readability Assessment in Filipino. In Artificial Intelligence in Education. Posters and Late Breaking Results, Workshops and Tutorials, Industry and Innovation Tracks, Practitioners' and Doctoral Consortium - 23rd International Conference, AIED 2022, Durham, UK, July 27-31, 2022, Proceedings, Part II (Lecture Notes in Computer Science, Vol. 13356), Maria Mercedes T. Rodrigo, Noburu Matsuda, Alexandra I. Cristea, and Vania Dimitrova (Eds.). Springer, 573--576. https: //doi.org/10.1007/978-3-031-11647-6_118
[18]
Joseph Marvin Imperial. 2021. BERT Embeddings for Automatic Readability Assessment. In Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP 2021), Held Online, 1-3 September, 2021, Galia Angelova, Maria Kunilovskaya, Ruslan Mitkov, and Ivelina Nikolova-Koleva (Eds.). INCOMA Ltd., 611--618. https://aclanthology.org/2021.ranlp-1.69
[19]
Shoaib Jameel and Xiaojun Qian. 2012. An Unsupervised Technical Readability Ranking Model by Building a Conceptual Terrain in LSI. In 2012 Eighth International Conference on Semantics, Knowledge and Grids. 39--46. https: //doi.org/10.1109/SKG.2012.20
[20]
Shoaib Jameel, Xiaojun Qian, and Wai Lam. 2012. $N$-Gram Fragment Sequence Based Unsupervised Domain-Specific Document Readability. In Proceedings of COLING 2012. The COLING 2012 Organizing Committee, Mumbai, India, 1309--1326.
[21]
Zhiwei Jiang, Qing Gu, Yafeng Yin, and Daoxu Chen. 2018. Enriching Word Embeddings with Domain Knowledge for Readability Assessment. In Proceedings of the 27th International Conference on Computational Linguistics, COLING 2018, Santa Fe, New Mexico, USA, August 20-26, 2018, Emily M. Bender, Leon Derczynski, and Pierre Isabelle (Eds.). Association for Computational Linguistics, 366--378. https://aclanthology.org/C18-1031/
[22]
Zhiwei Jiang, Qing Gu, Yafeng Yin, Jianxiang Wang, and Daoxu Chen. 2019. GRAW: A two-view graph propagation method with word coupling for readability assessment. J. Assoc. Inf. Sci. Technol. 70, 5 (2019), 433--447. https: //doi.org/10.1002/asi.24123
[23]
Zhiwei Jiang, Gang Sun, Qing Gu, Tao Bai, and Daoxu Chen. 2015. A Graph-based Readability Assessment Method using Word Coupling. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, EMNLP 2015, Lisbon, Portugal, September 17-21, 2015, Lluís Màrquez, Chris Callison-Burch, Jian Su, Daniele Pighin, and Yuval Marton (Eds.). The Association for Computational Linguistics, 411--420. https://doi.org/10.18653/v1/d15-1047
[24]
J Peter Kincaid, Robert P Fishburne Jr, Richard L Rogers, and Brad S Chissom. 1975. Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel. Technical Report. Naval Technical Training Command Millington TN Research Branch.
[25]
Bruce W. Lee, Yoo Sung Jang, and Jason Hyung-Jong Lee. 2021. Pushing on Text Readability Assessment: A Transformer Meets Handcrafted Linguistic Features. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih (Eds.). Association for Computational Linguistics, 10669--10686. https://doi.org/10.18653/v1/2021.emnlp-main.834
[26]
Bruce W. Lee and Jason Hyung-Jong Lee. 2020. LXPER Index 2.0: Improving Text Readability Assessment for L2 English Learners in South Korea. CoRR abs/2010.13374 (2020). arXiv:2010.13374 https://arxiv.org/abs/2010.13374
[27]
Justin Lee and Sowmya Vajjala. 2022. A Neural Pairwise Ranking Model for Readability Assessment. arXiv:2203.07450 [cs] (March 2022). arXiv:2203.07450 [cs]
[28]
Justin Lee and Sowmya Vajjala. 2022. A Neural Pairwise Ranking Model for Readability Assessment. In Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, May 22-27, 2022, Smaranda Muresan, Preslav Nakov, and Aline Villavicencio (Eds.). Association for Computational Linguistics, 3802--3813. https://doi.org/10.18653/v1/2022.findings-acl.300
[29]
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv:1907.11692 [cs]
[30]
Bertha A Lively and Sidney L Pressey. 1923. A method for measuring the vocabulary burden of textbooks. Educational administration and supervision 9, 7 (1923), 389--398.
[31]
Ion Madrazo Azpiazu and Maria Soledad Pera. 2020. An Analysis of Transfer Learning Methods for Multilingual Readability Assessment. In Adjunct Publication of the 28th ACM Conference on User Modeling, Adaptation and Personalization (Genoa, Italy) (UMAP '20 Adjunct). Association for Computing Machinery, New York, NY, USA, 95--100. https://doi.org/10.1145/3386392.3397605
[32]
Matej Martinc, Senja Pollak, and Marko Robnik-?ikonja. 2021. Supervised and Unsupervised Neural Approaches to Text Readability. Computational Linguistics 47, 1 (April 2021), 141--179. https://doi.org/10.1162/coli_a_00398
[33]
G.H McLaughlin. 1969. SMOG grading-a new readability formula. Journal of Reading, 12(8) (1969), 639--646.
[34]
Hamid Mohammadi and Seyed Hossein Khasteh. 2019. Text as Environment: A Deep Reinforcement Learning Text Readability Assessment Model. CoRR abs/1912.05957 (2019). arXiv:1912.05957 http://arxiv.org/abs/1912.05957
[35]
Neil Newbold, Harry McLaughlin, and Lee Gillam. 2010. Rank by Readability: Document Weighting for Information Retrieval. In Advances in Multi-disciplinary Retrieval, David Hutchison, Takeo Kanade, Josef Kittler, Jon M. Kleinberg, Friedemann Mattern, John C. Mitchell, Moni Naor, Oscar Nierstrasz, C. Pandu Rangan, Bernhard Steffen, Madhu Sudan, Demetri Terzopoulos, Doug Tygar, Moshe Y. Vardi, Gerhard Weikum, Hamish Cunningham, Allan Hanbury, and Stefan Rüger (Eds.). Vol. 6107. Springer Berlin Heidelberg, Berlin, Heidelberg, 20--30. https://doi.org/10.1007/978-3-642-13084-7_3
[36]
Florian Pickelmann, Michael Färber, and Adam Jatowt. 2023. Ablesbarkeitsmesser: A System for Assessing the Readability of German Text. In Advances in Information Retrieval - 45th European Conference on Information Retrieval, ECIR 2023, Dublin, Ireland, April 2-6, 2023, Proceedings, Part III (Lecture Notes in Computer Science, Vol. 13982), Jaap Kamps, Lorraine Goeuriot, Fabio Crestani, Maria Maistro, Hideo Joho, Brian Davis, Cathal Gurrin, Udo Kruschwitz, and Annalina Caputo (Eds.). Springer, 288--293. https://doi.org/10.1007/978-3-031-28241-6_28
[37]
Xinying Qiu, Yuan Chen, Hanwu Chen, Jian-Yun Nie, Yuming Shen, and Dawei Lu. 2021. Learning Syntactic Dense Embedding with Correlation Graph for Automatic Readability Assessment. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, Online, 3013--3025. https://doi.org/10.18653/v1/ 2021.acl-long.235
[38]
Andreas Schlapbach, Frank Wettstein, and Horst Bunke. 2008. Estimating the readability of handwritten text - a Support Vector Regression based approach. In 19th International Conference on Pattern Recognition (ICPR 2008), December 8-11, 2008, Tampa, Florida, USA. IEEE Computer Society, 1--4. https://doi.org/10.1109/ ICPR.2008.4761907
[39]
E A Smith and R J Senter. 1967. Automated Readability Index. AMRL-TR. Aerospace Medical Research Laboratories (U.S.) (1967), 1--14.
[40]
Sowmya Vajjala and Ivana Lučić. 2018. OneStopEnglish corpus: A new corpus for automatic readability assessment and text simplification. In Proceedings of the thirteenth workshop on innovative use of NLP for building educational applications. 297--304.
[41]
Tim vor der Brück. 2009. Approximation of the Parameters of a Readability Formula by Robust Regression. In Machine Learning and Data Mining in Pattern Recognition, 6th International Conference, MLDM 2009, Leipzig, Germany, July 2009, Poster Proceedings, Petra Perner (Ed.). ibai Publishing, 115--125.
[42]
Menglin Xia, Ekaterina Kochmar, and Ted Briscoe. 2016. Text Readability Assessment for Second Language Learners. In Proceedings of the 11th Workshop on Innovative Use of NLP for Building Educational Applications. Association for Computational Linguistics, San Diego, CA, 12--22. https://doi.org/10.18653/v1/W16-0502
[43]
Haoran Xie, Minhong Wang, Di Zou, and Fu Lee Wang. 2019. A Personalized Task Recommendation System for Vocabulary Learning Based on Readability and Diversity. In Blended Learning: Educational Innovation for Personalized Learning, Simon K. S. Cheung, Lap-Kei Lee, Ivana Simonova, Tomas Kozel, and Lam-For Kwok (Eds.). Vol. 11546. Springer International Publishing, Cham, 82--92. https://doi.org/10.1007/978-3-030-21562-0_7
[44]
Wei Xu, Chris Callison-Burch, and Courtney Napoles. 2015. Problems in Current Text Simplification Research: New Data Can Help. Transactions of the Association for Computational Linguistics 3 (05 2015), 283--297. https://doi.org/10.1162/tacl_a_00139 arXiv:https://direct.mit.edu/tacl/article-pdf/doi/10.1162/tacl_a_00139/1566780/tacl_a_00139.pdf

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
SIGIR '23: Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval
July 2023
3567 pages
ISBN:9781450394086
DOI:10.1145/3539618
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 18 July 2023

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. multi-signal learning
  2. pairwise ranking
  3. readability assessment
  4. unsupervised ranking

Qualifiers

  • Research-article

Funding Sources

Conference

SIGIR '23
Sponsor:

Acceptance Rates

Overall Acceptance Rate 792 of 3,983 submissions, 20%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 169
    Total Downloads
  • Downloads (Last 12 months)89
  • Downloads (Last 6 weeks)2
Reflects downloads up to 18 Nov 2024

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media