Abstract
Sentence compression is an important problem in natural language processing with wide applications in text summarization, search engine and human–AI interaction system etc. In this paper, we design a hybrid extractive sentence compression model combining a probability language model and a parse tree language model for compressing sentences by guaranteeing the syntax correctness of the compression results. Our compression model is formulated as an integer linear programming problem, which can be rewritten as a difference-of-convex (DC) programming problem based on the exact penalty technique. We use a well-known efficient DC algorithm—DCA to handle the penalized problem for local optimal solutions. Then a hybrid global optimization algorithm combining DCA with a parallel branch-and-bound framework, namely PDCABB, is used for finding global optimal solutions. Numerical results demonstrate that our sentence compression model can provide excellent compression results evaluated by F-score, and indicate that PDCABB is a promising algorithm for solving our sentence compression model.
Similar content being viewed by others
Notes
\(\left[ \!\left[ m,n\right] \!\right] \) with \(m\le n\) stands for the set of integers between m and n.
References
Bertsekas, D.P.: Nonlinear Programming, 2nd edn. MIT Press, Cambridge (1999)
Bird, S., Klein, E., Loper, E.: Natural Language Processing with Python: Analyzing Text with the Natural Language Toolkit. O’Reilly Media Inc, Newton (2009)
Chopra, S., Auli, M., Rush, A.M.: Abstractive sentence summarization with attentive recurrent neural networks. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 93–98 (2016)
Clarke, J., Lapata, M.: Global inference for sentence compression: an integer linear programming approach. J. Artif. Intell. Res. 31, 399–429 (2008)
Cohn, T., Lapata, M.: Sentence compression beyond word deletion. In: Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008), pp. 137–144 (2008)
Filippova, K., Alfonseca, E., Colmenares, C.A., Kaiser, Ł., Vinyals, O.: Sentence compression by deletion with LSTMS. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 360–368 (2015)
Filippova, K., Altun, Y.: Overcoming the lack of parallel data in sentence compression. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pp. 1481–1491 (2013)
Gurobi optimizer reference manual. https://www.gurobi.com
Hori, C., Furui, S.: Speech summarization: an approach through word extraction and a method for evaluation. IEICE Trans. Inf. Syst. 87(1), 15–25 (2004)
Jing, H.Y.: Sentence reduction for automatic text summarization. In: Sixth Applied Natural Language Processing Conference, pp. 310–315 (2000)
Knight, K., Marcu, D.: Summarization beyond sentence extraction: a probabilistic approach to sentence compression. Artif. Intell. 139(1), 91–107 (2002)
Le Thi, H.A., Le, H.M., Pham, D.T., Bouvry, P.: Solving the perceptron problem by deterministic optimization approach based on DC programming and DCA. In: 2009 7th IEEE International Conference on Industrial Informatics, pp. 222–226. IEEE (2009)
Le Thi, H.A., Moeini, M., Pham, D.T.: Portfolio selection under downside risk measures and cardinality constraints based on DC programming and DCA. Comput. Manag. Sci. 6(4), 459–475 (2009)
Le Thi, H.A., Nguyen, Q.T., Nguyen, H.T., Pham, D.T.: Solving the earliness tardiness scheduling problem by DC programming and DCA. Math. Balk. 23(3–4), 271–288 (2009)
Le Thi, H.A., Pham, D.T.: A continuous approach for globally solving linearly constrained quadratic zero-one programming problems. Optimization 50(1–2), 93–120 (2001)
Le Thi, H.A., Pham, D.T.: The DC (difference of convex functions) programming and DCA revisited with DC models of real world nonconvex optimization problems. Ann. Oper. Res. 133(3), 23–46 (2005)
Le Thi, H.A., Pham, D.T.: DC programming and DCA: thirty years of developments. Math. Program. 169(1), 5–68 (2018)
Le Thi, H.A., Pham, D.T., Huynh, V.N.: Exact penalty and error bounds in DC programming. J. Glob. Optim. 52(3), 509–535 (2012)
Le Thi, H.A., Pham, D.T., Muu, L.D.: Exact penalty in DC programming. Vietnam J. Math. 22(2), 169–178 (1999)
Manning, C.D., Surdeanu, M., Bauer, J., Finkel, J.R., Bethard, S., McClosky, D.: The stanford corenlp natural language processing toolkit. In: Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 55–60 (2014)
McDonald, R.: Discriminative sentence compression with soft syntactic evidence. In: 11th Conference of the European Chapter of the Association for Computational Linguistics (2006)
Michele, B., Vibhu, O.M., Michael, J.W.: Headline generation based on statistical translation. In: Proceedings of the 38th Annual Meeting on Association for Computational Linguistics, pp. 318–325. Association for Computational Linguistics (2000)
Niu, Y.S.: Programmation dc et dca en optimisation combinatoire et optimisation polynomiale via les techniques de sdp. Ph.D. thesis, INSA de Rouen, France (2010)
Niu, Y.S.: PDCABB: a parallel mixed-integer nonlinear optimization solver (parallel DCA-BB global optimization algorithm). https://github.com/niuyishuai/PDCABB (2017)
Niu, Y.S.: On difference-of-SOS and difference-of-convex-SOS decompositions for polynomials. arXiv preprint arXiv:1803.09900 (2018)
Niu, Y.S.: A parallel branch and bound with DC algorithm for mixed integer optimization. In: The 23rd International Symposium in Mathematical Programming (ISMP2018), Bordeaux, France (2018)
Niu, Y.S., Pham, D.T.: A dc programming approach for mixed-integer linear programs. In: International Conference on Modelling, Computation and Optimization in Information Systems and Management Sciences, pp. 244–253. Springer (2008)
Niu, Y.S., Pham, D.T.: Efficient dc programming approaches for mixed-integer quadratic convex programs. In: Proceedings of the International Conference on Industrial Engineering and Systems Management (IESM2011), Metz, France, pp. 222–231 (2011)
Niu, Y.S., You, Y., Liu, W.Z.: Parallel DC cutting plane algorithms for mixed binary linear program. In: Proceedings of the World Congress on Global Optimization (WCGO2019), pp. 330–340 (2019)
NLTK 3.2.5: The natural language toolkit. http://www.nltk.org
Over, P., Dang, H., Harman, D.: DUC in context. Inf. Process. Manag. 43(6), 1506–1520 (2007)
Pham, D.T., Le Thi, H.A.: Convex analysis approach to D.C. programming: theory, algorithm and applications. Acta Math. Vietnam. 22(1), 288–355 (1997)
Pham, D.T., Le Thi, H.A.: A D.C. optimization algorithm for solving the trust-region subproblem. SIAM J. Optim. 8(2), 476–505 (1998)
Pham, D.T., Le Thi, H.A., Pham, V.N., Niu, Y.S.: DC programming approaches for discrete portfolio optimization under concave transaction costs. Optim. Lett. 10(2), 261–282 (2016)
Rockafellar, R.T.: Convex Analysis. Princeton University Press, Princeton (1970)
Rush, A.M., Chopra, S., Weston, J.: A neural attention model for abstractive sentence summarization. arXiv preprint arXiv:1509.00685 (2015)
Schleich, J., Le Thi, H.A., Bouvry, P.: Solving the minimum m-dominating set problem by a continuous optimization approach based on DC programming and DCA. J. Comb. Optim. 24(4), 397–412 (2012)
Stanford CoreNLP, natural language software. https://stanfordnlp.github.io/CoreNLP/
Zajic, D., Dorr, B.J., Schwartz, R.: Bbn/umd at duc-2004: Topiary. In: Proceedings of the HLT-NAACL 2004 Document Understanding Workshop, Boston pp. 112–119 (2004)
Acknowledgements
The authors are supported by the National Natural Science Foundation of China (Grant 11601327) and the National “985” Key Program of China (Grant WF220426001).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix
A List of labels for POS tagging and parsing
Some labels used for sentence parsing and POS tagging are listed in Table 4. Some notations refer to POS tags of the Penn Treebank Tagset [30], others are introduced by ourselves. For example, notation “ADJ” refers to adjective tag cases “JJ”, “JJR” and “JJS” in the Penn Treebank Tagset, but “ADJP” is a notation that we defined to denote ‘adjective phrase’.
B List of CFG grammar and compression rules for statements
See Table 5.
Rights and permissions
About this article
Cite this article
Niu, YS., You, Y., Xu, W. et al. A difference-of-convex programming approach with parallel branch-and-bound for sentence compression via a hybrid extractive model. Optim Lett 15, 2407–2432 (2021). https://doi.org/10.1007/s11590-020-01695-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11590-020-01695-9