Nothing Special   »   [go: up one dir, main page]

Skip to main content

A Comparison of CNN and Conventional Descriptors for Word Spotting Approach: Application to Handwritten Document Image Retrieval

  • Conference paper
  • First Online:
Innovative Systems for Intelligent Health Informatics (IRICT 2020)

Abstract

Natural images are easier to represent in feature space than textual images due to the reduced complexity and thus do not require greater learning capacity. The visual information representation is an important step in content-based image retrieval (CBIR) systems, used for searching relevant visual information in large image datasets. The extraction of discriminant features can be carried out using two approaches. Manual (conventional CBIR), based on preselected features (colors, shapes, or textures) and automatic (modern CBIR) based on auto-extracted features using deep learning models. This second approach is more robust to the complexity relative to textual images, which require a deep representation reaching the semantics of text in the image. DIRS (Document Image Retrieval Systems) are CBIR systems related to documents images that propose a set of efficient word-spotting techniques, such as the interest points based techniques, which offer an effective local image representation. This paper presents an overview of existing word retrieval techniques and a comparison of our two proposed word-spotting approaches (interest points and CNN description), applied on handwritten documents. The results obtained on degraded and old Bentham datasets are compared with those of the literature.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Rath, T.M., Manmatha, R.: Features for word spotting in historical manuscripts. In: Proceedings of the Seventh International Conference on Document Analysis and Recognition, 2003, pp. 218–222. IEEE Computer Society, Edinburgh, UK (2003). https://doi.org/10.1109/ICDAR.2003.1227662

  2. Rodriguez, J.A., Perronnin, F.: Local Gradient Histogram Features for Word Spotting in Unconstrained Handwritten Documents. 1st ICFHR, p. 6 (2008)

    Google Scholar 

  3. Rusiñol, M., Lladós, J.: Efficient logo retrieval through hashing shape context descriptors. In: Proceedings of the 8th IAPR International Workshop on Document Analysis Systems - DAS’10. pp. 215–222. ACM Press, Boston, Massachusetts (2010). https://doi.org/10.1145/1815330.1815358

  4. Wang, P.: Historical Handwriting Representation Model Dedicated to Word Spotting Application (2014)

    Google Scholar 

  5. Zagoris, K., Pratikakis, I., Gatos, B.: Unsupervised word spotting in historical handwritten document images using document-oriented local features. IEEE Trans. on Image Process. 26, 4032–4041 (2017). https://doi.org/10.1109/TIP.2017.2700721

    Article  MathSciNet  MATH  Google Scholar 

  6. Konidaris, T., Kesidis, A.L., Gatos, B.: A segmentation-free word spotting method for historical printed documents. Pattern Anal. Appl. 19, 963–976 (2016). https://doi.org/10.1007/s10044-015-0476-0

    Article  MathSciNet  Google Scholar 

  7. Zagoris, K., Pratikakis, I., Gatos, B.: Unsupervised word spotting in historical handwritten document images using document-oriented local features. IEEE Trans. Image Process. 26, 4032–4041 (2017). https://doi.org/10.1109/TIP.2017.2700721

    Article  MathSciNet  MATH  Google Scholar 

  8. Sudholt, S., Fink, G.A.: PHOCNet: a deep convolutional neural network for word spotting in handwritten documents. In: 2016 15th International Conference on Frontiers in Handwriting Recognition (ICFHR), pp. 277–282. IEEE, Shenzhen, China (2016). https://doi.org/10.1109/ICFHR.2016.0060

  9. Zhong, Z., Pan, W., Jin, L., Mouchere, H., Viard-Gaudin, C.: SpottingNet: learning the similarity of word images with convolutional neural network for word spotting in handwritten historical documents. In: 2016 15th International Conference on Frontiers in Handwriting Recognition (ICFHR), pp. 295–300. IEEE, Shenzhen, China (2016). https://doi.org/10.1109/ICFHR.2016.0063

  10. Rothacker, L., Rusinol, M., Fink, G.A.: Bag-of-Features HMMs for segmentation-free word spotting in handwritten documents. In: 2013 12th International Conference on Document Analysis and Recognition. pp. 1305–1309. IEEE, Washington, DC, USA (2013). https://doi.org/10.1109/ICDAR.2013.264

  11. Wiggers, K.L., Britto, A.S., Heutte, L., Koerich, A.L., Oliveira, L.S.: Image retrieval and pattern spotting using siamese neural network. In: 2019 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Budapest, Hungary (2019). https://doi.org/10.1109/IJCNN.2019.8852197

  12. Benabdelaziz, R., Gaceb, D., Haddad, M.: Word spotting based on bispace similarity for visual information retrieval in handwritten document images. Int. J. Comput. Vis. Image Process. 9, 38–58 (2019). https://doi.org/10.4018/IJCVIP.2019070103

    Article  Google Scholar 

  13. Benabdelaziz, R., Gaceb, D., Haddad, M.: Word-Spotting approach using transfer deep learning of a CNN network. In: 020 1st International Conference on Communications, Control Systems and Signal Processing (CCSSP). pp. 219–224. IEEE, EL OUED, Algeria (2020). https://doi.org/10.1109/CCSSP49278.2020.9151583

  14. Puigcerver, J., Toselli, A.H., Vidal, E.: ICDAR2015 competition on keyword spotting for handwritten documents. In: 2015 13th International Conference on Document Analysis and Recognition (ICDAR). pp. 1176–1180. IEEE, Tunis, Tunisia (2015). https://doi.org/10.1109/ICDAR.2015.7333946

  15. Retsinas, G., Louloudis, G., Stamatopoulos, N., Gatos, B.: Keyword spotting in handwritten documents using projections of oriented gradients. In: 2016 12th IAPR Workshop on Document Analysis Systems (DAS), pp. 411–416. IEEE, Santorini, Greece (2016). https://doi.org/10.1109/DAS.2016.61

  16. Sfikas, G., Retsinas, G., Gatos, B.: Zoning aggregated hypercolumns for keyword spotting. In: 2016 15th International Conference on Frontiers in Handwriting Recognition (ICFHR), pp. 283–288. IEEE, Shenzhen, China (2016). https://doi.org/10.1109/ICFHR.2016.0061

  17. Sudholt, S., Rothacker, L., Fink, G.A.: Learning local image descriptors for word spotting. In: 2015 13th International Conference on Document Analysis and Recognition (ICDAR), pp. 651–655. IEEE, Tunis, Tunisia (2015). https://doi.org/10.1109/ICDAR.2015.7333842

  18. Ghosh, S.K., Valveny, E.: A sliding window framework for word spotting based on word attributes. In: Paredes, R., Cardoso, J.S., Pardo, X.M. (eds.) Pattern Recognition and Image Analysis. Lecture Notes in Computer Science, vol. 9117, pp. 652–661. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-19390-8_73

    Chapter  Google Scholar 

  19. Zagoris, K., Pratikakis, I., Gatos, B.: Segmentation-based historical handwritten word spotting using document-specific local features. In: 2014 14th International Conference on Frontiers in Handwriting Recognition, pp. 9–14. IEEE, Greece (2014). https://doi.org/10.1109/ICFHR.2014.10

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ryma Benabdelaziz .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Benabdelaziz, R., Gaceb, D., Haddad, M. (2021). A Comparison of CNN and Conventional Descriptors for Word Spotting Approach: Application to Handwritten Document Image Retrieval. In: Saeed, F., Mohammed, F., Al-Nahari, A. (eds) Innovative Systems for Intelligent Health Informatics. IRICT 2020. Lecture Notes on Data Engineering and Communications Technologies, vol 72. Springer, Cham. https://doi.org/10.1007/978-3-030-70713-2_99

Download citation

Publish with us

Policies and ethics