Next Article in Journal
A Robust Tensor-Based Submodule Clustering for Imaging Data Using l12 Regularization and Simultaneous Noise Recovery via Sparse and Low Rank Decomposition Approach
Next Article in Special Issue
Digital Hebrew Paleography: Script Types and Modes
Previous Article in Journal
Liouville Integrability in a Four-Dimensional Model of the Visual Cortex
Previous Article in Special Issue
Attention-Based Fully Gated CNN-BGRU for Russian Handwritten Text
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Word Spotting as a Service: An Unsupervised and Segmentation-Free Framework for Handwritten Documents

by
Konstantinos Zagoris
1,
Angelos Amanatiadis
2,* and
Ioannis Pratikakis
3
1
Department of Computer Science, Neapolis University, Pafos 8042, Cyprus
2
Department of Production and Management Engineering, Democritus University of Thrace, 67132 Xanthi, Greece
3
Department of Electrical and Computer Engineering, Democritus University of Thrace, 67132 Xanthi, Greece
*
Author to whom correspondence should be addressed.
Submission received: 15 November 2021 / Revised: 10 December 2021 / Accepted: 15 December 2021 / Published: 17 December 2021

Abstract

:
Word spotting strategies employed in historical handwritten documents face many challenges due to variation in the writing style and intense degradation. In this paper, a new method that permits efficient and effective word spotting in handwritten documents is presented that relies upon document-oriented local features that take into account information around representative keypoints and a matching process that incorporates a spatial context in a local proximity search without using any training data. The method relies on a document-oriented keypoint and feature extraction, along with a fast feature matching method. This enables the corresponding methodological pipeline to be both effectively and efficiently employed in the cloud so that word spotting can be realised as a service in modern mobile devices. The effectiveness and efficiency of the proposed method in terms of its matching accuracy, along with its fast retrieval time, respectively, are shown after a consistent evaluation of several historical handwritten datasets.

1. Introduction

The constant increase in power for mobile-edge computing, along with the availability of various cloud platforms, permits the development of services that were not feasible before for mobile device users [1]. A representative service of this kind is word spotting, which allows for users to retrieve query words of interest in handwritten document images without the need for transcription [2].
The original idea of word spotting was introduced in [3]. It is equivalent to a content-based image retrieval process since it is based on low-level pattern matching, wherein a search in a set of unindexed documents is applied using a query word. In a word spotting context, the image content is the only information source that is used. The outcome of the search is a ranked list of word images that has been retrieved by taking into account the similarity with the query word image. Dealing with document images, word spotting can be viewed as the task of identifying specific locations on a document image that have a high probability to correspond to a queried word image without explicitly recognizing it.
This work is based on an approach that is termed as segmentation-free because the search space is the complete document image, which is contrary to the case that is termed as segmentation-based, wherein the search space is a set of segmented word images. Although a segmentation-free approach is more challenging due to the unconstrained search space that should be dealt with, it has the potential to result in an improved performance in the case of considerable degradation in the document, where the required word segmentation step will introduce many errors, leading to an undesirable erroneous word detection.
For both the aforementioned strategies, the core operational pipeline used relies upon two main components: features extraction and matching. Although there is a plethora of systems that are tailored for both modern [4,5] and historical machine-printed documents [6,7,8,9], in the case of handwritten documents, very few of them can be used effectively due to severe degradation, text layout complexity and variation in the writing style.
The proposed work is an extended version of the work presented in [10,11] using document-oriented local features that consider the information around representative keypoints aiming to address the problem of accessing big data. In this respect, we have introduced a two-folded novelty that comprises a new indexing method, along with a new matching method. Furthermore, we offer a new perspective in the use of word spotting as a service through a mobile device and a cloud service.
Figure 1 shows the communication between a client device and the word spotting. The blue arrow points to the components of the OFFLINE procedure, whereas the green arrow points to the components of the ONLINE procedure, which is described in Section 3. First, the document image is analyzed during the OFFLINE procedure, and the DoLF local keypoints and features are calculated. Then, taking into account DoLF, the corresponding indexed features are created, which are used in the feature matching step during the ONLINE procedure (green arrow) using the corresponding features of the query word image.

2. Related Work

Previous approaches to word spotting can be distinguished as supervised or unsupervised based on the need for training data. Furthermore, each approach can be characterized by its dependency on segmentation, resulting in it being considered as either segmentation-based or segmentation-free.
Initial efforts in segmentation-based word spotting [12,13,14,15] used preprocessing techniques, such as binarization and document layout analysis, followed by word image segmentation. Next, for each segmented word image, a set of features is extracted. Finally, taking into account the extracted features, the similarity between the query word image and each segmented word image that has been found in a collection of document images is measured.
Certain works [16,17] on unsupervised word spotting do not require segmentation at either the textline or word level, and are thus denoted as segmentation-free. Instead, they target detecting the most informative parts in the image based on the gradient orientation, and apply a matching procedure based on naive elastic matching. Their main disadvantages are the sensitivity to different writing variations and matching words with different font sizes. Moreover, their matching engine is time-consuming, making its use impractical for large datasets and very costly to provide as a service.
Another similar approach [18] presented a template matching scheme that relies upon a block-based document image descriptor. However, the use of various queries result in introducing much noise in the final merging state from the different queries, especially in the case that variations in the words’ writing and size exist. In the work of Kovalchuk et al. [19], a comparative approach is presented for which they avoid the production of different variations by resizing them to fit a fixed-size rectangle, and, in the sequel, proceed with a HOG and LBP descriptor extraction. Finally, for the retrieval, a nearest-neighbour search is performed, followed by a suppression of extra overlapping candidates.
Apart from the methods mentioned above, which operate in an unsupervised context, many proposed procedures use training data to learn similarities. One such method is Almazan et al. [20], which does not use only word images but also text strings that are both embedded in a common vectorial subspace wherein images and strings representing the same word are found in proximity, which permits treating recognition and retrieval tasks as the nearest neighbour problem. However, the main drawback is the need for many training samples to create a suitable subspace, making it sensitive for unknown words.
Another strategy, aside from using words or the whole document, is by detecting lines. These approaches use Hidden Markov Models (HMMs) models [21] to spot the words inside the lines, embedding a discriminative stage in HMM, such as a support vector machine [22], a neural network [23] or deep learning network architectures [24]. Although they provide more reliable results than the other approaches, they have many drawbacks. Firstly, they required huge amounts of training data, and some cannot identify words that are not present in the training set. Moreover, their dependency on textline segmentation has the same drawbacks as the word segmentation does for degraded documents.
In the work of Llados et al. [25], the performance of various word descriptors in a Bag of Visual Words (BoVW) context is evaluated. They concluded that the statistical approach of the BoVW produces the best performance, although the descriptor storage has significant memory requirements.
Finally, Zagoris et al. [10] presented an unsupervised word detection method that employed novel local features, namely, Document-oriented Local Features (DoLF), whose texture information is considered in a wider neighborhood spatial context. Both segmentation-based and segmentation-free scenarios have been considered. Although they provide superior results for an unsupervised approach, their matching procedure is not efficient for large document sets, and thus they are prohibited to being provided as a cloud service.
This paper proposes a holistic unsupervised segmentation-free method for word spotting, addressing the limitations of the works mentioned above that are suitable for cloud service. It employs the DoLF local features from [10] on preprocessing documents using a completely different and dynamic matching procedure suitable for efficient results on big datasets without compromising the effectiveness.

3. Proposed Methodology

Figure 2 shows the proposed methodology. First, the DoLF local features are calculated and detected on a preprocessing document image. Afterwards, an indexing procedure uses the calculated DoLF and creates a set of data structures that allow for efficient and effective word spotting. Next, the user selects the query word image on which the DoLF local features are calculated. Finally, a novel matching procedure called Quantitative Near Neighborhood Search (QNNS) detects visually similar regions to this query word image on documents and presents them to the user.

3.1. Preprocessing and Local Points Calculation

The initial step contains a pre-processing step that aims to enhance the contrast between the foreground and background pixels of the document image and to keep the valuable edge boundary information. This enhances the effectiveness of the subsequent calculation of the gradient-based keypoints and descriptors. The required pre-processing step is the only operation that runs on-board on the client application, and comprises two steps. The first step consists of a contrast normalization method [26], which deals with the illumination changes imposed by the mobile device and the user. Subsequently, the gradient vectors I x and I y of the document image I are calculated and filtered by a high pass filter for eliminating any remaining background noise. The filter thresholds are calculated dynamically based on the Otsu algorithm for minimizing the intra-class variance between the foreground and background pixel clusters.
The two filtered gradient images, as shown in Figure 3a,b, are then encoded in a JavaScript Object Notation (JSON) format and passed to the server application for the keypoint extraction.
For the keypoint and feature extraction, we adapted a version of the Document-oriented Local Features (DoLF) proposed by the authors [10]. While most works that use local features are based on the Scale Invariant Feature Transform (SIFT) [27], which is initially used for natural images, in comparison with the document images, they have many structural differences that create problems, such as:
  • The erroneous local points detection between the document lines due to the image pyramid scaling;
  • The invariant properties of those descriptors that amplify noise. As stated in [17], features that are invariant to rotation result in a worse performance compared to features that depend on rotation. Furthermore, they adhere to the observation that invariant-to-rotation features are more sensitive to the noise.
On the contrary, the DoLF exhibit some desirable characteristics that make them suitable for the proposed method, such as:
  • They take into consideration the handwritten document particularities;
  • They provide consistency between different handwritten writing variations, as shown in Figure 4;
  • Their descriptors contain texture information in a spatial context, which is suitable in dealing with a document collection created by different writers, containing significant writing style variations.
The DoLF computation comprises two parts: the keypoint detection and the feature calculation around it. Subsequently, a brief description of the keypoints and their descriptors is provided. For a detailed description, the work in [10,28] should be consulted.
The next step involves the linear quantization of the gradient orientation. This step aims to label the changes to the writing direction, as these points consist of important and descriptive information. Figure 3d shows the output for the quantization of the gradient orientation values. Each color represents a different quantization level. The current work uses the default four different quantization levels as presented in [10,28].
Next, for each quantization level, the Connected Components (CC) are detected. These CCs represent chunks of strokes that correspond to different writing directions. The final local points are the center of gravity of each remaining CC. An example of these keypoints is shown in Figure 3e. The keypoint calculation method can detect meaningful points of the characters that reside in the documents independently of its scale. Moreover, it provides some consistency between different handwritten writing variations.
The next step involves the calculation of the features around the detected keypoints. The descriptor is calculated upon a scale-invariance window size around the detected local points, as shown in Figure 5a. The window size is defined dynamically by calculating the mean brightness in different window sizes and selecting the one that has the maximum value, as described in [29,30].
Then, the selected window size is divided into 16 cells, where a four-bin histogram is calculated, representing each cell (each bin corresponds to a quantization level). Each pixel inside a cell accumulates a vote in the corresponding angle histogram bin, as shown in Figure 5b. The strength of voting depends on the pixel’s magnitude of the gradient vector.
All histograms are concatenated in a single 64-bin histogram and normalized by its norm. Finally, all values above 0.2 are set to 0.2 and are re-normalized again to minimize the illumination effect in the descriptor [27]. The final descriptor is shown in Figure 5c.

3.2. Features Indexing Method

The sheer magnitude of the detected local points and their descriptors increases the costs in terms of time, memory and storage requirements, rendering them unusable for providing a service. To solve this issue, we transform the information from the DoLF to several different structures that permit efficient word spotting without compromising on the effectiveness.
The proposed indexing method comprises two distinct steps:
  • A descriptor quantization using multiple Bag of Visual Words (BoVW) to decompose and compress its information;
  • Three different memory and storage structures for very quickly segmentation-free word spotting during the client request.
Figure 6 shows the overall architecture of the quantization based on Bag of Visual Words.

3.2.1. Descriptor Quantization

At this point, the descriptor is split into four different parts. This is because each part describes a different spatial part of the descriptors. Next, four different codebooks are created for the different parts of the descriptor.
The size of each codebook corresponds to the descriptor precision, but with increasing processing and storage costs. For our experiments, a codebook of size 16 is used. The final quantized descriptor is a four-bin histogram, with each bin in the range of 0–15 values.

3.2.2. Indexing Data Structures

Then, three data structures are created to facilitate the low retrieval times, incorporating the information from the local points location and the corresponding quantized descriptor. The memory invert file structure, the descriptors storage structure and the spatial hash structure are shown in Figure 7.
The Keypoints Storage Structure (KSS) is the simplest and most prominent data structure. It contains all of the calculated keypoints and their descriptors. It is a struct array in which the array index corresponds to the keypoints I d . Each struct (72 bytes) holds the keypoint location (X, Y), the document that resides ( D o c I D ) and its quantized descriptor.
The Memory Invert File Structure (MIFS) resides in memory. It is an array of 64-bit integer lists (8 bytes). Each descriptor corresponds to a hash id l, which is calculated from the following hash function:
l = B O W 1 K 3 + B O W 2 K 2 + B O W 3 K + B O W 4
where B O W i corresponds to each quantized value of the descriptors (Figure 8a) and K is equal to the codebook size. In our proposed method, K = 16 .
The MIFS is a dictionary-based structure that maps a list of keypoints K with the same hash value l. Thus, the MIFS primary function is to retrieve a list of descriptors D that have the same hash value l.
The Spatial Hash Structure (SHS) assists the detection of every keypoint that resides in a specific location inside a document, as it provides information about the spatial distribution of the local points inside a document. Its primary function is to retrieve a list of descriptors D whose local points location is close to each other inside the document.
This is achieved by using the following hash function:
h ( x , y , n ) = n A 2 + y A + x
where A is any custom-defined number that is greater than the maximum width and height document inside a collection.
The above hash function takes the location of the keypoint (x, y) and the document n as an input, resulting in the h a s h I D computation of the corresponding local point.
The retrieval of all of the DoLF in proximity to the point ( x 1 , y 1 ) is achieved by:
First, calculating the h a s h I d s for all of the local points residing inside the space ( x ± d x , y ± d y ) for the document n. The d x and d y denote the distance space from the local point ( x 1 , y 1 ) . Next, all of the descriptor IDs are retrieved from the SHS that corresponds to that specific h a s h I d s. Finally, through the KSS, all of the descriptor information is available.
These three storage structs are the minimum required information to be stored from a documents’ collection. Only the MIFS needs to reside in memory; the other two can be stored in disks and accessed when needed.

3.3. Feature Matching

Feature matching is the only procedure that affects the user, as it commences during the word spotting search. Initially, the query image is analyzed, and the DoLF is calculated. From them, the quantized descriptors and the corresponding hash values l are calculated based on the previous trained BoVW. Finally, the hash values l are calculated based on the Equation (1) for each DoLF.
Through the MIFS, all of the descriptor IDs with the same hash value l with the query l value are retrieved. Finally, their locations and descriptors (denoted as d e s c r i p t o r s n ) are retrieved through KSS.
The feature matching goal is to identify those keypoints with a similar spatial distribution and descriptors with the query keypoints.
First, the nearest keypoint Q k c from the ( c x , c y ) point is identified. The ( c x , c y ) is denoted as the mean center ( c x , c y ) of the keypoints set in the query word image, and is calculated as:
( c x , c y ) = i = 1 k p x i k , i = 1 k p y i k
where p x i , p y i denote the location of the ith keypoint and k denotes the total number of the keypoints in a word image.
The next step involves identifying the most similar local points with the Q K c from the retrieved d e s c r i p t o r s n .
In our implementation, the Euclidean Distance ( E D ) is used, and the top N matches that are kept denote those that have a distance from the query keypoint Q k c feature equal to those that have a distance from the query keypoint Q k c feature lower than a threshold t. This threshold is experimentally defined and controls the time expense of the search in the document space. Each keypoint that belongs to the top N matches is a document candidate coordinate origin similar to the query image.
The spatial NNS for each keypoint that resides on the query image is addressed in the next stage. The spatial NNS is realized in a search area around each point. During the search, if there are one or more keypoints in the proximity of the query keypoint under consideration, the Euclidean distance between their descriptors is calculated and the minimum distance is kept. The previous procedure is repeated for each keypoint in the query image. The final similarity measure is the average of all of the minimum distances. In the case that a local point in its proximity does not exist, then the query local point is ignored.
As a final stage, the system presents all of the word images based on an ascending sort order of the calculated similarity measure. The server-side returns the user’s results in ascending order in an Extensible Markup Language (XML) format.

4. Experimental Results

The proposed method and corresponding service were evaluated on the following four different handwritten document collections:
  • English dataset that contains 115 pages/images and 15,923 words;
  • German dataset that contains 100 pages/images and 15,579 words;
  • Finnish dataset that contains 56 pages/images and 16,201 words;
  • Greek dataset that contains 120 pages/images and 18,809 words.
Figure 9 shows a representative document image from each dataset.
Initially, the proposed method is evaluated against the original method [10] that used the DoLF local points, which was superior to most unsupervised word spotting approaches. Table 1 shows that the proposed method and the original achieve a similar performance, while the novel indexing scheme used in the proposed method results in a compression of the available information from DoLF, thus enabling an efficient and effective word spotting on big datasets.
In the sequel, our experimental work contains a comparison between the proposed method and the original method in terms of time, memory and storage requirements, which are essential for reducing the running cost for the word spotting service. Since the original method has extensive storage and memory requirements when used in a production environment, it becomes unsuitable for large datasets. Table 2 shows that the proposed method provides a massive reduction in the retrieval time per query, memory requirement per document and storage requirement without compromising its performance, as shown in Table 1.
Finally, aiming to demonstrate how the proposed method will scale up based on the dataset size, we record the time, memory and storage requirements with different dataset sizes, namely, 50, 5000 and 50,000 documents, respectively. Table 3 shows that the retrieval time per query increases in a non-linear manner, making the search feasible in terms of time consumption for large scale datasets.
Table 4 shows the results for the ICDAR 2015 Competition segmentation-free TRACK I.B., which involves 1421 queries. It is worth noting that the marginally better performance of the PRG group uses a word segmentation algorithm, creating a dependency from the context of a segmentation-based scenario. Therefore, with a more complicated document layout, the PRG group’s performance will decrease considerably. However, in our proposed segmentation-free method, there is no performance loss despite extensive compression due to the proposed indexing.

5. Conclusions

In this paper, we present an effective and efficient word spotting method for handwritten documents that permits supporting a cloud-based service-oriented perspective in the use of word spotting. After employing a consistent experimental work, we have shown that the proposed method considerably decreases the retrieval time and the overall memory and storage requirements without a substantial performance loss. The improved efficiency, along with the appealing effectiveness achieved, enables word spotting to be addressed as a mobile device service that relies on the lack of any training dependency and its novel fast and spatial feature matching process.

Author Contributions

Conceptualization, K.Z., A.A. and I.P.; methodology, K.Z., A.A. and I.P.; software, K.Z. and A.A.; validation, K.Z. and A.A.; resources, I.P.; writing—original draft preparation, K.Z. and A.A.; writing—review and editing, K.Z., A.A. and I.P.; supervision, I.P.; funding acquisition, I.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been co-financed by the European Union and Greek national funds through the Operational Program Competitiveness, Entrepreneurship and Innovation, under the call RESEARCH-CREATE-INNOVATE (project code: T1EDK-01939).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

We would like to thank NVIDIA Corporation, which kindly donated the Titan X GPU, that has been used for this research.

Conflicts of Interest

The funders had no role in the design of the study; in the collection, analyses or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Corcoran, P.; Datta, S.K. Mobile-edge computing and the Internet of Things for consumers: Extending cloud computing and services to the edge of the network. IEEE Consum. Electron. Mag. 2016, 5, 73–74. [Google Scholar] [CrossRef]
  2. Pratikakis, I.; Zagoris, K.; Gatos, B.; Louloudis, G.; Stamatopoulos, N. ICFHR 2014 competition on handwritten keyword spotting (H-KWS 2014). In Proceedings of the 14th International Conference on Frontiers in Handwriting Recognition, Hersonissos, Greece, 1–4 September 2014; pp. 814–819. [Google Scholar]
  3. Manmatha, R.; Han, C.; Riseman, E.M. Word spotting: A new approach to indexing handwriting. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 18–20 June 1996; pp. 631–637. [Google Scholar]
  4. Hassan, E.; Chaudhury, S.; Gopal, M. Word shape descriptor-based document image indexing: A new DBH-based approach. Int. J. Doc. Anal. Recognit. 2013, 16, 227–246. [Google Scholar] [CrossRef]
  5. Zagoris, K.; Kavallieratou, E.; Papamarkos, N. A Document Image Retrieval System. Eng. Appl. Artif. Intell. 2010, 23, 872–879. [Google Scholar] [CrossRef]
  6. Kesidis, A.L.; Galiotou, E.; Gatos, B.; Pratikakis, I. A word spotting framework for historical machine-printed documents. Int. J. Doc. Anal. Recognit. 2011, 14, 131–144. [Google Scholar] [CrossRef]
  7. Konidaris, T.; Gatos, B.; Perantonis, S.J.; Kesidis, A. Keyword matching in historical machine-printed documents using synthetic data, word portions and dynamic time warping. In Proceedings of the The Eighth IAPR International Workshop on Document Analysis Systems, Nara, Japan, 16–19 September 2008; pp. 539–545. [Google Scholar]
  8. Zirari, F.; Ennaji, A.; Nicolas, S.; Mammass, D. A methodology to spot words in historical arabic documents. In Proceedings of the ACS International Conference on Computer Systems and Applications, Ifrane, Morocco, 27–30 May 2013; pp. 1–4. [Google Scholar]
  9. Khurshid, K.; Faure, C.; Vincent, N. Word spotting in historical printed documents using shape and sequence comparisons. Pattern Recognit. 2012, 45, 2598–2609. [Google Scholar] [CrossRef]
  10. Zagoris, K.; Pratikakis, I.; Gatos, B. Unsupervised word spotting in historical handwritten document images using document-oriented local features. IEEE Trans. Image Process. 2017, 26, 4032–4041. [Google Scholar] [CrossRef] [PubMed]
  11. Amanatiadis, A.; Zagoris, K.; Pratikakis, I. Word Spotting as a Service for Handwritten Documents. In Proceedings of the IEEE International Conference on Consumer Electronics, Las Vegas, NV, USA, 10–12 January 2021; pp. 1–2. [Google Scholar]
  12. Rath, T.M.; Manmatha, R. Features for word spotting in historical manuscripts. In Proceedings of the Seventh International Conference on Document Analysis and Recognition, Edinburgh, UK, 6 August 2003; pp. 218–222. [Google Scholar]
  13. Rath, T.M.; Manmatha, R. Word image matching using dynamic time warping. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Madison, WI, USA, 18–20 June 2003; Volume 2, p. II. [Google Scholar]
  14. Lavrenko, V.; Rath, T.M.; Manmatha, R. Holistic word recognition for handwritten historical documents. In Proceedings of the 1st International Workshop on Document Image Analysis for Libraries, Palo Alto, CA, USA, 23–24 January 2004; pp. 278–287. [Google Scholar]
  15. Zagoris, K.; Kavallieratou, E.; Papamarkos, N. Image retrieval systems based on compact shape descriptor and relevance feedback information. J. Vis. Commun. Image Represent. 2011, 22, 378–390. [Google Scholar] [CrossRef]
  16. Leydier, Y.; Lebourgeois, F.; Emptoz, H. Text search for medieval manuscript images. Pattern Recognit. 2007, 40, 3552–3567. [Google Scholar] [CrossRef]
  17. Leydier, Y.; Ouji, A.; LeBourgeois, F.; Emptoz, H. Towards an omnilingual word retrieval system for ancient manuscripts. Pattern Recognit. 2009, 42, 2089–2105. [Google Scholar] [CrossRef]
  18. Gatos, B.; Pratikakis, I. Segmentation-free word spotting in historical printed documents. In Proceedings of the 10th International Conference on Document Analysis and Recognition, Barcelona, Spain, 26–29 July 2009; pp. 271–275. [Google Scholar]
  19. Kovalchuk, A.; Wolf, L.; Dershowitz, N. A simple and fast word spotting method. In Proceedings of the 14th International Conference on Frontiers in Handwriting Recognition, Hersonissos, Greece, 1–4 September 2014; pp. 3–8. [Google Scholar]
  20. Almazan, J.; Gordo, A.; Fornes, A.; Valveny, E. Word Spotting and Recognition with Embedded Attributes. IEEE Tran. Pattern Anal. Mach. Intell. 2014, 36, 2552–2566. [Google Scholar] [CrossRef] [PubMed]
  21. Fischer, A.; Keller, A.; Frinken, V.; Bunke, H. Lexicon-free handwritten word spotting using character HMMs. Pattern Recognit. Lett. 2012, 33, 934–942. [Google Scholar] [CrossRef]
  22. Huang, B.Q.; Du, C.; Zhang, Y.; Kechadi, M. A hybrid HMM-SVM method for online handwriting symbol recognition. In Proceedings of the Sixth International Conference on Intelligent Systems Design and Applications, Jian, China, 16–18 October 2006; Volume 1, pp. 887–891. [Google Scholar]
  23. Espana-Boquera, S.; Castro-Bleda, M.; Gorbe-Moya, J.; Zamora-Martinez, F. Improving Offline Handwritten Text Recognition with Hybrid HMM/ANN Models. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 767–779. [Google Scholar] [CrossRef] [PubMed]
  24. Thomas, S.; Chatelain, C.; Heutte, L.; Paquet, T.; Kessentini, Y. A deep HMM model for multiple keywords spotting in handwritten documents. Pattern Anal. Appl. 2015, 18, 1003–1015. [Google Scholar] [CrossRef] [Green Version]
  25. Lladós, J.; Rusinol, M.; Fornés, A.; Fernández, D.; Dutta, A. On the influence of word representations for handwritten word spotting in historical documents. Int. J. Pattern Recognit. Artif. Intell. 2012, 26, 1263002. [Google Scholar] [CrossRef] [Green Version]
  26. Retsinas, G.; Louloudis, G.; Stamatopoulos, N.; Gatos, B. Efficient Learning-Free Keyword Spotting. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1587–1600. [Google Scholar] [CrossRef] [PubMed]
  27. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  28. Zagoris, K.; Pratikakis, I.; Gatos, B. Segmentation-Based Historical Handwritten Word Spotting Using Document-Specific Local Features. In Proceedings of the 14th International Conference on Frontiers in Handwriting Recognition (ICFHR), Hersonissos, Greece, 1–4 September 2014; pp. 9–14. [Google Scholar] [CrossRef]
  29. Lindeberg, T. Feature Detection with Automatic Scale Selection. Int. J. Comput. Vis. 1998, 30, 79–116. [Google Scholar] [CrossRef]
  30. Lindeberg, T. Scale-Space Theory in Computer Vision; Springer: New York, NY, USA, 2013; Volume 256. [Google Scholar]
  31. Aldavert, D.; Rusinol, M.; Toledo, R.; Llados, J. Integrating Visual and Textual Cues for Query-by-String Word Spotting. In Proceedings of the 12th International Conference on Document Analysis and Recognition (ICDAR), Washington, DC, USA, 25–28 August 2013; pp. 511–515. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The proposed word spotting client-server pipeline. Blue and green arrows denote the indexing and the query processing sequence, respectively.
Figure 1. The proposed word spotting client-server pipeline. Blue and green arrows denote the indexing and the query processing sequence, respectively.
Jimaging 07 00278 g001
Figure 2. The proposed word spotting architecture. The OFFLINE procedure is only performed once to create the appropriate data structures. The ONLINE process is visible to the user and executes when locating the word.
Figure 2. The proposed word spotting architecture. The OFFLINE procedure is only performed once to create the appropriate data structures. The ONLINE process is visible to the user and executes when locating the word.
Jimaging 07 00278 g002
Figure 3. The proposed word spotting indexing and matching pipeline for the Greek handwritten word of ’Aristotle’. (a) Query image; (b) filtered gradient image I x ; (c) filtered gradient image I y ; (d) quantization of the gradient orientation; (e) keypoints; (f) query keypoint (yellow) and the ellipse area enclosed keypoints (blue) that should be matched during the matching process.
Figure 3. The proposed word spotting indexing and matching pipeline for the Greek handwritten word of ’Aristotle’. (a) Query image; (b) filtered gradient image I x ; (c) filtered gradient image I y ; (d) quantization of the gradient orientation; (e) keypoints; (f) query keypoint (yellow) and the ellipse area enclosed keypoints (blue) that should be matched during the matching process.
Jimaging 07 00278 g003
Figure 4. Presentation of the Dolf’s location for various handwritten words: (a) ‘adjuncto’; (b) ‘adjunctur’; (c,d) ‘andere’; (e,f) ‘however’.
Figure 4. Presentation of the Dolf’s location for various handwritten words: (a) ‘adjuncto’; (b) ‘adjunctur’; (c,d) ‘andere’; (e,f) ‘however’.
Jimaging 07 00278 g004
Figure 5. Feature extraction: (a) scale invariant window size definition; (b) features calculation; (c) descriptor structure.
Figure 5. Feature extraction: (a) scale invariant window size definition; (b) features calculation; (c) descriptor structure.
Jimaging 07 00278 g005
Figure 6. The overall architecture for the quantization of the DoLF descriptors using multiple Bag of Visual Words.
Figure 6. The overall architecture for the quantization of the DoLF descriptors using multiple Bag of Visual Words.
Jimaging 07 00278 g006
Figure 7. Memory and storage structures and their relationships.
Figure 7. Memory and storage structures and their relationships.
Jimaging 07 00278 g007
Figure 8. (a) The visual representation for the descriptor quantization using four different Bag of Visual Words; (b) the descriptor histogram, before and after applying quantization.
Figure 8. (a) The visual representation for the descriptor quantization using four different Bag of Visual Words; (b) the descriptor histogram, before and after applying quantization.
Jimaging 07 00278 g008
Figure 9. Representative document images from (a) English dataset; (b) German dataset; (c) Finnish dataset; (d) Greek dataset.
Figure 9. Representative document images from (a) English dataset; (b) German dataset; (c) Finnish dataset; (d) Greek dataset.
Jimaging 07 00278 g009
Table 1. Experimental results for four handwritten segmentation-free datasets.
Table 1. Experimental results for four handwritten segmentation-free datasets.
MethodEnglishGermanFinnishGreek
P@5MAPP@5MAPP@5MAPP@5MAP
Original [10]0.350.220.590.420.580.430.390.31
Current0.350.220.570.380.560.390.400.32
Table 2. Time, memory and storage requirements.
Table 2. Time, memory and storage requirements.
MethodRetrieval Time per Query (sec)Memory Requirement per Document (KB)Storage Requirement per Document (KB)
Original [10]15.841980019800
Current0.66491676
Table 3. Comparative evaluation results for different size datasets.
Table 3. Comparative evaluation results for different size datasets.
Database (Documents)Retrieval Time per Query (sec)Overall Memory Requirement (MB)Overall Storage Requirement (MB)
500.612.169
50000.892132693
50,0001.14485843
Table 4. Experimental results for Track I.B.—training and segmentation-free (ICDAR15 competition).
Table 4. Experimental results for Track I.B.—training and segmentation-free (ICDAR15 competition).
MethodP@5MAP
PRG Group0.3760.293
CVC Group [31]0.1500.116
Competition Baseline [5]0.1500.116
Original [10]0.3870.326
Proposed Method0.35210.282
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zagoris, K.; Amanatiadis, A.; Pratikakis, I. Word Spotting as a Service: An Unsupervised and Segmentation-Free Framework for Handwritten Documents. J. Imaging 2021, 7, 278. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7120278

AMA Style

Zagoris K, Amanatiadis A, Pratikakis I. Word Spotting as a Service: An Unsupervised and Segmentation-Free Framework for Handwritten Documents. Journal of Imaging. 2021; 7(12):278. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7120278

Chicago/Turabian Style

Zagoris, Konstantinos, Angelos Amanatiadis, and Ioannis Pratikakis. 2021. "Word Spotting as a Service: An Unsupervised and Segmentation-Free Framework for Handwritten Documents" Journal of Imaging 7, no. 12: 278. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7120278

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop