Document Recognition in the Cultural Heritage: Methods and Applications

A special issue of Journal of Imaging (ISSN 2313-433X). This special issue belongs to the section "Document Analysis and Processing".

Deadline for manuscript submissions: closed (31 December 2022) | Viewed by 16473

Special Issue Editors


E-Mail Website
Guest Editor
Division of Visual Information and Interaction, Department of Information Technology, Uppsala University, Uppsala, Sweden
Interests: computer vision and image processing, especially for applications in microscopy, aerial photography, face and object recognition and hand written text recognition
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
1. Department InGeo, University "G. d'Annunzio" Chieti-Pescara, 65127 Pescara, Italy
2. Research Laboratory "Hugo Gernsback", Telematic University Leonardo da Vinci, 66010 Torrevecchia Teatina, Italy
Interests: artificial intelligence; pattern recognition; data analysis; human-computer interaction
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

The field of document recognition in the cultural heritage is rapidly progressing, wherein vast amounts of ancient manuscripts in libraries and other institutions across the world are increasingly undergoing digitization and transcription. However, the automatic recognition of patterns in ancient manuscripts to render them readable, searchable and understandable remains a challenging task. This can be due to degradations including ink bleed-through, ink corrosion, stains on paper or parchment, difficulty in the character discrimination, elements different from the text, such as images, etc. that limit the effectiveness of existing techniques. In recent times, machine learning, and deep learning-based methods in particular, have achieved significant performance improvements in document recognition. The aim of this Special Issue is to present recent advances in methods and applications for document recognition in the cultural heritage, attracting research papers from a wide array of disciplines, including machine learning, pattern recognition, image analysis, and digital humanities. The focus is to highlight advances in the research topics representing document recognition from a broad perspective, and to promote new algorithms and methodologies for better reflection of the state-of-the-art in this field.

We are interested in original research that addresses the various issues in document recognition. Topics include, but are not limited to, the following:

Handwritten document analysis;
 
Writer recognition;

Text line extraction or segmentation;

Document image binarization;

Background noise removal;

Word spotting;

Word recognition;

Automatic recognition and transcription of manuscripts;

Active learning for handwritten text recognition;

Scribe identification;

Dating of historical manuscripts;

Document layout analysis;  

Human–document interaction;

Feature extraction and representation;

Digital humanities applications;

Image processing, classification, and retrieval.

Prof. Anders Hast
Dr. Alessia Amelio
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Journal of Imaging is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1800 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Published Papers (5 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

24 pages, 6904 KiB  
Article
A Quality, Size and Time Assessment of the Binarization of Documents Photographed by Smartphones
by Rodrigo Bernardino, Rafael Dueire Lins and Ricardo da Silva Barboza
J. Imaging 2023, 9(2), 41; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging9020041 - 13 Feb 2023
Cited by 2 | Viewed by 1424
Abstract
Smartphones with an in-built camera are omnipresent today in the life of over eighty percent of the world’s population. They are very often used to photograph documents. Document binarization is a key process in many document processing platforms. This paper assesses the quality, [...] Read more.
Smartphones with an in-built camera are omnipresent today in the life of over eighty percent of the world’s population. They are very often used to photograph documents. Document binarization is a key process in many document processing platforms. This paper assesses the quality, file size and time performance of sixty-eight binarization algorithms using five different versions of the input images. The evaluation dataset is composed of deskjet, laser and offset printed documents, photographed using six widely-used mobile devices with the strobe flash off and on, under two different angles and four shots with small variations in the position. Besides that, this paper also pinpoints the algorithms per device that may provide the best visual quality-time, document transcription accuracy-time, and size-time trade-offs. Furthermore, an indication is also given on the “overall winner” that would be the algorithm of choice if one has to use one algorithm for a smartphone-embedded application. Full article
Show Figures

Figure 1

22 pages, 5554 KiB  
Article
Digital Hebrew Paleography: Script Types and Modes
by Ahmad Droby, Irina Rabaev, Daria Vasyutinsky Shapira, Berat Kurar Barakat and Jihad El-Sana
J. Imaging 2022, 8(5), 143; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging8050143 - 21 May 2022
Cited by 3 | Viewed by 3355
Abstract
Paleography is the study of ancient and medieval handwriting. It is essential for understanding, authenticating, and dating historical texts. Across many archives and libraries, many handwritten manuscripts are yet to be classified. Human experts can process a limited number of manuscripts; therefore, there [...] Read more.
Paleography is the study of ancient and medieval handwriting. It is essential for understanding, authenticating, and dating historical texts. Across many archives and libraries, many handwritten manuscripts are yet to be classified. Human experts can process a limited number of manuscripts; therefore, there is a need for an automatic tool for script type classification. In this study, we utilize a deep-learning methodology to classify medieval Hebrew manuscripts into 14 classes based on their script style and mode. Hebrew paleography recognizes six regional styles and three graphical modes of scripts. We experiment with several input image representations and network architectures to determine the appropriate ones and explore several approaches for script classification. We obtained the highest accuracy using hierarchical classification approach. At the first level, the regional style of the script is classified. Then, the patch is passed to the corresponding model at the second level to determine the graphical mode. In addition, we explore the use of soft labels to define a value we call squareness value that indicates the squareness/cursiveness of the script. We show how the graphical mode labels can be redefined using the squareness value. This redefinition increases the classification accuracy significantly. Finally, we show that the automatic classification is on-par with a human expert paleographer. Full article
Show Figures

Figure 1

13 pages, 6982 KiB  
Article
Word Spotting as a Service: An Unsupervised and Segmentation-Free Framework for Handwritten Documents
by Konstantinos Zagoris, Angelos Amanatiadis and Ioannis Pratikakis
J. Imaging 2021, 7(12), 278; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7120278 - 17 Dec 2021
Cited by 2 | Viewed by 2573
Abstract
Word spotting strategies employed in historical handwritten documents face many challenges due to variation in the writing style and intense degradation. In this paper, a new method that permits efficient and effective word spotting in handwritten documents is presented that relies upon document-oriented [...] Read more.
Word spotting strategies employed in historical handwritten documents face many challenges due to variation in the writing style and intense degradation. In this paper, a new method that permits efficient and effective word spotting in handwritten documents is presented that relies upon document-oriented local features that take into account information around representative keypoints and a matching process that incorporates a spatial context in a local proximity search without using any training data. The method relies on a document-oriented keypoint and feature extraction, along with a fast feature matching method. This enables the corresponding methodological pipeline to be both effectively and efficiently employed in the cloud so that word spotting can be realised as a service in modern mobile devices. The effectiveness and efficiency of the proposed method in terms of its matching accuracy, along with its fast retrieval time, respectively, are shown after a consistent evaluation of several historical handwritten datasets. Full article
Show Figures

Figure 1

23 pages, 9083 KiB  
Article
Attention-Based Fully Gated CNN-BGRU for Russian Handwritten Text
by Abdelrahman Abdallah, Mohamed Hamada and Daniyar Nurseitov
J. Imaging 2020, 6(12), 141; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6120141 - 18 Dec 2020
Cited by 26 | Viewed by 4098
Abstract
This article considers the task of handwritten text recognition using attention-based encoder–decoder networks trained in the Kazakh and Russian languages. We have developed a novel deep neural network model based on a fully gated CNN, supported by multiple bidirectional gated recurrent unit (BGRU) [...] Read more.
This article considers the task of handwritten text recognition using attention-based encoder–decoder networks trained in the Kazakh and Russian languages. We have developed a novel deep neural network model based on a fully gated CNN, supported by multiple bidirectional gated recurrent unit (BGRU) and attention mechanisms to manipulate sophisticated features that achieve 0.045 Character Error Rate (CER), 0.192 Word Error Rate (WER), and 0.253 Sequence Error Rate (SER) for the first test dataset and 0.064 CER, 0.24 WER and 0.361 SER for the second test dataset. Our proposed model is the first work to handle handwriting recognition models in Kazakh and Russian languages. Our results confirm the importance of our proposed Attention-Gated-CNN-BGRU approach for training handwriting text recognition and indicate that it can lead to statistically significant improvements (p-value < 0.05) in the sensitivity (recall) over the tests dataset. The proposed method’s performance was evaluated using handwritten text databases of three languages: English, Russian, and Kazakh. It demonstrates better results on the Handwritten Kazakh and Russian (HKR) dataset than the other well-known models. Full article
Show Figures

Figure 1

17 pages, 7136 KiB  
Article
CleanPage: Fast and Clean Document and Whiteboard Capture
by Jane Courtney
J. Imaging 2020, 6(10), 102; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6100102 - 01 Oct 2020
Cited by 2 | Viewed by 3704
Abstract
The move from paper to online is not only necessary for remote working, it is also significantly more sustainable. This trend has seen a rising need for the high-quality digitization of content from pages and whiteboards to sharable online material. However, capturing this [...] Read more.
The move from paper to online is not only necessary for remote working, it is also significantly more sustainable. This trend has seen a rising need for the high-quality digitization of content from pages and whiteboards to sharable online material. However, capturing this information is not always easy nor are the results always satisfactory. Available scanning apps vary in their usability and do not always produce clean results, retaining surface imperfections from the page or whiteboard in their output images. CleanPage, a novel smartphone-based document and whiteboard scanning system, is presented. CleanPage requires one button-tap to capture, identify, crop, and clean an image of a page or whiteboard. Unlike equivalent systems, no user intervention is required during processing, and the result is a high-contrast, low-noise image with a clean homogenous background. Results are presented for a selection of scenarios showing the versatility of the design. CleanPage is compared with two market leader scanning apps using two testing approaches: real paper scans and ground-truth comparisons. These comparisons are achieved by a new testing methodology that allows scans to be compared to unscanned counterparts by using synthesized images. Real paper scans are tested using image quality measures. An evaluation of standard image quality assessments is included in this work, and a novel quality measure for scanned images is proposed and validated. The user experience for each scanning app is assessed, showing CleanPage to be fast and easier to use. Full article
Show Figures

Figure 1

Back to TopTop