Data Quality Theory and Applications

A special issue of Algorithms (ISSN 1999-4893). This special issue belongs to the section "Databases and Data Structures".

Deadline for manuscript submissions: closed (30 March 2020) | Viewed by 5121

Special Issue Editor


E-Mail Website
Guest Editor
Department of Telecommunications and Information Processing, Ghent University, Ghent, Belgium
Interests: data quality; databases; uncertainty modelling

Special Issue Information

Dear Colleagues,

Data quality is an ever-emerging topic in computer science that deals with measurement and improvement of the quality of data stored in databases. As techniques on data quality are being developed in the fields of databases, machine learning, and statistics, data quality is a part of computer science that certainly has a wide application range. With the increasing volumes of databases, the need for techniques that deal with data quality more efficiently is bigger than ever. Moreover, new database models are challenging us to find adapted techniques for data quality handling.

I invite you to submit high-quality papers to this Special Issue on “Data Quality” with subjects covering the entire range from theory to application. The following is a (non-exhaustive) list of topics of interests:

  • Discovery of (Conditional) Functional Dependencies,
  • Discovery of Denial Constraints,
  • Edit rules for categorical and continuous data,
  • Efficient solutions to the Error localization problem,
  • Measurement procedures for data quality,
  • Record linkage and data fusion,
  • Imputation techniques,
  • Uncertainty models related to quality of data,
  • Techniques related to NoSQL databases, and
  • Incremental approaches for the whole measurement process.

Dr. Antoon Bronselaer
Guest Editor

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Algorithms is an international peer-reviewed open access monthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • data quality measurement
  • imputation
  • edit rules
  • dependencies
  • efficient approaches
  • NoSQL databases

Published Papers (1 paper)

Order results
Result details
Select all
Export citation of selected articles as:

Research

18 pages, 3042 KiB  
Article
How to Inspect and Measure Data Quality about Scientific Publications: Use Case of Wikipedia and CRIS Databases
by Otmane Azeroual and Włodzimierz Lewoniewski
Algorithms 2020, 13(5), 107; https://0-doi-org.brum.beds.ac.uk/10.3390/a13050107 - 26 Apr 2020
Cited by 5 | Viewed by 4871
Abstract
The quality assurance of publication data in collaborative knowledge bases and in current research information systems (CRIS) becomes more and more relevant by the use of freely available spatial information in different application scenarios. When integrating this data into CRIS, it is necessary [...] Read more.
The quality assurance of publication data in collaborative knowledge bases and in current research information systems (CRIS) becomes more and more relevant by the use of freely available spatial information in different application scenarios. When integrating this data into CRIS, it is necessary to be able to recognize and assess their quality. Only then is it possible to compile a result from the available data that fulfills its purpose for the user, namely to deliver reliable data and information. This paper discussed the quality problems of source metadata in Wikipedia and CRIS. Based on real data from over 40 million Wikipedia articles in various languages, we performed preliminary quality analysis of the metadata of scientific publications using a data quality tool. So far, no data quality measurements have been programmed with Python to assess the quality of metadata from scientific publications in Wikipedia and CRIS. With this in mind, we programmed the methods and algorithms as code, but presented it in the form of pseudocode in this paper to measure the quality related to objective data quality dimensions such as completeness, correctness, consistency, and timeliness. This was prepared as a macro service so that the users can use the measurement results with the program code to make a statement about their scientific publications metadata so that the management can rely on high-quality data when making decisions. Full article
(This article belongs to the Special Issue Data Quality Theory and Applications)
Show Figures

Figure 1

Back to TopTop