Next Article in Journal
Quantifying Reliable Walking Activity with a Wearable Device in Aged Residential Care: How Many Days Are Enough?
Next Article in Special Issue
A Novel BN Learning Algorithm Based on Block Learning Strategy
Previous Article in Journal
Cooperative Localization and Time Synchronization Based on M-VMP Method
Previous Article in Special Issue
PORF-DDPG: Learning Personalized Autonomous Driving Behavior with Progressively Optimized Reward Function
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Digital Forensics of Scanned QR Code Images for Printer Source Identification Using Bottleneck Residual Block

1
School of Electronic Information, Wuhan University, Wuhan 430072, China
2
School of Mathematics and Physics, Wuhan Institute of Technology, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Submission received: 19 August 2020 / Revised: 1 November 2020 / Accepted: 3 November 2020 / Published: 5 November 2020
(This article belongs to the Special Issue Artificial Intelligence and Sensors)

Abstract

:
With the rapid development of information technology and the widespread use of the Internet, QR codes are widely used in all walks of life and have a profound impact on people’s work and life. However, the QR code itself is likely to be printed and forged, which will cause serious economic losses and criminal offenses. Therefore, it is of great significance to identify the printer source of QR code. A method of printer source identification for scanned QR Code image blocks based on convolutional neural network (PSINet) is proposed, which innovatively introduces a bottleneck residual block (BRB). We give a detailed theoretical discussion and experimental analysis of PSINet in terms of network input, the first convolution layer design based on residual structure, and the overall architecture of the proposed convolution neural network (CNN). Experimental results show that the proposed PSINet in this paper can obtain extremely excellent printer source identification performance, the accuracy of printer source identification of QR code on eight printers can reach 99.82%, which is not only better than LeNet and AlexNet widely used in the field of digital image forensics, but also exceeds state-of-the-art deep learning methods in the field of printer source identification.

1. Introduction

With the popularity of smart phones and 4G networks, QR codes are widely used in logistics and transportation, financial institutions, advertising and marketing, security, and anti-counterfeiting [1,2,3,4,5]. It records data symbol information according to certain rules and distributes black and white images on a plane, and it has the advantages of low printing cost, easy production, and durability.
At present, printers are office equipment commonly used by institutions, enterprises and individuals, and digital printing is also widely used in people’s daily lives. As an important part of digital content, QR codes are usually printed on newspapers, magazines, advertisements, books, packaging, personal business cards, and other carriers. Users can scan the QR codes with a mobile phone camera to achieve the purpose of information acquisition, website redirection, advertisement push, anti-counterfeiting traceability, mobile payment, information technology, etc. [6,7,8,9] that uses the QR codes as the entrance has attracted more and more mobile phone users and businesses, and has become the current development trend. However, the QR code has the risk of being forged during the printing process, which is likely to cause economic disputes and criminal cases, resulting in huge economic losses and serious crimes. Therefore, it is necessary to identify the source of the printer. In the printing process of digital content, drifting motor and precision gear have created pattern information from the printer as the inherent information of device, so that each printer has a characteristics signature based on the corresponding fluctuation of the developed toner on the printed page, and each printer has its own unique texture features [10]. According to the above principles, the source forensic of printed documents can be carried out.
The currently widely used printed source identification techniques are pattern based shallow learning [11,12,13]. Tsai [14] et al. used Grey-Level Co-occurrence Matrix (GLCM) and discrete wavelet variation methods to extract the characteristics of scanned Chinese characters to find the printer source; Choi [15] et al. used noisy texture analysis and support vector machine (SVM) to detect the printer source of unknown images; Zhou [16] et al. proposed a text-independent printed identification method based on texture synthesis, Fast Fourier Transform (FFT) and GLCM were used for feature extraction, and SVM was used for classification. Similar methods have been developed by Mikkilineni [17], Gebhardt J [18], and Fang T [19]. However, the above methods require the participation of experts in the process of feature extraction, feature selection and classifier design. In addition, the entire process has to be repeated several times to obtain generalized results, which will undoubtedly consume time.
In recent years, deep learning has been widely used, and CNN has also made breakthroughs in image classification, image recognition and other issues [20,21,22,23,24,25,26]. As a new field of machine learning, deep learning uses the structure of a deep neural network with multiple convolution layers, which automatically learns the features of large-scale input data, and forms the bottom features into more abstract high-level features to represent the attribute categories or features of training data, and then represents the attribute categories or features of training data, so as to achieve higher recognition accuracy. Different from the traditional pattern based shallow learning, the structure of deep learning model contains more layers, and the number of hidden layers is usually above 5. Furthermore, deep learning can extract features layer by layer and transform data samples in the original feature space into a new feature space to represent the initial data, making classification or prediction problems easier to implement. Comparing with manually designed feature extraction methods, the data features obtained by deep learning models can better represent the rich internal information of big data.
In the studies of printer source identification based on deep learning, scientists and researchers often directly introduce CNNs which are commonly used in the field of computer vision to classify printers. However, there are essential differences between the printer source identification and classification in computer vision. In the field of computer vision, the morphological differences between categories are larger, which can be distinguished by human eyes; but in the field of printer source identification, the morphological differences between categories are extremely small, and the differences between categories exist in a weak form [27]. At present, the commonly used CNNs in the field of printer source identification of digital documents are generally designed based on LeNet [28] or AlexNet [29], because compared with VGGNet [30], GoogLeNet [31], ResNet [32], and other CNNs with more layers, they have lower structural complexity, have better tentative conditions for forensic problems with fewer data sets, and are more suitable for small-size image input.
In addition to the above general classical CNNs, researchers have also proposed a series of printer source identification methods of digital documents based on deep learning. Anselmo Ferreira [10] designed a deep learning method for laser printer attributes based on the analysis of small patches representing text characteristics. The proposed method uses multiple CNNs to learn multiple representations of the same character, and then the features are fused to improve the identification accuracy. However, this method will undoubtedly consume a lot of time and energy in the process of data collection and processing. Do-Guk Kim [33] proposed a cascaded learning method based on generative adversarial networks (GANs) and convolutional neural networks (CNNs) to identify the source color laser printer of photographed color documents. This method uses Simulated + Unsupervised (S + U) learning based on a GAN-alike framework for the decomposition of halftone color channels and proposed a CNN for source color laser printer identification, but it cannot be used for printer source identification of scanned images, because there are differences in the intensity of the illumination and clarity between the photographing environment and the scanning environment. Min-Jen Tsai [34] designed a CNN with seven layers to identify the printer source of text and images collected under microscopes and scanners, so as to achieve the same effect as pattern-based shallow learning, and effectively avoided the shortcomings of pattern based shallow learning method of manually designing feature extractor. However, there was no detailed analysis and discussion on the design of CNN structure in the field of printer source identification, such as the data input size of the network, the selection of the kernel size in the convolutional layer, etc.
The above forensic methods are mainly designed for grayscale text documents or color-captured image, but there are few studies on the source printer identification of widely used black and white QR codes. In addition, previous work rarely introduces breakthrough structures in deep learning into the field of printer source forensics, and achieves excellent identification results while ensuring low structural complexity of CNNs. Therefore, it is a work of great research value and practical significance to apply the widely and far-reaching structure of deep learning to the field of printed source identification of QR codes simply and effectively, at the same time, it has also become an urgent requirement for the legality of identification evidence in criminal and civil case in the judicial system.
In this study, a printer source identification network (PSINet) based on BRB, which is customized for the black and white image of the QR code, is proposed, and exploratory experiments have been done in terms of data input size, convolution kernel size, and CNN architecture; its purpose is to obtain the highest and most stable identification effect. Note that we do not limit the specific content of the QR code, and we assume that one of all the printers we use is a certified official printer and that the other printers may come from the companies different from the official printer, or they may come from the same company where the certified official printer is located, but the specific models are different.
The structure of this paper is as follows. Section 2 introduces the materials and method. Section 3 is the experimental results and analysis. Section 4 summarizes the paper.

2. Materials and Methods

2.1. The Printer Source Identification of QR Codes Using CNNs

The diagram shown in Figure 1 illustrates the printer source identification process of QR codes using CNNs, which can be divided into five stages such as printing process, scanning process, QR code extraction, dataset production, and training and testing of CNNs.
(1)
Printing process. Eight printers were used to print the same batch of QR code documents.
(2)
Scanning process. The printed QR code paper documents were scanned by the same scanner to generate digital images.
(3)
QR code extraction. It can be seen from Figure 1 that there are many QR codes in the scanned digital images and a single QR code needs to be extracted. In this study, OpenCV, a cross-platform computer vision and machine learning software library based on Berkeley Software Distribution (BSD) license, is used; the main extraction steps include graying, binarization, morphological opening operation, image reverse operation, finding contours, and setting the threshold range to extraction QR code contour area. These operations are well known in the field of image processing and are described in detail in Section 2.2.
(4)
Dataset production. Adjust the size of all the extracted QR codes to 256 × 256, and then make the calibration dataset, validation set, and prediction set in the ration of 3:1:1. Then, block all the QR code images in the data set to a size of M × N.
(5)
Training and testing of CNNs. Design PSINet, then the QR code blocks are input for training and testing.
Figure 1. The overall flow chart of printer source identification process of QR codes using convolution neural networks (CNNs).
Figure 1. The overall flow chart of printer source identification process of QR codes using convolution neural networks (CNNs).
Sensors 20 06305 g001

2.2. QR Code Extraction

(1)
Grayscale.
The scanned QR code images have three channels: R, G, and B, so it needs to be grayed. The human eye has the highest sensitivity to green and lowest sensitivity to blue. Therefore, the weighted average method is used to weight the RGB components of QR code image with a certain weight to obtain a reasonable gray image. The Open Source Computer Vision Library (OpenCV) is used, and the grayscale function is as follows,
G r a y ( i , j ) = 0.299 × R ( i , j ) + 0.587 × G ( i , j ) + 0.114 × B ( i , j )
where R, G, and B represent the brightness of red, green, and blue channels, respectively.
(2)
Otsu’s segmentation.
Otsu method can adaptively segment the QR code and background according to the gray characteristics of the image to maximize their inter-class variance. The threshold in the experiment is set to 127 to obtain a good segmentation effect.
(3)
Morphological opening operation.
The process of first corrosion and then expansion is adopted, which can remove isolated connected domains and fill the inside of the QR Code. The OpenCV library is used and the shape of the kernel of the structure element is set to rectangle, and the size of the kernel is 30 × 30.
(4)
Image reversal.
The formula of image reversal is as follows,
g ( x , y ) = 255 f ( x , y )
where f(x, y) represents the gray value of (x, y) and g(x, y) represents the gray value of (x, y); after image reversal, the QR code is changed into white and the background area is changed into black.
(5)
QR code extraction.
The OpenCV library is used to find the contour of the QR codes, and the QR code is extracted by setting the threshold of the area and height of the contour. The formula is as follows,
R e c t a r e a > 20 , 000 , 100 < R e c t h e i g h t < 500
where R e c t a r e a represents the area of the contour of QR code and R e c t h e i g h t represents the height of the contour of QR code.

2.3. PSINet: A Printer Source Identification Network for QR Codes Based on Bottleneck Residual Block

CNN is mainly composed of convolution layers, pooling layers, ReLU layers, and fully connected layers. When designing a CNN, multiple convolution layers are often used to extract deep features of images, and batch normalization layer is generally fused with the convolutional layer, and its main function is to alleviate the vanishing gradient and gradient explosion, and to accelerate the training speed. Pooling layer reduces the dimensionality of the feature map while maintaining important information, it is sandwiched between continuous convolutional layers to compress the amount of data and parameters, thereby preventing overfitting. The function of ReLU layer is to perform nonlinear mapping of the results of convolutional layer to increase the nonlinear expression ability of CNN. The fully-connected layer serves as the classifier in the entire CNN and represents the learned distribution features into the sample label space.
In 2015, He Kaiming, a researcher at Microsoft Research Institute, proposed a residual network with 152 layers and won the championship in the image recognition project of the ImageNet Large Scale Visual Recognition Challenge (ILSVRC). The residual network proposed the bottleneck residual block (BRB) structure, as shown in Figure 2. BRB is composed of 1 × 1 convolution layer, ReLU layer, 3 × 3 convolution layer, ReLU layer, 1 × 1 convolution layer, ReLU layer. The number of feature maps of the three convolutional layers in BRB is 64, 64, 256, respectively. By introducing identify mapping into CNN, the problem of gradient diffusion in deep CNN can be effectively solved, and the parameters are reduced while the depth of CNN is increased, which further improves the performance of CNN.
In this study, we were inspired by a lot of research work on the application of residual networks in the field of computer vision. At the same time, considering that BRB has a strong feature representation capability, we first introduced the BRB into the field of printer source identification of digital content, and designed a CNN named PSINet to identify the printer source of QR codes. Figure 3 is the overall framework of PSINet, which mainly follows the structure of convolution layer, pooling layer, BRBs, pooling layer, and fully connected layer. Each convolutional layer is followed by a batch normalization (BN) layer and a ReLU layer. The number of BRBs and the final structure of PSINet will be determined through the experiment. It is worth mentioning that the global average pooling layer is used before the fully connected layer, which can effectively prevent overfitting. The size of the convolutional kernel of the global average pooling layer is determined according to the size of feature map output by the previous convolution layer, and stride is 1; the sizes of the other pooling layers are 5 × 5, and the strides are 2; as the number of printers used is 8, the output categories of the fully connected layer is set to 8; finally, the softmax layer maps the multiple scalars output by the fully connected layer into probability distributions. The formula is as follows,
f ( z i ) = e z i i = 1 n e z i
where Z i represents input, n represents categories, and f ( z i ) represents output.
The hyperparameter settings of PSINet are as follows; training epoch is set to 30 and the number of iterations per epoch is 576; therefore, a total of 7280 iterations are performed, which can ensure that training curve fully converges; the solver type is set to stochastic gradient descent (SGD); the base learning rate is set to 0.01, the policy is set to exponential decay, and gamma is set to 0.99; momentum is set to 0.9 and weight decay is set to 0.0001.

3. Experimental Results and Analysis

3.1. Development Environment and Tools

We used a personal computer (Intel® CoreTM i7-7700HQ CPU @2.80 GHz (Intel Corporation, Santa Clara, CA, USA), GeForce GTX 1070 (NVIDIA Ltd., Santa Clara, CA, USA) and 16 GB RAM with Windows10 operating system) to develop the proposed PSINet; the software development environment of QR code extraction and segmentation program is Visual Studio 2015 (Microsoft Ltd., Chicago, IL, USA) and OpenCV-3.4.0 (Intel Ltd., Santa Clara, CA, USA); the deep learning framework used is Caffe (Berkeley Artificial Intelligence Research, San Francisco, CA, USA) [35] and Nvidia Digits-6.1.1 (NVIDIA Ltd., Santa Clara, CA, USA) [36].

3.2. Dataset Production

The extracted single QR code sample is shown in Figure 4, then the QR code data set are made. In this study, 8 printers are used; the specific brands and models are shown in Table 1. The number of QR codes is 240, after printing by 8 printers, the total number becomes 1920, and they are scanned into digital images by the same scanner with a resolution of 300 dpi. Then, the sizes of all QR codes are unified to 256 × 256, and the calibration set, validation set, and prediction set are made according to the ratio of 3:1:1, and each QR code image is divided into blocks with size of M × N. Figure 5 shows the block images of some QR codes printed by eight printers.

3.3. Determination of the Specific Structure of PSINet

First of all, when CNN is applied to the printer source identification of digit content, the image input size will have an impact on the performance of CNN to extract the inherent micro-features in the printing process. In this study, the printed and scanned QR code images are divided into blocks with sizes of 48 × 48, 56 × 56, and 64 × 64, respectively, and the block size that is most suitable for the QR code printer source identification is determined through experimental comparison. Then, the size of the convolution kernel also has an important influence on the accuracy of CNN. It can be seen from the application of residual network in the field of computer vision that when the BRB architecture is stable, setting the convolution kernel size of the first convolution layer to 7 × 7 has a better effect. In the field of printer source identification, the input image size is usually small, and the differences between categories are relatively subtle. Correspondingly, the impact of the size of the convolution kernel in the first convolutional layer on image feature extraction ability also requires experimentation. In this study, convolutional kernels with size of 3 × 3, 5 × 5, 7 × 7 were selected respectively. Finally, the number of CNN layers also has a very important impact on the identification accuracy. PSINet is mainly designed based on BRB, in this paper, we set the number of BRBs to 1, 2, and 3, respectively, at the same time, and the layers of PSINet are 7, 11, and 15, respectively. The identification accuracy of PSINet under different image input sizes, different convolution kernel sizes of first convolution layer, and different convolution layers is shown in Table 2.
It is well known that a single dataset is not representative, so in order to determine the size of QR code image block, the convolutional kernel size of the first convolution layer and the number of layers of PSINet, we use two indicators of average accuracy and standard deviation to determine the final parameter size. The specific process is as follows.
When the image size is 48 × 48, 56 × 56, 64 × 64, the average accuracy is (99.03 ± 1.26)%, (99.44 ± 0.66)%, and (99.56 ± 0.47)%, respectively. It can be seen that when the image block is 64 × 64, the average accuracy is the highest, and the standard deviation is the smallest, indicating that the proposed PSINet has the best identification performance and stability. Therefore, the QR code image block with size of 64 × 64 is selected as the input.
Then, the size of the convolution kernel of the first convolution layer is determined through experiments. When the size of the convolutional kernel is 3 × 3, 5 × 5, 7 × 7, respectively, the average accuracy is (99.37 ± 0.86)%, (99.33 ± 0.61)%, and (99.33 ± 0.99)%, respectively. It can be seen that when the size of the convolution kernel is 3 × 3, the average accuracy is the highest, which is 0.04% higher than the 5 × 5 convolution kernel and 7 × 7 convolution kernel, but the advantage is very weak; on the contrary, the difference in standard deviation is obvious: the standard deviation of the 5 × 5 convolution kernel is the smallest, with a value of 0.61%, which is 0.25% and 0.38% lower than 3 × 3 convolution kernel and 7 × 7 convolution kernel, respectively. The above analysis shows that while the 5 × 5 convolution kernel obtains similar accuracy, its stability can be significantly better that other convolution kernels, so it is selected.
Finally, the number of layers of PSINet is determined. When the number of layers of PSINet is 7, 11, and 15, the average accuracy rates are (99.81 ± 0.14)%, (99.69 ± 0.12)%, and (98.52 ± 0.95)%, respectively; it can be seen that when the number of layers of PSINet is 7 layers, the average accuracy is the highest, which is 0.12% and 0.29% higher than that of 11 layers and 15 layers, respectively; in terms of standard deviation, the standard deviation of PSINet with 11 layers is the smallest, which is 0.12%, but PSINet with 7 layers obtained a very close value of 0.14%.The PSINet with 15 layers performs the worst in terms of average accuracy and standard deviation. In summary, PSINet with 7 layers is selected.
It can also be seen from the experimental results that if we choose Canon iR-ADV C50455051 UFR 2 as the official printer, regardless of whether the suspected printer is from Canon’s other models, such as Canon iR-ADV C7260270 UFR 2, or from other companies’ printers, such as Epson L310 Series and RICOH’s Aficio MP7502, the PSINet we proposed can effectively identify.

3.4. Comparison of PSINet with Other Deep Learning Methods

3.4.1. Accuracy Comparison of Different CNNs

In this section, PSINet is compared with other CNN methods in identification accuracy to verify its effectiveness. It is well known that in the field of printer source identification, the size of input image is relatively small, and it can be seen from Section 3.3 that the shallow convolutional neural network is more suitable for printer source identification. Based on the above considerations, LeNet and AlexNet, which are currently widely used in the field of digital image forensics, are selected for comparison. In addition, there is CNN method specifically proposed for the field of printer source identification, such as Min-Jen Tsai’s [28], which is also used for comparison to verify the effectiveness and superiority of PSINet. It is worth mentioning that Anselmo Ferreira’s [26] and Do-Guk Kim’s [27] are not considered because their data input methods and data collection scenarios are quite different from that of PSINet. The size of the QR code image blocks used in the experiment are 64 × 64, after the blocks are divided according to the above size, there are 18,432 images in the calibration set, 6144 images in the validation set and 6144 images in the prediction set. They are evenly distributed in 8 types of printers, that is, each printer prints the same amount.
First, a confusion matrix is used to measure the accuracy of classification model, the confusion matrix of LeNet, AlexNet, Min-Jen Tsai, and PSINet is shown in Figure 6, where the columns represent predicted labels and the rows represent true labels, the values on the diagonal indicates the number of correct prediction, and non-diagonal elements are the part of prediction errors. The higher the value on the diagonal of the confusion matrix, the more accurate the prediction result. It can be seen that the value on the diagonal of the confusion matrix of PSINet is the largest, followed by AlexNet, then Min-Jen Tsai’s, and finally LeNet. Among them, PSINet has a high number of correct identification on all eight types of printers, while the number of correct identification of AlexNet on class 1 and class 2 printers in slightly lower than that of PSINet, the performance of Min-Jen Tsai’s on class 2 printer is significantly lower than that of PSINet and AlexNet, only 581 were correctly identified, LeNet’s test on class 2 printer is undoubtedly the worst, only 229 were correctly judged.
In addition to using the confusion matrix to represent the number of correct and wrong predictions for each category, the accuracy index is also used. The accuracy distributions of four CNNs are show in Figure 7 and Table 3, which, respectively, reflect the identification accuracy of each type of printer and average accuracy. In terms of the identification accuracy of each type of printer, PSINet performed better than LeNet, AlexNet, and Min-Jen Tsai’s in class 0, 1, 2, 3, 4, and 5 printers. The performance on the 6th printer is the same as LeNet, reaching 100.00%, 1.56% higher that AlexNet and 2.34% higher than AlexNet. The identification accuracy of four CNN methods on the 7th printer all reach 100.00%. In terms of average recognition accuracy, the identification accuracy of LeNet is 90.95%, that of AlexNet is 98.27%, and that of Min-Jen Tsai’s is 95.75%, PSINet achieves 99.82%, which is better than LeNet, AlexNet, and Min-Jen Tsai’s by 8.87%, 1.55%, and 4.07%, respectively. The effectiveness and superiority of PSINet are verified.

3.4.2. Comparison of Inference Time of Four CNN Models

This section mainly compares the inference time used by the four CNNs on the prediction set. Generally speaking, in the case of the same amount of data and the same accuracy, the less the inference time, the better the performance of the CNN model. The prediction set has 6144 QR code blocks with 64 × 64 sizes, the result is shown in Table 4.
It can be seen from Table 4 that the inference time of Min-Jen Tsai’s method is 11 s, which is the least among the four CNN methods; then, there is LeNet and PSINet, their inference time is 13 s; AlexNet has the longest inference time, which is 24 s. Combined with the identification accuracy, it can be seen that although the inference time of PSINet is 2 s longer than that of Min-Jen Tsai’s, the identification accuracy is 4.07% higher; PSINet and LeNet have the same inference time, but the identification accuracy of PSINet is 8.87% higher than that of LeNet; the inference of PSINet is 11 s less than that of AlexNet, and the identification accuracy of PSINet is 1.55% higher than that of AlexNet. Therefore, the performance of PSINet is quite excellent.

3.4.3. Comparison of Computational Cost of Four CNN Models

This section is a comparison between the computational costs of four CNN models, and their model sizes are shown in Table 5. The large the CNN model, the greater the number of parameters, and the higher the computational cost. It can be seen that the model sizes of LeNet, AlexNet, Min-Jen Tsai, and PSINet are 16.2 MB, 77 MB, 0.43 MB, and 5.18 MB, respectively. Comprehensive analysis shows that the model size of Min-Jen Tsai is the smallest, about one-twelfth of that of PSINet, and its average accuracy is 4.07% lower than that of PSINet; the model size of LeNet is 3.12 times that of PSINet, but the average accuracy is 8.87% lower than that of PSINet; the model size of AlexNet is 14.85 times that of PSINet, and the accuracy is 1.62% lower than that of PSINet. Overall, PSINet’s performance in terms of computational cost is still relatively good, but further optimization is worth encouraging.

4. Conclusions

Aiming at the problem that QR codes are easily copied and forged in the printing process, a printer source identification method of QR code image blocks based on convolution neural network (PSINet) is proposed in this paper. PSINet innovatively introduces the bottleneck residual block (BRB) into the field of printer source identification, and the data input, the size of the convolution kernel, and the CNN architecture are compared and analyzed in detail. Eight printers are used for experiments, and the results showed that the PSINet proposed in this paper has better performance than other CNNs.
With current solutions to the printer source identification of scanned QR code images achieving high classification results, we believe that it is time to meet greater challenges. For instance, in the case of digital camera shooting, the pattern noise of the camera itself, as well as the blur and reflection that often appear during the shooting process, will make the effect recognition difficult. The related printer source identification problem remains to be studied in the future. In addition, we also plan to experiment on other forms of documents such as color images and grayscale text. Finally, as the service life of printer increases, its internal mechanical components, such as precision gears and drifting motors, will also wear out, causing changes in the characteristic signature of the printer that may affect the identification results. We plan to investigate the above issue in the future.

Author Contributions

Conceptualization, Z.G. and H.Z.; methodology, Z.G.; software, Z.G.; validation, Z.G.; formal analysis, Z.G. and X.X.; investigation, Z.G. and C.Y.; resources, H.Z.; data curation, Z.G. and H.Z.; writing—original draft preparation, Z.G.; writing—review and editing, C.Y., X.X., Z.Z., and J.J.; visualization, Z.G.; supervision, H.Z.; project administration, H.Z.; funding acquisition X.W. and H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Key Research and Development Program of China (No 2016YFC1401100) and the National Natural Science Foundation of China (No 61771352 and 41474128).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nofal, R.M. Initiating android phone technology using QR codes to make innovative functional clothes. Int. J. Cloth. Sci. Technol. 2020. [Google Scholar] [CrossRef]
  2. Tara Spalla, K. Using QR Codes on Professional Posters to Increase Engagement and Understanding. Nurs. Educ. 2020, 45, 219. [Google Scholar]
  3. Li, H.-S.; Xue, F.; Xia, H.-Y. Style transfer for QR code. Multimed. Tools Appl. 2020, 1–14. [Google Scholar] [CrossRef]
  4. Yuan, T.; Wang, Y.; Xu, K.; Martin, R.R.; Hu, S.-M. Two-Layer QR Codes. IEEE Trans. Image Process. 2019, 28, 4413–4428. [Google Scholar] [CrossRef] [PubMed]
  5. Alan, C.; Gustavo, M.; Leire, M.B. QRcode, hashtag or audio watermark? A case study on second screening. Multimed. Tools. Appl. 2017, 76, 7519–7534. [Google Scholar]
  6. Yu, R.; Duan, X.; Jiao, B. Design and implement of mobile equipment management system based on QRcode. J. Phys. Conf. Ser. 2017, 887. [Google Scholar] [CrossRef] [Green Version]
  7. Li, C.; Shao, P. Precision UAV Parcel Delivery Using QR Code Recognition. J. Aeronaut. Astronaut. Aviat. 2019, 51, 275–289. [Google Scholar]
  8. Chen, R.; Yu, Y.; Xu, X.; Wang, L.; Zhao, H.; Tan, H.-Z. Adaptive Binarization of QR Code Images for Fast Automatic Sorting in Warehouse Systems. Sensors 2019, 19, 5466. [Google Scholar] [CrossRef] [Green Version]
  9. Gammer, N.; Cherrett, T.; Gutteridge, C. Disseminating real-time bus arrival information via QRcode tagged bus stops: A case study of user take-up and reaction in Southampton, UK. J. Transp. Geogr. 2014, 34, 254–261. [Google Scholar] [CrossRef]
  10. Ferreira, A.; Bondi, L.; Baroffio, L.; Bestagini, P.; Huang, J.; Dos Santos, J.A.; Tubaro, S.; Rocha, A.D.R. Data-Driven Feature Characterization Techniques for Laser Printer Attribution. IEEE Trans. Inf. Forensics Secur. 2017, 12, 1860–1872. [Google Scholar] [CrossRef]
  11. Mikkilineni, A.K.; Chiang, P.-J.; Ali, G.N.; Chiu, G.T.C.; Allebach, J.P.; Iii, E.J.D. Printer identification based on graylevel co-occurrence features for security and forensic applications. In Proceedings of the International Conference on Security, Steganography, and Watermarking of Multimedia Contents VII, San Jose, CA, USA, 17–20 January 2005; Volume 5681, pp. 430–440. [Google Scholar]
  12. Mikkilineni, A.; Arslan, O.; Chiang, P. Printer Forensics using SVM Techniques. In Proceedings of the NIP21: 21st International Conference on Digital Printing Technologies—Final Program and Proceedings, Baltimore, MD, USA, 18–23 September 2005; pp. 223–226. [Google Scholar]
  13. Tsai, M.-J.; Yuadi, I. Digital forensics of microscopic images for printed source identification. Multimed. Tools Appl. 2018, 77, 8729–8758. [Google Scholar] [CrossRef]
  14. Tsai, M.-J.; Liu, J.; Yin, J.-S.; Yuadi, I. Digital Forensics of Printed Source Identification for Chinese Characters. Multimed. Tools Appl. 2014, 73, 2129–2155. [Google Scholar] [CrossRef]
  15. Choi, J.-H.; Lee, H.-Y.; Lee, H.-K. Color laser printer forensic based on noisy feature and support vector machine classifier. Multimed. Tools Appl. 2013, 67, 363–382. [Google Scholar] [CrossRef]
  16. Zhou, Q.; Yan, Y.; Fang, T.; Luo, X.; Chen, Q. Text-independent printer identification based on texture synthesis. Multimed. Tools Appl. 2016, 75, 5557–5580. [Google Scholar] [CrossRef]
  17. Mikkilineni, A.; Khanna, N.; Delp, E. Texture Based Attacks on Intrinsic Signature Based Printer Identification. In Proceedings of the Media Forensics and Security II, Part of the IS&T-SPIE Electronic Imaging Symposium, San Jose, CA, USA, 18–20 January 2010. [Google Scholar]
  18. Gebhardt, J.; Goldstein, M.; Shafait, F.; Dengel, A. Document Authentication Using Printing Technique Features and Unsupervised Anomaly Detection. In Proceedings of the 2013 12th International Conference on Document Analysis and Recognition, Washington, DC, USA, 25–28 August 2013; pp. 479–483. [Google Scholar]
  19. Fang, T.; Chen, Q.; Yan, Y. Laser Print Document Identification Based on Gabor Feature and Sparse Representation Classification. Geomat. Inf. Sci. Wuhan Univ. 2016, 41, 1550–1555. [Google Scholar]
  20. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  21. Sun, Y.; Wang, X.; Tang, X. Deep Learning Face Representation by Joint Identification-Verification. Adv. Neural Inf. Process. Syst. 2014, 27, 1988–1996. [Google Scholar]
  22. Theodoridis, S. Neural Networks and Deep Learning. In Machine Learning; Theodoridis, S., Ed.; Department of Informatics and Telecommunications, National and Kapodistrian University of Athens: Athens, Greece, 2015; pp. 875–936. [Google Scholar]
  23. Sun, M.; Deng, Y.; Li, M.; Jiang, H.; Huang, H.; Liao, W.; Liu, Y.; Yang, J.; Li, Y. Extraction and Analysis of Blue Steel Roofs Information Based on CNN Using Gaofen-2 Imageries. Sensors 2020, 20, 4655. [Google Scholar] [CrossRef] [PubMed]
  24. Tsai, C.-Y.; Chen, H.-W. SurfNetv2: An Improved Real-Time SurfNet and Its Applications to Defect Recognition of Calcium Silicate Boards. Sensors 2020, 20, 4356. [Google Scholar] [CrossRef]
  25. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep Learning for Computer Vision: A Brief Review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef] [PubMed]
  26. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martínez, V.; Garcia-Rodriguez, J. A Review on Deep Learning Techniques Applied to Semantic Segmentation. arXiv 2017, arXiv:1704.06857. [Google Scholar]
  27. Alessandro, P. An Overview on Image Forensic. ISRN Signal Process. 2013, 2013, 1–22. [Google Scholar]
  28. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  29. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  30. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
  31. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2014; pp. 1–9. [Google Scholar]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  33. Kim, D.-G.; Hou, J.-U.; Lee, H.-K. Learning deep features for source color laser printer identification based on cascaded learning. Neurocomputing 2019, 365, 219–228. [Google Scholar] [CrossRef] [Green Version]
  34. Tsai, M.-J.; Tao, Y.-H.; Yuadi, I. Deep learning for printed document source identification. Signal. Process. Image Commun. 2019, 70, 184–198. [Google Scholar] [CrossRef]
  35. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional Architecture for Fast Feature Embedding. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 675–678. [Google Scholar]
  36. Luke, Y.; Julie, B. DIGITS: The Deep learning GPU Training System. In Proceedings of the International Conference on Machine Learning (ICML), Lille, France, 6–11 July 2015; pp. 1–4. [Google Scholar]
Figure 2. Bottleneck residual block (BRB).
Figure 2. Bottleneck residual block (BRB).
Sensors 20 06305 g002
Figure 3. The identification process of PSINet.
Figure 3. The identification process of PSINet.
Sensors 20 06305 g003
Figure 4. The extracted QR code image.
Figure 4. The extracted QR code image.
Sensors 20 06305 g004
Figure 5. Image samples for QR codes printed by eight printers. (a) Canon iR-ADV C50455051 UFR2, (b) Canon iR-ADV C7260270 UFR2 (Canon Inc., Tokyo, Japan), (c) Epson L310 Series(Seiko Epson Ltd., Suwa, Japan), (d) Fuji XEROX DocuCentre S2110, (e) RICOH Aficio MP7502(Ricoh Ltd., Tokyo, Japan), (f) RICOH Pro8100s, (g) Samsung K2200 series(Samsung Electronics Ltd., Seoul, Korea), and (h) TOSHIBA e-STUDIO2051C-11606695(Toshiba, Tokyo, Japan).
Figure 5. Image samples for QR codes printed by eight printers. (a) Canon iR-ADV C50455051 UFR2, (b) Canon iR-ADV C7260270 UFR2 (Canon Inc., Tokyo, Japan), (c) Epson L310 Series(Seiko Epson Ltd., Suwa, Japan), (d) Fuji XEROX DocuCentre S2110, (e) RICOH Aficio MP7502(Ricoh Ltd., Tokyo, Japan), (f) RICOH Pro8100s, (g) Samsung K2200 series(Samsung Electronics Ltd., Seoul, Korea), and (h) TOSHIBA e-STUDIO2051C-11606695(Toshiba, Tokyo, Japan).
Sensors 20 06305 g005
Figure 6. Graphical confusion matrices of several CNN methods.
Figure 6. Graphical confusion matrices of several CNN methods.
Sensors 20 06305 g006
Figure 7. Comparison of the accuracy of the four CNN methods.
Figure 7. Comparison of the accuracy of the four CNN methods.
Sensors 20 06305 g007
Table 1. Printer brand and models used in this study.
Table 1. Printer brand and models used in this study.
No.BrandModel
0CanoniR-ADV C50455051 UFR 2
1CanoniR-ADV C7260270 UFR 2
2EpsonL310 Series
3FujiXEROX DocuCentre S2110
4RICOHAficio MP7502
5RICOHPro 8100s
6SamsungK2200 series
7TOSHIBAe-STUDIO2051C-11606695
Table 2. The printer identification accuracy of PSINet under different image input sizes, different convolution kernel sizes, and different convolution layers.
Table 2. The printer identification accuracy of PSINet under different image input sizes, different convolution kernel sizes, and different convolution layers.
Image Size7 Layers11 Layers15 Layers
3 × 35 × 57 × 73 × 35 × 57 × 73 × 35 × 57 × 7
48 × 4899.90%99.49%99.77%99.84%99.77%99.59%97.75%98.41%96.76%
56 × 5699.92%99.89%99.93%99.85%99.71%99.66%98.00%98.93%99.06%
64 × 6499.89%99.82%99.71%99.46%99.63%99.72%99.69%98.36%99.73%
Table 3. The accuracy of the four CNN methods.
Table 3. The accuracy of the four CNN methods.
MethodsIdentification Accuracy (%)
01234567Ave
LeNet99.48%99.22%29.82%99.22%100.00%99.87%100.00%100.00%90.95%
AlexNet99.87%95.18%94.53%98.44%99.74%100.00%98.44%100.00%98.27%
Min-Jen Tsai’s99.09%97.40%75.65%96.61%99.74%99.87%97.66%100.00%95.75%
PSINet (ours)100.00%100.00%98.57%100.00%100.00%100.00%100.00%100.00%99.82%
Table 4. Comparison of inference time of four CNNs.
Table 4. Comparison of inference time of four CNNs.
CNNsLeNetAlexNetMin-Jen Tsai’sPSINet
Inference Time(s)13241113
Table 5. Comparison of computational costs of four CNN models.
Table 5. Comparison of computational costs of four CNN models.
CNNsLeNetAlexNetPSDIPSINet
Size (MB)16.277.00.435.18
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Guo, Z.; Zheng, H.; You, C.; Xu, X.; Wu, X.; Zheng, Z.; Ju, J. Digital Forensics of Scanned QR Code Images for Printer Source Identification Using Bottleneck Residual Block. Sensors 2020, 20, 6305. https://0-doi-org.brum.beds.ac.uk/10.3390/s20216305

AMA Style

Guo Z, Zheng H, You C, Xu X, Wu X, Zheng Z, Ju J. Digital Forensics of Scanned QR Code Images for Printer Source Identification Using Bottleneck Residual Block. Sensors. 2020; 20(21):6305. https://0-doi-org.brum.beds.ac.uk/10.3390/s20216305

Chicago/Turabian Style

Guo, Zhongyuan, Hong Zheng, Changhui You, Xiaohang Xu, Xiongbin Wu, Zhaohui Zheng, and Jianping Ju. 2020. "Digital Forensics of Scanned QR Code Images for Printer Source Identification Using Bottleneck Residual Block" Sensors 20, no. 21: 6305. https://0-doi-org.brum.beds.ac.uk/10.3390/s20216305

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop