Next Issue
Volume 7, November
Previous Issue
Volume 7, September
 
 

J. Imaging, Volume 7, Issue 10 (October 2021) – 27 articles

Cover Story (view full-size image): With the assistance of deep learning, mass spectrometry imaging can be used to detect cancer and therefore serve as a perioperative tissue assessment tool in surgery. Achieving this, however, requires the development of a database of mass spectrometry signals and their corresponding pathology labels. Assigning correct labels necessitates precise spatial registration of histopathology and mass spectrometry data. This is a challenging task due to the domain differences and noisy nature of images. In this study, we create a registration framework for mass spectrometry and pathology images as a contribution to the development of perioperative tissue assessment. In doing so, we explore two opportunities in deep learning for medical image registration, namely, unsupervised, multi-modal deformable image registration and registration evaluation. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
35 pages, 3397 KiB  
Article
A Hybrid Robust Image Watermarking Method Based on DWT-DCT and SIFT for Copyright Protection
by Mohamed Hamidi, Mohamed El Haziti, Hocine Cherifi and Mohammed El Hassouni
J. Imaging 2021, 7(10), 218; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100218 - 19 Oct 2021
Cited by 10 | Viewed by 2113
Abstract
In this paper, a robust hybrid watermarking method based on discrete wavelet transform (DWT), discrete cosine transform (DCT), and scale-invariant feature transformation (SIFT) is proposed. Indeed, it is of prime interest to develop robust feature-based image watermarking schemes to withstand both image processing [...] Read more.
In this paper, a robust hybrid watermarking method based on discrete wavelet transform (DWT), discrete cosine transform (DCT), and scale-invariant feature transformation (SIFT) is proposed. Indeed, it is of prime interest to develop robust feature-based image watermarking schemes to withstand both image processing attacks and geometric distortions while preserving good imperceptibility. To this end, a robust watermark is embedded in the DWT-DCT domain to withstand image processing manipulations, while SIFT is used to protect the watermark from geometric attacks. First, the watermark is embedded in the middle band of the discrete cosine transform (DCT) coefficients of the HL1 band of the discrete wavelet transform (DWT). Then, the SIFT feature points are registered to be used in the extraction process to correct the geometric transformations. Extensive experiments have been conducted to assess the effectiveness of the proposed scheme. The results demonstrate its high robustness against standard image processing attacks and geometric manipulations while preserving a high imperceptibility. Furthermore, it compares favorably with alternative methods. Full article
(This article belongs to the Special Issue Deep Learning for Visual Contents Processing and Analysis)
Show Figures

Figure 1

24 pages, 1432 KiB  
Review
A Review of Modern Thermal Imaging Sensor Technology and Applications for Autonomous Aerial Navigation
by Tran Xuan Bach Nguyen, Kent Rosser and Javaan Chahl
J. Imaging 2021, 7(10), 217; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100217 - 19 Oct 2021
Cited by 21 | Viewed by 4250
Abstract
Limited navigation capabilities of many current robots and UAVs restricts their applications in GPS denied areas. Large aircraft with complex navigation systems rely on a variety of sensors including radio frequency aids and high performance inertial systems rendering them somewhat resistant to GPS [...] Read more.
Limited navigation capabilities of many current robots and UAVs restricts their applications in GPS denied areas. Large aircraft with complex navigation systems rely on a variety of sensors including radio frequency aids and high performance inertial systems rendering them somewhat resistant to GPS denial. The rapid development of computer vision has seen cameras incorporated into small drones. Vision-based systems, consisting of one or more cameras, could arguably satisfy both size and weight constraints faced by UAVs. A new generation of thermal sensors is available that are lighter, smaller and widely available. Thermal sensors are a solution to enable navigation in difficult environments, including in low-light, dust or smoke. The purpose of this paper is to present a comprehensive literature review of thermal sensors integrated into navigation systems. Furthermore, the physics and characteristics of thermal sensors will also be presented to provide insight into challenges when integrating thermal sensors in place of conventional visual spectrum sensors. Full article
(This article belongs to the Special Issue Formal Verification of Imaging Algorithms for Autonomous System)
Show Figures

Figure 1

24 pages, 2863 KiB  
Article
Flexible Krylov Methods for Edge Enhancement in Imaging
by Silvia Gazzola, Sebastian James Scott and Alastair Spence
J. Imaging 2021, 7(10), 216; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100216 - 18 Oct 2021
Cited by 1 | Viewed by 1992
Abstract
Many successful variational regularization methods employed to solve linear inverse problems in imaging applications (such as image deblurring, image inpainting, and computed tomography) aim at enhancing edges in the solution, and often involve non-smooth regularization terms (e.g., total variation). Such regularization methods can [...] Read more.
Many successful variational regularization methods employed to solve linear inverse problems in imaging applications (such as image deblurring, image inpainting, and computed tomography) aim at enhancing edges in the solution, and often involve non-smooth regularization terms (e.g., total variation). Such regularization methods can be treated as iteratively reweighted least squares problems (IRLS), which are usually solved by the repeated application of a Krylov projection method. This approach gives rise to an inner–outer iterative scheme where the outer iterations update the weights and the inner iterations solve a least squares problem with fixed weights. Recently, flexible or generalized Krylov solvers, which avoid inner–outer iterations by incorporating iteration-dependent weights within a single approximation subspace for the solution, have been devised to efficiently handle IRLS problems. Indeed, substantial computational savings are generally possible by avoiding the repeated application of a traditional Krylov solver. This paper aims to extend the available flexible Krylov algorithms in order to handle a variety of edge-enhancing regularization terms, with computationally convenient adaptive regularization parameter choice. In order to tackle both square and rectangular linear systems, flexible Krylov methods based on the so-called flexible Golub–Kahan decomposition are considered. Some theoretical results are presented (including a convergence proof) and numerical comparisons with other edge-enhancing solvers show that the new methods compute solutions of similar or better quality, with increased speedup. Full article
(This article belongs to the Special Issue Inverse Problems and Imaging)
Show Figures

Figure 1

13 pages, 5192 KiB  
Article
A Combined Radiomics and Machine Learning Approach to Distinguish Clinically Significant Prostate Lesions on a Publicly Available MRI Dataset
by Leandro Donisi, Giuseppe Cesarelli, Anna Castaldo, Davide Raffaele De Lucia, Francesca Nessuno, Gaia Spadarella and Carlo Ricciardi
J. Imaging 2021, 7(10), 215; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100215 - 18 Oct 2021
Cited by 21 | Viewed by 2361
Abstract
Although prostate cancer is one of the most common causes of mortality and morbidity in advancing-age males, early diagnosis improves prognosis and modifies the therapy of choice. The aim of this study was the evaluation of a combined radiomics and machine learning approach [...] Read more.
Although prostate cancer is one of the most common causes of mortality and morbidity in advancing-age males, early diagnosis improves prognosis and modifies the therapy of choice. The aim of this study was the evaluation of a combined radiomics and machine learning approach on a publicly available dataset in order to distinguish a clinically significant from a clinically non-significant prostate lesion. A total of 299 prostate lesions were included in the analysis. A univariate statistical analysis was performed to prove the goodness of the 60 extracted radiomic features in distinguishing prostate lesions. Then, a 10-fold cross-validation was used to train and test some models and the evaluation metrics were calculated; finally, a hold-out was performed and a wrapper feature selection was applied. The employed algorithms were Naïve bayes, K nearest neighbour and some tree-based ones. The tree-based algorithms achieved the highest evaluation metrics, with accuracies over 80%, and area-under-the-curve receiver-operating characteristics below 0.80. Combined machine learning algorithms and radiomics based on clinical, routine, multiparametric, magnetic-resonance imaging were demonstrated to be a useful tool in prostate cancer stratification. Full article
(This article belongs to the Special Issue Radiomics and Texture Analysis in Medical Imaging)
Show Figures

Figure 1

23 pages, 3571 KiB  
Article
CasTabDetectoRS: Cascade Network for Table Detection in Document Images with Recursive Feature Pyramid and Switchable Atrous Convolution
by Khurram Azeem Hashmi, Alain Pagani, Marcus Liwicki, Didier Stricker and Muhammad Zeshan Afzal
J. Imaging 2021, 7(10), 214; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100214 - 16 Oct 2021
Cited by 16 | Viewed by 2665
Abstract
Table detection is a preliminary step in extracting reliable information from tables in scanned document images. We present CasTabDetectoRS, a novel end-to-end trainable table detection framework that operates on Cascade Mask R-CNN, including Recursive Feature Pyramid network and Switchable Atrous Convolution in the [...] Read more.
Table detection is a preliminary step in extracting reliable information from tables in scanned document images. We present CasTabDetectoRS, a novel end-to-end trainable table detection framework that operates on Cascade Mask R-CNN, including Recursive Feature Pyramid network and Switchable Atrous Convolution in the existing backbone architecture. By utilizing a comparativelyightweight backbone of ResNet-50, this paper demonstrates that superior results are attainable without relying on pre- and post-processing methods, heavier backbone networks (ResNet-101, ResNeXt-152), and memory-intensive deformable convolutions. We evaluate the proposed approach on five different publicly available table detection datasets. Our CasTabDetectoRS outperforms the previous state-of-the-art results on four datasets (ICDAR-19, TableBank, UNLV, and Marmot) and accomplishes comparable results on ICDAR-17 POD. Upon comparing with previous state-of-the-art results, we obtain a significant relative error reduction of 56.36%, 20%, 4.5%, and 3.5% on the datasets of ICDAR-19, TableBank, UNLV, and Marmot, respectively. Furthermore, this paper sets a new benchmark by performing exhaustive cross-datasets evaluations to exhibit the generalization capabilities of the proposed method. Full article
(This article belongs to the Section Document Analysis and Processing)
Show Figures

Figure 1

23 pages, 7547 KiB  
Article
Mitral Valve Segmentation Using Robust Nonnegative Matrix Factorization
by Hannah Dröge, Baichuan Yuan, Rafael Llerena, Jesse T. Yen, Michael Moeller and Andrea L. Bertozzi
J. Imaging 2021, 7(10), 213; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100213 - 16 Oct 2021
Cited by 3 | Viewed by 2352
Abstract
Analyzing and understanding the movement of the mitral valve is of vital importance in cardiology, as the treatment and prevention of several serious heart diseases depend on it. Unfortunately, large amounts of noise as well as a highly varying image quality make the [...] Read more.
Analyzing and understanding the movement of the mitral valve is of vital importance in cardiology, as the treatment and prevention of several serious heart diseases depend on it. Unfortunately, large amounts of noise as well as a highly varying image quality make the automatic tracking and segmentation of the mitral valve in two-dimensional echocardiographic videos challenging. In this paper, we present a fully automatic and unsupervised method for segmentation of the mitral valve in two-dimensional echocardiographic videos, independently of the echocardiographic view. We propose a bias-free variant of the robust non-negative matrix factorization (RNMF) along with a window-based localization approach, that is able to identify the mitral valve in several challenging situations. We improve the average f1-score on our dataset of 10 echocardiographic videos by 0.18 to a f1-score of 0.56. Full article
(This article belongs to the Special Issue Inverse Problems and Imaging)
Show Figures

Figure 1

27 pages, 9722 KiB  
Article
Bayesian Activity Estimation and Uncertainty Quantification of Spent Nuclear Fuel Using Passive Gamma Emission Tomography
by Ahmed Karam Eldaly, Ming Fang, Angela Di Fulvio, Stephen McLaughlin, Mike E. Davies, Yoann Altmann and Yves Wiaux
J. Imaging 2021, 7(10), 212; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100212 - 14 Oct 2021
Cited by 2 | Viewed by 2023
Abstract
In this paper, we address the problem of activity estimation in passive gamma emission tomography (PGET) of spent nuclear fuel. Two different noise models are considered and compared, namely, the isotropic Gaussian and the Poisson noise models. The problem is formulated within a [...] Read more.
In this paper, we address the problem of activity estimation in passive gamma emission tomography (PGET) of spent nuclear fuel. Two different noise models are considered and compared, namely, the isotropic Gaussian and the Poisson noise models. The problem is formulated within a Bayesian framework as a linear inverse problem and prior distributions are assigned to the unknown model parameters. In particular, a Bernoulli-truncated Gaussian prior model is considered to promote sparse pin configurations. A Markov chain Monte Carlo (MCMC) method, based on a split and augmented Gibbs sampler, is then used to sample the posterior distribution of the unknown parameters. The proposed algorithm is first validated by simulations conducted using synthetic data, generated using the nominal models. We then consider more realistic data simulated using a bespoke simulator, whose forward model is non-linear and not available analytically. In that case, the linear models used are mis-specified and we analyse their robustness for activity estimation. The results demonstrate superior performance of the proposed approach in estimating the pin activities in different assembly patterns, in addition to being able to quantify their uncertainty measures, in comparison with existing methods. Full article
(This article belongs to the Special Issue Inverse Problems and Imaging)
Show Figures

Figure 1

15 pages, 2018 KiB  
Review
Clinical Molecular Imaging for Atherosclerotic Plaque
by Anton Kondakov and Vladimir Lelyuk
J. Imaging 2021, 7(10), 211; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100211 - 13 Oct 2021
Cited by 8 | Viewed by 2155
Abstract
Atherosclerosis is a well-known disease leading to cardiovascular events, including myocardial infarction and ischemic stroke. These conditions lead to a high mortality rate, which explains the interest in their prevention, early detection, and treatment. Molecular imaging is able to shed light on the [...] Read more.
Atherosclerosis is a well-known disease leading to cardiovascular events, including myocardial infarction and ischemic stroke. These conditions lead to a high mortality rate, which explains the interest in their prevention, early detection, and treatment. Molecular imaging is able to shed light on the basic pathophysiological processes, such as inflammation, that cause the progression and instability of plaque. The most common radiotracers used in clinical practice can detect increased energy metabolism (FDG), macrophage number (somatostatin receptor imaging), the intensity of cell proliferation in the area (labeled choline), and microcalcifications (fluoride imaging). These radiopharmaceuticals, especially FDG and labeled sodium fluoride, can predict cardiovascular events. The limitations of molecular imaging in atherosclerosis include low uptake of highly specific tracers, possible overlap with other diseases of the vessel wall, and specific features of certain tracers’ physiological distribution. A common protocol for patient preparation, data acquisition, and quantification is needed in the area of atherosclerosis imaging research. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

16 pages, 5451 KiB  
Article
Design of Flexible Hardware Accelerators for Image Convolutions and Transposed Convolutions
by Cristian Sestito, Fanny Spagnolo and Stefania Perri
J. Imaging 2021, 7(10), 210; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100210 - 12 Oct 2021
Cited by 7 | Viewed by 1904
Abstract
Nowadays, computer vision relies heavily on convolutional neural networks (CNNs) to perform complex and accurate tasks. Among them, super-resolution CNNs represent a meaningful example, due to the presence of both convolutional (CONV) and transposed convolutional (TCONV) layers. While the former exploit multiply-and-accumulate (MAC) [...] Read more.
Nowadays, computer vision relies heavily on convolutional neural networks (CNNs) to perform complex and accurate tasks. Among them, super-resolution CNNs represent a meaningful example, due to the presence of both convolutional (CONV) and transposed convolutional (TCONV) layers. While the former exploit multiply-and-accumulate (MAC) operations to extract features of interest from incoming feature maps (fmaps), the latter perform MACs to tune the spatial resolution of the received fmaps properly. The ever-growing real-time and low-power requirements of modern computer vision applications represent a stimulus for the research community to investigate the deployment of CNNs on well-suited hardware platforms, such as field programmable gate arrays (FPGAs). FPGAs are widely recognized as valid candidates for trading off computational speed and power consumption, thanks to their flexibility and their capability to also deal with computationally intensive models. In order to reduce the number of operations to be performed, this paper presents a novel hardware-oriented algorithm able to efficiently accelerate both CONVs and TCONVs. The proposed strategy was validated by employing it within a reconfigurable hardware accelerator purposely designed to adapt itself to different operating modes set at run-time. When characterized using the Xilinx XC7K410T FPGA device, the proposed accelerator achieved a throughput of up to 2022.2 GOPS and, in comparison to state-of-the-art competitors, it reached an energy efficiency up to 2.3 times higher, without compromising the overall accuracy. Full article
(This article belongs to the Special Issue Image Processing Using FPGAs 2021)
Show Figures

Figure 1

15 pages, 1049 KiB  
Article
Signal Retrieval from Non-Sinusoidal Intensity Modulations in X-ray and Neutron Interferometry Using Piecewise-Defined Polynomial Function
by Simon Pinzek, Alex Gustschin, Tobias Neuwirth, Alexander Backs, Michael Schulz, Julia Herzen and Franz Pfeiffer
J. Imaging 2021, 7(10), 209; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100209 - 11 Oct 2021
Cited by 2 | Viewed by 1755
Abstract
Grating-based phase-contrast and dark-field imaging systems create intensity modulations that are usually modeled with sinusoidal functions to extract transmission, differential-phase shift, and scatter information. Under certain system-related conditions, the modulations become non-sinusoidal and cause artifacts in conventional processing. To account for that, we [...] Read more.
Grating-based phase-contrast and dark-field imaging systems create intensity modulations that are usually modeled with sinusoidal functions to extract transmission, differential-phase shift, and scatter information. Under certain system-related conditions, the modulations become non-sinusoidal and cause artifacts in conventional processing. To account for that, we introduce a piecewise-defined periodic polynomial function that resembles the physical signal formation process, modeling convolutions of binary periodic functions. Additionally, we extend the model with an iterative expectation-maximization algorithm that can account for imprecise grating positions during phase-stepping. We show that this approach can process a higher variety of simulated and experimentally acquired data, avoiding most artifacts. Full article
(This article belongs to the Special Issue X-ray Digital Radiography and Computed Tomography)
Show Figures

Figure 1

25 pages, 11730 KiB  
Article
A Semiautomatic Multi-Label Color Image Segmentation Coupling Dirichlet Problem and Colour Distances
by Giacomo Aletti, Alessandro Benfenati and Giovanni Naldi
J. Imaging 2021, 7(10), 208; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100208 - 07 Oct 2021
Cited by 11 | Viewed by 1868
Abstract
Image segmentation is an essential but critical component in low level vision, image analysis, pattern recognition, and now in robotic systems. In addition, it is one of the most challenging tasks in image processing and determines the quality of the final results of [...] Read more.
Image segmentation is an essential but critical component in low level vision, image analysis, pattern recognition, and now in robotic systems. In addition, it is one of the most challenging tasks in image processing and determines the quality of the final results of the image analysis. Colour based segmentation could hence offer more significant extraction of information as compared to intensity or texture based segmentation. In this work, we propose a new local or global method for multi-label segmentation that combines a random walk based model with a direct label assignment computed using a suitable colour distance. Our approach is a semi-automatic image segmentation technique, since it requires user interaction for the initialisation of the segmentation process. The random walk part involves a combinatorial Dirichlet problem for a weighted graph, where the nodes are the pixel of the image, and the positive weights are related to the distances between pixels: in this work we propose a novel colour distance for computing such weights. In the random walker model we assign to each pixel of the image a probability quantifying the likelihood that the node belongs to some subregion. The computation of the colour distance is pursued by employing the coordinates in a colour space (e.g., RGB, XYZ, YCbCr) of a pixel and of the ones in its neighbourhood (e.g., in a 8–neighbourhood). The segmentation process is, therefore, reduced to an optimisation problem coupling the probabilities from the random walker approach, and the similarity with respect the labelled pixels. A further investigation involves an adaptive preprocess strategy using a regression tree for learning suitable weights to be used in the computation of the colour distance. We discuss the properties of the new method also by comparing with standard random walk and kmeans approaches. The experimental results carried on the White Blood Cell (WBC) dataset and GrabCut datasets show the remarkable performance of the proposed method in comparison with state-of-the-art methods, such as normalised random walk and normalised lazy random walk, with respect to segmentation quality and computational time. Moreover, it reveals to be very robust with respect to the presence of noise and to the choice of the colourspace. Full article
(This article belongs to the Special Issue Advancing Color Image Processing)
Show Figures

Figure 1

15 pages, 3660 KiB  
Article
Three-Color Balancing for Color Constancy Correction
by Teruaki Akazawa, Yuma Kinoshita, Sayaka Shiota and Hitoshi Kiya
J. Imaging 2021, 7(10), 207; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100207 - 06 Oct 2021
Cited by 4 | Viewed by 1901
Abstract
This paper presents a three-color balance adjustment for color constancy correction. White balancing is a typical adjustment for color constancy in an image, but there are still lighting effects on colors other than white. Cheng et al. proposed multi-color balancing to improve the [...] Read more.
This paper presents a three-color balance adjustment for color constancy correction. White balancing is a typical adjustment for color constancy in an image, but there are still lighting effects on colors other than white. Cheng et al. proposed multi-color balancing to improve the performance of white balancing by mapping multiple target colors into corresponding ground truth colors. However, there are still three problems that have not been discussed: choosing the number of target colors, selecting target colors, and minimizing error which causes computational complexity to increase. In this paper, we first discuss the number of target colors for multi-color balancing. From our observation, when the number of target colors is greater than or equal to three, the best performance of multi-color balancing in each number of target colors is almost the same regardless of the number of target colors, and it is superior to that of white balancing. Moreover, if the number of target colors is three, multi-color balancing can be performed without any error minimization. Accordingly, we propose three-color balancing. In addition, the combination of three target colors is discussed to achieve color constancy correction. In an experiment, the proposed method not only outperforms white balancing but also has almost the same performance as Cheng’s method with 24 target colors. Full article
(This article belongs to the Special Issue Intelligent Media Processing)
Show Figures

Figure 1

16 pages, 3150 KiB  
Article
A Data-Centric Augmentation Approach for Disturbed Sensor Image Segmentation
by Andreas Roth, Konstantin Wüstefeld and Frank Weichert
J. Imaging 2021, 7(10), 206; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100206 - 06 Oct 2021
Cited by 4 | Viewed by 2930
Abstract
In the context of sensor-based data analysis, the compensation of image artifacts is a challenge. When the structures of interest are not clearly visible in an image, algorithms that can cope with artifacts are crucial for obtaining the desired information. Thereby, the high [...] Read more.
In the context of sensor-based data analysis, the compensation of image artifacts is a challenge. When the structures of interest are not clearly visible in an image, algorithms that can cope with artifacts are crucial for obtaining the desired information. Thereby, the high variation of artifacts, the combination of different types of artifacts, and their similarity to signals of interest are specific issues that have to be considered in the analysis. Despite the high generalization capability of deep learning-based approaches, their recent success was driven by the availability of large amounts of labeled data. Therefore, the provision of comprehensive labeled image data with different characteristics of image artifacts is of importance. At the same time, applying deep neural networks to problems with low availability of labeled data remains a challenge. This work presents a data-centric augmentation approach based on generative adversarial networks that augments the existing labeled data with synthetic artifacts generated from data not present in the training set. In our experiments, this augmentation leads to a more robust generalization in segmentation. Our method does not need additional labeling and does not lead to additional memory or time consumption during inference. Further, we find it to be more effective than comparable augmentations based on procedurally generated artifacts and the direct use of real artifacts. Building upon the improved segmentation results, we observe that our approach leads to improvements of 22% in the F1-score for an evaluated detection problem. Having achieved these results with an example sensor, we expect increased robustness against artifacts in future applications. Full article
Show Figures

Figure 1

21 pages, 6021 KiB  
Article
Novel Texture Feature Descriptors Based on Multi-Fractal Analysis and LBP for Classifying Breast Density in Mammograms
by Haipeng Li, Ramakrishnan Mukundan and Shelley Boyd
J. Imaging 2021, 7(10), 205; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100205 - 06 Oct 2021
Cited by 7 | Viewed by 3270
Abstract
This paper investigates the usefulness of multi-fractal analysis and local binary patterns (LBP) as texture descriptors for classifying mammogram images into different breast density categories. Multi-fractal analysis is also used in the pre-processing step to segment the region of interest (ROI). We use [...] Read more.
This paper investigates the usefulness of multi-fractal analysis and local binary patterns (LBP) as texture descriptors for classifying mammogram images into different breast density categories. Multi-fractal analysis is also used in the pre-processing step to segment the region of interest (ROI). We use four multi-fractal measures and the LBP method to extract texture features, and to compare their classification performance in experiments. In addition, a feature descriptor combining multi-fractal features and multi-resolution LBP (MLBP) features is proposed and evaluated in this study to improve classification accuracy. An autoencoder network and principal component analysis (PCA) are used for reducing feature redundancy in the classification model. A full field digital mammogram (FFDM) dataset, INBreast, which contains 409 mammogram images, is used in our experiment. BI-RADS density labels given by radiologists are used as the ground truth to evaluate the classification results using the proposed methods. Experimental results show that the proposed feature descriptor based on multi-fractal features and LBP result in higher classification accuracy than using individual texture feature sets. Full article
(This article belongs to the Special Issue Advances in Image Analysis: Shapes, Textures and Multifractals)
Show Figures

Figure 1

11 pages, 1769 KiB  
Article
Masked Face Analysis via Multi-Task Deep Learning
by Vatsa S. Patel, Zhongliang Nie, Trung-Nghia Le and Tam V. Nguyen
J. Imaging 2021, 7(10), 204; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100204 - 05 Oct 2021
Cited by 5 | Viewed by 2376
Abstract
Face recognition with wearable items has been a challenging task in computer vision and involves the problem of identifying humans wearing a face mask. Masked face analysis via multi-task learning could effectively improve performance in many fields of face analysis. In this paper, [...] Read more.
Face recognition with wearable items has been a challenging task in computer vision and involves the problem of identifying humans wearing a face mask. Masked face analysis via multi-task learning could effectively improve performance in many fields of face analysis. In this paper, we propose a unified framework for predicting the age, gender, and emotions of people wearing face masks. We first construct FGNET-MASK, a masked face dataset for the problem. Then, we propose a multi-task deep learning model to tackle the problem. In particular, the multi-task deep learning model takes the data as inputs and shares their weight to yield predictions of age, expression, and gender for the masked face. Through extensive experiments, the proposed framework has been found to provide a better performance than other existing methods. Full article
(This article belongs to the Special Issue Imaging Studies for Face and Gesture Analysis)
Show Figures

Figure 1

17 pages, 6238 KiB  
Article
Combined Mass Spectrometry and Histopathology Imaging for Perioperative Tissue Assessment in Cancer Surgery
by Laura Connolly, Amoon Jamzad, Martin Kaufmann, Catriona E. Farquharson, Kevin Ren, John F. Rudan, Gabor Fichtinger and Parvin Mousavi
J. Imaging 2021, 7(10), 203; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100203 - 04 Oct 2021
Cited by 4 | Viewed by 2383
Abstract
Mass spectrometry is an effective imaging tool for evaluating biological tissue to detect cancer. With the assistance of deep learning, this technology can be used as a perioperative tissue assessment tool that will facilitate informed surgical decisions. To achieve such a system requires [...] Read more.
Mass spectrometry is an effective imaging tool for evaluating biological tissue to detect cancer. With the assistance of deep learning, this technology can be used as a perioperative tissue assessment tool that will facilitate informed surgical decisions. To achieve such a system requires the development of a database of mass spectrometry signals and their corresponding pathology labels. Assigning correct labels, in turn, necessitates precise spatial registration of histopathology and mass spectrometry data. This is a challenging task due to the domain differences and noisy nature of images. In this study, we create a registration framework for mass spectrometry and pathology images as a contribution to the development of perioperative tissue assessment. In doing so, we explore two opportunities in deep learning for medical image registration, namely, unsupervised, multi-modal deformable image registration and evaluation of the registration. We test this system on prostate needle biopsy cores that were imaged with desorption electrospray ionization mass spectrometry (DESI) and show that we can successfully register DESI and histology images to achieve accurate alignment and, consequently, labelling for future training. This automation is expected to improve the efficiency and development of a deep learning architecture that will benefit the use of mass spectrometry imaging for cancer diagnosis. Full article
Show Figures

Figure 1

20 pages, 2342 KiB  
Article
An Automated Approach for Electric Network Frequency Estimation in Static and Non-Static Digital Video Recordings
by Georgios Karantaidis and Constantine Kotropoulos
J. Imaging 2021, 7(10), 202; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100202 - 02 Oct 2021
Cited by 3 | Viewed by 1952
Abstract
Electric Network Frequency (ENF) is embedded in multimedia recordings if the recordings are captured with a device connected to power mains or placed near the power mains. It is exploited as a tool for multimedia authentication. ENF fluctuates stochastically around its nominal frequency [...] Read more.
Electric Network Frequency (ENF) is embedded in multimedia recordings if the recordings are captured with a device connected to power mains or placed near the power mains. It is exploited as a tool for multimedia authentication. ENF fluctuates stochastically around its nominal frequency at 50/60 Hz. In indoor environments, luminance variations captured by video recordings can also be exploited for ENF estimation. However, the various textures and different levels of shadow and luminance hinder ENF estimation in static and non-static video, making it a non-trivial problem. To address this problem, a novel automated approach is proposed for ENF estimation in static and non-static digital video recordings. The proposed approach is based on the exploitation of areas with similar characteristics in each video frame. These areas, called superpixels, have a mean intensity that exceeds a specific threshold. The performance of the proposed approach is tested on various videos of real-life scenarios that resemble surveillance from security cameras. These videos are of escalating difficulty and span recordings from static ones to recordings, which exhibit continuous motion. The maximum correlation coefficient is employed to measure the accuracy of ENF estimation against the ground truth signal. Experimental results show that the proposed approach improves ENF estimation against the state-of-the-art, yielding statistically significant accuracy improvements. Full article
(This article belongs to the Special Issue Image and Video Forensics)
Show Figures

Figure 1

14 pages, 826 KiB  
Article
Sparsity-Based Recovery of Three-Dimensional Photoacoustic Images from Compressed Single-Shot Optical Detection
by Dylan Green, Anne Gelb and Geoffrey P. Luke
J. Imaging 2021, 7(10), 201; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100201 - 02 Oct 2021
Cited by 2 | Viewed by 2067
Abstract
Photoacoustic (PA) imaging combines optical excitation with ultrasonic detection to achieve high-resolution imaging of biological samples. A high-energy pulsed laser is often used for imaging at multi-centimeter depths in tissue. These lasers typically have a low pulse repetition rate, so to acquire images [...] Read more.
Photoacoustic (PA) imaging combines optical excitation with ultrasonic detection to achieve high-resolution imaging of biological samples. A high-energy pulsed laser is often used for imaging at multi-centimeter depths in tissue. These lasers typically have a low pulse repetition rate, so to acquire images in real-time, only one pulse of the laser can be used per image. This single pulse necessitates the use of many individual detectors and receive electronics to adequately record the resulting acoustic waves and form an image. Such requirements make many PA imaging systems both costly and complex. This investigation proposes and models a method of volumetric PA imaging using a state-of-the-art compressed sensing approach to achieve real-time acquisition of the initial pressure distribution (IPD) at a reduced level of cost and complexity. In particular, a single exposure of an optical image sensor is used to capture an entire Fabry–Pérot interferometric acoustic sensor. Time resolved encoding as achieved through spatial sweeping with a galvanometer. This optical system further makes use of a random binary mask to set a predetermined subset of pixels to zero, thus enabling recovery of the time-resolved signals. The Two-Step Iterative Shrinking and Thresholding algorithm is used to reconstruct the IPD, harnessing the sparsity naturally occurring in the IPD as well as the additional structure provided by the binary mask. We conduct experiments on simulated data and analyze the performance of our new approach. Full article
(This article belongs to the Special Issue Inverse Problems and Imaging)
Show Figures

Figure 1

18 pages, 26877 KiB  
Article
Single-Input Multi-Output U-Net for Automated 2D Foetal Brain Segmentation of MR Images
by Andrik Rampun, Deborah Jarvis, Paul D. Griffiths, Reyer Zwiggelaar, Bryan W. Scotney and Paul A. Armitage
J. Imaging 2021, 7(10), 200; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100200 - 01 Oct 2021
Cited by 6 | Viewed by 2002
Abstract
In this work, we develop the Single-Input Multi-Output U-Net (SIMOU-Net), a hybrid network for foetal brain segmentation inspired by the original U-Net fused with the holistically nested edge detection (HED) network. The SIMOU-Net is similar to the original U-Net but it has a [...] Read more.
In this work, we develop the Single-Input Multi-Output U-Net (SIMOU-Net), a hybrid network for foetal brain segmentation inspired by the original U-Net fused with the holistically nested edge detection (HED) network. The SIMOU-Net is similar to the original U-Net but it has a deeper architecture and takes account of the features extracted from each side output. It acts similar to an ensemble neural network, however, instead of averaging the outputs from several independently trained models, which is computationally expensive, our approach combines outputs from a single network to reduce the variance of predications and generalization errors. Experimental results using 200 normal foetal brains consisting of over 11,500 2D images produced Dice and Jaccard coefficients of 94.2 ± 5.9% and 88.7 ± 6.9%, respectively. We further tested the proposed network on 54 abnormal cases (over 3500 images) and achieved Dice and Jaccard coefficients of 91.2 ± 6.8% and 85.7 ± 6.6%, respectively. Full article
(This article belongs to the Special Issue Current Methods in Medical Image Segmentation)
Show Figures

Figure 1

13 pages, 1675 KiB  
Article
Fast Energy Dependent Scatter Correction for List-Mode PET Data
by Juan Manuel Álvarez-Gómez, Joaquín Santos-Blasco, Laura Moliner Martínez and María José Rodríguez-Álvarez
J. Imaging 2021, 7(10), 199; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100199 - 30 Sep 2021
Cited by 4 | Viewed by 1673
Abstract
Improvements in energy resolution of modern positron emission tomography (PET) detectors have created opportunities to implement energy-based scatter correction algorithms. Here, we use the energy information of auxiliary windows to estimate the scatter component. Our method is directly implemented in an iterative reconstruction [...] Read more.
Improvements in energy resolution of modern positron emission tomography (PET) detectors have created opportunities to implement energy-based scatter correction algorithms. Here, we use the energy information of auxiliary windows to estimate the scatter component. Our method is directly implemented in an iterative reconstruction algorithm, generating a scatter-corrected image without the need for sinograms. The purpose was to implement a fast energy-based scatter correction method on list-mode PET data, when it was not possible to use an attenuation map as a practical approach for the scatter degradation. The proposed method was evaluated using Monte Carlo simulations of various digital phantoms. It accurately estimated the scatter fraction distribution, and improved the image contrast in the simulated studied cases. We conclude that the proposed scatter correction method could effectively correct the scattered events, including multiple scatters and those originated in sources outside the field of view. Full article
Show Figures

Figure 1

15 pages, 1202 KiB  
Article
Semi-Supervised Domain Adaptation for Holistic Counting under Label Gap
by Mattia Litrico, Sebastiano Battiato, Sotirios A. Tsaftaris and Mario Valerio Giuffrida
J. Imaging 2021, 7(10), 198; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100198 - 29 Sep 2021
Cited by 1 | Viewed by 2049
Abstract
This paper proposes a novel approach for semi-supervised domain adaptation for holistic regression tasks, where a DNN predicts a continuous value yR given an input image x. The current literature generally lacks specific domain adaptation approaches for this task, as [...] Read more.
This paper proposes a novel approach for semi-supervised domain adaptation for holistic regression tasks, where a DNN predicts a continuous value yR given an input image x. The current literature generally lacks specific domain adaptation approaches for this task, as most of them mostly focus on classification. In the context of holistic regression, most of the real-world datasets not only exhibit a covariate (or domain) shift, but also a label gap—the target dataset may contain labels not included in the source dataset (and vice versa). We propose an approach tackling both covariate and label gap in a unified training framework. Specifically, a Generative Adversarial Network (GAN) is used to reduce covariate shift, and label gap is mitigated via label normalisation. To avoid overfitting, we propose a stopping criterion that simultaneously takes advantage of the Maximum Mean Discrepancy and the GAN Global Optimality condition. To restore the original label range—that was previously normalised—a handful of annotated images from the target domain are used. Our experimental results, run on 3 different datasets, demonstrate that our approach drastically outperforms the state-of-the-art across the board. Specifically, for the cell counting problem, the mean squared error (MSE) is reduced from 759 to 5.62; in the case of the pedestrian dataset, our approach lowered the MSE from 131 to 1.47. For the last experimental setup, we borrowed a task from plant biology, i.e., counting the number of leaves in a plant, and we ran two series of experiments, showing the MSE is reduced from 2.36 to 0.88 (intra-species), and from 1.48 to 0.6 (inter-species). Full article
(This article belongs to the Special Issue Transfer Learning Applications for Real-World Imaging Problems)
Show Figures

Figure 1

38 pages, 13223 KiB  
Review
Roadmap on Recent Progress in FINCH Technology
by Joseph Rosen, Simon Alford, Vijayakumar Anand, Jonathan Art, Petr Bouchal, Zdeněk Bouchal, Munkh-Uchral Erdenebat, Lingling Huang, Ayumi Ishii, Saulius Juodkazis, Nam Kim, Peter Kner, Takako Koujin, Yuichi Kozawa, Dong Liang, Jun Liu, Christopher Mann, Abhijit Marar, Atsushi Matsuda, Teruyoshi Nobukawa, Takanori Nomura, Ryutaro Oi, Mariana Potcoava, Tatsuki Tahara, Bang Le Thanh and Hongqiang Zhouadd Show full author list remove Hide full author list
J. Imaging 2021, 7(10), 197; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100197 - 29 Sep 2021
Cited by 42 | Viewed by 5547
Abstract
Fresnel incoherent correlation holography (FINCH) was a milestone in incoherent holography. In this roadmap, two pathways, namely the development of FINCH and applications of FINCH explored by many prominent research groups, are discussed. The current state-of-the-art FINCH technology, challenges, and future perspectives of [...] Read more.
Fresnel incoherent correlation holography (FINCH) was a milestone in incoherent holography. In this roadmap, two pathways, namely the development of FINCH and applications of FINCH explored by many prominent research groups, are discussed. The current state-of-the-art FINCH technology, challenges, and future perspectives of FINCH technology as recognized by a diverse group of researchers contributing to different facets of research in FINCH have been presented. Full article
Show Figures

Figure 1

10 pages, 22083 KiB  
Article
Variational Anisotropic Gradient-Domain Image Processing
by Ivar Farup
J. Imaging 2021, 7(10), 196; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100196 - 29 Sep 2021
Viewed by 1538
Abstract
Gradient-domain image processing is a technique where, instead of operating directly on the image pixel values, the gradient of the image is computed and processed. The resulting image is obtained by reintegrating the processed gradient. This is normally done by solving the Poisson [...] Read more.
Gradient-domain image processing is a technique where, instead of operating directly on the image pixel values, the gradient of the image is computed and processed. The resulting image is obtained by reintegrating the processed gradient. This is normally done by solving the Poisson equation, most often by means of a finite difference implementation of the gradient descent method. However, this technique in some cases lead to severe haloing artefacts in the resulting image. To deal with this, local or anisotropic diffusion has been added as an ad hoc modification of the Poisson equation. In this paper, we show that a version of anisotropic gradient-domain image processing can result from a more general variational formulation through the minimisation of a functional formulated in terms of the eigenvalues of the structure tensor of the differences between the processed gradient and the gradient of the original image. Example applications of linear and nonlinear local contrast enhancement and colour image Daltonisation illustrate the behaviour of the method. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

3 pages, 171 KiB  
Editorial
Editorial for Special Issue “Fine Art Pattern Extraction and Recognition”
by Fabio Bellavia, Giovanna Castellano and Gennaro Vessio
J. Imaging 2021, 7(10), 195; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100195 - 29 Sep 2021
Viewed by 1857
Abstract
Cultural heritage, especially the fine arts, plays an invaluable role in the cultural, historical, and economic growth of our societies [...] Full article
(This article belongs to the Special Issue Fine Art Pattern Extraction and Recognition)
31 pages, 6589 KiB  
Article
Spatially Coherent Clustering Based on Orthogonal Nonnegative Matrix Factorization
by Pascal Fernsel
J. Imaging 2021, 7(10), 194; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100194 - 28 Sep 2021
Cited by 2 | Viewed by 1562
Abstract
Classical approaches in cluster analysis are typically based on a feature space analysis. However, many applications lead to datasets with additional spatial information and a ground truth with spatially coherent classes, which will not necessarily be reconstructed well by standard clustering methods. Motivated [...] Read more.
Classical approaches in cluster analysis are typically based on a feature space analysis. However, many applications lead to datasets with additional spatial information and a ground truth with spatially coherent classes, which will not necessarily be reconstructed well by standard clustering methods. Motivated by applications in hyperspectral imaging, we introduce in this work clustering models based on Orthogonal Nonnegative Matrix Factorization (ONMF), which include an additional Total Variation (TV) regularization procedure on the cluster membership matrix to enforce the needed spatial coherence in the clusters. We propose several approaches with different optimization techniques, where the TV regularization is either performed as a subsequent post-processing step or included into the clustering algorithm. Finally, we provide a numerical evaluation of 12 different TV regularized ONMF methods on a hyperspectral dataset obtained from a matrix-assisted laser desorption/ionization imaging measurement, which leads to significantly better clustering results compared to classical clustering models. Full article
(This article belongs to the Special Issue Inverse Problems and Imaging)
Show Figures

Figure 1

15 pages, 3009 KiB  
Article
Detection of Manipulated Face Videos over Social Networks: A Large-Scale Study
by Federico Marcon, Cecilia Pasquini and Giulia Boato
J. Imaging 2021, 7(10), 193; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100193 - 28 Sep 2021
Cited by 9 | Viewed by 2280
Abstract
The detection of manipulated videos represents a highly relevant problem in multimedia forensics, which has been widely investigated in the last years. However, a common trait of published studies is the fact that the forensic analysis is typically applied on data prior to [...] Read more.
The detection of manipulated videos represents a highly relevant problem in multimedia forensics, which has been widely investigated in the last years. However, a common trait of published studies is the fact that the forensic analysis is typically applied on data prior to their potential dissemination over the web. This work addresses the challenging scenario where manipulated videos are first shared through social media platforms and then are subject to the forensic analysis. In this context, a large scale performance evaluation has been carried out involving general purpose deep networks and state-of-the-art manipulated data, and studying different effects. Results confirm that a performance drop is observed in every case when unseen shared data are tested by networks trained on non-shared data; however, fine-tuning operations can mitigate this problem. Also, we show that the output of differently trained networks can carry useful forensic information for the identification of the specific technique used for visual manipulation, both for shared and non-shared data. Full article
(This article belongs to the Special Issue Image and Video Forensics)
Show Figures

Figure 1

11 pages, 5206 KiB  
Article
High Spatial-Resolution Digital Phase-Stepping Shearography
by Awatef Rashid Al Jabri, Kazi Monowar Abedin and Sheikh Mohammed Mujibur Rahman
J. Imaging 2021, 7(10), 192; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging7100192 - 27 Sep 2021
Cited by 6 | Viewed by 1610
Abstract
Digital phase-stepping shearography is a speckle interferometric technique that uses laser speckles to generate the phase map of the displacement derivatives of a stressed object, and hence can map the stresses of a deformed object directly. Conventional digital phase-stepping shearography relies on the [...] Read more.
Digital phase-stepping shearography is a speckle interferometric technique that uses laser speckles to generate the phase map of the displacement derivatives of a stressed object, and hence can map the stresses of a deformed object directly. Conventional digital phase-stepping shearography relies on the use of video cameras of relatively lower resolution, in the order of 5 megapixels or lower, operating at a video rate. In the present work, we propose a novel method of performing high spatial resolution phase stepping shearography. This method uses a 24 megapixel still digital imaging device (DSLR camera) and a Michelson-type shearing arrangement with an edge-clamped, center-loaded plate. Different phase-stepping algorithms were used, and all successfully generated shearograms. The system enabled extremely high-resolution phase maps to be generated from relatively large deformations applied to the test plate. Quantitative comparison of the maximum achieved spatial resolution is made with the video-rate cameras used in conventional shearography. By switching from conventional (video) imaging methods to still imaging methods, significantly higher spatial resolution (by about 5 times) can be achieved in actual phase-stepping shearography, which is of great usefulness in industrial non-destructive testing (NDT). Full article
(This article belongs to the Special Issue Digital Holography: Development and Application)
Show Figures

Figure 1

Previous Issue
Next Issue
Back to TopTop