Next Issue
Volume 6, October
Previous Issue
Volume 6, August
 
 

J. Imaging, Volume 6, Issue 9 (September 2020) – 18 articles

Cover Story (view full-size image): In modern production processes, non-destructive testing (NDT) is becoming increasingly important. One emerging photonic NDT method is active thermography, where components are heated by an optical excitation while the temperature field is observed with an infrared camera. Thus, the temporal evolution of the 2D temperature data (thermal video) is measured and defects that alter the local heat diffusion can be identified by analyzing this time-dependent thermal sequence. We propose an algorithm that can be applied for defect segmentation and depth estimation, even for complex-shaped geometries, while keeping the computational cost low. We implement our algorithm by adapting the well-known thermographic signal reconstruction (TSR) method and compare the results to state-of-the-art thermographic methods using a composite component. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
11 pages, 5850 KiB  
Tutorial
Lensless Three-Dimensional Quantitative Phase Imaging Using Phase Retrieval Algorithm
by Vijayakumar Anand, Tomas Katkus, Denver P. Linklater, Elena P. Ivanova and Saulius Juodkazis
J. Imaging 2020, 6(9), 99; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090099 - 20 Sep 2020
Cited by 22 | Viewed by 4032
Abstract
Quantitative phase imaging (QPI) techniques are widely used for the label-free examining of transparent biological samples. QPI techniques can be broadly classified into interference-based and interferenceless methods. The interferometric methods which record the complex amplitude are usually bulky with many optical components and [...] Read more.
Quantitative phase imaging (QPI) techniques are widely used for the label-free examining of transparent biological samples. QPI techniques can be broadly classified into interference-based and interferenceless methods. The interferometric methods which record the complex amplitude are usually bulky with many optical components and use coherent illumination. The interferenceless approaches which need only the intensity distribution and works using phase retrieval algorithms have gained attention as they require lesser resources, cost, space and can work with incoherent illumination. With rapid developments in computational optical techniques and deep learning, QPI has reached new levels of applications. In this tutorial, we discuss one of the basic optical configurations of a lensless QPI technique based on the phase-retrieval algorithm. Simulative studies on QPI of thin, thick, and greyscale phase objects with assistive pseudo-codes and computational codes in Octave is provided. Binary phase samples with positive and negative resist profiles were fabricated using lithography, and a single plane and two plane phase objects were constructed. Light diffracted from a point object is modulated by phase samples and the corresponding intensity patterns are recorded. The phase retrieval approach is applied for 2D and 3D phase reconstructions. Commented codes in Octave for image acquisition and automation using a web camera in an open source operating system are provided. Full article
(This article belongs to the Special Issue Current Highlights and Future Applications of Computational Imaging)
Show Figures

Figure 1

19 pages, 5993 KiB  
Article
Visualization of the Anisotropy of the Velocity Dispersion and Characteristics of the Multi-Velocity Continuum in the Regions of Multi-Stream Flows of Gas-Dust Media with Polydisperse Dust
by Mikhail A. Bezborodov, Mikhail A. Eremin, Vitaly V. Korolev, Ilya G. Kovalenko and Elena V. Zhukova
J. Imaging 2020, 6(9), 98; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090098 - 17 Sep 2020
Cited by 1 | Viewed by 1764
Abstract
Collisionless media devoid of intrinsic stresses, for example, a dispersed phase in a multiphase medium, have a much wider variety of space-time structures and features formed in them than collisional media, for example, a carrier, gas, or liquid phase. This is a consequence [...] Read more.
Collisionless media devoid of intrinsic stresses, for example, a dispersed phase in a multiphase medium, have a much wider variety of space-time structures and features formed in them than collisional media, for example, a carrier, gas, or liquid phase. This is a consequence of the fact that evolution in such media occurs in phase space, i.e., in a space of greater dimensions than the usual coordinate space. As a consequence, the process of the formation of features in collisionless media (clustering or vice versa, a loss of continuity) can occur primarily in the velocity space, which, in contrast to the features in the coordinate space (folds, caustics, or voids), is poorly observed directly. To identify such features, it is necessary to use visualization methods that allow us to consider, in detail, the evolution of the medium in the velocity space. This article is devoted to the development of techniques that allow visualizing the degree of anisotropy of the velocity fields of collisionless interpenetrating media. Simultaneously tracking the behavior of different fractions in such media is important, as their behavior can be significantly different. We propose three different techniques for visualizing the anisotropy of velocity fields using the example of two- and three-continuum dispersed media models. We proposed the construction of spatial distributions of eccentricity fields (scalar fields), or fields of principal directions of the velocity dispersion tensor (tensor fields). In the first case, we used some simple eccentricity functions for dispersion tensors for two fractions simultaneously, which we call surrogate entropy. In the second case, to visualize the anisotropy of the velocity fields of three fractions simultaneously, we used an ordered array (3-vector) of eccentricities for the color representation through decomposition in three basic colors. In the case of a multi-stream flow, we used cluster analysis methods to identify sections of a multi-stream flow (beams) and used glyphs to visualize the entire set of beams (vector-tensor fields). Full article
Show Figures

Graphical abstract

16 pages, 4118 KiB  
Article
Understanding the Effects of Optimal Combination of Spectral Bands on Deep Learning Model Predictions: A Case Study Based on Permafrost Tundra Landform Mapping Using High Resolution Multispectral Satellite Imagery
by Md Abul Ehsan Bhuiyan, Chandi Witharana, Anna K. Liljedahl, Benjamin M. Jones, Ronald Daanen, Howard E. Epstein, Kelcy Kent, Claire G. Griffin and Amber Agnew
J. Imaging 2020, 6(9), 97; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090097 - 17 Sep 2020
Cited by 21 | Viewed by 5273
Abstract
Deep learning (DL) convolutional neural networks (CNNs) have been rapidly adapted in very high spatial resolution (VHSR) satellite image analysis. DLCNN-based computer visions (CV) applications primarily aim for everyday object detection from standard red, green, blue (RGB) imagery, while earth science remote sensing [...] Read more.
Deep learning (DL) convolutional neural networks (CNNs) have been rapidly adapted in very high spatial resolution (VHSR) satellite image analysis. DLCNN-based computer visions (CV) applications primarily aim for everyday object detection from standard red, green, blue (RGB) imagery, while earth science remote sensing applications focus on geo object detection and classification from multispectral (MS) imagery. MS imagery includes RGB and narrow spectral channels from near- and/or middle-infrared regions of reflectance spectra. The central objective of this exploratory study is to understand to what degree MS band statistics govern DLCNN model predictions. We scaffold our analysis on a case study that uses Arctic tundra permafrost landform features called ice-wedge polygons (IWPs) as candidate geo objects. We choose Mask RCNN as the DLCNN architecture to detect IWPs from eight-band Worldview-02 VHSR satellite imagery. A systematic experiment was designed to understand the impact on choosing the optimal three-band combination in model prediction. We tasked five cohorts of three-band combinations coupled with statistical measures to gauge the spectral variability of input MS bands. The candidate scenes produced high model detection accuracies for the F1 score, ranging between 0.89 to 0.95, for two different band combinations (coastal blue, blue, green (1,2,3) and green, yellow, red (3,4,5)). The mapping workflow discerned the IWPs by exhibiting low random and systematic error in the order of 0.17–0.19 and 0.20–0.21, respectively, for band combinations (1,2,3). Results suggest that the prediction accuracy of the Mask-RCNN model is significantly influenced by the input MS bands. Overall, our findings accentuate the importance of considering the image statistics of input MS bands and careful selection of optimal bands for DLCNN predictions when DLCNN architectures are restricted to three spectral channels. Full article
(This article belongs to the Special Issue Robust Image Processing)
Show Figures

Figure 1

15 pages, 4668 KiB  
Article
Extension of the Thermographic Signal Reconstruction Technique for an Automated Segmentation and Depth Estimation of Subsurface Defects
by Alexander Schager, Gerald Zauner, Günther Mayr and Peter Burgholzer
J. Imaging 2020, 6(9), 96; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090096 - 11 Sep 2020
Cited by 7 | Viewed by 2646
Abstract
With increased use of light-weight materials with low factors of safety, non-destructive testing becomes increasingly important. Thanks to the advancement of infrared camera technology, pulse thermography is a cost efficient way to detect subsurface defects non-destructively. However, currently available evaluation algorithms have either [...] Read more.
With increased use of light-weight materials with low factors of safety, non-destructive testing becomes increasingly important. Thanks to the advancement of infrared camera technology, pulse thermography is a cost efficient way to detect subsurface defects non-destructively. However, currently available evaluation algorithms have either a high computational cost or show poor performance if any geometry other than the most simple kind is surveyed. We present an extension of the thermographic signal reconstruction technique which can automatically segment and image defects from sound areas, while also estimating the defect depth, all with low computational cost. We verified our algorithm using real world measurements and compare our results to standard active thermography algorithms with similar computational complexity. We found that our algorithm can detect defects more accurately, especially when more complex geometries are examined. Full article
Show Figures

Figure 1

16 pages, 814 KiB  
Review
Deep Learning-Based Crowd Scene Analysis Survey
by Sherif Elbishlawi, Mohamed H. Abdelpakey, Agwad Eltantawy, Mohamed S. Shehata and Mostafa M. Mohamed
J. Imaging 2020, 6(9), 95; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090095 - 11 Sep 2020
Cited by 24 | Viewed by 3862
Abstract
Recently, our world witnessed major events that attracted a lot of attention towards the importance of automatic crowd scene analysis. For example, the COVID-19 breakout and public events require an automatic system to manage, count, secure, and track a crowd that shares the [...] Read more.
Recently, our world witnessed major events that attracted a lot of attention towards the importance of automatic crowd scene analysis. For example, the COVID-19 breakout and public events require an automatic system to manage, count, secure, and track a crowd that shares the same area. However, analyzing crowd scenes is very challenging due to heavy occlusion, complex behaviors, and posture changes. This paper surveys deep learning-based methods for analyzing crowded scenes. The reviewed methods are categorized as (1) crowd counting and (2) crowd actions recognition. Moreover, crowd scene datasets are surveyed. In additional to the above surveys, this paper proposes an evaluation metric for crowd scene analysis methods. This metric estimates the difference between calculated crowed count and actual count in crowd scene videos. Full article
Show Figures

Figure 1

14 pages, 6189 KiB  
Article
body2vec: 3D Point Cloud Reconstruction for Precise Anthropometry with Handheld Devices
by Magda Alexandra Trujillo-Jiménez, Pablo Navarro, Bruno Pazos, Leonardo Morales, Virginia Ramallo, Carolina Paschetta, Soledad De Azevedo, Anahí Ruderman, Orlando Pérez, Claudio Delrieux and Rolando Gonzalez-José
J. Imaging 2020, 6(9), 94; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090094 - 11 Sep 2020
Cited by 11 | Viewed by 4262
Abstract
Current point cloud extraction methods based on photogrammetry generate large amounts of spurious detections that hamper useful 3D mesh reconstructions or, even worse, the possibility of adequate measurements. Moreover, noise removal methods for point clouds are complex, slow and incapable to cope with [...] Read more.
Current point cloud extraction methods based on photogrammetry generate large amounts of spurious detections that hamper useful 3D mesh reconstructions or, even worse, the possibility of adequate measurements. Moreover, noise removal methods for point clouds are complex, slow and incapable to cope with semantic noise. In this work, we present body2vec, a model-based body segmentation tool that uses a specifically trained Neural Network architecture. Body2vec is capable to perform human body point cloud reconstruction from videos taken on hand-held devices (smartphones or tablets), achieving high quality anthropometric measurements. The main contribution of the proposed workflow is to perform a background removal step, thus avoiding the spurious points generation that is usual in photogrammetric reconstruction. A group of 60 persons were taped with a smartphone, and the corresponding point clouds were obtained automatically with standard photogrammetric methods. We used as a 3D silver standard the clean meshes obtained at the same time with LiDAR sensors post-processed and noise-filtered by expert anthropological biologists. Finally, we used as gold standard anthropometric measurements of the waist and hip of the same people, taken by expert anthropometrists. Applying our method to the raw videos significantly enhanced the quality of the results of the point cloud as compared with the LiDAR-based mesh, and of the anthropometric measurements as compared with the actual hip and waist perimeter measured by the anthropometrists. In both contexts, the resulting quality of body2vec is equivalent to the LiDAR reconstruction. Full article
(This article belongs to the Special Issue 3D and Multimodal Image Acquisition Methods)
Show Figures

Figure 1

19 pages, 912 KiB  
Article
Unsupervised Mitral Valve Tracking for Disease Detection in Echocardiogram Videos
by Kazi Tanzeem Shahid and Ioannis Schizas
J. Imaging 2020, 6(9), 93; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090093 - 09 Sep 2020
Cited by 2 | Viewed by 2445
Abstract
In this work, a novel algorithmic scheme is developed that processes echocardiogram videos, and tracks the movement of the mitral valve leaflets, and thereby estimates whether the movement is symptomatic of a healthy or diseased heart. This algorithm uses automatic Otsu’s thresholding to [...] Read more.
In this work, a novel algorithmic scheme is developed that processes echocardiogram videos, and tracks the movement of the mitral valve leaflets, and thereby estimates whether the movement is symptomatic of a healthy or diseased heart. This algorithm uses automatic Otsu’s thresholding to find a closed boundary around the left atrium, with the basic presumption that it is situated in the bottom right corner of the apical 4 chamber view. A centroid is calculated, and protruding prongs are taken within a 40-degree cone above the centroid, where the mitral valve is located. Binary images are obtained from the videos where the mitral valve leaflets have different pixel values than the cavity of the left atrium. Thus, the points where the prongs touch the valve will show where the mitral valve leaflets are located. The standard deviation of these points is used to calculate closeness of the leaflets. The estimation of the valve movement across subsequent frames is used to determine if the movement is regular, or affected by heart disease. Tests conducted with numerous videos containing both healthy and diseased hearts attest to our method’s efficacy, with a key novelty in being fully unsupervised and computationally efficient. Full article
Show Figures

Figure 1

17 pages, 573 KiB  
Article
Comparative Study of First Order Optimizers for Image Classification Using Convolutional Neural Networks on Histopathology Images
by Ibrahem Kandel, Mauro Castelli and Aleš Popovič
J. Imaging 2020, 6(9), 92; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090092 - 08 Sep 2020
Cited by 43 | Viewed by 5418
Abstract
The classification of histopathology images requires an experienced physician with years of experience to classify the histopathology images accurately. In this study, an algorithm was developed to assist physicians in classifying histopathology images; the algorithm receives the histopathology image as an input and [...] Read more.
The classification of histopathology images requires an experienced physician with years of experience to classify the histopathology images accurately. In this study, an algorithm was developed to assist physicians in classifying histopathology images; the algorithm receives the histopathology image as an input and produces the percentage of cancer presence. The primary classifier used in this algorithm is the convolutional neural network, which is a state-of-the-art classifier used in image classification as it can classify images without relying on the manual selection of features from each image. The main aim of this research is to improve the robustness of the classifier used by comparing six different first-order stochastic gradient-based optimizers to select the best for this particular dataset. The dataset used to train the classifier is the PatchCamelyon public dataset, which consists of 220,025 images to train the classifier; the dataset is composed of 60% positive images and 40% negative images, and 57,458 images to test its performance. The classifier was trained on 80% of the images and validated on the rest of 20% of the images; then, it was tested on the test set. The optimizers were evaluated based on their AUC of the ROC curve. The results show that the adaptative based optimizers achieved the highest results except for AdaGrad that achieved the lowest results. Full article
(This article belongs to the Special Issue Deep Learning in Medical Image Analysis)
Show Figures

Figure 1

28 pages, 2182 KiB  
Article
On the Use of Deep Active Semi-Supervised Learning for Fast Rendering in Global Illumination
by Ibtissam Constantin, Joseph Constantin and André Bigand
J. Imaging 2020, 6(9), 91; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090091 - 08 Sep 2020
Cited by 2 | Viewed by 2279
Abstract
Convolution neural networks usually require large labeled data-sets to construct accurate models. However, in many real-world scenarios, such as global illumination, labeling data are a time-consuming and costly human intelligent task. Semi-supervised learning methods leverage this issue by making use of a small [...] Read more.
Convolution neural networks usually require large labeled data-sets to construct accurate models. However, in many real-world scenarios, such as global illumination, labeling data are a time-consuming and costly human intelligent task. Semi-supervised learning methods leverage this issue by making use of a small labeled data-set and a larger set of unlabeled data. In this paper, our contributions focus on the development of a robust algorithm that combines active and deep semi-supervised convolution neural network to reduce labeling workload and to accelerate convergence in case of real-time global illumination. While the theoretical concepts of photo-realistic rendering are well understood, the increased need for the delivery of highly dynamic interactive content in vast virtual environments has increased recently. Particularly, the quality measure of computer-generated images is of great importance. The experiments are conducted on global illumination scenes which contain diverse distortions. Compared with human psycho-visual thresholds, the good consistency between these thresholds and the learning models quality measures can been seen. A comparison has also been made with SVM and other state-of-the-art deep learning models. We do transfer learning by running the convolution base of these models over our image set. Then, we use the output features of the convolution base as input to retrain the parameters of the fully connected layer. The obtained results show that our proposed method provides promising efficiency in terms of precision, time complexity, and optimal architecture. Full article
Show Figures

Figure 1

15 pages, 26387 KiB  
Article
The Colour of the Night Sky
by Zoltán Kolláth, Dénes Száz, Kai Pong Tong and Kornél Kolláth
J. Imaging 2020, 6(9), 90; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090090 - 05 Sep 2020
Cited by 6 | Viewed by 7388
Abstract
The measurement of night sky quality has become an important task in night sky conservation. Modern measurement techniques involve mainly a calibrated digital camera or a spectroradiometer. However, panchromatic devices are still prevalent to this day, even in the absence of determining the [...] Read more.
The measurement of night sky quality has become an important task in night sky conservation. Modern measurement techniques involve mainly a calibrated digital camera or a spectroradiometer. However, panchromatic devices are still prevalent to this day, even in the absence of determining the spectral information of the night sky. In the case of multispectral measurements, colour information is currently presented in multiple ways. One of the most frequently used metrics is correlated colour temperature (CCT), which is not without its limitation for the purpose of describing especially the colour of natural night sky. Moreover, visually displaying the colour of the night sky in a quantitatively meaningful way has not attracted sufficient attention in the community of astronomy and light pollution research—most photographs of the night sky are post-processed in a way for aesthetic attractiveness rather than accurate representation of the night sky. The spectrum of the natural night sky varies in a wide range depending on solar activity and atmospheric properties. The most noticeable variation in the visible range is the variation of the atomic emission lines, primarily the green oxygen and orange sodium emission. Based on the accepted models of night sky emission, we created a random spectral database which represents the possible range of night sky radiance distribution. We used this spectral database as a learning set, to create a colour transformation between different colour spaces. The spectral sensitivity of some digital cameras is also used to determine an optimal transformation matrix from camera defined coordinates to real colours. The theoretical predictions were extended with actual spectral measurements in order to test the models and check the local constituents of night sky radiance. Here, we present an extended modelling of night sky colour and recommendations of its consistent measurement, as well as methods of visualising the colour of night sky in a consistent way, namely using the false colour enhancement. Full article
Show Figures

Figure 1

15 pages, 1057 KiB  
Article
An Experimental Comparison between Deep Learning and Classical Machine Learning Approaches for Writer Identification in Medieval Documents
by Nicole Dalia Cilia, Claudio De Stefano, Francesco Fontanella, Claudio Marrocco, Mario Molinara and Alessandra Scotto di Freca
J. Imaging 2020, 6(9), 89; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090089 - 04 Sep 2020
Cited by 9 | Viewed by 3135
Abstract
In the framework of palaeography, the availability of both effective image analysis algorithms, and high-quality digital images has favored the development of new applications for the study of ancient manuscripts and has provided new tools for decision-making support systems. The quality of the [...] Read more.
In the framework of palaeography, the availability of both effective image analysis algorithms, and high-quality digital images has favored the development of new applications for the study of ancient manuscripts and has provided new tools for decision-making support systems. The quality of the results provided by such applications, however, is strongly influenced by the selection of effective features, which should be able to capture the distinctive aspects to which the paleography expert is interested in. This process is very difficult to generalize due to the enormous variability in the type of ancient documents, produced in different historical periods with different languages and styles. The effect is that it is very difficult to define standard techniques that are general enough to be effectively used in any case, and this is the reason why ad-hoc systems, generally designed according to paleographers’ suggestions, have been designed for the analysis of ancient manuscripts. In recent years, there has been a growing scientific interest in the use of techniques based on deep learning (DL) for the automatic processing of ancient documents. This interest is not only due to their capability of designing high-performance pattern recognition systems, but also to their ability of automatically extracting features from raw data, without using any a priori knowledge. Moving from these considerations, the aim of this study is to verify if DL-based approaches may actually represent a general methodology for automatically designing machine learning systems for palaeography applications. To this purpose, we compared the performance of a DL-based approach with that of a “classical” machine learning one, in a particularly unfavorable case for DL, namely that of highly standardized schools. The rationale of this choice is to compare the obtainable results even when context information is present and discriminating: this information is ignored by DL approaches, while it is used by machine learning methods, making the comparison more significant. The experimental results refer to the use of a large sets of digital images extracted from an entire 12th-century Bibles, the “Avila Bible”. This manuscript, produced by several scribes who worked in different periods and in different places, represents a severe test bed to evaluate the efficiency of scribe identification systems. Full article
(This article belongs to the Special Issue Recent Advances in Historical Document Processing)
Show Figures

Figure 1

12 pages, 5699 KiB  
Article
Marching Cubes and Histogram Pyramids for 3D Medical Visualization
by Porawat Visutsak
J. Imaging 2020, 6(9), 88; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090088 - 03 Sep 2020
Cited by 3 | Viewed by 4105
Abstract
This paper aims to implement histogram pyramids with marching cubes method for 3D medical volumetric rendering. The histogram pyramids are used for feature extraction by segmenting the image into the hierarchical order like the pyramid shape. The histogram pyramids can decrease the number [...] Read more.
This paper aims to implement histogram pyramids with marching cubes method for 3D medical volumetric rendering. The histogram pyramids are used for feature extraction by segmenting the image into the hierarchical order like the pyramid shape. The histogram pyramids can decrease the number of sparse matrixes that will occur during voxel manipulation. The important feature of the histogram pyramids is the direction of segments in the image. Then this feature will be used for connecting pixels (2D) to form up voxel (3D) during marching cubes implementation. The proposed method is fast and easy to implement and it also produces a smooth result (compared to the traditional marching cubes technique). The experimental results show the time consuming for generating 3D model can be reduced by 15.59% in average. The paper also shows the comparison between the surface rendering using the traditional marching cubes and the marching cubes with histogram pyramids. Therefore, for the volumetric rendering such as 3D medical models and terrains where a large number of lookups in 3D grids are performed, this method is a particularly good choice for generating the smooth surface of 3D object. Full article
(This article belongs to the Special Issue Robust Image Processing)
Show Figures

Figure 1

23 pages, 11781 KiB  
Article
Spatial Spectral Band Selection for Enhanced Hyperspectral Remote Sensing Classification Applications
by Ruben Moya Torres, Peter W.T. Yuen, Changfeng Yuan, Johathan Piper, Chris McCullough and Peter Godfree
J. Imaging 2020, 6(9), 87; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090087 - 31 Aug 2020
Cited by 7 | Viewed by 3157
Abstract
Despite the numerous band selection (BS) algorithms reported in the field, most if not all have exhibited maximal accuracy when more spectral bands are utilized for classification. This apparently disagrees with the theoretical model of the ‘curse of dimensionality’ phenomenon, without apparent explanations. [...] Read more.
Despite the numerous band selection (BS) algorithms reported in the field, most if not all have exhibited maximal accuracy when more spectral bands are utilized for classification. This apparently disagrees with the theoretical model of the ‘curse of dimensionality’ phenomenon, without apparent explanations. If it were true, then BS would be deemed as an academic piece of research without real benefits to practical applications. This paper presents a spatial spectral mutual information (SSMI) BS scheme that utilizes a spatial feature extraction technique as a preprocessing step, followed by the clustering of the mutual information (MI) of spectral bands for enhancing the efficiency of the BS. Through the SSMI BS scheme, a sharp ’bell’-shaped accuracy-dimensionality characteristic that peaks at about 20 bands has been observed for the very first time. The performance of the proposed SSMI BS scheme has been validated through 6 hyperspectral imaging (HSI) datasets (Indian Pines, Botswana, Barrax, Pavia University, Salinas, and Kennedy Space Center (KSC)), and its classification accuracy is shown to be approximately 10% better than seven state-of-the-art BS schemes (Saliency, HyperBS, SLN, OCF, FDPC, ISSC, and Convolution Neural Network (CNN)). The present result confirms that the high efficiency of the BS scheme is essentially important to observe and validate the Hughes’ phenomenon in the analysis of HSI data. Experiments also show that the classification accuracy can be affected by as much as approximately 10% when a single ‘crucial’ band is included or missed out for classification. Full article
Show Figures

Figure 1

21 pages, 5596 KiB  
Article
Realistic Dynamic Numerical Phantom for MRI of the Upper Vocal Tract
by Joe Martin, Matthieu Ruthven, Redha Boubertakh and Marc E. Miquel
J. Imaging 2020, 6(9), 86; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090086 - 27 Aug 2020
Cited by 3 | Viewed by 4052
Abstract
Dynamic and real-time MRI (rtMRI) of human speech is an active field of research, with interest from both the linguistics and clinical communities. At present, different research groups are investigating a range of rtMRI acquisition and reconstruction approaches to visualise the speech organs. [...] Read more.
Dynamic and real-time MRI (rtMRI) of human speech is an active field of research, with interest from both the linguistics and clinical communities. At present, different research groups are investigating a range of rtMRI acquisition and reconstruction approaches to visualise the speech organs. Similar to other moving organs, it is difficult to create a physical phantom of the speech organs to optimise these approaches; therefore, the optimisation requires extensive scanner access and imaging of volunteers. As previously demonstrated in cardiac imaging, realistic numerical phantoms can be useful tools for optimising rtMRI approaches and reduce reliance on scanner access and imaging volunteers. However, currently, no such speech rtMRI phantom exists. In this work, a numerical phantom for optimising speech rtMRI approaches was developed and tested on different reconstruction schemes. The novel phantom comprised a dynamic image series and corresponding k-space data of a single mid-sagittal slice with a temporal resolution of 30 frames per second (fps). The phantom was developed based on images of a volunteer acquired at a frame rate of 10 fps. The creation of the numerical phantom involved the following steps: image acquisition, image enhancement, segmentation, mask optimisation, through-time and spatial interpolation and finally the derived k-space phantom. The phantom was used to: (1) test different k-space sampling schemes (Cartesian, radial and spiral); (2) create lower frame rate acquisitions by simulating segmented k-space acquisitions; (3) simulate parallel imaging reconstructions (SENSE and GRAPPA). This demonstrated how such a numerical phantom could be used to optimise images and test multiple sampling strategies without extensive scanner access. Full article
Show Figures

Graphical abstract

17 pages, 5285 KiB  
Article
Efficient Deconvolution Architecture for Heterogeneous Systems-on-Chip
by Stefania Perri, Cristian Sestito, Fanny Spagnolo and Pasquale Corsonello
J. Imaging 2020, 6(9), 85; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090085 - 25 Aug 2020
Cited by 6 | Viewed by 2573
Abstract
Today, convolutional and deconvolutional neural network models are exceptionally popular thanks to the impressive accuracies they have been proven in several computer-vision applications. To speed up the overall tasks of these neural networks, purpose-designed accelerators are highly desirable. Unfortunately, the high computational complexity [...] Read more.
Today, convolutional and deconvolutional neural network models are exceptionally popular thanks to the impressive accuracies they have been proven in several computer-vision applications. To speed up the overall tasks of these neural networks, purpose-designed accelerators are highly desirable. Unfortunately, the high computational complexity and the huge memory demand make the design of efficient hardware architectures, as well as their deployment in resource- and power-constrained embedded systems, still quite challenging. This paper presents a novel purpose-designed hardware accelerator to perform 2D deconvolutions. The proposed structure applies a hardware-oriented computational approach that overcomes the issues of traditional deconvolution methods, and it is suitable for being implemented within any virtually system-on-chip based on field-programmable gate array devices. In fact, the novel accelerator is simply scalable to comply with resources available within both high- and low-end devices by adequately scaling the adopted parallelism. As an example, when exploited to accelerate the Deep Convolutional Generative Adversarial Network model, the novel accelerator, running as a standalone unit implemented within the Xilinx Zynq XC7Z020 System-on-Chip (SoC) device, performs up to 72 GOPs. Moreover, it dissipates less than 500mW@200MHz and occupies 5.6%, 4.1%, 17%, and 96%, respectively, of the look-up tables, flip-flops, random access memory, and digital signal processors available on-chip. When accommodated within the same device, the whole embedded system equipped with the novel accelerator performs up to 54 GOPs and dissipates less than 1.8W@150MHz. Thanks to the increased parallelism exploitable, more than 900 GOPs can be executed when the high-end Virtex-7 XC7VX690T device is used as the implementation platform. Moreover, in comparison with state-of-the-art competitors implemented within the Zynq XC7Z045 device, the system proposed here reaches a computational capability up to 20% higher, and saves more than 60% and 80% of power consumption and logic resources requirement, respectively, using 5.7× fewer on-chip memory resources. Full article
Show Figures

Figure 1

23 pages, 4384 KiB  
Article
An Efficient and Lightweight Illumination Model for Planetary Bodies Including Direct and Diffuse Radiation
by Marco Scharringhausen and Lars Witte
J. Imaging 2020, 6(9), 84; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090084 - 24 Aug 2020
Viewed by 2406
Abstract
We present a numerical illumination model to calculate direct as well as diffuse or Hapke scattered radiation scenarios on arbitrary planetary surfaces. This includes small body surfaces such as main belt asteroids as well as e.g., the lunar surface. The model is based [...] Read more.
We present a numerical illumination model to calculate direct as well as diffuse or Hapke scattered radiation scenarios on arbitrary planetary surfaces. This includes small body surfaces such as main belt asteroids as well as e.g., the lunar surface. The model is based on the ray tracing method. This method is not restricted to spherical or ellipsoidal shapes but digital terrain data of arbitrary spatial resolution can be fed into the model. Solar radiation is the source of direct radiation, wavelength-dependent effects (e.g. albedo) can be accounted for. Mutual illumination of individual bodies in implemented (e.g. in binary or multiple systems) as well as self-illumination (e.g. crater floors by crater walls) by diffuse or Hapke radiation. The model is validated by statistical methods. A χ2 test is utilized to compare simulated images with DAWN images acquired during the survey phase at small body 4 Vesta and to successfully prove its validity. Full article
Show Figures

Figure 1

15 pages, 4051 KiB  
Article
Investigating the Performance of Generative Adversarial Networks for Prostate Tissue Detection and Segmentation
by Ufuk Cem Birbiri, Azam Hamidinekoo, Amélie Grall, Paul Malcolm and Reyer Zwiggelaar
J. Imaging 2020, 6(9), 83; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090083 - 24 Aug 2020
Cited by 11 | Viewed by 4850
Abstract
The manual delineation of region of interest (RoI) in 3D magnetic resonance imaging (MRI) of the prostate is time-consuming and subjective. Correct identification of prostate tissue is helpful to define a precise RoI to be used in CAD systems in clinical practice during [...] Read more.
The manual delineation of region of interest (RoI) in 3D magnetic resonance imaging (MRI) of the prostate is time-consuming and subjective. Correct identification of prostate tissue is helpful to define a precise RoI to be used in CAD systems in clinical practice during diagnostic imaging, radiotherapy and monitoring the progress of disease. Conditional GAN (cGAN), cycleGAN and U-Net models and their performances were studied for the detection and segmentation of prostate tissue in 3D multi-parametric MRI scans. These models were trained and evaluated on MRI data from 40 patients with biopsy-proven prostate cancer. Due to the limited amount of available training data, three augmentation schemes were proposed to artificially increase the training samples. These models were tested on a clinical dataset annotated for this study and on a public dataset (PROMISE12). The cGAN model outperformed the U-Net and cycleGAN predictions owing to the inclusion of paired image supervision. Based on our quantitative results, cGAN gained a Dice score of 0.78 and 0.75 on the private and the PROMISE12 public datasets, respectively. Full article
(This article belongs to the Special Issue MIUA2019)
Show Figures

Figure 1

13 pages, 4237 KiB  
Article
Detection of HER2 from Haematoxylin-Eosin Slides Through a Cascade of Deep Learning Classifiers via Multi-Instance Learning
by David La Barbera, António Polónia, Kevin Roitero, Eduardo Conde-Sousa and Vincenzo Della Mea
J. Imaging 2020, 6(9), 82; https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging6090082 - 23 Aug 2020
Cited by 16 | Viewed by 4586
Abstract
Breast cancer is the most frequently diagnosed cancer in woman. The correct identification of the HER2 receptor is a matter of major importance when dealing with breast cancer: an over-expression of HER2 is associated with aggressive clinical behaviour; moreover, HER2 targeted therapy results [...] Read more.
Breast cancer is the most frequently diagnosed cancer in woman. The correct identification of the HER2 receptor is a matter of major importance when dealing with breast cancer: an over-expression of HER2 is associated with aggressive clinical behaviour; moreover, HER2 targeted therapy results in a significant improvement in the overall survival rate. In this work, we employ a pipeline based on a cascade of deep neural network classifiers and multi-instance learning to detect the presence of HER2 from Haematoxylin–Eosin slides, which partly mimics the pathologist’s behaviour by first recognizing cancer and then evaluating HER2. Our results show that the proposed system presents a good overall effectiveness. Furthermore, the system design is prone to further improvements that can be easily deployed in order to increase the effectiveness score. Full article
(This article belongs to the Special Issue Deep Learning in Medical Image Analysis)
Show Figures

Figure 1

Previous Issue
Next Issue
Back to TopTop