Next Article in Journal
Object Detection for Agricultural Vehicles: Ensemble Method Based on Hierarchy of Classes
Previous Article in Journal
Research on Speech Synthesis Based on Mixture Alignment Mechanism
Previous Article in Special Issue
Failure Severity Prediction for Protective-Coating Disbondment via the Classification of Acoustic Emission Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Time-Distributed Framework for 3D Reconstruction Integrating Fringe Projection with Deep Learning

by
Andrew-Hieu Nguyen
1 and
Zhaoyang Wang
2,*
1
Neuroimaging Research Branch, National Institute on Drug Abuse, National Institutes of Health, Baltimore, MD 21224, USA
2
Department of Mechanical Engineering, The Catholic University of America, Washington, DC 20064, USA
*
Author to whom correspondence should be addressed.
Submission received: 21 June 2023 / Revised: 7 August 2023 / Accepted: 18 August 2023 / Published: 20 August 2023
(This article belongs to the Special Issue Intelligent Sensing and Automatic Device for Industrial Process)

Abstract

:
In recent years, integrating structured light with deep learning has gained considerable attention in three-dimensional (3D) shape reconstruction due to its high precision and suitability for dynamic applications. While previous techniques primarily focus on processing in the spatial domain, this paper proposes a novel time-distributed approach for temporal structured-light 3D shape reconstruction using deep learning. The proposed approach utilizes an autoencoder network and time-distributed wrapper to convert multiple temporal fringe patterns into their corresponding numerators and denominators of the arctangent functions. Fringe projection profilometry (FPP), a well-known temporal structured-light technique, is employed to prepare high-quality ground truth and depict the 3D reconstruction process. Our experimental findings show that the time-distributed 3D reconstruction technique achieves comparable outcomes with the dual-frequency dataset ( p = 0.014) and higher accuracy than the triple-frequency dataset ( p = 1.029 × 10 9 ), according to non-parametric statistical tests. Moreover, the proposed approach’s straightforward implementation of a single training network for multiple converters makes it more practical for scientific research and industrial applications.

1. Introduction

Three-dimensional (3D) reconstruction, a subfield within computer vision, has gained exceptional popularity as a measurement tool in recent decades, owing to its inherent advantages in capturing real-world objects’ visual appearance and geometric shape. The process of 3D reconstruction involves using computer vision algorithms and image processing techniques to analyze a set of representative two-dimensional (2D) images and generate a 3D digital point-cloud model of an object or scene. The demand for 3D shape reconstruction is evident across many applications in various fields such as vision-guided robots, visual inspection, face recognition, autonomous navigation, medical imaging, driverless vehicles, 3D entertainment, archaeology, and gaming [1,2,3,4,5].
There are two primary categories of 3D shape reconstruction techniques: active methods and passive methods. Typical active methods encompass time-of-flight, structured light, optical interferometry, laser scanning, computed tomography, etc. [6,7,8,9,10]. On the other hand, popular passive methods comprise stereo vision, photogrammetry, shape from motion, shape from defocus, etc. [11,12,13,14,15]. Active techniques, as opposed to passive ones that solely rely on the natural texture information captured, project known patterns onto the target of interest and observe their deformation, enabling highly accurate depth measurements. Among these active methods, structured-light 3D reconstruction techniques have become increasingly popular in industrial applications due to their extraordinary accuracy and reliability. Figure 1 showcases a typical 3D reconstruction system and the utilization of illuminated patterns for precise 3D reconstruction. A key drawback of such a technique is that its measurement speed is slow when high accuracy is desired since multiple images are required and complicated computations are involved. This well-known limitation has troubled the technical community for many years, until recently when artificial intelligence (AI) provided opportunities to tackle it.
The present world is in an era of big data with tremendous amounts of information and data generated every second, presenting a considerable challenge for relevant personnel in integrating and efficiently utilizing this abundance of data. In recent years, the rise of AI has helped cope with the problem. AI technologies have empowered machines to perform tasks previously considered beyond human capabilities. Deep learning, a collection of learning algorithms and statistical models derived from AI, emulates the human brain’s cognitive processes in acquiring knowledge. It encompasses two primary approaches: supervised learning and unsupervised learning [16,17]. While unsupervised learning has gained recent attention and demonstrated promising results in various domains (e.g., object recognition, image segmentation, anomaly detection, image retrieval, image compression, image generation, etc.) [18,19,20,21], supervised learning remains pivotal in most deep learning work and applications. Crucial factors contributing to the extensive utilization of supervised learning include the availability of large-scale labeled datasets, task-specific learning, higher performance, broader applications, and higher interpretability. Advances in technology have facilitated the collection and annotation of massive amounts of data from various sources. These labeled datasets enable deep learning models to discern complex patterns and exhibit strong generalization capabilities when faced with new and unseen examples.
One of the most significant impacts of deep learning techniques has been in the field of computer vision. Incorporating deep learning has greatly influenced 3D reconstruction methods, leading to substantial advancements. Leveraging its ability to comprehend intricate patterns and representations from extensive datasets, deep learning has brought a transformative shift in 3D reconstruction. Its application spans different phases of the reconstruction workflow, encompassing fundamental feature learning and more complex tasks such as dense 3D reconstruction, shape completion, surface reconstruction, and single-view and multi-view reconstruction. Deep learning techniques can potentially optimize the efficiency of the process, enabling real-time or high-speed 3D reconstruction at a super-resolution level [22,23,24]. Various output representations can be employed in deep learning techniques for 3D object reconstruction, including volumetric representations, surface-based representations, and intermediate representations [25]. Sun et al. introduced a NeuralRecon framework for real-time scene reconstruction using a learning-based TSDF fusion module [26]. Additionally, Zhao et al. proposed a method that can accelerate the 3D reconstruction up to 10 Hz using fully-connected conditional random fields model [27]. To address computational cost and memory efficiency issues, a method named occupancy networks proposed a new representation for 3D output training with reduced memory footprint [28]. Three-dimensional reconstruction via deep learning has also found key applications in augmented reality (AR) tasks. For instance, Park et al. developed a smart and user-centric task assistance method that combines instance segmentation and deep learning-based object detection to reconstruct 2.5D and 3D replicas in wearable AR smart glasses [29]. In addition, 3D reconstruction through deep learning has been applied in various indoor mapping applications using mobile devices [30,31,32].
Deep learning has also emerged as an AI-assisted tool in the field of experimental mechanics and metrology, where precision is vital. It simplifies traditional techniques while ensuring consistent accuracy and allows real-time or high-speed measurements. In recent years, there has been a growing interest in integrating deep learning with the aforementioned structured-light technique, which is popular in a few fields, including optics, experimental mechanics, metrology, and computer vision, to achieve accurate 3D shape measurement and 3D reconstruction. This combination can substantially simplify and enhance conventional techniques while maintaining stable accuracy [33,34,35,36,37]. It holds promise for numerous scientific and engineering applications where accurate and efficient 3D reconstruction is paramount.
Among various structured light techniques, fringe projection profilometry (FPP) is the most widely used technique in combination with the deep learning method for 3D reconstruction [38,39,40,41,42]. The integrated approaches can be broadly categorized into fringe-to-depth and fringe-to-phase techniques. In the fringe-to-depth approach, a direct conversion of the captured fringe pattern(s) to the desired depth information is accomplished using convolutional neural networks (CNNs). This process is analogous to the image-to-image transformation in computer vision applications. By training CNN models on appropriate datasets, the fringe patterns can be effectively mapped to corresponding depth values, enabling accurate 3D reconstruction [43,44,45,46,47,48]. On the other hand, the fringe-to-phase approach exploits the multi-stage nature of the FPP. It involves transforming the fringe pattern(s) into intermediate results, which ultimately enable the acquisition of precise phase distributions. These phase distributions and camera calibration information are then utilized to achieve accurate 3D reconstruction [49,50,51,52,53,54,55].
In general, the fringe-to-phase approaches tend to yield more detailed 3D reconstruction results than the fringe-to-depth counterpart. This is primarily attributed to its incorporating additional phase calculations and utilizing parameter information obtained through camera calibration. Over the past few years, fringe-to-phase approaches, which focus on obtaining precise unwrapped phase distributions, have undergone notable developments in several aspects. These advancements include the employment of single or multiple input(s)/output(s), the introduction of reference planes, the implementation of multi-stage networks, the utilization of combined red-green-blue (RGB) color fringe images, and the use of coded patterns, among others [56,57,58]. Regardless of the specific variations, it is evident that the integration primarily relies on choosing single or multiple inputs. The subsequent training of the network(s) and the output(s) definition can be determined based on the researcher’s preferences and interests. In addition to several advanced fringe-to-phase techniques that utilize single-shot input and single network, alternative deep learning-based approaches have employed multi-shot inputs with multi-stage networks [59,60,61]. As an example, Yu et al. [62] introduced a concept where a single or two fringe patterns are transformed into multiple phase-shifted fringe patterns using multiple FTPNet networks. Liang et al. [63] utilized a similar autoencoder-based network in a two-step training process to derive the unwrapped phase from the segmented wrapped phase. In other studies, the researchers [57,64] employed two subnetworks with cosine fringe pattern and multi-code/reference pattern to obtain the wrapped phase and fringe orders. The work reported in [65,66] followed a framework comprising two deep neural networks, aiming to enhance the quality of the fringe pattern and accurately determine the numerator and denominator through denoising patterns. Machineni et al. [67] presented an end-to-end deep learning-based framework for 3D object profiling, and the method encompassed a two-stage process involving a synthesis network and a phase estimation network. Its notable drawbacks and limitations include the need for multiple training hardwares, extended training duration, a higher number of learning parameters, and a sequential process.
Drawing upon the advancements in single-shot 3D reconstruction techniques and recognizing the limitations of multi-stage multi-shot approaches, this paper presents a proof-of-concept 3D reconstruction method. The proposed approach utilizes a single network and employs a time-distributed wrapper to handle multiple inputs. The technique employs a time-distributed framework to convert multiple fringe images into intermediate results of numerators and denominators of arctangent functions, enabling the subsequent acquisition of phase distributions and 3D shape information. Unlike stacking multiple inputs and outputs in the spatial domain of the training vector, the proposed approach encodes multiple inputs and their corresponding outputs into temporal slices of the training vector. Similar to training and prediction using the spatial vector, the proposed framework can predict the intermediate results for unseen objects once the training process is successfully completed.
It should be emphasized that the classic FPP technique serves a dual purpose in this study. First, it prepares training data with ground-truth labels for the learning process. Second, it plays a crucial role in the subsequent process of obtaining the phase distributions and final 3D point cloud after the deep learning prediction. Given that the temporal FPP technique involves capturing multiple fringe images over a span of time, the proposed time-distributed framework is a well-suited approach for effectively handling and converting multiple inputs within the reconstruction process. The proposed technique brings several noteworthy contributions in comparison with previous fringe-to-phase methods:
  • It introduces a single network instead of relying on multiple subnetworks for a multi-stage process.
  • It presents a proof-of-concept 3D reconstruction approach where multiple inputs are stacked in the temporal domain vector rather than the spatial domain vector.
  • The data labeling process is simplified, with multiple inputs and corresponding outputs consolidated into a single training vector instead of separate vectors.
  • It maintains the accuracy advantages of the classic FPP method while reducing the number of required fringe patterns.
The remaining sections of this paper are structured as follows. Section 2 provides an overview of the FPP technique and presents the proposed framework for phase measurement. In Section 3, various experiments are conducted to assess the effectiveness of the proposed approach. Section 4 presents discussions and further analysis of the results, while Section 5 offers a concise summary of the proposed work.

2. Materials and Methods

The process of FPP 3D imaging involves two main steps. First, evenly spaced fringe patterns are projected onto the surface of the target, and the surface profile is encoded in the distorted fringe patterns. A camera then captures the patterns for subsequent 3D decoding. This decoding process comprises four key sub-steps: phase extraction, phase unwrapping, depth determination, and 3D reconstruction. It is worth noting that the proposed time-distributed framework specifically focuses on converting fringe-pattern images into their corresponding numerators and denominators in the phase determination function. Nevertheless, it should also be emphasized that the subsequent phase determination and 3D reconstruction still rely on the conventional FPP technique. Therefore, providing an overview of the classic FPP technique is essential before discussing the proposed time-distributed network.

2.1. Temporal Structured-Light Technique: Fringe Projection Profilometry

The temporal-based FPP technique involves projecting a series of fringe patterns onto the surface of the target object. The fringe patterns used in this technique can be described as uniform, with consistent characteristics across the entire projection:
I j i ( u , v ) = I 0 + I 0 cos ϕ i ( u , v ) + δ j
where I represents the intensity of the projected input at a specific pixel location ( u , v ) ; the subscript j denotes the order of the phase-shifted image, with j ranging from 1 to 4 in the case of a four-step phase-shifting algorithm; and superscript i implies the ith frequency. The intensity modulation is represented by the constant value I 0 , typically set to 127.5 . The fringe phase ϕ can be expressed as ϕ i ( u , v ) = 2 π f i u W , where f i corresponds to the fringe frequency defined as the number of fringes in the entire pattern, and W represents the width of the pattern. Moreover, the phase-shift amount δ is given by δ j = ( j 1 ) π 2 .
In practice, the fringe patterns captured from the synchronous camera are distinct from the generated fringe patterns and can be elaborated as follows [68]:
I j i ( u , v ) = I a ( u , v ) + I b ( u , v ) cos ϕ i ( u , v ) + δ j
where I, I a , and I b represent the pixel intensities of the captured patterns, the intensity background, and the fringe amplitude at a specific pixel location ( u , v ) . The value of ϕ i ( u , v ) can be computed using the standard phase-shifting algorithm. In this study, we utilize the four-step phase-shifting algorithm, and the determination of ϕ i w ( u , v ) is given by the following equation [69]:
ϕ i w ( u , v ) = arctan I 4 i ( u , v ) I 2 i ( u , v ) I 1 i ( u , v ) I 3 i ( u , v ) = arctan N i D i
where N and D denote the numerator and denominator of the arctangent function, respectively. Hereinafter, the pixel coordinate ( u , v ) will be omitted to streamline the subsequent equations. The result obtained from Equation (3) lies within the range of [ π , π ) , and to obtain the true phase, it is necessary to unwrap ϕ i w . In the context of FPP 3D imaging, the multi-frequency phase-shifting algorithm is widely recognized for its ability to handle geometric discontinuities and situations involving overlapping objects with varying height or depth information.
In our proposed approach, we utilize the dual-frequency four-step (DFFS) phase-shifting scheme, which involves two frequencies ( f 1 and f 2 ), as well as the triple-frequency four-step (TFFS) scheme, which incorporates three frequencies ( f 1 , f 2 , and f 3 ). These schemes are employed to obtain high-quality unwrapped phase maps and serve as the ground-truth labels for training the proposed time-distributed network.
When using the DFFS phase-shifting scheme, the unwrapped phase can be obtained by satisfying the condition f 2 f 1 = 1 . In such cases, the equations governing the unwrapped phase can be expressed as follows [70]:
ϕ 12 u w = ϕ 2 w ϕ 1 w + 0 , ϕ 2 w ϕ 1 w 2 π , ϕ 2 w < ϕ 1 w ϕ = ϕ 2 u w = ϕ 2 w + INT ϕ 12 u w f 2 ϕ 2 w 2 π 2 π
where ϕ 1 w and ϕ 2 w are the wrapped phases of two frequencies f 1 and f 2 , respectively. The initial unwrapped phase, ϕ 12 u w , is derived from the pattern with only one fringe. However, due to the noise caused by the frequency mismatch between f 1 and f 2 , ϕ 12 u w cannot be directly used. Instead, it serves as the interfering unwrapped phase for the hierarchical phase-unwrapping process of ϕ 2 u w . The final unwrapped phase, denoted as ϕ , corresponds to the phase distribution of the highest fringe frequency. This study utilizes two frequencies, f 1 = 79 and f 2 = 80 , in accordance with the requirements of the DFFS scheme. Figure 2a illustrates the flowchart of the DFFS phase-shifting scheme.
In the TFFS scheme, as depicted in Figure 2b, if the three frequencies fulfill the condition ( f 3 f 2 ) ( f 2 f 1 ) = 1 , where ( f 3 f 2 ) > ( f 2 f 1 ) > 0 , the unwrapped phase of the fringe patterns with the highest frequency can be computed using the following hierarchical equations [71,72]:
ϕ 12 w = ϕ 2 w ϕ 1 w + 0 ϕ 2 w ϕ 1 w 2 π ϕ 2 w < ϕ 1 w ϕ 23 w = ϕ 3 w ϕ 2 w + 0 ϕ 3 w ϕ 2 w 2 π ϕ 3 w < ϕ 2 w ϕ 123 = ϕ 23 w ϕ 12 w + 0 ϕ 23 w ϕ 12 w 2 π ϕ 23 w < ϕ 12 w ϕ 23 = ϕ 23 w + INT ϕ 123 ( f 3 f 2 ) ϕ 23 w 2 π 2 π ϕ = ϕ 3 u w = ϕ 3 w + INT ϕ 23 f 3 f 3 f 2 ϕ 3 w 2 π 2 π
where ϕ with superscript w and u w are the wrapped phase and unwrapped phase, respectively. The function “INT” rounds the value to the nearest integer. The term ϕ m n represents the difference between ϕ m and ϕ n , where ( f n f m ) corresponds to the number of wrapped fringes in the phase map. The algorithm’s core principle is based on the fact that ϕ 123 is both wrapped and unwrapped due to the presence of only one fringe in the pattern. This property enables a hierarchical phase-unwrapping process that connects ϕ 123 and ϕ 3 through ϕ 23 . The phase distribution of the highest-frequency fringe patterns, ϕ 3 , is utilized for the final phase determination as it provides the highest level of accuracy. In the TFFS scheme, the chosen frequencies are 61, 70, and 80. These specific frequencies were selected to maintain a balanced hierarchical calculation with a ratio of 1:10:80.
Ultimately, the FPP 3D imaging technique is employed to directly reconstruct the height/depth information from the unwrapped phase obtained from Equation (4) or Equation (5). The equation governing the retrieval of the depth map from ϕ can be derived as described in [69]:
z = c P 1 P 2 d P 1 P 2 c = 1 c 1 c 2 c 3 c 17 c 18 c 19 d = d 0 d 1 d 2 d 3 c 17 d 18 d 19 P 1 = 1 ϕ u u ϕ v v ϕ u 2 u 2 ϕ u v u v ϕ v 2 v 2 ϕ P 2 = u 3 u 3 ϕ u 2 v u 2 v ϕ u v 2 u v 2 ϕ v 3 v 3 ϕ .
The equation for determining the height or depth value z at a specific pixel coordinate ( u , v ) involves using triangulation parameters. These parameters, denoted as c 1 to c 19 and d 0 to d 19 , are obtained through a system calibration process.
This study used a set of 31 sculptures showing various surface shapes, as well as 10 objects commonly found in laboratories, including gauge block, tape measure, corded telephone, remote control, ping-pong ball, electronic charger, glue bottle, calibration board, rotary fan, and balloon [33]. Each object was arbitrarily positioned many times in the field of view to serve as multiple different targets. In addition, two or multiple objects were randomly grouped together to form new objects for the dataset generation.
The DFFS datasets consisted of a total of 2048 scenes with a resolution of 640 × 448  [39,70]. Each scene involved the projection of 8 uniform sinusoidal four-step phase-shifted images, with two frequencies of f 1 = 79 and f 2 = 80 , by the projector. Simultaneously, the camera captured 8 corresponding images. During the data labeling process, the first image of each frequency, namely I 1 79 and I 1 80 , was selected as the temporal input slices. The corresponding output of numerators and denominators, represented as N 79 , D 79 , N 80 , and D 80 , was generated using all 8 captured images and Equation (3). Figure 3a illustrates examples of the input–output pairs used for the proposed time-distributed framework with the DFFS datasets.
Likewise, the TFFS datasets consisted of 1500 data samples with the resolution of 640 × 352  [71,72], with each scene capturing a total of 12 images. These four-step phase-shifted images employed three frequencies: f 1 = 61 , f 2 = 70 , and f 3 = 80 . Figure 3b shows two examples of input–output pairs generated for the TFFS datasets.

2.2. Time-Distributed Framework for Temporal Fringe-Pattern Transformation

The primary aim of the proposed time-distributed (TD) framework remains consistent with previous fringe-to-phase approaches, focusing on the determination of phase distributions for 3D shape measurement. However, the specific goal of this framework is to showcase a proof-of-concept image-to-image conversion using deep learning techniques for the temporal FPP technique.
Time-distributed is a term commonly employed in Recurrent Neural Networks (RNNs) or sequence-to-sequence models, where it is utilized in the context of sequential data, such as a sequence of images. In the context of the temporal FPP technique, which involves multiple fringe patterns captured at different time steps, the time-distributed concept allows using the same network parameters (weights and biases) to process each individual input separately. This ensures that the network can extract consistent features, such as phase-shifted information, from each time step while facilitating the learning of temporal dependencies, such as consecutive frequencies.
Figure 4a and Figure 5a present the workflow of the proposed TD framework, which is specifically designed for converting sequential fringe-to-phase data. The goal of this framework is to train the model to convert the given fringe patterns into their corresponding phase-shifted information, namely the numerators and denominators. However, unlike the conventional approach that combines all spatial and temporal information in the spatial domain, as depicted in Figure 4c and Figure 5b, the TD framework differentiates and distributes the spatial and temporal information into two distinct learning concepts. The first concept involves extracting features, such as performing the fringe-to-ND (F2ND) conversion, for each individual frame within the time steps. This is illustrated by each row or the horizontal direction in the figures. The second concept focuses on applying the same feature extraction process to consecutive temporal frequencies represented in the vertical direction. By segregating and distributing the spatial and temporal information in this manner, the TD framework enables effective and efficient learning of the desired features.
In this study, the TD framework utilizes a widely used network architecture called UNet for image-to-image conversion [73]. The network consists of an encoder and a decoder path with symmetric concatenation for accurate feature transformation. The encoder path employs ten convolution layers and four max-pooling layers, reducing the resolution but increasing the filter depth. The decoder path includes eight convolution layers and four transposed convolution layers, enriching the input feature maps to higher resolution while decreasing the filter depths. A 1 × 1 convolution layer at the end of the decoder path leads to the numerator and denominator outputs. The proposed framework employs a linear activation function and mean-squared error (MSE) loss for training, considering the continuous nature of the output variables. Details of the network architecture are explained in detail in our previous works [70,71,72].
In Figure 4 and Figure 5, the TD framework utilizes a single network, where the same weights and biases are applied for feature extraction across the temporal slices. The dashed line in these figures represents the TD concept. Two approaches for implementing the TD concept in the deep learning network are introduced: TD Layer and TD Module. In the TD Layer approach, the TD wrapper is applied to each layer of the learning model, as shown in Figure 4a and Figure 5a. The TD wrapper encapsulates the entire network model in the TD Module approach, as depicted in Figure 4b. Although the F2ND conversion task remains the same, it is valuable to investigate the framework’s performance using different implementations. In Keras implementation, the TD Layer and TD Module can be better understood through the following examples:
  • TD Layer
    output = keras.layers.TimeDistributed(keras.layers.Conv2D(…))(input)
  • TD Module
    module = keras.Model(network_input, network_output)
    output = keras.layers.TimeDistributed(module)(input)
To compare the performance of the framework with previous methods using the spatial domain, a popular spatial F2ND approach is employed, where all the input and output data are organized in the spatial slices, as shown in Figure 4c and Figure 5b. The input–output pair selected for this framework is a commonly used combination in the field. The input consists of consecutive fringe patterns captured at different time steps, each with a distinct frequency. The corresponding output comprises the numerators and denominators associated with these fringe patterns.
The preparation of multidimensional data format for the TD network differs from that of a regular spatial convolution network. In the TD network, the input, output, and internal hidden layers are represented as five-dimensional tensors with shapes s , t , h , w , c , where s indicates the number of data samples, t denotes the timeframe of each different frequency, h and w represent the height and width of the input, output, or feature maps at the sub-scale resolution layer, respectively, and c is the channel or filter depth. In this study, t is set as 2 and 3 for the DFFS and TFFS schemes, respectively. Moreover, c is set to 1 for the input of a single grayscale image and 2 for the output of the numerator and denominator at each timestep. Clear visualization of this multidimensional data is explained in detail and depicted in Figure 4 and Figure 5.
Hyperparameter tunning: The convolution layers are employed with a LeakyReLU function, introducing a small negative coefficient of 0.1 to address the zero-gradient problem. Additionally, a dropout function with a rate of 0.2 is incorporated between the encoder and the two decoder paths to enhance robustness. The model is trained for 1000 epochs with a mini-batch size of 2, using the Adam optimizer with an initial learning rate of 0.0001 for the first 800 epochs. Afterward, a step decay schedule is implemented to gradually reduce the learning rate for better convergence [74]. To prevent overfitting, various data augmentation techniques, including ZCA whitening, brightness, and contrast augmentation, are employed. During training, the mean squared error (MSE) is used as the evaluation metric, and Keras callbacks like History and ModelCheckpoint are utilized to monitor training progress and save the best model.

3. Experiments and Results

The performance of the proposed TD framework was evaluated through a range of quantitative and qualitative analyses. Firstly, the quantitative assessment included using two image quality metrics, namely Structural Similarity Index Measure (SSIM) and Peak Signal-to-Noise Ratio (PSNR), to evaluate the predicted numerators and denominators. Additionally, four error metrics and three accuracy metrics were employed to verify the depth accuracy of the proposed technique. Secondly, qualitative comparisons were made by visually examining the 3D shape reconstructions of test objects generated using the TD Layer, TD Module, and a comparable F2ND approach. These analyses provided a comprehensive evaluation of the performance of the proposed TD framework.
The datasets were captured using an RVBUST RVC-X mini 3D camera (Figure 1b), which provides an ideal camera–projector–target triangulation setup. The training process utilized multiple GPU nodes available in the Biowulf cluster of the High-Performance Computing group at the National Institutes of Health. The main GPUs used were 4 × NVIDIA A100 GPUs with 80 GB VRAM and 4 × NVIDIA V100-SXM2 GPUs with 32 GB VRAM. To optimize performance, Nvidia CUDA Toolkit 11.2.2 and cuDNN v8.1.0.77 were installed on these units. The network architecture was constructed using TensorFlow v2.8.2 and Keras v2.8.0, popular open-source deep learning frameworks and Python libraries known for their user-friendly nature.

3.1. Quantitative Evaluation of TD Layer, TD Module, and Spatial F2ND in DFFS and TFFS Datasets

Upon the completion of training in the TD framework, the predicted numerators and denominators are further processed using the classic FPP technique to derive the unwrapped phase distributions and 3D depth/shape information. It is important to note that the TD framework’s primary task is converting fringe patterns to their corresponding numerators or denominators, also known as the F2ND conversion or image-to-image conversion. To quantitatively evaluate the accuracy of the reconstructed numerators and denominators, SSIM and PSNR metrics were utilized. These metrics provide valuable insights into the similarity and fidelity of the reconstructed results, enabling a quantitative evaluation of the performance of the TD framework.
Figure 6 showcases the predicted output of an unseen test object utilizing the DFFS datasets, accompanied by the corresponding evaluation metrics. Upon careful examination, it may initially appear challenging to visually discern any noticeable disparities between the predicted numerators/denominators and the ground truth counterparts. However, an in-depth analysis of the structural similarity index (SSIM), ranging from 0.998 to 1.000, and the peak signal-to-noise ratio (PSNR), which consistently hovers around 40, provides valuable insights. These metrics collectively suggest that the reconstructed images resemble the reference ground-truth images, affirming their high degree of fidelity and accuracy. The TD framework demonstrates comparable performance to the spatial F2ND approach, confirming its effectiveness in capturing spatial information for accurate predictions.
The depth measurement accuracy is an essential quantitative measure for evaluating the FPP 3D imaging technique. In this study, various error and accuracy metrics commonly employed for assessing monocular depth reconstruction are utilized. These metrics are calculated by comparing the predicted depth map with the ground-truth depth map. The proposed TD Layer, TD Module, and the spatial F2ND approach are subjected to quantitative evaluation using these metrics in both DFFS and TFFS datasets. The evaluation encompasses four error metrics and three accuracy metrics, which provide a comprehensive assessment of the performance of the different approaches:
  • Absolute relative error (rel): 1 n i = 1 n z i ^ z i z i ^
  • Root-mean-square error (rms): 1 n i = 1 n z i ^ z i 2
  • Average l o g 10 error (log): 1 n i = 1 n l o g 10 ( z i ^ ) l o g 10 ( z i )
  • Root-mean-square log error (rms log):
    1 n i = 1 n l o g 10 ( z i ^ ) l o g 10 ( z i )
  • Threshold accuracy: δ = ( z i ^ z i , z i z i ^ ) < t h r ;
    t h r 1.25 , 1.25 2 , 1.25 3
where z i ^ and z i represent the ground-truth depth determined in Equation (6) and the predicted depth at valid ith pixel, respectively. The key quantitative analyses are presented in Table 1. Upon examining the DFFS datasets, it is evident that the spatial F2ND approach demonstrates slightly superior performance compared with the proposed TD Layer and TD Module approaches. Nevertheless, the differences in performance are negligible as all the metrics exhibit similar values. Notably, the TD Layer and TD Module approaches outperform the spatial F2ND approach in the TFFS datasets, as observed in the error and accuracy metrics. These quantitative metrics provide evidence that the proposed techniques not only serve as a proof of concept but also yield comparable or slightly improved results compared with the state-of-the-art techniques used in previous studies.
To ascertain the distinctions among the proposed TD Layer, TD Module, and spatial F2ND approaches in terms of accuracy, additional statistical analyses were performed. The non-parametric Kruskal–Wallis H-test was selected for this task, utilizing the mean absolute error (MAE) values as test samples. These MAE values represent the disparities between the ground-truth depths and the predicted depths generated by each approach (TD Layer, TD Module, and spatial F2ND).
The outcomes of the Kruskal–Wallis H-test revealed significant error differences among the three groups for both the DFFS dataset (H = 8.532, p = 0.014) and the TFFS dataset (H = 21.144, p = 1.029 × 10 9 ). This statistical analysis provides evidence of the notable variations in accuracy between the three approaches.

3.2. 3D Reconstruction from DFFS Phase-Shifting Scheme via Time-Distributed Concept

Visual comparisons of the 3D shape surfaces were conducted to further assess the proposed techniques’ performance. The depth maps obtained from the ground truth, TD Layer, TD Module, and the comparable spatial F2ND approach were analyzed for differences. This visual evaluation provides additional insights into the accuracy and quality of the reconstructed 3D shape surfaces.
The 3D reconstruction of three different objects is showcased in Figure 7, with each object corresponding to a single scene. The first and second columns of the figure display the original image and an example input image, respectively. The subsequent four columns present the 3D reconstructions obtained from the ground truth, TD Layer, TD Module, and the comparable spatial F2ND approach. It should be noted that the scenes have been cropped and zoomed in to enhance visibility and facilitate a comparative analysis of the results. Upon visual inspection of the figure, it is evident that all three comparable techniques exhibit a high degree of similarity to the ground truth, with no significant degradation in the quality of the reconstructed results. However, a closer examination reveals that the 3D reconstruction outcomes obtained using the TD Layer exhibit a certain level of blurring, resulting in less detailed representations. Conversely, the spatial F2ND approach demonstrates more intricate joint structures in the reconstructed 3D surfaces. This observation aligns with the quantitative findings presented in Table 1, where the spatial F2ND approach demonstrates slightly superior performance.
Furthermore, the visual evaluation involves the reconstruction of scenes with multiple objects. The first two rows of Figure 8 showcase four scenes, each featuring distinct objects with varying heights and depths. It is worth mentioning that in the traditional FPP technique, obtaining continuous phase map distributions for separated objects poses a challenge due to the presence of discontinuous fringe order. The shadows in the background of the scenes provide valuable visual cues for observing the differences in depth between the objects, which contribute to the challenges associated with determining phase distributions and fringe order ambiguity. The reconstruction of the scenes reaffirms that both the TD Module and spatial F2ND approaches offer more detailed results than the TD Layer approach while maintaining overall similarity in terms of the shapes. To enhance the visibility of depth differences among the subjects, the grid pattern and view angle were adjusted during the scene reconstruction process.

3.3. 3D Reconstruction from TFFS Phase-Shifting Scheme via Time-Distributed Concept

The TFFS datasets were utilized to evaluate the efficacy and feasibility of the proposed techniques in terms of 3D reconstruction. The 3D reconstruction of various techniques for a single object is depicted in Figure 9. At first glance, the reconstructed results closely resemble the ground truth, making it challenging to discern any notable differences. The reconstructed scenes exhibit similar shapes and depth information, suggesting that these techniques can accurately capture the underlying 3D structure. However, upon closer examination, the TD Layer technique stands out for its ability to capture finer details, particularly in the contoured and concave regions of the shape. This indicates that the TD Layer approach excels in preserving intricate features, resulting in a more faithful representation of the object’s surface.
Subsequently, the 3D reconstruction process was extended to encompass four distinct unseen scenes, each featuring multiple objects. The scenes were carefully configured from various angles to accentuate the differences in depth among the objects, a characteristic that is further emphasized by the presence of shadowed backgrounds. The obtained results in Figure 10 reveal that, while some minor discrepancies and variations near the object edges are observed, the reconstructed objects’ overall shape and intricate details are largely preserved and closely resemble the ground truth 3D representations. Despite the inherent challenges associated with accurately capturing depth information and intricate object surfaces, the proposed techniques effectively capture the main features and structures, demonstrating their ability to provide reliable and faithful 3D reconstructions.

4. Discussion

This paper explores the novel concept of a time-distributed wrapper to integrate the FPP technique with deep learning, specifically focusing on the F2ND transformation. The performance of the proposed approach is evaluated through comprehensive quantitative and qualitative analyses using TFFS and DFFS datasets. These analyses encompass comparisons of image quality, depth differences, and the visual appearance of the 3D reconstructions.
Overall, the proposed TD Layer and TD Module approaches demonstrate promising performance in terms of both quantitative measures and visual assessments. While the spatial F2ND technique may show slightly better results in certain quantitative metrics, the differences are marginal. The visual comparisons reveal that the proposed TD techniques can accurately capture the shapes and depth information of the objects, although the TD Layer technique may exhibit some blurring effects. These findings indicate that the TD Layer and TD Module approaches are viable alternatives to the traditional spatial F2ND technique, offering competitive performance in 3D reconstruction tasks.
It should be noted that alternative output vectors, such as multiple phase-shifted fringe images or wrapped phases with different frequencies, can be used instead of numerators and denominators. However, recent studies [39,70,71,75] have demonstrated that the spatial F2ND approach yields similar results to the fringe-to-fringe approach while requiring less storage space due to fewer channels in the output vector. Moreover, the fringe-to-wrapped phase approach is not considered ideal as it produces inferior results compared with the spatial F2ND approach.
Despite introducing the new concept of the time-distributed wrapper for the temporal FPP technique, the manuscript also acknowledges certain drawbacks and limitations. One limitation arises from the requirement of equal depth channels in both the input and output vectors. The time-distributed network cannot be trained if the depth channels differ across different timeframes. For instance, in the DFFS dataset, the first temporal output slice includes both numerators and denominators (i.e., [s,0,h,w,0] and [s,0,h,w,1]). In contrast, the second temporal output slice only consists of a single fringe order map [75] or a single coarse map [39] (i.e., [s,0,h,w,0]), resulting in a missing channel in the second temporal output slice.
The previously mentioned limitation raises a question regarding the possibility of utilizing different output formats in the proposed approach of the TD framework. The answer is affirmative, provided that the depth channels in both the input and output vectors are consistent. Figure 11 showcases a potential application of the TD framework, where different output formats in the FPP technique are employed. The figure illustrates that the channel depth balance in the temporal slice remains at 1, utilizing either the pair of wrapped phase and fringe order or the pair of wrapped phase and coarse map. However, as stated earlier, using the wrapped phase typically leads to poor 3D reconstruction outcomes. Hence, it has been excluded from this investigation.
Although the proposed technique may not have been able to perform extensive comparisons with other well-established 3D reconstruction methods in diverse fields like image processing and computer vision, it has successfully carved out a unique niche in the narrower domain of optics and experimental mechanics. Notably, integrating the Fringe Projection technique and deep learning sets this approach apart as a novel and innovative 3D reconstruction technique, overcoming the limitations and weaknesses of previous multi-stage and multi-network approaches.
Moreover, the application of TimeDistributed Layer in this specific field is relatively scarce, highlighting the significance of our proposed technique as a pioneering example for a simple yet essential task such as image-to-image transformation. By showcasing the potential of the TimeDistributed concept, our work can inspire further exploration and adoption of this technique in various other fields, ultimately contributing to advancing 3D reconstruction and deep learning applications. One compelling application for the TimeDistributed Layer lies in reconstructing dynamic augmented reality (AR) views, incorporating time-oriented data. Leveraging the overlapping four-dimensional (4D) representations at different time viewpoints can effectively address occlusion issues in the real scene, resulting in improved and comprehensive visualizations [76,77]. Moreover, the TimeDistributed Layer shows promise in determining camera motion and pose for feature tracking in AR applications, enabling incremental motion estimates at various points in the time series [78,79]. Another intriguing use case is AR-based 3D scene reconstruction via the structure from motion (SFM) technique, which establishes relationships between different images [80,81]. These applications exemplify the versatility and potential of the TimeDistributed Layer, indicating its relevance beyond the specific field of 3D shape reconstruction.
Future research could focus on refining the TD techniques to address the minor discrepancies observed near the object edges and improve the detail level in the reconstructed 3D surfaces. Additionally, exploring the application of the proposed TD framework in other domains or extending it to handle more complex scenes with occlusions and varying lighting conditions could be valuable directions for future investigations. Exploring more advanced network models [82,83,84,85] (e.g., Attention UNet, R2U-Net, ResUNet, U 2 -Net, etc.) as alternatives to UNet for achieving even higher accuracy in shape measurement could be an exciting avenue for future research. As a preliminary step, we have conducted initial experiments with the proposed technique using the Attention UNet model, and the results have been summarized in Table 2. However, to draw definitive conclusions, a more comprehensive investigation is necessary in the future to make an accurate comparison. The preliminary findings indicate differing outcomes for the DFFS and TFFS datasets, with improved accuracy observed in the TFFS dataset, while there is a slight reduction in accuracy for the DFFS dataset.

5. Conclusions

In summary, this manuscript presents a novel time-distributed framework for 3D reconstruction by integrating fringe projection technique and deep learning. The proposed framework uses a single network and a time-distributed wrapper to convert fringe patterns to their corresponding numerators and denominators. Unlike previous approaches employing multi-stage or spatial networks, this framework utilizes the same network parameters to ensure consistent feature learning across time steps. It enables the learning of temporal dependencies among different phase-shifting frequencies. Quantitative evaluations and qualitative 3D reconstructions were conducted to validate the proposed technique, highlighting its potential for industrial applications and its contribution as a novel concept in scientific research.

Author Contributions

Conceptualization, A.-H.N.; methodology, A.-H.N. and Z.W.; software, A.-H.N. and Z.W.; validation, A.-H.N.; formal analysis, A.-H.N. and Z.W.; investigation, A.-H.N.; resources, A.-H.N.; data curation, A.-H.N. and Z.W.; writing—original draft preparation, A.-H.N. and Z.W.; writing—review and editing, A.-H.N. and Z.W.; visualization, A.-H.N.; project administration, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

This work utilized the computational resources of the NIH HPC Biowulf cluster (http://hpc.nih.gov, accessed on 15 June 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Su, X.; Zhang, Q. Dynamic 3-D shape measurement method: A review. Opt. Lasers Eng. 2010, 48, 191–204. [Google Scholar] [CrossRef]
  2. Bennani, H.; McCane, B.; Corwall, J. Three-dimensional reconstruction of In Vivo human lumbar spine from biplanar radiographs. Comput. Med. Imaging Graph. 2022, 96, 102011. [Google Scholar] [CrossRef]
  3. Huang, S.; Xu, K.; Li, M.; Wu, M. Improved Visual Inspection through 3D Image Reconstruction of Defects Based on the Photometric Stereo Technique. Sensors 2019, 19, 4970. [Google Scholar] [CrossRef]
  4. Bruno, F.; Bruno, S.; Sensi, G.; Luchi, M.; Mancuso, S.; Muzzupappa, M. From 3D reconstruction to virtual reality: A complete methodology for digital archaeological exhibition. J. Cult. Herit. 2010, 11, 42–49. [Google Scholar] [CrossRef]
  5. Nguyen, H.; Kieu, H.; Wang, Z.; Le, H.N.D. Three-dimensional facial digitization using advanced digital image correlation. Appl. Opt. 2015, 57, 2188–2196. [Google Scholar] [CrossRef]
  6. Geng, J. Structured-light 3D surface imaging: A tutorial. Adv. Opt. Photonics 2011, 3, 128–160. [Google Scholar] [CrossRef]
  7. Zhang, S. High-speed 3D shape measurement with structured light methods: A review. Opt. Lasers Eng. 2018, 106, 119–131. [Google Scholar] [CrossRef]
  8. Nguyen, H.; Ly, K.; Nguyen, T.; Wang, Y.; Wang, Z. MIMONet: Structured-light 3D shape reconstruction by a multi-input multi-output network. Appl. Opt. 2021, 60, 5134–5144. [Google Scholar] [CrossRef]
  9. Remondino, F.; El-Hakim, S. Image-based 3D Modelling: A Review. Photogramm. Rec. 2006, 21, 269–291. [Google Scholar] [CrossRef]
  10. Sansoni, G.; Trebeschi, M.; Docchio, F. State-of-The-Art and Applications of 3D Imaging Sensors in Industry, Cultural Heritage, Medicine, and Criminal Investigation. Sensors 2009, 9, 568–601. [Google Scholar] [CrossRef]
  11. Tippetts, B.; Lee, D.; Lillywhite, K.; Archibald, J. Review of stereo vision algorithms and their suitability for resource-limited systems. J. Real-Time Image Process. 2016, 11, 5–25. [Google Scholar] [CrossRef]
  12. Lazaros, N.; Sirakoulis, G.; Gasteratos, A. Review of Stereo Vision Algorithms: From Software to Hardware. Int. J. Optomechatronics 2008, 2, 435–462. [Google Scholar] [CrossRef]
  13. Lin, H.; Nie, L.; Song, Z. A single-shot structured light means by encoding both color and geometrical features. Pattern Recognit. 2016, 54, 178–189. [Google Scholar] [CrossRef]
  14. Gu, F.; Song, Z.; Zhao, Z. Single-Shot Structured Light Sensor for 3D Dense and Dynamic Reconstruction. Sensors 2020, 20, 1094. [Google Scholar] [CrossRef]
  15. Nguyen, H.; Wang, Z.; Jones, P.; Zhao, B. 3D shape, deformation, and vibration measurements using infrared Kinect sensors and digital image correlation. Appl. Opt. 2017, 56, 9030–9037. [Google Scholar] [CrossRef] [PubMed]
  16. Love, B. Comparing supervised and unsupervised category learning. Psychon. Bull. Rev. 2002, 9, 829–835. [Google Scholar] [CrossRef] [PubMed]
  17. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2016, 521, 436–444. [Google Scholar] [CrossRef]
  18. Casolla, G.; Cuomo, S.; Di Cola, V.S.; Piccialli, F. Exploring Unsupervised Learning Techniques for the Internet of Things. IEEE Trans. Industr. Inform. 2020, 16, 2621–2628. [Google Scholar] [CrossRef]
  19. Libbrecht, M.; Noble, W. Machine learning applications in genetics and genomics. Nat. Rev. Genet. 2015, 16, 321–332. [Google Scholar] [CrossRef]
  20. Hofmann, T. Exploring Unsupervised Learning Techniques for the Internet of Things. Mach. Learn. 2001, 42, 177–196. [Google Scholar] [CrossRef]
  21. Yang, Y.; Liao, Y.; Meng, G.; Lee, J. A hybrid feature selection scheme for unsupervised learning and its application in bearing fault diagnosis. Expert. Syst. Appl. 2011, 38, 11311–11320. [Google Scholar] [CrossRef]
  22. Fu, K.; Peng, J.; He, Q.; Zhang, H. Single image 3D object reconstruction based on deep learning: A review. Multimed. Tools Appl. 2020, 80, 463–498. [Google Scholar] [CrossRef]
  23. Zhang, Y.; Liu, Z.; Liu, T.; Peng, B.; Li, X. RealPoint3D: An Efficient Generation Network for 3D Object Reconstruction from a Single Image. IEEE Access 2019, 7, 57539–75749. [Google Scholar] [CrossRef]
  24. Minaee, S.; Liang, X.; Yan, S. Modern Augmented Reality: Applications, Trends, and Future Directions. arXiv 2022, arXiv:2202.09450. [Google Scholar] [CrossRef]
  25. Han, X.F.; Laga, H.; Bennamoun, M. Image-Based 3D Object Reconstruction: State-of-the-Art and Trends in the Deep Learning Era. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1578–1604. [Google Scholar] [CrossRef]
  26. Sun, J.; Xie, Y.; Chen, L.; Zhou, X.; Bao, H. NeuralRecon: Real-Time Coherent 3D Reconstruction from Monocular Video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 15593–15602. [Google Scholar] [CrossRef]
  27. Zhao, C.; Sun, L.; Stolkin, R. A fully end-to-end deep learning approach for real-time simultaneous 3D reconstruction and material recognition. In Proceedings of the 18th International Conference on Advanced Robotics (ICAR), Hong Kong, China, 10–12 July 2017; pp. 75–82. [Google Scholar] [CrossRef]
  28. Mescheder, L.; Oechsle, M.; Niemeyer, M.; Nowozin, S.; Geiger, A. Occupancy Networks: Learning 3D Reconstruction in Function Space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4455–4465. [Google Scholar] [CrossRef]
  29. Park, K.; Kim, M.; Choi, S.; Lee, J. Deep learning-based smart task assistance in wearable augmented reality. Robot. Comput. Integr. Manuf. 2020, 63, 101887. [Google Scholar] [CrossRef]
  30. Manni, A.; Oriti, D.; Sanna, A.; Pace, F.; Manuri, F. Snap2cad: 3D indoor environment reconstruction for AR/VR applications using a smartphone device. Comput. Graph. 2021, 100, 116–124. [Google Scholar] [CrossRef]
  31. Chen, J.; Kira, Z.; Cho, Y.K. Deep Learning Approach to Point Cloud Scene Understanding for Automated Scan to 3D Reconstruction. J. Comput. Civ. Eng. 2019, 33, 04019027. [Google Scholar] [CrossRef]
  32. Yang, X.; Zhuo, L.; Jiang, H.; Tang, Z.; Wang, Y.; Bao, H.; Zhang, G. Mobile3DRecon: Real-time Monocular 3D Reconstruction on a Mobile Phone. IEEE Trans. Vis. Comput. Graph. 2020, 26, 3446–3456. [Google Scholar] [CrossRef]
  33. Nguyen, H.; Wang, Y.; Wang, Z. Single-Shot 3D Shape Reconstruction Using Structured Light and Deep Convolutional Neural Networks. Sensors 2020, 20, 3718. [Google Scholar] [CrossRef]
  34. Jeught, S.; Dirckx, J. Deep neural networks for single shot structured light profilometry. Opt. Express 2019, 27, 17091–17101. [Google Scholar] [CrossRef]
  35. Fanello, S.; Rhemann, C.; Tankovich, V.; Kowdle, A.; Escolano, S.; Kim, D.; Izadi, S. Hyperdepth: Learning depth from structured light without matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5441–5450. [Google Scholar] [CrossRef]
  36. Tang, S.; Zhang, X.; Song, Z.; Song, L.; Zeng, H. Robust pattern decoding in shape-coded structured light. Opt. Lasers Eng. 2017, 96, 50–62. [Google Scholar] [CrossRef]
  37. Du, Q.; Liu, R.; Guan, B.; Pan, Y.; Sun, S. Stereo-Matching Network for Structured Light. IEEE Signal Process. Lett. 2019, 26, 164–168. [Google Scholar] [CrossRef]
  38. Yang, G.; Wang, Y. Three-dimensional measurement of precise shaft parts based on line structured light and deep learning. Measurement 2022, 191, 110837. [Google Scholar] [CrossRef]
  39. Nguyen, A.; Ly, K.; Lam, V.; Wang, Z. Generalized Fringe-to-Phase Framework for Single-Shot 3D Reconstruction Integrating Structured Light with Deep Learning. Sensors 2023, 23, 4209. [Google Scholar] [CrossRef] [PubMed]
  40. Wang, F.; Wang, C.; Guan, Q. Single-shot fringe projection profilometry based on deep learning and computer graphics. Opt. Express 2021, 29, 8024–8040. [Google Scholar] [CrossRef] [PubMed]
  41. Jia, T.; Liu, Y.; Yuan, X.; Li, W.; Chen, D.; Zhang, Y. Depth measurement based on a convolutional neural network and structured light. Meas. Sci. Technol. 2022, 33, 025202. [Google Scholar] [CrossRef]
  42. Nguyen, M.; Ghim, Y.; Rhee, H. DYnet++: A deep learning based single-shot phase-measuring deflectometry for the 3D measurement of complex free-form surfaces. IEEE Trans. Ind. Electron. 2023, 71, 2112–2121. [Google Scholar] [CrossRef]
  43. Zhu, X.; Han, Z.; Zhang, Z.; Song, L.; Wang, H.; Guo, Q. PCTNet: Depth estimation from single structured light image with a parallel CNN-transformer network. Meas. Sci. Technol. 2023, 34, 085402. [Google Scholar] [CrossRef]
  44. Ravi, V.; Gorthi, R. LiteF2DNet: A lightweight learning framework for 3D reconstruction using fringe projection profilometry. Appl. Opt. 2023, 62, 3215–3224. [Google Scholar] [CrossRef]
  45. Wang, L.; Lu, D.; Tao, J.; Qiu, R. Single-shot structured light projection profilometry with SwinConvUNet. Opt. Eng. 2022, 61, 114101. [Google Scholar] [CrossRef]
  46. Nguyen, A.; Sun, B.; Li, C.; Wang, Z. Different structured-light patterns in single-shot 2D-to-3D image conversion using deep learning. Appl. Opt. 2022, 61, 10105–10115. [Google Scholar] [CrossRef] [PubMed]
  47. Nguyen, H.; Ly, K.L.; Tran, T.; Wang, Y.; Wang, Z. hNet: Single-shot 3D shape reconstruction using structured light and h-shaped global guidance network. Results Opt. 2021, 4, 100104. [Google Scholar] [CrossRef]
  48. Nguyen, H.; Tran, T.; Wang, Y.; Wang, Z. Three-dimensional Shape Reconstruction from Single-shot Speckle Image Using Deep Convolutional Neural Networks. Opt. Lasers Eng. 2021, 143, 106639. [Google Scholar] [CrossRef]
  49. Wan, M.; Kong, L.; Peng, X. Single-Shot Three-Dimensional Measurement by Fringe Analysis Network. Photonics 2023, 10, 417. [Google Scholar] [CrossRef]
  50. Xu, M.; Zhang, Y.; Wan, Y.; Luo, L.; Peng, J. Single-Shot Multi-Frequency 3D Shape Measurement for Discontinuous Surface Object Based on Deep Learning. Photonics 2023, 14, 328. [Google Scholar] [CrossRef]
  51. Wu, Z.; Wang, J.; Jiang, X.; Fan, L.; Wei, C.; Yue, H.; Liu, Y. High-precision dynamic three-dimensional shape measurement of specular surfaces based on deep learning. Opt. Express 2023, 31, 17437–17449. [Google Scholar] [CrossRef]
  52. Liu, X.; Yang, L.; Chu, X.; Zhuo, L. A novel phase unwrapping method for binocular structured light 3D reconstruction based on deep learning. Optik 2023, 279, 170727. [Google Scholar] [CrossRef]
  53. Yu, H.; Chen, X.; Huang, R.; Bai, L.; Zheng, D.; Han, J. Untrained deep learning-based phase retrieval for fringe projection profilometry. Opt. Lasers Eng. 2023, 164, 107483. [Google Scholar] [CrossRef]
  54. Song, J.; Liu, K.; Sowmya, A.; Sun, C. Super-Resolution Phase Retrieval Network for Single-Pattern Structured Light 3D Imaging. IEEE Trans. Image. Process. 2022, 32, 537–549. [Google Scholar] [CrossRef]
  55. Nguyen, H.; Nicole, D.; Li, H.; Wang, Y.; Wang, Z. Real-time 3D shape measurement using 3LCD projection and deep machine learning. Apt. Opt. 2019, 58, 7100–7109. [Google Scholar] [CrossRef] [PubMed]
  56. Li, Y.; Qian, J.; Feng, S.; Chen, Q.; Zuo, C. Composite fringe projection deep learning profilometry for single-shot absolute 3D shape measurement. Opt. Express 2022, 30, 3424–3442. [Google Scholar] [CrossRef] [PubMed]
  57. Li, W.; Yu, J.; Gai, S.; Da, F. Absolute phase retrieval for a single-shot fringe projection profilometry based on deep learning. Opt. Eng. 2021, 60, 064104. [Google Scholar] [CrossRef]
  58. Bai, S.; Luo, X.; Xiao, K.; Tan, C.; Song, W. Deep absolute phase recovery from single-frequency phase map for handheld 3D measurement. Opt. Commun. 2022, 512, 128008. [Google Scholar] [CrossRef]
  59. Xu, M.; Zhang, Y.; Wang, N.; Luo, L.; Peng, J. Single-shot 3D shape reconstruction for complex surface objects with colour texture based on deep learning. J. Mod. Opt. 2022, 69, 941–956. [Google Scholar] [CrossRef]
  60. Dong, Y.; Yang, X.; Wu, H.; Chen, X.; Xi, J. Lightweight and edge-preserving speckle matching network for precise single-shot 3D shape measurement. Measurement 2023, 210, 112549. [Google Scholar] [CrossRef]
  61. Li, Y.; Guo, W.; Shen, J.; Wu, Z.; Zhang, Q. Motion-Induced Phase Error Compensation Using Three-Stream Neural Networks. Appl. Sci. 2022, 12, 8114. [Google Scholar] [CrossRef]
  62. Yu, H.; Chen, X.; Zhang, Z.; Zuo, C.; Zhang, Y.; Zheng, D.; Han, J. Dynamic 3-D measurement based on fringe-to-fringe transformation using deep learning. Opt. Express 2020, 28, 9405–9418. [Google Scholar] [CrossRef]
  63. Liang, J.; Zhang, J.; Shao, J.; Song, B.; Yao, B.; Liang, R. Deep Convolutional Neural Network Phase Unwrapping for Fringe Projection 3D Imaging. Sensors 2020, 20, 3691. [Google Scholar] [CrossRef]
  64. Yao, P.; Gai, S.; Chen, Y.; Chen, W.; Da, F. A multi-code 3D measurement technique based on deep learning. Opt. Lasers Eng. 2021, 143, 106623. [Google Scholar] [CrossRef]
  65. Wang, J.; Li, Y.; Ji, Y.; Qian, J.; Che, Y.; Zuo, C.; Chen, Q.; Feng, S. Deep Learning-Based 3D Measurements with Near-Infrared Fringe Projection. Sensors 2022, 22, 6469. [Google Scholar] [CrossRef] [PubMed]
  66. You, D.; Zhu, J.; Duan, Z.; You, Z.; Cheng, P. One-shot fringe pattern analysis based on deep learning image d. Opt. Eng. 2021, 60, 124113. [Google Scholar] [CrossRef]
  67. Machineni, R.; Spoorthi, G.; Vengala, K.; Gorthi, S.; Gorthi, R. End-to-end deep learning-based fringe projection framework for 3D profiling of objects. Comp. Vis. Imag. Underst. 2020, 199, 103023. [Google Scholar] [CrossRef]
  68. Nguyen, H.; Nguyen, D.; Wang, Z.; Kieu, H.; Le, M. Real-time, high-accuracy 3D imaging and shape measurement. Appl. Opt. 2015, 54, A9–A17. [Google Scholar] [CrossRef]
  69. Nguyen, H.; Liang, J.; Wang, Y.; Wang, Z. Accuracy assessment of fringe projection profilometry and digital image correlation techniques for three-dimensional shape measurements. J. Phys. Photonics 2021, 3, 014004. [Google Scholar] [CrossRef]
  70. Nguyen, A.; Ly, K.; Li, C.; Wang, Z. Single-shot 3D shape acquisition using a learning-based structured-light technique. Appl. Opt. 2022, 61, 8589–8599. [Google Scholar] [CrossRef]
  71. Nguyen, H.; Wang, Z. Accurate 3D Shape Reconstruction from Single Structured-Light Image via Fringe-to-Fringe Network. Photonics 2021, 8, 459. [Google Scholar] [CrossRef]
  72. Nguyen, H.; Novak, E.; Wang, Z. Accurate 3D reconstruction via fringe-to-phase network. Measurement 2022, 190, 110663. [Google Scholar] [CrossRef]
  73. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar] [CrossRef]
  74. Keras. ExponentialDecay. Available online: https://keras.io/api/optimizers/learning_rate_schedules/ (accessed on 13 April 2023).
  75. Nguyen, A.; Rees, O.; Wang, Z. Learning-based 3D imaging from single structured-light image. Graph. Models 2023, 126, 101171. [Google Scholar] [CrossRef]
  76. Zollmann, S.; Kalkofen, D.; Hoppe, C.; Kluckner, S.; Bischof, H.; Reitmayr, G. Interactive 4D overview and detail visualization in augmented reality. In Proceedings of the IIEEE International Symposium on Mixed and Augmented Reality (ISMAR), Atlanta, GA, USA, 5–8 November 2012; pp. 167–176. [Google Scholar] [CrossRef]
  77. Tian, Y.; Long, Y.; Xia, D.; Yao, H.; Zhang, J. Handling occlusions in augmented reality based on 3D reconstruction method. Neurocomputing 2015, 156, 96–104. [Google Scholar] [CrossRef]
  78. Xu, K.; Chia, K.; Cheok, A. Real-time camera tracking for marker-less and unprepared augmented reality environments. Image Vis. Comput. 2008, 26, 673–689. [Google Scholar] [CrossRef]
  79. Castle, R.; Klein, G.; Murray, D. Wide-area augmented reality using camera tracking and mapping in multiple regions. Comput. Vis. Image. Underst. 2011, 115, 854–867. [Google Scholar] [CrossRef]
  80. Zollmann, S.; Hoppe, C.; Kluckner, S.; Poglitsch, C.; Bischof, H.; Reitmayr, G. Augmented Reality for Construction Site Monitoring and Documentation. Proc. IEEE 2014, 102, 137–154. [Google Scholar] [CrossRef]
  81. Collins, T.; Pizarro, D.; Gasparini, S.; Bourdel, N.; Chauvet, P.; Canis, M.; Calvet, L.; Bartoli, A. Augmented Reality Guided Laparoscopic Surgery of the Uterus. IEEE Trans. Med. Imaging 2021, 40, 371–380. [Google Scholar] [CrossRef]
  82. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  83. Alom, M.Z.; Yakopcic, C.; Hasan, M.; Taha, T.M.; Asari, V.K. Recurrent residual U-Net for medical image segmentation. J. Med. Imaging 2019, 6, 014006. [Google Scholar] [CrossRef] [PubMed]
  84. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. ResUNet-a: A deep learning framework for semantic segmentation of remotely sensed data. ISPRS J. Photogramm. Remote Sens. 2020, 6, 94–114. [Google Scholar] [CrossRef]
  85. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
Figure 1. (a) Illustration of a 3D reconstruction system and process; (b) an RVBUST RVC 3D Camera employed in this work.
Figure 1. (a) Illustration of a 3D reconstruction system and process; (b) an RVBUST RVC 3D Camera employed in this work.
Sensors 23 07284 g001
Figure 2. Flowchart of the FPP 3D imaging technique with DFFS (a) and TFFS (b) phase-shifting schemes.
Figure 2. Flowchart of the FPP 3D imaging technique with DFFS (a) and TFFS (b) phase-shifting schemes.
Sensors 23 07284 g002
Figure 3. Examplars of input–output pair in (a) DFFS datasets and (b) TFFS datasets.
Figure 3. Examplars of input–output pair in (a) DFFS datasets and (b) TFFS datasets.
Sensors 23 07284 g003
Figure 4. (a,b) Time-distributed concept for DFFS phase-shifting scheme, and (c) the comparable spatial F2ND approach.
Figure 4. (a,b) Time-distributed concept for DFFS phase-shifting scheme, and (c) the comparable spatial F2ND approach.
Sensors 23 07284 g004
Figure 5. (a) Time-distributed concept for TFFS phase-shifting scheme, and (b) the comparable spatial F2ND approach.
Figure 5. (a) Time-distributed concept for TFFS phase-shifting scheme, and (b) the comparable spatial F2ND approach.
Sensors 23 07284 g005
Figure 6. Evaluation of image quality metrics (SSIM and PSNR) for predicted numerators and denominators.
Figure 6. Evaluation of image quality metrics (SSIM and PSNR) for predicted numerators and denominators.
Sensors 23 07284 g006
Figure 7. 3D shape reconstruction of a single-object scene using DFFS datasets.
Figure 7. 3D shape reconstruction of a single-object scene using DFFS datasets.
Sensors 23 07284 g007
Figure 8. 3D shape reconstruction of a scene with multiple objects using DFFS datasets.
Figure 8. 3D shape reconstruction of a scene with multiple objects using DFFS datasets.
Sensors 23 07284 g008
Figure 9. 3D shape reconstruction of a single-object scene using TFFS datasets.
Figure 9. 3D shape reconstruction of a single-object scene using TFFS datasets.
Sensors 23 07284 g009
Figure 10. 3D shape reconstruction of a scene with multiple objects using TFFS datasets.
Figure 10. 3D shape reconstruction of a scene with multiple objects using TFFS datasets.
Sensors 23 07284 g010
Figure 11. Potential application of TD framework with different output formats in FPP technique.
Figure 11. Potential application of TD framework with different output formats in FPP technique.
Sensors 23 07284 g011
Table 1. Quantitative analysis comparing TD and spatial F2ND approaches.
Table 1. Quantitative analysis comparing TD and spatial F2ND approaches.
DatasetMethodError (Lower Is Better)Accuracy (Higher Is Better)
relrmslogrms log δ < 1.25 δ < 1.25 2 δ < 1.25 3
DFFSTD Layer0.0041.3120.0040.05994.1%96.6%98.2%
TD Module0.0041.2160.0040.05594.9%97.0%98.4%
Spatial F2ND0.0040.8560.0020.04497.9%98.7%99.2%
TFFSTD Layer0.0030.2130.0020.03799.4%99.5%99.5%
TD Module0.0030.1760.0020.03596.9%97.0%97.2%
Spatial F2ND0.0051.0560.0020.03896.8%96.9%97.0%
Table 2. Initial quantitative evaluation of TD Module and spatial F2ND techniques using the internal Attention UNet network.
Table 2. Initial quantitative evaluation of TD Module and spatial F2ND techniques using the internal Attention UNet network.
DatasetMethodError (Lower Is Better)Accuracy (Higher Is Better)
relrmslogrms log δ < 1.25 δ < 1.25 2 δ < 1.25 3
DFFSAttention TD Module0.0031.3340.0050.06093.6%96.3%98.1%
Attention F2ND0.0051.3450.0040.05894.1%96.4%98.0%
TFFSAttention TD Module0.0030.1500.0020.03597.0%97.1%97.3%
Attention F2ND0.0050.9410.0020.04096.9%97.0%97.1%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nguyen, A.-H.; Wang, Z. Time-Distributed Framework for 3D Reconstruction Integrating Fringe Projection with Deep Learning. Sensors 2023, 23, 7284. https://0-doi-org.brum.beds.ac.uk/10.3390/s23167284

AMA Style

Nguyen A-H, Wang Z. Time-Distributed Framework for 3D Reconstruction Integrating Fringe Projection with Deep Learning. Sensors. 2023; 23(16):7284. https://0-doi-org.brum.beds.ac.uk/10.3390/s23167284

Chicago/Turabian Style

Nguyen, Andrew-Hieu, and Zhaoyang Wang. 2023. "Time-Distributed Framework for 3D Reconstruction Integrating Fringe Projection with Deep Learning" Sensors 23, no. 16: 7284. https://0-doi-org.brum.beds.ac.uk/10.3390/s23167284

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop