Next Article in Journal
Using Paper Texture for Choosing a Suitable Algorithm for Scanned Document Image Binarization
Next Article in Special Issue
Path Tracing vs. Volume Rendering Technique in Post-Surgical Assessment of Bone Flap in Oncologic Head and Neck Reconstructive Surgery: A Preliminary Study
Previous Article in Journal
Using an Ultrasound Tissue Phantom Model for Hybrid Training of Deep Learning Models for Shrapnel Detection
Previous Article in Special Issue
Augmenting Performance: A Systematic Review of Optical See-Through Head-Mounted Displays in Surgery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

X23D—Intraoperative 3D Lumbar Spine Shape Reconstruction Based on Sparse Multi-View X-ray Data

1
Research in Orthopedic Computer Science, Balgrist University Hospital, University of Zurich, 8008 Zurich, Switzerland
2
Department of Orthopedics, Balgrist University Hospital, University of Zurich, 8008 Zurich, Switzerland
*
Author to whom correspondence should be addressed.
Submission received: 30 June 2022 / Revised: 7 September 2022 / Accepted: 27 September 2022 / Published: 2 October 2022

Abstract

:
Visual assessment based on intraoperative 2D X-rays remains the predominant aid for intraoperative decision-making, surgical guidance, and error prevention. However, correctly assessing the 3D shape of complex anatomies, such as the spine, based on planar fluoroscopic images remains a challenge even for experienced surgeons. This work proposes a novel deep learning-based method to intraoperatively estimate the 3D shape of patients’ lumbar vertebrae directly from sparse, multi-view X-ray data. High-quality and accurate 3D reconstructions were achieved with a learned multi-view stereo machine approach capable of incorporating the X-ray calibration parameters in the neural network. This strategy allowed a priori knowledge of the spinal shape to be acquired while preserving patient specificity and achieving a higher accuracy compared to the state of the art. Our method was trained and evaluated on 17,420 fluoroscopy images that were digitally reconstructed from the public CTSpine1K dataset. As evaluated by unseen data, we achieved an 88% average F1 score and a 71% surface score. Furthermore, by utilizing the calibration parameters of the input X-rays, our method outperformed a counterpart method in the state of the art by 22% in terms of surface score. This increase in accuracy opens new possibilities for surgical navigation and intraoperative decision-making solely based on intraoperative data, especially in surgical applications where the acquisition of 3D image data is not part of the standard clinical workflow.

1. Introduction

Spinal surgeries are challenging procedures due to their technical complexity and proximity to vital organs such as the spinal cord, nerves, and the aorta [1]. Several surgeon- and patient-specific factors have been reported in the prior art as the root cause for postoperative complications in such interventions [2], ranging from pedicle screw malplacement [3] to cage malpositioning [4]. In particular, the placement of pedicle screws, which is one of the most common steps in spine surgery, requires highly accurate surgical execution. Studies have shown that pedicle screw malplacement is often evaluated based on the distance between the breaching screw’s edge and the adjacent bone cortex [5]. The distances are classified into minor (<2 mm), moderate (2–4 mm), and severe (>4 mm) cases. According to literature such as [6,7], screw perforations exceeding 4 mm can cause postoperative complications. Furthermore, it has been shown that lumbar pedicle screw malplacement rates can range from 5% to 41% depending on the insertion method, the vertebral level, and the assessment protocol [3]. Pedicle screw malplacement bears the risk of neural injuries with notable negative consequences [8]. Such screw malplacements can lead to revision surgeries in up to 6% of patients [9]. Moreover, in an analysis of cage placement in posterior lumbar interbody fusion surgeries, it was identified that factors such as cage size, shape, and position can determine the lordosis in such surgeries [10]. To this end, cage malpositioning has been reported to cause nerve root compression and neurological complications [11].
Given that potential screw breaches and cage malpositionings should be ideally determined and avoided intraoperatively to prevent subsequent revision surgeries, most surgeons rely on visual assessment of multi-view (i.e., anterior–posterior and lateral views) intraoperative 2D X-rays and fluoroscopy-guided surgery as their predominant aid for surgical guidance and error prevention.
Because 2D X-ray imaging projects the 3D surgical scene onto a 2D plane, this approximation remains a poor representation of the real 3D shape of the anatomy and lacks appropriate spatial feedback. As shown in [12,13], the detection of potential pedicle screw malplacements based on planar X-ray imagery is of low accuracy (ranging from 63% to 83%) and is highly dependent on the evaluator, spinal level, and the screw material [14].
Computer-assisted surgical (CAS) navigation techniques have been introduced as means to facilitate and standardize such interventions by providing 3D intraoperative spatial guidance and were shown to result in substantially higher pedicle screw-insertion accuracy compared to free-hand and conventional 2D fluoroscopy-guided methods [3,15,16]. Furthermore, CAS technologies have also been used for cage-insertion purposes [17] and have been shown to result in improved performance [18]. The most common CAS systems still rely on preoperative 3D imaging [19] and preoperative planning, which allows for the generation of patient-specific models (bio-mechanical digital twins), which can be used as the basis for computer simulations helping to determine and optimize surgery plans. To provide intraoperative spatial guidance, these plans need to be registered to the patient’s anatomy. Although registration of preoperative data to the anatomy is a well-established concept that can be achieved through feature-based [20,21], intensity-based [22,23], or data-driven algorithms [24], it remains a challenging and sometimes error-prone process that often requires meticulous manual input [25,26]. Requiring a registration process between the preoperative and intraoperative data remains the major bottleneck in the widespread adoption of the CAS systems [22]. Additionally, due to the use of preoperatively created plans that are generated before the intervention, such systems usually fail to adapt to unforeseen intraoperative changes such as anatomical alterations and are not a suitable option in non-elective urgent surgeries where surgeons must resort to conventional, non-navigating techniques. Those and other contributing factors, such as the added cost and surgery time, have all contributed to the limited uptake of CAS systems in the community, the usage of which has been reported to be on the order of 11% of the surgeries conducted worldwide [27].
In light of the abovementioned limitations, new opportunities arise when full-fledged surgical navigation can be accomplished only based on intraoperative data and without registration. The generation of 3D representations of the anatomy instantaneously and directly in the operating room plays an essential role here. Currently, two categories of solutions exist to enable the intraoperative-only generation of 3D representations of anatomical shapes. Methods of the first category create representations by using 3D intraoperative imaging, whereas the second category consists of data-driven approaches that adapt generic atlases according to intraoperative 2D imaging.
The first category requires the use of mobile C-arm devices equipped with cone-beam computed tomography (CBCT) functionality, enabling the acquisition of 3D volumetric images during the operation. Combined with optical tracking systems, this technology allows for registration-free surgical navigation [28]. Apart from the high price of CBCT devices [16], the image acquisition process itself leads to a considerable increase in surgery time [29], and additional ionizing radiation [30].
One prominent example in the family of data-driven approaches are statistical shape models (SSMs), which capture the average shape of anatomies and their variations in atlases constructed from pre-existing databases of often healthy anatomy [31,32]. Although a variety of SSM techniques exist in the prior art, a majority of such methods still require a registration step to match the adapted atlas to the intraoperative data [33,34,35] when using them for CAS purposes. To overcome the need for performing registration, recent machine learning (ML) algorithms have been proposed that can learn to create 3D anatomical representations based on different modalities such as 2D input images [36,37,38] or ultrasound [39]. Once trained, data-driven approaches are capable of generating intraoperative 3D representations without relying on patient-specific preoperative data by fitting the trained model or atlas to fluoroscopy data [40]. A growing number of publications exist on reconstructions of patients’ 3D shapes based on 2D imaging data (i.e., [38]). Those methods process a fixed number of images from fixed poses (e.g., two bilateral X-rays taken from anterior–posterior and lateral viewpoints in [38]). Due to the end-to-end training of those networks and a limited number of pose-restricted images, a pixel-wise correspondence to the 3D shape is learned. Unseen parts of the 3D shape can be reconstructed through the underlying generalization of the network. To the best of our knowledge, most of the existing methods in this field are focused on diagnostic purposes and have not been designed and validated for intraoperative applications (i.e., CAS). Furthermore, a general issue with the data-driven methods is the limited patient specificity due to the strong prior assumptions of the models, which makes it difficult to generalize to all possible pathologies, such as bone deformities, fractures, and tumors [41]. This characteristic has been recently demonstrated on end-to-end 3D reconstruction networks that failed to sufficiently include the spatial information of the image data and tended to produce an average shape representation [42].
Multi-view fluoroscopy images are ubiquitously available in many operating rooms; however, they are currently mainly used for qualitative visual assessment purposes. In hopes of utilizing the full untapped potential of such images. In this study, we introduce a novel shape estimation approach, which we call X23D, that aims at extracting patient-specific 3D shape representations of the lumbar spine by using multi-view sparse X-ray data and the corresponding image calibration parameters as input. For this, we adapted a learnt stereo machine (LSM) [43] and trained it on a large dataset of digitally reconstructed radiographs (DRRs) generated from a computed tomography (CT) dataset. The X23D model takes 2D fluoroscopy images and their corresponding pose information as inputs and reconstructs accurate 3D shapes from a fewer number of 2D images than classic approaches [43], while conforming to the underlying geometric constraints. Because our network structure takes image projection matrices as input, anatomical shape priors can be incorporated while preserving features visible in the underlying 2D images [42]. In the context of our target application (i.e., spinal fusion surgeries), this translates into the capability of estimating high-quality 3D representations of the lumbar vertebrae in a purely intraoperative fashion and relying neither on additional imaging nor tracking apparatus for 3D reconstruction and registration purposes, respectively. This algorithm can pave the way toward a fully fetched surgical navigation solution that can provide surgery guidance tailored to intraoperative conditions. By using pedicle screw placement as an example, surgical navigation can in future work be achieved by utilizing our 3D representation with intraoperative planning to calculate ideal drilling trajectories. These plans can then, for example, be further visualized in an augmented manner, such as in [44], or be used for robotic executed or supported procedures, such as in [45].
In this paper, we lay the foundation for the design and the training of the aforementioned algorithm and evaluate its performance under different imaging conditions and against a state-of-the-art method.

2. Materials and Methods

Figure 1 demonstrates the intended flow of information through our X23D algorithm during the inference stage. Assuming that our X23D algorithm is previously trained, it accepts a number of 2D X-ray images cropped down to the vertebra of interest, together with the projection matrices. For this, a pre-processing module is introduced that: (a) localizes the lumbar vertebrae in the input X-rays (and optionally segments the bone projections) and (b) calibrates the input X-rays to estimate their projection matrices including the intrinsic and the extrinsic imaging parameters. As the output of the X23D algorithm, the 3D shape representations of the individual lumbar vertebrae are generated.
The methodology section is organized as follows. First, we discuss the required pre-processing of the inputs before explaining the network architecture. Afterward, we elaborate on the data generation, including the X-ray simulation and the choice of viewing angles. In the end, we point out the train–validation–test split with its data augmentation and evaluation of the X23D’s performance in different conditions.

2.1. Input Data and Pre-Processing

The raw X-ray images first go through a calibration module that estimates the corresponding imaging parameters at the time of acquisition. The choice of the calibration method depends on the X23D’s use case. In a navigation setup, where the pose of the imaging device can be tracked through the use of a navigation apparatus (i.e., optical or electromagnetic tracking systems), methods such as [46] can be used. Alternatively, in setups where the use of optical tracking apparatus is rendered infeasible, precisely fabricated calibration phantoms can be used [47,48,49]. Furthermore, AI-based algorithms for C-arm calibration are currently unveiling [50,51]. In our previous work [49,51], calibration was achieved by using phantom-based and AI-based approaches, respectively, which in [49] resulted in a 0.4 mm average calibration accuracy. Therefore, the X-ray calibration module was not included in the scope of this study. After the X-ray calibration is completed for the ith image, the projection matrices P i contain the intrinsic K i and the extrinsic [ R i | t i ] imaging parameters, all expressed in the world coordinate frame:
P i = K i R i [ I | t i ] .
To crop the raw input X-rays in a way that they only contain the lumbar vertebra of interest, an X-ray localization module was needed during the pre-processing stage to extract the 2D bounding boxes for each vertebra. Previously, localization of vertebrae levels in X-ray images has been achieved through work such as [52,53,54,55]. Similar to X-ray calibration, X-ray localization was not covered in this study, given the existence of such methods in the prior art. A similar line of methodologies exists that can segment the projection of vertebrae on input X-rays (e.g., [56,57]). This can be fed into the X23D algorithm as an auxiliary input to assist the reconstruction task (depicted by dashed lines shown in Figure 1).
Due to the cropping introduced during the localization process, the P-matrices had to be adapted. Cropping to the localized vertebra results in a shift inside the image plane that can be expressed by
P i ^ = Q i P i , Q i = 1 0 t i x 0 1 t i y 0 0 1 ,
where P ^ is the adjusted P-matrix and t x , t y are the amount of horizontal and vertical shift introduced during the cropping stage, which are determined based on the position of the vertebral bounding box in the original full-sized X-rays.

2.2. X23D Network Architecture

As its input, the X23D network accepts i localized X-ray images representing the anatomy from multiple views. Each image is resized to 224 × 224 pixels and normalized with respect to intensity to a range of [0, 1] (Figure 2). The first stage of the network is represented by a truncated UNet [58] (channels = [(1), 32, 64, 128, 256]) with skip connections that result in 32 2D feature maps w i of size 112 × 112 . Those feature maps are then back-projected into the 3D space by utilizing the adjusted projection matrices P i ^ , resulting each in a 128 × 128 × 128 3D feature grid W i . The following back-projection strategy is applied to avoid ray-tracing every 2D feature in the extracted 2D feature maps w i from i viewpoints onto the associated 3D feature grids W i . All points from the 3D feature grids X W i are projected into the 2D feature map space x w i by using x = P ^ i X . The resulting points x can thus be represented in a continuous 2D space, compared to the discrete locations of the 2D feature maps w. To assign values to the back-projected points, the discrete points of the 2D feature maps were bi-linearly interpolated. This differential back-propagation is similar to [43] and allows forward and backward passes through the neural network.
Because high-resolution feature grids are required in our application to produce high-quality 3D output shapes, the gated recurrent unit (GRU) used in [43] proved to be too memory-intensive. Hence, we replaced it with an averaging strategy, which allowed us to achieve an output resolution of 128 × 128 × 128 voxels on our available hardware (more on the used hardware in Section 2.6). The final stage was a refiner network based on a UNet architecture [58] (channels = [16, 32, 64, 128, 256]) with skip connections. The output of the finally reconstructed 3D shape was represented in the form of a 128 × 128 × 128 voxel grid. Rectified linear units have been used for all activation functions except the last one, which used sigmoid functions.

2.3. Data Generation and Training

Large datasets of annotated images were needed to train the X23D network (Section 2.2). Given that access to corresponding X-ray + CT scan datasets along with the annotations is a tedious undertaking, we opted for simulating the required X-ray images based on a large number of patient CT scans. To this end, we used the CTSpine1K dataset [59], which consists of 1005 CT scans of patients in DICOM format with 4712 lumbar vertebrae and their corresponding segmentation masks. The dataset was filtered according to our inclusion and exclusion criteria. Inclusion criteria were a resolution better than 1 mm in all the three axes, CT volumes bigger than 128 voxels on each axis, and the presence of all five lumbar vertebrae. Cases with pathological bone fusion (synostosis) were excluded. After filtering, the dataset consisted of 871 individual patient CT scans with a total of 4355 vertebrae from which the input–output training pairs were generated (Section 2.3.1).

2.3.1. X-ray Simulation

By using the collected 3D patient dataset along with the corresponding 3D segmentations, 2D X-rays and all the required 2D annotations were automatically generated in a custom-made software. The core component of the software integrates a method for generating DRR [60], which can be summarized as follows. Taking as input an object (i.e., a CT scan) with its X-ray attenuation map A : R 3 R , our software generates a DRR image I through
I ( p ) = A ( T 1 L ( p , s ) ) d s ,
where I ( p ) is the intensity of the DRR at image point p, T : R 3 R 3 is the transformation between the object and the image plane, and L ( p , s ) is the ray initiating from the X-ray source and intersecting with the image plane at point p parameterized by s. The interface of our custom-developed software is shown in more details in Figure 3.
The outputs of the DRR creation and annotation software were as follows.
  • Full-size 2D DRR images generated from different view angles (more details about the view angles are provided in Section 2.3.2).
  • The corresponding P-matrices expressing the intrinsic and the extrinsic parameters.
  • 2D vertebral bounding boxes.
  • 2D vertebral segmentation masks.
  • Automatically localized 2D DRRs that were cropped down to a single vertebra using the bounding boxes in item 3.
  • Automatically segmented localized 2D DRRs consisting of the bone and the background classes.
  • Adjusted P ^ -matrices expressing the cropping effect.
The automatically cropped DRRs and the corresponding adjusted P ^ -matrices were used to train the X23D network (Section 2.2). A detailed illustration of the training pipeline for the X23D network is provided in Figure 4.

2.3.2. View Angles

The DRR approach was selected because it enabled us to simulate 2D X-rays from arbitrary viewpoints around the patient. However, poses were limited to those that can be acquired with a C-arm in an operating room, namely varying anterior–posterior, lateral, and oblique views. Therefore, four categories of viewing angles were selected:
  • AP: included three images—one strict anterior–posterior image and two images with ± 15 deviation in the sagittal plane.
  • Lateral: included four images—two strict lateral images from each side and an additional two images with ± 20 deviation in the coronal plane.
  • Oblique: included four images—two oblique shots with ± 20 separation from the AP in the sagittal plane and two additional shots with ± 35 separation from the AP in the same fashion.
  • Miscellaneous: included nine shots from poses with deviations in two planes that could not be assigned to the previously defined categories.
The 871 CTs were rotated such that each CT was in a supine position. There was intentionally no further alignment or registration to introduce additional variance within the extrinsic poses. The intrinsics K only depended on the DRR size, which was fixed to 224 × 224 pixels and the focal length. To mimic different C-arm designs, three focal lengths (850, 900, and 1000) were chosen randomly for every new DRR.
The X23D algorithm was tested on two input configurations, namely by using four or eight X-rays as inputs. According to a discussion with our surgeons, the four-view configuration was defined as the baseline, given that it was deemed clinically preferable in terms of surgery time and radiation exposure. When training the X23D network with the four-view configuration, we randomly selected one shot per view angle category (i.e., AP, lateral, oblique, and miscellaneous). For the eight-view setup, we selected the first four shots following the same approach as in the four-view setup and the second four shots were chosen randomly from all the view angle categories combined.

2.3.3. Train–Validation–Test Split

The dataset was split after random shuffling into 70% training (611 patients), 20% validation (173 patients), and 10% (87 patients) test data. Given the fact that we produced 20 DRRs for every patient from all the abovementioned viewing angels, a total of 17,420 full-size DRRs were created for all the patients within our dataset (87,100 localized DRRs).

2.4. Performance Evaluation

In addition to using the F1 score and IoU for measuring the performance of X23D, we followed [42,62] and report the surface score, which is the harmonic mean of precision and recall at a given threshold (in our case d = 1 % of the side length of the reconstruction volume)
S ( d ) = 2 · P r e c i s i o n ( d ) R e c a l l ( d ) P r e c i s i o n ( d ) + R e c a l l ( d ) .
To calculate S ( d ) , the voxelized predictions and the ground truth were converted to meshes R and G , respectively, after which points were sampled from their surfaces. The precision and recall was then calculated as two asymmetric distances, where precision is the aggregated distances from points r R to g G
P r e c i s i o n ( d ) = 1 n R r R min g G g r < d
and recall from g G to r R
R e c a l l ( d ) = 1 n G g G min r R g r < d ,
with n being the number of sampled points. Compared to the F1 score, which is a function of the overlap of target and prediction volumes, the surface score is a suitable metric for comparing fine surface details. To have a better understanding of the clinical applicability of our reconstructions and have a physical estimate of X23D’s performance, we further calculated the robust Hausdorff Distance with a 95% percentile (HD95) and the average surface distance (ASD).
Note that parts of the adjacent vertebrae and the surrounding soft tissue were present in the localized DRRs that were used to train the X23D model in Section 2.3. By using the DRR creation and annotation software (Section 2.3.1), we created two additional datasets of localized 2D DRRs that were also segmented to foreground (i.e., bone) and background (i.e., other tissues) classes. This was done for two different reasons. Counterpart 3D reconstruction models, such as Pix2Vox++ [63], generally require pre-segmented 2D inputs to reconstruct the corresponding 3D shape of the segmented area. Therefore, to have a benchmark for comparing our end-to-end algorithm to such models, we needed segmented input DRRs. Furthermore, given that the simulated DRR images that we have used to train our reconstruction network can be inevitably different in terms of image and radiological properties compared to real intraoperative X-rays, the performance of the reconstruction network may also differ when confronted with real X-ray or simulated DRR images. As a potential remedy, an investigation was performed to evaluate the feasibility of our algorithm when used on real X-ray images that have previously been segmented to foreground (bone) and background (other tissues) classes. Given that such segmentation masks will result in similar input images for our reconstruction algorithm regardless of fidelity of the underlying image (i.e., DRRs or real X-rays), we created the aforementioned segmented datasets by using DRR images. By performing this analysis, we assessed the expected performance of our reconstruction method on real X-rays that are accompanied with corresponding segmentation masks. To this end, we created two additional datasets consisting of the aforementioned segmented 2D DRRs and used them to train additional two X23D versions that could accept segmented 2D inputs, as follows.
  • Segmented grayscale: these localized images were created by applying the 2D segmentation masks (as described in item 6 in the list above) onto the unsegmented localized images (Figure 5b).
  • Segmented binary: these localized images were created by thresholding the segmented grayscale images (Figure 5c).
Figure 4 shows the use of the segmented grayscale and segmented binary as additional inputs for the X23D architecture, indicated by dashed lines.
Specific to the segmented binary dataset, a data augmentation process was implemented based on applying 3D elastic deformation on the underlying CT scans before creating the 2D DRR images, resulting in a doubling of the amount of 2D data (i.e., segmented binary augmented). A displacement for each voxel was interpolated from a coarse grid by using cubic B-splines with six control points. Each control point was displaced by a random amount between 0 and 18 voxels. Note that the data augmentation was only applied to the segmented binary dataset given that such augmentations would have introduced unrealistic DRR projections due to the deformation of surrounding tissue.

2.5. Statistical Evaluation

Apart from reporting mean and standard deviation of the results, inferential statistical tests were performed to compare the results of our algorithm to the ones from the benchmark method (i.e., Pix2Vox++). Furthermore, we ran statistical testing to determine whether or not there was a statistically significant difference between the reconstructions achieved by segmented versus unsegmented 2D input. After initial investigation of the reconstruction results, we determined that they do not follow a normal distribution; thus, we applied a Mann–Whitney U tests for all of our inferential statistics. An α -value of 0.05 was selected and Bonferroni-corrected.

2.6. Coding Platform, Hardware and Training Hyperparameters

For the DRR creation and annotation software, the Insight Toolkit (ITK) framework [64] and the SimpleITK framework [65] were used. In addition, PySide6 and QT were used to create the interface. The elastic deformations were applied by using TorchIO’s [66] RandomElasticDeformation function. The network was built with PyTorch [67] and trained on a GPU cluster consisting of multiple Nvidia RTX 2080 Ti with 11 GB VRAM and Nvidia GTX 1080 TI with 11 GB VRAM. Training of the networks required between 30 and 40 epochs with a learning rate of 10 4 . Early stopping was applied to regularize the training. Due to the large output volume size of 128 × 128 × 128 voxels, we leveraged multiple GPUs and PyTorch Lightning’s [68] accumulated gradients, which resulted in an effective batch size of 16. In the eight-view configuration, training on eight GPUs lasted approximately seven days.
For the calculation of the metrics we used Deepmind’s surface-distance repository (https://github.com/deepmind/surface-distance accessed on 5 September 2022). For the statistical evaluation SciPy [69] and R have been used.

3. Results

Table 1 reports the performance of our proposed method on unseen data (unseen patients and unseen DRRs) by using the F1, IoU, and surface scores. This table also reflects on the performance of our algorithm with pre-segmented input X-rays. Additionally, we have compared the results of X23D to the results achieved by the Pix2Vox++ network.
On unsegmented vertebrae and with four views, our method outperformed the four- and eight-view versions of Pix2Vox++, with a 0.88 average F1 score. Our network’s performance increased slightly when pre-segmentation was available, which was indicated by an increase in the F1 score from 0.88 to 0.89 in the four-view setup. A more notable performance difference can be seen in the surface scores, where our network performed 22% better than Pix2Vox++.
According to the statistical test carried out, all of our methods perform significantly better than the benchmark method Pix2Vox++ regardless of the number of views. Our main approach with four unsegmented views does not perform significantly differently from the segmented binary approach in terms of F1 score or surface score. Additionally, there is no significant difference comparing the surface score results of our main approach (unsegmented four-view) to the augmented segmented binary approach.
Table 2 shows the results of the evaluation of the unsegmented four-view approach from Table 1 broken down to the vertebrae level L1 to L5. The F1 scores deviated only slightly from the average across all levels. L1 and L2 resulted in similar performances with only slightly above average F1 scores. However, the surface scores of both were 5% better than average. Performance drops gradually towards L5, with the surface score for L4 being 3% and L5 15% below the average.
Figure 6 shows sample results of our reconstruction algorithm trained on four views of unsegmented and grayscale segmented data, respectively. Slices of the axial and sagittal planes from the 3D reconstructions in Figure 6 are shown in Figure 7 for better comparison. In Figure 8, visualizations of precision and recall are shown with a threshold of 2% for two unseen test cases.
The reconstruction of a vertebra, including loading of the images, took on average 4.2 s for the four-view arrangement and 7.9 s for the eight-view arrangement. These values were measured during the inference of 100 vertebrae on a workstation with NVIDIA RTX 2080 Super and an i7-8700 CPU @ 3.20 GHz.

4. Discussion

This study showed that accurate patient-specific 3D representations of the spine anatomy can be generated from a sparse set of four X-ray images. We introduced X23D as a complete pipeline for data processing, network training and inference, the performance of which was tested on unseen data. Compared to the state-of-the-art 3D reconstruction algorithms such as Pix2Vox++, our method achieved a 22% higher surface score, which can be attributed to the exploitation of the spatial information present in the projection matrices. Compared to the status quo in conventional spinal surgeries where on average 54 X-ray views are acquired [70], and CBCT-based operations where 100–120 fluoroscopy views are needed [29], our method required substantially fewer X-rays to produce accurate 3D representations.
The X23D algorithm achieved high-quality solutions even in cases where no pre-segmentation of the input X-rays was available. In such cases, the 3D shape reconstruction accuracy of X23D was 0.88 in a four-view setup as assessed by the F1 score. The only principal requirement was the availability of localization of individual vertebral levels on the input X-rays, which has been presented in studies such as [54]. In cases where a pre-segmentation of vertebrae were present, we demonstrated a slight statistical improvement in X23D’s reconstructions, where the F1 score and the surface score were improved by an additional 1% and 3%, respectively. The segmented binary augmented approach did not perform significantly different than our main approach (unsegmented four-views). This speaks for the application of our method on real X-ray images. The result for the individual levels shows a clear drop in performance for L5. This can be attributed to the fact that L1 to L4 look more similar than L5. More training data could allow training of models for individual vertebral levels, which could increase the overall performance.
Despite higher F1 scores, this study was not able to show a significant improvement of the eight-view variants over their four-view variants in terms of surface score. One potential reason might be that the current network architecture did not have enough capacity to reveal its full potential given more views. This might be connected with the simplistic merging stage chosen in the herein implementation of the network to reduce the memory consumption to an acceptable level. Alternative architectures such as attention mechanisms will be evaluated in future work. Upon close visual inspection of the precision-recall plots (as example shown in Figure 8), anatomical regions of high clinical importance such as the spinal canal, vertebra body, and pedicles were areas with high reconstruction precision, which is especially important for applications such as pedicle screw placement surgery.
The inference time of 4.2 s for the four-view and 7.9 s for the eight-view constellation, respectively, show the potential of our algorithm for intraoperative use cases. In comparison to prior art on spinal reconstruction were computation time off approximately four seconds has been reported (e.g., [35]), our approach has a rather comparable computation time. Note that neither the network nor the data loader are optimized for speed yet. Therefore, further speed gains are expected to enable intraoperative lumbar spine reconstruction in a few seconds.
This study had limitations. Although stemming from real patient CT scans, the X-ray images used for training and testing the X23D algorithm were simulated by using a custom-made software to streamline the data collection, annotation and curation in light of the practical hurdles along the collection of paired real CT and X-ray datasets. Prior work, such as [71,72], show that clinically feasible simulations of X-rays can be achieved by more advanced simulation methods. Such X-ray simulation algorithms will be implemented into the X23D pipeline in the future phases of this project. Nevertheless, in an effort to demonstrate the potential of the current X23D network when faced with real X-ray data, we explored its reconstruction accuracy when 2D segmentation masks are available. This mimics a scenario when a real X-ray is first segmented and later fed to the X23D algorithm. By performing such additional evaluations, we showed that our approach can be extended to a segmentation and reconstruction pipeline that can achieve similar 3D reconstruction performance on real X-rays. In the future, an in vivo data collection study will allow us to collect real X-rays with corresponding calibration data by using our calibration phantom. The reconstruction accuracy depends on the quality of the calibration. However, a comprehensive sensitivity analysis of the calibration parameters was beyond the scope of this work, but is planned for the future. Nevertheless, we performed a preliminary sensitivity analysis to determine the applicability of our method. Perturbations of up to 50 mm were applied separately to the focal length parameters of either anterior–posterior or lateral images. For perturbed lateral or anterior–posterior images, a monotonous performance decrease in the F1 score of up to 3% and 2% respectively, was observed. The surface scored decreased monotonously by up to 7% for lateral and 5% for anterior–posterior perturbed images.

5. Conclusions

In this work, an end-to-end approach to the 3D reconstruction of vertebra of the lumbar spine was presented. Our method requires only sparse 2D fluoroscopic data to achieve accurate reconstructions, making it a prime candidate for intraoperative applications. This was achieved by explicitly incorporating the calibration parameters as network inputs, which can be considered as one of our key contributions. We see big potential in using our method as a component in the next generation of computer-aided interventions and surgical navigation, especially in surgical applications in which the acquisition of 3D image data is not part of the clinical workflow.

Author Contributions

Conceptualization, S.J. and H.E.; methodology, S.J., H.E. and P.F.; software, S.J., C.J. and H.E.; validation, S.J. and H.E.; formal analysis, S.J., C.J. and H.E.; investigation, S.J.; resources, M.F. and P.F.; data curation, S.J., C.J. and H.E.; writing—original draft preparation, S.J.; writing—review and editing, H.E. and P.F.; visualization, S.J.; technical supervision, H.E. and P.F.; clinical supervision: M.F.; project administration, H.E. and S.J.; funding acquisition, M.F. and P.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Monique Dornonville de la Cour foundation, Switzerland.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The code presented in this study is available on request from the corresponding author. The code is not publicly available due to a pending patent application. The described dataset is available here: https://github.com/MIRACLE-Center/CTSpine1K (accessed on 13 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ASD Average Surface Distance
CAS Computer-Assisted Surgical
CBCT Cone-Beam Computed Tomography
CTComputed Tomography
DRR Digitally Reconstructed Radiograph
GRU Gated Recurrent Unit
ITKInsight Toolkit
LSMLearnt Stereo Machines
ML Machine Learning
SSM Statistical Shape Model

References

  1. Medress, Z.A.; Jin, M.C.; Feng, A.; Varshneya, K.; Veeravagu, A. Medical malpractice in spine surgery: A review. Neurosurg. Focus FOC 2020, 49, E16. [Google Scholar] [CrossRef] [PubMed]
  2. Farshad, M.; Bauer, D.E.; Wechsler, C.; Gerber, C.; Aichmair, A. Risk factors for perioperative morbidity in spine surgeries of different complexities: A multivariate analysis of 1009 consecutive patients. Spine J. 2018, 18, 1625–1631. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Gelalis, I.D.; Paschos, N.K.; Pakos, E.E.; Politis, A.N.; Arnaoutoglou, C.M.; Karageorgos, A.C.; Ploumis, A.; Xenakis, T.A. Accuracy of pedicle screw placement: A systematic review of prospective in vivo studies comparing free hand, fluoroscopy guidance and navigation techniques. Eur. Spine J. 2012, 21, 247–255. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Hu, Y.H.; Niu, C.C.; Hsieh, M.K.; Tsai, T.T.; Chen, W.J.; Lai, P.L. Cage positioning as a risk factor for posterior cage migration following transforaminal lumbar interbody fusion—An analysis of 953 cases. BMC Musculoskelet. Disord. 2019, 20, 260. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Amato, V.; Giannachi, L.; Irace, C.; Corona, C. Accuracy of pedicle screw placement in the lumbosacral spine using conventional technique: Computed tomography postoperative assessment in 102 consecutive patients. J. Neurosurg. Spine 2010, 12, 306–313. [Google Scholar] [CrossRef]
  6. Gertzbein, S.D.; Robbins, S.E. Accuracy of pedicular screw placement in vivo. Spine 1990, 15, 11–14. [Google Scholar] [CrossRef]
  7. Laine, T.; Mäkitalo, K.; Schlenzka, D.; Tallroth, K.; Poussa, M.; Alho, A. Accuracy of pedicle screw insertion: A prospective CT study in 30 low back patients. Eur. Spine J. 1997, 6, 402–405. [Google Scholar] [CrossRef] [Green Version]
  8. Hicks, J.M.; Singla, A.; Shen, F.H.; Arlet, V. Complications of pedicle screw fixation in scoliosis surgery: A systematic review. Spine 2010, 35, E465–E470. [Google Scholar] [CrossRef]
  9. Nevzati, E.; Marbacher, S.; Soleman, J.; Perrig, W.N.; Diepers, M.; Khamis, A.; Fandino, J. Accuracy of pedicle screw placement in the thoracic and lumbosacral spine using a conventional intraoperative fluoroscopy-guided technique: A national neurosurgical education and training center analysis of 1236 consecutive screws. World Neurosurg. 2014, 82, 866–871. [Google Scholar] [CrossRef]
  10. Landham, P.R.; Don, A.S.; Robertson, P.A. Do position and size matter? An analysis of cage and placement variables for optimum lordosis in PLIF reconstruction. Eur. Spine J. 2017, 26, 2843–2850. [Google Scholar] [CrossRef]
  11. Kraiwattanapong, C.; Arnuntasupakul, V.; Kantawan, R.; Keorochana, G.; Lertudomphonwanit, T.; Sirijaturaporn, P.; Thonginta, M. Malposition of Cage in Minimally Invasive Oblique Lumbar Interbody Fusion. Case Rep. Orthop. 2018, 2018, 9142074. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Learch, T.J.; Massie, J.B.; Pathria, M.N.; Ahlgren, B.A.; Garfin, S.R. Assessment of pedicle screw placement utilizing conventional radiography and computed tomography: A proposed systematic approach to improve accuracy of interpretation. Spine 2004, 29, 767–773. [Google Scholar] [CrossRef] [PubMed]
  13. Ferrick, M.R.; Kowalski, J.M.; Simmons, E.D., Jr. Reliability of roentgenogram evaluation of pedicle screw position. Spine 1997, 22, 1249–1252. [Google Scholar] [CrossRef] [PubMed]
  14. Choma, T.J.; Denis, F.; Lonstein, J.E.; Perra, J.H.; Schwender, J.D.; Garvey, T.A.; Mullin, W.J. Stepwise methodology for plain radiographic assessment of pedicle screw placement: A comparison with computed tomography. Clin. Spine Surg. 2006, 19, 547–553. [Google Scholar] [CrossRef]
  15. Mason, A.; Paulsen, R.; Babuska, J.M.; Rajpal, S.; Burneikiene, S.; Nelson, E.L.; Villavicencio, A.T. The accuracy of pedicle screw placement using intraoperative image guidance systems: A systematic review. J. Neurosurg. Spine 2014, 20, 196–203. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Tonetti, J.; Boudissa, M.; Kerschbaumer, G.; Seurat, O. Role of 3D intraoperative imaging in orthopedic and trauma surgery. Orthop. Traumatol. Surg. Res. 2020, 106, S19–S25. [Google Scholar] [CrossRef] [PubMed]
  17. Sati, M.; Bourquin, Y.; Berlemann, U.; Nolte, L.P. Computer-assisted technology for spinal cage delivery. Oper. Tech. Orthop. 2000, 10, 69–76. [Google Scholar] [CrossRef]
  18. Strong, M.J.; Yee, T.J.; Khalsa, S.S.S.; Saadeh, Y.S.; Swong, K.N.; Kashlan, O.N.; Szerlip, N.J.; Park, P.; Oppenlander, M.E. The feasibility of computer-assisted 3D navigation in multiple-level lateral lumbar interbody fusion in combination with posterior instrumentation for adult spinal deformity. Neurosurg. Focus 2020, 49, E4. [Google Scholar] [CrossRef]
  19. Wang, M.; Li, D.; Shang, X.; Wang, J. A review of computer-assisted orthopaedic surgery systems. Int. J. Med. Robot. Comput. Assist. Surg. 2020, 16, 1–28. [Google Scholar] [CrossRef]
  20. Guéziec, A.; Kazanzides, P.; Williamson, B.; Taylor, R.H. Anatomy-based registration of CT-scan and intraoperative X-ray images for guiding a surgical robot. IEEE Trans. Med. Imaging 1998, 17, 715–728. [Google Scholar] [CrossRef]
  21. Sundar, H.; Khamene, A.; Xu, C.; Sauer, F.; Davatzikos, C. A novel 2D-3D registration algorithm for aligning fluoro images with 3D pre-op CT/MR images. In Proceedings of the Medical Imaging 2006: Visualization, Image-Guided Procedures, and Display, San Diego, CA, USA, 11–16 February 2006; Volume 6141, p. 61412K. [Google Scholar]
  22. Esfandiari, H.; Anglin, C.; Guy, P.; Street, J.; Weidert, S.; Hodgson, A.J. A comparative analysis of intensity-based 2D–3D registration for intraoperative use in pedicle screw insertion surgeries. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1725–1739. [Google Scholar] [CrossRef] [PubMed]
  23. Penney, G.P.; Batchelor, P.G.; Hill, D.L.; Hawkes, D.J.; Weese, J. Validation of a two-to three-dimensional registration algorithm for aligning preoperative CT images and intraoperative fluoroscopy images. Med. Phys. 2001, 28, 1024–1032. [Google Scholar] [CrossRef] [PubMed]
  24. Miao, S.; Wang, Z.J.; Liao, R. A CNN regression approach for real-time 2D/3D registration. IEEE Trans. Med. Imaging 2016, 35, 1352–1363. [Google Scholar] [CrossRef] [PubMed]
  25. Varnavas, A.; Carrell, T.; Penney, G. Fully automated 2D–3D registration and verification. Med. Image Anal. 2015, 26, 108–119. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, J.N.; Fan, Y.; Hao, D.J. Risk factors for robot-assisted spinal pedicle screw malposition. Sci. Rep. 2019, 9, 3025. [Google Scholar] [CrossRef] [Green Version]
  27. Härtl, R.; Lam, K.S.; Wang, J.; Korge, A.; Kandziora, F.; Audigé, L. Worldwide survey on the use of navigation in spine surgery. World Neurosurg. 2013, 79, 162–172. [Google Scholar] [CrossRef]
  28. Tian, N.F.; Huang, Q.S.; Zhou, P.; Zhou, Y.; Wu, R.K.; Lou, Y.; Xu, H.Z. Pedicle screw insertion accuracy with different assisted methods: A systematic review and meta-analysis of comparative studies. Eur. Spine J. 2011, 20, 846–859. [Google Scholar] [CrossRef] [Green Version]
  29. Beck, M.; Mittlmeier, T.; Gierer, P.; Harms, C.; Gradl, G. Benefit and accuracy of intraoperative 3D-imaging after pedicle screw placement: A prospective study in stabilizing thoracolumbar fractures. Eur. Spine J. 2009, 18, 1469–1477. [Google Scholar] [CrossRef] [Green Version]
  30. Costa, F.; Cardia, A.; Ortolina, A.; Fabio, G.; Zerbi, A.; Fornari, M. Spinal Navigation: Standard Preoperative Versus Intraoperative Computed Tomography Data Set Acquisition for Computer-Guidance SystemRadiological and Clinical Study in 100 Consecutive Patients. Spine 2011, 36, 2094–2098. [Google Scholar] [CrossRef]
  31. Hollenbeck, J.F.; Cain, C.M.; Fattor, J.A.; Rullkoetter, P.J.; Laz, P.J. Statistical shape modeling characterizes three-dimensional shape and alignment variability in the lumbar spine. J. Biomech. 2018, 69, 146–155. [Google Scholar] [CrossRef]
  32. Furrer, P.R.; Caprara, S.; Wanivenhaus, F.; Burkhard, M.D.; Senteler, M.; Farshad, M. Patient-specific statistical shape modeling for optimal spinal sagittal alignment in lumbar spinal fusion. Eur. Spine J. 2021, 30, 2333–2341. [Google Scholar] [CrossRef] [PubMed]
  33. Yu, W.; Tannast, M.; Zheng, G. Non-rigid free-form 2D–3D registration using a B-spline-based statistical deformation model. Pattern Recognit. 2017, 63, 689–699. [Google Scholar] [CrossRef] [Green Version]
  34. Baka, N.; Kaptein, B.L.; de Bruijne, M.; van Walsum, T.; Giphart, J.; Niessen, W.J.; Lelieveldt, B.P. 2D–3D shape reconstruction of the distal femur from stereo X-ray imaging using statistical shape models. Med. Image Anal. 2011, 15, 840–850. [Google Scholar] [CrossRef] [PubMed]
  35. Aubert, B.; Vazquez, C.; Cresson, T.; Parent, S.; de Guise, J.A. Toward automated 3D spine reconstruction from biplanar radiographs using CNN for statistical spine model fitting. IEEE Trans. Med. Imaging 2019, 38, 2796–2806. [Google Scholar] [CrossRef] [PubMed]
  36. Shen, L.; Zhao, W.; Xing, L. Patient-specific reconstruction of volumetric computed tomography images from a single projection view via deep learning. Nat. Biomed. Eng. 2019, 3, 880–888. [Google Scholar] [CrossRef]
  37. Ying, X.; Guo, H.; Ma, K.; Wu, J.; Weng, Z.; Zheng, Y. X2CT-GAN: Reconstructing CT from biplanar X-rays with generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 10619–10628. [Google Scholar]
  38. Kasten, Y.; Doktofsky, D.; Kovler, I. End-to-end convolutional neural network for 3D reconstruction of knee bones from bi-planar X-ray images. In Proceedings of the International Workshop on Machine Learning for Medical Image Reconstruction, Lima, Peru, 8 October 2020; Springer: Cham, Switzerland. [Google Scholar]
  39. Li, R.; Niu, K.; Wu, D.; Vander Poorten, E. A Framework of Real-time Freehand Ultrasound Reconstruction based on Deep Learning for Spine Surgery. In Proceedings of the 10th Conference on New Technologies for Computer and Robot Assisted Surgery, Barcelona, Spain, 28–30 September 2020. [Google Scholar]
  40. Uneri, A.; Goerres, J.; De Silva, T.; Jacobson, M.W.; Ketcha, M.D.; Reaungamornrat, S.; Kleinszig, G.; Vogt, S.; Khanna, A.J.; Wolinsky, J.P.; et al. Deformable 3D-2D Registration of Known Components for Image Guidance in Spine Surgery. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016, Athens, Greece, 17–21 October 2016; Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 124–132. [Google Scholar]
  41. Dufour, P.A.; Abdillahi, H.; Ceklic, L.; Wolf-Schnurrbusch, U.; Kowal, J. Pathology hinting as the combination of automatic segmentation with a statistical shape model. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Nice, France, 1–5 October 2012; Springer: Berlin, Germany, 2012; pp. 599–606. [Google Scholar]
  42. Tatarchenko, M.; Richter, S.R.; Ranftl, R.; Li, Z.; Koltun, V.; Brox, T. What do single-view 3d reconstruction networks learn? In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3405–3414. [Google Scholar]
  43. Kar, A.; Häne, C.; Malik, J. Learning a multi-view stereo machine. Adv. Neural Inf. Process. Syst. 2017, 2017, 365–376. [Google Scholar]
  44. Liebmann, F.; Roner, S.; von Atzigen, M.; Scaramuzza, D.; Sutter, R.; Snedeker, J.; Farshad, M.; Fürnstahl, P. Pedicle screw navigation using surface digitization on the Microsoft HoloLens. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1157–1165. [Google Scholar] [CrossRef]
  45. Lieberman, I.H.; Kisinde, S.; Hesselbacher, S. Robotic-assisted pedicle screw placement during spine surgery. JBJS Essent. Surg. Tech. 2020, 10, e0020. [Google Scholar] [CrossRef]
  46. Amiri, S.; Wilson, D.R.; Masri, B.A.; Anglin, C. A low-cost tracked C-arm (TC-arm) upgrade system for versatile quantitative intraoperative imaging. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 695–711. [Google Scholar] [CrossRef]
  47. Chintalapani, G.; Jain, A.K.; Burkhardt, D.H.; Prince, J.L.; Fichtinger, G. CTREC: C-arm tracking and reconstruction using elliptic curves. In Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Anchorage, AK, USA, 23–28 June 2008; IEEE: Washingotn, USA, 2008; pp. 1–7. [Google Scholar]
  48. Navab, N.; Bani-Hashemi, A.R.; Mitschke, M.M.; Holdsworth, D.W.; Fahrig, R.; Fox, A.J.; Graumann, R. Dynamic geometrical calibration for 3D cerebral angiography. In Proceedings of the Medical Imaging 1996: Physics of Medical Imaging, Newport Beach, CA, USA, 10–15 February 1996; SPIE: Bellingham, WA, USA, 1996; Volume 2708, pp. 361–370. [Google Scholar]
  49. Esfandiari, H.; Martinez, J.F.; Alvarez, A.G.; Guy, P.; Street, J.; Anglin, C.; Hodgson, A.J. An automated, robust and closed form mini-RSA system for intraoperative C-Arm calibration. In Proceedings of the CARS 2017—Computer Assisted Radiology and Surgery, Barcelona, Spain, 20–24 June 2017; Volume 12 (Suppl. S1), pp. S37–S38. [Google Scholar] [CrossRef] [Green Version]
  50. Kausch, L.; Thomas, S.; Kunze, H.; Privalov, M.; Vetter, S.; Franke, J.; Mahnken, A.H.; Maier-Hein, L.; Maier-Hein, K. Toward automatic C-arm positioning for standard projections in orthopedic surgery. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 1095–1105. [Google Scholar] [CrossRef]
  51. Esfandiari, H.; Andreß, S.; Herold, M.; Böcker, W.; Weidert, S.; Hodgson, A.J. A Deep Learning Approach for Single Shot C-Arm Pose Estimation. CAOS 2020, 4, 69–73. [Google Scholar]
  52. Al Arif, S.M.R.; Knapp, K.; Slabaugh, G. Fully automatic cervical vertebrae segmentation framework for X-ray images. Comput. Methods Programs Biomed. 2018, 157, 95–111. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Sa, R.; Owens, W.; Wiegand, R.; Chaudhary, V. Fast scale-invariant lateral lumbar vertebrae detection and segmentation in X-ray images. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1054–1057. [Google Scholar]
  54. Wang, L.; Xie, C.; Lin, Y.; Zhou, H.Y.; Chen, K.; Cheng, D.; Dubost, F.; Collery, B.; Khanal, B.; Khanal, B.; et al. Evaluation and comparison of accurate automated spinal curvature estimation algorithms with spinal anterior-posterior X-ray images: The AASCE2019 challenge. Med. Image Anal. 2021, 72, 102115. [Google Scholar] [CrossRef] [PubMed]
  55. Otake, Y.; Schafer, S.; Stayman, J.; Zbijewski, W.; Kleinszig, G.; Graumann, R.; Khanna, A.; Siewerdsen, J. Automatic localization of vertebral levels in x-ray fluoroscopy using 3D-2D registration: A tool to reduce wrong-site surgery. Phys. Med. Biol. 2012, 57, 5485. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Mushtaq, M.; Akram, M.U.; Alghamdi, N.S.; Fatima, J.; Masood, R.F. Localization and Edge-Based Segmentation of Lumbar Spine Vertebrae to Identify the Deformities Using Deep Learning Models. Sensors 2022, 22, 1547. [Google Scholar] [CrossRef]
  57. Kim, K.C.; Cho, H.C.; Jang, T.J.; Choi, J.M.; Seo, J.K. Automatic detection and segmentation of lumbar vertebrae from X-ray images for compression fracture evaluation. Comput. Methods Programs Biomed. 2021, 200, 105833. [Google Scholar] [CrossRef]
  58. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  59. Deng, Y.; Wang, C.; Hui, Y.; Li, Q.; Li, J.; Luo, S.; Sun, M.; Quan, Q.; Yang, S.; Hao, Y.; et al. CTSpine1K: A Large-Scale Dataset for Spinal Vertebrae Segmentation in Computed Tomography. arXiv 2021, arXiv:2105.14711. [Google Scholar]
  60. Goitein, M.; Abrams, M.; Rowell, D.; Pollari, H.; Wiles, J. Multi-dimensional treatment planning: II. Beam’s eye-view, back projection, and projection through CT sections. Int. J. Radiat. Oncol. Biol. Phys. 1983, 9, 789–797. [Google Scholar] [CrossRef]
  61. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, September 6–12 2014; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  62. Knapitsch, A.; Park, J.; Zhou, Q.Y.; Koltun, V. Tanks and temples: Benchmarking large-scale scene reconstruction. ACM Trans. Graph. 2017, 36, 1–13. [Google Scholar] [CrossRef]
  63. Xie, H.; Yao, H.; Zhang, S.; Zhou, S.; Sun, W. Pix2Vox++: Multi-scale context-aware 3D object reconstruction from single and multiple images. Int. J. Comput. Vis. 2020, 128, 2919–2935. [Google Scholar] [CrossRef]
  64. McCormick, M.M.; Liu, X.; Ibanez, L.; Jomier, J.; Marion, C. ITK: Enabling reproducible research and open science. Front. Neuroinformatics 2014, 8, 13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Lowekamp, B.; Chen, D.; Ibanez, L.; Blezek, D. The Design of SimpleITK. Front. Neuroinformatics 2013, 7, 45. [Google Scholar] [CrossRef] [PubMed]
  66. Pérez-García, F.; Sparks, R.; Ourselin, S. TorchIO: A Python library for efficient loading, preprocessing, augmentation and patch-based sampling of medical images in deep learning. Comput. Methods Programs Biomed. 2021, 208, 106236. [Google Scholar] [CrossRef] [PubMed]
  67. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 8024–8035. [Google Scholar]
  68. Falcon, W.; Borovec, J.; Wälchli, A.; Eggert, N.; Schock, J.; Jordan, J.; Skafte, N.; Bereznyuk, V.; Harris, E.; Murrell, T.; et al. Pytorchlightning/pytorch-lightning: 0.7.6 Release. Zenodo, 15 May 2020. [Google Scholar] [CrossRef]
  69. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [Green Version]
  70. Tunçer, N.; Kuyucu, E.; Sayar, Ş.; Polat, G.; Erdil, İ.; Tuncay, İ. Orthopedic surgeons’ knowledge regarding risk of radiation exposition: A survey analysis. SICOT-J 2017, 3, 29. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  71. Moore, C.; Liney, G.; Beavis, A.; Saunderson, J. A method to produce and validate a digitally reconstructed radiograph-based computer simulation for optimisation of chest radiographs acquired with a computed radiography imaging system. Br. J. Radiol. 2011, 84, 890–902. [Google Scholar] [CrossRef] [Green Version]
  72. Shiode, R.; Kabashima, M.; Hiasa, Y.; Oka, K.; Murase, T.; Sato, Y.; Otake, Y. 2D–3D reconstruction of distal forearm bone from actual X-ray images of the wrist using convolutional neural networks. Sci. Rep. 2021, 11, 15249. [Google Scholar] [CrossRef]
Figure 1. Inference pipeline of the proposed 3D shape reconstruction algorithm. Raw X-rays from multiple views are passed to a pre-processing module that estimates the imaging parameters (i.e., P-matrices) and crops the input images to individual lumbar vertebrae. The cropped images, along with their corresponding adjusted P-matrices are then passed to the X23D network to generate the 3D reconstructions. Segmentation masks can optionally be applied to the localized vertebrae before processing.
Figure 1. Inference pipeline of the proposed 3D shape reconstruction algorithm. Raw X-rays from multiple views are passed to a pre-processing module that estimates the imaging parameters (i.e., P-matrices) and crops the input images to individual lumbar vertebrae. The cropped images, along with their corresponding adjusted P-matrices are then passed to the X23D network to generate the 3D reconstructions. Segmentation masks can optionally be applied to the localized vertebrae before processing.
Jimaging 08 00271 g001
Figure 2. Pipeline of the proposed X23D shape reconstruction algorithm.
Figure 2. Pipeline of the proposed X23D shape reconstruction algorithm.
Jimaging 08 00271 g002
Figure 3. The interface of our custom-made DRR generation and annotation software. Intrinsic and extrinsic camera parameters can be adjusted on the left panel. DRR parameters such as image size and spacing can be configured. The middle part of the interface allows for applying segmentation filters on a vertebral level. Additionally, all the parameters can be passed to the back-end for batch processing of multiple CTs. The output of the software are DRRs with corresponding projection matrices, bounding boxes, and segmented areas in the COCO dataset’s format [61].
Figure 3. The interface of our custom-made DRR generation and annotation software. Intrinsic and extrinsic camera parameters can be adjusted on the left panel. DRR parameters such as image size and spacing can be configured. The middle part of the interface allows for applying segmentation filters on a vertebral level. Additionally, all the parameters can be passed to the back-end for batch processing of multiple CTs. The output of the software are DRRs with corresponding projection matrices, bounding boxes, and segmented areas in the COCO dataset’s format [61].
Jimaging 08 00271 g003
Figure 4. Training pipeline of the proposed X23D shape reconstruction algorithm. The 3D input, consisting of CTs and 3D segmentation masks, is passed to a DRR creation and annotation tool. The DRR tool outputs 2D DRRs, bounding boxes, segmentation masks of vertebrae, as well as the corresponding P-matrices. The bounding boxes and the DRRs are further processed to localized vertebrae regions before they are passed with the P-matrices to the training of the X23D network.
Figure 4. Training pipeline of the proposed X23D shape reconstruction algorithm. The 3D input, consisting of CTs and 3D segmentation masks, is passed to a DRR creation and annotation tool. The DRR tool outputs 2D DRRs, bounding boxes, segmentation masks of vertebrae, as well as the corresponding P-matrices. The bounding boxes and the DRRs are further processed to localized vertebrae regions before they are passed with the P-matrices to the training of the X23D network.
Jimaging 08 00271 g004
Figure 5. DRRs generated to train and evaluate our method showing an L1 vertebra. (a) Unsegmented. (b) Grayscale segmented using the segmentation masks in the patient CT scan. (c) Binarized version of (b).
Figure 5. DRRs generated to train and evaluate our method showing an L1 vertebra. (a) Unsegmented. (b) Grayscale segmented using the segmentation masks in the patient CT scan. (c) Binarized version of (b).
Jimaging 08 00271 g005
Figure 6. Visualized output of the X23D algorithm with unseen input. The upper row shows the predictions for L1 to L5 using four unsegmented images as input. The middle row shows the corresponding ground truth. The bottom row shows an overlay of prediction and ground truth.
Figure 6. Visualized output of the X23D algorithm with unseen input. The upper row shows the predictions for L1 to L5 using four unsegmented images as input. The middle row shows the corresponding ground truth. The bottom row shows an overlay of prediction and ground truth.
Jimaging 08 00271 g006
Figure 7. The 2D slices extracted from the results reported in Figure 6. The ground truth is depicted in white. Blue shows results using four unsegmented unseen views as input. Magenta shows results using four unseen grayscale segmented views.
Figure 7. The 2D slices extracted from the results reported in Figure 6. The ground truth is depicted in white. Blue shows results using four unsegmented unseen views as input. Magenta shows results using four unseen grayscale segmented views.
Jimaging 08 00271 g007
Figure 8. Visualization of the surface score on two X23D reconstructions each using four-view unseen and unsegmented inputs. Distance is expressed in color with distances bigger than 2% of the volume side length being black. (a) Visualization of point distances to compute the precision. (a) Distances measured from the target to the closest predicted point. (b) Visualization of point distances to compute the recall. (b) The distance from the prediction to the closest target point. Darker areas in (a) (precision) indicate areas that were reconstructed but are not present in the target (as seen in the spinous process and transverse area of (A). Darker areas in (b) (recall) indicate areas that are missing in the prediction. All points with distances bigger than a 1% threshold (defined in Section 2.4), ranging from red to black, contributed to lowering either recall or precision. The spinal canal was reconstructed correctly in both examples despite not being visible in input images. (B) Overall good precision, recall and, hence, surface score.
Figure 8. Visualization of the surface score on two X23D reconstructions each using four-view unseen and unsegmented inputs. Distance is expressed in color with distances bigger than 2% of the volume side length being black. (a) Visualization of point distances to compute the precision. (a) Distances measured from the target to the closest predicted point. (b) Visualization of point distances to compute the recall. (b) The distance from the prediction to the closest target point. Darker areas in (a) (precision) indicate areas that were reconstructed but are not present in the target (as seen in the spinous process and transverse area of (A). Darker areas in (b) (recall) indicate areas that are missing in the prediction. All points with distances bigger than a 1% threshold (defined in Section 2.4), ranging from red to black, contributed to lowering either recall or precision. The spinal canal was reconstructed correctly in both examples despite not being visible in input images. (B) Overall good precision, recall and, hence, surface score.
Jimaging 08 00271 g008
Table 1. Performance of our X23D algorithm. Best results presented in bold font. Quantitative comparisons are provided when the X23D network was trained on segmented and unsegmented 2D inputs. Additionally, the performance of X23D is compared to the results achieved through Pix2Vox++ trained on segmented inputs.
Table 1. Performance of our X23D algorithm. Best results presented in bold font. Quantitative comparisons are provided when the X23D network was trained on segmented and unsegmented 2D inputs. Additionally, the performance of X23D is compared to the results achieved through Pix2Vox++ trained on segmented inputs.
UnsegmentedSegmentedPix2Vox++ (Seg.)
GrayscaleBinaryBinary-Aug.
#Views48484448
F1 score0.88 ± 0.030.89 ± 0.040.89 ± 0.030.91 ± 0.020.90 ± 0.020.91 ± 0.020.85 ± 0.050.86 ± 0.06
Surface score0.71 ± 0.110.67 ± 0.120.73 ± 0.100.69 ± 0.100.70 ± 0.100.72 ± 0.090.58 ± 0.120.58 ± 0.12
IoU0.89 ± 0.030.89 ± 0.030.90 ± 0.020.91 ± 0.020.90 ± 0.030.91 ± 0.020.86 ± 0.040.87 ± 0.04
HD95 (mm)2.83 ± 1.723.21 ± 1.842.31 ± 0.943.79 ± 4.512.43 ± 0.922.20 ± 0.663.74 ± 2.013.76 ± 1.90
ASD (mm)0.73 ± 0.360.84 ± 0.430.67 ± 0.211.02 ± 0.660.73± 0.200.73 ± 0.181.01 ± 0.440.99 ± 0.39
Table 2. Performance of our X23D algorithm evaluated on individual vertebral levels by using four unsegmented views as input. Best results presented in bold font.
Table 2. Performance of our X23D algorithm evaluated on individual vertebral levels by using four unsegmented views as input. Best results presented in bold font.
4-Views
L1L2L3L4L5
F1 score0.89 ± 0.030.90 ± 0.020.89 ± 0.030.88 ± 0.030.86 ± 0.04
Surface score0.75 ± 0.090.76 ± 0.090.73 ± 0.090.69 ± 0.110.62 ± 0.11
IoU0.90 ± 0.020.90 ± 0.020.89 ± 0.020.89 ± 0.030.87 ± 0.03
HD95 (mm)2.39 ± 1.052.39 ± 1.122.87 ± 1.793.10 ± 2.463.38 ± 1.60
ASD (mm)0.64 ± 0.180.62 ± 0.160.67 ± 0.160.81 ± 0.650.90 ± 0.30
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jecklin, S.; Jancik, C.; Farshad, M.; Fürnstahl, P.; Esfandiari, H. X23D—Intraoperative 3D Lumbar Spine Shape Reconstruction Based on Sparse Multi-View X-ray Data. J. Imaging 2022, 8, 271. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging8100271

AMA Style

Jecklin S, Jancik C, Farshad M, Fürnstahl P, Esfandiari H. X23D—Intraoperative 3D Lumbar Spine Shape Reconstruction Based on Sparse Multi-View X-ray Data. Journal of Imaging. 2022; 8(10):271. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging8100271

Chicago/Turabian Style

Jecklin, Sascha, Carla Jancik, Mazda Farshad, Philipp Fürnstahl, and Hooman Esfandiari. 2022. "X23D—Intraoperative 3D Lumbar Spine Shape Reconstruction Based on Sparse Multi-View X-ray Data" Journal of Imaging 8, no. 10: 271. https://0-doi-org.brum.beds.ac.uk/10.3390/jimaging8100271

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop