Next Article in Journal
Numerical and Experimental Validation of a Supersonic Mixing Layer Facility
Previous Article in Journal
Nanostructured Na2CaP2O7: A New and Efficient Catalyst for One-Pot Synthesis of 2-Amino-3-Cyanopyridine Derivatives and Evaluation of Their Antibacterial Activity
Previous Article in Special Issue
Autonomous Prostate Segmentation in 2D B-Mode Ultrasound Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic 3D MRI-Ultrasound Registration for Image Guided Arthroscopy

1
Department of Electrical Engineering, Indian Institute of Technology Madras, Chennai 600036, India
2
Healthcare Technology Innovation Centre, Indian Institute of Technology Madras, Chennai 600036, India
3
Centre for Biomedical Technologies, Queensland University of Technology, Brisbane 4000, Australia
4
School of Mechanical, Medical and Process Engineering, Queensland University of Technology, Brisbane 4000, Australia
5
Department of Orthopaedic Surgery, Hakodate Central General Hospital, Sapporo 060-8604, Japan
6
Department of Orthopaedic Surgery, Hyogo College of Medicine, Nishinomiya 663-8501, Japan
7
School of Electrical Engineering and Robotics, Queensland University of Technology, Brisbane 4000, Australia
8
School of Clinical Sciences, Queensland University of Technology, Brisbane 4000, Australia
*
Author to whom correspondence should be addressed.
Submission received: 28 April 2022 / Revised: 26 May 2022 / Accepted: 26 May 2022 / Published: 28 May 2022
(This article belongs to the Special Issue Innovations in Ultrasound Imaging for Medical Diagnosis)

Abstract

:
Registration of partial view intra-operative ultrasound (US) to pre-operative MRI is an essential step in image-guided minimally invasive surgery. In this paper, we present an automatic, landmark-free 3D multimodal registration of pre-operative MRI to 4D US (high-refresh-rate 3D-US) for enabling guidance in knee arthroscopy. We focus on the problem of initializing registration in the case of partial views. The proposed method utilizes a pre-initialization step of using the automatically segmented structures from both modalities to achieve a global geometric initialization. This is followed by computing distance maps of the procured segmentations for registration in the distance space. Following that, the final local refinement between the MRI-US volumes is achieved using the LC 2 (Linear correlation of linear combination) metric. The method is evaluated on 11 cases spanning six subjects, with four levels of knee flexion. A best-case error of 1.41 mm and 2.34 and an average registration error of 3.45 mm and 7.76 is achieved in translation and rotation, respectively. An inter-observer variability study is performed, and a mean difference of 4.41 mm and 7.77 is reported. The errors obtained through the developed registration algorithm and inter-observer difference values are found to be comparable. We have shown that the proposed algorithm is simple, robust and allows for the automatic global registration of 3D US and MRI that can enable US based image guidance in minimally invasive procedures.

1. Introduction

Knee arthroscopy is a minimally invasive surgery (MIS) commonly used to treat intra-articular knee disorders [1]. The procedure is performed by inserting an arthroscope and a surgical tool into the knee joint through small incisions. The setup allows the surgeon to inspect the joint through the arthroscopic view displayed on a screen while altering the leg position to allow the scope to capture the required tissues. Despite offering several clinical advantages, arthroscopic procedures face long-standing challenges such as physically-demanding ergonomics during patient manipulation, lack of depth perception, limited field of view, and counter-intuitive hand-eye coordination between scope and surgical instruments.Even with skilled operator manipulation, a complex procedure like this can cause unintended trauma and post-operative complications [2,3,4].
To overcome these setbacks, an image guidance platform which uses both intra-operative and pre-operative imaging, in particular US and MRI, is proposed in [5]. This image guidance platform enables the surgeon to track and visualize both the target tissue distribution and the position of the surgical tool(s) in real-time. Intra-operative US is currently the only non-invasive, non-ionizing, imaging modality compatible with the knee arthroscopy surgical setup that can offer real-time imaging for guidance during the surgery [6,7]. However, US can only offer a limited field of view [8], with complex image interpretation. On the other hand, MRI provides a superior soft-tissue contrast, high resolution, as well as field of view covering the entire surgical site (total knee). Pre-operative MRI allows the surgeon to derive and use the information missing from the intra-operative US, enabling a comprehensive view of the bony and soft tissues. Therefore, the combination of both the modalities would synergistically integrate high contrast, large field-of-view, semantically labelable pre-operative MRI data into the fast, intra-operative, partial view 3D US and enable image-guided robotic surgery or MIS. This fusion not only provides a better visualization of structures of interest but also safely enhances the level of automation of MIS. This requires a quick and automatic registration of intra-operative US with pre-operative MRI.
Though relatively new for arthroscopy, multi-modal registration of US and MRI has been previously studied in the field of brain surgery where intra-operative US is used to track brain shift and tumor resection. The MICCAI Challenge CuRIOUS provided a public platform to benchmark MRI-US registration on the RESECT dataset [9]. A total of six fully automated registration algorithms are presented in this challenge [10]. These algorithms were tested on 12 patient volumes and the least registration error, which is 1.57 ± 0.96 mm, was reported by the team ImFusion who performed rigid registration using the LC 2 metric [11]. Another recent work [12] proposed combined segmentation based MRI–US registration approach in image-guided neurosurgery. Completely automatic registration of US and MRI volumes is a complex problem with no satisfactory solution so far [11]. It is primarily because of the entirely different principles used to capture the information in the two modalities [11]. Hence, the direct registration of the two modalities is not easily achievable in the image intensity domain using traditional metrics like the sum of squared distances, mutual information [13] or correlation [14]. Gradient-based methods have also been used, but they tend to discard valuable MRI intensity information [15]. Specifically targeting the multi-modal registration, a similarity measure, LC 2 , was proposed in [11], which is invariant to local changes with respect to brightness, contrast, etc. The metric was initially designed for US-CT [16] and later adapted to US-MRI registration.
The other challenging aspect of automating the MRI-US registration is the initialization. In standard clinical practice, close initialization is obtained by selecting 3D landmarks. Defining these landmarks requires not only tedious manual intervention but also keen knowledge of the targeted anatomy and clear geometrical understanding, especially in the case of 3D US, where there is limited volumetric coverage. Hence, to automate the registration process, instead of manually identifying landmarks in 3D, we show that obtaining coarse segmentations and using them for global initialization is a much more convenient alternative [17].
Segmentation-based initialization has been studied for prostate biopsy in [18], where registration was done between trans-rectal US and the corresponding MRI volumes. Even so, the target organ complexity or limited field-of-view of US could demand further regularization of initialization. Hence, we introduce a strong two-level initialization procedure which utilizes structural information of the tissues segmented followed by the distance maps, which were proven to be useful for correspondence estimation [18,19,20].
In this work, we present the first attempt to automatically register intra-operative 4D US to pre-operative 3D MRI as a first step towards the feasibility of autonomous or robotic-assisted knee arthroscopy. Initialization of registration requires strong anatomical priors, which we draw from automatic segmentation of structures from both the modalities. Given the cross-modal nature and partial view obtained in the intra-operative modality (as compared to the view available in pre-operative MRI), an alternative domain that utilizes both the computed priors is adopted—the Euclidean distance transform of segmented masks. This choice addresses the dual needs of working with cross-modal registration, and partial views. The registration of segmented masks in distance space gives us the initial transform for the registration of original volumes. A globally convergent similarity metric, LC 2 [11], is used for this registration.

2. Data

2.1. Ultrasound Data

Ultrasound volumes were acquired by an experienced orthopaedic surgeon F.S. using a Philips EpiQ7 US system and a Philips VL13-5 US probe. SonoCT real-time compound imaging technology and XRES image processing were selected on the US system to enhance image quality. Image acquisition settings have been optimized by a US specialist to work best for knee structure visualization: 13 MHz probe frequency, 4 cm penetration depth and far field focus. Given the restriction of movement due to the presence of surgical tools through the medial/lateral patellar portals, the US probe is positioned on the patellar tendon and parallel to the principal axis of tibia, as shown in Figure 1. 4D US sequences were acquired simulating several possible surgical scenarios for the four flexion angles, 0 , 30 , 60 and 90 [21]. In this paper, we use 4D US static sequences which were acquired when the leg is fixed at one of the reference knee flexion angles and the probe positioned on the patellar tendon, with the scanning plane parallel to the tibial shaft.
The neutral leg position, where the femoral shaft and tibial tubercle were aligned, was designated as 0 flexion. The flexion angles were then achieved by bending the leg such that the angle between femur and tibia with the neutral position was half the total angle. For angles other than 0 , the US scans were recorded with the leg placed on the custom made supporting cushion. Eleven 4D US static sequences comprising 61 3D US volumes of the anterior knee portion were collected from six healthy subjects at different flexion angles. Table 1 shows the subject wise angle information. Each 3D volume is of 272 × 510 × 256 voxels, corresponding approximately to a size of 4 cm × 4 cm × 5 cm, i.e., each voxel has the dimensions of about 0.14 mm × 0.09 mm × 0.21 mm. The Queensland University of Technology Ethics Committee (No. 1700001110) granted the ethics approval for the US and MRI data collection. The study objective and measurement procedure were explained to each volunteer and their written informed consent is procured.

2.2. MRI Data

MRI images were collected from the same volunteers from whom the US volumes were acquired. The volunteers’ knees were imaged using a 3T MRI system (Siemens Magnetom 3T Prisma, Erlangen, Germany). The MRI scans were acquired with knee positioned at 0 , 30 , 60 and 90 knee flexion for each volunteer. For the 0 knee flexion, the scan is acquired with the volunteer in the supine position. For the other knee flexion angles, the volunteer lay in the left/right lateral recumbent position for the corresponding leg to be scanned, using the cushion specifically designed to keep the leg at the desired angle. A total of 32 3D MRI volumes were obtained each of 256 × 256 × 192 voxels, where each voxel is of size 0.66 mm × 0.66 mm × 0.69 mm.

3. Methods

For the multi-modal volumetric registration of US with MRI, we propose a two-level initialization i.e., pre-initialisation using structural information followed by segmentation based registration in distance space. This, in turn, initialises US for the final step, LC 2 registration. The flow of the proposed algorithm is shown in Figure 2, and the detailed step-wise procedures are discussed in this section.

3.1. Segmentation

For the multi-modal registration we do in this paper, where we have non-corresponding and unequal number of landmarks, we use segmentations to aid the initialization (Appendix A). The choice of the segmentation algorithm and the number of the anatomical structures segmented generally depend on the specific application and target anatomy. These segmentations can easily be obtained by coarse annotations or fully automatic methods.

3.1.1. Segmentation of Femoral Cartilage from US

We adopt deep-learning based segmentation of femoral cartilage from the US volumes. Deep-learning algorithms provide non-iterative and rapid inference, which is beneficial since the US data are intra-operative and requires near-real-time processing for our target application. We trained a Mask R-CNN instance segmentation network [22] to simultaneously segment the medial and lateral femoral cartilage, also achieving identification along with region delineation [23]. The ground truth labels are marked interactively by the authors and validated by an expert clinician (refer Appendix A.1).

3.1.2. Segmentation of Femoral Cartilage from MRI

Automatic dense labeling of MRI subjects is done by using atlas-based segmentation. An open-source SPL Knee Atlas developed by The Surgical Planning Laboratory at Brigham and Women’s Hospital, Harvard Medical School [24], is elastically deformed onto our MRI volumes. The atlas is an MRI derived atlas generated through semi-automated segmentation algorithms and 3D image reconstruction techniques. Deformable registration of the atlas data onto the subject volume is performed using ANTs (Advanced Normalization Tools). Applying diffeomorphisms, the altas based segmentation computes a reversible transformation composed of a global affine transform and a dense vector field denoting the deformation suffered by the atlas to morph onto the subject MRI image [25]. Figure 3 shows the segmentation output achieved after the deformable registration is applied to the atlas.

3.2. Registration

3.2.1. Data Pre-Processing

Both MRI and US volumes are resampled to obtain an isotropic spacing, i.e., a spacing of 0.33 is chosen and set in all directions for both the modalities. Following that, the US volumes are modified (right-to-left changed to left-to-right or vice versa) in such a way so as to match the anatomical directions of MRI volumes. For example, conventions for left leg are, from the inferior apex of the patella to the superior surface of the tibia, i.e., along the direction of patellar ligament in sagittal plane and from medial femoral cartilage to lateral femoral cartilage in the transverse plane. As an effect of resampling, the size of the MRI image is very large compared to that of the US. It is required for both the modalities to have comparable sizes for close registration and hence an appropriate ROI is chosen from the MRI volume.

3.2.2. Pre-Initialization

During the MIS, US probe is positioned on the patellar tendon and fat pad as shown in Figure 1. We model the US view in the MRI as an oriented bounding box of the anterior knee anatomy. We use the fat pad structure to arrive at a cuboidal region that is close to the view obtained by the US, for the initialization. The infrapatellar fat pad segmentation is obtained by using the method in Section 3.1.2. The principal axes of the fat pad, which are the normals to the faces of the bounding box, are computed from the segmentation, using principal component analysis (PCA) on the points on the fat pad surface.
We now use anatomical information to align the US beam direction to one of the axes. Given the probe placement on the patellar tendon, the fat pad, which is immediately beneath, has almost similar orientation. This enables us to align the inferio-superior direction to one of the fat pad axes directly. The centroid, C, of the fat pad, projected outward gives a point on the anterior fat pad surface, which is our initialization for probe center. Of the three eigenvectors obtained, the vector, v s i , which goes from superior to inferior direction of the fat pad, gives the rotation matrix for the US volume. Let p 1 and p 2 be the boundary points of the fat pad that are intercepted by v s i . The anterior point, p 1 , gives the required translation for the US volume (refer Appendix B).
Figure 4 shows the fat pad segmentation, the boundary points intercepted by v s i , and the principal axes of fat pad. O is the centre of the cropped MRI image. The rotation matrix given by v s i and the translation vector given by p 1 together give us the initial transform of the US.

3.2.3. Initialization by Euclidean Distance Transform

Registering the obtained set of label masks (Section 3.1.1 and Section 3.1.2) extracted from US and MRI volumes can be posed as a pseudo/mono-modal registration problem for which there are many traditional registration approaches available. However, these methods tend to face issues like incorrect error estimation and underperformance due to a lack of efficient optimization. Since labels are encoded with numerical values, one incorrectly registered voxel with label difference of two gives the same error as two incorrectly registered voxels with label difference of one leading to incorrect estimation of registration error. Metrics like Dice are good for finding an overlap of segmentations but might not be sufficient as a metric for registration, since there are equivalences of values at multiple shifts, which would imply multiple local minima in the error surface, which could trap iterative methods. In addition, in this case, a bad initialization might lead to low capture range as the homogeneous label regions such as the background label, do not produce meaningful information for the optimization and therefore result in poor registration [17]. Furthermore, the structures in MRI are complete, whereas, in US, it is dependent on the placement of the probe, so complete volume overlap of the two masks is not expectable from a practical standpoint. Considering these issues, in our method, we transform the masks to a geodesic, using distance transform which makes the support region of the mask larger, thereby increasing capture range. Hence, this enables working with partial segmentations. Each label map is considered as a different binary image, I n , which has two classes foreground f g and background b g .
Distance transform assigns a number to a voxel which is the distance from that voxel to the nearest non-zero voxel. Euclidean distance transform ϕ is applied on the pre-initialized I n in Equation (1),
ϕ m , n = ϕ ( I m , n ) a n d ϕ f , n = ϕ ( I f , n )
where ϕ m , n and ϕ f , n are distance maps of moving and fixed modalities respectively, and are derived using Equation (2),
ϕ ( I n ( X ) ) = 0 , I n ( X ) f g , n m i n ( X X 0 ) , I n ( X 0 ) f g , n a n d I n ( X ) b g , n
X X 0 is the Euclidean distance between two points. This transform is applied on all the label classes separately. The initial transform is obtained by minimizing the following Equation (3):
m i n T S E ( 3 ) n = 1 N S f | ( ϕ m , n . T ) ( x ) ϕ l , n ( x ) | 2 d x
where T S E ( 3 ) is the required rigid transformation to initialize ϕ m , n , S f is the set of coarse segmentations in fixed volume and N is the total number of segmentations. Since the distance transform of a label map is also dependent on the distance between the segmentation and boundary, the ROI of MRI should be selected accordingly for the minimization problem to work more efficiently. Intuitively, the dimensions and centre for cropping the ROI are selected based on the length and width of femoral cartilage, f c . The centre of the segmented femoral cartilage is set as the centre for ROI, and the dimensions are determined by Equation (4):
X M R I = l e n g t h o f f c M R I l e n g t h o f f c U S · X U S
where X M R I denotes the required x , y , z dimensions of ROI, and X U S denotes the x , y , z dimensions of US volume and length of femoral cartilage is the length along that particular axis.

3.2.4. Registration

For the registration of the intra-operative US volumes to pre-operative MRI volumes, we use LC 2 , a multi-modal similarity metric [11,16]. It is a powerful measure which is invariant to local changes such as brightness and contrast and is proven to work for US-MRI registration [11]. LC 2 correlates both the MRI intensity values and the spatial gradient magnitude with the US. We use this cost function along with optimal patch size to perform the registration. During the registration, it is necessary to look out for and eliminate the unwanted parameter updates in order to avoid unstable results. Therefore, we optimize all the parameters of the obtained rigid transform after every evaluation. It is difficult to compute the analytic derivative of LC 2 which is of least square fitting and hence, as in [11], we use Bound Optimization by Quadratic Approximation (BOBYQA) [26] optimization, which is capable of internally creating its own derivative approximations resulting in fewer evaluations than any other direct search methods.

4. Results and Discussion

To evaluate the performance of the proposed algorithm, an experienced orthopaedic surgeon F.S., manually registered the MRI-US pairs used in this study. The registrations were performed in ImFusion software. The MRI-US volumes were rotated/translated in a common reference system by selecting the rotation/translation values associated with each volume along three orthogonal planes. The software allowed users to visualize the two volumes simultaneously directly while the rotation/translation values were modified by the user.
The MRI volume is kept fixed, while the US volume is rigidly moved to overlay the tibia and the femur surfaces in the two modalities along the three orthogonal planes. The transformation obtained as a result of this process which the surgeon (F.S.) performed manually is considered as the ground truth transform. For every registration, the three translation (mm) and rotation (degree) parameters were obtained from the transform. A similar kind of transform is obtained as a result of registering the two modalities using the proposed algorithm. In-order-to calculate the registration error, the MRI volume is fixed in the position selected in the manual registration performed by the surgeon and the transformation parameters of the US volumes (translation and rotation) are compared as in [27]. The norm of the difference vector of the parameters, computed using Equations (5) and (6), give the translation (t) and rotation (r) error values for each MRI-US pair,
t = ( x G T x o ) 2 + ( y G T y o ) 2 + ( z G T z o ) 2
r = ( θ G T θ o ) y z 2 + ( θ G T θ o ) x z 2 + ( θ G T θ o ) x y 2
where subscript G T refers to the ground truth parameter value, and subscript o refers to the outputted parameter value from the developed registration method. x, y, z represent translation parameters in each plane and θ represents rotation parameter for y z , x z , x y planes.
The proposed algorithm is tested on 11 MRI-US pairs obtained from six volunteers at four different knee flexion angles. The transform obtained at the end of each of the three stages of the algorithm is compared against the ground truth transform, and the translation/rotation error is computed using Equations (5) and (6). Table 2 shows the detailed evaluated error values at each stage from the registration method developed. To calculate the initial error, the transforms of US and MRI volumes are set to identity. It can be clearly seen from Table 2 that there is a decreasing trend in the registration error and that each stage of the registration contributes notably to reduce this error. The pre-initialization step alone, in particular, significantly brings down the initial error by 76.8%, thereby assisting the subsequent steps in the algorithm to achieve global registration more efficiently.
Furthermore, to estimate the importance of including the distance transform step, a cost function analysis is performed. The final cost function, which determines the accuracy of the registration, LC 2 , is plotted against any two parameters from the initial transform. We selected the rotation x, rotation z parameters from the output transforms of both pre-in and DT stages, varied their values within the range ±10 and plotted against the LC 2 metric for a given subject. The resulting curves (Figure 5) revealed that the inclusion of distance transform produces a more peaked objective surface (Figure 5b), which would enable better convergence, whereas, without the distance transform (i.e., using LC 2 directly on volumes), the objective surface is undulated with no clear optimum (Figure 5a). In Figure 5b, a global peak is seen with maximum LC 2 value of 0.2, whereas, in Figure 5a, the maximum LC 2 value is observed to be a little lesser with minor local optimum. It can therefore be inferred that incorporating the distance transform step in the algorithm helps achieve a more accurate global registration.
To demonstrate the step-wise decreasing trend in the registration error from Table 2, the registration output at the end of each of the four stages for subject 5 overlaid on their respective ground truths is shown in Figure 6. Figure 6a shows the initial gap between the initial position of the US image and the ground-truth solution provided by the expert and Figure 6b shows the position after the pre-initialization, and it can be seen that this step provides a good initial alignment. The last two steps, Figure 6c,d, help with further fine-tuning and lowering the registration error. It can be spotted from Figure 6d that the patella, the medial femoral cartilage and the patellar shadow of the output red slice overlap almost perfectly with blue ground truth slice. The same registration result is overlaid onto the respective MRI of subject 5 and shown in Figure 7. Furthermore, to assess the efficiency of LC 2 , randomized trials were executed on all 11 data sets, repeating the rigid registration for randomly displaced starting positions, i.e., within the range of ±10 mm/ for all six parameters resulting in a total of 1100 registration attempts. Tracing the RMS error and the value of LC 2 metric, we observed that, in many cases, LC 2 is resilient to perturbations in initialization, and hence is an appropriate measure especially for a multi-modal 3D MRI-US registration in a realistic clinical setup. The results for all subjects are shown in Figure 8, including the percentage of converged executions.
To assess the reproducibility of the manual registrations here used as ground truths, a second experienced orthopaedic surgeon (Y.T. 5 ) performed the manual registrations for a subset of the MRI-US pairs (8 out of 11). The second surgeon utilized the MRI alignment provided by the first expert and then performed the registration between the MRI-US pairs independently from the first surgeon, following the same procedure. Since for each MRI-US pair registered by both surgeons, the MRI volume is fixed in a common position, the difference between the US volume translation/rotation values selected by the two surgeons is utilized to quantify the inter-observer variability as shown in [27]. The inter-observer reproducibility study results reported in Table 3 show that the mean difference between the manual registrations performed by the two surgeons is 4.41 mm ± 1.88 mm for translation and 7.77 ± 2.80 for rotation. The difference in these registrations could be because the rigid match itself might be insufficient for alignment between the knee US and MRI volumes. This is possibly a result of minor leg movements from the reference position during the MRI acquisition. Moreover, during the US acquisition, it is also possible to have small shifts of the patella due to the pressure exerted by the probe that could have possibly affected the overlap of this bone in the two modalities. This effect might have introduced a degree of subjectivity in finding the best compromise to align the bone-like surfaces in the two modalities resulting in a possible mismatch among the operators.
In both our achieved results (Table 2) and inter-operator study (Table 3), it can be observed that the registration error increases as the angle of flexion increases. It can be highlighted from comparing these tables that the registration error of the proposed automated algorithm for almost all tested cases is either smaller or comparable to the inter-observer registration difference (p-value < 0.002 ) and thus our registration method can be considered comparable to clinical standards. Figure 9 shows the registration output of subject 6 in frontal, axial and lateral planes for all the four flexion angles. It can clearly be seen from the overlapping images that the registration error (in particular the rotation) keeps increasing as the flexion angle is increased. A possible explanation for this could be, as the flexion angle increases, the lateral and medial femoral cartilage, which are the basis of our algorithm, shift posteriorly and go behind the patella and hence cannot clearly be distinguished in the US volume. In addition, as the knee flexion increases, the area of femoral cartilage in the knee gap decreases, thereby hindering the accuracy of initialization.

5. Conclusions

Automatic registration of pre-operative MRI with intra-operative US enables real-time guidance during the surgery and can enhance the quality and safety of image-guided knee arthroscopy. In this paper, we have introduced a fully automated registration algorithm consisting of a segmentation based pre-initialization, followed by a distance transform step to achieve global initialization. This is followed by registration with an LC 2 similarity metric which uses the total amount of available image information from both the modalities by locally matching US intensities to both MRI intensity and gradient magnitude. Our method is evaluated on 11 MRI-US pairs from six volunteers at different knee flexion angles and resulted in a mean registration error of 3.45 mm and 7.76 in translation and rotation, respectively. In addition, an inter-observer variability study is performed, comparing the registrations performed by two experienced surgeons and resulted in a mean difference between the registrations of 4.41 mm and 7.77 . Although the registration error of our algorithm increased with the angle of flexion of the knee, mean registration error of the proposed method is smaller than inter-observer variability (p-value < 0.002 ). We have shown that the proposed automated 3D multi-modal registration of US and MRI is globally convergent, robust and efficient. For the US-MRI volumetric registration, the proposed method can form the starting point for developing more complex transforms, including deformable, and can be explored as part of the future work.

Author Contributions

Conceptualization, K.R.; Data curation, F.S. and T.Y.; Formal analysis, G.K.; Funding acquisition, A.K.P. and M.S.; Investigation, M.A.; Methodology, G.K.; Project administration, A.K.P. and M.S.; Resources, M.A.; Supervision, K.R. and D.F.; Validation, J.S.; Visualization, G.K.; Writing—original draft, G.K.; Writing—review and editing, M.A. and K.R. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Indo-Australian Biotechnology Fund through Department of Biotechnology, Government of India (Intelligent Robotic Imaging System for Keyhole Surgeries) under Grant No. AISRF53820.

Institutional Review Board Statement

The Queensland University of Technology Ethics Committee (No. 1700001110) granted the ethics approval for the Ultrasound and MRI data collection.

Informed Consent Statement

The study objective and measurement procedure were explained to each participant and their written informed consent was procured.

Data Availability Statement

The data presented in this study are available on request from the corresponding author, D.F.

Acknowledgments

The authors would like to thank Rajeswaran Rangasami, diagnostic radiologist and senior consultant at Sri Ramachandra Medical Centre, and professor in Radiology at Sri Ramachandra Institute for Higher Education and Research, Chennai, India, for assisting in annotation of the US volumes. Maria Antico acknowledges her fellowship provided by the Centre for Biomedical Technologies, Queensland University of Technology, Australia.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Segmentations

In order to automate our whole registration process, we generate the required segmentations using completely automatic and robust methods. Let V f : S f R and V m : S m R denote the fixed ( V f ) and moving ( V m ) volumes, respectively, where S f and S m are set of all possible segmentations and S f , S m R 3 as shown in Equation (A1). The first step is to create N coarse segmentations for both V f and V m i.e., we take two partitions of S f and S m ,
n = 1 N S f , n S f and n = 1 N S m , n S m

Appendix A.1. Segmentation of Femoral Cartilage from US

For the Mask R-CNN deep-learning technique, we adopt segmentation of femoral cartilage from US volumes, and the ground truth labels are marked interactively by the authors and validated by an expert clinician. The annotation of the medial and the lateral femoral cartilage is done using the segment editor GUI in Slicer-3D The contours are drawn on the highest resolution plane, i.e., the sagittal plane of US volumes while parallelly visualizing its axial and coronal projections. The US images are split with a 3:2:2 ratio for training, validation and testing, respectively. The network is initially pretrained with both COCO and ImageNet datasets. For training the network, all hyperparameters are tuned and set experimentally. Figure A1 shows the end-to-end pipeline of the working of Mask R-CNN. A sample of segmented output along with its ground truth is shown in Figure A1g.
Figure A1. Flow diagram of the network (a) original image; (b) pre-processed image; (c) feature-maps; (d) proposed ROIs; (e) ROI Alignment output; (f) boundary-box regressor; (g) predicted mask (cyan) along with the ground truth (white) (Conv.—convolutional layers, FC—Fully Convolutional Network, ROI—Region of Interest).
Figure A1. Flow diagram of the network (a) original image; (b) pre-processed image; (c) feature-maps; (d) proposed ROIs; (e) ROI Alignment output; (f) boundary-box regressor; (g) predicted mask (cyan) along with the ground truth (white) (Conv.—convolutional layers, FC—Fully Convolutional Network, ROI—Region of Interest).
Applsci 12 05488 g0a1

Appendix B. Pre-Initialization

The principal axes of the fat pad, which are the normals to the faces of the bounding box, are computed from the segmentation, using principal component analysis (PCA) on the points on the fat pad surface. Let the data matrix with n surface points be, P = [ p 1 , p 2 , p 3 p n ] R 3 × n where each p i is a boundary voxel. The eigenvectors of the covariance, C o v , of the data matrix give the principal axes Equation (A2),
( C o v λ I ) v = 0 ; v T . v = 1
where v R 3 is the eigenvector, and the λ is the eigenvalue. Of the three eigenvectors obtained, the vector, v s i , which goes from superior to inferior direction of the fat pad, gives the rotation matrix for the US volume. In addition, the point on the boundary of the fat pad through which v s i passes, i.e., intercepts 0 or 180 angle, gives the required translation for the US volume. This point can be calculated by computing the dot product of v s i and all boundary voxels, Equation (A3).
cos θ = ( p i c ) . v s i | p i c | . | v s i | ; p i P a n d C = ( x ^ , y ^ , z ^ ) ,
where centroid C is the mean of all points in P. There will be two closest points on the boundary which intercept 0 and 180 with v s i , and the one farther away from the centre of the MRI image is the point that gives the translation vector for the US. Let p 1 and p 2 be the obtained closest points, and O be the centre of the cropped MRI image. The required point p is determined by Equation (A4),
p = p 1 , | | p 1 O | | > | | p 2 O | | p 2 , | | p 1 O | | < | | p 2 O | |
The rotation matrix given by v s i and the translation vector given by obtained point p together give us the initial transform of the US.

References

  1. Schena, A.; Ross, G. Knee Arthroscopy: Technique and Normal Anatomy. In Knee Arthroscopy; Springer: New York, NY, USA, 2009; pp. 1–10. [Google Scholar] [CrossRef]
  2. Jaiprakash, A.; O’Callaghan, W.B.; Whitehouse, S.L.; Pandey, A.; Wu, L.; Roberts, J.; Crawford, R.W. Orthopaedic surgeon attitudes towards current limitations and the potential for robotic and technological innovation in arthroscopic surgery. J. Orthop. Surg. 2017, 25, 2309499016684993. [Google Scholar] [CrossRef] [PubMed]
  3. Price, A.J.; Erturan, G.; Akhtar, K.; Judge, A.; Alvand, A.; Rees, J.L. Evidence-based surgical training in orthopaedics. Bone Jt. J. 2015, 97-B, 1309–1315. [Google Scholar] [CrossRef]
  4. Banach, A.; Strydom, M.; Jaiprakash, A.; Carneiro, G.; Eriksson, A.; Crawford, R.; McFadyen, A. Visual Localisation for Knee Arthroscopy. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 2137–2145. [Google Scholar] [CrossRef] [PubMed]
  5. Wu, L.; Jaiprakash, A.; Pandey, A.K.; Fontanarosa, D.; Jonmohamadi, Y.; Antico, M.; Strydom, M.; Razjigaev, A.; Sasazawa, F.; Roberts, J.; et al. 29-Robotic and Image-Guided Knee Arthroscopy. In Handbook of Robotic and Image-Guided Surgery; Abedin-Nasab, M.H., Ed.; Elsevier: Amsterdam, The Netherlands, 2020; pp. 493–514. [Google Scholar] [CrossRef]
  6. Antico, M.; Sasazawa, F.; Wu, L.; Jaiprakash, A.; Roberts, J.; Crawford, R.; Pandey, A.K.; Fontanarosa, D. Ultrasound guidance in minimally invasive robotic procedures. Med. Image Anal. 2019, 54, 149–167. [Google Scholar] [CrossRef]
  7. Chopra, S.S.; Hünerbein, M.; Eulenstein, S.; Lange, T.; Schlag, P.M.; Beller, S. Development and validation of a three-dimensional ultrasound based navigation system for tumor resection. Eur. J. Surg. Oncol. 2008, 34, 456–461. [Google Scholar] [CrossRef]
  8. Kim, K.; Choi, H. High-efficiency high-voltage class F amplifier for high-frequency wireless ultrasound systems. PLoS ONE 2021, 16, e0249034. [Google Scholar] [CrossRef] [PubMed]
  9. Xiao, Y.; Fortin, M.; Unsgård, G.; Rivaz, H.; Reinertsen, I. REtroSpective Evaluation of Cerebral Tumors (RESECT): A clinical database of pre-operative MRI and intra-operative ultrasound in low-grade glioma surgeries. Med. Phys. 2017, 44, 3875–3882. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Xiao, Y.; Rivaz, H.; Chabanas, M.; Fortin, M.; Machado, I.; Ou, Y.; Heinrich, M.P.; Schnabel, J.A.; Zhong, X.; Maier, A.; et al. Evaluation of MRI to Ultrasound Registration Methods for Brain Shift Correction: The CuRIOUS2018 Challenge. IEEE Trans. Med. Imaging 2020, 39, 777–786. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Wein, W.; Ladikos, A.; Fuerst, B.; Shah, A.; Sharma, K.; Navab, N. Global Registration of Ultrasound to MRI Using the LC2 Metric for Enabling Neurosurgical Guidance. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2013, Nagoya, Japan, 22–26 September 2013; Mori, K., Sakuma, I., Sato, Y., Barillot, C., Navab, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 34–41. [Google Scholar] [CrossRef] [Green Version]
  12. Nitsch, J.; Klein, J.; Dammann, P.; Wrede, K.; Gembruch, O.; Moltz, J.; Meine, H.; Sure, U.; Kikinis, R.; Miller, D. Automatic and efficient MRI-US segmentations for improving intraoperative image fusion in image-guided neurosurgery. Neuroimage Clin. 2019, 22, 101766. [Google Scholar] [CrossRef] [PubMed]
  13. Huang, X.; Hill, N.A.; Ren, J.; Guiraudon, G.; Boughner, D.; Peters, T.M. Dynamic 3D Ultrasound and MR Image Registration of the Beating Heart. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2005, Palm Springs, CA, USA, 26–29 October 2005; Duncan, J.S., Gerig, G., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 171–178. [Google Scholar] [CrossRef] [Green Version]
  14. Penney, G.; Blackall, J.; Hamady, M.; Sabharwal, T.; Adam, A.; Hawkes, D. Registration of freehand 3D ultrasound and magnetic resonance liver images. Med. Image Anal. 2004, 8, 81–91. [Google Scholar] [CrossRef] [PubMed]
  15. Nigris, D.D.; Collins, D.L.; Arbel, T. Fast rigid registration of pre-operative magnetic resonance images to intra-operative ultrasound for neurosurgery based on high confidence gradient orientations. Int. J. Comput. Assist. Radiol. Surg. 2013, 8, 649–661. [Google Scholar] [CrossRef] [PubMed]
  16. Wein, W.; Brunke, S.; Khamene, A.; Callstrom, M.R.; Navab, N. Automatic CT-ultrasound registration for diagnostic imaging and image-guided intervention. Med. Image Anal. 2008, 12, 577–585. [Google Scholar] [CrossRef] [PubMed]
  17. Rackerseder, J.; Baust, M.; Göbl, R.; Navab, N.; Hennersperger, C. Initialize Globally Before Acting Locally: Enabling Landmark-Free 3D US to MRI Registration. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2018, Granada, Spain, 16–20 September 2018; Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 827–835. [Google Scholar] [CrossRef] [Green Version]
  18. Fedorov, A.; Khallaghi, S.; Sánchez, C.A.; Lasso, A.; Fels, S.; Tuncali, K.; Sugar, E.N.; Kapur, T.; Zhang, C.; Wells, W.; et al. Open-source image registration for MRI–TRUS fusion-guided prostate interventions. Int. J. Comput. Assist. Radiol. Surg. 2015, 10, 925–934. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Slavcheva, M.; Kehl, W.; Navab, N.; Ilic, S. SDF-2-SDF: Highly Accurate 3D Object Reconstruction. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 680–696. [Google Scholar] [CrossRef]
  20. Canalini, L.; Klein, J.; Miller, D.; Kikinis, R. Registration of Ultrasound Volumes Based on Euclidean Distance Transform. In Large-Scale Annotation of Biomedical Data and Expert Label Synthesis and Hardware Aware Learning for Medical Imaging and Computer Assisted Intervention; Zhou, L., Heller, N., Shi, Y., Xiao, Y., Sznitman, R., Cheplygina, V., Mateus, D., Trucco, E., Hu, X.S., Chen, D., et al., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 127–135. [Google Scholar] [CrossRef] [Green Version]
  21. Antico, M.; Sasazawa, F.; Takeda, Y.; Jaiprakash, A.T.; Wille, M.L.; Pandey, A.K.; Crawford, R.; Fontanarosa, D. 4D Ultrasound-Based Knee Joint Atlas for Robotic Knee Arthroscopy: A Feasibility Study. IEEE Access 2020, 8, 146331–146341. [Google Scholar] [CrossRef]
  22. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar] [CrossRef]
  23. Kompella, G.; Antico, M.; Sasazawa, F.; Jeevakala, S.; Ram, K.; Fontanarosa, D.; Pandey, A.K.; Sivaprakasam, M. Segmentation of Femoral Cartilage from Knee Ultrasound Images Using Mask R-CNN. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 966–969. [Google Scholar] [CrossRef] [Green Version]
  24. Richolt, J.; Jakab, M.; Kikinis, R. MRI-Based Atlas of the Knee; Technical Report; Surgical Planning Laboratory, Department of Radiology, Brigham and Women’s Hospital, Harvard Medical School: Boston, MA, USA, 2015. [Google Scholar]
  25. Avants, B.B.; Tustison, N.J.; Stauffer, M.; Song, G.; Wu, B.; Gee, J.C. The Insight ToolKit image registration framework. Front. Neuroinform. 2014, 8, 44. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Powell, M.J.D. The BOBYQA Algorithm for Bound Constrained Optimization without Derivatives; University of Cambridge: Cambridge, UK, 2009. [Google Scholar]
  27. Vaarkamp, J. Reproducibility of interactive registration of 3D CT and MR pediatric treatment planning head images. J. Appl. Clin. Med. Phys. 2001, 2, 131–137. [Google Scholar] [CrossRef] [PubMed]
Figure 1. US probe positioning during the knee scan with the corresponding schematic in sagittal view.
Figure 1. US probe positioning during the knee scan with the corresponding schematic in sagittal view.
Applsci 12 05488 g001
Figure 2. Flowchart of the proposed algorithm. (a,b) input MRI and US volumes; (c) segmented femoral cartilage from MRI using the atlas based deformable registration performed by ANTs (Advanced Normalization Tools); (d) segmented femoral cartilage from US using the Mask R-CNN framework; (eg) registration output after each stage of the algorithm (US slice shown in blue, overlaid on a respective MRI slice).
Figure 2. Flowchart of the proposed algorithm. (a,b) input MRI and US volumes; (c) segmented femoral cartilage from MRI using the atlas based deformable registration performed by ANTs (Advanced Normalization Tools); (d) segmented femoral cartilage from US using the Mask R-CNN framework; (eg) registration output after each stage of the algorithm (US slice shown in blue, overlaid on a respective MRI slice).
Applsci 12 05488 g002
Figure 3. Coronal, axial, sagittal views of the segmentations obtained from knee MRI using deformable registration.
Figure 3. Coronal, axial, sagittal views of the segmentations obtained from knee MRI using deformable registration.
Applsci 12 05488 g003
Figure 4. Illustration of fat-pad along with its eigenvectors and obtained boundary points.
Figure 4. Illustration of fat-pad along with its eigenvectors and obtained boundary points.
Applsci 12 05488 g004
Figure 5. Surface plots of LC 2 cost function against rotation parameters when (a) registration is performed directly after Pre-in, without including the DT step and (b) registration is performed with the DT step.
Figure 5. Surface plots of LC 2 cost function against rotation parameters when (a) registration is performed directly after Pre-in, without including the DT step and (b) registration is performed with the DT step.
Applsci 12 05488 g005
Figure 6. A sagittal slice is extracted from a 3D US volume and output at four different stages of the method is shown along with the ground truth. Algorithm’s output is shown in red, and the ground truth slice is shown in blue. From top-left: (a) initial orientation; (b) output after pre-initialization; (c) output after applying distance transforms; (d) final output after implementing LC 2 registration.
Figure 6. A sagittal slice is extracted from a 3D US volume and output at four different stages of the method is shown along with the ground truth. Algorithm’s output is shown in red, and the ground truth slice is shown in blue. From top-left: (a) initial orientation; (b) output after pre-initialization; (c) output after applying distance transforms; (d) final output after implementing LC 2 registration.
Applsci 12 05488 g006
Figure 7. Final registration output of Subject 5: US aligned on respective MRI.
Figure 7. Final registration output of Subject 5: US aligned on respective MRI.
Applsci 12 05488 g007
Figure 8. Random study registration results with 100 randomly initialized transformations for 11 subject cases along with the percentage of converged optimizations.
Figure 8. Random study registration results with 100 randomly initialized transformations for 11 subject cases along with the percentage of converged optimizations.
Applsci 12 05488 g008
Figure 9. Output of proposed registration method for subject 6 overlaid on its respective ground truth for all four of the flexion angles. The output of the proposed algorithm is shown in red and ground truth for the study is shown in blue.
Figure 9. Output of proposed registration method for subject 6 overlaid on its respective ground truth for all four of the flexion angles. The output of the proposed algorithm is shown in red and ground truth for the study is shown in blue.
Applsci 12 05488 g009
Table 1. Scans acquired at different angles of flexion for different subjects.
Table 1. Scans acquired at different angles of flexion for different subjects.
0 30 60 90
Volunteer 1Volunteer 3Volunteer 3Volunteer 3
Volunteer 2Volunteer 6Volunteer 6Volunteer 6
Volunteer 4
Volunteer 5
Volunteer 6
Table 2. Table showing a decreasing trend in the translation and rotation error (left to right) for each subject computed after each step of the proposed registration algorithm. In-Err: Initial error, Pre-in: Error after pre-initialization, DT: Error after applying Distance Transform, LC 2 : error after final LC 2 registration.
Table 2. Table showing a decreasing trend in the translation and rotation error (left to right) for each subject computed after each step of the proposed registration algorithm. In-Err: Initial error, Pre-in: Error after pre-initialization, DT: Error after applying Distance Transform, LC 2 : error after final LC 2 registration.
AnglePatient No.Translation Error (mm)Rotation Error ( )Final Error(t,r)
In-ErrPre-inDTLC 2 In-ErrPre-inDTLC 2
0 Subject 134.729.836.042.0073.7020.1328.675.502.00 mm, 5.50
0 Subject 235.3910.5914.623.9462.1421.0114.478.783.94 mm, 8.78
0 Subject 439.1013.452.671.7260.288.944.933.911.72 mm, 3.91
0 Subject 532.097.433.881.4765.8813.2917.264.151.47 mm, 4.15
0 Subject 631.6811.0417.742.4367.8333.1513.418.672.43 mm, 8.67
30 Subject 326.1412.051.431.4145.3732.132.412.341.41 mm, 2.34
30 Subject 631.8930.069.194.8025.8835.2912.569.034.80 mm, 9.03
60 Subject 327.9917.758.705.0955.217.8115.806.435.09 mm, 6.43
60 Subject 627.717.497.456.0027.9418.709.705.806.00 mm, 5.80
90 Subject 333.1322.795.363.6325.8233.2926.4615.073.63 mm, 15.07
90 Subject 634.5012.153.586.0325.0912.5317.6216.046.03 mm, 16.04
Table 3. Inter-operator variability study.
Table 3. Inter-operator variability study.
AngleSubject Δ Translation Δ Rotation
0 Subject 33.00 mm4.12
0 Subject 53.89 mm6.51
0 Subject 64.63 mm9.24
30 Subject 63.64 mm4.05
60 Subject 37.94 mm9.06
60 Subject 61.67 mm11.99
90 Subject 34.77 mm9.84
90 Subject 65.80 mm7.28
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kompella, G.; Singarayan, J.; Antico, M.; Sasazawa, F.; Yu, T.; Ram, K.; Pandey, A.K.; Fontanarosa, D.; Sivaprakasam, M. Automatic 3D MRI-Ultrasound Registration for Image Guided Arthroscopy. Appl. Sci. 2022, 12, 5488. https://0-doi-org.brum.beds.ac.uk/10.3390/app12115488

AMA Style

Kompella G, Singarayan J, Antico M, Sasazawa F, Yu T, Ram K, Pandey AK, Fontanarosa D, Sivaprakasam M. Automatic 3D MRI-Ultrasound Registration for Image Guided Arthroscopy. Applied Sciences. 2022; 12(11):5488. https://0-doi-org.brum.beds.ac.uk/10.3390/app12115488

Chicago/Turabian Style

Kompella, Gayatri, Jeevakala Singarayan, Maria Antico, Fumio Sasazawa, Takeda Yu, Keerthi Ram, Ajay K. Pandey, Davide Fontanarosa, and Mohanasankar Sivaprakasam. 2022. "Automatic 3D MRI-Ultrasound Registration for Image Guided Arthroscopy" Applied Sciences 12, no. 11: 5488. https://0-doi-org.brum.beds.ac.uk/10.3390/app12115488

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop