Next Article in Journal
Quantification of the Monomer Compositions of Poly(3-hydroxybutyrate-co-3-hydroxyvalerate) and Poly(3-hydroxyvalerate) by Alkaline Hydrolysis and Using High-Performance Liquid Chromatography
Next Article in Special Issue
A New Breakpoint to Classify 3D Voxels in MRI: A Space Transform Strategy with 3t2FTS-v2 and Its Application for ResNet50-Based Categorization of Brain Tumors
Previous Article in Journal
Concept for the Treatment of Class III Anomalies with a Skeletally Anchored Appliance Fabricated in the CAD/CAM Process—The MIRA Appliance
Previous Article in Special Issue
CNN-Based Identification of Parkinson’s Disease from Continuous Speech in Noisy Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Augmented Reality Surgical Navigation System Integrated with Deep Learning

1
Department of Electrical Engineering, College of Engineering, Chang Gung University, Kwei-Shan, Taoyuan 333, Taiwan
2
Department of Nuclear Medicine, Linkou Chang Gung Memorial Hospital, Taoyuan 333, Taiwan
3
Department of Neurosurgery, Keelung Chang Gung Memorial Hospital, Keelung 204, Taiwan
4
Department of Electrical and Electronic Engineering, College of Engineering, Yonsei University, Seodaemun-gu, Seoul 03722, Republic of Korea
5
Department of Artificial Intelligence, College of Software & Convergence Technology, Daeyang AI Center, Sejong University, Seoul 05006, Republic of Korea
6
Department of Electrical Engineering, National Taiwan University, Taipei 106, Taiwan
7
New Taipei City Tucheng Hospital, Tao-Yuan, Tucheng, New Taipei City 236, Taiwan
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(5), 617; https://doi.org/10.3390/bioengineering10050617
Submission received: 13 March 2023 / Revised: 11 May 2023 / Accepted: 17 May 2023 / Published: 20 May 2023
(This article belongs to the Special Issue Artificial Intelligence in Biomedical Diagnosis and Prognosis)

Abstract

:
Most current surgical navigation methods rely on optical navigators with images displayed on an external screen. However, minimizing distractions during surgery is critical and the spatial information displayed in this arrangement is non-intuitive. Previous studies have proposed combining optical navigation systems with augmented reality (AR) to provide surgeons with intuitive imaging during surgery, through the use of planar and three-dimensional imagery. However, these studies have mainly focused on visual aids and have paid relatively little attention to real surgical guidance aids. Moreover, the use of augmented reality reduces system stability and accuracy, and optical navigation systems are costly. Therefore, this paper proposed an augmented reality surgical navigation system based on image positioning that achieves the desired system advantages with low cost, high stability, and high accuracy. This system also provides intuitive guidance for the surgical target point, entry point, and trajectory. Once the surgeon uses the navigation stick to indicate the position of the surgical entry point, the connection between the surgical target and the surgical entry point is immediately displayed on the AR device (tablet or HoloLens glasses), and a dynamic auxiliary line is shown to assist with incision angle and depth. Clinical trials were conducted for EVD (extra-ventricular drainage) surgery, and surgeons confirmed the system’s overall benefit. A “virtual object automatic scanning” method is proposed to achieve a high accuracy of 1 ± 0.1 mm for the AR-based system. Furthermore, a deep learning-based U-Net segmentation network is incorporated to enable automatic identification of the hydrocephalus location by the system. The system achieves improved recognition accuracy, sensitivity, and specificity of 99.93%, 93.85%, and 95.73%, respectively, representing a significant improvement from previous studies.

Graphical Abstract

1. Introduction

In recent years, several studies have proposed applying augmented reality (AR), virtual reality (VR), and artificial intelligence (AI) technologies with medicine, producing promising results but also exhibiting some limitations. For instance, AR or VR technology can be a promising tool for complex procedures, especially in maxillofacial surgery, to ensure predictable and safe outcomes [1]. AR-based surgical navigation techniques can be broadly categorized into three types: projection navigation [2], optical positioning navigation [3,4,5,6], and image positioning navigation [7,8,9,10]. Projection navigation involves direct projection onto the surgical area [9] or a semi-transparent mirror for superimposition [11].
In terms of projection navigation, Tabrizi et al. [2] proposed a direct projection navigation approach using a projector to project onto the Region of Interest (ROI) for brain tumor removal. However, this approach causes curvature distortion and reduces accuracy. Fida et al. [12] applied volume subtraction navigation to knee joint transplantation (KNS), projecting the image onto a semi-transparent mirror for superimposition, with two optical cameras to achieve better depth detection and put the real-time image into the tracking learning detection (TLD) method [13] to track specific objects. The system processes the frames using volume subtraction to remove redundant images and improve the accuracy of KNS surgery. Nonetheless, due to space limitations, volume subtraction navigation is not suitable for neurosurgical procedures such as extra-ventricular drainage (EVD). Regarding optical positioning navigation, a positioning device (NDI Polaris Vicra and a cursor ball), along with a screen (a head-mounted device (HMD)), were utilized by Chen et al. [3]. The system determines the current virtual image’s position based on the distance between the positioning devices and the cursor ball, and subsequently applies iterative closest point (ICP) and manual adjustment to superimpose the virtual image onto the physical image.
Concerning image positioning navigation, Müller et al. [5] suggested fiducial marker image navigation, which employs a lens to recognize fiducial points and generate virtual images for percutaneous nephrolithotomy (PCNL). Nevertheless, the reference point placement before surgery is time-consuming (99 s), and the average error of 2.5 mm is unsuitable for neurosurgical procedures (such as EVD).
According to Prakosa et al. [14], the use of AR guidance in a virtual heart can increase catheter navigation accuracy, leading to a potential effect on ventricular tachycardia (VT) termination. In their study, the use of AR resulted in significantly smaller errors in 54 VT 1 ablation simulations compared to those without AR. The accuracy of VT termination may also improve with substrate characteristics. Tu et al. [15] proposed a HoloLens-to-world registration method that utilizes an external EM tracker and a customized registration cube to improve depth perception and reduce registration errors. The preclinical phantom experiments showed reprojection errors along the X, Y, and Z axes to be 1.55 ± 0.27 mm, 1.71 ± 0.40 mm, and 2.84 ± 0.78 mm, respectively. The end-to-end evaluation method indicated the distance error was 1.61 ± 0.44 mm, and the 3D angle error was 1.46 ± 0.46°. However, the accuracy performances of these methods are still not satisfactory, making them unsuitable for current surgical operations.
Over the past decade, there has been significant attention given to advances in deep learning methods for various medical imaging applications such as diagnosis, segmentation, and detection [16,17,18,19]. Convolutional neural networks (CNNs) have been widely used in deep learning for image-based prediction [20,21,22]. Developers have also incorporated decoder paths into the CNN to segment objects within input images, resulting in the generation of segmentation maps with pixel-wise predictions that match the input size. This approach, known as encoder-decoder networks or U-Net [23], is commonly employed in most medical imaging applications that use supervised learning, which requires labeled references during the training phase.
Long et al. [24] proposed a fully convolutional network that takes the input of the arbitrary size and produces correspondingly-sized output with efficient inference and learning. They adapted several classification networks (AlexNet [20], the VGG net [21], and GoogLeNet [22]) into fully convolutional networks and transferred their learned representations by fine-tuning [25] to the segmentation task.
Yigin et al. [26] explored the effectiveness of commonly used morphological parameters in hydrocephalus diagnosis. For this purpose, they compared the effect of six common morphometric parameters (Frontal Horns’ Length (FHL), Maximum Lateral Length (MLL), Biparietal Diameter (BPD), Evans’ Ratio (ER), Cella Media Ratio (CMR), and Frontal Horns’ Ratio (FHR)) in terms of their importance in predicting hydrocephalus using a Random Forest classifier.
Martin et al. [27] proposed a method using Compositional Pattern Producing Network (CPPN) to enable Fully Convolutional Networks (FCN) to learn cerebral ventricular system (CVS) location. To address the ventriculomegaly problem that arises in the clinical routine, dilation of the CVS is required.
However, current AI methods cannot automatically locate hydrocephalus. To address this issue, this paper proposes a comprehensive solution to obtain the surgical target, scalpel entry point, and scalpel direction, and automatically locate hydrocephalus. The proposed approach includes virtual object automatic scanning operation navigation to improve accuracy and the use of a tablet computer lens to align two custom image targets for locating the virtual head and virtual scalpel. The improved U-Net [28] is also utilized to recommend target points, resulting in a surgery efficiency and accuracy rate of 99.93%. Clinical trials were conducted for EVD (extra-ventricular drainage) surgery, and surgeons have confirmed the feasibility of the system. Overall, the proposed approach has the potential to enhance surgical outcomes and improve patient care in the field of neurosurgery.

2. Materials and Methods

2.1. System Overview

The proposed system comprises an AR device (Surface Pro 7 (Microsoft Co., Redmond, WA, USA)), printed images for head and scalpel positioning, an upright tripod, a flat clamp, and a medical septum cover. The tablet tracks the feature points of the two positioning images to display the virtual image correctly (see Figure 1a). The system follows the flow chart depicted in Figure 1b, providing multiple functions such as automatic scanning of DICOM-formatted virtual objects, selection of surgical targets, generation of surgical trajectories, and color-assisted calibration of surgical entry points to aid surgeons during EVD surgery. Table 1 defines the symbols and parameters used in the proposed method, while Table 2 presents the main abbreviations employed in this article.

2.2. Virtual Object Automatic Scanning

The patient’s head CT scans are converted into a virtual head object with Avizo software (Waltham, MA, USA) [29], which provides an image analysis platform for identifying the locations of the scalp, skull, and hydrocephalus. The principle of the automatic scanning method is based on a trigger function. Virtual planes that are perpendicular to the x, y, and z axes in Unity are utilized to scan virtual objects. The entry and exit points (or two collision points) are obtained when these virtual planes enter and leave the virtual object or when they enter the object from two opposite directions, resulting in a total of six reference points.
As an example, RPxR (reference point on the right of the X-axis) is obtained when the axial plane enters the virtual head object from the right, while RPxL (reference point on the left of the X-axis) is obtained when the axial plane enters the virtual scalp from the left. RPxR and RPx serve as two reference points on the X-axis for displaying real-time DICOM (Digital Imaging and Communications in Medicine) images.
The system performs simultaneous virtual scalp scans along the three-axis (sagittal, frontal, horizontal, and lateral) to obtain a total of six reference points, which are displayed as virtual head objects as RPxR, RPxL, RPyT, RPyB, RPzT, and RPzB. These reference points play a critical role in locating the DICOM plane, which significantly impacts target accuracy (refer to Section 2.5 for detailed methodology). After completing the calibration, an accurate augmented reality environment is generated to assist with EVD surgical navigation.

2.3. Using Machine Learning to Predict and Recommend Surgical Target

To predict the location of the hydrocephalus and identify the surgical target, the system selected the connection area with the largest hydrocephalus, computed the center point of this area, and marked it as the final recommended surgical target point. The system’s connection areas were selected based on the segmented output of U-Net, with the largest connected pixels chosen using the bwconncomp function in Matlab. This function is designed to identify and count connected components in binary segmented images.
DICOM files were utilized to conduct cross-validation on ten groups of patients, with the database divided into five folds. Each group was assigned a number from 1 to 10, and two groups of patient DICOM data were used as testing sets, while the remaining groups served as training sets. Consequently, there were five folds with five distinct testing sets, with four folds (consisting of eight groups of patients’ DICOM data) as the training set and the other fold (the DICOM data of the other two groups) as the test set. The training set included label data and normalized data from the eight patient groups, while the test set used the label and normalized data from the other two groups. The normalized data was used as the basis for the final accuracy test.

2.4. Manual Operation Target Point Positioning

After superimposing the virtual head onto the real head, DICOM images in the horizontal and lateral planes are displayed on the tablet. This allows the surgeon to individually select and confirm the accuracy of the target position. Once the target position is confirmed, the relative position is then converted into conversion space, and the specific DICOM slice containing the ideal target ( N t a r g e t -th slice) can be obtained from the total number of DICOM slices N t o t a l . The DICOM image is displayed in the lower left of the screen, with the ideal target position displayed as P o s t a r g e t 2 D X t a r g e t 2 D , Y t a r g e t 2 D . The length and width of the DICOM image are I d i c o m X 2 D and I d i c o m Y 2 D , respectively. The origin point P o s o 3 D 0,0 , 0 is located in the lower left of the head, with the length of the head along the X, Y, and Z axes being I X 3 D , I Y 3 D , I Z 3 D , respectively. The target position in space is denoted by
P o s t a r g e t 3 D X t a r g e t 2 D × I X 3 D I d i c o m X 2 D , I Y 3 D N t o t a l × N t a r g e t 1 , Y t a r g e t 2 D × I Z 3 D I d i c o m Y 2 D

2.5. DICOM Image Real-Time Display and Selection of Target Point

After automatic scanning and obtaining the six reference points on the three axes, the longest distance of the reference point on each axis is calculated as Disx (the distance on the x-axis), Disy (the distance on the y-axis), and Disz (the distance on the z-axis). The resulting values are then divided by the number of DICOM slices on the corresponding axis, including Numx, Numy, and Numz. The resulting values are the thicknesses on the x, y, and z axes, denoted as Tx, Ty, and Tz, respectively.
T x = D i s x ÷ N u m x
T y = D i s y ÷ N u m y
T z = D i s z ÷ N u m z
The distance between the scalpel’s edge point ( E p ) and R P x L is divided by T x to determine the corresponding DICOM slice on the x-axis, known as T r u e X . This algorithm is repeated for the y and z axes.
T r u e X = E p R P x L ÷ T x , E p R P x L
T r u e Y = E p R P y L ÷ T y , E p R P y L
T r u e Z = ( E p R P z L ) ÷ T z , E p R P z L
Once T r u e X and T r u e Z have been calculated, the Unity Web Request function is utilized to present a real-time DICOM image display (Figure 2a) in combination with augmented reality. This allows surgeons to access the display without having to look away to a separate screen. Surgeons can then simply tap on the screen to select the ideal target (Figure 2b).

2.6. Generation of Surgical Trajectory and Calibration of Entry Point Angle

Once the target point is selected, a surgical simulation trajectory is generated, connecting the scalpel’s edge to the target point. The surgeon then confirms this trajectory by pressing the function button, which generates the final trajectory connecting the surgical entry point to the target point (Figure 3a). To ensure maximum accuracy of the location and path of the surgical entry point, color-assisted angle calibration is used. The color of the trajectory changes from red (Figure 3b) to yellow (Figure 3c) to green (Figure 3d), providing high color accuracy during surgery.

3. Experiments Studies and Tests

To demonstrate the feasibility of the method proposed in Section 2, prosthesis experiments were first conducted in the laboratory using the proposed method. Subsequently, clinical feasibility tests are carried out in hospitals. A Surface 7 tablet was used as the AR device in both test reports. Furthermore, Hololens 2 smart glasses are currently the most popular advanced medical AR HMD devices. A detailed explanation of the potential outcomes when substituting the AR devices with the Hololens 2 smart glasses is provided.

3.1. Experiment of the Virtual Object Automatic Scanning

To evaluate the accuracy of the proposed virtual object automatic scanning method, DICOM data from ten patients were utilized. The automatic scanning error was determined by measuring the distance between the predicted axis plane (automatically selected by the deep learning system) and the actual axis plane (where the target point was located). The minimum error, maximum error, lower quartile, median, and upper quartile, as well as the target point error established by the software Avizo, were also calculated. The virtual model of the point was imported into 3ds Max (Autodesk Inc., San Francisco, CA, USA) for anchor point processing, ensuring constant relative positions of the scalpel and the target point, which facilitated error calculation.
Additionally, it should be noted that the predicted axis plane was obtained by selecting the center point of the largest connection area as the target point after predicting the contour of the ventricle using the deep learning model. The actual axis plane was obtained by extracting the center point of the 3D object of the target point, which is created in 3ds Max from the DICOM data. Before this, the 3D object of the ventricle was generated in 3ds Max, followed by the 3D object of the target point that corresponds to the doctor-marked target point on the DICOM.

3.2. Experiment of Machine Learning to Predict and Recommend Surgical Target

To predict the location of hydrocephalus, U-Net (the original model) was employed for deep learning to maximize accuracy and minimize loss by setting 30 epochs, and identifying all ventricular regions in the patient’s DICOM images (Figure 4b). The Labeled ventricle contour (green) and the predicted ventricle contour (red) were drawn using Matlab (MathWorks Inc., Natick, MA, USA) (Figure 4a). Finally, the average sensitivity, specificity, and accuracy for predicting the location of hydrocephalus were calculated, and these data (sensitivity, specificity, and accuracy) are the result of comparing the contour with the pixel-by-pixel method.
The U-Net (the original model) architecture consists of two paths, the encoder, and the decoder. The encoder path, comprising convolutional and pooling layers, is responsible for extracting and learning contextual features. Conversely, the decoder path, which includes transpose convolution and up-sampling layers, aims to transfer these learned features into a single prediction layer of the same size as the input image, known as the dense prediction or segmentation map. Each convolution layer is appended with batch normalization and ReLU activation function, except for the last layer, which produces a binary output using sigmoid activation. The entire network uses a convolution kernel of size 3 × 3 and stride of 1 with feature maps of 32, 64, 128, 256, and 320 across all resolution levels.
Hyper-parameters used in this study were a learning rate of 0.003, a batch size of 20, and a total of 30 epochs. The objective was to minimize the overall loss of each pixel by computing the dice loss function between the segmented map and labeled reference, and the Adam optimizer was utilized to optimize the weight parameters in each layer.

3.3. Test of Clinical Feasibility

The feasibility of the proposed system was tested at various stages of clinical implementation, beginning with the conversion of DICOM images from patients into 3D models. To assess the clinical feasibility of each step, a pre-operative simulation was conducted in the operating room approximately 1 h before the surgery. Figure 5 illustrates the setup of the system in the operating room and its operation by a neurosurgeon. Specifically, Figure 5a shows the superimposed 3D model of the patient’s head position, while Figure 5b shows the DICOM-selected target position on the display. Figure 5c depicts the alignment position and angle following entry point selection, and Figure 5d shows the completed alignment. Following the entire process, an experienced surgeon concluded that the system concept is feasible for clinical use.

3.4. Test of Hololens 2 Feasibility

In order to test whether our proposed method is accurate on HoloLens 2, we designed a separate accuracy experiment specifically for HoloLens 2. A solid sponge brick was used for flower arrangement, and the target point was set as the middle point at the bottom of the sponge brick. A virtual sponge brick model of the same size was created in Unity, and the target point, insertion point, and guide path were set. The experimenters wore Hololens 2 and inserted the real navigation stick into the sponge brick through the virtual guide path seen in the Hololens 2 screen to test whether it could accurately reach the target point.
To perform the experiment, the experimenter needed to superimpose the real sponge brick and the virtual model (Figure 6), align the navigation stick with the path of the virtual guide (Figure 6b), insert the sponge brick along the guiding direction, and pass the navigation stick through the sponge bricks. The difference between the “true target position” and “the final position where the experimenter arrived at the real sponge brick using the navigation stick” was measured to calculate the error distance.

4. Results

4.1. Results of the Virtual Object Automatic Scanning

The virtual object automatic scanning error (Figure 7a) was calculated by determining the distance between the axis plane that is automatically selected by the system and the axis plane where the actual target point is located. The average automatic scanning error was 1.008 mm with a deviation of 0.001 mm. The minimum and maximum errors were 0.978 mm and 1.039 mm, respectively. Due to the small deviation, the lower quartile, median, and upper quartile are represented by a straight line in the box plot. The box plot indicates two outliers with values of 0.978 mm and 1.039 mm, respectively. The target point error (Figure 7b) was determined using Avizo software.
To facilitate alignment in Unity, the anchor points of the two virtual objects were adjusted to the same position. Subsequently, the distance between the real target point and the virtual target point was used to obtain the target point error, which was found to be 1 mm with a deviation of 0.1 mm. The minimum and maximum errors were 0.495 mm and 1.21 mm, respectively, while the lower quartile, median, and upper quartile were 0.88 mm, 0.98 mm, and 1.12 mm, respectively. Stability tests (Figure 7c) were conducted in a total of 20 phantom trials. A script was written to record the center position of the scalp generated by the image target of the head every 1 s for 1 min. The stability was measured in 3 dimensions and normalization was performed afterward. The average stability and deviation were 0.076 mm and 0.052 mm, respectively.

4.2. Results of the Machine Learning to Predict and Recommend Surgical Target

The proposed system utilizes machine learning (specifically, the U-Net model) to predict and recommend surgical targets. This system was tested on 10 hydrocephalic patients, and the results indicated an average sensitivity of 93.85%, specificity of 95.73%, and accuracy of 99.93% in predicting the location of hydrocephalus.
The U-Net model generates a binary mask output, with ones indicating the ventricular region and zeros indicating other parts of the image. By comparing the output prediction to the labeled reference, true positive (TP), true negative (TN), false positive (FP), and false negative (FN) values can be computed. This allows for the calculation of sensitivity (TP/(TP + FN)), specificity (TN/(TN + FP)), and accuracy ((TP + TN)/(TP + FP + TN + FN)). Notably, the labeled reference of the ventricular region is available for experimental data, enabling the calculation of these indices in a similar manner.
The system enhances location prediction for hydrocephalus in terms of accuracy, sensitivity, and specificity. As shown in Table 3, the hydrocephalus prediction function presented in this paper can more accurately predict the location of hydrocephalus and provide surgeons with better surgical target recommendations, regardless of accuracy, sensitivity, and specificity.

4.3. Results of the Proposed System

The proposed approach exhibits fewer image navigation limitations and lower costs than optical navigation. A virtual object automatic scanning method is proposed to reduce calibration time in the preoperative stage, taking only 4 s. This represents an 87%, 96%, and 98% reduction in time compared to Konishi et al.’s ultrasonic scanning [33], Müller et al.’s reference registration method [5], and Tabrizi et al.’s projection registration method [2], respectively. Additionally, the proposed method achieves higher accuracy, with a range of 1 ± 0.1 mm, surpassing Ieiri et al.’s optical positioning navigation (18.8 ± 8.56 mm) [34], Deng et al.’s image positioning navigation (2.5 mm) [7], and Frantz et al.’s AR goggles surgical navigation system (1.41 mm) [9]. The proposed system leverages virtual object automatic scanning and image positioning for EVD surgical navigation, providing improved equipment cost, registration time, and accuracy.
Table 4 shows that the proposed image positioning method offers cost savings in comparison to the other three positioning methods, while also improving registration time and target accuracy. The registration time of 4 s is achieved through the virtual object automatic scanning method, while the accuracy of 1 ± 0.1 mm is obtained from the “3.1. Prosthesis experiment.” The proposed system provides superior target accuracy performance, primarily due to the virtual object automatic scanning method that offers accurate reference points, and all functions are performed within the system. This indicates that compared to other research methods, there are no external factors that may impact the accuracy of the target.

4.4. Results of the Hololens 2 Feasibility

Table 5 presents the accuracy results of five experiments conducted by five male experimenters aged 22 to 25, indicating that the impact of visual assistance with HoloLens 2 can differ significantly among users. Consequently, software feedback is essential for the navigation stick assistance method. However, as illustrated in Figure 8, the augmented reality performance of HoloLens 2 in tracking spatial images lacks stability, leading us to abandon the use of HoloLens 2 in the clinical feasibility test. For the clinical feasibility testing, a Microsoft Surface Pro 7 tablet was ultimately used.

5. Discussion

5.1. Comparing Augmented Reality Medical Guidance Techniques

5.1.1. Image Positioning vs. Optical Positioning

Compared to the current optical navigation method, the proposed method in this paper offers significant advantages in terms of intuition, mobility, accessibility, and cost-effectiveness. Most current image-guided surgical navigation methods combine optical navigation with a navigation stick tracked by a cursor ball and display navigation information on an external screen.
Regarding intuition, our proposed method provides surgeons with intuitive spatial information through AR perspective display. In terms of mobility, the current optical navigation method requires a specific operating room, whereas our system can be used for surgical navigation in different fields with only a 10-min setup time. Furthermore, the proposed system is more accessible and cost-effective than the optical navigation method due to its lower equipment and practice costs. Augmented reality is an excellent solution for guiding surgery in areas with insufficient medical resources.

5.1.2. Our Method vs. Other Method

Currently, several advanced augmented reality methods show promise for surgical navigation [36,37,38,39], but they still have limitations. Gavaghan et al. [36] used a projector to project liver blood vessels onto the liver surface, which has good accuracy but lacks deep spatial information and may not be suitable for guiding surgery. Kenngott et al. [37] proposed a method that provides three-dimensional anatomical information and has undergone clinical feasibility testing, but this method only offers viewing functions and lacks other auxiliary guidance. Heinrich et al. [38] and Hecht et al. [39] both provided visual aids for guided injections but lack feedback.
In comparison to the current augmented reality methods for medical guidance [36,37,38,39], the proposed method in this paper exhibits significant advantages in accuracy, provision of anatomical information, stereoscopic display, path navigation, visual feedback, and clinical suitability, as outlined in Table 6. As a result, the proposed method outperforms the current methods in all aspects.

5.1.3. Tablets vs. Smart Glasses

The proposed system was implemented on a Microsoft Surface Pro 7 tablet and Microsoft HoloLens 2 smart glasses to compare their performance in terms of stability, flexibility, and information richness. The performance metrics are presented in Table 7.
In terms of stability, the Surface Pro 7 displays the head model and maintains the navigation stick’s stability well in a fixed position. On the other hand, the HoloLens 2 shows good stability for the head model in a fixed position, but its field of view changes with user movement, resulting in increased offset error. Additionally, the HoloLens 2 exhibits noticeable visual instabilities when tracking the navigation stick in motion.
Concerning flexibility, the Surface Pro 7 requires an additional stand that limits the viewing angle, while the HoloLens 2 has superior flexibility. Regarding comfort, the Surface Pro 7 is more comfortable as physicians do not need to wear the HoloLens 2, which can increase head weight, eye pressure, and visual interference.
Regarding information display richness, the HoloLens 2 can set windows in space to display DICOM information in a larger, clearer, and more persistent way. In contrast, the Surface Pro 7 can only display information on a single screen. Moreover, to avoid blocking the surgical guidance image, the DICOM image display must be canceled after selecting the target on the Surface Pro 7, preventing simultaneous display. Although multiple DICOM images can be superimposed and displayed on the head at the same time, the visual effect is not comfortable. Directly locking and displaying the target in an AR manner is a relatively simple visual effect after judging the target position.
Therefore, despite the HoloLens 2’s flexibility and complete information display advantages, guidance accuracy is the most critical factor, making the Surface Pro 7 the ideal platform for implementation.

5.2. The Impact of Viewing Angle Changes on the Coordinates

To discuss the intrinsic error of Vuforia, the influence of the head model and the tip position of the navigation stick on the coordinate position at different visual angles was tested. Figure 9 presents the results of testing commonly used navigation stick angles (60~140°) and head model recognition, graphing their influence on coordinates under reasonable usage angles (60~120°). The X coordinate of the navigation bar is found to be significantly affected by the viewing angle, but within the most frequently used range of angles (80~100°), the error is only ±1 mm, and there is no significant effect on the Y and Z coordinates, with most errors outside the outliers within ±1 mm. Additionally, to examine the impact of changes in the viewing angle of the head model identification map, the system was tested in the range of 60–120 degrees in 10-degree increments, as both the head model identification map and the camera’s viewing angle are fixed values.

5.3. Sterile Environment for Surgery

To ensure suitability for clinical applications, the proposed system must be able to function in a sterile environment. As such, a layer of the surgical cell membrane is covered on the identification map, and the tablet is wrapped in a plastic sleeve, allowing it to remain operational without compromising sterility.

5.4. Clinical Significance and Limitation

In summary, previously proposed methods have not provided a comprehensive solution for accurately guiding surgical targets, scalpel entry points, and scalpel orientation in brain surgery. The proposed approach aims to address these shortcomings. To ensure ease of portability, a tablet PC is used as the primary AR device in the proposed system. The DICOM data processing takes approximately one hour to complete the system update. Surgeons can use the proposed system before and during surgery for real-time guidance on surgical targets, entry points, and scalpel paths. In terms of precision, the proposed system has an average spatial error of 1 ± 0.1 mm, which is a significant improvement over many previous methods. The system achieves improved recognition accuracy, sensitivity, and specificity, with values of 99.93%, 93.85%, and 95.73%, respectively, marking a significant improvement over previous studies. Smart glasses are not recommended for the proposed AR system due to their potential to introduce significant errors, as accuracy and stability are important considerations.

6. Conclusions

This study combined virtual object automatic scanning with deep learning and augmented reality to improve surgeons’ surgical procedures. U-Net was utilized for deep learning to predict the location of hydrocephalus, reducing pre-operative time requirements and increasing surgical precision. Augmented reality overlays virtual images directly on the patient’s head, allowing for intuitive guidance in locating surgical target points and trajectory guidance for improved accuracy in EVD surgery. The proposed system also employed color coding for angle correction at the surgical entry point, allowing for more intuitive and accurate operations. The developed EVD surgical navigation system using virtual object automatic scanning and augmented reality shows improved accuracy, registration time, and surgical costs. Future work will focus on exploring the use of smart glasses for collaborative operations and conducting clinical trials for intraoperative navigation to enhance the clinical utility of the proposed system.

Author Contributions

Conceptualization, S.-Y.C. and H.-L.L.; methodology, S.-Y.C., L.-S.L., C.-W.L. and M.A.A.-m.,; software, L.-S.L. and C.-W.L.; validation, S.-Y.C., M.A.A.-m., J.-L.Y. and P.-Y.C.; formal analysis, S.-Y.C. and M.A.A.-m.; investigation, S.-Y.C., L.-S.L. and C.-W.L.; resources, J.-L.Y. and P.-Y.C.; data curation, S.-Y.C., L.-S.L. and C.-W.L.; writing—original draft preparation, L.-S.L. and C.-W.L.; writing—review and editing, S.-Y.C. and M.A.A.-m.; visualization, S.-Y.C., L.-S.L. and C.-W.L.; supervision, S.-Y.C., D.-H.K. and M.A.A.-m.; project administration, S.-Y.C.; funding acquisition, S.-Y.C. and K.-C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the Ministry of Science and Technology under Grant MOST 111-2221-E-182-020 and by the CGMH project under Grant CMRPD2M0021 and BMRPB46.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Chang Gung Medical Foundation (with IRB No.: 201900856B0A3, and Date of Approval: 2019/07/16).

Informed Consent Statement

Not applicable for studies not involving humans.

Data Availability Statement

The statistical data presented in this study are available in Table 3 and Table 4. The datasets used and/or analyzed during the current study are available from the corresponding author upon request. These data are not publicly available due to privacy and ethical reasons.

Acknowledgments

This work was partially supported by the Ministry of Science and Technology under Grant MOST 111-2221-E-182-020 and in part by the CGMH project under Grant CMRPD2M0021 and BMRPB46.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Joda, T.; Gallucci, G.; Wismeijer, D.; Zitzmann, N. Augmented and virtual reality in dental medicine: A systematic review. Comput. Biol. Med. 2019, 108, 93–100. [Google Scholar] [CrossRef] [PubMed]
  2. Tabrizi, L.; Mahvash, M. Augmented reality–guided neurosurgery: Accuracy and intraoperative application of an image projection technique. J. Neurosurg. 2015, 123, 206–211. [Google Scholar] [CrossRef]
  3. Chen, X.; Xu, L.; Wang, Y.; Wang, H.; Wang, F.; Zeng, X.; Wang, Q.; Egger, J. Development of a surgical navigation system based on augmented reality using an optical see-through head-mounted display. J. Biomed. Inform. 2015, 55, 124–131. [Google Scholar] [CrossRef]
  4. Lai, M.; Skyrman, S.; Shan, C.; Babic, D.; Homan, R.; Edström, E.; Persson, O.; Urström, G.; Elmi-Terander, A.; Hendriks, B.; et al. Fusion of augmented reality imaging with the endoscopic view for endonasal skull base surgery; A novel application for surgical navigation based on intraoperative cone beam computed tomography and optical tracking. PLoS ONE 2020, 15, e0227312. [Google Scholar]
  5. Müller, M.; Rassweiler, M.; Klein, J.; Seitel, A.; Gondan, M.; Baumhauer, M.; Teber, D.; Rassweiler, J.; Meinzer, H.; Maier-Hein, L. Mobile augmented reality for computer-assisted percutaneous nephrolithotomy. Int. J. Comput. Assist. Radiol. Surg. 2013, 8, 663–675. [Google Scholar] [CrossRef]
  6. Wang, J.; Suenaga, H.; Hoshi, K.; Yang, L.; Kobayashi, E.; Sakuma, I.; Liao, H. Augmented reality navigation with automatic marker-free image registration using 3-D image overlay for dental surgery. IEEE Trans. Biomed. Eng. 2014, 61, 1295–1304. [Google Scholar] [CrossRef]
  7. Conrad, C.; Fusaglia, M.; Peterhans, M.; Lu, H.; Weber, S.; Gayet, B. Augmented reality navigation surgery facilitates laparoscopic rescue of failed portal vein embolization. J. Am. Coll. Surg. 2016, 223, 31–34. [Google Scholar] [CrossRef]
  8. Deng, W.; Deng, F.; Wang, M.; Song, Z. Easy-to-use augmented reality neuronavigation using a wireless tablet PC. Stereotact. Funct. Neurosurg. 2014, 92, 17–24. [Google Scholar] [CrossRef] [PubMed]
  9. Frantz, T.; Jansen, B.; Duerinck, J.; Vandemeulebroucke, J. Augmenting Microsoft’s HoloLens with vuforia tracking for neuronavigation. Healthc. Technol. Lett. 2018, 5, 221–225. [Google Scholar] [CrossRef] [PubMed]
  10. Vandoormaal, T.; Vandoormaal, J.; Mensink, T. Clinical accuracy of holographic navigation using point-based registration on augmented-reality glasses. Oper. Neurosurg. 2019, 17, 588–593. [Google Scholar] [CrossRef]
  11. Montani, S.; Striani, M. Artificial intelligence in clinical decision support: A focused literature survey. Yearb. Med. Inform. 2019, 28, 120. [Google Scholar] [CrossRef] [PubMed]
  12. Fida, B.; Cutolo, F.; Franco, G.; Ferrari, M.; Ferrari, V. Augmented reality in open surgery. Updat. Surg. 2018, 70, 389–400. [Google Scholar] [CrossRef]
  13. Kalal, Z.; Mikolajczyk, K.; Matas, J. Tracking-learning-detection. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 1409–1422. [Google Scholar] [CrossRef]
  14. Prakosa, A.; Southworth, M.K.; Silva, J.A.; Trayanova, N.A. Impact of augmented-reality improvement in ablation catheter navigation as assessed by virtual-heart simulations of ventricular tachycardia ablation. Comput. Biol. Med. 2021, 133, 104366. [Google Scholar] [CrossRef]
  15. Tu, P.; Gao, Y.; Lungu, A.; Li, D.; Wang, H.; Chen, X. Augmented reality based navigation for distal interlocking of intramedullary nails utilizing Microsoft HoloLens 2. Comput. Biol. Med. 2021, 133, 104402. [Google Scholar] [CrossRef] [PubMed]
  16. Shoeibi, A.; Khodatars, M.; Jafari, M.; Ghassemi, N.; Moridian, P.; Alizadehsani, R.; Ling, S.H.; Khosravi, A.; Alinejad-Rokny, H.; Lam, H.K.; et al. Diagnosis of brain diseases in fusion of neuroimaging modalities using deep learning: A review. Inf. Fusion 2023, 93, 85–117. [Google Scholar] [CrossRef]
  17. Shoeibi, A.; Ghassemi, N.; Khodatars, M.; Moridian, P.; Khosravi, A.; Zare, A.; Gorriz, J.M.; Chale-Chale, A.H.; Khadem, A.; Acharya, U.R. Automatic diagnosis of schizophrenia and attention deficit hyperactivity disorder in rs-fMRI modality using convolutional autoencoder model and interval type-2 fuzzy regression. Cogn. Neurodyn. 2022, 1–23. [Google Scholar] [CrossRef]
  18. Al-Masni, M.A.; Kim, W.-R.; Kim, E.Y.; Noh, Y.; Kim, D.-H. Automated detection of cerebral microbleeds in MR images: A two-stage deep learning approach. NeuroImage Clin. 2020, 28, 102464. [Google Scholar] [CrossRef]
  19. Al-Masni, M.A.; Kim, D.-H. CMM-Net: Contextual multi-scale multi-level network for efficient biomedical image segmentation. Sci. Rep. 2021, 11, 10191. [Google Scholar] [CrossRef]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1–9. [Google Scholar] [CrossRef]
  21. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  22. Liu, C.S.W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  23. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the 8th International Conference on Medical Image Computing and Computer-Assisted Intervention–(MICCAI), Munich, Germany, 5–9 October 2015; Part III 18. pp. 234–241. [Google Scholar]
  24. Long, L.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2015, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  25. Donahue, J.; Jia, Y.; Vinyals, O.; Hoffman, J.; Zhang, N.; Tzeng, E.; Darrell, T. Decaf: A deep convolutional activation feature for generic visual recognition. In Proceedings of the 31st International Conference on Machine Learning, Beijing, China, 21–26 June 2014; pp. 647–655. [Google Scholar]
  26. Yigin, B.O.; Algin, O.; Saygili, G. Comparison of morphometric parameters in prediction of hydrocephalus using random forests. Comput. Biol. Med. 2020, 116, 103547. [Google Scholar] [CrossRef] [PubMed]
  27. Martin, M.; Sciolla, B.; Sdika, M.; Quetin, P.; Delachartre, P. Automatic segmentation and location learning of neonatal cerebral ventricles in 3D ultrasound data combining CNN and CPPN. Comput. Biol. Med. 2021, 131, 104268. [Google Scholar] [CrossRef] [PubMed]
  28. Li, H.; Li, A.; Wang, M. A novel end-to-end brain tumor segmentation method using improved fully convolutional networks. Comput. Biol. Med. 2019, 108, 150–160. [Google Scholar] [CrossRef] [PubMed]
  29. Westenberger, P. Avizo—Three-Dimensional Visualization Framework. In Geoinformatics 2008—Data to Knowledge; USGS: Reston, VA, USA, 2008; pp. 13–14. [Google Scholar]
  30. Prevedello, L.; Erdal, B.; Ryu, J.; Little, K.; Demirer, M.; Qian, S.; White, R. Automated critical test findings identification and online notification system using artificial intelligence in imaging. Radiology 2017, 285, 923–931. [Google Scholar] [CrossRef]
  31. Rau, A.; Kim, S.; Yang, S.; Reisert, M.; Kellner, E.; Duman, I.; Stieltjes, B.; Hohenhaus, M.; Beck, J.; Urbach, H.; et al. SVM-Based Normal Pressure Hydrocephalus Detection. Clin. Neuroradiol. 2021, 31, 1029–1035. [Google Scholar] [CrossRef] [PubMed]
  32. Sahli, H.; Mouelhi, A.; Sayadi, M.; Rachdi, R. Discriminant textural feature selection and classification for a computerized fetal hydrocephalus detection. In Proceedings of the 2018 IEEE International Conference on Image Processing, Applications and Systems (IPAS), Sophia Antipolis, France, 12–14 December 2018; pp. 232–237. [Google Scholar]
  33. Konishi, K.; Nakamoto, M.; Kakeji, Y.; Tanoue, K.; Kawanaka, H.; Yamaguchi, S.; Ieiri, S.; Sato, Y.; Maehara, Y.; Tamura, S.; et al. A real-time navigation system for laparoscopic surgery based on three-dimensional ultrasound using magneto-optic hybrid tracking configuration. Int. J. Comput. Assist. Radiol. Surg. 2007, 2, 1–10. [Google Scholar] [CrossRef]
  34. Ieiri, S.; Uemura, M.; Konishi, K.; Souzaki, R.; Nagao, Y.; Tsutsumi, N.; Akahoshi, T.; Ohuchida, K.; Ohdaira, T.; Tomikawa, M.; et al. Augmented reality navigation system for laparoscopic splenectomy in children based on preoperative CT image using optical tracking device. Pediatr. Surg. Int. 2012, 28, 341–346. [Google Scholar] [CrossRef]
  35. Duan, W.; Zhang, J.; Zhang, L.; Lin, Z.; Chen, Y.; Hao, X.; Wang, Y.; Zhang, H. Evaluation of an artificial intelligent hydrocephalus diagnosis model based on transfer learning. Medicine 2020, 99, e21229. [Google Scholar] [CrossRef]
  36. Gavaghan, K.A.; Peterhans, M.; Oliveira-Santos, T.; Weber, S. A portable image overlay projection device for computer-aided open liver surgery. IEEE Trans. Biomed. Eng. 2011, 58, 1855–1864. [Google Scholar] [CrossRef]
  37. Kenngott, H.; Preukschas, A.; Wagner, M.; Nickel, F.; Müller, M.; Bellemann, N.; Stock, C.; Fangerau, M.; Radeleff, B.; Kauczor, H.; et al. Mobile, real-time, and point-of-care augmented reality is robust, accurate, and feasible: A prospective pilot study. Surg. Endosc. 2018, 32, 2958–2967. [Google Scholar] [CrossRef] [PubMed]
  38. Heinrich, F.; Schwenderling, L.; Becker, M.; Skalej, M.; Hansen, C. HoloInjection: Augmented reality support for CT-guided spinal needle injections. Healthc. Technol. Lett. 2019, 6, 165–171. [Google Scholar] [CrossRef] [PubMed]
  39. Hecht, R.; Li, M.; de Ruiter, Q.M.; Pritchard, W.; Li, X.; Krishnasamy, V.; Saad, W.; Karanian, J.W.; Wood, B. Smartphone augmented reality CT-based platform for needle insertion guidance: A phantom study. Cardiovasc. Interv. Radiol. 2020, 43, 756–764. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Surgical navigation overview and process flow chart. (a) Hardware overview of surgical navigation, and (b) flow chart of surgical navigation.
Figure 1. Surgical navigation overview and process flow chart. (a) Hardware overview of surgical navigation, and (b) flow chart of surgical navigation.
Bioengineering 10 00617 g001
Figure 2. Screen display of target selection. (a) Before target selection, and (b) after target selection.
Figure 2. Screen display of target selection. (a) Before target selection, and (b) after target selection.
Bioengineering 10 00617 g002
Figure 3. Calibration of surgical trajectory and entry point. (a) Surgery trajectory, (b) deviation from the entry point, (c) approach to the trajectory, and (d) calibrated trajectory.
Figure 3. Calibration of surgical trajectory and entry point. (a) Surgery trajectory, (b) deviation from the entry point, (c) approach to the trajectory, and (d) calibrated trajectory.
Bioengineering 10 00617 g003
Figure 4. Comparison of contour and indication of the predicted target. (a) Contour comparison, and (b) predicted target indication.
Figure 4. Comparison of contour and indication of the predicted target. (a) Contour comparison, and (b) predicted target indication.
Bioengineering 10 00617 g004
Figure 5. Clinical feasibility test. (a) Superimposition of the 3D model onto the patient’s head, (b) selection of target point, (c) calibration of trajectory angle, and (d) completion of calibration.
Figure 5. Clinical feasibility test. (a) Superimposition of the 3D model onto the patient’s head, (b) selection of target point, (c) calibration of trajectory angle, and (d) completion of calibration.
Bioengineering 10 00617 g005aBioengineering 10 00617 g005b
Figure 6. Accuracy experiment of AR overlay using HoloLens 2. (a) Superimposing the real sponge brick and the virtual model, and (b) aligning the navigation stick with the path of the virtual guide.
Figure 6. Accuracy experiment of AR overlay using HoloLens 2. (a) Superimposing the real sponge brick and the virtual model, and (b) aligning the navigation stick with the path of the virtual guide.
Bioengineering 10 00617 g006
Figure 7. Boxplot analysis of automatic scanning and target point error. (a) Automatic scanning error, (b) target point error, and (c) line graph illustrating stability test on a virtual phantom.
Figure 7. Boxplot analysis of automatic scanning and target point error. (a) Automatic scanning error, (b) target point error, and (c) line graph illustrating stability test on a virtual phantom.
Bioengineering 10 00617 g007
Figure 8. Spatial image tracking using augmented reality representation on HoloLens 2. (a) Imprecise, and (b) Precise.
Figure 8. Spatial image tracking using augmented reality representation on HoloLens 2. (a) Imprecise, and (b) Precise.
Bioengineering 10 00617 g008
Figure 9. Effect of Viewing Angle Changes on Coordinate Accuracy. (a) X-coordinate error at various navigation stick angles, (b) Y-coordinate error at various navigation stick angles, (c) Z-coordinate error at various navigation stick angles, (d) X-coordinate error at various phantom angles, (e) Y-coordinate error at various phantom angles, and (f) Z-coordinate error at various phantom angles.
Figure 9. Effect of Viewing Angle Changes on Coordinate Accuracy. (a) X-coordinate error at various navigation stick angles, (b) Y-coordinate error at various navigation stick angles, (c) Z-coordinate error at various navigation stick angles, (d) X-coordinate error at various phantom angles, (e) Y-coordinate error at various phantom angles, and (f) Z-coordinate error at various phantom angles.
Bioengineering 10 00617 g009aBioengineering 10 00617 g009b
Table 1. Notations.
Table 1. Notations.
NotationDefinition
D i s x / D i s y / D i s z Distance on the x/y/z axis of the DICOM object
E p The edge point of the scalpel
l d i c o m X 2 D / l d i c o m Y 2 D The length/width of the DICOM image
l X 3 D / l Y 3 D / l Z 3 D The length of the head along the x/y/z axes
N t a r g e t The number of the specific DICOM slice with the ideal target
N t o t a l The total amount of DICOM slices
N u m x / N u m y / N u m z The number of DICOM slices on the x/y/z axis
P o s t a r g e t 2 D ( X t a r g e t 2 D , Y t a r g e t 2 D ) The ideal 2D target position
P o s t a r g e t 3 D ( x , y , z ) The ideal 3D target position
P o s 0 3 D ( 0 , 0 , 0 ) The origin point
R P x L / R P x R The reference point on the left/right of the x-axis
T x / T y / T z The thickness on the x/y/z axis of DICOM slices
T r u e X / T r u e Y / T r u e Z The x/y/z-axis displayed the DICOM slice
Table 2. Definition of common abbreviations.
Table 2. Definition of common abbreviations.
AbbreviationMeaning
AR/VRAugmented reality/Virtual reality
AR devicesAugmented reality devices, which include mobile devices and head-mounted displays, that are capable of running augmented reality software
CNNConvolutional neural network. A type of artificial neural network that belongs to the family of deep learning algorithms
EVDExtra-ventricular drainage
DICOMDigital Imaging and Communications in Medicine
DICOM filesFiles that have been saved in the Digital Imaging and Communications in Medicine (DICOM) format, typically obtained from computed tomography (CT) or magnetic resonance imaging (MRI) procedures.
DICOM dataData that has been formatted in accordance with the Digital Imaging and Communications in Medicine (DICOM) standard
DICOM imageA 2D image that has been saved in the Digital Imaging and Communications in Medicine (DICOM) format and can be displayed on a screen or monitor
DICOM sliceA single two-dimensional image of a patient’s anatomy that has been acquired using imaging modalities such as X-ray, CT scan, MRI, or ultrasound.
HMDHead-mounted device
Virtual imageComputer-generated image that cannot exist in the physical world and can only be viewed on a display monitor
Virtual objectA digital object that lacks a physical presence and can only be created and manipulated within software
U-NetU-Net is a convolutional neural network that was originally developed at the Computer Science Department of the University of Freiburg for biomedical image segmentation. The original U-Net model, which consists of an encoder and decoder, is widely used for tasks such as medical image segmentation, as presented in this paper.
Table 3. Comparison of accuracy, sensitivity, and specificity.
Table 3. Comparison of accuracy, sensitivity, and specificity.
MethodAccuracy (%)Sensitivity (%)Specificity (%)
Proposed99.9393.8595.73
Surgeon85–97N/AN/A
[30]N/A90.085.0
[31]91.8091.8494.77
[32]N/A80.090.0
[2]94.093.694.4
[10]93.0N/AN/A
Table 4. Comparison of registration time, average error, and standard deviation.
Table 4. Comparison of registration time, average error, and standard deviation.
MethodRegistration TimeAverage Error
(mm)
Standard Deviation
(mm)
Proposed4 s10.1
[7]30 s1.960.87
[8]99 s2.5N/A
[35]N/A2.1N/A
[9]228 s1.20.54
[34]N/A3N/A
[13]N/A2N/A
[36]N/A2N/A
[33]N/A1.90.45
[4]N/A2.82.7
[5]300 s7.9N/A
Table 5. Accuracy results of HoloLens 2 feasibility test.
Table 5. Accuracy results of HoloLens 2 feasibility test.
ExperimenterABCDEAverage
Error (mm)5372136
Table 6. Comparison of properties between our method and other methods for using augmented reality in medical guidance.
Table 6. Comparison of properties between our method and other methods for using augmented reality in medical guidance.
Properties[36][37][38][39]Proposed
DeviceHandheld projectoriPadHoloLens 1Mobile phoneSurface Pro 7
Accuracyprojection error:
1.3 ± 0.9 mm
Prosthesis experiment:
2.8 ± 2.7 mm
Average insertion angle difference:
6.35 degree
Prosthesis experiment:
2.7 ± 2.6 mm
Prosthesis experiment:
1.0 ± 0.1 mm
Anatomy information providingVV V
3D display providing VVVV
Path navigation providing VVV
Visual feedback providing V
Clinical feasibilityVV VV
Table 7. Feature comparison of Surface Pro 7 and HoloLens 2.
Table 7. Feature comparison of Surface Pro 7 and HoloLens 2.
FeaturesSurface Pro 7Hololens 2
StabilityBetter
Flexibility Better
ComfortBetter
Information richness Better
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chiou, S.-Y.; Liu, L.-S.; Lee, C.-W.; Kim, D.-H.; Al-masni, M.A.; Liu, H.-L.; Wei, K.-C.; Yan, J.-L.; Chen, P.-Y. Augmented Reality Surgical Navigation System Integrated with Deep Learning. Bioengineering 2023, 10, 617. https://doi.org/10.3390/bioengineering10050617

AMA Style

Chiou S-Y, Liu L-S, Lee C-W, Kim D-H, Al-masni MA, Liu H-L, Wei K-C, Yan J-L, Chen P-Y. Augmented Reality Surgical Navigation System Integrated with Deep Learning. Bioengineering. 2023; 10(5):617. https://doi.org/10.3390/bioengineering10050617

Chicago/Turabian Style

Chiou, Shin-Yan, Li-Sheng Liu, Chia-Wei Lee, Dong-Hyun Kim, Mohammed A. Al-masni, Hao-Li Liu, Kuo-Chen Wei, Jiun-Lin Yan, and Pin-Yuan Chen. 2023. "Augmented Reality Surgical Navigation System Integrated with Deep Learning" Bioengineering 10, no. 5: 617. https://doi.org/10.3390/bioengineering10050617

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop