Next Article in Journal
YouTube as a Source of Patient Information Regarding Exercises and Compensated Maneuvers for Dysphagia
Previous Article in Journal
When Nurses Become Patients. Validation of the Content of the Diagnostic Label Professional Traumatic Grief
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reliability and Quality of YouTube Videos on Ultrasound-Guided Brachial Plexus Block: A Programmatical Review

1
Department of Anesthesiology and Pain Medicine, Gachon University Gil Medical Center, Incheon 21565, Korea
2
Department of Anesthesiology and Pain Medicine, CHA Bundang Medical Center, CHA University School of Medicine, Seongnam 13496, Korea
3
Department of Biomedical Informatics, School of Medicine, Pusan National University, Yangsan 50612, Korea
4
Division of Vascular Surgery, Department of Surgery, Gachon University Gil Medical Center, Incheon 21565, Korea
*
Author to whom correspondence should be addressed.
Healthcare 2021, 9(8), 1083; https://doi.org/10.3390/healthcare9081083
Submission received: 1 June 2021 / Revised: 9 August 2021 / Accepted: 18 August 2021 / Published: 23 August 2021

Abstract

:
Background: Ultrasound-guided regional anesthesia has gained popularity over the last decade. This study aimed to assess whether YouTube videos sufficiently serve as an adjunctive tool for learning how to perform an ultrasound-guided brachial plexus block (BPB). Methods: All YouTube videos were classified, based on their sources, as either academic, manufacturer, educational, or individual videos. The metrics, accuracy, utility, reliability (using the Journal of American Medical Association Score benchmark criteria (JAMAS)), and educational quality (using the Global Quality Score (GQS) and Brachial Plexus Block Specific Quality Score (BSQS)) were validated. Results: Here, 175 videos were included. Academic (1.19 ± 0.62, mean ± standard deviation), manufacturer (1.17 ± 0.71), and educational videos (1.15 ± 0.76) had better JAMAS accuracy and reliability than individual videos (0.26 ± 0.67) (p < 0.001). Manufacturer (11.22 ± 1.63) and educational videos (10.33 ± 3.34) had a higher BSQS than individual videos (7.32 ± 4.20) (p < 0.001). All sources weakly addressed the equipment preparation and post-procedure questions after BSQS analysis. Conclusions: The reliability and quality of ultrasound-guided BPB videos differ depending on their source. As YouTube is a useful educational platform for learners and teachers, global societies of regional anesthesiologists should set a standard for videos.

1. Introduction

Regional anesthesia provides benefits for patients by reducing acute, chronic postoperative pain, postoperative nausea and vomiting, and pulmonary complications [1,2,3,4]. Ultrasound-guided regional anesthesia (UGRA) has grown in popularity over the last decade, and several advancements have led to an increase in its effectiveness and safety [5,6,7,8]. As ultrasonography (US) provides direct visualization of the needle pathway, target nerve, surrounding tissues, and local anesthetic spread around the nerve, and anesthesiologists can rapidly and more accurately perform the nerve block [9,10]. Although there is no definitive evidence that UGRA reduces peripheral nerve injury compared with the traditional nerve stimulation techniques, it has been reported that UGRA reduced the incidence of local anesthetic systemic toxicity [11,12,13] and the frequency of pneumothorax associated with US-guided supraclavicular blocks [14,15,16]. The majority of the teaching physicians believe that US increases the efficacy and safety of regional anesthesia, and recommend that UGRA be included in the teaching programs of residents and fellows [17].
The use of ultrasound is highly dependent on the operator. Despite advances in ultrasound technology, some students may find it harder to master [18,19]. Besides comprehension of the general principles of US and anatomy, UGRA requires new skills, including image interpretation, needle beam alignment, and needle trajectory-tracking [20,21]. Sites et al. [21] analyzed 520 nerve block procedures performed by anesthesia residents and found that the most common errors, such as failure to visualize the needle, need to be addressed in training programs. For procedural safety, it has been consistently suggested that other educational tools are required in addition to the conventional ultrasound workshop [22].
YouTube was established as a social platform in 2005, and has since become a popular educational platform. Rapp et al. [23] showed that YouTube was the most preferred source of surgical videos for medical students, surgical residents, and specialists. A recent survey conducted on surgeons regarding laparoscopic surgery showed that more than 86% of the trainees routinely watched online surgical videos on YouTube to learn or perfect their surgical technique [24]. Similar to laparoscopic surgery, the hand−eye−screen coordination required during UGRA requires practice, as hand and needle movements occur in three different axes, whereas the ultrasound image is only presented in two dimensions.
Several researchers have evaluated the quality of YouTube videos for medical information and skills [25,26,27,28,29,30,31,32]. They found some videos on YouTube to be educationally useful; however, a large portion of videos had a lower quality of education and some even inaccurate information [25,29,33]. This is mostly due to the lack of a review process, similar to the scientific literature publication, uncertainty of sources, and their reliability. Despite the shortcomings of YouTube videos, there is no doubt that many medical students and residents currently still seek educational information on YouTube, and that well-designed educational videos may enhance their learning, thereby serving as a highly effective educational tool [34,35,36,37]. It has been demonstrated that adding online learning methods, such as video materials, to the text learning-based pedagogical approach can improve education [37]. Thus, we aimed to investigate whether YouTube can serve as an adjunctive tool for learning the complex process of UGRA. Several studies have investigated the quality of UGRA YouTube videos. Tewfik et al. [38] showed that user-uploaded videos on YouTube had less educational characteristics than those by the anesthesia society websites. Selvi et al. [39] scored the videos of brachial plexus block (BPB) on YouTube using their own questionnaires. However, to the best of our knowledge, there has been no in-depth analysis of the quality and sources. Therefore, the objectives of this study were as follows: (1) identify the upload sources and characteristics of the YouTube videos, (2) investigate the quality of videos using three different score instruments measuring reliability and educational values, (3) examine the difference in the quality of videos between different sources, and (4) present future directions for high-quality educational videos of UGRA. We focused on the ultrasound-guided BPB, which is the most representative and common type of UGRA.

2. Materials and Methods

2.1. Search Strategy and Inclusion Criteria

YouTube (www.youtube.com) was systematically searched in July 2020 via the YouTube Data API and Google Apps Script. The code for Google App Script can be found at “https://github.com/igreg1221/YouTube-API-Search-”. We identified YouTube videos uploaded from 1 January 2005 to 31 December 2019 with the terms “brachial plexus block”, “interscalene block”, “supraclavicular block”, “infraclavicular block”, and “axillary block” in the title. Videos that demonstrated the procedures using US were included. The exclusion criteria were videos without audio or text, videos with non-English narration or captions, PowerPoint presentation slides, animations, videos that did not use the US in the procedures, procedures not performed on humans, videos only demonstrating surgical procedures, and videos with a resolution lower than 360p (480 × 360 pixels).

2.2. Video Metrics

Using the unique video identifier acquired from the Google Apps Script, the following video parameters were extracted using the R package “tuber” (https://CRAN.R-project.org/package=tuber (accessed on July, 2020)): (1) title, (2) video duration, (3) date of publication, (4) number of views, (5) number of likes, (6) number of dislikes, (7) number of comments, (8) resolution, (9) like ratio (Like × 100/(Like + Dislike)), (10) view ratio (number of views/days), and (11) Video Power Index (VPI) (like ratio × view ratio/100). The VPI, which was first described by Erdem et al. [40], was used to evaluate the popularity of the videos. All videos were classified based on their source, which was (1) individual (independent medical doctor without any affiliation mentioned, (2) academic (hospital or university affiliation pertaining to authors), (3) manufacturer (US manufacturer affiliation pertaining to authors), and (4) educational (corporation videos for educational purpose, i.e., The New York School of Regional Anesthesia (NYSORA), Ultrasound-Guided Regional Anesthesia and Pain Medicine (USRA)).

2.3. Video Accuracy, Utility and Reliability

The accuracy, utility, and reliability of each video were evaluated according to the Journal of American Medical Association Score (JAMAS) benchmark criteria on a scale of 0 to 4, as suggested by Silberg et al. [41]. The JAMAS benchmark criteria (Table S1) consists of four individual criteria, with one point assigned for each criterion as a non-specific assessment of the source’s reliability. A score of four indicates a higher accuracy, utility, and reliability, whereas a score of 0 indicates poor accuracy, utility, and reliability.

2.4. Educational Value of Video

We used two scoring systems to assess the educational value of the videos. The Global Quality Score (GQS) [40,42] provides a non-specific assessment of the educational value, with five criteria (Table S2). The Global Quality Score has a scale of 0 to 5, and a higher score indicates higher educational quality. For a more specific quality assessment of the videos, two anesthesiologists (N.R.C. and J.J.P.) who routinely perform BPB in daily practice, utilized a Brachial Plexus Block Specific Quality Score (BSQS) (Table 1) based on the Miller’s Anesthesia textbook [43], recommendations from the American and European society of regional anesthesia and pain therapy joint committee [20], and published articles [44,45,46,47]. To assess the extent of the procedure-specific knowledge and skills the videos contained, we evaluated the BSQS from the procedure preparation to post-procedural steps. The BSQS consisted of pre-procedure, equipment preparation, intra-procedure, and post-procedure scores. All details are provided in Table 1. Briefly, for pre-procedure scores, videos mentioned the approaches and indicated the procedures, targeted dermatome, and the patient’s position. For equipment preparation, the equipment for ultrasound-guided BPB with aseptic fashion was assessed. For the intra-procedure scores, videos showed the sonographic view of the procedure with anatomical landmarks, needling techniques, and needle tip confirmation. Lastly, for post-procedure scores, the videos mentioned how to verify the success of procedures and types of complications. Of the 16 criteria, five indispensable steps were assessed, and the two scores were weighted. The videos were given 1 point if the instructions were presented orally or in text. All scorings were performed independently by two authors. Videos with different scores were reassessed until a consensus was reached. To visualize the BSQS scores according to video sources, we have shown the percentage of BSQS questions as a heatmap.

2.5. Statistical Analysis

The R program (Version 3.6.0, R Foundation for Statistical Computing, Vienna, Austria) was used for the statistical analyses. Descriptive statistics were used to quantify the video characteristics. As the parameters did not show a normal distribution, the Kruskal−Wallis test was used for the intergroup comparisons and the Mann−Whitney U test was used to identify the group that caused the difference. All data are presented as mean ± standard deviation (median). The significance level was set at p < 0.05. Spearman rank correlation was used to identify the correlations between the variables. Cohen ƙ and intraclass correlation coefficients were calculated to evaluate the degree among the raters using R package “irr” (https://cran.r-project.org/web/packages/irr/index.html (accessed on August 2020)) with a single-measurement, absolute-agreement, and two-way mixed effects model [48].

3. Results

3.1. Overall Video Metrics

The five search terms yielded 799 unique video identifiers through the YouTube Data API and Google Apps Script (Figure 1). After excluding 170 duplicates, 629 videos were watched in full detail. A total of 444 videos met the exclusion criteria, and the most common reason for exclusion was the lack of audio or captions in the videos (n = 217). Finally, 175 videos were analyzed for their quality assessment. Overall, 50.3% (n = 88) of the videos were individual, 20.6% (n = 36) were academic, 10.3% (n = 18) were manufacturer, and 18.9% (n = 33) were educational videos (Table 2). All variables of 175 videos, including the unique identifier, BSQS, GQS, JAMAS, and VPI, among others, are provided in the Table S3. The mean video duration was 247.67 ± 208.40 (191) s. The mean number of views was 18,907.94 ± 46,373.86 (785) times. The mean view ratio was 7.73 ± 16.78 (0.53). The mean number of days since upload was 1948.99 ± 1079.98 (1883) days. The mean number of comments was 1.38 ± 3.15 (0). Videos received an average of 50.30 ± 121.30 (3) likes and 3.21 ± 7.20 (0) dislikes, with a mean-like ratio of 92.33 ± 17.28 (100). The mean VPI was 8.82 ± 17.38 (0.88), JAMAS was 0.71 ± 0.82 (1), and the GQS and BSQS scores were 1.68 ± 0.74 (2) and 8.58 ± 4.03 (9), respectively. The JAMAS, GQS, and BSQS scores were determined through consensus when the scores from two anesthesiologists showed any discrepancy. Before consensus, the intraclass correlation coefficients calculated for BSQS, GQS, and JAMAS were 0.862 (95% confidence interval (CI)-0.818 to 0.895), 0.802 (95% CI-0.742 to 0.849), and 0.883 (95% CI-0.844 to 0.912).
The numbers of views, likes, dislikes, comments, and VPI were significantly different across the sources (Table 2). The count of views was highest for manufacturer videos and statistically higher than that of individual and academic videos (p < 0.001). Educational videos had more likes than individual and academic videos (p < 0.001). Manufacturer and educational videos had more dislikes than individual videos (p < 0.001). The VPI, as a measurement of popularity, was higher on the manufacturer and educational videos than that for the individual videos (p < 0.001). The length of the video was not different among sources (p = 0.32). The JAMAS benchmark criteria were used to assess the accuracy, utility, and reliability. Academic (1.19 ± 0.62), manufacturer (1.17 ± 0.71), and educational videos (1.15 ± 0.76) had higher JAMAS scores than individual videos (0.26 ± 0.67) (p < 0.001). The GQS and BSQS were used for the non-specific and specific assessment of educational quality, respectively. Manufacturer (11.22 ± 1.63) and educational videos (10.33 ± 3.34) had a higher BSQS than individual videos (7.32 ± 4.20) (p < 0.001). The mean GQS did not statistically differ based on the video source (p = 0.18). In the correlation analysis of BSQS, GQS, and JAMAS, BSQS and GQS showed a strong correlation (ρ = 0.74, p < 0.001) (Multimedia Figure S1). Scores and other quantitative variables (view, like, dislike, VPI, among others) showed a positive correlation; however, the like ratio (Like × 100 / (Like + Dislike)) showed a negative correlation with all other variables.

3.2. Further Analysis of Reliability and Quality Assessment Scores

We determined the percentage of BSQS questions that were addressed per source and depicted it as a heatmap (Figure 2). Most notably, all sources weakly addressed the equipment preparation (Q4–Q8) and post-procedure questions (Q15–16). The intra-procedure questions (Q9–Q14) were the most addressed. To investigate the difference by year for BSQS, GQS, and JAMAS, we analyzed the changes in the scores by year according to the sources (Figure 3). For BSQS and GQS, no apparent changes were observed for all sources. In JAMAS, all videos except individual videos showed a tendency for incline by year.
The decline of BSQS of individual and academic videos by year was significantly lower than that of the manufacturer and educational videos. Due to an increase in awareness of intellectual property rights and the reinforcement of restrictions, the standardized reporting system has become more common when uploading medical information on YouTube. As a result, the JAMAS score seems to have increased by year on all videos except the individual videos.

4. Discussion

We analyzed the educational quality of BPB videos on YouTube and found that manufacturer and educational videos had higher scores in all aspects; however, all sources had common deficiencies in their contents and did not increase the learner’s participation. The coronavirus disease-19 (COVID-19) pandemic has led to a reduction in the number of elective surgeries being performed. As a result, the opportunity for fellows to practice regional anesthesia has also been reduced [49,50,51]. Moreover, due to social distancing and other local policies regarding the size of meetings or gatherings, conferences and hands-on courses have been canceled or substituted for remote video and audio conferences. Although there is no substitute for expert guidance by experienced instructors in clinical practice, a new learning tool may be required to compensate for the lack of conventional education in this era [52]. Many universities have launched their channels, including medicine and science, encouraging students and teachers to cultivate a student−teacher coordinated effort and enable real-time feedback from students [53,54]. To this end, we investigated the educational quality of YouTube videos demonstrating a BPB by performing a systematic search and applying multiple scoring systems.
When assessing the source, the manufacturer and educational videos had a higher BSQS, GQS, JAMAS score, and VPI compared with individual videos. In contrast, academic videos only had a higher JAMAS score compared with individual videos. The reliability and quality of the manufacturer videos were better than that of the individual ones; they focused on US probe usage and intra-procedure techniques, such as needle tip visualization. Despite these limitations, for most manufacturer’s videos, well-informed experts performed and explained the procedure. The educational videos were mostly uploaded by NYSORA and Ultrasound for Regional Anesthesia (URSA). These websites are well-organized with highly accurate educational content; novice trainees can learn basic anatomy and clinical skills for UGRA from NYSORA and URSA [55,56,57,58,59,60,61]. We observed several common features in all videos; (1) the BSQS lacked the equipment preparation and post-procedure information and (2) there were very few comments (1.38 ± 3.15 (median: 0)). We weighted the two scores on all questions in post-procedure scores as observing success and complications is as important as the procedure itself. However, most videos, regardless of their source, scarcely mentioned the post-procedural information. These common deficiencies should be addressed when making ultrasound-guided BPB videos. Furthermore, the lack of comments indicates that videos did not attract students’ participation and may not play an active role on social media outreach.
The influence of videos in medical education is likely to increase. Studies have pointed out the benefits of multimedia-enhanced teaching; it significantly improves surgical performance and understanding of complex temporal and spatial events [62,63,64]. Furthermore, various trials based on video-based coaching have been performed to improve the surgical techniques [65,66,67,68]. Mota et al. [69] investigated the difference in the characteristics of video usage between residents and specialists. Interestingly, they showed that residents used YouTube more significantly than the specialists. They preferred a more easily accessible information tool, with feedback, comments, and various points of view on each topic; several of these are the main advantages of YouTube [70]. With increasing access to the Internet and the widespread use of mobile devices, trainees with limited access to information on novel techniques and technologies may also be able to access the instructional material. However, these videos do not undergo a peer-review process and are only screened for copyright infringement, not for their educational value and quality. Many authors reported poor video quality, inaccurate information, incomprehensible or lack of audio, and the lack of background patient information on videos for common surgical procedures such as appendectomy and cholecystectomy [71]. Many authors agree to use YouTube videos as an educational tool; however, prior to utilizing these videos, they caution against misleading information [32]. Urgent guideline advocacy is required for publishing educational videos from each society of medical specialties [72].
Despite these drawbacks of YouTube videos, the affordability, easy access, and the ability to interact with the global community firmly establishes YouTube as an educational learning tool. In the COVID-19 pandemic era, the challenging conditions will undoubtedly minimize the resident teaching and exposure to regional anesthesia [73,74]. A dramatic drop in exposure to regional anesthesia training among anesthesia residents may potentially lead to negative effects in the future. Medina et al. [75] suggested strategies such as watching didactic material and high-quality videos to maintain relevant education and training in regional anesthesia procedures.
Our study programmatically reviewed ultrasound-guided BPB videos on YouTube and evaluated the reliability and quality of videos. YouTube videos on UGRA procedures must be validated, and the global community of regional anesthesiologists should play an active role. Our reporting on the common deficiencies in the contents of videos can help improve the future video quality of regional anesthesia techniques. Moreover, using the social media platform, communication between uploaders and learners should be encouraged to maximize the effectiveness of learning on YouTube. YouTube has the potential to be the largest educational platform; therefore, regional anesthesiologists worldwide should assess the quality of these videos and promote effective communication through YouTube. Our results can be applied to any field of medical education, especially those that require repetitive practices and hand−eye−screen coordination, such as hybrid (open and endovascular) vascular surgery. Our study has limitations. Metrics, such as likes, dislikes, and the length of the video do not fully represent the viewer’s response. The average percent viewed may show how long viewers watched the total length of the video; however, YouTube Data API does not provide these due to privacy issues. We were unable to conclude that high-scored videos helped viewers improve their knowledge and skills. In a future study, this should be implemented to make a video that satisfies all the scores we used, and then release it on YouTube to see if it helps viewers improve UGRA learning.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/healthcare9081083/s1, Table S1: The Journal of American Medical Association Score Benchmark Criteria, Table S2: Global Quality Score, Table S3: Metrics, Reliability, and Quality Assessments of included 175 videos, Figure S1: Heatmap of Spearman rank-order correlation coefficients matrix between scores and quantitative variables (−1:1 shown by color legend). The number in the box represents the correlation coefficient and the color in the box is intensified as the coefficient increases. The correlation coefficient with statistical insignificance remains without color. Significance level is p < 0.05.

Author Contributions

N.R.C. and D.S.K. made substantial contributions to the conception and design. N.R.C. and J.J.P. made the questionnaires (Brachial Plexus Block Specific Quality Score) and collected data. N.R.C. and Y.H.K. performed the statistical analysis. N.R.C. wrote the manuscript and D.S.K. made critical revisions to the manuscript. J.H.C. made contributions to the revision substantially. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by grants from the National Research Foundation of Korea (NRF), the Korean government (NRF-2020R1A2C1102433), the Young Medical Scientist Research Grant through the Daewoong Foundation (DY20111P), and the Korea Medical Institute.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors have no competing interests.

Abbreviations

BPB: Brachial Plexus Block; BSQS: Brachial Plexus Block Specific Quality Score; COVID-19: Coronavirus Disease-19; GQS: Global Quality Score; JAMAS: Journal of American Medical Association Score; NYSORA: New York School of Regional Anesthesia; UGRA: Ultrasound-guided regional anesthesia; US: Ultrasonography; USRA: Ultrasound-Guided Regional Anesthesia and Pain Medicine; VPI: Video Power Index.

References

  1. Gan, T.J.; Diemunsch, P.; Habib, A.S.; Kovac, A.; Kranke, P.; Meyer, T.A.; Watcha, M.; Chung, F.; Angus, S.; Apfel, C.C.; et al. Consensus Guidelines for the Management of Postoperative Nausea and Vomiting. Anesth. Analg. 2014, 118, 85–113. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. van Lier, F.; van der Geest, P.J.; Hoeks, S.E.; van Gestel, Y.R.; Hol, J.W.; Sin, D.D.; Stolker, R.J.; Poldermans, D. Epidural Analgesia is Associated with Improved Health Outcomes of Surgical Patients with Chronic Obstructive Pulmonary Disease. Anesthesiology 2011, 115, 315–321. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Smith, L.M.; Cozowicz, C.; Uda, Y.; Memtsoudis, S.G.; Barrington, M.J. Neuraxial and Combined Neuraxial/General Anesthesia Compared to General Anesthesia for Major Truncal and Lower Limb Surgery: A Systematic Review and Meta-analysis. Anesth. Analg. 2017, 125, 1931–1945. [Google Scholar] [CrossRef] [PubMed]
  4. Andreae, M.H.; Andreae, D.A. Regional Anaesthesia to Prevent Chronic Pain after Surgery: A Cochrane Systematic Review and Meta-Analysis. Br. J. Anaesth. 2013, 111, 711–720. [Google Scholar] [CrossRef] [Green Version]
  5. Haskins, S.C.; Tanaka, C.Y.; Boublik, J.; Wu, C.L.; Sloth, E. Focused Cardiac Ultrasound for the Regional Anesthesiologist and Pain Specialist. Reg. Anesth. Pain Med. 2017, 42, 632–644. [Google Scholar] [CrossRef] [PubMed]
  6. Haskins, S.C.; Tsui, B.C.; Nejim, J.A.; Wu, C.L.; Boublik, J. Lung Ultrasound for the Regional Anesthesiologist and Acute Pain Specialist. Reg. Anesth. Pain Med. 2017, 42, 289–298. [Google Scholar] [CrossRef] [PubMed]
  7. Neal, J.M.; Gravel Sullivan, A.; Rosenquist, R.W.; Kopacz, D.J. Regional Anesthesia and Pain Medicine: US Anesthesiology Resident Training-The Year 2015. Reg. Anesth. Pain Med. 2017, 42, 437–441. [Google Scholar] [CrossRef] [PubMed]
  8. Neal, J.M.; Brull, R.; Horn, J.L.; Liu, S.S.; McCartney, C.J.; Perlas, A.; Salinas, F.V.; Tsui, B.C. The Second American Society of Regional Anesthesia and Pain Medicine Evidence-Based Medicine Assessment of Ultrasound-Guided Regional Anesthesia: Executive Summary. Reg. Anesth. Pain Med. 2016, 41, 181–194. [Google Scholar] [CrossRef] [PubMed]
  9. Lee, M.H.; Kim, K.Y.; Song, J.H.; Jung, H.J.; Lim, H.K.; Lee, D.I.; Cha, Y.D. Minimal Volume of Local Anesthetic Required for an Ultrasound-Guided SGB. Pain Med. 2012, 13, 1381–1388. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, S.S. Evidence Basis for Ultrasound-Guided Block Characteristics Onset, Quality, and Duration. Reg. Anesth. Pain Med. 2016, 41, 205–220. [Google Scholar] [CrossRef]
  11. Orebaugh, S.L.; Kentor, M.L.; Williams, B.A. Adverse Outcomes Associated with Nerve Stimulator-Guided and Ultrasound-Guided Peripheral Nerve Blocks by Supervised Trainees: Update of a Single-Site Database. Reg. Anesth. Pain Med. 2012, 37, 577–582. [Google Scholar] [CrossRef]
  12. Lecours, M.; Lévesque, S.; Dion, N.; Nadeau, M.J.; Dionne, A.; Turgeon, A.F. Complications of Single-Injection Ultrasound-Guided Infraclavicular Block: A Cohort Study. Can. J. Anaesth. 2013, 60, 244–252. [Google Scholar] [CrossRef] [Green Version]
  13. Liu, S.S.; Gordon, M.A.; Shaw, P.M.; Wilfred, S.; Shetty, T.; Yadeau, J.T. A Prospective Clinical Registry of Ultrasound-Guided Regional Anesthesia for Ambulatory Shoulder Surgery. Anesth. Analg. 2010, 111, 617–623. [Google Scholar] [CrossRef]
  14. Tsui, B.C.; Doyle, K.; Chu, K.; Pillay, J.; Dillane, D. Case Series: Ultrasound-Guided Supraclavicular Block Using a Curvilinear Probe in 104 Day-Case Hand Surgery Patients. Can. J. Anaesth. 2009, 56, 46–51. [Google Scholar] [CrossRef]
  15. Abell, D.J.; Barrington, M.J. Pneumothorax After Ultrasound-Guided Supraclavicular Block: Presenting Features, Risk, and Related Training. Reg. Anesth. Pain Med. 2014, 39, 164–167. [Google Scholar] [CrossRef]
  16. Neal, J.M. Ultrasound-Guided Regional Anesthesia and Patient Safety: Update of an Evidence-Based Analysis. Reg. Anesth. Pain Med. 2016, 41, 195–204. [Google Scholar] [CrossRef] [PubMed]
  17. Helwani, M.A.; Saied, N.N.; Asaad, B.; Rasmussen, S.; Fingerman, M.E. The Current Role of Ultrasound Use in Teaching Regional Anesthesia: A Survey of Residency Programs in the United States. Pain Med. 2012, 13, 1342–1346. [Google Scholar] [CrossRef]
  18. Bodenham, A.R. Editorial II: Ultrasound Imaging by Anaesthetists: Training and Accreditation Issues. Br. J. Anaesth. 2006, 96, 414–417. [Google Scholar] [CrossRef] [Green Version]
  19. Ramlogan, R.; Manickam, B.; Chan, V.W.; Liang, L.; Adhikary, S.D.; Liguori, G.A.; Hargett, M.J.; Brull, R. Challenges and Training Tools Associated with the Practice of Ultrasound-Guided Regional Anesthesia: A Survey of the American Society of Regional Anesthesia and Pain Medicine. Reg. Anesth. Pain Med. 2010, 35, 224–226. [Google Scholar] [CrossRef]
  20. Sites, B.D.; Chan, V.W.; Neal, J.M.; Weller, R.; Grau, T.; Koscielniak-Nielsen, Z.J.; Ivani, G. The American Society of Regional Anesthesia and Pain Medicine and the European Society of Regional Anaesthesia and Pain Therapy Joint Committee Recommendations for Education and Training in Ultrasound-Guided Regional Anesthesia. Reg. Anesth. Pain Med. 2009, 34, 40–46. [Google Scholar] [CrossRef] [PubMed]
  21. Sites, B.D.; Spence, B.C.; Gallagher, J.D.; Wiley, C.W.; Bertrand, M.L.; Blike, G.T. Characterizing Novice Behavior Associated with Learning Ultrasound-Guided Peripheral Regional Anesthesia. Reg. Anesth. Pain Med. 2007, 32, 107–115. [Google Scholar] [CrossRef] [PubMed]
  22. Das Adhikary, S.; Karanzalis, D.; Liu, W.R.; Hadzic, A.; McQuillan, P.M. A Prospective Randomized Study to Evaluate a New Learning Tool for Ultrasound-Guided Regional Anesthesia. Pain Med. 2017, 18, 856–865. [Google Scholar] [CrossRef] [Green Version]
  23. Rapp, A.K.; Healy, M.G.; Charlton, M.E.; Keith, J.N.; Rosenbaum, M.E.; Kapadia, M.R. YouTube is the Most Frequently Used Educational Video Source for Surgical Preparation. J. Surg. Educ. 2016, 73, 1072–1076. [Google Scholar] [CrossRef] [PubMed]
  24. Celentano, V.; Smart, N.; Cahill, R.A.; McGrath, J.S.; Gupta, S.; Griffith, J.P.; Acheson, A.G.; Cecil, T.D.; Coleman, M.G. Use of Laparoscopic Videos Amongst Surgical Trainees in the United Kingdom. Surgeon 2019, 17, 334–339. [Google Scholar] [CrossRef]
  25. Mueller, S.M.; Jungo, P.; Cajacob, L.; Schwegler, S.; Itin, P.; Brandt, O. The Absence of Evidence is Evidence of Non-Sense: Cross-Sectional Study on the Quality of Psoriasis-Related Videos on YouTube and Their Reception by Health Seekers. J. Med. Internet Res. 2019, 21, e11935. [Google Scholar] [CrossRef]
  26. Gao, X.; Hamzah, S.H.; Yiu, C.K.; McGrath, C.; King, N.M. Dental Fear and Anxiety in Children and Adolescents: Qualitative Study Using YouTube. J. Med. Internet Res. 2013, 15, e29. [Google Scholar] [CrossRef] [PubMed]
  27. Azer, S.A.; Algrain, H.A.; AlKhelaif, R.A.; AlEshaiwi, S.M. Evaluation of the Educational Value of YouTube Videos About Physical Examination of the Cardiovascular and Respiratory Systems. J. Med. Internet Res. 2013, 15, e241. [Google Scholar] [CrossRef] [PubMed]
  28. Ferhatoglu, M.F.; Kartal, A.; Ekici, U.; Gurkan, A. Evaluation of the Reliability, Utility, and Quality of the Information in Sleeve Gastrectomy Videos Shared on Open Access Video Sharing Platform YouTube. Obes. Surg. 2019, 29, 1477–1484. [Google Scholar] [CrossRef] [PubMed]
  29. Mueller, S.M.; Hongler, V.N.S.; Jungo, P.; Cajacob, L.; Schwegler, S.; Steveling, E.H.; Manjaly Thomas, Z.R.; Fuchs, O.; Navarini, A.; Scherer, K.; et al. Fiction, Falsehoods, and Few Facts: Cross-Sectional Study on the Content-Related Quality of Atopic Eczema-Related Videos on YouTube. J. Med. Internet Res. 2020, 22, e15599. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Syed-Abdul, S.; Fernandez-Luque, L.; Jian, W.S.; Li, Y.C.; Crain, S.; Hsu, M.H.; Wang, Y.C.; Khandregzen, D.; Chuluunbaatar, E.; Nguyen, P.A.; et al. Misleading Health-Related Information Promoted Through Video-Based Social Media: Anorexia on YouTube. J. Med. Internet Res. 2013, 15, e30. [Google Scholar] [CrossRef]
  31. Basch, C.H.; Hillyer, G.C.; Meleo-Erwin, Z.C.; Jaime, C.; Mohlman, J.; Basch, C.E. Preventive Behaviors Conveyed on YouTube to Mitigate Transmission of COVID-19: Cross-Sectional Study. JMIR. Public Health Surveill. 2020, 6, e18807. [Google Scholar] [CrossRef] [PubMed]
  32. Van den Eynde, J.; Crauwels, A.; Demaerel, P.G.; Van Eycken, L.; Bullens, D.; Schrijvers, R.; Toelen, J. YouTube Videos as a Source of Information About Immunology for Medical Students: Cross-Sectional Study. JMIR. Med. Educ. 2019, 5, e12605. [Google Scholar] [CrossRef]
  33. Farag, M.; Bolton, D.; Lawrentschuk, N. Use of YouTube as a Resource for Surgical Education-Clarity or Confusion. Eur. Urol. Focus. 2020, 6, 445–449. [Google Scholar] [CrossRef] [PubMed]
  34. Hsin, W.J.; Cigas, J. Short Videos Improve Student Learning in Online Education. J. Comput. Sci. Coll. 2013, 28, 253–259. [Google Scholar]
  35. Rackaway, C. Video Killed the Textbook Star? Use of Multimedia Supplements to Enhance Student Learning. J. Political Sci. Educ. 2012, 8, 189–200. [Google Scholar] [CrossRef]
  36. Schmid, R.F.; Bernard, R.M.; Borokhovski, E.; Tamim, R.M.; Abrami, P.C.; Surkes, M.A.; Wade, C.A.; Woods, J. The Effects of Technology Use in Postsecondary Education: A Meta-Analysis of Classroom Applications. Comput. Educ. 2014, 72, 271–291. [Google Scholar] [CrossRef]
  37. Stockwell, B.R.; Stockwell, M.S.; Cennamo, M.; Jiang, E. Blended Learning Improves Science Education. Cell 2015, 162, 933–936. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Tewfik, G.L.; Work, A.N.; Shulman, S.M.; Discepola, P. Objective Validation of YouTube™ Educational Videos for the Instruction of Regional Anesthesia Nerve Blocks: A Novel Approach. BMC. Anesthesiol. 2020, 20, 168. [Google Scholar] [CrossRef]
  39. Selvi, O.; Tulgar, S.; Senturk, O.; Topcu, D.I.; Ozer, Z. [YouTube as an Informational Source for Brachial Plexus Blocks: Evaluation of Content and Educational Value]. Braz. J. Anesthesiol. 2019, 69, 168–176. [Google Scholar] [CrossRef] [PubMed]
  40. Erdem, M.N.; Karaca, S. Evaluating the Accuracy and Quality of the Information in Kyphosis Videos Shared on YouTube. Spine (Phila Pa 1976) 2018, 43, E1334–E1339. [Google Scholar] [CrossRef]
  41. Silberg, W.M.; Lundberg, G.D.; Musacchio, R.A. Assessing, Controlling, and Assuring the Quality of Medical Information on the Internet: Caveant Lector et Viewor—Let the Reader and Viewer Beware. JAMA 1997, 277, 1244–1245. [Google Scholar] [CrossRef] [PubMed]
  42. Singh, A.G.; Singh, S.; Singh, P.P. YouTube for Information on Rheumatoid Arthritis—A Wakeup Call. J. Rheumatol. 2012, 39, 899–903. [Google Scholar] [CrossRef] [PubMed]
  43. Miller, R.D. Miller’s Anesthesia, 9th ed.; Elseiver: Amsterdam, The Netherlands, 2020. [Google Scholar]
  44. Kannan, S.; Marri, S.; Sivasubramaniam, S. USG for Brachial Plexus Block. Anaesth. Pain Intensive Care 2015, 19, 333–340. [Google Scholar]
  45. Marhofer, P.; Willschke, H.; Kettner, S.C. Ultrasound-guided Upper Extremity Blocks—Tips and Tricks to Improve the Clinical Practice. Paediatr. Anaesth. 2012, 22, 65–71. [Google Scholar] [CrossRef]
  46. Raju, P.K.B.C.; Coventry, D.M. Ultrasound-Guided Brachial Plexus Blocks. Contin. Educ. Anaesth. Crit. Care Pain 2013, 14, 185–191. [Google Scholar] [CrossRef] [Green Version]
  47. Sandhu, N.S.; Capan, L.M. Ultrasound-guided Infraclavicular Brachial Plexus Block. Br. J. Anaesth. 2002, 89, 254–259. [Google Scholar] [CrossRef] [Green Version]
  48. Koo, T.K.; Li, M.Y. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. J. Chiropr. Med. 2016, 15, 155–163. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Diaz, A.; Sarac, B.A.; Schoenbrunner, A.R.; Janis, J.E.; Pawlik, T.M. Elective Surgery in the Time of COVID-19. Am. J. Surg. 2020, 219, 900–902. [Google Scholar] [CrossRef]
  50. Dowdell, J.E.; Louie, P.K.; Virk, S.; McCarthy, M.H.; Sandhu, H.S.; Qureshi, S.A.; Albert, T.J.; Kim, H.J. Spine Fellowship Training Reorganizing During a Pandemic: Perspectives from a Tertiary Orthopedic Specialty Center in the Epicenter of Outbreak. Spine J. 2020, 20, 1381–1385. [Google Scholar] [CrossRef] [PubMed]
  51. White, E.M.; Shaughnessy, M.P.; Esposito, A.C.; Slade, M.D.; Korah, M.; Yoo, P.S. Surgical Education in the Time of COVID: Understanding the Early Response of Surgical Training Programs to the Novel Coronavirus Pandemic. J. Surg. Educ. 2021, 78, 412–421. [Google Scholar] [CrossRef]
  52. Osgerby, J.; Rush, D. An Exploratory Case Study Examining Undergraduate Accounting Students’ Perceptions of Using Twitter as a Learning Support Tool. Int. J. Manag. Educ. 2015, 13, 337–348. [Google Scholar] [CrossRef]
  53. Duncan, I.; Yarwood-Ross, L.; Haigh, C. YouTube as a Source of Clinical Skills Education. Nurse Educ. Today 2013, 33, 1576–1580. [Google Scholar] [CrossRef]
  54. Torres-Ramírez, M.; García-Domingo, B.; Aguilera, J.; De La Casa, J. Video-Sharing Educational Tool Applied to the Teaching in Renewable Energy Subjects. Comput. Educ. 2014, 73, 160–177. [Google Scholar] [CrossRef]
  55. Abhinaya, R.J.; Venkatraman, R.; Matheswaran, P.; Sivarajan, G. A Randomised Comparative Evaluation of Supraclavicular and Infraclavicular Approaches to Brachial Plexus Block for Upper Limb Surgeries Using Both Ultrasound and Nerve Stimulator. Indian J. Anaesth. 2017, 61, 581–586. [Google Scholar] [CrossRef] [PubMed]
  56. Ferede, Y.A.; Nigatu, Y.A.; Agegnehu, A.F.; Mustofa, S.Y. Practice of Spinal Anesthesia Among Anesthetists in the Operation Room of Referral Hospital: Cross-Sectional Study. Int. J. Surg. Open 2020, 27, 145–148. [Google Scholar] [CrossRef]
  57. Greengrass, R.A.; Narouze, S.; Bendtsen, T.F.; Hadzic, A. Cervical Plexus and Greater Occipital Nerve Blocks: Controversies and Technique Update. Reg. Anesth. Pain Med. 2019, 44, 623–626. [Google Scholar] [CrossRef]
  58. Jones, M.R.; Novitch, M.B.; Sen, S.; Hernandez, N.; De Haan, J.B.; Budish, R.A.; Bailey, C.H.; Ragusa, J.; Thakur, P.; Orhurhu, V.; et al. Upper Extremity Regional Anesthesia Techniques: A Comprehensive Review for Clinical Anesthesiologists. Best Pract. Res. Clin. Anaesthesiol. 2020, 34, e13–e29. [Google Scholar] [CrossRef]
  59. Rao Kadam, V.; Ludbrook, G.; van Wijk, R.; Hewett, P.; Thiruvenkatarajan, V.; Moran, J.; Williams, P. Ultrasound-guided Transmuscular Quadratus Lumborum Block Catheter Technique: A Reply. Anaesthesia 2020, 75, 414–415. [Google Scholar] [CrossRef]
  60. Scala, V.A.; Lee, L.S.K.; Atkinson, R.E. Implementing Regional Nerve Blocks in Hip Fracture Programs: A Review of Regional Nerve Blocks, Protocols in the Literature, and the Current Protocol at The Queen’s Medical Center in Honolulu, HI. Hawaii J. Health Soc. Welf. 2019, 78, 11–15. [Google Scholar]
  61. Tucker, R.V.; Huang, R.; Peterson, W.J.; Munzer, B.W.; Thiessen, M. An Ultrasound-Guided Regional Anesthesia Elective for Emergency Medicine Residents. J. Educ. Teach. Emerg. Med. 2021, 6. [Google Scholar] [CrossRef]
  62. Mayer, R.E. Applying the Science of Learning: Evidence-Based Principles for the Design of Multimedia Instruction. Am. Psychol. 2008, 63, 760–769. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. Pape-Koehler, C.; Immenroth, M.; Sauerland, S.; Lefering, R.; Lindlohr, C.; Toaspern, J.; Heiss, M. Multimedia-based Training on Internet Platforms Improves Surgical Performance: A Randomized Controlled Trial. Surg. Endosc. 2013, 27, 1737–1747. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  64. Friedl, R.; Höppler, H.; Ecard, K.; Scholz, W.; Hannekum, A.; Ochsner, W.; Stracke, S. Multimedia-Driven Teaching Significantly Improves Students’ Performance When Compared with a Print Medium. Ann. Thorac. Surg. 2006, 81, 1760–1766. [Google Scholar] [CrossRef]
  65. Singh, P.; Aggarwal, R.; Tahir, M.; Pucher, P.H.; Darzi, A. A Randomized Controlled Study to Evaluate the Role of Video-Based Coaching in Training Laparoscopic Skills. Ann. Surg. 2015, 261, 862–869. [Google Scholar] [CrossRef] [PubMed]
  66. Bonrath, E.M.; Dedy, N.J.; Gordon, L.E.; Grantcharov, T.P. Comprehensive Surgical Coaching Enhances Surgical Skill in the Operating Room: A Randomized Controlled Trial. Ann. Surg. 2015, 262, 205–212. [Google Scholar] [CrossRef] [PubMed]
  67. Karam, M.D.; Thomas, G.W.; Koehler, D.M.; Westerlind, B.O.; Lafferty, P.M.; Ohrt, G.T.; Marsh, J.L.; Van Heest, A.E.; Anderson, D.D. Surgical Coaching from Head-Mounted Video in the Training of Fluoroscopically Guided Articular Fracture Surgery. J. Bone Jt. Surg. Am. 2015, 97, 1031–1039. [Google Scholar] [CrossRef] [PubMed]
  68. Cole, S.J.; Mackenzie, H.; Ha, J.; Hanna, G.B.; Miskovic, D. Randomized controlled trial on the effect of coaching in simulated laparoscopic training. Surg. Endosc. 2014, 28, 979–986. [Google Scholar] [CrossRef] [PubMed]
  69. Mota, P.; Carvalho, N.; Carvalho-Dias, E.; João Costa, M.; Correia-Pinto, J.; Lima, E. Video-Based Surgical Learning: Improving Trainee Education and Preparation for Surgery. J. Surg. Educ. 2018, 75, 828–835. [Google Scholar] [CrossRef] [Green Version]
  70. Drozd, B.; Couvillon, E.; Suarez, A. Medical YouTube Videos and Methods of Evaluation: Literature Review. JMIR Med. Educ. 2018, 4, e3. [Google Scholar] [CrossRef] [Green Version]
  71. Lee, J.S.; Seo, H.S.; Hong, T.H. YouTube as a Potential Training Method for Laparoscopic Cholecystectomy. Ann. Surg. Treat. Res. 2015, 89, 92–97. [Google Scholar] [CrossRef]
  72. Tulgar, S.; Selvi, O.; Serifsoy, T.E.; Senturk, O.; Ozer, Z. [YouTube as an Information Source of Spinal Anesthesia, Epidural Anesthesia and Combined Spinal and Epidural Anesthesia]. Rev. Bras. Anestesiol. 2017, 67, 493–499. [Google Scholar] [CrossRef] [PubMed]
  73. Ashokka, B.; Chakraborty, A.; Subramanian, B.J.; Karmakar, M.K.; Chan, V. Reconfiguring the Scope and Practice of Regional Anesthesia in a Pandemic: The COVID-19 Perspective. Reg. Anesth. Pain Med. 2020, 45, 536–543. [Google Scholar] [CrossRef] [PubMed]
  74. Lie, S.A.; Wong, S.W.; Wong, L.T.; Wong, T.G.L.; Chong, S.Y. Practical Considerations for Performing Regional Anesthesia: Lessons Learned from the Covid-19 Pandemic. Can. J. Anaesth. 2020, 67, 885–892. [Google Scholar] [CrossRef] [Green Version]
  75. Ríos Medina, A.M.; Caicedo Salazar, J. COVID-19 and Education in Regional Anesthesia. Reg. Anesth. Pain Med. 2021, 46, 550. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flowchart of the video search and screening process: video metrics, reliability, overall quality, and specific quality of videos by source.
Figure 1. Flowchart of the video search and screening process: video metrics, reliability, overall quality, and specific quality of videos by source.
Healthcare 09 01083 g001
Figure 2. Heatmap of brachial plexus block specific quality scores according to video sources. The X-axis represents each question and Y-axis represents the video source. The number in the box represents the percentage of questions addressed. The color is intensified as the percentage increases.
Figure 2. Heatmap of brachial plexus block specific quality scores according to video sources. The X-axis represents each question and Y-axis represents the video source. The number in the box represents the percentage of questions addressed. The color is intensified as the percentage increases.
Healthcare 09 01083 g002
Figure 3. Changes of BSQS, GQS, and JAMAS by year according to sources. The smoothed loess regression represents the changes of scores from 2005 to 2019.
Figure 3. Changes of BSQS, GQS, and JAMAS by year according to sources. The smoothed loess regression represents the changes of scores from 2005 to 2019.
Healthcare 09 01083 g003
Table 1. Brachial plexus block specific quality score.
Table 1. Brachial plexus block specific quality score.
Brachial Plexus Block Specific Quality ScoreScore
Pre-Procedure Scores
Q1. Mentioned which procedure1
Q2. Mentioned targeted dermatome or indication2
Q3. Patient position1
Equipment preparation
Q4. High-frequency linear probe1
Q5. Disinfectant solution1
Q6. Local anesthetics (which and how much)1
Q7. Needle gauze1
Q8. Sterile gel and sterile probe cover1
Intra-Procedure Scores
Q9. Probe placement1
Q10. Anatomical landmarks2
Q11. Important vessels and structures1
Q12. Needling technique; in-plane or out-of-plane1
Q13. Needle tip confirmation (negative aspiration or small amount injection or nerve stimulation use)2
Q14. Spread of local anesthetics1
Post-Procedure Scores
Q15. Dermatome check or nerve stimulation use2
Q16. Complications2
Total21
Table 2. Video metrics, reliability, overall quality, and specific quality of the videos based on the source.
Table 2. Video metrics, reliability, overall quality, and specific quality of the videos based on the source.
IndividualAcademicManufacturerEducationalp-ValuePost Hoc. Tukey’s Test
Videos, n (%)88 (50.29)36 (20.57)18 (10.29)33 (18.86)
Years, median
(min, max)
2015 (2008, 2019)2015 (2008, 2019)2014 (2010, 2017)2015 (2009, 2019)
Video metrics, mean ± SD (median)
Views3472.78 ± 10,333.14 (458)21,120.58 ± 47,242.49 (509)47,873.61 ± 87,483.11 (9675)41,855.12 ± 56,751.61 (19,418)<0.001Manufacturer > Individual, Academic; Education > Academic
Likes10.95 ± 33.80 (2)31.11 ± 69.70 (2)119 ± 205.87 (18)138.70 ± 183.52 (45)<0.001Educational > Individual, Academic
Dislikes0.74 ± 2.35 (0)3.44 ± 6.81 (0)7.5 ± 13.03 (2)7.21 ± 9.05 (4)<0.001Manufacturer, Educational > Individual
Comments0.59 ± 1.64 (0)1.44 ± 2.67 (0)2.11 ± 4.78 (0)3.03 ± 4.69 (1)<0.001Educational > Individual
Length (seconds)247.65 ± 212.64 (193)238.56 ± 237.33 (166)195.89 ± 72.47 (162)285.94 ± 213.82 (234)0.32
VPI1.59 ± 3.35 (0.38)7.80 ± 14.32 (0.82)20.81 ± 25.28 (13.19)21.35 ± 24.75 (10.49)<0.001Manufacturer, Educational > Individual
Reliability, mean ± SD (median)
JAMAS0.26 ± 0.67 (0)1.19 ± 0.62 (1)1.17 ± 0.71 (1)1.15 ± 0.76 (1)<0.001Academic, Manufacturer, Educational > Individual
Overall quality, mean ± SD (median)
GQS1.58 ± 0.71 (1)1.69 ± 0.89 (1)1.83 ± 0.38 (2)1.85 ± 0.82 (2)0.1087
Specific quality, mean ± SD (median)
BSQS7.32 ± 4.20 (7)8.72 ± 3.84 (8.5)11.22 ± 1.63 (11)10.33 ± 3.34 (10)<0.001Manufacturer, Educational > Individual
BSQS—Brachial Plexus Block Specific Quality Score; GQS—Global Quality Score; JAMAS—Journal of American Medical Association; SD—standard deviation; VPI—Video Power Index.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cho, N.R.; Cha, J.H.; Park, J.J.; Kim, Y.H.; Ko, D.S. Reliability and Quality of YouTube Videos on Ultrasound-Guided Brachial Plexus Block: A Programmatical Review. Healthcare 2021, 9, 1083. https://doi.org/10.3390/healthcare9081083

AMA Style

Cho NR, Cha JH, Park JJ, Kim YH, Ko DS. Reliability and Quality of YouTube Videos on Ultrasound-Guided Brachial Plexus Block: A Programmatical Review. Healthcare. 2021; 9(8):1083. https://doi.org/10.3390/healthcare9081083

Chicago/Turabian Style

Cho, Noo Ree, Jeong Ho Cha, Jeong Jun Park, Yun Hak Kim, and Dai Sik Ko. 2021. "Reliability and Quality of YouTube Videos on Ultrasound-Guided Brachial Plexus Block: A Programmatical Review" Healthcare 9, no. 8: 1083. https://doi.org/10.3390/healthcare9081083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop