Scoring metrics for assessing skills in arthroscopic rotator cuff repair: performance comparison study of novice and expert surgeons


Demirel D., Palmer B., Sundberg G., Karaman B., Halic T., Kockara S., ...More

International Journal of Computer Assisted Radiology and Surgery, vol.17, no.10, pp.1823-1835, 2022 (SCI-Expanded) identifier identifier identifier

  • Publication Type: Article / Article
  • Volume: 17 Issue: 10
  • Publication Date: 2022
  • Doi Number: 10.1007/s11548-022-02683-3
  • Journal Name: International Journal of Computer Assisted Radiology and Surgery
  • Journal Indexes: Science Citation Index Expanded (SCI-EXPANDED), Scopus, EMBASE, INSPEC, MEDLINE
  • Page Numbers: pp.1823-1835
  • Keywords: Arthroscopic surgery performance metrics, Deep learning, Surgeon skills, Rotator cuff, Arthroscopy
  • Erzincan Binali Yildirim University Affiliated: Yes

Abstract

© 2022, CARS.Purpose: We aim to develop quantitative performance metrics and a deep learning model to objectively assess surgery skills between the novice and the expert surgeons for arthroscopic rotator cuff surgery. These proposed metrics can be used to give the surgeon an objective and a quantitative self-assessment platform. Methods: Ten shoulder arthroscopic rotator cuff surgeries were performed by two novices, and fourteen were performed by two expert surgeons. These surgeries were statistically analyzed. Two existing evaluation systems: Basic Arthroscopic Knee Skill Scoring System (BAKSSS) and the Arthroscopic Surgical Skill Evaluation Tool (ASSET), were used to validate our proposed metrics. In addition, a deep learning-based model called Automated Arthroscopic Video Evaluation Tool (AAVET) was developed toward automating quantitative assessments. Results: The results revealed that novice surgeons used surgical tools approximately 10% less effectively and identified and stopped bleeding less swiftly. Our results showed a notable difference in the performance score between the experts and novices, and our metrics successfully identified these at the task level. Moreover, the F1-scores of each class are found as 78%, 87%, and 77% for classifying cases with no-tool, electrocautery, and shaver tool, respectively. Conclusion: We have constructed quantitative metrics that identified differences in the performances of expert and novice surgeons. Our ultimate goal is to validate metrics further and incorporate these into our virtual rotator cuff surgery simulator (ViRCAST), which has been under development. The initial results from AAVET show that the capability of the toolbox can be extended to create a fully automated performance evaluation platform.