Scoring metrics for assessing skills in arthroscopic rotator cuff repair: performance comparison study of novice and expert surgeons


Demirel D., Palmer B., Sundberg G., Karaman B., Halic T., Kockara S., ...Daha Fazla

International Journal of Computer Assisted Radiology and Surgery, cilt.17, sa.10, ss.1823-1835, 2022 (SCI-Expanded) identifier identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 17 Sayı: 10
  • Basım Tarihi: 2022
  • Doi Numarası: 10.1007/s11548-022-02683-3
  • Dergi Adı: International Journal of Computer Assisted Radiology and Surgery
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Scopus, EMBASE, INSPEC, MEDLINE
  • Sayfa Sayıları: ss.1823-1835
  • Anahtar Kelimeler: Arthroscopic surgery performance metrics, Deep learning, Surgeon skills, Rotator cuff, Arthroscopy
  • Erzincan Binali Yıldırım Üniversitesi Adresli: Evet

Özet

© 2022, CARS.Purpose: We aim to develop quantitative performance metrics and a deep learning model to objectively assess surgery skills between the novice and the expert surgeons for arthroscopic rotator cuff surgery. These proposed metrics can be used to give the surgeon an objective and a quantitative self-assessment platform. Methods: Ten shoulder arthroscopic rotator cuff surgeries were performed by two novices, and fourteen were performed by two expert surgeons. These surgeries were statistically analyzed. Two existing evaluation systems: Basic Arthroscopic Knee Skill Scoring System (BAKSSS) and the Arthroscopic Surgical Skill Evaluation Tool (ASSET), were used to validate our proposed metrics. In addition, a deep learning-based model called Automated Arthroscopic Video Evaluation Tool (AAVET) was developed toward automating quantitative assessments. Results: The results revealed that novice surgeons used surgical tools approximately 10% less effectively and identified and stopped bleeding less swiftly. Our results showed a notable difference in the performance score between the experts and novices, and our metrics successfully identified these at the task level. Moreover, the F1-scores of each class are found as 78%, 87%, and 77% for classifying cases with no-tool, electrocautery, and shaver tool, respectively. Conclusion: We have constructed quantitative metrics that identified differences in the performances of expert and novice surgeons. Our ultimate goal is to validate metrics further and incorporate these into our virtual rotator cuff surgery simulator (ViRCAST), which has been under development. The initial results from AAVET show that the capability of the toolbox can be extended to create a fully automated performance evaluation platform.