Surgical skill levels: Classification and analysis using deep neural network model and motion signals

Research output: Contribution to journalArticleResearchpeer-review

2 Citations (Scopus)

Abstract

Background and Objectives: Currently, the assessment of surgical skills relies primarily on the observations of expert surgeons. This may be time-consuming, non-scalable, inconsistent and subjective. Therefore, an automated system that can objectively identify the actual skills level of a junior trainee is highly desirable. This study aims to design an automated surgical skills evaluation system. Methods: We propose to use a deep neural network model that can analyze raw surgical motion data with minimal preprocessing. A platform with inertial measurement unit sensors was developed and participants with different levels of surgical experience were recruited to perform core open surgical skills tasks. JIGSAWS a publicly available robot based surgical training dataset was used to evaluate the generalization of our deep network model. 15 participants (4 experts, 4 intermediates and 7 novices)were recruited into the study. Results: The proposed deep model achieved an accuracy of 98.2%. With comparison to JIGSAWS; our method outperformed some existing approaches with an accuracy of 98.4%, 98.4% and 94.7% for suturing, needle-passing, and knot-tying, respectively. The experimental results demonstrated the applicability of this method in both open surgery and robot-assisted minimally invasive surgery. Conclusions: This study demonstrated the potential ability of the proposed deep network model to learn the discriminative features between different surgical skills levels.

Original languageEnglish
Pages (from-to)1-8
Number of pages8
JournalComputer Methods and Programs in Biomedicine
Volume177
DOIs
Publication statusPublished - 1 Aug 2019

Keywords

  • Deep neural network
  • Hand motion signals
  • Surgical education
  • Surgical skill assessment

Cite this