Proposing novel methods for gynecologic surgical action recognition on laparoscopic videos
Autor: | Parastoo Dezyani, Toktam Khatibi |
---|---|
Rok vydání: | 2020 |
Předmět: |
Endoscope
medicine.diagnostic_test Computer Networks and Communications business.industry Computer science Surgical action Feature extraction ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION 020207 software engineering 02 engineering and technology Machine learning computer.software_genre Task (project management) Hardware and Architecture 0202 electrical engineering electronic engineering information engineering Media Technology medicine Gynecologic laparoscopy Feature (machine learning) Artificial intelligence business Laparoscopy computer Software |
Zdroj: | Multimedia Tools and Applications. 79:30111-30133 |
ISSN: | 1573-7721 1380-7501 |
DOI: | 10.1007/s11042-020-09540-y |
Popis: | Laparoscopy or minimally-invasive surgery (MIS) is performed by inserting a camera called endoscope inside the body to display the surgical actions online with the ability to record and archive the video. Recognizing the surgical actions automatically from the laparoscopic videos have many applications such as surgical skill assessment, teaching purposes, and workflow recognition but is a challenging task. The main aim of this study is proposing novel automatic methods for surgical action recognition from the laparoscopic video frames. For this purpose, three different scenarios are designed, evaluated and compared using 5-fold cross validation strategy. The first and the second scenarios are based on deep neural networks and combination of pre-trained CNNs and conventional machine learning models, respectively. The last scenario combines handcraft feature extraction, pre-trained CNNs, feature engineering based on complex networks and conventional classifiers. Dataset analyzed in this study is ITEC LapGyn4 Gynecologic Laparoscopy Image dataset. Experimental results show that the second and the third scenarios have highly desirable performance for multi-instance surgical action recognition with the average accuracy of 99.20 and AUC of 99.12. On the other hand, for single-instance surgical action recognition, the third scenario outperforms the compared ones with the average accuracy of 99.05 and AUC of 96.41. Moreover, different feature sets in the third scenario are ranked and assigned the importance score based on “Mean Decrease of Accuracy” measure. The first-ranked features are the deep features extracted from our proposed CNNs in the first scenario and the second-ranked ones are the features engineered from the complex networks. |
Databáze: | OpenAIRE |
Externí odkaz: |