دورية أكاديمية

Evaluating robotic-assisted partial nephrectomy surgeons with fully convolutional segmentation and multi-task attention networks.

التفاصيل البيبلوغرافية
العنوان: Evaluating robotic-assisted partial nephrectomy surgeons with fully convolutional segmentation and multi-task attention networks.
المؤلفون: Wang Y; Department of Computer Science, Southern Methodist University, Dallas, USA., Wu Z; Department of Computer Science, Southern Methodist University, Dallas, USA., Dai J; Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA., Morgan TN; Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA., Garbens A; Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA., Kominsky H; Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA., Gahan J; Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA., Larson EC; Department of Computer Science, Southern Methodist University, Dallas, USA. eclarson@smu.edu.
المصدر: Journal of robotic surgery [J Robot Surg] 2023 Oct; Vol. 17 (5), pp. 2323-2330. Date of Electronic Publication: 2023 Jun 27.
نوع المنشور: Journal Article
اللغة: English
بيانات الدورية: Publisher: Springer Country of Publication: England NLM ID: 101300401 Publication Model: Print-Electronic Cited Medium: Internet ISSN: 1863-2491 (Electronic) Linking ISSN: 18632483 NLM ISO Abbreviation: J Robot Surg Subsets: MEDLINE
أسماء مطبوعة: Original Publication: London : Springer
مواضيع طبية MeSH: Robotic Surgical Procedures*/methods , Laparoscopy* , Surgeons*, Humans ; Nephrectomy/education
مستخلص: We use machine learning to evaluate surgical skill from videos during the tumor resection and renography steps of a robotic assisted partial nephrectomy (RAPN). This expands previous work using synthetic tissue to include actual surgeries. We investigate cascaded neural networks for predicting surgical proficiency scores (OSATS and GEARS) from RAPN videos recorded from the DaVinci system. The semantic segmentation task generates a mask and tracks the various surgical instruments. The movements from the instruments found via semantic segmentation are processed by a scoring network that regresses (predicts) GEARS and OSATS scoring for each subcategory. Overall, the model performs well for many subcategories such as force sensitivity and knowledge of instruments of GEARS and OSATS scoring, but can suffer from false positives and negatives that would not be expected of human raters. This is mainly attributed to limited training data variability and sparsity.
(© 2023. The Author(s).)
References: Brian D (2015) Robotic surgery-a personal view of the past, present and future. Int J Adv Robot Syst 12(5):54. (PMID: 10.5772/60118)
Dhanesh R, Taylor Graham W (2017) Deep multimodal learning: a survey on recent advances and trends’. IEEE Signal Process Mag 34(6):96–108. (PMID: 10.1109/MSP.2017.2738401)
Chen Y (2015) Convolutional neural network for sentence classification. Master’s thesis. University of Waterloo.
Martin JA, Glenn R, Richard R, Helen M, John M, Carol H, Brown M (1997) Objective structured assessment of technical skill (OSATS) for surgical residents. Br J Surg 84(2):273–278. (PMID: 9052454)
Johnson Brett A, Matthew T, Steinberg Ryan L, Matthew K, Bradly M, Gahan Jeffrey C (2019) Design and validation of a lowcost, high-fidelity model for urethrovesical anastomosis in radical prostatectomy. J Endourol 33(4):331–336. (PMID: 10.1089/end.2018.087130734578)
Yihao W, Jessica D, Morgan Tara N, Mohamed E, Garbens Alaina Q, Xingming SR, Jeffrey G, Larson Eric C (2022) Evaluating robotic-assisted surgery training videos with multi-task convolutional neural networks. J Robot Surg 16(4):917–925. https://doi.org/10.1007/s11701-021-01316-2. (PMID: 10.1007/s11701-021-01316-2)
Ashish V, Noam S, Niki P, Jakob U, Llion J, Gomez Aidan N, Łukasz K, Illia P (2017) Attention is all you need. Advances in neural information processing systems. Springer, p 30.
Sammit G, Wu Z, Wang Y, Wu Z, Kamata A, Nese J, Larson EC (2022) Automated prosody classification for oral reading fluency with quadratic kappa loss and attentive X-vectors. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 3613–3617.
Zijian Z, Sandrine V, Ying W, Faliang C, Ruijian L (2017) Tracking-by-detection of surgical instruments in minimally invasive surgery via the convolutional neural network deep learning-based method. Comput Assist Surg 22(1):26–35.
Hei L, Khurshid G, Jia D (2017) Surgeon technical skill assessment using computer vision based analysis. Machine learning for healthcare conference. PMLR, pp 88–99.
Lee D, Won YH, Kwon H, Hyoun-Joong K, Eun LK, Chan KH (2020) Evaluation of surgical skills during robotic surgery by deep learning-based multiple surgical instrument tracking in training and actual operations. J Clin Med 9(6):1964. (PMID: 10.3390/jcm9061964325859537355689)
Jeffrey G, Ryan S, Garbens Alaina Q, Xingming LE (2020) MP34-06 machine learning using a multi-task convolutional neural networks can accurately assess robotic skills. J Urol 203(4):e505–e505.
Garcia-Garcia A, Orts-Escolano S, Oprea S, Villena-Martinez V, Garcia-Rodriguez J (2017) A review on deep learning techniques applied to semantic segmentation. ArXiv preprint arXiv:1704.06857.
Wang P, Chen P, Yuan Y, Liu D, Huang Z, Hou X, Cottrell G (2018) Understanding convolution for semantic segmentation. In: IEEE Winter Conference on Applications of Computer Vision (WACV), IEEE, pp 1451–1460.
Giorgia M, Leonardo T, Luca U, Francesco P, Enrico V (2023) A multi-task convolutional neural network for semantic segmentation and event detection in laparoscopic surgery. J Personal Med 13(3):413. (PMID: 10.3390/jpm13030413)
Dhanesh R, Taylor Graham W (2017) Deep multimodal learning: a survey on recent advances and trends. IEEE Signal Process Mag 34(6):96–108. (PMID: 10.1109/MSP.2017.2738401)
de La Torre J, Puig D, Valls A (2018) Weighted kappa loss function for multi-class classification of ordinal data in deep learning. Pattern Recognit Lett 105:144–154. (PMID: 10.1016/j.patrec.2017.05.018)
Rui F, Alberto B, James P, Alexandre M, Ben VC, Vander SJ, Angelo M, Gallagher Anthony G et al (2023) International expert consensus on metric-based characterization of robot-assisted partial nephrectomy. Eur Urol Focus 9(2):388–395. (PMID: 10.1016/j.euf.2022.09.017)
Ronneberger O, Fischer P, Brox T (2015) U-Net: convolutional networks for biomedical image segmentation. In: CoRR abs/1505.04597 . arXiv: 1505.04597.
Shen Z, Bello I, Vemulapalli R, Jia X, Chen C-H (2020) Global self-attention networks for image recognition. In: arXiv preprint arXiv:2010.03019.
Chen C-FR, Fan Q, Panda R (2021) Crossvit: cross attention multi-scale vision transformer for image classification. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp 357–366.
Snyder D, Garcia-Romero D, Sell G, Povey D, Khudanpur S (2018) X-vectors: robust DNN embeddings for speaker recognition. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, pp 5329–5333.
Okabe K, Koshinaka T, Shinoda K (2018) Attentive statistics pooling for deep speaker embedding. In: arXiv preprint arXiv:1803.10963.
Connor S, Khoshgoftaar Taghi M (2019) A survey on image data augmentation for deep learning. J Big Data 6(1):1–48.
He K, Gkioxari G, Dollár P, Girshick R (2017) Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp 2961–2969.
Bradley E, Robert T (1997) Improvements on cross-validation: the 632+ bootstrap method. J Am Stat Assoc 92(438):548–560.
فهرسة مساهمة: Keywords: Convolutional network; Multi-task learning; Self-attention; Surgical assessment
تواريخ الأحداث: Date Created: 20230627 Date Completed: 20230911 Latest Revision: 20230922
رمز التحديث: 20230922
مُعرف محوري في PubMed: PMC10492672
DOI: 10.1007/s11701-023-01657-0
PMID: 37368225
قاعدة البيانات: MEDLINE
الوصف
تدمد:1863-2491
DOI:10.1007/s11701-023-01657-0