Skip to Main Content (Press Enter)

Logo UNIMORE
  • ×
  • Home
  • Corsi
  • Insegnamenti
  • Professioni
  • Persone
  • Pubblicazioni
  • Strutture
  • Terza Missione
  • Attività
  • Competenze

UNI-FIND
Logo UNIMORE

|

UNI-FIND

unimore.it
  • ×
  • Home
  • Corsi
  • Insegnamenti
  • Professioni
  • Persone
  • Pubblicazioni
  • Strutture
  • Terza Missione
  • Attività
  • Competenze
  1. Pubblicazioni

Improving Augmented Reality Through Deep Learning: Real-time Instrument Delineation in Robotic Renal Surgery

Articolo
Data di Pubblicazione:
2023
Citazione:
Improving Augmented Reality Through Deep Learning: Real-time Instrument Delineation in Robotic Renal Surgery / De Backer, P.; Van Praet, C.; Simoens, J.; Peraire Lores, M.; Creemers, H.; Mestdagh, K.; Allaeys, C.; Vermijs, S.; Piazza, P.; Mottaran, A.; Bravi, C. A.; Paciotti, M.; Sarchi, L.; Farinha, R.; Puliatti, S.; Cisternino, F.; Ferraguti, F.; Debbaut, C.; De Naeyer, G.; Decaestecker, K.; Mottrie, A.. - In: EUROPEAN UROLOGY. - ISSN 0302-2838. - 84:1(2023), pp. 86-91. [10.1016/j.eururo.2023.02.024]
Abstract:
Several barriers prevent the integration and adoption of augmented reality (AR) in robotic renal surgery despite the increased availability of virtual three-dimensional (3D) models. Apart from correct model alignment and deformation, not all instruments are clearly visible in AR. Superimposition of a 3D model on top of the surgical stream, including the instruments, can result in a potentially hazardous surgical situation. We demonstrate real-time instrument detection during AR-guided robot-assisted partial nephrectomy and show the generalization of our algorithm to AR-guided robot-assisted kidney transplantation. We developed an algorithm using deep learning networks to detect all nonorganic items. This algorithm learned to extract this information for 65 927 manually labeled instruments on 15 100 frames. Our setup, which runs on a standalone laptop, was deployed in three different hospitals and used by four different surgeons. Instrument detection is a simple and feasible way to enhance the safety of AR-guided surgery. Future investigations should strive to optimize efficient video processing to minimize the 0.5-s delay currently experienced. General AR applications also need further optimization, including detection and tracking of organ deformation, for full clinical implementation.
Tipologia CRIS:
Articolo su rivista
Keywords:
Augmented reality; Deep learning; Instrument segmentation; Kidney transplantation; Partial nephrectomy; Real time; Renal cell carcinoma; Robotic surgery; Three-dimensional models
Elenco autori:
De Backer, P.; Van Praet, C.; Simoens, J.; Peraire Lores, M.; Creemers, H.; Mestdagh, K.; Allaeys, C.; Vermijs, S.; Piazza, P.; Mottaran, A.; Bravi, C. A.; Paciotti, M.; Sarchi, L.; Farinha, R.; Puliatti, S.; Cisternino, F.; Ferraguti, F.; Debbaut, C.; De Naeyer, G.; Decaestecker, K.; Mottrie, A.
Autori di Ateneo:
FERRAGUTI Federica
Puliatti Stefano
Link alla scheda completa:
https://iris.unimore.it/handle/11380/1360486
Link al Full Text:
https://iris.unimore.it//retrieve/handle/11380/1360486/705544/1-s2.0-S0302283823026337-main.pdf
Pubblicato in:
EUROPEAN UROLOGY
Journal
  • Utilizzo dei cookie

Realizzato con VIVO | Designed by Cineca | 26.5.0.0