Inpainting surgical occlusion from laparoscopic video sequences for robot-assisted interventions.

computer vision deep learning inpainting optical-flow poisson blending segmentation

Journal

Journal of medical imaging (Bellingham, Wash.)
ISSN: 2329-4302
Titre abrégé: J Med Imaging (Bellingham)
Pays: United States
ID NLM: 101643461

Informations de publication

Date de publication:
Jul 2023
Historique:
received: 03 04 2023
revised: 21 07 2023
accepted: 28 07 2023
pmc-release: 29 08 2024
medline: 31 8 2023
pubmed: 31 8 2023
entrez: 31 8 2023
Statut: ppublish

Résumé

Medical technology for minimally invasive surgery has undergone a paradigm shift with the introduction of robot-assisted surgery. However, it is very difficult to track the position of the surgical tools in a surgical scene, so it is crucial to accurately detect and identify surgical tools. This task can be aided by deep learning-based semantic segmentation of surgical video frames. Furthermore, due to the limited working and viewing areas of these surgical instruments, there is a higher chance of complications from tissue injuries (e.g., tissue scars and tears). With the aid of digital inpainting algorithms, we present an application that uses image segmentation to remove surgical instruments from laparoscopic/endoscopic video. We employ a modified U-Net architecture (U-NetPlus) to segment the surgical instruments. It consists of a redesigned decoder and a pre-trained VGG11 or VGG16 encoder. The decoder was modified by substituting an up-sampling operation based on nearest-neighbor interpolation for the transposed convolution operation. Furthermore, these interpolation weights do not need to be learned to perform upsampling, which eliminates the artifacts generated by the transposed convolution. In addition, we use a very fast and adaptable data augmentation technique to further enhance performance. The instrument segmentation mask is filled in (i.e., inpainted) by the tool removal algorithms using the previously acquired tool segmentation masks and either previous instrument-containing frames or instrument-free reference frames. We have shown the effectiveness of the proposed surgical tool segmentation/removal algorithms on a robotic instrument dataset from the MICCAI 2015 and 2017 EndoVis Challenge. We report a 90.20% DICE for binary segmentation, a 76.26% DICE for instrument part segmentation, and a 46.07% DICE for instrument type (i.e., all instruments) segmentation on the MICCAI 2017 challenge dataset using our U-NetPlus architecture, outperforming the results of earlier techniques used and tested on these data. In addition, we demonstrated the successful execution of the tool removal algorithm from surgical tool-free videos that contained moving surgical tools that were generated artificially. Our application successfully separates and eliminates the surgical tool to reveal a view of the background tissue that was otherwise hidden by the tool, producing results that are visually similar to the actual data.

Identifiants

pubmed: 37649957
doi: 10.1117/1.JMI.10.4.045002
pii: 23078GR
pmc: PMC10462486
doi:

Types de publication

Journal Article

Langues

eng

Pagination

045002

Subventions

Organisme : NIGMS NIH HHS
ID : R35 GM128877
Pays : United States

Informations de copyright

© 2023 Society of Photo-Optical Instrumentation Engineers (SPIE).

Références

Annu Rev Biomed Eng. 2017 Jun 21;19:221-248
pubmed: 28301734
IEEE Trans Pattern Anal Mach Intell. 2017 Apr;39(4):640-651
pubmed: 27244717
Int J Biomed Imaging. 2016;2016:7952541
pubmed: 26904103
Phys Rep. 2014 Oct 10;543(2):61-162
pubmed: 25267872
Annu Int Conf IEEE Eng Med Biol Soc. 2019 Jul;2019:7205-7211
pubmed: 31947497
Int J Comput Assist Radiol Surg. 2016 Oct;11(10):1927-36
pubmed: 27038964
World J Urol. 2018 Apr;36(4):537-541
pubmed: 29427003
J Laparoendosc Adv Surg Tech A. 1997 Dec;7(6):369-73
pubmed: 9449087
Proc SPIE Int Soc Opt Eng. 2021 Feb;11598:
pubmed: 34079156
Neurospine. 2019 Mar;16(1):6-14
pubmed: 31618807

Auteurs

S M Kamrul Hasan (SMK)

Rochester Institute of Technology, Biomedical Modeling, Visualization, and Image-guided Navigation (BiMVisIGN) Lab, Rochester, New York, United States.
Rochester Institute of Technology, Center for Imaging Science, Rochester, New York, United States.

Richard A Simon (RA)

Rochester Institute of Technology, Biomedical Modeling, Visualization, and Image-guided Navigation (BiMVisIGN) Lab, Rochester, New York, United States.
Rochester Institute of Technology, Biomedical Engineering, Rochester, New York, United States.

Cristian A Linte (CA)

Rochester Institute of Technology, Biomedical Modeling, Visualization, and Image-guided Navigation (BiMVisIGN) Lab, Rochester, New York, United States.
Rochester Institute of Technology, Center for Imaging Science, Rochester, New York, United States.
Rochester Institute of Technology, Biomedical Engineering, Rochester, New York, United States.

Classifications MeSH