A deep learning framework for quality assessment and restoration in video endoscopy

Published in Medical Image Analysis, 2020

Recommended citation: Sharib Ali, Felix Zhou, Adam Bailey, Barbara Braden, James E. East, Xin Lu, Jens Rittscher. (2021). "A deep learning framework for quality assessment and restoration in video endoscopy." Medical Image Analysis Vol 68, pg. 101900. https://doi.org/10.1016/j.media.2020.101900

Abstract: Endoscopy is a routine imaging technique used for both diagnosis and minimally invasive surgical treatment. Artifacts such as motion blur, bubbles, specular reflections, floating objects and pixel saturation impede the visual interpretation and the automated analysis of endoscopy videos. Given the widespread use of endoscopy in different clinical applications, robust and reliable identification of such artifacts and the automated restoration of corrupted video frames is a fundamental medical imaging problem. Existing state-of-the-art methods only deal with the detection and restoration of selected artifacts. However, typically endoscopy videos contain numerous artifacts which motivates to establish a comprehensive solution.

In this paper, a fully automatic framework is proposed that can: 1) detect and classify six different artifacts, 2) segment artifact instances that have indefinable shapes, 3) provide a quality score for each frame, and 4) restore partially corrupted frames. To detect and classify different artifacts, the proposed framework exploits fast, multi-scale and single stage convolution neural network detector. In addition, we use an encoder-decoder model for pixel-wise segmentation of irregular shaped artifacts. A quality score is introduced to assess video frame quality and to predict image restoration success. Generative adversarial networks with carefully chosen regularization and training strategies for discriminator-generator networks are finally used to restore corrupted frames.

The detector yields the highest mean average precision (mAP) of 45.7 and 34.7, respectively for 25% and 50% IoU thresholds, and the lowest computational time of 88 ms allowing for near real-time processing. The restoration models for blind deblurring, saturation correction and inpainting demonstrate significant improvements over previous methods. On a set of 10 test videos, an average of 68.7% of video frames successfully passed the quality score (>=0.9) after applying the proposed restoration framework thereby retaining 25% more frames compared to the raw videos. The importance of artifacts detection and their restoration on improved robustness of image analysis methods is also demonstrated in this work.

[Paper]

Bibtex
@article{Ali2021ADL, title={A deep learning framework for quality assessment and restoration in video endoscopy}, author={Sharib Ali and Felix Y Zhou and Adam Bailey and Barbara Braden and James Edward East and Xin Lu and J. Rittscher}, journal={Medical image analysis}, year={2021}, volume={68}, pages={ 101900}}