Open Access

Automatic Segmentation and Inpainting of Specular Highlights for Endoscopic Imaging

EURASIP Journal on Image and Video Processing20102010:814319

DOI: 10.1155/2010/814319

Received: 30 April 2010

Accepted: 2 December 2010

Published: 13 December 2010

Abstract

Minimally invasive medical procedures have become increasingly common in today's healthcare practice. Images taken during such procedures largely show tissues of human organs, such as the mucosa of the gastrointestinal tract. These surfaces usually have a glossy appearance showing specular highlights. For many visual analysis algorithms, these distinct and bright visual features can become a significant source of error. In this article, we propose two methods to address this problem: (a) a segmentation method based on nonlinear filtering and colour image thresholding and (b) an efficient inpainting method. The inpainting algorithm eliminates the negative effect of specular highlights on other image analysis algorithms and also gives a visually pleasing result. The methods compare favourably to the existing approaches reported for endoscopic imaging. Furthermore, in contrast to the existing approaches, the proposed segmentation method is applicable to the widely used sequential RGB image acquisition systems.

1. Introduction

Due to reduced patient recovery time and mortality rate, minimally invasive medical procedures have become increasingly common in today's healthcare practice. Consequently, technological research related to this class of medical procedures is becoming more widespread. Since many minimally invasive procedures are guided through optical imaging systems, it is a commonly investigated question, what kind of sensible information may be automatically extracted from these image data and how this information may be used to improve guidance systems or procedure analysis and documentation. Research topics in this context are, among others, robot-assisted guidance and surgery [17], automated documentation [810] or registration of the optically acquired images or videos to image data obtained from preprocedure X-ray, computed tomography (CT), magnetic resonance imaging (MRI) and other medical image acquisition techniques [1115].

A key technological advancement that has contributed to the success of minimally invasive procedures is video endoscopy. Endoscopy is the most commonly used method for image-guided minimally invasive procedures, for example, colonoscopy, bronchoscopy, laparoscopy, rhinoscopy. An endoscope is a flexible tube fitted with a camera and an illumination unit at the tip. Depending on the type of procedure the tube is inserted into the human body through either a natural orifice or a small incision. During the procedure, the performing physician can observe the endoscopic video data in real-time on a monitor.

Images and videos from minimally invasive medical procedures largely show tissues of human organs, such as the mucosa of the gastrointestinal tract. These surfaces usually have a glossy appearance showing specular highlights due to specular reflection of the light sources. Figure 1 shows example images extracted from different domains with typical specular highlights. These image features can negatively affect the perceived image quality [16]. Furthermore, for many visual analysis algorithms, these distinct and bright visual features can become a significant source of error. Since the largest image gradients can usually be found at the edges of specular highlights, they may interfere with all gradient-based computer vision and image analysis algorithms. Similarly, they may also affect texture based approaches. On the contrary, specular highlights hold important information about the surface orientation, if the relative locations of the camera and the illumination unit are known. Detecting specular highlights may therefore improve the performance of 3D reconstruction algorithms.
Figure 1

Examples of images from minimally invasive medical procedures showing specular highlights. (a) Laparoscope image of the appendix, (b) Colonoscopic image with specularity and colour channel misalignment due to sequential RGB endoscopic system, (c) Colonoscopic image showing a colonic polyp.

Our area of research is the analysis of endoscopic video data, in particular from colonoscopy procedures. Colonoscopy is a video endoscopy of the large intestine and the currently preferred method for colorectal cancer screening. Common topics in colonoscopic imaging research are, among others, the detection of polyps and colorectal cancer [1720], temporal segmentation and summarisation of colonoscopy procedures [2123], image classification [2426], image quality enhancement [27] and automated procedure quality assessment [28, 29].

Segmentation of specular highlights may be beneficial in many of these topics. An example is the automatic detection of colorectal polyps. Colorectal polyps can develop into cancer if they are not detected and removed. Figure 1(c) shows an example of a typical colonic polyp. Texture is one of the important characteristics that are used in their detection. The specular highlights on the polyp can affect texture features obtained from the polyp surface and may therefore impede robust detection. A negative effect of specular highlights was also reported by Oh et al. [26], in the context of the detection of indistinct frames in colonoscopic videos. The term indistinct refers to blurry images that occur when the camera is too close to the intestinal mucosa or is covered by liquids.

In this paper, we propose: (a) a method for segmentation of specular highlights based on nonlinear filtering and colour image thresholding and (b) an efficient inpainting method that alters the specular regions in a way that eliminates the negative effect on most algorithms and also gives a visually pleasing result. We also present an application of these methods in improvement of colour channel misalignment artefacts removal.

For many applications, the segmentation will be sufficient, since the determined specular areas can simply be omitted in further computations. For others, it might be necessary or more efficient to inpaint the highlights. For example the colour misalignment artefacts as shown in Figure 1(b) is a major hindrance in many processing algorithms, for example, automated polyp detection. In order to remove these artefacts the endoscope camera motion needs to be estimated. Feature point detection and matching are two pivotal steps in most camera motion estimation algorithm. Due to the invariance of positions in different colour channels of the images similar to the one shown in Figure 1(b), the specular highlights creates a major problem for any feature matching algorithm and consequently for the camera motion estimation algorithm.

The paper is organised as follows. Section 2 takes a look at related work in segmentation of specular highlights, before the proposed approach is explained in detail in Section 3. The evaluation of the segmentation method is presented in Section 4. The proposed inpainting approach is described in Section 5 along with a brief look at the literature on the topic. In Section 6 we show how removal of specular highlights facilitates better performance of other processing algorithms with the example of colour channel misalignment artefacts. Section 7 concludes the paper and gives an outlook on future work.

2. Related Specular Highlights Segmentation Methods

There exist a number of approaches to segment specular highlights in images, usually either by detecting grey scale intensity jumps [30, 31] or sudden colour changes [32, 33] in an image. This can be seen as detecting the instances, when the image properties violate the assumption of diffuse reflection. The problem is also closely related to the detection of defects in still images or videos, which has been studied extensively (for an overview, see [34]).

The segmentation and inpainting of specular highlights was found to be beneficial in the context of indistinct frame detection in colonoscopic videos [26]. Furthermore, Cao et al. [35], detected specular highlights to facilitate the segmentation process in their algorithm for better detection of medical instruments in endoscopic images. However, this approach inherently detects only specular highlights of a specific size.

The algorithm presented in [26] detects specular highlights of all sizes and incorporates the idea of detecting absolutely bright regions in a first step and relatively bright regions in a second step. This idea fits the problem well, as most of the specular highlights appear saturated white or contain at least one saturated colour channel, while some, usually relatively small reflections are not as bright and appear as light grey or coloured spots. Figure 2 illustrates those different types of specular highlights.
Figure 2

Example illustrating absolutely bright (green) and relatively bright (yellow) specular highlights.

In their approach, Oh et al. [26], first converted the image to the HSV colour space (Hue, Saturation, Value). To obtain the absolutely bright regions, they used two thresholds, and , on value ( ) and saturation ( ), respectively, and classified a pixel at location as absolutely bright, if it satisfied the following conditions:
(1)

After this step, the image was segmented into regions of similar colour and texture using the image segmentation algorithm presented in [36], which involves colour quantisation and region growing and merging at multiple scales. Within those regions, relatively bright pixels were found using (1) with the same saturation threshold and a value threshold , computed for each region using the 75th percentile and the interquartile range of the values in that region. The union of the set of the absolutely bright pixels as computed in the first step and the set of the relatively bright pixels as obtained through the second step are considered as the set of the specular highlight pixels.

A disadvantage of this method is the high computational cost of the segmentation algorithm. Another issue is the choice of the colour space. Many endoscopy units nowadays use sequential RGB image acquisition. In this technique, the colour image is composed of three monochromatic images taken at different time instances under subsequent red, green and blue illumination. While this allows for an increase in image resolution, it has the disadvantage that fast camera motion leads to misalignment of the colour channels (Figure 1(b)). Consequently, specular highlights can appear either white or highly saturated red, green or blue. The fact that the method presented in [26] only detects specular highlights by thresholding the value and saturation channels, makes it less applicable to sequential RGB systems. In Section 4 we evaluate the proposed method against the one proposed by Oh et al. which we implemented as described in [26].

3. Proposed Specular Highlights Segmentation Method

The proposed segmentation approach comprises two separate modules that make use of two related but different characteristics of specular highlights.

3.1. Module 1

The first module uses colour balance adaptive thresholds to determine the parts of specular highlights that show a too high intensity to be part of the nonspecular image content. It assumes that the colour range of the nonspecular image content is well within the dynamic range of the image sensor. The automatic exposure correction of endoscope systems is generally reliable in this respect, so the image very rarely shows significant over- or underexposure. In order to maintain compatibility with sequential RGB imaging systems, we need to detect specular highlights even if they only occur in one colour channel. While this suggests 3 independent thresholds for each of the 3 colour channels, we set one fixed grey scale threshold and compute the colour channel thresholds using available image information.

More specifically, the colour channels may have intensity offsets due to colour balancing. At the same time the actual intensity of the specular highlights can be above the point of saturation of all three colour channels. Therefore, we normalise the green and blue colour channels, and , according to the ratios of the 95th percentiles of their intensities to the 95th percentile of the grey scale intensity for every image, which we computed as , with being the red colour channel. Using such high percentiles compensates for colour balance issues only if they show in the very high intensity range, which results in a more robust detection for varying lighting and colour balance. The reason why we use the grey scale intensity as a reference instead of the dominating red channel is the fact that intense reddish colours are very common in colonoscopic videos and therefore a red intensity close to saturation occurs not only in connection with specular highlights. We compute the colour balance ratios as follows:
(2)
with being the 95th percentile. Using these ratios, any given pixel is marked as a possible specular highlight when the following condition is met:
(3)

3.2. Module 2

The second module compares every given pixel to a smoothed nonspecular surface colour at the pixel position, which is estimated from local image statistics. This module is aimed at detecting the less intense parts of the specular highlights in the image. Looking at a given pixel, the underlying nonspecular surface colour could be estimated as a colour representative of an area surrounding the pixel, if it was known that this area does not contain specular highlights or at least which pixels in the area lie on specular highlights. Although we do not know this exactly, we can obtain a good estimate using global image thresholding and an outlier resilient estimation of the representative colour. Once this representative colour is computed, we determine the class of the current pixel from its dissimilarity to this colour.

The algorithm is initialised by an image thresholding step similar to the one in the first module: Using a slightly lower threshold , pixels with high intensity are detected using the condition in (3). The pixels meeting this condition are likely to belong to specular highlights, which is one part of the information we need. The actual computation of the representative colour is performed by a modified median filter. Similar nonlinear filters have been successfully used in defect detection in images and video (see, e.g., [37, 38]), which is a closely related problem. The median filter was chosen for its robustness in the presence of outliers and its edge preserving character, both of which make it an ideal choice for this task.

We incorporate the information about the location of possible specular highlights into the median filter by filling each detected specular region with the centroid of the colours of the pixels in an area within a fixed distance range from the contour of the region. We isolate this area of interest by exclusive disjunction of the masks obtained from two different dilation operations on the mask of possible specular highlight locations. For the dilation we use disk shaped structuring elements with radii of 2 pixels and 4 pixels, respectively. The same concept of filling of the specular highlights is also used in the proposed image inpainting method, which is described in Section 5.

We then perform median filtering on this modified image. Filling possible specular highlights with a representative colour of their surrounding effectively prevents the filtered image to appear too bright in regions where specular highlights cover a large area. Smaller specular highlights are effectively removed by the median filter when using a relatively large window size . Figure 3 shows an example of the output of the median filter.
Figure 3

Example of a colonoscopic image before and after median filtering.

Following this, specular highlights are found as positive colour outliers by comparing the pixel values in the input and the median filtered image. For this comparison, several distance measures and ratios are possible. Examples of such measures are the euclidean distance in RGB space or the infinity norm of the differences. During evaluation we found that the maximal ratio of the three colour channel intensities in the original image and the median filtered image produces optimal results. For each pixel location , this intensity ratio is computed as
(4)
with , , and being the intensities of the red, green and blue colour channel in the median filtered image, respectively. Here again, varying colour balance and contrast can lead to large variations of this characteristic for different images. These variations are compensated using a contrast coefficient , which is calculated for each of the 3 colour channels for every given image as
(5)
with being the sample mean of all pixel intensities in colour channel and being the sample standard deviation. Using these coefficients, we modify (4) to obtain the contrast compensated intensity ratio as follows:
(6)
Using a threshold for this relative measure, the pixel at location is then classified as a specular highlight pixel, if
(7)

At this point the outputs of the first and second module are joined by logical disjunction of the resulting masks. The two modules complement each other well: The first module uses a global threshold and can therefore only detect the very prominent and bright specular highlights. The less prominent ones are detected by the second module by looking at relative features compared to the underlying surface colour. With a higher dynamic range of the image sensor, the second module alone would lead to good results. However, since the sensor saturates easily, the relative prominence of specular highlights becomes less intense the brighter a given area of an image is. It is these situations in which the first module still allows detection.

3.3. Postprocessing

During initial tests we noticed that some bright regions in the image are mistaken for specular highlights by the algorithm presented so far. In particular, the mucosal surface in the close vicinity of the camera can appear saturated without showing specular reflection and may therefore be picked up by the detection algorithm. To address this problem, we made use of the property, that the image area surrounding the contour of specular highlights generally shows strong image gradients. Therefore, we compute the mean of the gradient magnitude in a stripe-like area within a fixed distance to the contours of the detected specular regions. Using this information, only those specular regions are retained, whose corresponding contour areas meet the condition
(8)
with being the grey scale gradient magnitude of the th out of pixels of the contour area corresponding to a given possible specular region. is a constant allowing to restrict the computation to larger specular regions, as the problem of nonspecular saturation occurs mainly in large uniform areas. The gradient is approximated by vertical and horizontal differences of directly neighbouring pixels. Figure 4 illustrates the idea. Using this approach, bright, nonspecular regions such as the large one on the right in Figure 4(a), can be identified as false detections.
Figure 4

Illustration of the area that is used for the gradient test. (a) original image. (b) detected specular highlights. (c) contour areas for the gradient test, (d) resulting specular highlights after the gradient test.

In the presence of strong noise it can happen that single isolated pixels are classified as specular highlights. These are at this stage removed by morphological erosion. The final touch to the algorithm is a slightly stronger dilation of the resulting binary mask, which extends the specular regions more than it would be necessary to compensate for the erosion. This step is motivated by the fact that the transition from specular to nonspecular areas is not a step function but spread due to blur induced by factors such as motion or residues on the camera lens. The mask is therefore slightly extended to better cover the spread out regions.

4. Evaluation of the Segmentation Method

In order to evaluate the proposed algorithm a large ground truth dataset was created by manually labelling a set of 100 images from 20 different colonoscopy videos. Since negative effects of specular highlights on image analysis algorithms are mostly due to the strong gradients along their contours, the gradient magnitudes were computed using a Sobel operator and overlayed on the images. This allowed the manual labelling to be very precise on the contours. Great care was taken in including the contours fully in the marked specular regions.

In order to compare the performance of the proposed algorithm with the state of the art, we implemented the approach proposed by Oh et al. as described in [26], which was also proposed for detection of specular highlights in endoscopic images. Both methods were assessed by their performance to classify the pixels of a given image into either specular highlight pixels or other pixels.

Using the aforementioned data set, we evaluated both methods using a cross-validation scheme where in each iteration the images of one video were used as the test set and the rest of the images were used as the training set. For each iteration we optimised the parameters of both the method in [26] and the proposed one using the training set and tested their performance on the test set. At any point no information about the test image was used in the optimizing process of the parameters. We chose two different cost scenarios to measure optimal performance: scenario A assigned equal costs (unit per misclassified pixel) to missed specular highlights and falsely detected specular highlights; scenario B assigned twice the cost to missed specular highlights (2 units per missed specular highlight pixel).

The results are reported in Tables 1 and 2 with the resulting cost and the commonly used measures accuracy, precision, sensitivity and specificity [39], for the two cost scenarios, averaged over the 20 cross-validation iterations. We report two different variants of the method in [26]. One is the original method as it was reported in [26]. The second method is equivalent to the first, followed by a dilation similar to one in the postprocessing step of the proposed method. This was considered appropriate and necessary for a better comparison of the two methods, because in our understanding of the extent of specular highlights, any image gradient increase due to the contours of the specular highlights is to be included during labelling, while the definition in [26] was motivated by a purely visual assessment. The overall improvement resulting from this modification, as it can be seen in Tables 1 and 2, supports this interpretation.
Table 1

Performance of the algorithm for equal costs of false positives and false negatives. Compared to the method in [26] with dilation the proposed method achieves a cost reduction of 28.16%.

Method

Cost

Accuracy [%]

Precision [%]

Sensitivity [%]

Specificity [%]

Method of Oh et al.

8070

96.83

87.76

37.27

99.25

Method of Oh et al. with Dilation

6473

97.35

86.66

53.34

99.14

Proposed Method

4650

98.33

81.29

75.31

99.28

Table 2

Performance of the algorithm for doubled costs of false negatives. Compared to the method in [26] with dilation the proposed method achieves a cost reduction of 31.03%.

Method

Cost

Accuracy [%]

Precision [%]

Sensitivity [%]

Specificity [%]

Method of Oh et al.

15400

96.70

86.15

39.94

99.01

Method of Oh et al. with Dilation

10271

97.05

68.85

69.09

98.13

Proposed Method

7084

97.90

70.23

83.78

98.51

It can be seen that the proposed method outperforms the one presented in [26] substantially with a cost reduction of 28.16% and 31.03% for cost scenario A and B, respectively. Furthermore, the proposed algorithm was able to process 2.34 frames per second on average on a 2.66 GHz Intel Core2Quad system—a speed improvement of a factor of 23.8 over the approach presented in [26], which is heavily constrained by its image segmentation algorithm. It took 10.18 seconds on average to process an image. The results are visually depicted in Figure 6.

While the parameters were optimised for each iteration of the cross-validation scheme, they varied only marginally. For images with similar dimensions (in the vicinity of ) to the ones used in this study, we recommend to use the following parameters for cost scenario A (cost scenario B): , , , median filter window size , , . The size of the structuring element for the dilation in the postprocessing step should be 3 and 5 for cost scenario A and B, respectively.

5. Inpainting of Specular Highlights

Image inpainting is the process of restoring missing data in still images and usually refers to interpolation of the missing pixels using information of the surrounding neighbourhood. An overview over the commonly used techniques can be found in [40] or, for video data, in [34].

For most applications in automated analysis of endoscopic videos, inpainting will not be necessary. The information about specular highlights will be used directly (in algorithms exploiting this knowledge), or the specular regions will simply be excluded from further processing. However, a study by Vogt et al. [16], suggests that well-inpainted endoscopic images are preferred by physicians over images showing specular highlights. Algorithms with the intention of visual enhancement may therefore benefit from a visually pleasing inpainting strategy, as well as algorithms working in the frequency domain. Vogt et al. also [16] proposed an inpainting method based on temporal information and can be only used for a sequence of frames in a video and not for isolated individual images.

An inpainting method was reported by Cao et al. in [35]. The authors replaced the pixels inside a sliding rectangular window by the average intensity of the window outline, once the window covered a specular highlight. The approach can not be used universally, as it is matched to the specular highlight segmentation algorithm presented in the same paper.

In [26], along with their specular highlight segmentation algorithm, the authors also reported an image inpainting algorithm, where they replaced each detected specular highlight by the average intensity on its contour. A problem with this approach is that the resulting hard transition between the inpainted regions and their surroundings may again lead to strong gradients.

In order to prevent these artefacts, in the proposed algorithm, the inpainting is performed on two levels. We first use the filling technique presented in Section 3, where we modify the image by replacing all detected specular highlights by the centroid colour of the pixels within a certain distance range of the outline (see above for details). Additionally, we filter this modified image using a Gaussian kernel ( ), which results in a strongly smoothed image free of specular highlights, which is similar to the median filtered image in the segmentation algorithm.

For the second level, the binary mask marking the specular regions in the image is converted to a smooth weighting mask. The smoothing is performed by adding a nonlinear decay to the contours of the specular regions. The weights of the pixels surrounding the specular highlights in the weighting mask are computed depending on their euclidean distance to the contour of the specular highlight region:
(9)

which can be interpreted as a logistic decay function in a window from to , mapped to a distance range from 0 to . The constant can be used to introduce a skew on the decay function. In the examples in this paper, we use the parameters , , and .

The resulting integer valued weighting mask (see, e.g., Figure 5(e)) is used to blend between the original image and the smoothed filled image . The smoothing of the mask results in a gradual transition between and . Figure 5 illustrates the approach by showing the relevant images and masks.
Figure 5

Stages of the inpainting algorithm. (a) Original image (b) Image section showing the specular highlights (c) Gaussian filtered, filled image section (d) Detected specular highlights (e) Weighting mask (f) Inpainted image section

Figure 6

Examples illustrating the performance of the specular highlight segmentation algorithm. Original images are shown in the first column. The second column contains the ground truth images, the third column shows the results of the method presented in [26] and in the fourth column the results achieved by the proposed algorithm are depicted.

The inpainted image is computed for all pixel locations using the following equation:
(10)

with for all pixel locations .

Figure 7 shows a number of images before and after inpainting and a comparison to inpainting method reported in [26]. It can be seen that the proposed inpainting method produces only minor artefacts for small specular highlights. Very large specular regions, however, appear strongly blurred. This is an obvious consequence from the Gaussian smoothing. For more visually pleasing results for large specular areas, it would be necessary to use additional features of the surroundings, such as texture or visible contours. However, such large specular regions are rare in clear colonoscopic images and errors arising from them can therefore usually be neglected. The performance of the combination of the presented segmentation and inpainting algorithms can be seen in an example video which is available online in the following website: http://www.scss.tcd.ie/Anarta.Ghosh/WEB_PAGE_SPECU_PAPER/.
Figure 7

Examples illustrating the performance of the inpainting algorithm. Original images are shown in the first column. The second column contains images which were inpainted using the proposed method and the third column shows the results of the method presented in [26]. The segmentation of specular highlights prior to inpainting was performed using the proposed segmentation algorithm.

6. Specular Highlights and Colour Channel Misalignment Artefacts

Sequential RGB image acquisition systems are very commonly used in endoscopy. In these systems the images corresponding to the red (R), the green (G) and the blue (B) colour channels are acquired at different time instances and merged to form the resulting video frame. However, an inherent technological shortcoming of such systems is: whenever the speed of the camera is high enough such that it moves significantly in the time interval between the acquisition instances of the images corresponding to two colour channels, they get misaligned in the resulting video frame, compare, Figure 1(b). This channel misalignment gives the images an unnatural, highly colourful, and stroboscopic appearance, which degrades the overall video quality of the minimally invasive procedures. Moreover, in endoscopic images, the colour is an invariant characteristic for a given status of the organ [41]. Malignant tumors are usually inflated and inflamed. This inflammation is usually reddish and more severe in colour than the surrounding tissues. Benign tumors exhibit less intense colours. Hence the colour is one of the important features used both in clinical and automated detection of lesions [42]. Consequently, removal of these artefacts is of high importance both from the clinical and the technical perspectives.

We developed an algorithm to remove these colour channel misalignment artefacts as follows. Let , , be the three colour channels of a given endoscopy video frame. The developed algorithm to remove the colour misalignment artefacts comprises the following key steps.
  1. (i)

    Compute the Kullback-Leibler divergence, , between the intensity histograms of the colour channels, denoted as: , , for all . is the intensity histogram corresponding to colour channel . Choose the colour channels and , for which the is minimum.

     
  2. (ii)

    Compute the homography ( ) between the chosen colour channels and , through feature matching. Assume linearity of motion and compute the homography between consecutive colour channels, , .

     
  3. (iii)

    Align all the colour channels by using the inverse homography, , .

     
We tested the algorithm with 50 colonoscopy video frames before (Dataset 1) and after (Dataset 2) removing specular highlights. The measures used to evaluate the algorithm are as follows: (a) percentage of images where colour channels were successfully realigned (SR), (b) percentage of images where colour channels were not successfully realigned but they were not distorted either (USRND), (c) percentage of images where colour channels were not successfully realigned moreover they were also distorted (USRD). Successful realignment and distortion of the images were evaluated using visual inspection. The results of the evaluation are shown in Table 3 and visualized in Figure 8. We see a substantial improvement when specular highlights are removed.
Table 3

Performance of the colour channel misalignment artefact removal algorithm in images before and after removing specular highlights. SR: percentage of images where the colour channels were successfully realigned. USRND: percentage of images where the colour channels were not successfully realigned, however they were not distorted. USRD: percentage of images where the colour channels were not successfully realigned and they were also distorted. Dataset 1: 50 colonoscopy video frame with colour channel misalignment. Dataset 2: Dataset 1 after specular highlights are removed by the proposed algorithm.

Dataset

SR [%]

USRND [%]

USRD [%]

Dataset 1

78

18

4

Dataset 2

84

14

2

Figure 8

The results of colour channel realignment algorithm in Datasets 1 (a, b) and 2 (c, d). (a, c): the original images. (b, d): the resulting images after the colour channel misalignment artefacts are removed.

7. Discussion

In this paper, we have presented methods for segmenting and inpainting specular highlights. We have argued that specular highlights can negatively affect the perceived image quality. Furthermore, they may be a significant source of error, especially for algorithms that make use of the gradient information in an image. The proposed segmentation approach showed a promising performance in the detailed evaluation. It performed favourably to the approach presented in [26] and avoids any initial image segmentation, thus resulting in significantly shorter computation time (a reduction by a factor of 23.8 for our implementation). Furthermore, in contrast to other approaches, the proposed segmentation method is applicable to the widely used sequential RGB image acquisition systems. In the sequential RGB endoscope, a very common problem is the colour channel misalignment artefacts. We developed a simple algorithm to remove these artefacts and tested it using colonoscopy video frames before and after removing specular highlight. A substantial improvement in the performance was observed when specular highlights are removed. The performance of the proposed inpainting approach was demonstrated on a set of images and compared to the inpainting method proposed in [26].

When using inpainting in practice, it is important to keep the users informed that specular highlights are being suppressed and to allow for disablement of this enhancement. For example, while inpainting of specular highlights may help in detecting polyps (both for human observers and algorithms) it could make their categorisation more difficult, as it alters the pit-pattern of the polyp in the vicinity of the specular highlight. Also, as it can be seen in the second row of Figure 7, inpainting can have a blurring effect on medical instruments. Explicit detection of medical instruments may allow to prevent these artefacts and will be considered in future studies.

Future work will also include a clinical study into whether endoscopists prefer inpainted endoscopic videos over standard ones. We will further investigate to what degree other image analysis algorithms for endoscopic videos benefit from using the proposed methods as preprocessing steps.

Declarations

Acknowledgments

This work has been supported by the Enterprise Ireland Endoview project CFTD-2008-204. Thea authors would also like to acknowledge the support from National Development Plan, 2007-2013, Ireland.

Authors’ Affiliations

(1)
School of Computer Science and Statistics, Trinity College

References

  1. Khan GN, Gillies DF: Vision based navigation system for an endoscope. Image and Vision Computing 1996,14(10):763-772. 10.1016/S0262-8856(96)01085-2View ArticleGoogle Scholar
  2. Kwoh CK, Khan GN, Gillies DF: Automated endoscope navigation and advisory system from medical imaging. Medical Imaging: Physiology and Function from Multidimensional Images, 1999, Proceedings of SPIE 3660: 214-224.Google Scholar
  3. Phee SJ, Ng WS, Chen IM, Seow-Choen F, Davies BL: Automation of colonoscopy part II. visual-control aspects: interpreting images with a computer to automatically maneuver the colonoscope. IEEE Engineering in Medicine and Biology Magazine 1998,17(3):81-88. 10.1109/51.677173View ArticleGoogle Scholar
  4. Sucar LE, Gillies DF: Knowledge-based assistant for colonscopy. Proceedings of the 3rd International Conference on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems (IEA/AIE '90), July 1990 665-672.Google Scholar
  5. Uecker DR, Lee C, Wang YF, Wang Y: Automated instrument tracking in robotically assisted laparoscopic surgery. Journal of Image Guided Surgery 1995,1(6):308-325. 10.1002/(SICI)1522-712X(1995)1:6<308::AID-IGS3>3.0.CO;2-EView ArticleGoogle Scholar
  6. Voros S, Long JA, Cinquin P: Automatic detection of instruments in laparoscopic images: a first step towards high-level command of robotic endoscopic holders. International Journal of Robotics Research 2007,26(11-12):1173-1190. 10.1177/0278364907083395View ArticleGoogle Scholar
  7. Wang YF, Uecker DR, Wang Y: A new framework for vision-enabled and robotically assisted minimally invasive surgery. Computerized Medical Imaging and Graphics 1998,22(6):429-437. 10.1016/S0895-6111(98)00052-4View ArticleGoogle Scholar
  8. Cao Y, Li D, Tavanapong W, Oh J, Wong J, de Groen PC: Parsing and browsing tools for colonoscopy videos. Proceedings of the 12th ACM International Conference on Multimedia (Multimedia '04), October 2004 844-851.View ArticleGoogle Scholar
  9. Cunha JPS, Coimbra M, Campos P, Soares JM: Automated topographic segmentation and transit time estimation in endoscopic capsule exams. IEEE Transactions on Medical Imaging 2008,27(1):19-27.View ArticleGoogle Scholar
  10. Iakovidis DK, Tsevas S, Maroulis D, Polydorou A: Unsupervised summarisation of capsule endoscopy video. Proceedings of the 4th International IEEE Conference Intelligent Systems (IS '08), September 2008 315-320.Google Scholar
  11. Burschka D, Li M, Taylor R, Hager GD: Scale-invariant registration of monocular endoscopic images to CT-scans for sinus surgery. Proceedings of the 7th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI '04), 2004, Lecture Notes in Computer Science 3217: 413-421.Google Scholar
  12. Gross P, Kitney RI, Claesen S, Halls JM: Mr-compatible endoscopy and tracking for image-guided surgery. Proceedings of the 15th International Congress and Exhibition of Computer Assisted Radiology and Surgery, 2001 1230: 1076-1082.Google Scholar
  13. Liu J, Yoo T, Subramanian K, Van Uitert R: A stable optic-flow based method for tracking colonoscopy images. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPR '08), June 2008 1-8.Google Scholar
  14. Mori K, Deguchi D, Hasegawa J, et al.: A method for tracking the camera motion of real endoscope by epipolar geometry analysis and virtual endoscopy system. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI '06), 2001 1-8.Google Scholar
  15. Wengert C, Cattin PC, Duff JM, Baur C, Sźekely G: Markerless endoscopic registration and referencing. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI '06), 2006 4190: 816-823.Google Scholar
  16. Vogt F, Paulus D, Heigl B, Vogelgsang C, Niemann H, Greiner G, Schick C: Making the invisible visible: highlight substitution by color light fields. Proceedings of the 1st European Conference on Colour in Graphics, Imaging, and Vision (CGIV '02), April 2002 352-357.Google Scholar
  17. Coimbra MT, Cunha JPS: MPEG-7 visual descriptors—contributions for automated feature extraction in capsule endoscopy. IEEE Transactions on Circuits and Systems for Video Technology 2006,16(5):628-636.View ArticleGoogle Scholar
  18. Esgiar AN, Naguib RNG, Sharif BS, Bennett MK, Murray A: Fractal analysis in the detection of colonic cancer images. IEEE Transactions on Information Technology in Biomedicine 2002,6(1):54-58. 10.1109/4233.992163View ArticleGoogle Scholar
  19. Karkanis SA, Iakovidis DK, Maroulis DE, Karras DA, Tzivras M: Computer aided tumor detection in endoscopic video using color wavelets features. IEEE Transactions on Information Technology in Biomedicine 2003,7(3):141-152. 10.1109/TITB.2003.813794View ArticleGoogle Scholar
  20. Maroulis DE, Iakovidis DK, Karkanis SA, Karras DA: Cold: a versatile detection system for colorectal lesions in endoscopy video-frames. Computer Methods and Programs in Biomedicine 2003,70(2):151-166. 10.1016/S0169-2607(02)00007-XView ArticleGoogle Scholar
  21. Cao Y, Tavanapong W, Kim K, Wong J, Oh J, de Groen PC: A framework for parsing colonoscopy videos for semantic units. Proceedings of the IEEE International Conference on Multimedia and Expo (ICME '04), June 2004 3: 1879-1882.Google Scholar
  22. Oh J, Rajbal MA, Muthukudage JK, Tavanapong W, Wong J, de Groen PC: Real-time phase boundary detection in colonoscopy videos. Proceedings of the 6th International Symposium on Image and Signal Processing and Analysis (ISPA '09), September 2009 724-729.Google Scholar
  23. Liu D, Cao Y, Tavanapong W, Wong J, Oh J, de Groen PC: Mining colonoscopy videos to measure quality of colonoscopic procedures. Proceedings of the 5th IASTED International Conference on Biomedical Engineering (BioMED '07), February 2007 409-414.Google Scholar
  24. Arnold M, Ghosh A, Lacey G, Patchett S, Mulcahy H: Indistinct frame detection in colonoscopy videos. Proceedings of the 13th International Machine Vision and Image Processing Conference (IMVIP '09), September 2009 47-52.Google Scholar
  25. Tjoa MP, Krishnan SM: Texture-based quantitative characterization and analysis of colonoscopic images. Proceedings of Annual International Conference of the IEEE Engineering in Medicine and Biology, October 2002, Houston, Tex, USA 2: 1090-1091.Google Scholar
  26. Oh J, Hwang S, Lee J, Tavanapong W, Wong J, de Groen PC: Informative frame classification for endoscopy video. Medical Image Analysis 2007,11(2):110-127. 10.1016/j.media.2006.10.003View ArticleGoogle Scholar
  27. Dahyot R, Vilariño F, Lacey G: Improving the quality of color colonoscopy videos. EURASIP Journal on Image and Video Processing 2008, 2008:-7.Google Scholar
  28. Liu D, Cao Y, Tavanapong W, Wong J, Oh J, de Groen PC: Quadrant coverage histogram: a new method for measuring quality of colonoscopic procedures. Proceedings of the 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society 2007, 2007: 3470-3473.View ArticleGoogle Scholar
  29. Oh J, Hwang S, Cao Y, Tavanapong W, Liu D, Wong J, de Groen PC: Measuring objective quality of colonoscopy. IEEE Transactions on Biomedical Engineering 2009,56(9):2190-2196.View ArticleGoogle Scholar
  30. Forbus K: Light source effects. Massachusetts Institute of Technology; 1977.Google Scholar
  31. Brelstaff G, Blake A: Detecting specular reflections using lambertian constraints. Proceedings of the 2nd International Conference on Computer Vision, 1988 297-302.Google Scholar
  32. Gershon R, Jepson AD, Tsotsos JK: The use of color in highlight identification. Proceedings of the 10th International Joint Conference on Artificial Intelligence, 1987 2: 752-754.Google Scholar
  33. Klinker G, Shafer S, Kanade T: Using a color reflection model to separate highlights from object color. Proceedings of the 1st International Conference on Computer Vision, 1987 145-150.Google Scholar
  34. Kokaram AC: On missing data treatment for degraded video and film archives: a survey and a new Bayesian approach. IEEE Transactions on Image Processing 2004,13(3):397-415. 10.1109/TIP.2004.823815View ArticleGoogle Scholar
  35. Cao Y, Liu D, Tavanapong W, Wong J, Oh J, de Groen PC: Computer-aided detection of diagnostic and therapeutic operations in colonoscopy videos. IEEE Transactions on Biomedical Engineering 2007,54(7):1268-1279.View ArticleGoogle Scholar
  36. Deng Y, Manjunath BS: Unsupervised segmentation of color-texture regions in images and video. IEEE Transactions on Pattern Analysis and Machine Intelligence 2001,23(8):800-810. 10.1109/34.946985View ArticleGoogle Scholar
  37. Decencière E: Motion picture restoration using morphological tools. In Mathematical Morphology and Its Applications to Image and Signal Processing. Edited by: Maragos P, Schafer RW, Butt MA. Kluwer Academic Publishers, Norwell, Mass, USA; 1996:361-368.Google Scholar
  38. Buisson O, Besserer B, Boukir S, Helt F: Deterioration detection for digital film restoration. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, June 1997 78-84.View ArticleGoogle Scholar
  39. Han J, Kamber M: Data Mining: concepts and Techniques. Morgan Kaufmann, San Francisco, Calif, USA; 2006.MATHGoogle Scholar
  40. Shih TK, Chang RC: Digital inpainting—survey and multilayer image inpainting algorithms. Proceedings of the 3rd International Conference on Information Technology and Applications (ICITA '05), July 2005 1: 15-24.View ArticleGoogle Scholar
  41. Paris Workshop Participants : The Paris endospcopic classification of superficial neoplastic lesions. Gastrointestinal Endoscopy 2003,58(6):3-23.View ArticleGoogle Scholar
  42. Karkanis SA, Iakovidis DK, Maroulis DE, Karras DA, Tzivras M: Computer aided tumor detection in endoscopic video using color wavelets features. IEEE Transactions on Information Technology in Biomedicine 2003,7(3):141-152. 10.1109/TITB.2003.813794View ArticleGoogle Scholar

Copyright

© Mirko Arnold et al. 2010

This article is published under license to BioMed Central Ltd. This is an open access paper distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.