A Benchmark Data Set to Evaluate the Illumination Robustness of Image Processing Algorithms for Object Segmentation and Classification

Arif Ul Maula Khan, Ralf Mikut, Markus Reischl, Arif Ul Maula Khan, Ralf Mikut, Markus Reischl

Abstract

Developers of image processing routines rely on benchmark data sets to give qualitative comparisons of new image analysis algorithms and pipelines. Such data sets need to include artifacts in order to occlude and distort the required information to be extracted from an image. Robustness, the quality of an algorithm related to the amount of distortion is often important. However, using available benchmark data sets an evaluation of illumination robustness is difficult or even not possible due to missing ground truth data about object margins and classes and missing information about the distortion. We present a new framework for robustness evaluation. The key aspect is an image benchmark containing 9 object classes and the required ground truth for segmentation and classification. Varying levels of shading and background noise are integrated to distort the data set. To quantify the illumination robustness, we provide measures for image quality, segmentation and classification success and robustness. We set a high value on giving users easy access to the new benchmark, therefore, all routines are provided within a software package, but can as well easily be replaced to emphasize other aspects.

Conflict of interest statement

Competing Interests: The authors have declared that no competing interests exist.

Figures

Fig 1. Benchmark data sets.
Fig 1. Benchmark data sets.
Overview of images in R = 4 scenes corresponding to increasing artifact and noise levels
Fig 2. Representation of ground truth objects…
Fig 2. Representation of ground truth objects using color shading and numbers as labels.
Left: Brightfield image X of a benchmark data set with marked object edges, right: Ground truth image Xtruth, with object types in numbers. Gray scales denote the value of xij,truth (0: black (background), 5: white)
Fig 3. Exemplary pipeline for the segmentation…
Fig 3. Exemplary pipeline for the segmentation of benchmark images.
Fig 4. Results of benchmark data set…
Fig 4. Results of benchmark data set r = 1.
Total quality criterion Q(r, b, n) vs. increasing artifact level A(r, b, n) for image series with stepwise addition of both shading and noise for each successive image. The first row of images indicates original images from data set r = 1. The second row shows corresponding segmentation and classification results using Otsu’s method. The third row shows corresponding segmentation and classification results using edge detection method. Brown color represents correct classification of the segmented BLOB w.r.t corresponding ground truth BLOB and light green color shows an erroneous classification object. Robustness values for Otsu thresholding and Sobel edge detection are Rotsu = 0.54 and Redge = 0.47 respectively.

References

    1. Zhang W, Morris Q, Chang R, Shai O, Bakowski M, Mitsakakis N, et al. (2004) The functional landscape of mouse gene expression. Journal of Biology 3: 21 10.1186/jbiol16
    1. Lieschke GJ, Currie PD (2007) Animal models of human disease: zebrafish swim into view. Nature Reviews Genetics 8: 353–67. 10.1038/nrg2091
    1. Nüsslein-Volhard C, Wieschaus E (1980) Mutations affecting segment number and polarity in drosophila. Nature 287: 795–801. 10.1038/287795a0
    1. Caselles V, Kimmel R, Sapiro G (1997) Geodesic active contours. International Journal of Computer Vision 22: 61–79. 10.1023/A:1007979827043
    1. Bock R, Grote H, Notz D, Regler M (2009) Data Analysis Techniques for High-Energy Physics Experiments. Academic Press.
    1. Boehme HJ, Baumann U, Corradini A, Gross H (1999) Person localization and posture recognition for human-robot-interaction In: Braffort A, editor, Gesture Based Communication in Human-Computer Interaction, Springer Verlag.
    1. Nickel K, Stiefelhagen R (2007) Visual recognition of pointing gestures for human–robot interaction. Image and Vision Computing 25: 1875–1884. 10.1016/j.imavis.2005.12.020
    1. Brosnan T, Sun D (2004) Improving quality inspection of food products by computer vision—a review. Journal of Food Engineering 61: 3–16. 10.1016/S0260-8774(03)00183-3
    1. Du C, Sun D (2004) Recent developments in the applications of image processing techniques for food quality evaluation. Trends in Food Science and Technology 15: 230–249. 10.1016/j.tifs.2003.10.006
    1. Kavoussanos M, Pouliezzos A (1998) An intelligent robotic system for depalletizing and emptying polyethylene sacks. International Journal of Advanced Manufacturing Technology 14: 358–362. 10.1007/BF01178915
    1. Patil G, Chaudhari D (2013) Sift based approach: Object recognition and localization for pick-and-place system. International Journal of Advanced Research in Computer Science and Software Engineering 3 (3): 196–201.
    1. Wang WH, Liu XY, Sun Y (2006) Robust contact detection in micromanipulation using computer vision microscopy. In: Proc., 28th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS). pp. 2219–2222.
    1. Malamas E, Petrakis E, Zervakis M, Petit L, Legat J (2003) A survey on industrial vision systems, applications and tools. Image and Vision Computing 21(2): 171–188. 10.1016/S0262-8856(02)00152-X
    1. Gelasca ED, Byun J, Obara B, Manjunath B (2008) Evaluation and benchmark for biological image segmentation. In: IEEE International Conference on Image Processing (ICIP). pp. 1816–1819.
    1. Ruusuvuori P, Lehmussola A, Selinummi J, Rajala T, Huttunen H, Yli-Harja O (2008) Benchmark set of synthetic images for validating cell image analysis algorithms. In: Proceedings of the 16th European Signal Processing Conference, EUSIPCO.
    1. Escalante HJ, Hernández CA, Gonzalez JA, López-López A, Montes M, Morales EF, et al. (2010) The segmented and annotated IAPR TC-12 benchmark. Computer Vision and Image Understanding 114: 419–428. 10.1016/j.cviu.2009.03.008
    1. Oh S, Hoogs A, Perera A, Cuntoor N, Chen CC (2011) A large-scale benchmark dataset for event recognition in surveillance video. In: Proc., IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, pp. 3153–3160.
    1. Guthaus MR, Ringenberg JS, Ernst D, Austin TM, Mudge T, Brown RB. (2001) MiBench: A free, commercially representative embedded benchmark suite. In: Proc., IEEE International Workshop on Workload Characterization (WWC-4). IEEE, pp. 3–14.
    1. Deng J, Dong W, Socher R, Li L, Li K, Fei-Fei L (2009) Imagenet: A large-scale hierarchical image database. In: Proc., IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Ieee, pp. 248–255.
    1. Ge F, Wang S, Liu T (2007) New benchmark for image segmentation evaluation. Journal of Electronic Imaging 16: 033011-1–033011-16.
    1. Stallkamp J, Schlipsing M, Salmen J, Igel C (2011) The German traffic sign recognition benchmark: A multi-class classification competition. In: Proc., International Joint Conference on Neural Networks (IJCNN). IEEE, pp. 1453–1460.
    1. Anderson E (1935) The irises of the Gaspé Peninsula. Bulletin of the American Iris Society 59: 2–5.
    1. Forina, M et al (1991). PARVUS—an extendible package for data exploration, classification, and correlation. Institute of Pharmaceutical and Food Analysis and Technologies, Via Brigata Salerno, 16147 Genoa, Italy.
    1. Stegmaier J, Otte JC, Kobitski A, Bartschat A, Garcia A, Mikut R, et al. (2014) Fast segmentation of stained nuclei in terabyte-scale, time resolved 3D microscopy image stacks. PLoS ONE 9: e90036 10.1371/journal.pone.0090036
    1. Otsu N (1979) A threshold selection method from gray-level histograms. IEEE Transactions on Systems, Man and Cybernetics 9: 62–66. 10.1109/TSMC.1979.4310076
    1. Sobel I, Feldman G (1968) A 3 × 3 isotropic gradient operator for image processing.
    1. Stegmaier J, Alshut R, Reischl M, Mikut R (2012) Information fusion of image analysis, video object tracking, and data mining of biological images using the open source MATLAB toolbox Gait-CAD. Biomedizinische Technik (Biomedical Engineering) 57 (S1): 458–461.

Source: PubMed

3
Iratkozz fel