• DocumentCode
    2425709
  • Title

    Wall-adherent cells segmentation based on cross-entropy and watershed transform

  • Author

    Fan, Di ; Cao, Maoyong ; Lv, Changzhi

  • Author_Institution
    Shandong Univ. of Sci. & Technol., Qingdao
  • fYear
    2008
  • fDate
    7-9 July 2008
  • Firstpage
    1703
  • Lastpage
    1707
  • Abstract
    The microscopic image processing technology is a new solving approach to segment and count wall-adherent cells in anti-virus experiment in vitro. But the segmentation is very stubborn because of the cellspsila multiformity.This paper presents a segmentation strategy based on cross-entropy and watershed transform to segment and count the wall-adherent cells. Firstly, top-hat transform is used to enhance the original cells microscopic image. Suppose the conditional distributions of object and background are modeled with normal distributions, maximum between-class cross-entropy threshold segments the image into binary one. Then morphological filters reduce the burrs and holes in binary image and watershed transform further segments the cells by single-pixel wide edges. Finally, the cells are counted by labeling them.The experiments show that this strategy is effective, simply and strongly adaptive to lighting. The segmentation boundaries are continuious and the cellspsila shapes are well kept.
  • Keywords
    image resolution; image segmentation; transforms; cell microscopic image; cross-entropy-watershed transform; microscopic image processing technology; morphological filters; normal distributions; segmentation strategy; single-pixel wide edges; top-hat transform; wall-adherent cells; wall-adherent cells segmentation; Cells (biology); Drugs; Filters; Gaussian distribution; Image edge detection; Image processing; Image segmentation; In vitro; Microscopy; Shape;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Audio, Language and Image Processing, 2008. ICALIP 2008. International Conference on
  • Conference_Location
    Shanghai
  • Print_ISBN
    978-1-4244-1723-0
  • Electronic_ISBN
    978-1-4244-1724-7
  • Type

    conf

  • DOI
    10.1109/ICALIP.2008.4590160
  • Filename
    4590160