Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Semi-Supervised Subspace Learning for Mumford-Shah Model Based Texture Segmentation

Open Access Open Access

Abstract

We propose a novel image segmentation model which incorporates subspace clustering techniques into a Mumford-Shah model to solve texture segmentation problems. While the natural unsupervised approach to learn a feature subspace can easily be trapped in a local solution, we propose a novel semi-supervised optimization algorithm that makes use of information derived from both the intermediate segmentation results and the regions-of-interest (ROI) selected by the user to determine the optimal subspaces of the target regions. Meanwhile, these subspaces are embedded into a Mumford-Shah objective function so that each segment of the optimal partition is homogeneous in its own subspace. The method outperforms standard Mumford-Shah models since it can separate textures which are less separated in the full feature space. Experimental results are presented to confirm the usefulness of subspace clustering in texture segmentation.

©2010 Optical Society of America

1. Introduction

Image segmentation is indispensable in applications that use imaging systems. It facilitates the extraction of useful information for subsequent high level image analysis. For instance, in pathological research, digital microscopy becomes increasingly popular since the introduction of high-throughput tissue microarrays (TMA) into bioimaging communities [1]. It is therefore crucial to segment each tissue image into a meaningful partition in an accurate, fast, automated and robust manner. In robotics, segmentation is often coupled with 3D scene reconstruction.

Texture may be defined as the variation of data at scales smaller than the scales of interest [2]. In segmentation, texture can be a useful cue for object recognition, but it can also be a nuisance since it creates many extra lines in the edge map. Most segmentation algorithms deal with texture by first transforming the intensity levels to high-dimensional feature vectors, followed by looking for regions which are homogeneous in the feature space. For example, texture features extracted by Gabor filters [3] or Laws’ filters [4] have been widely used [5, 6, 7, 8]. Numerous approaches for texture segmentation exist: clustering [5], region-growing [9], graph-partitioning [10] and optimization model-based [11, 12, 13, 14]. Recent surveys in this topic can be found in [15] and [16].

1.1. Motivation

In the literature, ample research has also been done on designing algorithms for segmenting non-texture images, i.e. images that can be well-approximated by piecewise smooth functions. Prominent examples include the Mumford-Shah model [17], watershed algorithm [18], snake/active contour models [19], and their thousands of descendants. Although many of these algorithms are originally designed for non-texture piecewise smooth images, almost all of them can be naturally generalized to handle textured images reasonably well. This is because of the optimism that when feature extraction is done properly, the feature vector over the image domain is more or less piecewise smooth. Of course, when dealing with high-dimensional feature spaces, some complications arise.

First of all, textured images and their transforms often have no sharp boundary. For example, features computed over a sliding window tend to be smeared even if the original intensity map has sharp edges. Second, the algorithm has to be robust to noise while still be able to capture small scale textures. Lastly, one has to face the feature selection problem described as followed.

1.2. Subspace Clustering

Different textures are better captured by different features. Hence, a high-dimensional feature space is usually considered for texture segmentation. However, for each given pixel the number of relevant features is usually small, making the data very sparse. It is well-known in high-dimensional data analysis that distances computed in high-dimension are less reliable since contributions from irrelevant dimensions become significant. Moreover, in practice users may want to include all kinds of features and let the algorithm decide which set of features are relevant. A successful approach to alleviate this problem is to use subspace clustering techniques so that objects are projected onto their own relevant subspace only.

Subspace clustering is a well-studied problem in data mining community [20, 21, 22, 23, 24]. For example, in clustering of text documents, each document is typically represented by a vector containing the (suitably normalized) frequency of a large universal set of keywords. Since documents in each cluster usually concentrate on a specific topic, they share a relatively small set of keywords. Hence different clusters are characterized by different sets of keywords. Another example is supplier relationship management where each supplier, and hence the cluster that it belongs, is associated with a set of items. In the context of subspace clustering, these are examples where different clusters live in different subspaces. In [25] Jing et al. proposed an entropy weighting k-means algorithm for subspace clustering of high-dimensional sparse data. They applied the algorithm to the problems of document clustering and supplier relationship management and yielded promising results showing the usefulness of the subspace selection procedure. The model they proposed is as follows. Let U = {uki}, c = {cij} and Λ = {λij}. The objective to optimize is given by

FUcΛ=i=1nj=1mλijk=1Nukifkjcij2+γi=1nj=1mλijlogλij.

Here, f k = (f k1,..., fkm) for k = 1,2,...,N are the N given m-dimensional data points; c i = (c i1,...,cim) for i = 1,2,...,n are the n cluster centers; λij is the weight of jth feature in the ith cluster; uki is the degree of membership of the kth point to the ith cluster; and γ > 0 is a parameter controlling the entropy of the feature weights. The unknowns U and Λ are required to satisfy the following constraints:

j=1mλij=1,fori=1,,n,
i=1nuki=1,fork=1,,N,
0λij1,fori=1,,n,andj=1,,m,
uki{0,1}.fork=1,,N,andi=1,,n.

This model simultaneously minimizes the within-cluster dispersion and maximizes the entropy of the weights to stimulate more features to contribute to the identification of clusters. Setting γ = 0 results in a trivial solution where for each cluster one particular feature gets a full weight of 1. In contrast, setting γ = ∞ reduces to the original k-means where all features are equally weighted.

While subspace clustering algorithms are very useful for clustering high-dimensional data, most existing ones are designed for generic data that they do not consider any geometric properties of the segments in the spatial domain, e.g. regularity of the boundary. It is thus attractive to incorporate geometric features into the k-means subspace clustering model.

1.3. Vector-Valued Mumford-Shah Model

Among the various segmentation approaches, optimization model-based methods often give very promising results. An advantage of this approach is that one can incorporate domain knowledge into the model explicitly through rigorous mathematical modeling. This makes such kind of methods highly adaptive to deal with various kinds of segmentation problems [26, 27, 28].

The Mumford-Shah (MS) model [17] is one of the best optimization models for non-texture segmentation. A distinctive feature is that it is region-based which allows it to segment objects without sharp edges. Some promising results in segmenting objects without edges are obtained in [29, 30]. For the same reason, it can group a cluster of smaller objects into a larger object. Moreover, the geometry of the segments is regularized explicitly. Thus it is robust to noise that it does not require any pre-smoothing, which can dislocate edges. These features are very attractive for texture segmentation. It is thus tempted to further develop the Mumford-Shah model to handle high-dimensional sparse data.

In particular, we consider the vector-valued piecewise constant version of the Mumford-Shah model by Sandberg et al. [31] which seeks to minimize the following objective

FMSCc=1mi=1nj=1mΩifjxycij2dxdy+μLength(C).

The set of curves C partition the image domain Ω into n mutually exclusive segments Ωi for i = 1,2,...,n. The idea is to partition the image so that each feature fj in each segment Ωi is well-approximated by a constant cij. The goodness-of-fit is measured by the fitting term ∑m j=1Ωi|fj(x,y)-cij|2 dxdy. On the other hand, a minimum description length principle is employed which requires the curves C to be as short as possible. This increases the robustness to noise and avoids spurious segments. The parameter μ > 0 controls the trade-off between the goodness-of-fit and the length of the curves C. It can be easily shown that for each fixed C, the optimal constant cij is given by the average of fj over Ωi.

In [12], Sandberg et al. considered the application of this model to texture segmentation using Gabor features. The result showed that Gabor features are able to capture the characteristics of various kinds of textures. However, after some features have been selected (either manually or automatically by some criteria), all the selected features are used to describe all pixels. Hence, as mentioned in Section 1.2, computing similarity between pixels is problematic and erroneous. In summary, the vector-valued active contour without edges model is natural for the use of texture segmentation. Together with Gabor features as input channels, the model is able to obtain good segmentation results for certain kind of images if relevant features are selected properly. However, when different regions lay on different subspaces, this approach might fail to retrieve a good result. In this case, subspace clustering technique is indispensable.

1.4. Objective

In this paper, we propose the use of subspace clustering techniques for texture segmentation. Our first contribution is a novel model, called the Semi-Supervised Subspace Mumford-Shah model, which combines subspace clustering and Mumford-Shah segmentation techniques. However the natural unsupervised learning of a feature subspace can easily be trapped in a local solution. Our second contribution is a novel semi-supervised algorithm which is much more robust for optimizing the proposed model.

We propose to incorporate the entropy weighting k-means subspace clustering model [25] into the Mumford-Shah model to facilitate feature selection, which is important for the often high-dimensional problem of texture segmentation. The two methods combine naturally since they share the same data fidelity term. We use a soft subspace clustering approach so that relevant features are given a larger weighting and vice versa.

In most k-means type algorithms, the solution obtained is only a local minimum of the objective function and depends highly on the initial guess. This problem is even more severe in subspace clustering since the objective involves extra variables (the weights) in which the objective is not jointly convex. While such an unsupervised learning of the segmentation and the weights fails to work, an alternative is to first use a supervised approach to learn the weights based on some sample textures and then use an unsupervised method to construct the segmentation. This supervised approach can greatly reduce the “number of local minima” making the computed solution more robust to the initial guess. But it assumes that the sample segments are good representatives everywhere in the image. This can be unrealistic in many real images.

To alleviate these problems, we propose a semi-supervised approach that makes use of information derived from both the intermediate segmentation results and the regions-of-interest selected by the user to obtain the weights of the features. We show that the proposed model can separate textures which are otherwise difficult in the original full feature space.

2. Algorithm

A work flow of the Semi-Supervised Subspace Mumford-Shah method is shown in Fig. 1. The details are described in the following sections.

 figure: Fig. 1.

Fig. 1. A work flow of the Semi-Supervised Subspace Mumford-Shah method. Features such as Gabor and RGB are first extracted from the image. Then a vector-valued Subspace Mumford-Shah model is used to segment the image, where the feature weights {λij} are learned with the help of the selected ROI’s.

Download Full Size | PDF

2.1. Feature Extraction

2.1.1. Gabor Features

Gabor filters have been widely used in extracting textured image features [3, 6, 12, 11, 13, 7] since they are designed to capture local patterns in different scales and orientations. A 2-D Gabor function G(x,y) is defined as [6]:

Gxy=12πσxσyexp[12(x2σx2+y2σy2)+2πiWx],

where W is the modulation frequency and i = √-1. Then a self-similar filter dictionary can be obtained by dilations and rotations:

Gpqxy=apG(x′,y),
x=ap(xcosθ+ysinθ),
y=ap(xsinθ+ycosθ),

where a > 1, p ∊ ℤ, q ∊ ℤ, θ = /K and K is the total number of orientations. If σx, σy, W are fixed, then the parameters of a Gabor filter are the orientation q and the scale p. By varying p and q, we can obtain a set of Gabor functions {GPjqj}j. Given an image u, its Gabor feature at location (x,y) is defined to be

fjxy=Ωux̂ŷGpjqjxx̂yŷ¯dx̂dŷ,

which is the modulus of the Gabor transform of u. Here indicates the complex conjugate of G. Thus, we can obtain a set of input channels {fj}j for texture segmentation.

Gabor transforms contain many redundant information since the support of their filter responses in the frequency domain are highly overlapped. To reduce the redundancy, Manjunath and Ma [6] designed a scheme to select an appropriate subset of filters which ensures that the half-peak-magnitude support of the filter responses in the frequency spectrum touch each other. We apply this scheme in our experiments for parameter setting.

2.1.2. Features Derived from a Texture Database

A distance measure between two images in an image database has been proposed in [6]. The distance is simple to compute and has been shown to be quite effective in the context of image retrieval. We propose a novel set of features which utilizes such a distance measure. The feature vector at a pixel location is the set of distances between the local patch around the pixel and each texture patch in a texture database. The details will be discussed as follows.

We here present the distance proposed in [6]. Given a set of Gabor features {fj} derived from an image u, we define for each Gabor feature the mean and standard deviation as follows:

μj(u)=1ΩΩfjxydxdy,σj(u)=1ΩΩ[fixyμj(u)]2dxdy.

We also denote the standard deviation of μj(∙) and σj(∙), calculated over the entire image database, by std(μj) and std(σj) respectively. Let u and v be two images (which may have different sizes). Let {fj} and {gj} be the Gabor features derived from u and v respectively. The distance between u and v is defined as

duvjμj(u)μj(v)std(μj)+σj(u)σj(v)std(σj).

Now, given an image u and a texture database {vk} where each vk for k = 1,2,..., K is a texture patch, we define a feature value fk(x,y) at pixel (x,y) by

fkxy=d(u(,x,y),vk)

for k = 1,2,..., K and for (x,y) ∊ Ω. Here u(∙,∙|x,y) is a local patch in u around (x,y). In some experiments below, we combine these K features with others such as Gabor features and Laws’ features to form a large feature space. We note that a subset of 64 in the Brodatz album [32] is used as the texture database, over which the quantities std(μj) and std(σj) are computed.

2.2. Semi-Supervised Subspace Mumford-Shah Segmentation Model

In this section, we present the proposed semi-supervised model. To better explain the ideas behind, we first introduce an unsupervised version.

Let C be a set of curves partitioning the image domain Ω into n mutually exclusive segments Ωi for i = 1,2,..., n. Let c = {cij} and Λ = {λij} for 1 ≤ in and 1 ≤ jm be two sets of scalars. The Mumford-Shah image segmentation model (2) and the k-means subspace clustering model (1) can be naturally combined as follows:

FUSSMSCcΛ=i=1nj=1mλijΩifixycij2dxdy+γi=1nj=1mλijlogλij+μLength(C).

We refer this model as the Unsupervised Subspace Mumford-Shah model (USSMS). This model inherits the essential features of models (2) and (1)—it enables cluster-wise feature selection and it regularizes the geometry of the segments in the spatial domain to avoid over-fitting. In this model, we seek for a segmentation of the image such that the jth feature value within the segment Ωi is well-approximated by a constant cij, provided that the weight λij is “significant”. If λij is close to zero, then the determination of the ith segment is insensitive to the jth feature. Thus, a key feature of the model is that each segment can use a different subspace in computing the within-segment dispersion. Therefore, it opens up the possibility of using an optimal subset of features for each segment.

Unfortunately, this objective function is highly non-convex which is difficult to optimize. We found empirically that many spurious local minima can exist. We tried several standard optimization methods as well as with different initial conditions but their performance has been disappointing. The results often depend very much on the initial guess. Thus this model can be difficult to use in practice. To resolve this problem, we propose the following Semi-Supervised Subspace Mumford-Shah model (SSSMS) which amounts to the minimization of the following objective function:

FSSSMSCcΛ=(1β)FUSSMSCcΛ+β[ΩkROIki=1nj=1mλijFij+γi=1nj=1mλijlogλij].

Here ROIk is the k-th user-specified region-of-interest (ROI) which serves as a sample of the k-th segment and

Fij=ROIifixyc˜ij2dxdy

is the fitting error of the jth feature over ROIi. The optimal fitting constant is given by

c˜ij=ROIifjxydxdyROIi.

The parameter β ∊ (0,1) is to control the degree of involvement of the ROI’s. In the algorithm presented in the next section, the parameter β is gradually decreased from 1 towards 0. Thus the ROI’s eventually drop out from the objective F SSSMS which converges to its unsupervised version F USSMS. This can make the final weights depend on the image globally thus preventing any bias introduced by the ROI’s. One may think of this approach as using the ROI’s to specify an initial guess for the optimization of the unsupervised model F USSMS.

Except for the introduction of the constant |Ω|/∑k|ROIk| to normalize the difference between the size of the ROI’s and Ω, the proposed objective function is essentially a convex combination of F USSMS (C, c, Λ) and F USSMS (, , Λ) where = {ij} and is the set of curves defining the ROI’s and is therefore determined. Roughly speaking, the idea of the model is that the given ROI’s (i.e. ) allow us to estimate the weights Λ more accurately. Once the weights are in place, the problem essentially reduces to the original Mumford-Shah problem. Of course, the Mumford-Shah model itself can possess multiple local minima, but the problem is much more amenable. See [33] for global optimization of the Mumford-Shah model.

2.3. Optimization of the Objective

To optimize the objective F SSSMS, we use an alternating minimization approach (cf. the second and third boxes in Fig. 1). The main steps are as follows. The parameter β is reduced every other M steps.

  1. Compute the constants using (5).
  2. Initialize β 0 = 0.99.
  3. Initialize C 0 using e.g. k-means and initialize Λ0 as {1m}.
  4. Repeat the followings steps until Cn stabilizes.
    1. c n+1=arg minc F SSSMS(Cn,c,Λn)
    2. Λn+1=arg minΛ F SSSMS(Cn,c n+1,Λ)
    3. C n+1 =arg minC F SSSMS(C,c n+1n+1)
    4. If n is a multiple of M, then set β n+1 = 0.9βn. Otherwise, set β n+1 = βn.

The details of the implementation of Steps 4.1)–4.3) are given in the Appendix.

3. Methods

Five real-life images are used for evaluation. The endometrial tissue image is from [34]. The rat myocardium muscle image and the cuttlefish image are obtained from the World Wide Web. The zebra image is from [12]. The stem cell images are obtained within the first and second authors’ institute. Each selected image has different textures, some of them have noticeable pattern (e.g. zebra) and some have not (e.g. cuttlefish). This allows us to test the effectiveness of the proposed feature selection procedure on different types of textures.

We use a large set of features to see how the algorithm determines the optimal weights. As a result, some less informative and less discriminative features are also included. In this way, we can compare the results with and without subspace clustering. We expect that if the number of irrelevant features is increased, then the original MS model might perform worse since the distance is less reliable. However, our SSSMS model is much less sensitive to irrelevant features.

In all tests, we use the parameter setting in [6] (i.e. (σx,σy) = (1.4054,1.8544) and W = 0.4). The parameter β is set to 0.99 initially and is periodically decreased (e.g., every other 100 steps) at a rate of 0.9, i.e., β → 0.9β. The parameters μ and γ are chosen manually. The detailed setting of other parameters are reported in Section 4 individually for each example. In all examples, we normalized the feature values to between 0 and 1.

To validate our method, we compare it with the Mumford-Shah segmentation model (2) proposed in [12]. For the MS model, we set an initial segmentation u 0 to be 1 (resp. 0) within the area of the ROI which represents foreground (resp. background) and randomly set a value from {0,1} elsewhere. We also set the initial cij to be the mean feature value of the j-th feature taken over the corresponding ROIi. For the SSSMS model, we use a random assignment as the initial segmentation u 0. Both methods use the same full feature spaces. In addition, we report the result of the MS model with nonlinear structure tensor features (MSNLST) [35] as input channels and an initial segmentation as described above for the MS model. In the last example, we use the k-means clustering result as an initial segmentation for all three stem cell images. The ROI’s are selected from one of the three images and are used for all the three images for segmentation. The percentage symmetric difference between the manual segmentation and the computed segmentation is used as a measure of segmentation error. In the first four real life examples, we tried a wide range of μ for the MS and the MSNLST models and reported the one with the least segmentation error. The errors of the models of SSSMS, MS and MSNLST are shown in Table 2. All algorithms are implemented in MATLAB 7. The tests are done with a Pentium D 3GHz machine.

4. Results

4.1. Endometrial tissue image

In this test, we use an endometrial tissue image with size 225 × 225 containing some endometrial glands and stroma. In pathology research, the ratio of the area of these two kinds of tissue is crucial for aiding the diagnosis of some diseases such as hyperplasia and extraperitoneal endometriosis. To segment the image, we use the RGB channels and twelve Gabor features with 2 scales and 6 different orientations (θ ∊ {0, π/6, π/3, π/2,2π/3,5π/6}). In Fig. 2, the results show that both the MS and the MSNLST models cannot find a clear boundary while the SSSMS model can. In the second row of the figure, we also show the top channel having the largest weighting for each region. The weights of the two most significant channels for each region are shown in Table 1.

 figure: Fig. 2.

Fig. 2. Segmentation of endometrial tissue: SSSMS and MS models with three RGB channels and twelve Gabor features as input features, and MSNLST with nonlinear structure tensor features. μ = 0.04 (SSSMS) and 0.17 (MS) and 0.002 (MSNLST), γ = 0.5.

Download Full Size | PDF

Tables Icon

Table 1. The weights of the top two channels for each region.

4.2. Muscle image

In this test, we use a transmission electron microscopy image of rat myocardium with size 300 × 300. Our aim is to separate the nucleus in the upper part of the image from the striated muscle. To achieve this, we use three Gabor features with the same scale but different orientations (θ ∊ {0,π/3,2π/3}) and 4 Brodatz texture similarity measures (see Section 2.1 for details). The user specifies one ROI for each kind of cells. In Fig. 3, the results show that both the MS and the MSNLST models cannot find a clear boundary while the SSSMS model can. In the figure, we also report for each region the top channel having the largest weighting. Note that in both cases, the top channel is a Brodatz texture similarity measure.

 figure: Fig. 3.

Fig. 3. Segmentation of a rat myocardium image: SSSMS and MS models with 3 Gabor features and 4 Brodatz texture similarity measures as input features, and MSNLST with nonlinear structure tensor features. μ = 0.5 (SSSMS) and 1.5 (MS) and 0.0085 (MSNLST), γ = 0.5.

Download Full Size | PDF

4.3. Cuttlefish camouflage

In this test, we use a cuttlefish image with size 260 × 260 containing two infant cuttlefishes where one is located in the center and (part of) the other is located in the upper left corner of the image. The cuttlefishes in the image can hardly be seen in the RGB space. It is because cuttlefish has the ability to regulate its disruptive skin patterns for camouflage on benthic substrate [36] such as sand in this case. To segment the image, we use 6 Gabor features with the same scale but different orientations (θ ∊ {0,π/6,π/3,π/2,2π/3,5π/6}), 14 Laws’ Mask features [4, 37] and 64 Brodatz texture similarity measures (see Section 2.1 for details). To optimize the performance of both the MS and the SSSMS models, we first select the top five channels which have the highest mean intensity differences between two selected ROI’s. In Fig. 4, the results show that the SSSMS model is more capable of distinguishing the two camouflaged cuttlefishes from the sand substrate than the MS and the MSNLST models. In the figure, we also report for each region the top channel having the largest weighting. Note that in both cases, the top channel is a Brodatz texture similarity measure.

4.4. Zebra image

The purpose of this experiment is to test all the three models on an image which has been used by many researchers to evaluate texture segmentation methods and serves as a benchmark for direct comparison with other models. In this test, we use a 200 × 200 image of two zebras with noticeable contrast in textures between object and background. We use 6 Gabor filters with the same scale and different orientations (θ ∊ {0, π/6, π/3, π/2, 2π/3, 5π/6}). The results are shown in Fig. 5. As expected, all of the models are able to obtain a good segmentation result. But the proposed model has an additional advantage that it can reveal the dominated features of each segment. For example, in the zebra region, the Gabor features with θ ∊ {π/6, π/3, π/2} have the highest weight. In the grassland region, all features have a similar weight. These observations are consistent with the textures of the zebra and the grassland.

 figure: Fig. 4.

Fig. 4. Segmentation of two cuttlefishes: SSSMS and MS models with 84 input features before feature selection process, and MSNLST with nonlinear structure tensor features. μ = 0.1 (both SSSMS and MS) and 0.006 (MSNLST), γ = 0.5.

Download Full Size | PDF

The percentage errors of the segmentations in Fig. 2, 3,4 and 5 (with respect to some manual segmentations) are shown in Table 2. We remark that each result of the MS and the MSNLST models in Fig. 2, 3, 4 and 5 has been optimized by choosing a parameter μ that minimizes the segmentation error.

Tables Icon

Table 2. Percentage errors (percentage symmetric difference) w.r.t. manual segmentations.

4.5. Stem cell images

In practice, given a set of images with similar textures, one may want to select ROI’s once from one image and use it for all images for segmentation. The aim of this experiment is to show whether this strategy works in our case. In this test, we use three mesenchymal stem cell images with size 305×305 containing some differentiated cells and some undifferentiated cells under different conditions. The ratio of the area of these two kinds of cells is crucial for understanding cell differentiation mechanisms. To segment the images, we use the red channel and six Gabor features with same scale but different orientations (θ ∊ {0,π/6,π/3,π/2,2π/3,5π/6}) and specify one ROI for each kind of cells in one image. In Fig. 6, the results show that the SSSMS model can find a clear boundary for all three images. In the figure, we also report the percentage symmetric difference between the results and our manual segmentations. Since the ROI’s were selected from the leftmost image, it has the minimal error as expected. However, the results of the other two images are still promising.

 figure: Fig. 5.

Fig. 5. Segmentation of two zebras: SSSMS and MS models with 6 Gabor transforms as input features, and MSNLST with nonlinear structure tensor features. μ = 0.02 (SSSMS) and 0.2 (MS) and 0.05 (MSNLST), γ = 50.

Download Full Size | PDF

5. Conclusion

In this paper, we propose a model which incorporates the concept of subspace clustering into the Mumford-Shah segmentation model and propose an improved semi-supervised optimization algorithm to optimize the model. The algorithm uses the ROI’s with a decreasing degree of involvement to avoid the problem of over-specification of the ROI’s. In our experiments, we show that even when the images have a very low contrast in textures, the algorithm can give much better results than the original Mumford-Shah model since irrelevant features are filtered which in effect increases the contrast between different segments.

 figure: Fig. 6.

Fig. 6. Segmentation of three stem cell images: SSSMS model with red channel and 6 Gabor features as input features. μ = 0.3, γ = 5.

Download Full Size | PDF

A. Implementation Details of Alternating Minimization Algorithm

In Section 2.3, there are three subproblems that need to be solved alternatively. For the sub-problem in Step 4.1), the optimal cij is simply the mean feature value of the jth feature taken over the ith segment, i.e.,

cij=ΩifjxydxdyΩi=ΩχixyfjxydxdyΩχixydxdy

where χi equals 1 in Ωi and 0 otherwise.

For the subproblem in Step 4.2), the optimal weight λij is given by the following closed-form formula:

λij=exp(Dijγ)k=1mexp(Dikγ),

where

Dij=(1β)Ωifjxycij2dxdy+βΩFijkROIk.

This formula can be easily derived using some arguments similar to those used in the proof of Theorem 1 in [25]. Thus the optimal weights are easily computed for any given segmentation and ROI’s.

For the subproblem in Step 4.3), a closed-form solution does not exist. We present a fast global minimization algorithm to solve it. The objective function F SSSMS is non-convex and therefore may possess many local minima (even for fixed constants c and fixed weights Λ). To globally minimize the objective F SSSMS, one can use the Stochastic Level Set Method proposed in [33]. Here we devise a fast algorithm for minimizing F SSSMS w.r.t. C based on the results of Chan et al. [38], Chambolle [39], Aujol et al. [40], and Bresson et al. [41]. An advantage of the method is that a global minimum w.r.t. the curves C for fixed c is obtained. Moreover, using a dual formulation, a fast iterative method is devised. The method works extremely well in the case of two-phase segmentation.

The idea is to first reformulate the non-convex problem into a convex minimization problem [38]:

min0u1{Ωrudxdy+μΩudxdy},

where

rxy=j=1m{λ1j[c1jfjxy]2λ2j[c2jfjxy]2}.

Then the problem is solved by solving its dual problem using a splitting method [40] and an iterative scheme [39]. It has been shown in [38] that by thresholding the optimal solution u of (6) at almost every t ∊ [0,1], we obtain a global optimal segmentation of F SSSMS (in the continuous setting). Here we fix t = 0.5. We refer the readers to [41] for details.

The algorithm for optimizing F SSSMS w.r.t. the segmentation with fixed c and Λ is outlined below. Here, θ is a small constant controlling the closeness between u and an auxiliary variable v, δ = 1/8 is a step size to ensure convergence, div is the discrete divergence operator, and ∇ is the discrete gradient operator.

  1. Initialize u 0. Set it to a binary image representing an initial segmentation obtained by e.g. the k-means clustering algorithm or set it to any available estimate obtained from e.g. the previous alternating minimization step or random assignment.
  2. Initialize p 0(x,y) = (0,0) for each pixel (x,y).
  3. Repeat the following steps until un converges. The segmentation is given by the binary image χn.
χn+1xy={1ifunxy>0.5,0otherwise
rn+1xy=j=1m{λ1j[c1jfjxy]2λ2j[c2jfjxy]2}
vn+1xy=min{max{un+1xyθrn+1xy,0},1}
pn+1xy=pnxy+δ[(divpnvn+1μθ)]xy1+δ[(divpnvn+1μθ)]xy
un+1xy=vn+1xyμθ(divpn+1)xy.

In practice, we can often speed up the convergence of the outer alternating minimization in Section 2.3 by iterating only once in the inner loop in the Step 3 of the above algorithm. In this case, the alternating minimization is equivalent to the above algorithm but with the update of c, Λ and β inserted immediately after the update χ n+1.

Acknowledgements

This work was supported in part by the Biomedical Research Council of A*STAR (Agency for Science, Technology and Research), Singapore, and in part by the National University of Singapore under Grant R-146-000-116-112.

References and links

1. A. Kneller, “The new age of bioimaging,” in “Paradigm,” (2006), pp. 18–25.

2. P. Petrou and P. G. Sevilla, Dealing with Texture (Wiley, 2006). [CrossRef]  

3. D. Gabor, “Theory of communication,” J. IEEE 93, 429–459 (1946).

4. K. Laws, “Texture energy measures,” Proc. of Image Understanding Workshop pp. 47–51 (1979).

5. A. Jain and F. Farrokhnia, “Unsupervised texture segmentation using Gabor filters,” Pattern. Recog. 24, 1167–1186 (1991). [CrossRef]  

6. B. Manjunath and W. Ma, “Texture features for browsing and retrieval of image data,” IEEE Trans. Pattern Anal. Mach. Intell. 18, 837–842 (1996). [CrossRef]  

7. B. Sharif, A. Ahmadian, M. Oghabian, and N. Izadi, “Texture segmentation of endometrial images for aiding diagnosis of hyperplasia,” Proceedings of the International Conference on Computer as a Tool 2, 983–986 (2005).

8. M. Datar, D. Padfield, and H. Cline, “Color and texture based segmentation of molecular pathology images using HSOMs,” Proceedings of the 5th IEEE International Symposium on Biomedical Imaging: From Nano to Macro pp. 292–295 (2008). [CrossRef]  

9. S. Belongie, C. Carson, H. Greenspan, and J. Malik, “Color- and texture-based image segmentation using EM and its application to content-based image retrieval,” Proc. IEEE Conf. on Computer Vision p. 675 (1998).

10. A. Barbu and S. Zhu, “Multigrid and multi-level Swendsen-Wang cuts for hierarchic graph partition,” Proc. IEEE Conf. on Computer Vision and Pattern Recognition 2, 731–738 (2004).

11. N. Paragios and R. Deriche, “Geodesic active regions and level set methods for supervised texture segmentation,” Int. J. Comput. Vision 46, 223–247 (2002). [CrossRef]  

12. B. Sandberg, T. Chan, and L. Vese, “A level-set and Gabor-based active contour algorithm for segmentation textured images,” Cam report, UCLA (2002).

13. C. Sagiv, N. Sochen, and Y. Zeevi, “Texture segmentation via a diffusion-segmentation scheme in the Gabor feature space,” Proc. of the 2nd Intl. Workshop on Texture Analysis and Synthesis (2002).

14. M. Rousson, T. Brox, and R. Deriche, “Active unsupervised texture segmentation on a diffusion based feature space,” Proc. of the 2003 IEEE Computer Vision and Pattern Recognition (2003).

15. T. Reed, “A review of recent texture segmentation and feature extraction techniques,” CVGIP: Image Understanding 57, 359–372 (1993). [CrossRef]  

16. M. Haindl and S. Mikeš, “Unsupervised texture segmentation,” in “Patten Recognition Techniques, Technology and Applications,” , P.-Y. Yin, ed. (I-Tech, 2008), pp. 227–248.

17. D. Mumford and J. Shah, “Optimal approximation by piecewise smooth functions and associated variational problems,” Commun. Pure Appl. Math. 42, 577–685 (1989). [CrossRef]  

18. S. Beucher and C. Lantuéjoul, Use of watersheds in contour detection,” Proc. International Workshop on Image Processing: Real-time Edge and Motion Detection/Estimation (1979). [PubMed]  

19. M. Kass, A. Witkin, and D. Terzopoulos, “Snakes: Active contour models,” International Journal of Computer Vision 1, 321–331 (1988). [CrossRef]  

20. R. Agrawal, J. Gehrke, D. Gunopulos, and P. Raghavan, “Automatic subspace clustering of high dimensional data for data mining applications,” Proc. of the 1998 ACM SIGMOD International Conference on Management of Data pp. 94–105 (1998).

21. C.-H. Cheng, A. W. Fu, and Y. Zhang, “Entropy-based subspace clustering for mining numerical data,” Proc. of the 5th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining pp. 84–93 (1999). [CrossRef]  

22. C. Aggarwal and P. Yu, “Finding generalized projected clusters in high dimensional spaces,” Proc. of the 2000 ACM SIGMOD International Conference on Management of Data pp. 70–81 (2000).

23. J. Yang, W. Wang, H. Wang, and P. Yu, “δ-clusters: capturing subspace correlation in a large data set,” Proc. of 18th Interational Conference on Data Engineering pp. 517–528 (2002). [CrossRef]  

24. J. Friedman and J. Meulman, “Clustering objects on subsets of attributes,” J. R. Statist. Soc. B 66, 815–849 (2004). [CrossRef]  

25. L. Jing, M. Ng, and J. Huang, “An entropy weighting k-means algorithm for subspace clustering of high-dimensional sparse data,” IEEE Trans. Knowledge and Data Engineering 19, 1026–1041 (2007). [CrossRef]  

26. C. Yap and H. Lee, “Identification of cell nucleus using a Mumford-Shah ellipse detector,” Proc. of ISVC 1, 582–593 (2008).

27. W. Yu, H. Lee, S. Hariharan, W. Bu, and S. Ahmed, “Level set segmentation of cellular images based on topological dependence,” Proc. of ISVC 1, 540–551 (2008).

28. W. Zhu, T. Chan, and S. Esedog¯lu, “Segmentation with depth: A level set approach,” SIAM J. Sci. Comput. 28, 1957–1973 (2006). [CrossRef]  

29. T. F. Chan and L. A. Vese, “Active contours without edges,” IEEE Tran. Image Process. 10, 266–277 (2001). [CrossRef]  

30. L. A. Vese and T. F. Chan, “A multiphase level set framework for image segmentation using the Mumford and Shah model,” Int. J. Comput. Vision 50, 271–293 (2002). [CrossRef]  

31. T. Chan, B. Sandberg, and L. Vese, “Active contours without edges for vector-valued images,” Journal of Visual Communication and Image Representation 11, 130–141 (2000). [CrossRef]  

32. P. Brodatz, Textures: A photographic album for artists and designers (Dover, New York, 1996).

33. Y. Law, H. Lee, and A. Yip, “A multiresolution stochastic level set method for Mumford-Shah image segmentation,” IEEE Transactions on Image Processing 17, 2289–2300 (2008). [CrossRef]   [PubMed]  

34. L. Roberts, J. Redan, and H. Reich, “Extraperitoneal endometriosis with catamenial pneumothoraces: A review of the literature,” Journal of the Society of Laparoendoscopic Surgeons 7, 371–375 (2003). [PubMed]  

35. T. Brox, M. Rousson, R. Deriche, and J. Weickert, “Unsupervised segmentation incorporating color, texture, and motion,” Proc. of the Intl. Conf. on Computer Analysis of Images and Patterns pp. 353–360 (2003). [CrossRef]  

36. C. Chiao and R. Hanlon, “Cuttlefish camouflage: Visual perception of size, contrast and number of white squares on artificial checkerboard substrata initiates disruptive coloration,” The Journal of Experimental Biology 204, 2119–2125 (2001). [PubMed]  

37. M. Rachidi, A. Chappard, C. Marchadier, C. Gadois, E. Lespessailles, and C. L. Benhamou, “Application of Laws’ masks to bone texture analysis: An innovative image analysis tool in osteoporosis,” Proceedings of the 5th IEEE International Symposium on Biomedical Imaging: From Nano to Macro pp. 1191–1194 (2008). [CrossRef]  

38. T. F. Chan, S. Esedoglu, and M. Nikolova, “Algorithms for finding global minimizers of denoising and segmentation models,” SIAM J. Appl. Math. 66, 1632–1648 (2006). [CrossRef]  

39. A. Chambolle, “An algorithm for total variation minimization and applications,” J. Math. Imaging Vision 20, 89–97 (2004). [CrossRef]  

40. J.-F. Aujol, G. Gilboa, T. Chan, and S. Osher, “Structure-texture image decomposition - modeling, algorithms, and parameter selection,” Int. J. Comput. Vis. 67, 111–136 (2006). [CrossRef]  

41. X. Bresson, S. Esedoglu, P. Vandergheynst, J.-P. Thiran, and S. Osher, “Fast global minimization of the active contour/snake model,” J. Math. Imaging Vis. 28, 151–167 (2007). [CrossRef]  

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (6)

Fig. 1.
Fig. 1. A work flow of the Semi-Supervised Subspace Mumford-Shah method. Features such as Gabor and RGB are first extracted from the image. Then a vector-valued Subspace Mumford-Shah model is used to segment the image, where the feature weights {λij } are learned with the help of the selected ROI’s.
Fig. 2.
Fig. 2. Segmentation of endometrial tissue: SSSMS and MS models with three RGB channels and twelve Gabor features as input features, and MSNLST with nonlinear structure tensor features. μ = 0.04 (SSSMS) and 0.17 (MS) and 0.002 (MSNLST), γ = 0.5.
Fig. 3.
Fig. 3. Segmentation of a rat myocardium image: SSSMS and MS models with 3 Gabor features and 4 Brodatz texture similarity measures as input features, and MSNLST with nonlinear structure tensor features. μ = 0.5 (SSSMS) and 1.5 (MS) and 0.0085 (MSNLST), γ = 0.5.
Fig. 4.
Fig. 4. Segmentation of two cuttlefishes: SSSMS and MS models with 84 input features before feature selection process, and MSNLST with nonlinear structure tensor features. μ = 0.1 (both SSSMS and MS) and 0.006 (MSNLST), γ = 0.5.
Fig. 5.
Fig. 5. Segmentation of two zebras: SSSMS and MS models with 6 Gabor transforms as input features, and MSNLST with nonlinear structure tensor features. μ = 0.02 (SSSMS) and 0.2 (MS) and 0.05 (MSNLST), γ = 50.
Fig. 6.
Fig. 6. Segmentation of three stem cell images: SSSMS model with red channel and 6 Gabor features as input features. μ = 0.3, γ = 5.

Tables (2)

Tables Icon

Table 1. The weights of the top two channels for each region.

Tables Icon

Table 2. Percentage errors (percentage symmetric difference) w.r.t. manual segmentations.

Equations (28)

Equations on this page are rendered with MathJax. Learn more.

F U c Λ = i = 1 n j = 1 m λ ij k = 1 N u ki f kj c ij 2 + γ i = 1 n j = 1 m λ ij log λ ij .
j = 1 m λ ij = 1 , for i = 1 , , n ,
i = 1 n u ki = 1 , for k = 1 , , N ,
0 λ ij 1 , for i = 1 , , n , and j = 1 , , m ,
u ki { 0,1 } . for k = 1 , , N , and i = 1 , , n .
F MS C c = 1 m i = 1 n j = 1 m Ω i f j x y c ij 2 dxdy + μ Length ( C ) .
G x y = 1 2 π σ x σ y exp [ 1 2 ( x 2 σ x 2 + y 2 σ y 2 ) + 2 πiW x ] ,
G pq x y = a p G ( x′ , y ) ,
x = a p ( x cos θ + y sin θ ) ,
y = a p ( x sin θ + y cos θ ) ,
f j x y = Ω u x ̂ y ̂ G p j q j x x ̂ y y ̂ ¯ d x ̂ d y ̂ ,
μ j ( u ) = 1 Ω Ω f j x y dxdy , σ j ( u ) = 1 Ω Ω [ f i x y μ j ( u ) ] 2 dxdy .
d u v j μ j ( u ) μ j ( v ) std ( μ j ) + σ j ( u ) σ j ( v ) std ( σ j ) .
f k x y = d ( u ( , x , y ) , v k )
F USSMS C c Λ = i = 1 n j = 1 m λ ij Ω i f i x y c ij 2 dxdy + γ i = 1 n j = 1 m λ ij log λ ij + μ Length ( C ) .
F SSSMS C c Λ = ( 1 β ) F USSMS C c Λ + β [ Ω k ROI k i = 1 n j = 1 m λ ij F ij + γ i = 1 n j = 1 m λ ij log λ ij ] .
F ij = ROI i f i x y c ˜ ij 2 dxdy
c ˜ ij = ROI i f j x y dxdy ROI i .
c ij = Ω i f j x y dxdy Ω i = Ω χ i x y f j x y dxdy Ω χ i x y dxdy
λ ij = exp ( D ij γ ) k = 1 m exp ( D ik γ ) ,
D ij = ( 1 β ) Ω i f j x y c ij 2 dxdy + β Ω F ij k ROI k .
min 0 u 1 { Ω rudxdy + μ Ω u dxdy } ,
r x y = j = 1 m { λ 1 j [ c 1 j f j x y ] 2 λ 2 j [ c 2 j f j x y ] 2 } .
χ n + 1 x y = { 1 if u n x y > 0.5 , 0 otherwise
r n + 1 x y = j = 1 m { λ 1 j [ c 1 j f j x y ] 2 λ 2 j [ c 2 j f j x y ] 2 }
v n + 1 x y = min { max { u n + 1 x y θ r n + 1 x y , 0 } , 1 }
p n + 1 x y = p n x y + δ [ ( div p n v n + 1 μθ ) ] x y 1 + δ [ ( div p n v n + 1 μθ ) ] x y
u n + 1 x y = v n + 1 x y μθ ( div p n + 1 ) x y .
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.