Learning to Segment from a Few Well-Selected Training Images

Learning to Segment from a Few Well-Selected Training Images Alireza Farhangfar FARHANG @ CS . UALBERTA . CA Russell Greiner GREINER @ CS . UALBERTA ...
Author: Kristian Morton
1 downloads 0 Views 935KB Size
Learning to Segment from a Few Well-Selected Training Images

Alireza Farhangfar FARHANG @ CS . UALBERTA . CA Russell Greiner GREINER @ CS . UALBERTA . CA Csaba Szepesv´ari SZEPESVA @ CS . UALBERTA . CA Department of Computing Science, University of Alberta, Edmonton, Alberta, Canada, T6G 2E8 Keywords: Active learning, classification, image segmentation, discriminative random field

Abstract We address the task of actively learning a segmentation system: given a large number of unsegmented images, and access to an oracle that can segment a given image, decide which images to provide, to quickly produce a segmenter (here, a discriminative random field) that is accurate over this distribution of images. We extend the standard models for active learner to define a system for this task that first selects the image whose expected label will reduce the uncertainty of the other unlabeled images the most, and then after greedily selects, from the pool of unsegmented images, the most informative image. The results of our experiments, over two real-world datasets (segmenting brain tumors within magnetic resonance images; and segmenting the sky in real images) show that training on very few informative images (here, as few as 2) can produce a segmenter that is as good as training on the entire dataset.

1. Introduction Many imaging tasks involve segmentation. For example, given a Magnetic Resonance (MR) image of the brain, it is important to find and segment any tumor region present. Many effective imaging systems involve a number of parameters that have to be adjusted; some of these systems therefore include a learning component that can learn effective parameters from a set of labeled (that is, segmented) images. In general, these systems require a large number of such labeled images to produce an effective segmenter. Fortunately, there are often a large number of available images — perhaps on the web, or in clinical databases. UnforAppearing in Proceedings of the 26th International Conference on Machine Learning, Montreal, Canada, 2009. Copyright 2009 by the author(s)/owner(s).

tunately, most such images are unlabeled, and worse, it can be expensive to obtain the labels (as this may require paying a medical doctor to label each image, which is costly in terms of both time and money). This often limits the amount of training data available, which can lead to an inferior segmentation system. An active learning process tries to address this problem by identifying which of the unlabeled images should be labeled. Section 2 overviews this body of work, to help motivate our approach. Section 3 then presents our actual active learning algorithm, L MU. It also describes how the underlying performance system — here using a Discriminative Random Field — segments the images. Section 4 shows the results of our experiments using this active learner on two real-world datasets: Finding the sky in the Geometric Context dataset [SKH08] and segmenting tumors within MR images of human brains [LWB+ 08]. In particular, we compare the segmentation performance (on hold-out images) of a segmenter trained on all labeled images, versus one trained using only the first k images selected by our active learner, for various values of k. (We also consider segmenters learned from k randomly-selected images.) We find that, surprisingly, the segmenter based on only k = 2 well-selected images is typically as good as the one based on all (here, 85) images! But only if the images are wellselected; the segmenter based on k = 2 random images is typically considerably inferior, as is one based on a larger number of random images. (See also our webpage [GFS09] for additional information — e.g., timing information, studies using the simpler CalTech101 images [FFFP04], etc.)

2. Related Work Most supervised learning systems are passive, in that they produce a classifier based only on the existing corpus of labeled training instances. By contrast, an active learner is able to extend this set of labeled instances by sequentially identifying an unlabeled instance and obtaining its label from an oracle, then adding the resulting labeled in-

Learning to Segment from a Few Well-Selected Training Images

stance to the training data. There are many results on active learning, most of which relate to the standard supervised learning framework, where the label for an instance x ∈ X is drawn from a small set of possible labels y ∈ Y (e.g., Y ={+1, -1}). Some of these algorithms select the instance x whose label y is most uncertain, based on the Pˆ (D) (y|x) probability distribution obtained by the current training data D. Freund et al. [FSST97] select the instance that has maximal disagreement from the current committee of classifiers. Many researchers, working with support vector machines, select the instance closest to the boundary [TK02, SC00, CCS00]. Lewis and Gale [LG94] use a probabilistic classifier, and select the most uncertain instance — i.e., the one whose conditional probability of +class given the features is closest to 0.5, assuming binary labels. Our L MU system (Section 3.2) uses a segmentationanalogue to this basic approach as part of its process. This most uncertain approach typically works well if the current parameters nicely approximate the conditional probability distribution of the entire data. Unfortunately, these parameters are often problematic as they are based on a very small training set. As the goal is to learn a classifier that works well on the distribution over X, it makes sense to consider the marginal distribution over unlabeled data P ( X ). This motivates a second class of approaches that use the pool of unlabeled instances. Some active learners use clustering algorithms to first group the unlabeled data. Nguyen and Smeulders [NS04] repeatedly cluster the unlabeled data and then request labels of one representative from each cluster. Xu et al. [XYT+ 03] request labels for the instances near the centers of cluster lying within the margin of support vector machine. Other systems explicitly use P(X); eg, Cohn et al. [CGJ96] and Zhang and Chen [ZC02] estimated and then used the density P ( X ) as weights for unlabeled data. Roy and McCallum [RM01] selected instances that reduce expected error over unlabeled data. Guo and Greiner [GG05] proposed an algorithm that selects the instance that provides the maximum conditional mutual information about the labels of unlabeled instances. Our L MU system also incorporates (a segmentation-analogue to) this idea, for its first iteration. The underlying domain of X is quite arbitrary, ranging from simple binary tuples to feature sets obtained from natural language texts. Some active learners deal with images, which are complicated due to their large dimensionality. Vijayanarasimhan and Grauman [VG08] proposed a framework to actively recognize objects inside the image (from a small predefined set of labels), using a mixture of weakly and strongly labeled images. Their method selects the partially labeled or unlabeled instance that minimizes the expected risk of other instances. The system by Collins

et al. [CDLFF08] actively selects the most uncertain instance, then uses boosting techniques to train a decision stump classifier to detect and recognize various objects. As noted above, most of the existing systems have been used primarily to learn a classifier that maps each instance to a simple label; even the imaging work mentioned above has focused on mapping each image to one of small set of labels Y . As images can often be recognized based only on a small set of extracted features, object recognition corresponds to a typical machine learning problem. There have been relatively little active learning research related to image segmentation, which requires producing a more complicated label: such systems map an image of n × m pixels to n × m individual (correlated) pixel-labels; if each pixellabel is binary, this means that Y = {+1, −1}n×m . Moreover, these pixel-labels are not independent of one another (e.g., if one pixel is a tumor, it is more likely that its neighbors are, as well). This forces a segmenter to consider the entire n × m image as an instance, rather than label one pixel at a time. Hence, notions like uncertainty and information content must be defined for the entire image.

3. Implementation This section overviews our basic system. We first describe the training and performance of the underlying segmentation system, based on Discriminative Random Fields (DRFs) [KH03]. (A DRF is a version of a conditional random field (CRF) that is designed to deal with variables that are organized in a 2-dimension grid.) We then summarize our L MU system, which actively learns the parameters for the DRF. For notation, we let I represent the set of n × m image pixels, x = {xi | i ∈ I} be an observed input image, where each xi is a vector describing pixel i (perhaps its intensity and texture) and y = {yi | i ∈ I} is the corresponding joint set of labels over all pixels of the image. We will assume the segmentation is binary over n × m images, where each yi ∈ {−1, +1} (e.g., is this pixel tumorous vs healthy), and the overall output y is n × m such bits. At any time, our system has a pool of unsegmented instances U , as well as a (possibly empty) pool of segmented instances L.1 Our active learner will sequentially select an unsegmented image u ∈ U , obtain its label (recall this label is a set of |u| bits — one for each pixel in u), and then move this now-labeled image from U to L. 1 Later, when we describe our experiments, we will also have a set of unlabeled test images T . Note that these sets, T , L, U , are disjoint. Here, and below, we will use labeled as a synonym for segmented and unlabeled for unsegmented.

Learning to Segment from a Few Well-Selected Training Images

3.1. Discriminative Random Field A discriminative random field (DRF) is a model of the conditional probability of a set of labels y given the observations x, here given by Pθ ( y | x) =

X 1 exp Φw (yi , x) Zθ (x) i∈I

In our active learning framework, we need to consider the effect of adding one more image from unlabeled set to the training set. Since the label y (u) of an unlabeled image x(u) is not known before presenting it to an expert, we use a conditional entropy term to express the likelihood associated with the unlabeled image as X RLu ( θ ) = Pθ ( y | x(u) ) log Pθ ( y | x(u) ) y

 +

XX

Ψv (yi , yj , x)

(1)

i∈I j∈Ni

where Ni are the 4 neighbors of i (north, east, south, west); Φw (yi , x) = log( 1+exp(−y1i wT hi (x)) ) is the association potential of pixel i that uses the association parameters w obtained from training data, and hi (x) is the feature vector for pixel i;2 Ψv (yi , yj , x) = yi yj vT µij (x) is the interaction potential that captures the spatial correlation with neighboring pixels using the interaction parameters v obtained from the training data; µij (x) = xi − xj is the difference between feature vectors in pixel i and j; and θ = [w, v] are the model parameters. The normalizing factor Zθ (x) =

X y

exp

X

Φw (yi , x)

i∈I

 +

XX

Ψv (yi , yj , x) (2)

i∈I j∈Ni

The overall objective function now consists of two terms: the first term deals with all labeled images in training set L and the second term is for an unlabeled image u. RLL+u ( θ )

=

RLL ( θ ) + γ RLu ( θ )

(4)

where the γ ∈ < parameter trades-off the two factors. We ∗ then seek the parameters θL+u that maximize Equation 4. 3.1.2 Useful Approximation: Given the lattice neighborhood structure of the DRF, it is intractable to compute the normalizing factor Zθ (x) in Equation 2. Following [KH03], we therefore incorporate the pseudolikelihood approximation, which assumes that the joint probability distribution of all pixel labels y can be approximated by the product of “local probabilities” of each pixel, which is based on only the observations of xi and the labels of the neighboring nodes yNi : Y Pˆθ (y|x) ≈ Pˆθ (yi |yNi , x) i∈I

insures that the DRF produces a probability. Given a set of parameters θ, we can compute the label y∗ for a given image x: y∗ = arg maxy Pθ ( y | x) (see Section 3.1.2). The challenge is learning the best values for these parameters, θ∗ . The rest of this subsection discusses how to compute these optimal parameters from a set of labeled images L and then for this L and a single unlabeled image u; it then provides a useful approximation to this computation, to avoid its inherent intractability. 3.1.1 Training: Typical supervised DRF training involves finding the parameters ∗ θL = arg max RLL ( θ ) θ

`∈L

Here, this hi (x) is just the information in xi . In general, it could also include information from some adjacent pixels, via some smoothing operator, etc.

1 exp ( Φw (yi , x) (5) zi (x) X + Ψv (yi , yj , x))

=

j∈Ni

where this zi (x) is a “local normalizing term”, which deals only with the ith pixel. Using the approximation in Equation 5, the entropy regularization term for (respectively) training set L and a single unlabeled image u is: X X (`) (`) (`) c L( θ ) = RL log Pˆθ (yi |yNi , xi ) (6) `∈L i∈I(`)

(3)

that maximize the log of the posterior probability over training set of labeled images L X RLL ( θ∗ ) = log Pθ∗ ( y(`) | x(`) ) 2

Pˆθ (yi |yNi , x)

c u( θ ) RL

=

XX

(u) (u) Pˆθ (yi |yNi , xi )

×

(7)

i∈I yi (u) (u) log Pˆθ (yi |yNi , xi )

Following Equation 4, we can combine these to form a single objective that combines the effect of the many labeled images L and the one unlabeled data u: c L+u ( θ ) RL

=

c L ( θ ) + γ RL c u( θ ) RL

(8)

Learning to Segment from a Few Well-Selected Training Images

In our experiments, we set γ = 1. We also used conjugate gradient to optimize the objective function in Equation 8. c u ( θ ) term from Equation 7 requires the label for The RL (u) unlabeled data that is not yet available, i.e., yNi is not known. An inference step is added to estimate the labels based on current parameters. The inference is based on iterative conditional probability (ICM) [Bes86], which sets the label for pixel as the maximum posterior probability:

approach does not consider the distribution over X, which means knowing more about this “boundary” point might not help identify that much wrt this X. This suggests an alternative approach: select the instance that would provide the maximum information about the labels of remaining unlabeled instances — i.e., the instance that most reduces the uncertainty (RU) of the other unlabeled images U − {u}:

yi∗ = arg max Pθ ( yi | yNi , x) yi

RU(U, L) =

where for each pixel i, we assume that the labels of its neighbors yNi are fixed to their current estimate. We then use yi∗ to compute the label for its neighbors. We repeat this process until every yi converges to its final value.

arg max H( YU | XU , L ) − u∈U

H( YU | XU , L + x(u) )

arg min H( YU | XU , L + x ) (13) u∈U X = arg min H( Y(v) | x(v) , L + x(u) )(14) =

u∈U

3.2. The L MU Active Learning Algorithm ≈ arg min

At each time, given U and L, our L MU active learner needs to select which unlabeled image u ∈ U to give to the oracle for labeling. (Recall that this now-labeled u will then be added to the set of labeled images L.) One option is to choose the most uncertain image: MU(U, L)

arg max H( Y(u) | x(u) , L )

=

u∈U

where H( Y | x, L )=−

X

P ( y | x, L ) log P ( y | x, L ) (9)

y∈Y

≈−

X

PˆθL (y|x) log PˆθL (y|x)

(10)

y∈Y

approximates the conditional entropy of the label Y given the image observations x, based on the current conditional probability, which is based on the training data L. (To explain Equation 10: As we use the data L to produce the parameters θL , we identify P ( y | x, L ) with PˆθL (y|x); see Equation 3 and relevant approximation.) The summation in Equation 10 is over all |Y | = 2m×n possible binary assignments to n × m pixels. We approximate this as the simple sum of the entropies of the labels yi of each pixel of the image, i ∈ I: ˆ Y | x, L ) = H( ˆ Y | x, θL ) = H( i X X h (11) − PˆθL (yi |xi ) log PˆθL (yi |xi ) i∈I yi ∈{±1}

We view this most-uncertain instance MU(U, L) as being at the boundary between positives and negative instances. Having the label for such boundary points will help us to define the boundary more precisely and consequently increase the classification accuracy. Of course, our active learner has to start with an empty L = {}; here the associated θ{} parameters are problematic. Moreover, this

(12)

(u)

u∈U

v∈U ;v6=u

X

ˆ Y(v) | x(v) , θ H( L+x(u) ) (15)

v∈U ;v6=u

Equation 13 follows from the observation that the first term of Equation 12 is constant for all instances u; Equation 14 uses the fact that the entropy of the set of independent images is just their sum; and Equation 15 again uses the apˆ defined in Equation 11. proximate conditional entropy H (This θL+x(u) is the solution to Equation 8.) Note this RU approach works even for L = {}; of course, it is computationally more expensive than MU approach. Our actual L MU system uses both approaches: Given a set of unlabeled images U and no labeled instances L = {}, it first uses RU to find the first instance u1 to label, then sets L = {u1 }. Thereafter, it used MU to find the second, third, and further images. See Figure 1.

4. Experiments To investigate the empirical performance of our active learning algorithm, we conducted a set of experiments on two challenging real-world problems: Finding the sky in the geometric context dataset and segmenting tumors in medical images. We also ran a scaling study, to see the influence of the size of each image on the number of images required to obtain good performance. 4.1. Finding the Sky in Color Images The geometric context dataset [SKH08] is a collection of 125 images, many very cluttered, that span a variety of natural, urban, and suburban sceneries. Our goal here is to find the sky within these images. Figure 2 shows three instances of images in this dataset. This task is challenging since the sky could be blue or white, clear or overcast, and worse, many scenes contain both sky and ocean, which are

Learning to Segment from a Few Well-Selected Training Images L MU( U : unsegmented images ) L := {} % L is initially empty % Compute u1 = RU(U, {}) for each unlabeled image u ∈ U c u( θ ) θu := arg maxθ RL % Equation 7 su := 0 for each other unlabeled image v ∈ U , v 6= u ˆ Y(v) | x(v) , θu ) % Equation 11 su += H( u1 := arg minu su y(u1 ) := Oracle(x(u1 ) ) L := (hx(u1 ) , y(u1 ) i) c L( θ ) θL := arg maxθ RL

% Get label % Equation 6

U := U − {u1 } for i = 2, ... % ui := MU(U, L) for each u ∈ U ˆ Y(u) | x(u) , θL ) tu = H( ui := arg maxu tu y(ui ) := Oracle(x(ui ) ) L := L + (hx(ui ) , y(ui ) i) c L( θ ) θL := arg maxθ RL

Figure 3. Accuracy of segmentation versus number of training data chosen from unlabeled set for geometric context dataset

% Equation 11 % Get label

beled training set L. We then train a segmenter on this augmented training set L+”labeled”u to produce θL+u , then test this system on 20 images in the test-set T , recording the average F-measure

% Equation 6

U := U − {ui } end Figure 1. Pseudo code for the L MU active learning algorithm

Figure 2. Sample “sky” images from Geometric context dataset

not easily separable based on their color. The original images were of various sizes; we downsized each to 65 × 65 pixels to make them uniform, and to make our computations more tractable. We partitioned these images into the unlabeled-set U with 85 images and the test-set T with 20 images. Features: We associate each pixel with twelve values: its 3 color intensities, its vertical position (as the sky is typically at the top of the image) and 8 texture values: We apply the MR8 filter banks [VZ02] (which contain filters at 6 different orientations, at 3 scales) to the region centered on each pixel, but record only the maximum filter response at each of the 6 orientations; we also include 2 isotropic features: Gaussian and a Laplacian of Gaussian. In each iteration of the active learning process, our L MU system identifies one specific image u from U , which is removed from U , labeled by an oracle, then added to the la-

F-measure

=

2 × (precision)(recall) (precision + recall)

where precision = tp/(tp + f p) and recall = tp/(tp + f n), where tp, f p, f n are true positive, false positive and false negative. Note this measure is problematic when tp = 0; see Footnote 3. In Figure 3, the horizontal axis represents the number of added unlabeled images and vertical axis is the average F-measure. The “all” line shows the results of training on (the oracle-labeled versions of) all 85 unlabeled images. Checking the “LMU” line, we see that the first carefully-selected image alone produced a classifier whose F-measure was 67%, and this accuracy improved to 81% by using the second image. Note this is only 2% below the accuracy obtained by training on all unlabeled data; moreover, this is statistically indistinguishable at the p < 0.05 level, based on a paired t-test. (The accuracy of the second iteration was significantly better than the first — paired ttest p < 0.05.) There is no significant change between the second iteration and subsequent iterations, which means that the first two images chosen by L MU are good enough to train the classifier. Those two images, by the way, are the left and middle images shown in Figure 2. Of course, it is possible that any two images would be sufficient. To test this, we randomly selected images for the training set; see the “random” line in Figure 3. Each point in this line is the average over 10 random choices. We see that this line is well below the LMU line. Moreover, the segmenter remains statistically inferior to the “all” line until observing (on average) 5 randomly-drawn images.

Learning to Segment from a Few Well-Selected Training Images

Figure 4. Sample images from brain tumor dataset, tumor is segmented in red

4.2. Finding Tumors in Brain Scans Here, we consider the challenge of finding tumors within a patient’s brain — that is, labeling each pixel in a magnetic resonance (MR) image as either tumorous or nontumorous. This task is crucial in surgical planning and radiation therapy, and currently requires a significant amount of manual work by human medical experts.

Figure 5. Accuracy of segmentation versus number of training data chosen from unlabeled set for brain tumor dataset

Here, we have 80 images (axial slices) from the brains of 16 patients. These are taken from different regions of the brains; in particular, no two images are adjacent to each other. We resized each image from 256 × 256 pixels to 65×65. Figure 4 shows three images from this dataset, with the tumor segmentation outlined in red. We again segment these images into an unlabeled set U , containing 71 images from 11 individuals, and a test set T , containing 9 images from the other 5 patients.3 Features: Most patient visits yield scans in 3 different MR modalities: T1, T2, and T1c (that is, T1 after the patient has received a contrasting agent). We identify each pixel i with a vector of 4 values, including the T2 value and the difference between T1c and T1. As each brain is somewhat symmetric around the sagittal plane, we also include the symmetry feature by computing the difference between intensities of pairs of symmetrical pixels with respect to the sagittal plane, for both T2 and T1c - T1 modalities. So we compute four features for each pixel. Figure 5 shows the results of actively selecting the training set. Actively training on one image in this dataset produces an average F-measure of 57%. This segmenter is as good as the one obtained using all of the data (paired t-test p < 0.05). (The specific image selected is the left one in Figure 4.) Training on the second L MU-selected image increased the average accuracy to 70%, which is higher than the 61% accuracy obtained using all of the data. (Note, however, that this is not statistically better, at p < 0.05.) Donmez and Carbonell [DC08] report a similar situation 3

The F-measure score is problematic if any test image has no tumor, as here there can be no true positives (tp = 0). Hence, to simplify our analysis, our T includes only images that contain some tumor. However, the unlabeled set U includes some images that have no tumor.

Figure 6. Accuracy of segmentation versus number of training data chosen from unlabeled set for geometric context dataset with images downsized to 32 × 32 pixels

(albeit in active sampling in rank learning on TREC 2004 dataset), noting that the performance of their active learning algorithm is sometimes better than the one obtained by training on all the data. Finally, as with the Sky data, on average the segmenters produced using the first several (here three) randomly drawn images were all significantly inferior to the “all” segmenter. 4.3. Scalability Study This subsection explores how our L MU scales with the size of the images. We therefore downsized the images in the geometric context dataset from 65 × 65 to 32 × 32, then repeated the same active learning process described above. The results, appearing in Figure 6, show essentially the same trend that appeared in Figure 3. Here, however, L MU required 7 images before it first obtained a segmenter whose performance was statistically “equivalent” to the one based on all of the data (paired t-test, p < 0.05).

Learning to Segment from a Few Well-Selected Training Images

this is just used to select the appropriate image to give to the oracle; the oracle then finds the parameters for the full DRF, based on equation 6. [GFS09] presents those results. We also experimented using several variants of our L MU, including one that used only RU throughout, and another that used only(a variant) of MU. [GFS09] presents those findings, which demonstrate that our LMU is superior.

5. Conclusions Figure 7. LMU of sunflower category in Caltech101 dataset

4.4. Discussion It is, at first, very surprising that one can produce an effective segmenter with so few images — here, only 2 for the (original) sky data, and 1 for the brain tumor data! Towards explaining this, note that each of these images is not really a single “point”, but is actually 65 × 65 ≈ 4,000 pixels, and each oracle-label is actually providing around 4000 bits. Hence, the 2 sky images is essentially 8000 bits, which is a lot of information. The results in the scaling studies are consistent with this conjecture: Here, we required 7 carefully-selected images to obtain the information needed to do well; notice that this 7 × (32 × 32) ≈ 2 × (65 × 65).4 As further support, consider the segmenters based on randomly-drawn images. While they were, typically, worse than ones based on images specified by L MU, we found that we were still getting good segmenters using only a few such random images. Again, this is consistent with the view that the label of each image is supplying a great deal of information — even if the image is drawn randomly. 4.5. Other Results The results shown above strongly suggest that very few images are sufficient to train a DRF-based segmentor, but only if they are well selected. We explored this claim over other datasets. Figure 7 shows the results of applying L MU on the set of sunflower objects from Caltech101 dataset [FFFP04]. Since the dataset is intended for object recognition task, it is relatively easy to segment the object from background, which is probably why L MU gets good accuracy after actively selecting only one image. We also explored many ways to reduce L MU’s computation complexity. For example, the L MU-LR system used logical regression to estimate the conditional probabilities (as if the pixels were independent) for the entropy function. Note 4

We are not claiming that 8000 bits is a magical number — instead, we are just observing that our active learner requires more small images than large images, which is consistent with the claim that the number of pixels being labeled seems significant.

While there are now many results in active learning, this is one of the first studies that considers the challenge of actively learning the parameters of a DRF-based segmenter. While our L MU system is based on standard “parts” — selecting the image with maximal uncertainty, and or that most reduces the uncertainty of other images — we found that this particular combination was effective for this task. We also found, to our surprise, that we could produce an effective segmenter using very few segmented images. Our studies support the claim that it may be because the label for a single image contains a great deal of information; i.e., corresponds to receiving many single-bit labels, in the standard active learning framework.

Acknowledgments All authors gratefully acknowledge the support from the Alberta Ingenuity Centre for Machine Learning (AICML), NSERC,and iCORE. We also thank Dr. Albert Murtha, as well as the Alberta Cancer Board, for providing the (segmented) brain tumor images. Csaba Szepesv´ari is on leave from MTA SZTAKI, Bp. Hungary.

References [Bes86]

J. Besag. On the statistical analysis of dirty pictures. Journal of Royal Statistical Society, Series B, 48:259–302, 1986.

[CCS00]

Colin Campbell, Nello Cristianini, and Alex Smola. Query learning with large margin classifiers. In Proceedings of the International Conference on Machine Learning (ICML), 2000.

[CDLFF08] Brendan Collins, Jia Deng, Kai Li, and Li FeiFei. Towards scalable dataset construction: An active learning approach. In Proceedings of the European Conference on Computer Vision (ECCV), 2008. [CGJ96]

David A. Cohn, Zoubin Ghahramani, and Michael I. Jordan. Active learning with statistical models. Journal of Artificial Intelligence Research, 4:129–145, 1996.

Learning to Segment from a Few Well-Selected Training Images

[DC08]

Pinar Donmez and Jaime G. Carbonell. Optimizing estimated loss reduction for active sampling in rank learning. In Proceedings of the International Conference on Machine Learning (ICML), 2008.

[FFFP04]

L. Fei-Fei, R. Fergus, and P. Perona. Learning generative visual models from few training examples: an incremental bayesian approach tested on 101 object categories. In Workshop on Generative-Model Based Vision (CVPR), 2004.

[FSST97]

[GFS09]

Yoav Freund, H. Sebastian Seung, Eli Shamir, and Naftali Tishby. Selective sampling using the query by committee algorithm. Machine Learning, 28:133–168, 1997. Russell Greiner, Alireza Farhangfar, and Csaba Szepesvari, 2009. http://www.cs.ualberta.ca/∼greiner/ RESEARCH/ActiveLearning4Segmentation/.

[GG05]

Yuhong Guo and Russ Greiner. Optimistic active learning using mutual information. In Proceedings of the Uncertainty in Artificial Intelligence(UAI), 2005.

[KH03]

Sanjiv Kumar and Martial Hebert. Discriminative random fields: A discriminative framework for contextual interaction in classification. In Proceedings of the International Conference on Computer Vision (ICCV), 2003.

[LG94]

David D. Lewis and William A. Gale. A sequential algorithm for training text classifiers. In Proceedings of the Special Interest Group on Information Retrieval (ACMSIGIR), 1994.

[LWB+ 08] ChiHoon Lee, Shaojun Wang, Matthew Brown, Albert Murtha, and Russell Greiner. Segmenting brain tumors using pseudoconditional random fields. In Proceedings of the 11th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), 2008. [NS04]

Hieu Nguyen and Arnold Smeulders. Active learning using pre-clustering. In Proceedings of the International Conference on Machine Learning (ICML), 2004.

[RM01]

Nicholas Roy and Andrew McCallum. Toward optimal active learning through sampling estimation of error reduction. In Proceedings of the International Conference on Machine Learning (ICML), 2001.

[SC00]

Greg Schohn and David Cohn. Less is more: Active learning with support vector machines. In Proceedings of the International Conference on Machine Learning (ICML), 2000.

[SKH08]

Martin Szummer, Pushmeet Kohli, and Derek Hoiem. Learning CRFs using graph cuts. In Proceedings of the European Conference on Computer Vision (ECCV), 2008.

[TK02]

Simon Tong and Daphne Koller. Support vector machine active learning with applications to text classification. Journal of Machine Learning Research, 2:45–66, 2002.

[VG08]

Sudheendra Vijayanarasimhan and Kristen Grauman. Multi-level active prediction of useful image annotations for recognition. In Proceedings of the Neural Information Processing Systems (NIPS), 2008.

[VZ02]

Manik Varma and Andrew Zisserman. Classifying images of materials: Achieving viewpoint and illumination independence. In Proceedings of the 7th European Conference on Computer Vision(ECCV), 2002.

[XYT+ 03] Zhao Xu, Kai Yu, Volker Tresp, Xiaowei Xu, and Jizhi Wang. Representative sampling for text classification using support vector machines. In Proceedings of the European Conference on Information Retrieval, 2003. [ZC02]

Cha Zhang and Tsuhan Chen. An active learning framework for content based information retrieval. IEEE trans. on Multimedia, Special Issue on Multimedia Database, 4:260– 268, 2002.

Suggest Documents