67
Active Learning Literature Survey Burr Settles Computer Sciences Technical Report 1648 University of Wisconsin–Madison Updated on: January 26, 2010

Active Learning Literature Survey

  • Upload
    butest

  • View
    994

  • Download
    0

Embed Size (px)

DESCRIPTION

 

Citation preview

Page 1: Active Learning Literature Survey

Active Learning Literature Survey

Burr Settles

Computer Sciences Technical Report 1648University of Wisconsin–Madison

Updated on: January 26, 2010

Page 2: Active Learning Literature Survey

Abstract

The key idea behind active learning is that a machine learning algorithm canachieve greater accuracy with fewer training labels if it is allowed to choose thedata from which it learns. An active learner may pose queries, usually in the formof unlabeled data instances to be labeled by an oracle (e.g., a human annotator).Active learning is well-motivated in many modern machine learning problems,where unlabeled data may be abundant or easily obtained, but labels are difficult,time-consuming, or expensive to obtain.

This report provides a general introduction to active learning and a survey ofthe literature. This includes a discussion of the scenarios in which queries canbe formulated, and an overview of the query strategy frameworks proposed inthe literature to date. An analysis of the empirical and theoretical evidence forsuccessful active learning, a summary of problem setting variants and practicalissues, and a discussion of related topics in machine learning research are alsopresented.

Page 3: Active Learning Literature Survey

Contents1 Introduction 3

1.1 What is Active Learning? . . . . . . . . . . . . . . . . . . . . . . 41.2 Active Learning Examples . . . . . . . . . . . . . . . . . . . . . 51.3 Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . 8

2 Scenarios 82.1 Membership Query Synthesis . . . . . . . . . . . . . . . . . . . . 92.2 Stream-Based Selective Sampling . . . . . . . . . . . . . . . . . 102.3 Pool-Based Sampling . . . . . . . . . . . . . . . . . . . . . . . . 11

3 Query Strategy Frameworks 123.1 Uncertainty Sampling . . . . . . . . . . . . . . . . . . . . . . . . 123.2 Query-By-Committee . . . . . . . . . . . . . . . . . . . . . . . . 153.3 Expected Model Change . . . . . . . . . . . . . . . . . . . . . . 183.4 Expected Error Reduction . . . . . . . . . . . . . . . . . . . . . . 193.5 Variance Reduction . . . . . . . . . . . . . . . . . . . . . . . . . 213.6 Density-Weighted Methods . . . . . . . . . . . . . . . . . . . . . 25

4 Analysis of Active Learning 264.1 Empirical Analysis . . . . . . . . . . . . . . . . . . . . . . . . . 274.2 Theoretical Analysis . . . . . . . . . . . . . . . . . . . . . . . . 28

5 Problem Setting Variants 305.1 Active Learning for Structured Outputs . . . . . . . . . . . . . . 305.2 Active Feature Acquisition and Classification . . . . . . . . . . . 325.3 Active Class Selection . . . . . . . . . . . . . . . . . . . . . . . 335.4 Active Clustering . . . . . . . . . . . . . . . . . . . . . . . . . . 33

6 Practical Considerations 346.1 Batch-Mode Active Learning . . . . . . . . . . . . . . . . . . . . 356.2 Noisy Oracles . . . . . . . . . . . . . . . . . . . . . . . . . . . . 366.3 Variable Labeling Costs . . . . . . . . . . . . . . . . . . . . . . . 376.4 Alternative Query Types . . . . . . . . . . . . . . . . . . . . . . 396.5 Multi-Task Active Learning . . . . . . . . . . . . . . . . . . . . . 426.6 Changing (or Unknown) Model Classes . . . . . . . . . . . . . . 436.7 Stopping Criteria . . . . . . . . . . . . . . . . . . . . . . . . . . 44

1

Page 4: Active Learning Literature Survey

7 Related Research Areas 447.1 Semi-Supervised Learning . . . . . . . . . . . . . . . . . . . . . 447.2 Reinforcement Learning . . . . . . . . . . . . . . . . . . . . . . 457.3 Submodular Optimization . . . . . . . . . . . . . . . . . . . . . . 467.4 Equivalence Query Learning . . . . . . . . . . . . . . . . . . . . 477.5 Model Parroting and Compression . . . . . . . . . . . . . . . . . 47

8 Conclusion and Final Thoughts 48

Bibliography 49

2

Page 5: Active Learning Literature Survey

1 IntroductionThis report provides a general review of the literature on active learning. Therehave been a host of algorithms and applications for learning with queries overthe years, and this document is an attempt to distill the core ideas, methods, andapplications that have been considered by the machine learning community. Tomake this survey more useful in the long term, an online version will be updatedand maintained indefinitely at:

http://active-learning.net/

When referring to this document, I recommend using the following citation:

Burr Settles. Active Learning Literature Survey. Computer Sciences Tech-nical Report 1648, University of Wisconsin–Madison. 2009.

An appropriate BIBTEX entry is:

@techreport{settles.tr09,Author = {Burr Settles},Institution = {University of Wisconsin--Madison},Number = {1648},Title = {Active Learning Literature Survey},Type = {Computer Sciences Technical Report},Year = {2009},

}

This document is written for a machine learning audience, and assumes the readerhas a working knowledge of supervised learning algorithms (particularly statisti-cal methods). For a good introduction to general machine learning, I recommendMitchell (1997) or Duda et al. (2001). I have strived to make this review as com-prehensive as possible, but it is by no means complete. My own research deals pri-marily with applications in natural language processing and bioinformatics, thusmuch of the empirical active learning work I am familiar with is in these areas.Active learning (like so many subfields in computer science) is rapidly growingand evolving in a myriad of directions, so it is difficult for one person to providean exhaustive summary. I apologize for any oversights or inaccuracies, and en-courage interested readers to submit additions, comments, and corrections to meat: [email protected].

3

Page 6: Active Learning Literature Survey

1.1 What is Active Learning?Active learning (sometimes called “query learning” or “optimal experimental de-sign” in the statistics literature) is a subfield of machine learning and, more gener-ally, artificial intelligence. The key hypothesis is that if the learning algorithm isallowed to choose the data from which it learns—to be “curious,” if you will—itwill perform better with less training. Why is this a desirable property for learningalgorithms to have? Consider that, for any supervised learning system to performwell, it must often be trained on hundreds (even thousands) of labeled instances.Sometimes these labels come at little or no cost, such as the the “spam” flag youmark on unwanted email messages, or the five-star rating you might give to filmson a social networking website. Learning systems use these flags and ratings tobetter filter your junk email and suggest movies you might enjoy. In these casesyou provide such labels for free, but for many other more sophisticated supervisedlearning tasks, labeled instances are very difficult, time-consuming, or expensiveto obtain. Here are a few examples:

• Speech recognition. Accurate labeling of speech utterances is extremelytime consuming and requires trained linguists. Zhu (2005a) reports thatannotation at the word level can take ten times longer than the actual au-dio (e.g., one minute of speech takes ten minutes to label), and annotatingphonemes can take 400 times as long (e.g., nearly seven hours). The prob-lem is compounded for rare languages or dialects.

• Information extraction. Good information extraction systems must be trainedusing labeled documents with detailed annotations. Users highlight entitiesor relations of interest in text, such as person and organization names, orwhether a person works for a particular organization. Locating entities andrelations can take a half-hour or more for even simple newswire stories (Set-tles et al., 2008a). Annotations for other knowledge domains may requireadditional expertise, e.g., annotating gene and disease mentions for biomed-ical information extraction usually requires PhD-level biologists.

• Classification and filtering. Learning to classify documents (e.g., articlesor web pages) or any other kind of media (e.g., image, audio, and videofiles) requires that users label each document or media file with particularlabels, like “relevant” or “not relevant.” Having to annotate thousands ofthese instances can be tedious and even redundant.

4

Page 7: Active Learning Literature Survey

Active learning systems attempt to overcome the labeling bottleneck by askingqueries in the form of unlabeled instances to be labeled by an oracle (e.g., a humanannotator). In this way, the active learner aims to achieve high accuracy usingas few labeled instances as possible, thereby minimizing the cost of obtaininglabeled data. Active learning is well-motivated in many modern machine learningproblems where data may be abundant but labels are scarce or expensive to obtain.Note that this kind of active learning is related in spirit, though not to be confused,with the family of instructional techniques by the same name in the educationliterature (Bonwell and Eison, 1991).

1.2 Active Learning Examples

machine learningmodel

LU

labeledtraining set

unlabeled pool

oracle (e.g., human annotator)

learn a model

select queries

Figure 1: The pool-based active learning cycle.

There are several scenarios in which active learners may pose queries, andthere are also several different query strategies that have been used to decide whichinstances are most informative. In this section, I present two illustrative examplesin the pool-based active learning setting (in which queries are selected from alarge pool of unlabeled instances U) using an uncertainty sampling query strategy(which selects the instance in the pool about which the model is least certain howto label). Sections 2 and 3 describe all the active learning scenarios and querystrategy frameworks in more detail.

5

Page 8: Active Learning Literature Survey

-3

-2

-1

0

1

2

3

-4 -2 0 2 4-3

-2

-1

0

1

2

3

-4 -2 0 2 4-3

-2

-1

0

1

2

3

-4 -2 0 2 4(a) (b) (c)

Figure 2: An illustrative example of pool-based active learning. (a) A toy data set of400 instances, evenly sampled from two class Gaussians. The instances arerepresented as points in a 2D feature space. (b) A logistic regression modeltrained with 30 labeled instances randomly drawn from the problem domain.The line represents the decision boundary of the classifier (70% accuracy). (c)A logistic regression model trained with 30 actively queried instances usinguncertainty sampling (90%).

Figure 1 illustrates the pool-based active learning cycle. A learner may beginwith a small number of instances in the labeled training set L, request labels forone or more carefully selected instances, learn from the query results, and thenleverage its new knowledge to choose which instances to query next. Once aquery has been made, there are usually no additional assumptions on the part ofthe learning algorithm. The new labeled instance is simply added to the labeledset L, and the learner proceeds from there in a standard supervised way. There area few exceptions to this, such as when the learner is allowed to make alternativetypes of queries (Section 6.4), or when active learning is combined with semi-supervised learning (Section 7.1).

Figure 2 shows the potential of active learning in a way that is easy to visu-alize. This is a toy data set generated from two Gaussians centered at (-2,0) and(2,0) with standard deviation σ = 1, each representing a different class distribu-tion. Figure 2(a) shows the resulting data set after 400 instances are sampled (200from each class); instances are represented as points in a 2D feature space. Ina real-world setting these instances may be available, but their labels usually arenot. Figure 2(b) illustrates the traditional supervised learning approach after ran-domly selecting 30 instances for labeling, drawn i.i.d. from the unlabeled pool U .The line shows the linear decision boundary of a logistic regression model (i.e.,where the posterior equals 0.5) trained using these 30 points. Notice that mostof the labeled instances in this training set are far from zero on the horizontal

6

Page 9: Active Learning Literature Survey

0.5

0.6

0.7

0.8

0.9

1

0 20 40 60 80 100

accu

racy

number of instance queries

uncertainty samplingrandom

Figure 3: Learning curves for text classification: baseball vs. hockey. Curves plot clas-sification accuracy as a function of the number of documents queried for two se-lection strategies: uncertainty sampling (active learning) and random sampling(passive learning). We can see that the active learning approach is superior herebecause its learning curve dominates that of random sampling.

axis, which is where the Bayes optimal decision boundary should probably be.As a result, this classifier only achieves 70% accuracy on the remaining unlabeledpoints. Figure 2(c), however, tells a different story. The active learner uses uncer-tainty sampling to focus on instances closest to its decision boundary, assuming itcan adequately explain those in other parts of the input space characterized by U .As a result, it avoids requesting labels for redundant or irrelevant instances, andachieves 90% accuracy with a mere 30 labeled instances.

Now let us consider active learning for a real-world learning task: text classifi-cation. In this example, a learner must distinguish between baseball and hockeydocuments from the 20 Newsgroups corpus (Lang, 1995), which consists of 2,000Usenet documents evenly divided between the two classes. Active learning al-gorithms are generally evaluated by constructing learning curves, which plot theevaluation measure of interest (e.g., accuracy) as a function of the number ofnew instance queries that are labeled and added to L. Figure 3 presents learningcurves for the first 100 instances labeled using uncertainty sampling and random

7

Page 10: Active Learning Literature Survey

sampling. The reported results are for a logistic regression model averaged overten folds using cross-validation. After labeling 30 new instances, the accuracy ofuncertainty sampling is 81%, while the random baseline is only 73%. As can beseen, the active learning curve dominates the baseline curve for all of the pointsshown in this figure. We can conclude that an active learning algorithm is superiorto some other approach (e.g., a random baseline like traditional passive supervisedlearning) if it dominates the other for most or all of the points along their learningcurves.

1.3 Further ReadingThis is the first large-scale survey of the active learning literature. One way to viewthis document is as a heavily annotated bibliography of the field, and the citationswithin a particular section or subsection of interest serve as good starting pointsfor further investigation. There have also been a few PhD theses over the yearsdedicated to active learning, with rich related work sections. In fact, this reportoriginated as a chapter in my PhD thesis (Settles, 2008), which focuses on activelearning with structured instances and potentially varied annotation costs. Also ofinterest may be the related work chapters of Tong (2001), which considers activelearning for support vector machines and Bayesian networks, Monteleoni (2006),which considers more theoretical aspects of active learning for classification, andOlsson (2008), which focuses on active learning for named entity recognition (atype of information extraction). Fredrick Olsson has also written a survey of activelearning specifically within the scope of the natural language processing (NLP)literature (Olsson, 2009).

2 ScenariosThere are several different problem scenarios in which the learner may be ableto ask queries. The three main settings that have been considered in the litera-ture are (i) membership query synthesis, (ii) stream-based selective sampling, and(iii) pool-based sampling. Figure 4 illustrates the differences among these threescenarios, which are explained in more detail in this section. Note that all thesescenarios (and the lion’s share of active learning work to date) assume that queriestake the form of unlabeled instances to be labeled by the oracle. Sections 6 and 5discuss some alternatives to this setting.

8

Page 11: Active Learning Literature Survey

instance space or input

distribution

Usample a largepool of instances

sample aninstance

model generatesa query de novo

model decides toquery or discard

model selectsthe best query

membership query synthesis

stream-based selective sampling

pool-based sampling query is labeledby the oracle

Figure 4: Diagram illustrating the three main active learning scenarios.

2.1 Membership Query SynthesisOne of the first active learning scenarios to be investigated is learning with mem-bership queries (Angluin, 1988). In this setting, the learner may request labelsfor any unlabeled instance in the input space, including (and typically assuming)queries that the learner generates de novo, rather than those sampled from someunderlying natural distribution. Efficient query synthesis is often tractable andefficient for finite problem domains (Angluin, 2001). The idea of synthesizingqueries has also been extended to regression learning tasks, such as learning topredict the absolute coordinates of a robot hand given the joint angles of its me-chanical arm as inputs (Cohn et al., 1996).

Query synthesis is reasonable for many problems, but labeling such arbitraryinstances can be awkward if the oracle is a human annotator. For example, Langand Baum (1992) employed membership query learning with human oracles totrain a neural network to classify handwritten characters. They encountered anunexpected problem: many of the query images generated by the learner con-tained no recognizable symbols, only artificial hybrid characters that had no nat-ural semantic meaning. Similarly, one could imagine that membership queriesfor natural language processing tasks might create streams of text or speech thatamount to gibberish. The stream-based and pool-based scenarios (described in thenext sections) have been proposed to address these limitations.

However, King et al. (2004, 2009) describe an innovative and promising real-world application of the membership query scenario. They employ a “robot scien-

9

Page 12: Active Learning Literature Survey

tist” which can execute a series of autonomous biological experiments to discovermetabolic pathways in the yeast Saccharomyces cerevisiae. Here, an instance is amixture of chemical solutions that constitute a growth medium, as well as a partic-ular yeast mutant. A label, then, is whether or not the mutant thrived in the growthmedium. All experiments are autonomously synthesized using an active learningapproach based on inductive logic programming, and physically performed us-ing a laboratory robot. This active method results in a three-fold decrease in thecost of experimental materials compared to naıvely running the least expensiveexperiment, and a 100-fold decrease in cost compared to randomly generated ex-periments. In domains where labels come not from human annotators, but fromexperiments such as this, query synthesis may be a promising direction for auto-mated scientific discovery.

2.2 Stream-Based Selective SamplingAn alternative to synthesizing queries is selective sampling (Cohn et al., 1990,1994). The key assumption is that obtaining an unlabeled instance is free (or in-expensive), so it can first be sampled from the actual distribution, and then thelearner can decide whether or not to request its label. This approach is sometimescalled stream-based or sequential active learning, as each unlabeled instance istypically drawn one at a time from the data source, and the learner must decidewhether to query or discard it. If the input distribution is uniform, selective sam-pling may well behave like membership query learning. However, if the distri-bution is non-uniform and (more importantly) unknown, we are guaranteed thatqueries will still be sensible, since they come from a real underlying distribution.

The decision whether or not to query an instance can be framed several ways.One approach is to evaluate samples using some “informativeness measure” or“query strategy” (see Section 3 for examples) and make a biased random deci-sion, such that more informative instances are more likely to be queried (Daganand Engelson, 1995). Another approach is to compute an explicit region of uncer-tainty (Cohn et al., 1994), i.e., the part of the instance space that is still ambiguousto the learner, and only query instances that fall within it. A naıve way of doingthis is to set a minimum threshold on an informativeness measure which definesthe region. Instances whose evaluation is above this threshold are then queried.Another more principled approach is to define the region that is still unknown tothe overall model class, i.e., to the set of hypotheses consistent with the current la-beled training set called the version space (Mitchell, 1982). In other words, if anytwo models of the same model class (but different parameter settings) agree on all

10

Page 13: Active Learning Literature Survey

the labeled data, but disagree on some unlabeled instance, then that instance lieswithin the region of uncertainty. Calculating this region completely and explicitlyis computationally expensive, however, and it must be maintained after each newquery. As a result, approximations are used in practice (Seung et al., 1992; Cohnet al., 1994; Dasgupta et al., 2008).

The stream-based scenario has been studied in several real-world tasks, includ-ing part-of-speech tagging (Dagan and Engelson, 1995), sensor scheduling (Kr-ishnamurthy, 2002), and learning ranking functions for information retrieval (Yu,2005). Fujii et al. (1998) employ selective sampling for active learning in wordsense disambiguation, e.g., determining if the word “bank” means land alongsidea river or a financial institution in a given context (only they study Japanese wordsin their work). The approach not only reduces annotation effort, but also limitsthe size of the database used in nearest-neighbor learning, which in turn expeditesthe classification algorithm.

It is worth noting that some authors (e.g., Thompson et al., 1999; Moskovitchet al., 2007) use “selective sampling” to refer to the pool-based scenario describedin the next section. Under this interpretation, the term merely signifies that queriesare made with a select set of instances sampled from a real data distribution.However, in most of the literature selective sampling refers to the stream-basedscenario described here.

2.3 Pool-Based SamplingFor many real-world learning problems, large collections of unlabeled data can begathered at once. This motivates pool-based sampling (Lewis and Gale, 1994),which assumes that there is a small set of labeled data L and a large pool of un-labeled data U available. Queries are selectively drawn from the pool, which isusually assumed to be closed (i.e., static or non-changing), although this is notstrictly necessary. Typically, instances are queried in a greedy fashion, accordingto an informativeness measure used to evaluate all instances in the pool (or, per-haps if U is very large, some subsample thereof). The examples from Section 1.2use this active learning setting.

The pool-based scenario has been studied for many real-world problem do-mains in machine learning, such as text classification (Lewis and Gale, 1994; Mc-Callum and Nigam, 1998; Tong and Koller, 2000; Hoi et al., 2006a), informationextraction (Thompson et al., 1999; Settles and Craven, 2008), image classificationand retrieval (Tong and Chang, 2001; Zhang and Chen, 2002), video classification

11

Page 14: Active Learning Literature Survey

and retrieval (Yan et al., 2003; Hauptmann et al., 2006), speech recognition (Turet al., 2005), and cancer diagnosis (Liu, 2004) to name a few.

The main difference between stream-based and pool-based active learning isthat the former scans through the data sequentially and makes query decisionsindividually, whereas the latter evaluates and ranks the entire collection beforeselecting the best query. While the pool-based scenario appears to be much morecommon among application papers, one can imagine settings where the stream-based approach is more appropriate. For example, when memory or processingpower may be limited, as with mobile and embedded devices.

3 Query Strategy FrameworksAll active learning scenarios involve evaluating the informativeness of unlabeledinstances, which can either be generated de novo or sampled from a given distribu-tion. There have been many proposed ways of formulating such query strategiesin the literature. This section provides an overview of the general frameworks thatare used. From this point on, I use the notation x∗A to refer to the most informativeinstance (i.e., the best query) according to some query selection algorithm A.

3.1 Uncertainty SamplingPerhaps the simplest and most commonly used query framework is uncertaintysampling (Lewis and Gale, 1994). In this framework, an active learner queriesthe instances about which it is least certain how to label. This approach is oftenstraightforward for probabilistic learning models. For example, when using aprobabilistic model for binary classification, uncertainty sampling simply queriesthe instance whose posterior probability of being positive is nearest 0.5 (Lewisand Gale, 1994; Lewis and Catlett, 1994).

For problems with three or more class labels, a more general uncertainty sam-pling variant might query the instance whose prediction is the least confident:

x∗LC = argmaxx

1− Pθ(y|x),

where y = argmaxy Pθ(y|x), or the class label with the highest posterior prob-ability under the model θ. One way to interpret this uncertainty measure is theexpected 0/1-loss, i.e., the model’s belief that it will mislabel x. This sort of strat-egy has been popular, for example, with statistical sequence models in information

12

Page 15: Active Learning Literature Survey

extraction tasks (Culotta and McCallum, 2005; Settles and Craven, 2008). Thisis because the most likely label sequence (and its associated likelihood) can beefficiently computed using dynamic programming.

However, the criterion for the least confident strategy only considers informa-tion about the most probable label. Thus, it effectively “throws away” informationabout the remaining label distribution. To correct for this, some researchers use adifferent multi-class uncertainty sampling variant called margin sampling (Schef-fer et al., 2001):

x∗M = argminx

Pθ(y1|x)− Pθ(y2|x),

where y1 and y2 are the first and second most probable class labels under themodel, respectively. Margin sampling aims to correct for a shortcoming in leastconfident strategy, by incorporating the posterior of the second most likely la-bel. Intuitively, instances with large margins are easy, since the classifier has littledoubt in differentiating between the two most likely class labels. Instances withsmall margins are more ambiguous, thus knowing the true label would help themodel discriminate more effectively between them. However, for problems withvery large label sets, the margin approach still ignores much of the output distri-bution for the remaining classes.

A more general uncertainty sampling strategy (and possibly the most popular)uses entropy (Shannon, 1948) as an uncertainty measure:

x∗H = argmaxx−∑i

Pθ(yi|x) logPθ(yi|x),

where yi ranges over all possible labelings. Entropy is an information-theoreticmeasure that represents the amount of information needed to “encode” a distri-bution. As such, it is often thought of as a measure of uncertainty or impurity inmachine learning. For binary classification, entropy-based sampling reduces tothe margin and least confident strategies above; in fact all three are equivalent toquerying the instance with a class posterior closest to 0.5. However, the entropy-based approach generalizes easily to probabilistic multi-label classifiers and prob-abilistic models for more complex structured instances, such as sequences (Settlesand Craven, 2008) and trees (Hwa, 2004).

Figure 5 visualizes the implicit relationship among these uncertainty mea-sures. In all cases, the most informative instance would lie at the center of thetriangle, because this represents where the posterior label distribution is most uni-form (and thus least certain under the model). Similarly, the least informativeinstances are at the three corners, where one of the classes has extremely high

13

Page 16: Active Learning Literature Survey

(a) least confident (b) margin (c) entropy

Figure 5: Heatmaps illustrating the query behavior of common uncertainty measures ina three-label classification problem. Simplex corners indicate where one labelhas very high probability, with the opposite edge showing the probability rangefor the other two classes when that label has very low probability. Simplexcenters represent a uniform posterior distribution. The most informative queryregion for each strategy is shown in dark red, radiating from the centers.

probability (and thus little model uncertainty). The main differences lie in therest of the probability space. For example, the entropy measure does not favorinstances where only one of the labels is highly unlikely (i.e., along the outerside edges), because the model is fairly certain that it is not the true label. Theleast confident and margin measures, on the other hand, consider such instancesto be useful if the model cannot distinguish between the remaining two classes.Empirical comparisons of these measures (e.g., Korner and Wrobel, 2006; Scheinand Ungar, 2007; Settles and Craven, 2008) have yielded mixed results, suggest-ing that the best strategy may be application-dependent (note that all strategiesstill generally outperform passive baselines). Intuitively, though, entropy seemsappropriate if the objective function is to minimize log-loss, while the other two(particularly margin) are more appropriate if we aim to reduce classification error,since they prefer instances that would help the model better discriminate amongspecific classes.

Uncertainty sampling strategies may also be employed with non-probabilisticclassifiers. One of the first works to explore uncertainty sampling used a decisiontree classifier (Lewis and Catlett, 1994). Similar approaches have been appliedto active learning with nearest-neighbor (a.k.a. “memory-based” or “instance-based”) classifiers (Fujii et al., 1998; Lindenbaum et al., 2004), by allowing eachneighbor to vote on the class label of x, with the proportion of these votes rep-resenting the posterior label probability. Tong and Koller (2000) also experiment

14

Page 17: Active Learning Literature Survey

with an uncertainty sampling strategy for support vector machines—or SVMs—that involves querying the instance closest to the linear decision boundary. Thislast approach is analogous to uncertainty sampling with a probabilistic binary lin-ear classifier, such as logistic regression or naıve Bayes.

So far we have only discussed classification tasks, but uncertainty samplingis also applicable in regression problems (i.e., learning tasks where the outputvariable is a continuous value rather than a set of discrete class labels). In thissetting, the learner simply queries the unlabeled instance for which the model hasthe highest output variance in its prediction. Under a Gaussian assumption, theentropy of a random variable is a monotonic function of its variance, so this ap-proach is very much in same the spirit as entropy-based uncertainty sampling forclassification. Closed-form approximations of output variance can be computedfor a variety of models, including Gaussian random fields (Cressie, 1991) and neu-ral networks (MacKay, 1992). Active learning for regression problems has a longhistory in the statistics literature, generally referred to as optimal experimentaldesign (Federov, 1972). Such approaches shy away from uncertainty sampling inlieu of more sophisticated strategies, which we will explore further in Section 3.5.

3.2 Query-By-CommitteeAnother, more theoretically-motivated query selection framework is the query-by-committee (QBC) algorithm (Seung et al., 1992). The QBC approach involvesmaintaining a committee C = {θ(1), . . . , θ(C)} of models which are all trained onthe current labeled set L, but represent competing hypotheses. Each committeemember is then allowed to vote on the labelings of query candidates. The mostinformative query is considered to be the instance about which they most disagree.

The fundamental premise behind the QBC framework is minimizing the ver-sion space, which is (as described in Section 2.2) the set of hypotheses that areconsistent with the current labeled training data L. Figure 6 illustrates the con-cept of version spaces for (a) linear functions and (b) axis-parallel box classifiersin different binary classification tasks. If we view machine learning as a searchfor the “best” model within the version space, then our goal in active learning isto constrain the size of this space as much as possible (so that the search can bemore precise) with as few labeled instances as possible. This is exactly what QBCaims to do, by querying in controversial regions of the input space. In order toimplement a QBC selection algorithm, one must:

15

Page 18: Active Learning Literature Survey

(a) (b)

Figure 6: Version space examples for (a) linear and (b) axis-parallel box classifiers. Allhypotheses are consistent with the labeled training data in L (as indicated byshaded polygons), but each represents a different model in the version space.

i. be able to construct a committee of models that represent different regionsof the version space, and

ii. have some measure of disagreement among committee members.

Seung et al. (1992) accomplish the first task simply by sampling a commit-tee of two random hypotheses that are consistent with L. For generative modelclasses, this can be done more generally by randomly sampling an arbitrary num-ber of models from some posterior distribution P (θ|L). For example, McCallumand Nigam (1998) do this for naıve Bayes by using the Dirichlet distribution overmodel parameters, whereas Dagan and Engelson (1995) sample hidden Markovmodels—or HMMs—by using the Normal distribution. For other model classes,such as discriminative or non-probabilistic models, Abe and Mamitsuka (1998)have proposed query-by-boosting and query-by-bagging, which employ the well-known ensemble learning methods boosting (Freund and Schapire, 1997) and bag-ging (Breiman, 1996) to construct committees. Melville and Mooney (2004) pro-pose another ensemble-based method that explicitly encourages diversity amongcommittee members. Muslea et al. (2000) construct a committee of two modelsby partitioning the feature space. There is no general agreement in the literatureon the appropriate committee size to use, which may in fact vary by model class orapplication. However, even small committee sizes (e.g., two or three) have beenshown to work well in practice (Seung et al., 1992; McCallum and Nigam, 1998;Settles and Craven, 2008).

16

Page 19: Active Learning Literature Survey

For measuring the level of disagreement, two main approaches have been pro-posed. The first is vote entropy (Dagan and Engelson, 1995):

x∗V E = argmaxx−∑i

V (yi)

Clog

V (yi)

C,

where yi again ranges over all possible labelings, and V (yi) is the number of“votes” that a label receives from among the committee members’ predictions,and C is the committee size. This can be thought of as a QBC generalization ofentropy-based uncertainty sampling. Another disagreement measure that has beenproposed is average Kullback-Leibler (KL) divergence (McCallum and Nigam,1998):

x∗KL = argmaxx

1

C

C∑c=1

D(Pθ(c)‖PC),

where:

D(Pθ(c)‖PC) =∑i

Pθ(c)(yi|x) logPθ(c)(yi|x)PC(yi|x)

.

Here θ(c) represents a particular model in the committee, and C represents the com-mittee as a whole, thus PC(yi|x) = 1

C

∑Cc=1 Pθ(c)(yi|x) is the “consensus” proba-

bility that yi is the correct label. KL divergence (Kullback and Leibler, 1951) isan information-theoretic measure of the difference between two probability dis-tributions. So this disagreement measure considers the most informative queryto be the one with the largest average difference between the label distributionsof any one committee member and the consensus. Other information-theoreticapproaches like Jensen-Shannon divergence have also been used to measure dis-agreement (Melville et al., 2005), as well as the other uncertainty sampling mea-sures discussed in Section 3.1, by pooling the model predictions to estimate classposteriors (Korner and Wrobel, 2006). Note also that in the equations above, suchposterior estimates are based on committee members that cast “hard” votes fortheir respective label predictions. They might also cast “soft” votes using theirposterior label probabilities, which in turn could be weighted by an estimate ofeach committee member’s accuracy.

Aside from the QBC framework, several other query strategies attempt to min-imize the version space as well. For example, Cohn et al. (1994) describe a selec-tive sampling algorithm that uses a committee of two neural networks, the “mostspecific” and “most general” models, which lie at two extremes the version spacegiven the current training set L. Tong and Koller (2000) propose a pool-based

17

Page 20: Active Learning Literature Survey

margin strategy for SVMs which, as it turns out, attempts to minimize the versionspace directly. The membership query algorithms of Angluin (1988) and Kinget al. (2004) can also be interpreted as synthesizing instances de novo that mostconstrain the size of the version space. However, Haussler (1994) shows that thesize of the version space can grow exponentially with the size of L. This meansthat, in general, the version space of an arbitrary model class cannot be explicitlyrepresented in practice. The QBC framework, rather, uses a committee to serve asa subset approximation.

QBC can also be employed in regression settings, i.e., by measuring disagree-ment as the variance among the committee members’ output predictions (Bur-bidge et al., 2007). Note, however, that there is no notion of “version space” formodels that produce continuous outputs, so the interpretation of QBC in regres-sion settings is a bit different. We can think of L as constraining the posterior jointprobability of predicted output variables and the model parameters, P (Y, θ|L)(note that this applies for both regression and classification tasks). By integratingover a set of hypotheses and identifying queries that lie in controversial regions ofthe instance space, the learner attempts to collect data that reduces variance overboth the output predictions and the parameters of the model itself (as opposedto uncertainty sampling, which focuses only on the output variance of a singlehypothesis).

3.3 Expected Model ChangeAnother general active learning framework uses a decision-theoretic approach, se-lecting the instance that would impart the greatest change to the current model ifwe knew its label. An example query strategy in this framework is the “expectedgradient length” (EGL) approach for discriminative probabilistic model classes.This strategy was introduced by Settles et al. (2008b) for active learning in themultiple-instance setting (see Section 6.4), and has also been applied to proba-bilistic sequence models like CRFs (Settles and Craven, 2008).

In theory, the EGL strategy can be applied to any learning problem wheregradient-based training is used. Since discriminative probabilistic models areusually trained using gradient-based optimization, the “change” imparted to themodel can be measured by the length of the training gradient (i.e., the vectorused to re-estimate parameter values). In other words, the learner should querythe instance x which, if labeled and added to L, would result in the new traininggradient of the largest magnitude. Let ∇`θ(L) be the gradient of the objectivefunction ` with respect to the model parameters θ. Now let ∇`θ(L ∪ 〈x, y〉) be

18

Page 21: Active Learning Literature Survey

the new gradient that would be obtained by adding the training tuple 〈x, y〉 to L.Since the query algorithm does not know the true label y in advance, we mustinstead calculate the length as an expectation over the possible labelings:

x∗EGL = argmaxx

∑i

Pθ(yi|x)∥∥∥∇`θ(L ∪ 〈x, yi〉)∥∥∥,

where ‖ · ‖ is, in this case, the Euclidean norm of each resulting gradient vector.Note that, at query time, ‖∇`θ(L)‖ should be nearly zero since ` converged atthe previous round of training. Thus, we can approximate ∇`θ(L ∪ 〈x, yi〉) ≈∇`θ(〈x, yi〉) for computational efficiency, because training instances are usuallyassumed to be independent.

The intuition behind this framework is that it prefers instances that are likelyto most influence the model (i.e., have greatest impact on its parameters), regard-less of the resulting query label. This approach has been shown to work well inempirical studies, but can be computationally expensive if both the feature spaceand set of labelings are very large. Furthermore, the EGL approach can be ledastray if features are not properly scaled. That is, the informativeness of a giveninstance may be over-estimated simply because one of its feature values is unusu-ally large, or the corresponding parameter estimate is larger, both resulting in agradient of high magnitude. Parameter regularization (Chen and Rosenfeld, 2000;Goodman, 2004) can help control this effect somewhat, and it doesn’t appear tobe a significant problem in practice.

3.4 Expected Error ReductionAnother decision-theoretic approach aims to measure not how much the model islikely to change, but how much its generalization error is likely to be reduced. Theidea it to estimate the expected future error of a model trained using L∪ 〈x, y〉 onthe remaining unlabeled instances in U (which is assumed to be representative ofthe test distribution, and used as a sort of validation set), and query the instancewith minimal expected future error (sometimes called risk). One approach is tominimize the expected 0/1-loss:

x∗0/1 = argminx

∑i

Pθ(yi|x)

(U∑u=1

1− Pθ+〈x,yi〉(y|x(u))

),

where θ+〈x,yi〉 refers to the the new model after it has been re-trained with thetraining tuple 〈x, yi〉 added to L. Note that, as with EGL in the previous section,

19

Page 22: Active Learning Literature Survey

we do not know the true label for each query instance, so we approximate usingexpectation over all possible labels under the current model θ. The objective hereis to reduce the expected total number of incorrect predictions. Another, lessstringent objective is to minimize the expected log-loss:

x∗log = argminx

∑i

Pθ(yi|x)

(−

U∑u=1

∑j

Pθ+〈x,yi〉(yj|x(u)) logPθ+〈x,yi〉(yj|x(u))

),

which is equivalent to reducing the expected entropy over U . Another interpreta-tion of this strategy is maximizing the expected information gain of the query x,or (equivalently) the mutual information of the output variables over x and U .

Roy and McCallum (2001) first proposed the expected error reduction frame-work for text classification using naıve Bayes. Zhu et al. (2003) combined thisframework with a semi-supervised learning approach (Section 7.1), resulting ina dramatic improvement over random or uncertainty sampling. Guo and Greiner(2007) employ an “optimistic” variant that biases the expectation toward the mostlikely label for computational convenience, using uncertainty sampling as a fall-back strategy when the oracle provides an unexpected labeling. This frameworkhas the dual advantage of being near-optimal and not being dependent on themodel class. All that is required is an appropriate objective function and a way toestimate posterior label probabilities. For example, strategies in this frameworkhave been successfully used with a variety of models including naıve Bayes (Royand McCallum, 2001), Gaussian random fields (Zhu et al., 2003), logistic regres-sion (Guo and Greiner, 2007), and support vector machines (Moskovitch et al.,2007). In theory, the general approach can be employed not only to minimize lossfunctions, but to optimize any generic performance measure of interest, such asmaximizing precision, recall, F1-measure, or area under the ROC curve.

In most cases, unfortunately, expected error reduction is also the most com-putationally expensive query framework. Not only does it require estimating theexpected future error over U for each query, but a new model must be incre-mentally re-trained for each possible query labeling, which in turn iterates overthe entire pool. This leads to a drastic increase in computational cost. For non-parametric model classes such as Gaussian random fields (Zhu et al., 2003), theincremental training procedure is efficient and exact, making this approach fairlypractical1. For a many other model classes, this is not the case. For example, abinary logistic regression model would require O(ULG) time complexity simply

1The bottleneck in non-parametric models generally not re-training, but inference.

20

Page 23: Active Learning Literature Survey

to choose the next query, where U is the size of the unlabeled pool U , L is thesize of the current training set L, and G is the number of gradient computationsrequired by the by optimization procedure until convergence. A classification taskwith three or more labels using a MaxEnt model (Berger et al., 1996) would re-quire O(M2ULG) time complexity, where M is the number of class labels. For asequence labeling task using CRFs, the complexity explodes to O(TMT+2ULG),where T is the length of an input sequence. Because of this, the applications ofthe expected error reduction framework have mostly only considered simple bi-nary classification tasks. Moreover, because the approach is often still impractical,researchers must resort to Monte Carlo sampling from the pool (Roy and McCal-lum, 2001) to reduce the U term in the previous analysis, or use approximatetraining techniques (Guo and Greiner, 2007) to reduce the G term.

3.5 Variance ReductionMinimizing the expectation of a loss function directly is expensive, and in generalthis cannot be done in closed form. However, we can still reduce generaliza-tion error indirectly by minimizing output variance, which sometimes does havea closed-form solution. Consider a regression problem, where the learning objec-tive is to minimize standard error (i.e., squared-loss). We can take advantage ofthe result of Geman et al. (1992), showing that a learner’s expected future errorcan be decomposed in the following way:

ET[(y − y)2|x

]= E

[(y − E[y|x])2

]+ (EL[y]− E[y|x])2

+ EL[(y − EL[y])2

],

where EL[·] is an expectation over the labeled set L, E[·] is an expectation overthe conditional density P (y|x), and ET is an expectation over both. Here alsoy is shorthand for the model’s predicted output for a given instance x, while yindicates the true label for that instance.

The first term on the right-hand side of this equation is noise, i.e., the varianceof the true label y given only x, which does not depend on the model or trainingdata. Such noise may result from stochastic effects of the method used to obtainthe labels, for example, or because the feature representation is inadequate. Thesecond term is the bias, which represents the error due to the model class itself,e.g., if a linear model is used to learn a function that is only approximately lin-ear. This component of the overall error is invariant given a fixed model class.

21

Page 24: Active Learning Literature Survey

The third term is the model’s variance, which is the remaining component of thelearner’s squared-loss with respect to the target function. Minimizing the vari-ance, then, is guaranteed to minimize the future generalization error of the model(since the learner itself can do nothing about the noise or bias components).

Cohn (1994) and Cohn et al. (1996) present the first statistical analyses ofactive learning for regression in the context of a robot arm kinematics problem,using the estimated distribution of the model’s output σ2

y . They show that thiscan be done in closed-form for neural networks, Gaussian mixture models, andlocally-weighted linear regression. In particular, for neural networks the outputvariance for some instance x can be approximated by (MacKay, 1992):

σ2y(x) ≈

[∂y

∂θ

]T [∂2

∂θ2Sθ(L)

]−1 [∂y

∂θ

]≈ ∇xTF−1∇x,

where Sθ(L) is the squared error of the current model θ on the training set L.In the equation above, the first and last terms are computed using the gradient ofthe model’s predicted output with respect to model parameters, written in short-hand as ∇x. The middle term is the inverse of a covariance matrix representing asecond-order expansion around the objective function S with respect to θ, writtenin shorthand as F . This is also known as the Fisher information matrix (Schervish,1995), and will be discussed in more detail later. An expression for 〈σ2

y〉+x canthen be derived, which is the estimated mean output variance across the inputdistribution after the model has been re-trained on query x and its correspond-ing label. Given the assumptions that the model’s prediction for x is fairly good,that ∇x is locally linear (true for most network configurations), and that varianceis Gaussian, variance can be estimated efficiently in closed form so that actualmodel re-training is not required; more gory details are given by Cohn (1994).The variance reduction query selection strategy then becomes:

x∗V R = argminx

〈σ2y〉+x.

Because this equation represents a smooth function that is differentiable withrespect to any query instance x in the input space, gradient methods can be usedto search for the best possible query that minimizes output variance, and there-fore generalization error. Hence, their approach is an example of query synthesis(Section 2.1), rather than stream-based or pool-based active learning.

This sort of approach is derived from statistical theories of optimal experi-mental design, or OED (Federov, 1972; Chaloner and Verdinelli, 1995). A key

22

Page 25: Active Learning Literature Survey

ingredient of these approaches is Fisher information, which is sometimes writtenI(θ) to make its relationship with model parameters explicit. Formally, Fisherinformation is the variance of the score, which is the partial derivative of the log-likelihood function with respect to the model parameters:

I(θ) = N

∫x

P (x)

∫y

Pθ(y|x)∂2

∂θ2logPθ(y|x),

where there are N independent samples drawn from the input distribution. Thismeasure is convenient because its inverse sets a lower bound on the variance of themodel’s parameter estimates; this result is known as the Cramer-Rao inequality(Cover and Thomas, 2006). In other words, to minimize the variance over itsparameter estimates, an active learner should select data that maximizes its Fisherinformation (or minimizes the inverse thereof). When there is only one parameterin the model, this strategy is straightforward. But for models of K parameters,Fisher information takes the form of a K ×K covariance matrix (denoted earlieras F ), and deciding what exactly to optimize is a bit tricky. In the OED literature,there are three types of optimal designs in such cases:

• A-optimality minimizes the trace of the inverse information matrix,

• D-optimality minimizes the determinant of the inverse matrix, and

• E-optimality minimizes the maximum eigenvalue of the inverse matrix.

E-optimality doesn’t seem to correspond to an obvious utility function, andis not often used in the machine learning literature, though there are some excep-tions (Flaherty et al., 2006). D-optimality, it turns out, is related to minimizingthe expected posterior entropy (Chaloner and Verdinelli, 1995). Since the deter-minant can be thought of as a measure of volume, the D-optimal design criterionessentially aims to minimize the volume of the (noisy) version space, with bound-aries estimated via entropy, which makes it somewhat analogous to the query-by-committee algorithm (Section 3.2).

A-optimal designs are considerably more popular, and aim to reduce the av-erage variance of parameter estimates by focusing on values along the diagonalof the information matrix. A common variant of A-optimal design is to mini-mize tr(AF−1)—the trace of the product of A and the inverse of the informa-tion matrix F—where A is a square, symmetric “reference” matrix. As a specialcase, consider a matrix of rank one: A = ccT, where c is some vector of length

23

Page 26: Active Learning Literature Survey

K (i.e., the same length as the model’s parameter vector). In this case we havetr(AF−1) = cTF−1c, and minimizing this value is sometimes called c-optimality.Note that, if we let c = ∇x, this criterion results in the equation for output vari-ance σ2

y(x) in neural networks defined earlier. Minimizing this variance measurecan be achieved by simply querying on instance x, so the c-optimal criterion canbe viewed as a formalism for uncertainty sampling (Section 3.1).

Recall that we are interested in reducing variance across the input distribution(not merely for a single point in the instance space), thus the A matrix should en-code the whole instance space. MacKay (1992) derived such solutions for regres-sion with neural networks, while Zhang and Oles (2000) and Schein and Ungar(2007) derived similar methods for classification with logistic regression. Con-sider letting the reference matrix A = IU(θ), i.e., the Fisher information of theunlabeled pool of instances U , and letting F = Ix(θ), i.e., the Fisher informa-tion of some query instance x. Using A-optimal design, we can derive the Fisherinformation ratio (Zhang and Oles, 2000):

x∗FIR = argminx

tr(IU(θ)Ix(θ)−1

).

The equation above provides us with a ratio given by the inner product of the twomatrices, which can be interpreted as the model’s output variance across the inputdistribution (as approximated by U) that is not accounted for by x. Querying theinstance which minimizes this ratio is then analogous to minimizing the futureoutput variance once x has been labeled, thus indirectly reducing generalizationerror (with respect to U). The advantage here over error reduction (Section 3.4) isthat the model need not be retrained: the information matrices give us an approx-imation of output variance that simulates retraining. Zhang and Oles (2000) andSchein and Ungar (2007) applied this sort of approach to text classification usingbinary logistic regression. Hoi et al. (2006a) extended this to active text classifica-tion in the batch-mode setting (Section 6.1) in which a set of queriesQ is selectedall at once in an attempt to minimize the ratio between IU(θ) and IQ(θ). Settlesand Craven (2008) have also generalized the Fisher information ratio approach toprobabilistic sequence models such as CRFs.

There are some practical disadvantages to these variance-reduction methods,however, in terms of computational complexity. Estimating output variance re-quires inverting a K × K matrix for each new instance, where K is the numberof parameters in the model θ, resulting in a time complexity of O(UK3), whereU is the size of the query pool U . This quickly becomes intractable for large K,which is a common occurrence in, say, natural language processing tasks. Paass

24

Page 27: Active Learning Literature Survey

and Kindermann (1995) propose a sampling approach based on Markov chains toreduce the U term in this analysis. For inverting the Fisher information matrix andreducing the K3 term, Hoi et al. (2006a) use principal component analysis to re-duce the dimensionality of the parameter space. Alternatively, Settles and Craven(2008) approximate the matrix with its diagonal vector, which can be inverted inonly O(K) time. However, these methods are still empirically much slower thansimpler query strategies like uncertainty sampling.

3.6 Density-Weighted MethodsA central idea of the estimated error and variance reduction frameworks is thatthey focus on the entire input space rather than individual instances. Thus, theyare less prone to querying outliers than simpler query strategies like uncertaintysampling, QBC, and EGL. Figure 7 illustrates this problem for a binary linearclassifier using uncertainty sampling. The least certain instance lies on the classi-fication boundary, but is not “representative” of other instances in the distribution,so knowing its label is unlikely to improve accuracy on the data as a whole. QBCand EGL may exhibit similar behavior, by spending time querying possible out-liers simply because they are controversial, or are expected to impart significantchange in the model. By utilizing the unlabeled pool U when estimating futureerrors and output variances, the estimated error and variance reduction strategiesimplicitly avoid these problems. We can also overcome these problems by mod-eling the input distribution explicitly during query selection.

The information density framework described by Settles and Craven (2008),and further analyzed in Chapter 4 of Settles (2008), is a general density-weightingtechnique. The main idea is that informative instances should not only be thosewhich are uncertain, but also those which are “representative” of the underlyingdistribution (i.e., inhabit dense regions of the input space). Therefore, we wish toquery instances as follows:

x∗ID = argmaxx

φA(x)×

(1

U

U∑u=1

sim(x, x(u))

.

Here, φA(x) represents the informativeness of x according to some “base” querystrategy A, such as an uncertainty sampling or QBC approach. The second termweights the informativeness of x by its average similarity to all other instancesin the input distribution (as approximated by U), subject to a parameter β that

25

Page 28: Active Learning Literature Survey

A

B

Figure 7: An illustration of when uncertainty sampling can be a poor strategy for classifi-cation. Shaded polygons represent labeled instances in L, and circles representunlabeled instances in U . Since A is on the decision boundary, it would bequeried as the most uncertain. However, querying B is likely to result in moreinformation about the data distribution as a whole.

controls the relative importance of the density term. A variant of this might firstcluster U and compute average similarity to instances in the same cluster.

This formulation was presented by Settles and Craven (2008), however it isnot the only strategy to consider density and representativeness in the literature.McCallum and Nigam (1998) also developed a density-weighted QBC approachfor text classification with naıve Bayes, which is a special case of informationdensity. Fujii et al. (1998) considered a query strategy for nearest-neighbor meth-ods that selects queries that are (i) least similar to the labeled instances in L,and (ii) most similar to the unlabeled instances in U . Nguyen and Smeulders(2004) proposed a density-based approach that first clusters instances and tries toavoid querying outliers by propagating label information to instances in the samecluster. Similarly, Xu et al. (2007) use clustering to construct sets of queries forbatch-mode active learning (Section 6.1) with SVMs. Reported results in all theseapproaches are superior to methods that do not consider density or representative-ness measures. Furthermore, Settles and Craven (2008) show that if densities canbe pre-computed efficiently and cached for later use, the time required to selectthe next query is essentially no different than the base informativeness measure(e.g., uncertainty sampling). This is advantageous for conducting active learninginteractively with oracles in real-time.

4 Analysis of Active LearningThis section discusses some of the empirical and theoretical evidence for how andwhen active learning approaches can be successful.

26

Page 29: Active Learning Literature Survey

4.1 Empirical AnalysisAn important question is: “does active learning work?” Most of the empiricalresults in the published literature suggest that it does (e.g., the majority of papersin the bibliography of this survey). Furthermore, consider that software companiesand large-scale research projects such as CiteSeer, Google, IBM, Microsoft, andSiemens are increasingly using active learning technologies in a variety of real-world applications2. Numerous published results and increased industry adoptionseem to indicate that active learning methods have matured to the point of practicaluse in many situations.

As usual, however, there are caveats. In particular, consider that a training setbuilt in cooperation with an active learner is inherently tied to the model that wasused to generate it (i.e., the class of the model selecting the queries). Therefore,the labeled instances are a biased distribution, not drawn i.i.d. from the underlyingnatural density. If one were to change model classes—as we often do in machinelearning when the state of the art advances—this training set may no longer be asuseful to the new model class (see Section 6.6 for more discussion on this topic).Somewhat surprisingly, Schein and Ungar (2007) showed that active learning cansometimes require more labeled instances than passive learning even when us-ing the same model class, in their case logistic regression. Guo and Schuurmans(2008) found that off-the-shelf query strategies, when myopically employed in abatch-mode setting (Section 6.1) are often much worse than random sampling.Gasperin (2009) reported negative results for active learning in an anaphora res-olution task. Baldridge and Palmer (2009) found a curious inconsistency in howwell active learning helps that seems to be correlated with the proficiency of theannotator (specifically, a domain expert was better utilized by an active learnerthan a domain novice, who was better suited to a passive learner).

Nevertheless, active learning does reduce the number of labeled instances re-quired to achieve a given level of accuracy in the majority of reported results(though, admittedly, this may be due to the publication bias). This is often trueeven for simple query strategies like uncertainty sampling. Tomanek and Ols-son (2009) report in a survey that 91% of researchers who used active learningin large-scale annotation projects had their expectations fully or partially met.Despite these findings, the survey also states that 20% of respondents opted notto use active learning in such projects, specifically because they were “not con-vinced that [it] would work well in their scenario.” This is likely because other

2Based on personal communication with (respectively): C. Lee Giles, David “Pablo” Cohn,Prem Melville, Eric Horvitz, and Balaji Krishnapuram.

27

Page 30: Active Learning Literature Survey

subtleties arise when using active learning in practice (implementation overheadamong them). Section 6 discusses some of the more problematic issues for real-world active learning.

4.2 Theoretical AnalysisA strong theoretical case for why and when active learning should work remainssomewhat elusive, although there have been some recent advances. In particular,it would be nice to have some sort of bound on the number of queries requiredto learn a sufficiently accurate model for a given task, and theoretical guaranteesthat this number is less than in the passive supervised setting. Consider the fol-lowing toy learning task to illustrate the potential of active learning. Suppose thatinstances are points lying on a one-dimensional line, and our model class is asimple binary thresholding function g parameterized by θ:

g(x; θ) =

{1 if x > θ, and0 otherwise.

According to the probably approximately correct (PAC) learning model (Valiant,1984), if the underlying data distribution can be perfectly classified by some hy-pothesis θ, then it is enough to draw O(1/ε) random labeled instances, where εis the maximum desired error rate. Now consider a pool-based active learningsetting, in which we can acquire the same number of unlabeled instances fromthis distribution for free (or very inexpensively), and only labels incur a cost. Ifwe arrange these points on the real line, their (unknown) labels are a sequence ofzeros followed by ones, and our goal is to discover the location at which the tran-sition occurs while paying for as few labels as possible. By conducting a simplebinary search through these unlabeled instances, a classifier with error less thanε can be achieved with a mere O(log 1/ε) queries—since all other labels can beinferred—resulting in an exponential reduction in the number of labeled instances.Of course, this is a simple, one-dimensional, noiseless, binary toy learning task.Generalizing this phenomenon to more interesting and realistic problem settingsis the focus of much theoretical work in active learning.

There have been some fairly strong results for the membership query scenario,in which the learner is allowed to create query instances de novo and acquire theirlabels (Angluin, 1988, 2001). However, such instances can be difficult for humansto annotate (Lang and Baum, 1992) and may result in querying outliers, since theyare not created according to the data’s underlying natural density. A great many

28

Page 31: Active Learning Literature Survey

applications for active learning assume that unlabeled data (drawn from a realdistribution) are available, so these results also have limited practical impact.

A stronger early theoretical result in the stream-based and pool-based scenar-ios is an analysis of the query-by-committee (QBC) algorithm by Freund et al.(1997). They show that, under a Bayesian assumption, it is possible to achievegeneralization error ε after seeing O(d/ε) unlabeled instances, where d is theVapnik-Chervonenkis (VC) dimension (Vapnik and Chervonenkis, 1971) of themodel space, and requesting only O(d log 1/ε) labels. This, like the toy exampleabove, is an exponential improvement over the typical O(d/ε) sample complexityof the supervised setting. This result can tempered somewhat by the computa-tional complexity of the QBC algorithm in certain practical situations, but Gilad-Bachrach et al. (2006) offer some improvements by limiting the version space viakernel functions.

Dasgupta et al. (2005) propose a variant of the perceptron update rule whichcan achieve the same label complexity bounds as reported for QBC. Interestingly,they show that a standard perceptron makes a poor active learner in general, re-quiring O(1/ε2) labels as a lower bound. The modified training update rule—originally proposed in a non-active setting by Blum et al. (1996)—is key in achiev-ing the exponential savings. The two main differences between QBC and theirapproach are that (i) QBC is more limited, requiring a Bayesian assumption forthe theoretical analysis, and (ii) QBC can be computationally prohibitive, whereasthe modified perceptron algorithm is much more lightweight and efficient, evensuitable for online learning.

In earlier work, Dasgupta (2004) also provided a variety of theoretical upperand lower bounds for active learning in the more general pool-based setting. Inparticular, if using linear classifiers the sample complexity can grow to O(1/ε) inthe worst case, which offers no improvement over standard supervised learning,but is also no worse. Encouragingly, Balcan et al. (2008) also show that, asymp-totically, certain active learning strategies should always better than supervisedlearning in the limit.

Most of these results have used theoretical frameworks similar to the standardPAC model, and necessarily assume that the learner knows the correct conceptclass in advance. Put another way, they assume that some model in our hypothe-sis class can perfectly classify the instances, and that the data are also noise-free.To address these limitations, there has been some recent theoretical work in ag-nostic active learning (Balcan et al., 2006), which only requires that unlabeledinstances are drawn i.i.d. from a fixed distribution, and even noisy distributionsare allowed. Hanneke (2007) extends this work by providing upper bounds on

29

Page 32: Active Learning Literature Survey

query complexity for the agnostic setting. Dasgupta et al. (2008) propose a some-what more efficient query selection algorithm, by presenting a polynomial-timereduction from active learning to supervised learning for arbitrary input distribu-tions and model classes. These agnostic active learning approaches explicitly usecomplexity bounds to determine which hypotheses still “look viable,” so to speak,and queries can be assessed by how valuable they are in distinguishing amongthese viable hypotheses. Methods such as these have attractive PAC-style con-vergence guarantees and complexity bounds that are, in many cases, significantlybetter than passive learning.

However, most positive theoretical results to date have been based on in-tractable algorithms, or methods otherwise too prohibitively complex and par-ticular to be used in practice. The few analyses performed on efficient algorithmshave assumed uniform or near-uniform input distributions (Balcan et al., 2006;Dasgupta et al., 2005), or severely restricted hypothesis spaces. Furthermore,these studies have largely only been for simple classification problems. In fact,most are limited to binary classification with the goal of minimizing 0/1-loss, andare not easily adapted to other objective functions that may be more appropriatefor many applications. Furthermore, some of these methods require an explicitenumeration over the version space, which is not only often intractable (see thediscussion at the end of Section 3.2), but difficult to even consider for complexlearning models (e.g., heterogeneous ensembles or structured prediction modelsfor sequences, trees, and graphs). However, some recent theoretical work has be-gun to address these issues, coupled with promising empirical results (Dasguptaand Hsu, 2008; Beygelzimer et al., 2009).

5 Problem Setting VariantsThis section discusses some of the generalizations and extensions of traditionalactive learning work into different problem settings.

5.1 Active Learning for Structured OutputsActive learning for classification tasks has been widely studied (e.g., Cohn et al.,1994; Zhang and Oles, 2000; Guo and Greiner, 2007). However, many impor-tant learning problems involve predicting structured outputs on instances, such assequences and trees. Figure 8 illustrates how, for example, an information extrac-tion problem can be viewed as a sequence labeling task. Let x = 〈x1, . . . , xT 〉

30

Page 33: Active Learning Literature Survey

start nullThe

org

ACME

Inc.

ofces

in

loc

announced...

Chicagonull

The ACME Inc. offices in Chicago announced ...

null null nullorg org loc

x =

y = ...

(a) (b)

Figure 8: An information extraction example viewed as a sequence labeling task.(a) A sample input sequence x and corresponding label sequence y. (b) A se-quence model represented as a finite state machine, illustrating the path of〈x,y〉 through the model.

be an observation sequence of length T with a corresponding label sequencey = 〈y1, . . . , yT 〉. Words in a sentence correspond to tokens in the input sequencex, which are mapped to labels in y. Figure 8(a) presents an example 〈x,y〉 pair.The labels indicate whether a given word belongs to a particular entity class ofinterest (org and loc in this case, for “organization” and “location,” respectively)or not (null).

Unlike simpler classification tasks, each instance x in this setting is not rep-resented by a single feature vector, but rather a structured sequence of featurevectors: one for each token (i.e., word). For example, the word “Madison” mightbe described by the features WORD=Madison and CAPITALIZED. However, itcan variously correspond to the labels person (“The fourth U.S. President JamesMadison...”), loc (“The city of Madison, Wisconsin...”), and org (“Madison de-feated St. Cloud in yesterday’s hockey match...”). The appropriate label for a to-ken often depends on its context in the sequence. For sequence-labeling problemslike information extraction, labels are typically predicted by a sequence modelbased on a probabilistic finite state machine, such as CRFs or HMMs. An exam-ple sequence model is shown in Figure 8(b).

Settles and Craven (2008) present and evaluate a large number of active learn-ing algorithms for sequence labeling tasks using probabilistic sequence modelslike CRFs. Most of these algorithms can be generalized to other probabilisticsequence models, such as HMMs (Dagan and Engelson, 1995; Scheffer et al.,2001) and probabilistic context-free grammars (Baldridge and Osborne, 2004;

31

Page 34: Active Learning Literature Survey

Hwa, 2004). Thompson et al. (1999) also propose query strategies for structuredoutput tasks like semantic parsing and information extraction using inductive logicprogramming methods.

5.2 Active Feature Acquisition and ClassificationIn some learning domains, instances may have incomplete feature descriptions.For example, many data mining tasks in modern business are characterized by nat-urally incomplete customer data, due to reasons such as data ownership, client dis-closure, or technological limitations. Consider a credit card company that wishesto model its most profitable customers; the company has access to data on clienttransactions using their own cards, but no data on transactions using cards fromother companies. Here, the task of the model is to classify a customer using in-complete purchase information as the feature set. Similarly, consider a learningmodel used in medical diagnosis which has access to some patient symptom in-formation, but not other data that require complex, expensive, or risky medicalprocedures. Here, the task of the model is to suggest a diagnosis using incompletepatient information as the feature set.

In these domains, active feature acquisition seeks to alleviate these problemsby allowing the learner to request more complete feature information. The as-sumption is that additional features can be obtained at a cost, such as leasing trans-action records from other credit card companies, or running additional diagnosticprocedures. The goal in active feature acquisition is to select the most informativefeatures to obtain during training, rather than randomly or exhaustively acquiringall new features for all training instances. Zheng and Padmanabhan (2002) pro-posed two “single-pass” approaches for this problem. In the first approach, theyattempt to impute the missing values, and then acquire the ones about which themodel has least confidence. As an alternative, they also consider imputing thesevalues, training a classifiers on the imputed training instances, and only acquiringfeature values for the instances which are misclassified. In contrast, incrementalactive feature acquisition may acquire values for a few salient features at a time,either by selecting a small batch of misclassified examples (Melville et al., 2004),or by taking a decision-theoretic approach and acquiring the feature values whichare expected to maximize some utility function (Saar-Tsechansky et al., 2009).

Similarly, work in active classification considers the case in which missingfeature values may be obtained during classification (test time) rather than duringtraining. Greiner et al. (2002) introduced this setting and provided a PAC-styletheoretical analysis of learning such classifiers given a fixed budget. Variants of

32

Page 35: Active Learning Literature Survey

naıve Bayes (Ling et al., 2004) and decision tree classifiers (Chai et al., 2004;Esmeir and Markovitch, 2008) have also been proposed to minimize costs at clas-sification time. Typically, these are evaluated in terms of their total cost (featureacquisition plus misclassification, which must be converted into the same cur-rency) as a function of the number of missing values. This is often flexible enoughto incorporate other types of costs, such as delays between query time and valueacquisition (Sheng and Ling, 2006). Another approach is to model the featureacquisition task as a sequence of decisions to either acquire more information orto terminate and make a prediction, using an HMM (Ji and Carin, 2007).

The difference between these learning settings and typical active learning isthat the “oracle” provides salient feature values rather than training labels. Sincefeature values can be highly variable in their acquisition costs (e.g., running twodifferent medical tests might provide roughly the same predictive power, whileone is half the cost of the other), some of these approaches are related in spirit tocost-sensitive active learning (see Section 6.3).

5.3 Active Class SelectionActive learning assumes that instances are freely or inexpensively obtained, andit is the labeling process that incurs a cost. Imagine the opposite scenario, how-ever, where a learner is allowed to query a known class label, and obtaining eachinstance incurs a cost. This fairly new problem setting is known as active classselection. Lomasky et al. (2007) propose several active class selection query al-gorithms for an “artificial nose” task, in which a machine learns to discriminatebetween different vapor types (the class labels) which must be chemically synthe-sized (to generate the instances). Some of their approaches show significant gainsover uniform class sampling, the “passive” learning equivalent.

5.4 Active ClusteringFor most of this survey, we assume that the learner to be “activized” is supervised,i.e., the task of the learner is to induce a function that accurately predicts a label yfor some new instance x. In contrast, a learning algorithm is called unsupervised ifits job is simply to organize a large amount of unlabeled data in a meaningful way.The main difference is that supervised learners try to map instances into a pre-defined vocabulary of labels, while unsupervised learners exploit latent structure

33

Page 36: Active Learning Literature Survey

in the data alone to find meaningful patterns3. Clustering algorithms are probablythe most common examples of unsupervised learning (e.g., see Chapter 10 ofDuda et al., 2001).

Since active learning generally aims to select data that will reduce the model’sclassification error or label uncertainty, unsupervised active learning may seem abit counter-intuitive. Nevertheless, Hofmann and Buhmann (1998) have proposedan active clustering algorithm for proximity data, based on an expected value ofinformation criterion. The idea is to generate (or subsample) the unlabeled in-stances in such a way that they self-organize into groupings with less overlap ornoise than for clusters induced using random sampling. The authors demonstrateimproved clusterings in computer vision and text retrieval tasks.

Some clustering algorithms operate under certain constraints, e.g., a user canspecify a priori that two instances must belong to the same cluster, or that two oth-ers cannot. Grira et al. (2005) Have explored an active variant of this approach forimage databases, where queries take the form of such “must-link” and “cannot-link” constraints on similar or dissimilar images. Huang and Mitchell (2006) ex-periment with interactively-obtained clustering constraints on both instances andfeatures, and Andrzejewski et al. (2009) address the analogous problem of incor-porating constraints on features in topic modeling (Steyvers and Griffiths, 2007),another popular unsupervised learning technique. Although these last two worksdo not solicit constraints in an active manner, one can easily imagine extendingthem to do so. Active variants for these unsupervised methods are akin to thework on active learning by labeling features discussed in Section 6.4, with thesubtle difference that constraints in the (semi-)supervised case are links betweenfeatures and labels, rather than features (or instances) with one another.

6 Practical ConsiderationsUntil very recently, most active learning research has focused on mechanisms forchoosing queries from the learner’s perspective. In essence, this body of workaddressed the question, “can machines learn with fewer training instances if theyask questions?” By and large, the answer to this question is “yes,” subject to someassumptions. For example, we often assume that there is a single oracle, or thatthe oracle is always correct, or that the cost for labeling queries is either free oruniformly expensive.

3Note that semi-supervised learning (Section 7.1) also tries to exploit the latent structure ofunlabeled data, but with the specific goal of improving label predictions.

34

Page 37: Active Learning Literature Survey

In many real-world situations these assumptions do not hold. As a result, theresearch question for active learning has shifted in recent years to “can machineslearn more economically if they ask questions?” This section describes several ofthe challenges for active learning in practice, and summarizes some the researchthat has addressed these issues to date.

6.1 Batch-Mode Active LearningIn most active learning research, queries are selected in serial, i.e., one at a time.However, sometimes the time required to induce a model is slow or expensive,as with large ensemble methods and many structured prediction tasks (see Sec-tion 5.1). Consider also that sometimes a distributed, parallel labeling environ-ment may be available, e.g., multiple annotators working on different labelingworkstations at the same time on a network. In both of these cases, selectingqueries in serial may be inefficient. By contrast, batch-mode active learning al-lows the learner to query instances in groups, which is better suited to parallellabeling environments or models with slow training procedures.

The challenge in batch-mode active learning is how to properly assemble theoptimal query setQ. Myopically querying the “Q-best” queries according to someinstance-level query strategy often does not work well, since it fails to considerthe overlap in information content among the “best” instances. To address this, afew batch-mode active learning algorithms have been proposed. Brinker (2003)considers an approach for SVMs that explicitly incorporates diversity among in-stances in the batch. Xu et al. (2007) propose a similar approach for SVM activelearning, which also incorporates a density measure (Section 3.6). Specifically,they query the centroids of clusters of instances that lie closest to the decisionboundary. Hoi et al. (2006a,b) extend the Fisher information framework (Sec-tion 3.5) to the batch-mode setting for binary logistic regression. Most of theseapproaches use greedy heuristics to ensure that instances in the batch are bothdiverse and informative, although Hoi et al. (2006b) exploit the properties of sub-modular functions (see Section 7.3) to find batches that are guaranteed to be near-optimal. Alternatively, Guo and Schuurmans (2008) treat batch construction forlogistic regression as a discriminative optimization problem, and attempt to con-struct the most informative batch directly. For the most part, these approachesshow improvements over random batch sampling, which in turn is generally bet-ter than simple “Q-best” batch construction.

35

Page 38: Active Learning Literature Survey

6.2 Noisy OraclesAnother strong assumption in most active learning work is that the quality oflabeled data is high. If labels come from an empirical experiment (e.g., in biologi-cal, chemical, or clinical studies), then one can usually expect some noise to resultfrom the instrumentation of experimental setting. Even if labels come from humanexperts, they may not always be reliable, for several reasons. First, some instancesare implicitly difficult for people and machines, and second, people can becomedistracted or fatigued over time, introducing variability in the quality of their an-notations. The recent introduction of Internet-based “crowdsourcing” tools suchas Amazon’s Mechanical Turk4 and the clever use of online annotation games5

have enabled some researchers to attempt to “average out” some of this noise bycheaply obtaining labels from multiple non-experts. Such approaches have beenused to produce gold-standard quality training sets (Snow et al., 2008) and alsoto evaluate learning algorithms on data for which no gold-standard labelings exist(Mintz et al., 2009; Carlson et al., 2010).

The question remains about how to use non-experts (or even noisy experts) asoracles in active learning. In particular, when should the learner decide to queryfor the (potentially noisy) label of a new unlabeled instance, versus querying forrepeated labels to de-noise an existing training instance that seems a bit off? Shenget al. (2008) study this problem using several heuristics that take into account es-timates of both oracle and model uncertainty, and show that data can be improvedby selective repeated labeling. However, their analysis assumes that (i) all oraclesare equally and consistently noisy, and (ii) annotation is a noisy process over someunderlying true label. Donmez et al. (2009) address the first issue by allowing an-notators to have different noise levels, and show that both true instance labels andindividual oracle qualities can be estimated (so long as they do not change overtime). They take advantage of these estimates by querying only the more reliableannotators in subsequent iterations active learning.

There are still many open research questions along these lines. For example,how can active learners deal with noisy oracles whose quality varies over time(e.g., after becoming more familiar with the task, or after becoming fatigued)?How might the effect of payment influence annotation quality (i.e., if you pay anon-expert twice as much, are they likely to try and be more accurate)? Whatif some instances are inherently noisy regardless of which oracle is used, andrepeated labeling is not likely to improve matters? Finally, in most crowdsourcing

4http://www.mturk.com5http://www.gwap.com

36

Page 39: Active Learning Literature Survey

environments the users are not necessarily available “on demand,” thus accurateestimates of annotator quality may be difficult to achieve in the first place, andmight possibly never be applicable again, since the model has no real choice overwhich oracles to use. How might the learner continue to make progress?

6.3 Variable Labeling CostsContinuing in the spirit of the previous section, in many applications there is vari-ance not only in label quality from one instance to the next, but also in the cost ofobtaining that label. If our goal in active learning is to minimize the overall cost oftraining an accurate model, then simply reducing the number of labeled instancesdoes not necessarily guarantee a reduction in overall labeling cost. One proposedapproach for reducing annotation effort in active learning involves using the cur-rent trained model to assist in the labeling of query instances by pre-labeling themin structured learning tasks like parsing (Baldridge and Osborne, 2004) or infor-mation extraction (Culotta and McCallum, 2005). However, such methods do notactually represent or reason about labeling costs. Instead, they attempt to reducecost indirectly by minimizing the number of annotation actions required for aquery that has already been selected.

Another group of cost-sensitive active learning approaches explicitly accountsfor varying label costs while selecting queries. Kapoor et al. (2007) propose adecision-theoretic approach that takes into account both labeling costs and mis-classification costs. In this setting, each candidate query is evaluated by summingits labeling cost with the future misclassification costs that are expected to be in-curred if the instance were added to the training set. Instead of using real costs,however, their experiments make the simplifying assumption that the cost of la-beling an instances is a linear function of its length (e.g., one cent per second forvoicemail messages). Furthermore, labeling and misclassification costs must bemapped into the same currency (e.g., $0.01 per second of annotation and $10 permisclassification), which may not be appropriate or straightforward for some ap-plications. King et al. (2004) use a similar decision-theoretic approach to reduceactual labeling costs. They describe a “robot scientist” which can execute a seriesof autonomous biological experiments to discover metabolic pathways, with theobjective of minimizing the cost of materials used (i.e., the cost of an experimentplus the expected total cost of future experiments until the correct hypothesis isfound). But here again, the cost of materials is fixed and known at the time ofexperiment (query) selection.

37

Page 40: Active Learning Literature Survey

In all the settings above, and indeed in most of the cost-sensitive active learn-ing literature (e.g., Margineantu, 2005; Tomanek et al., 2007), the cost of anno-tating an instance is still assumed to be fixed and known to the learner beforequerying. Settles et al. (2008a) propose a novel approach to cost-sensitive activelearning in settings where annotation costs are variable and not known, for exam-ple, when the labeling cost is a function of elapsed annotation time. They learna regression cost-model (alongside the active task-model) which tries to predictthe real, unknown annotation cost based on a few simple “meta features” on theinstances. An analysis of four data sets using real-world human annotation costsreveals the following (Settles et al., 2008a):

• In some domains, annotation costs are not (approximately) constant acrossinstances, and can vary considerably. This result is also supported by thesubsequent findings of others, working on different learning tasks (Aroraet al., 2009; Vijayanarasimhan and Grauman, 2009a).

• Consequently, active learning approaches which ignore cost may performno better than random selection (i.e., passive learning).

• The cost of annotating an instance may not be intrinsic, but may insteadvary based on the person doing the annotation. This result is also supportedby the findings of Ringger et al. (2008) and Arora et al. (2009).

• The measured cost for an annotation may include stochastic components.In particular, there are at least two types of noise which affect annotationspeed: jitter (minor variations due to annotator fatigue, latency, etc.) andpause (major variations that should be shorter under normal circumstances).

• Unknown annotation costs can sometimes be accurately predicted, even af-ter seeing only a few training instances. This result is also supported bythe findings of Vijayanarasimhan and Grauman (2009a). Moreover, theselearned cost-models are significantly more accurate than simple cost heuris-tics (e.g., a linear function of document length).

While empirical experiments show that learned cost-models can be trained topredict accurate annotation times, further work is warranted to determine howsuch approximate, predicted labeling costs can be utilized effectively by cost-sensitive active learning systems. Settles et al. show that simply dividing theinformativeness measure (e.g., entropy) by the cost is not necessarily an effective

38

Page 41: Active Learning Literature Survey

cost-reducing strategy for several natural language tasks when compared to ran-dom sampling (even if true costs are known). However, results from Haertel et al.(2008) suggest that this heuristic, which they call return on investment (ROI),is sometimes effective for part-of-speech tagging, although like most work theyuse a fixed heuristic cost model. Vijayanarasimhan and Grauman (2009a) alsodemonstrate potential cost savings in active learning using predicted annotationcosts in a computer vision task using a decision-theoretic approach. It is unclearwhether these disparities are intrinsic, task-specific, or simply a result of differingexperimental assumptions.

Even among methods that do not explicitly reason about annotation cost, sev-eral authors have found that alternative query types (such as labeling featuresrather than instances, see Section 6.4) can lead to reduced annotation costs forhuman oracles (Raghavan et al., 2006; Druck et al., 2009; Vijayanarasimhan andGrauman, 2009a). Interestingly, Baldridge and Palmer (2009) used active learn-ing for morpheme annotation in a rare-language documentation study, using twolive human oracles (one expert and one novice) interactively “in the loop.” Theyfound that the best query strategy differed between the two annotators, in terms ofreducing both labeled corpus size and annotation costs. The domain expert was amore efficient oracle with an uncertainty-based active learner, but semi-automatedannotations—intended to assist in the labeling process—were of little help. Thenovice, however, was more efficient with a passive learner (selecting passages atrandom), but semi-automated annotations were in this case beneficial.

6.4 Alternative Query TypesMost work in active learning assumes that a “query unit” is of the same type asthe target concept to be learned. In other words, if the task is to assign class labelsto text documents, the learner must query a document and the oracle provides itslabel. What other forms might a query take?

Settles et al. (2008b) introduce an alternative query scenario in the context ofmultiple-instance active learning. In multiple-instance (MI) learning, instancesare grouped into bags (i.e., multi-sets), and it is the bags, rather than instances,that are labeled for training. A bag is labeled negative if and only if all of itsinstances are negative. A bag is labeled positive, however, if at least one of itsinstances is positive (note that positive bags may also contain negative instances).A naıve approach to MI learning is to view it as supervised learning with one-sided noise (i.e., all negative instances are truly negative, but some positives areactually negative). However, special MI learning algorithms have been developed

39

Page 42: Active Learning Literature Survey

bag: image = { instances: segments } bag: document = { instances: passages }

(a) (b)

Figure 9: Multiple-instance active learning. (a) In content-based image retrieval, imagesare represented as bags and instances correspond to segmented image regions.An active MI learner may query which segments belong to the object of in-terest, such as the gold medal shown in this image. (b) In text classification,documents are bags and the instances represent passages of text. In MI ac-tive learning, the learner may query specific passages to determine if they arerepresentative of the positive class at hand.

to learn from labeled bags despite this ambiguity. The MI setting was formalizedby Dietterich et al. (1997) in the context of drug activity prediction, and has sincebeen applied to a wide variety of tasks including content-based image retrieval(Maron and Lozano-Perez, 1998; Andrews et al., 2003; Rahmani and Goldman,2006) and text classification (Andrews et al., 2003; Ray and Craven, 2005).

Figure 9 illustrates how the MI representation can be applied to (a) content-based image retrieval (CBIR) and to (b) text classification. For the CBIR task,images are represented as bags and instances correspond to segmented regions ofthe image. A bag representing a given image is labeled positive if the image con-tains some object of interest. The MI paradigm is well-suited to this task becauseonly a few regions of an image may represent the object of interest, such as thegold medal in Figure 9(a). An advantage of the MI representation here is that it issignificantly easier to label an entire image than it is to label each segment, or evena subset of the image segments. For the text classification task, documents can berepresented as bags and instances correspond to short passages (e.g., paragraphs)that comprise each document. The MI representation is compelling for classifi-cation tasks for which document labels are freely available or cheaply obtained

40

Page 43: Active Learning Literature Survey

(e.g., from online indexes and databases), but the target concept is represented byonly a few passages.

For MI learning tasks such as these, it is possible to obtain labels both at thebag level and directly at the instance level. Fully labeling all instances, how-ever, is expensive. Often the rationale for formulating the learning task as anMI problem is that it allows us to take advantage of coarse labelings that maybe available at low cost, or even for free. In MI active learning, however, thelearner is sometimes allowed to query for labels at a finer granularity than the tar-get concept, e.g., querying passages rather than entire documents, or segmentedimage regions rather than entire images. Settles et al. (2008b) focus on this type ofmixed-granularity active learning with a multiple-instance generalization of logis-tic regression. Vijayanarasimhan and Grauman (2009a,b) have extended the ideato SVMs for the image retrieval task, and also explore an approach that interleavesqueries at varying levels of granularity and cost.

Another alternative setting is to query on features rather than (or in additionto) instances. Raghavan et al. (2006) have proposed one such approach, tandemlearning, which can incorporate feature feedback in traditional classification prob-lems. In their work, a text classifier may interleave instance-label queries withfeature-salience queries (e.g., “is the word puck a discriminative feature for clas-sifying sports documents?”). Values for the salient features are then amplified ininstance feature vectors to reflect their relative importance. Raghavan et al. re-ported that interleaving such queries is very effective for text classification, andalso found that words (or features) are often much easier for human annotatorsto label in empirical user studies. Note, however, that these “feature labels” onlyimply their discriminative value and do not tie features to class labels directly.

In recent years, several new methods have been developed for incorporatingfeature-based domain knowledge into supervised and semi-supervised learning(e.g., Haghighi and Klein, 2006; Druck et al., 2008). In this line of work, usersmay specify a set of constraints between features and labels, e.g., “95% of thetime, when the word puck is observed in a document, the class label is hockey.”The learning algorithm then tries to find a set of model parameters that match ex-pected label distributions over the unlabeled pool U against these user-specifiedpriors (for details, see Druck et al., 2008; Mann and McCallum, 2008). Inter-estingly, Mann and McCallum found that specifying many imprecise constraintsis more effective than fewer more precise ones, suggesting that human-specifiedfeature labels (however noisy) are useful if there are enough of them. This begsthe question of how to actively solicit these constraints.

41

Page 44: Active Learning Literature Survey

Druck et al. (2009) propose and evaluate a variety of active query strategiesaimed at gathering useful feature-label constraints. They show that active featurelabeling is more effective than either “passive” feature labeling (using a varietyof strong baselines) or instance-labeling (both passive and active) for two infor-mation extraction tasks. These results held true for both simulated and interac-tive human-annotator experiments. Liang et al. (2009) present a more principledapproach to the problem, grounded in Bayesian experimental design (see Sec-tion 3.5). However, this method is intractable for most real-world problems, andthey also resort to heuristics in practice. Sindhwani et al. (2009) have also ex-plored interleaving class-label queries for both instances and features, which theyrefer to as active dual supervision, in a semi-supervised graphical model.

6.5 Multi-Task Active LearningThe typical active learning setting assumes that there is only one learner tryingto solve a single task. In many real-world problems, however, the same datainstances may be labeled in multiple ways for different subtasks. In such cases, itis likely most economical to label a single instance for all subtasks simultaneously.Therefore, multi-task active learning algorithms assume that a single query willbe labeled for multiple tasks, and attempt to assess the informativeness of a querywith respect to all the learners involved.

Reichart et al. (2008) study a two-task active learning scenario for naturallanguage parsing and named entity recognition (NER), a form of information ex-traction. They propose two methods for actively learning both tasks in tandem.The first is alternating selection, which allows the parser to query sentences inone iteration, and then the NER system to query instances in the next. The secondis rank combination, in which both learners rank the query candidates in the poolindependently, and instances with the highest combined rank are selected for la-beling. In both cases, uncertainty sampling is used as the base selection strategyfor each learner. As one might expect, these methods outperform passive learn-ing for both subtasks, while learning curves for each individual subtask are not asgood as they would have been in the single-task active setting.

Qi et al. (2008) study a different multi-task active learning scenario, in whichimages may be labeled for several binary classification tasks in parallel. For ex-ample, an image might be labeled as containing a beach, sunset, mountain,field, etc., which are not all mutually exclusive; however, they are not entirelyindependent, either. The beach and sunset labels may be highly correlated, forexample, so a simple rank combination might over-estimate the informativeness

42

Page 45: Active Learning Literature Survey

of some instances. They propose and evaluate a novel Bayesian approach, whichtakes into account the mutual information among labels.

6.6 Changing (or Unknown) Model ClassesAs mentioned in Section 4.1, a training set built via active learning comes from abiased distribution, which is implicitly tied to the class of the model used in se-lecting the queries. This can be an issue if we wish to re-use this training data withmodels of a different type, or if we do not even know the appropriate model class(or feature set) for the task to begin with. Fortunately, this is not always a problem.For example, Lewis and Catlett (1994) showed that decision tree classifiers canstill benefit significantly from a training set constructed by an active naıve Bayeslearner using uncertainty sampling. Tomanek et al. (2007) also showed that infor-mation extraction data gathered by a MaxEnt model using QBC can be effectivelyre-used to train CRFs, maintaining cost savings compared with random sampling.Hwa (2001) successfully re-used natural language parsing data selected by onetype of parser to train other types of parsers.

However, Baldridge and Osborne (2004) encountered the exact opposite prob-lem when re-using data selected by one parsing model to train a variety of otherparsers. As an alternative, they perform active learning using a heterogeneousensemble composed of different parser types, and also use semi-automated label-ing to cut down on human annotation effort. This approach helped to reduce thenumber of training examples required for each parser type compared with pas-sive learning. Similarly, Lu and Bongard (2009) employed active learning witha heterogeneous ensemble of neural networks and decision trees, when the moreappropriate model was not known in advance. Their ensemble approach is ableto simultaneously select informative instances for the overall model, as well asbias the constituent weak learners toward the more appropriate model class asit learns. Sugiyama and Rubens (2008) have experimented with an ensemble oflinear regression models using different feature sets, to study cases in which theappropriate feature set is not yet decided upon.

This section brings up a very important issue for active learning in practice. Ifthe best model class and feature set happen to be known in advance—or if theseare not likely to change much in the future—then active learning can probably besafely used. Otherwise, random sampling (at least for pilot studies, until the taskcan be better understood) may be more advisable than taking one’s chances onactive learning with an inappropriate learning model. One viable active approach

43

Page 46: Active Learning Literature Survey

seems to be the use of heterogeneous ensembles in selecting queries, but there isstill much work to be done in this direction.

6.7 Stopping CriteriaA potentially important element of interactive learning applications in general isknowing when to stop learning. One way to think about this is the point at whichthe cost of acquiring new training data is greater than the cost of the errors madeby the current model. Another view is how to recognize when the accuracy ofa learner has reached a plateau, and acquiring more data is likely a waste of re-sources. Since active learning is concerned with improving accuracy while re-maining sensitive to data acquisition costs, it is natural to think about devising a“stopping criterion” for active learning, i.e., a method by which an active learnermay decide to stop asking questions in order to conserve resources.

Several such stopping criteria for active learning have been proposed (Vla-chos, 2008; Bloodgood and Shanker, 2009; Olsson and Tomanek, 2009). Thesemethods are all fairly similar, generally based on the notion that there is an intrin-sic measure of stability or self-confidence within the learner, and active learningceases to be useful once that measure begins to level-off or degrade. Such self-stopping methods seem like a good idea, and may be applicable in certain situ-ations. However, in my own experience, the real stopping criterion for practicalapplications is based on economic or other external factors, which likely comewell before an intrinsic learner-decided threshold.

7 Related Research AreasResearch in active learning is driven by two key ideas: (i) the learner should notbe strictly passive, and (ii) unlabeled data are often readily available or easilyobtained. There are a few related research areas with rich literature as well.

7.1 Semi-Supervised LearningActive learning and semi-supervised learning (for a good introduction, see Zhu,2005b) both traffic in making the most out of unlabeled data. As a result, thereare a few conceptual overlaps between the two areas that are worth considering.For example, a very basic semi-supervised technique is self-training (Yarowsky,1995), in which the learner is first trained with a small amount of labeled data, and

44

Page 47: Active Learning Literature Survey

then used to classify the unlabeled data. Typically the most confident unlabeledinstances, together with their predicted labels, are added to the training set, andthe process repeats. A complementary technique in active learning is uncertaintysampling (see Section 3.1), where the instances about which the model is leastconfident are selected for querying.

Similarly, co-training (Blum and Mitchell, 1998) and multi-view learning (deSa, 1994) use ensemble methods for semi-supervised learning. Initially, separatemodels are trained with the labeled data (usually using separate, conditionallyindependent feature sets), which then classify the unlabeled data, and “teach”the other models with a few unlabeled examples (using predicted labels) aboutwhich they are most confident. This helps to reduce the size of the version space,i.e., the models must agree on the unlabeled data as well as the labeled data.Query-by-committee (see Section 3.2) is an active learning compliment here, asthe committee represents different parts of the version space, and is used to querythe unlabeled instances about which they do not agree.

Through these illustrations, we see that active learning and semi-supervisedlearning attack the same problem from opposite directions. While semi-supervisedmethods exploit what the learner thinks it knows about the unlabeled data, activemethods attempt to explore the unknown aspects6. It is therefore natural to thinkabout combining the two. Some example formulations of semi-supervised activelearning include McCallum and Nigam (1998), Muslea et al. (2000), Zhu et al.(2003), Zhou et al. (2004), Tur et al. (2005), Yu et al. (2006), and Tomanek andHahn (2009).

7.2 Reinforcement LearningIn reinforcement learning (Sutton and Barto, 1998), the learner interacts with theworld via “actions,” and tries to find an optimal policy of behavior with respectto “rewards” it receives from the environment. For example, consider a machinethat is learning how to play chess. In a supervised setting, one might providethe learner with board configurations from a database of chess games along withlabels indicating which moves ultimately resulted in a win or loss. In a rein-forcement setting, however, the machine actually plays the game against real orsimulated opponents (Baxter et al., 2001). Each board configuration (state) allowsfor certain moves (actions), which result in rewards that are positive (e.g., cap-

6One might make the argument that active methods also “exploit” what is known rather than“exploring,” by querying about what isn’t known. This is a minor semantic issue.

45

Page 48: Active Learning Literature Survey

turing the opponent’s queen) or negative (e.g., having its own queen taken). Thelearner aims to improve as it plays more games.

The relationship with active learning is that, in order to perform well, thelearner must be proactive. It is easy to converge on a policy of actions that haveworked well in the past but are sub-optimal or inflexible. In order to improve,a reinforcement learner must take risks and try out actions for which it is uncer-tain about the outcome, just as an active learner requests labels for instances it isuncertain how to label. This is often called the “exploration-exploitation” trade-off in the reinforcement learning literature. Furthermore, Mihalkova and Mooney(2006) consider an explicitly active reinforcement learning approach which aimsto reduce the number of actions required to find an optimal policy.

7.3 Submodular OptimizationRecently, there has been a growing interest in submodular functions (Nemhauseret al., 1978) in machine learning research. Submodularity is a property of setfunctions that intuitively formalizes the idea of “diminishing returns.” That is,adding some instance x to the set A provides more gain in terms of the targetfunction than adding x to a larger set A′, where A ⊆ A′. Informally, since A′ isa superset of A and already contains more information, adding x will not help asmuch. More formally, a set function F is submodular if it satisfies the property:

F (A ∪ {x})− F (A) ≥ F (A′ ∪ {x})− F (A′),

or, equivalently:

F (A) + F (B) ≥ F (A ∪ B) + F (A ∩ B),

for any two sets A and B. The key advantage of submodularity is that, for mono-tonically non-decreasing submodular functions where F (∅) = 0, a greedy algo-rithm for selecting N instances guarantees a performance of (1− 1/e)× F (S∗N),where F (S∗N) is the value of the optimal set of size N . In other words, using agreedy algorithm to optimize a submodular function gives us a lower-bound per-formance guarantee of around 63% of optimal; in practice these greedy solutionsare often within 90% of optimal (Krause, 2008).

In learning settings where there is a fixed budget on gathering data, it is ad-vantageous to formulate (or approximate) the objective function for data selectionas a submodular function, because it guarantees near-optimal results with signif-

46

Page 49: Active Learning Literature Survey

icantly less computational effort7. The relationship to active learning is simple:both aim to maximize some objective function while minimizing data acquisitioncosts (or remaining within a budget). Active learning strategies do not optimize tosubmodular functions in general, but Guestrin et al. (2005) show that maximizingmutual information among sensor locations using Gaussian processes (analogousto active learning by expected error reduction, see Section 3.4) can be approxi-mated with a submodular function. Similarly, Hoi et al. (2006b) formulate theFisher information ratio criterion (Section 3.5) for binary logistic regression as asubmodular function, for use with batch-mode active learning (Section 6.1).

7.4 Equivalence Query LearningAn area closely related to active learning is learning with equivalence queries(Angluin, 1988). Similar to membership query learning (Section 2.1), here thelearner is allowed to synthesize queries de novo. However, instead of generatingan instance to be labeled by the oracle (or any other kind of learning constraint),the learner instead generates a hypothesis of the target concept class, and the or-acle either confirms or denies that the hypothesis is correct. If it is incorrect, theoracle should provide a counter-example, i.e., an instance that would be labeleddifferently by the true concept and the query hypothesis.

There seem to be few practical applications of equivalence query learning,because an oracle often does not know (or cannot provide) an exact descriptionof the concept class for most real-world problems. Otherwise, it would be suffi-cient to create an “expert system” by hand and machine learning is not required.However, it is an interesting intellectual exercise, and learning from combinedmembership and equivalence queries is in fact the basis of a popular inductivelogic game called Zendo8.

7.5 Model Parroting and CompressionDifferent machine learning algorithms possess different properties. In some cases,it is desirable to induce a model using one type of model class, and then “trans-fer” that model’s knowledge to a model of a different class with another set ofproperties. For example, artificial neural networks have been shown to achieve

7Many interesting set optimization problems are NP-hard, and can thus scale exponentially. Sogreedy approaches are usually more efficient.

8http://www.wunderland.com/icehouse/Zendo/

47

Page 50: Active Learning Literature Survey

better generalization accuracy than decision trees for many applications. How-ever, decision trees represent symbolic hypotheses of the learned concept, and aretherefore much more comprehensible to humans, who can inspect the logical rulesand understand what the model has learned. Craven and Shavlik (1996) proposedthe TREPAN (Trees Parroting Networks) algorithm to extract highly accurate de-cision trees from trained artificial neural networks (or similarly opaque modelclasses, such as ensembles), providing comprehensible, symbolic interpretations.Several others (Bucilua et al., 2006; Liang et al., 2008) have adapted this ideato “compress” large, computationally expensive model classes (such as complexensembles or structured-output models) into smaller, more efficient model classes(such as neural networks or simple linear classifiers).

These approaches can be thought of as active learning methods where the ora-cle is in fact another machine learning model (i.e., the one being parroted or com-pressed) rather than, say, a human annotator. In both cases, the “oracle model” canbe trained using a small set of the available labeled data, and the “parrot model” isallowed to query the the oracle model for (i) the labels of any unlabeled data thatis available, or (ii) synthesize new instances de novo. These two model parrotingand compression approaches correspond to the pool-based and membership queryscenarios for active learning, respectively.

8 Conclusion and Final ThoughtsActive learning is a growing area of research in machine learning, no doubt fueledby the reality that data is increasingly easy or inexpensive to obtain but difficult orcostly to label for training. Over the past two decades, there has been much workin formulating and understanding the various ways in which queries are selectedfrom the learner’s perspective (Sections 2 and 3). This has generated a lot ofevidence that the number of labeled examples necessary to train accurate modelscan be effectively reduced in a variety of applications (Section 4).

Drawing on these foundations, the current surge of research seems to be aimedat applying active learning methods in practice, which has introduced many im-portant problem variants and practical concerns (Sections 5 and 6). So this is aninteresting time to be involved in machine learning and active learning in partic-ular, as some basic questions have been answered but many more still remain.These issues span interdisciplinary topics from learning to statistics, cognitivescience, and human-computer interaction to name a few. It is my hope that thissurvey is an effective summary for researchers (like you) who have an interest

48

Page 51: Active Learning Literature Survey

in active learning, helping to identify novel opportunities and solutions for thispromising area of science and technology.

AcknowledgementsThis survey began as a chapter in my PhD thesis. During that phase of my career,I am indebted to my advisor Mark Craven and committee members Jude Shavlik,Xiaojin “Jerry” Zhu, David Page, and Lewis Friedland, who offered valuable feed-back and encouraged me to expand this into a general resource for the machinelearning community. My own research and thinking on active learning has alsobeen shaped by collaborations with several others, including Andrew McCallum,Gregory Druck, and Soumya Ray.

The insights and organization of ideas in the survey are not wholly my own,but draw from the conversations I’ve had with numerous researchers in the field.After putting out the first draft of this document, I received nearly a hundredemails with additions, corrections, and new perspectives, which have all beenwoven into the fabric of this revision; I thank everyone who took (and continuesto take) the time to share your thoughts. In particular, I would like to thank (inalphabetical order): Jason Baldridge, Aron Culotta, Pinar Donmez, Russ Greiner,Carlos Guestrin, Robbie Haertel, Steve Hanneke, Ashish Kapoor, John Lang-ford, Percy Liang, Prem Melville, Tom Mitchell, Clare Monteleoni, Ray Mooney,Foster Provost, Eric Ringger, Teddy Seidenfeld, Katrin Tomanek, and other col-leagues who have discussed active learning with me, both online and in person.

ReferencesN. Abe and H. Mamitsuka. Query learning strategies using boosting and bagging.

In Proceedings of the International Conference on Machine Learning (ICML),pages 1–9. Morgan Kaufmann, 1998.

S. Andrews, I. Tsochantaridis, and T. Hofmann. Support vector machines formultiple-instance learning. In Advances in Neural Information Processing Sys-tems (NIPS), volume 15, pages 561–568. MIT Press, 2003.

D. Andrzejewski, X. Zhu, and M. Craven. Incorporating domain knowledge intotopic modeling via Dirichlet forest priors. In Proceedings of the InternationalConference on Machine Learning (ICML), pages 25–32. ACM Press, 2009.

49

Page 52: Active Learning Literature Survey

D. Angluin. Queries and concept learning. Machine Learning, 2:319–342, 1988.

D. Angluin. Queries revisited. In Proceedings of the International Conference onAlgorithmic Learning Theory, pages 12–31. Springer-Verlag, 2001.

S. Arora, E. Nyberg, and C.P. Rose. Estimating annotation cost for active learningin a multi-annotator environment. In Proceedings of the NAACL HLT Workshopon Active Learning for Natural Language Processing, pages 18–26. ACL Press,2009.

M.F. Balcan, A. Beygelzimer, and J. Langford. Agnostic active learning. In Pro-ceedings of the International Conference on Machine Learning (ICML), pages65–72. ACM Press, 2006.

M.F. Balcan, S. Hanneke, and J. Wortman. The true sample complexity of activelearning. In Proceedings of the Conference on Learning Theory (COLT), pages45–56. Springer, 2008.

J. Baldridge and M. Osborne. Active learning and the total cost of annotation.In Proceedings of the Conference on Empirical Methods in Natural LanguageProcessing (EMNLP), pages 9–16. ACL Press, 2004.

J. Baldridge and A. Palmer. How well does active learning actually work? Time-based evaluation of cost-reduction strategies for language documentation. InProceedings of the Conference on Empirical Methods in Natural LanguageProcessing (EMNLP), pages 296–305. ACL Press, 2009.

J. Baxter, A. Tridgell, and L. Weaver. Reinforcement learning and chess. InJ. Furnkranz and M. Kubat, editors, Machines that Learn to Play Games, pages91–116. Nova Science Publishers, 2001.

A.L. Berger, V.J. Della Pietra, and S.A. Della Pietra. A maximum entropy ap-proach to natural language processing. Computational Linguistics, 22(1):39–71, 1996.

A. Beygelzimer, S. Dasgupta, and J. Langford. Importance-weighted active learn-ing. In Proceedings of the International Conference on Machine Learning(ICML), pages 49–56. ACM Press, 2009.

50

Page 53: Active Learning Literature Survey

M. Bloodgood and V. Shanker. A method for stopping active learning based onstabilizing predictions and the need for user-adjustable stopping. In Proceed-ings of the Conference on Natural Language Learning (CoNLL), pages 39–47.ACL Press, 2009.

A. Blum and T. Mitchell. Combining labeled and unlabeled data with co-training.In Proceedings of the Conference on Learning Theory (COLT), pages 92–100.Morgan Kaufmann, 1998.

A. Blum, A. Frieze, R. Kannan, and S. Vempala. A polynomial-time algorithmfor learning noisy linear threshold functions. In Proceedings of the IEEE Sym-posium on the Foundations of Computer Science, page 330. IEEE Press, 1996.

C. Bonwell and J. Eison. Active Learning: Creating Excitement in the Classroom.Jossey-Bass, 1991.

L. Breiman. Bagging predictors. Machine Learning, 24(2):123–140, 1996.

K. Brinker. Incorporating diversity in active learning with support vector ma-chines. In Proceedings of the International Conference on Machine Learning(ICML), pages 59–66. AAAI Press, 2003.

C. Bucilua, R. Caruana, and A. Niculescu-Mizil. Model compression. In Pro-ceedings of the International Conference on Knowledge Discovery and DataMining (KDD), pages 535–541. ACM Press, 2006.

R. Burbidge, J.J. Rowland, and R.D. King. Active learning for regression basedon query by committee. In Proceedings of Intelligent Data Engineering andAutomated Learning (IDEAL), pages 209–218. Springer, 2007.

A. Carlson, J. Betteridge, R. Wang, E.R. Hruschka Jr, and T. Mitchell. Coupledsemi-supervised learning for information extraction. In Proceedings of the In-ternational Conference on Web Search and Data Mining (WSDM). ACM Press,2010.

X. Chai, L. Deng, Q. Yang, and C.X. Ling. Test-cost sentistive naive Bayes clas-sification. In Proceedings of the IEEE Conference on Data Mining (ICDM),pages 51–5. IEEE Press, 2004.

K. Chaloner and I. Verdinelli. Bayesian experimental design: A review. StatisticalScience, 10(3):237–304, 1995.

51

Page 54: Active Learning Literature Survey

S.F. Chen and R. Rosenfeld. A survey of smoothing techniques for ME models.IEEE Transactions on Speech and Audio Processing, 8(1):37–50, 2000.

D. Cohn. Neural network exploration using optimal experiment design. In Ad-vances in Neural Information Processing Systems (NIPS), volume 6, pages679–686. Morgan Kaufmann, 1994.

D. Cohn, L. Atlas, R. Ladner, M. El-Sharkawi, R. Marks II, M. Aggoune, andD. Park. Training connectionist networks with queries and selective sampling.In Advances in Neural Information Processing Systems (NIPS). Morgan Kauf-mann, 1990.

D. Cohn, L. Atlas, and R. Ladner. Improving generalization with active learning.Machine Learning, 15(2):201–221, 1994.

D. Cohn, Z. Ghahramani, and M.I. Jordan. Active learning with statistical models.Journal of Artificial Intelligence Research, 4:129–145, 1996.

T.M. Cover and J.A. Thomas. Elements of Information Theory. Wiley, 2006.

M. Craven and J. Shavlik. Extracting tree-structured representations of trainednetworks. In Advances in Neural Information Processing Systems (NIPS), vol-ume 8, pages 24–30. MIT Press, 1996.

N. Cressie. Statistics for Spatial Data. Wiley, 1991.

A. Culotta and A. McCallum. Reducing labeling effort for stuctured predictiontasks. In Proceedings of the National Conference on Artificial Intelligence(AAAI), pages 746–751. AAAI Press, 2005.

I. Dagan and S. Engelson. Committee-based sampling for training probabilisticclassifiers. In Proceedings of the International Conference on Machine Learn-ing (ICML), pages 150–157. Morgan Kaufmann, 1995.

S. Dasgupta. Analysis of a greedy active learning strategy. In Advances in Neu-ral Information Processing Systems (NIPS), volume 16, pages 337–344. MITPress, 2004.

S. Dasgupta and D.J. Hsu. Hierarchical sampling for active learning. In Pro-ceedings of the International Conference on Machine Learning (ICML), pages208–215. ACM Press, 2008.

52

Page 55: Active Learning Literature Survey

S. Dasgupta, A. Kalai, and C. Monteleoni. Analysis of perceptron-based activelearning. In Proceedings of the Conference on Learning Theory (COLT), pages249–263. Springer, 2005.

S. Dasgupta, D. Hsu, and C. Monteleoni. A general agnostic active learning al-gorithm. In Advances in Neural Information Processing Systems (NIPS), vol-ume 20, pages 353–360. MIT Press, 2008.

V.R. de Sa. Learning classification with unlabeled data. In Advances in NeuralInformation Processing Systems (NIPS), volume 6, pages 112–119. MIT Press,1994.

T. Dietterich, R. Lathrop, and T. Lozano-Perez. Solving the multiple-instanceproblem with axis-parallel rectangles. Artificial Intelligence, 89:31–71, 1997.

P. Donmez, J. Carbonell, and J. Schneider. Efficiently learning the accuracy oflabeling sources for selective sampling. In Proceedings of the InternationalConference on Knowledge Discovery and Data Mining (KDD), pages 259–268.ACM Press, 2009.

G. Druck, G. Mann, and A. McCallum. Learning from labeled features usinggeneralized expectation criteria. In Proceedings of the ACM SIGIR Conferenceon Research and Development in Information Retrieval, pages 595–602. ACMPress, 2008.

G. Druck, B. Settles, and A. McCallum. Active learning by labeling features.In Proceedings of the Conference on Empirical Methods in Natural LanguageProcessing (EMNLP), pages 81–90. ACL Press, 2009.

R. Duda, P. Hart, and D. Stork. Pattern Classification. Wiley-Interscience, 2001.

S. Esmeir and S. Markovitch. Anytime induction of cost-sensitive trees. In Ad-vances in Neural Information Processing Systems (NIPS), volume 20, pages425–432. MIT Press, 2008.

V. Federov. Theory of Optimal Experiments. Academic Press, 1972.

P. Flaherty, M. Jordan, and A. Arkin. Robust design of biological experiments. InAdvances in Neural Information Processing Systems (NIPS), volume 18, pages363–370. MIT Press, 2006.

53

Page 56: Active Learning Literature Survey

Y. Freund and R.E. Schapire. A decision-theoretic generalization of on-line learn-ing and an application to boosting. Journal of Computer and System Sciences,55(1):119–139, 1997.

Y. Freund, H.S. Seung, E. Shamir, and N. Tishby. Selective samping using thequery by committee algorithm. Machine Learning, 28:133–168, 1997.

A. Fujii, T. Tokunaga, K. Inui, and H. Tanaka. Selective sampling for example-based word sense disambiguation. Computational Linguistics, 24(4):573–597,1998.

C. Gasperin. Active learning for anaphora resolution. In Proceedings of theNAACL HLT Workshop on Active Learning for Natural Language Processing,pages 1–8. ACL Press, 2009.

S. Geman, E. Bienenstock, and R. Doursat. Neural networks and the bias/variancedilemma. Neural Computation, 4:1–58, 1992.

R. Gilad-Bachrach, A. Navot, and N. Tishby. Query by committee made real. InAdvances in Neural Information Processing Systems (NIPS), volume 18, pages443–450. MIT Press, 2006.

J. Goodman. Exponential priors for maximum entropy models. In Proceedings ofHuman Language Technology and the North American Association for Compu-tational Linguistics (HLT-NAACL), pages 305–312. ACL Press, 2004.

R. Greiner, A. Grove, and D. Roth. Learning cost-sensitive active classifiers.Artificial Intelligence, 139:137–174, 2002.

N. Grira, M. Crucianu, and N. Boujemaa. Active semi-supervised fuzzy cluster-ing for image database categorization. In Proceedings the ACM Workshop onMultimedia Information Retrieval (MIR), pages 9–16. ACM Press, 2005.

C. Guestrin, A. Krause, and A.P. Singh. Near-optimal sensor placements in gaus-sian processes. In Proceedings of the International Conference on MachineLearning (ICML), pages 265–272. ACM Press, 2005.

Y. Guo and R. Greiner. Optimistic active learning using mutual information. InProceedings of International Joint Conference on Artificial Intelligence (IJ-CAI), pages 823–829. AAAI Press, 2007.

54

Page 57: Active Learning Literature Survey

Y. Guo and D. Schuurmans. Discriminative batch mode active learning. In Ad-vances in Neural Information Processing Systems (NIPS), number 20, pages593–600. MIT Press, Cambridge, MA, 2008.

R. Haertel, K. Seppi, E. Ringger, and J. Carroll. Return on investment for activelearning. In Proceedings of the NIPS Workshop on Cost-Sensitive Learning,2008.

A. Haghighi and D. Klein. Prototype-driven learning for sequence models. InProceedings of the North American Association for Computational Linguistics(NAACL), pages 320–327. ACL Press, 2006.

S. Hanneke. A bound on the label complexity of agnostic active learning. In Pro-ceedings of the International Conference on Machine Learning (ICML), pages353–360. ACM Press, 2007.

A. Hauptmann, W. Lin, R. Yan, J. Yang, and M.Y. Chen. Extreme video retrieval:joint maximization of human and computer performance. In Proceedings of theACM Workshop on Multimedia Image Retrieval, pages 385–394. ACM Press,2006.

D. Haussler. Learning conjunctive concepts in structural domains. MachineLearning, 4(1):7–40, 1994.

T. Hofmann and J.M. Buhmann. Active data clustering. In Advances in NeuralInformation Processing Systems (NIPS), volume 10, pages 528—534. MorganKaufmann, 1998.

S.C.H. Hoi, R. Jin, and M.R. Lyu. Large-scale text categorization by batch modeactive learning. In Proceedings of the International Conference on the WorldWide Web, pages 633–642. ACM Press, 2006a.

S.C.H. Hoi, R. Jin, J. Zhu, and M.R. Lyu. Batch mode active learning and itsapplication to medical image classification. In Proceedings of the InternationalConference on Machine Learning (ICML), pages 417–424. ACM Press, 2006b.

Y. Huang and T. Mitchell. Text clustering with extended user feedback. In Pro-ceedings of the ACM SIGIR Conference on Research and Development in In-formation Retrieval, pages 413–420. ACM Press, 2006.

55

Page 58: Active Learning Literature Survey

R. Hwa. On minimizing training corpus for parser acquisition. In Proceedingsof the Conference on Natural Language Learning (CoNLL), pages 1–6. ACLPress, 2001.

R. Hwa. Sample selection for statistical parsing. Computational Linguistics, 30(3):73–77, 2004.

S. Ji and L. Carin. Cost-sensitive feature acquisition and classification. PatternRecognition, 40:1474–1485, 2007.

A. Kapoor, E. Horvitz, and S. Basu. Selective supervision: Guiding super-vised learning with decision-theoretic active learning,. In Proceedings of In-ternational Joint Conference on Artificial Intelligence (IJCAI), pages 877–882.AAAI Press, 2007.

R.D. King, K.E. Whelan, F.M. Jones, P.G. Reiser, C.H. Bryant, S.H. Muggleton,D.B. Kell, and S.G. Oliver. Functional genomic hypothesis generation and ex-perimentation by a robot scientist. Nature, 427(6971):247–52, 2004.

R.D. King, J. Rowland, S.G. Oliver, M. Young, W. Aubrey, E. Byrne, M. Liakata,M. Markham, P. Pir, L.N. Soldatova, A. Sparkes, K.E. Whelan, and A. Clare.The automation of science. Science, 324(5923):85–89, 2009.

C. Korner and S. Wrobel. Multi-class ensemble-based active learning. In Pro-ceedings of the European Conference on Machine Learning (ECML), pages687–694. Springer, 2006.

A. Krause. Optimizing Sensing: Theory and Applications. PhD thesis, CarnegieMellon University, 2008.

V. Krishnamurthy. Algorithms for optimal scheduling and management of hiddenmarkov model sensors. IEEE Transactions on Signal Processing, 50(6):1382–1397, 2002.

S. Kullback and R.A. Leibler. On information and sufficiency. Annals of Mathe-matical Statistics, 22:79–86, 1951.

K. Lang. Newsweeder: Learning to filter netnews. In Proceedings of the Inter-national Conference on Machine Learning (ICML), pages 331–339. MorganKaufmann, 1995.

56

Page 59: Active Learning Literature Survey

K. Lang and E. Baum. Query learning can work poorly when a human oracleis used. In Proceedings of the IEEE International Joint Conference on NeuralNetworks, pages 335–340. IEEE Press, 1992.

D. Lewis and J. Catlett. Heterogeneous uncertainty sampling for supervised learn-ing. In Proceedings of the International Conference on Machine Learning(ICML), pages 148–156. Morgan Kaufmann, 1994.

D. Lewis and W. Gale. A sequential algorithm for training text classifiers. InProceedings of the ACM SIGIR Conference on Research and Development inInformation Retrieval, pages 3–12. ACM/Springer, 1994.

P. Liang, H. Daume, and D. Klein. Structure compilation: Trading structure forfeatures. In Proceedings of the International Conference on Machine Learning(ICML), pages 592–599. ACM Press, 2008.

P. Liang, M.I. Jordan, and D. Klein. Learning from measurements in exponentialfamilies. In Proceedings of the International Conference on Machine Learning(ICML), pages 641–648. ACM Press, 2009.

M. Lindenbaum, S. Markovitch, and D. Rusakov. Selective sampling for nearestneighbor classifiers. Machine Learning, 54(2):125–152, 2004.

C.X. Ling, Q. Yang, J. Wang, and S. Zhang. Decision trees with minimal costs.In Proceedings of the International Conference on Machine Learning (ICML),pages 483–486. ACM Press, 2004.

Y. Liu. Active learning with support vector machine applied to gene expressiondata for cancer classification. Journal of Chemical Information and ComputerSciences, 44:1936–1941, 2004.

R. Lomasky, C.E. Brodley, M. Aernecke, D. Walt, and M. Friedl. Active classselection. In Proceedings of the European Conference on Machine Learning(ECML), pages 640–647. Springer, 2007.

Z. Lu and J. Bongard. Exploiting multiple classifier types with active learning.In Proceedings of the Conference on Genetic and Evolutionary Computation(GECCO), pages 1905–1906. ACM Press, 2009.

D. MacKay. Information-based objective functions for active data selection. Neu-ral Computation, 4(4):590–604, 1992.

57

Page 60: Active Learning Literature Survey

G. Mann and A. McCallum. Generalized expectation criteria for semi-supervisedlearning of conditional random fields. In Proceedings of the Association forComputational Linguistics (ACL). ACL Press, 2008.

D. Margineantu. Active cost-sensitive learning. In Proceedings of InternationalJoint Conference on Artificial Intelligence (IJCAI), pages 1622–1623. AAAIPress, 2005.

O. Maron and T. Lozano-Perez. A framework for multiple-instance learning. InAdvances in Neural Information Processing Systems (NIPS), volume 10, pages570–576. MIT Press, 1998.

A. McCallum and K. Nigam. Employing EM in pool-based active learning fortext classification. In Proceedings of the International Conference on MachineLearning (ICML), pages 359–367. Morgan Kaufmann, 1998.

P. Melville and R. Mooney. Diverse ensembles for active learning. In Proceedingsof the International Conference on Machine Learning (ICML), pages 584–591.Morgan Kaufmann, 2004.

P. Melville, M. Saar-Tsechansky, F. Provost, and R. Mooney. Active feature-valueacquisition for classifier induction. In Proceedings of the IEEE Conference onData Mining (ICDM), pages 483–486. IEEE Press, 2004.

P. Melville, S.M. Yang, M. Saar-Tsechansky, and R. Mooney. Active learning forprobability estimation using Jensen-Shannon divergence. In Proceedings of theEuropean Conference on Machine Learning (ECML), pages 268–279. Springer,2005.

L. Mihalkova and R. Mooney. Using active relocation to aid reinforcement learn-ing. In Proccedings of the Florida Artificial Intelligence Research Society(FLAIRS), pages 580–585. AAAI Press, 2006.

M. Mintz, S. Bills, R. Snow, and D. Jurafsky. Distant supervision for relationextraction without labeled data. In Proceedings of the Association for Compu-tational Linguistics (ACL), 2009.

T. Mitchell. Generalization as search. Artificial Intelligence, 18:203–226, 1982.

T. Mitchell. Machine Learning. McGraw-Hill, 1997.

58

Page 61: Active Learning Literature Survey

C. Monteleoni. Learning with Online Constraints: Shifting Concepts and ActiveLearning. PhD thesis, Massachusetts Institute of Technology, 2006.

R. Moskovitch, N. Nissim, D. Stopel, C. Feher, R. Englert, and Y. Elovici. Im-proving the detection of unknown computer worms activity using active learn-ing. In Proceedings of the German Conference on AI, pages 489–493. Springer,2007.

I. Muslea, S. Minton, and C.A. Knoblock. Selective sampling with redundantviews. In Proceedings of the National Conference on Artificial Intelligence(AAAI), pages 621–626. AAAI Press, 2000.

G.L. Nemhauser, L.A. Wolsey, and M.L. Fisher. An analysis of approximationsfor maximizing submodular set functions. Mathematical Programming, 14(1):265–294, 1978.

H.T. Nguyen and A. Smeulders. Active learning using pre-clustering. In Pro-ceedings of the International Conference on Machine Learning (ICML), pages79–86. ACM Press, 2004.

F. Olsson. Bootstrapping Named Entity Recognition by Means of Active MachineLearning. PhD thesis, University of Gothenburg, 2008.

F. Olsson. A literature survey of active machine learning in the context of naturallanguage processing. Technical Report T2009:06, Swedish Institute of Com-puter Science, 2009.

F. Olsson and K. Tomanek. An intrinsic stopping criterion for committee-basedactive learning. In Proceedings of the Conference on Computational NaturalLanguage Learning (CoNLL), pages 138–146. ACL Press, 2009.

G. Paass and J. Kindermann. Bayesian query construction for neural networkmodels. In Advances in Neural Information Processing Systems (NIPS), vol-ume 7, pages 443–450. MIT Press, 1995.

G.J. Qi, X.S. Hua, Y. Rui, J. Tang, and H.J. Zhang. Two-dimensional active learn-ing for image classification. In Proceedings of the Conference on ComputerVision and Pattern Recognition (CVPR), 2008.

H. Raghavan, O. Madani, and R. Jones. Active learning with feedback on bothfeatures and instances. Journal of Machine Learning Research, 7:1655–1686,2006.

59

Page 62: Active Learning Literature Survey

R. Rahmani and S.A. Goldman. MISSL: Multiple-instance semi-supervised learn-ing. In Proceedings of the International Conference on Machine Learning(ICML), pages 705–712. ACM Press, 2006.

S. Ray and M. Craven. Supervised versus multiple instance learning: An empir-ical comparison. In Proceedings of the International Conference on MachineLearning (ICML), pages 697–704. ACM Press, 2005.

R. Reichart, K. Tomanek, U. Hahn, and A Rappoport. Multi-task active learningfor linguistic annotations. In Proceedings of the Association for ComputationalLinguistics (ACL), pages 861–869. ACL Press, 2008.

E. Ringger, M. Carmen, R. Haertel, K. Seppi, D. Lonsdale, P. McClanahan, J. Car-roll, and N. Ellison. Assessing the costs of machine-assisted corpus annotationthrough a user study. In Proceedings of the International Conference on Lan-guage Resources and Evaluation (LREC). European Language Resources As-sociation, 2008.

N. Roy and A. McCallum. Toward optimal active learning through sampling es-timation of error reduction. In Proceedings of the International Conference onMachine Learning (ICML), pages 441–448. Morgan Kaufmann, 2001.

M. Saar-Tsechansky, P. Melville, and F. Provost. Active feature-value acquisition.Management Science, 55(4):664–684, 2009.

T. Scheffer, C. Decomain, and S. Wrobel. Active hidden Markov models for infor-mation extraction. In Proceedings of the International Conference on Advancesin Intelligent Data Analysis (CAIDA), pages 309–318. Springer-Verlag, 2001.

A.I. Schein and L.H. Ungar. Active learning for logistic regression: An evaluation.Machine Learning, 68(3):235–265, 2007.

M.J. Schervish. Theory of Statistics. Springer, 1995.

B. Settles. Curious Machines: Active Learning with Structured Instances. PhDthesis, University of Wisconsin–Madison, 2008.

B. Settles and M. Craven. An analysis of active learning strategies for sequencelabeling tasks. In Proceedings of the Conference on Empirical Methods in Nat-ural Language Processing (EMNLP), pages 1069–1078. ACL Press, 2008.

60

Page 63: Active Learning Literature Survey

B. Settles, M. Craven, and L. Friedland. Active learning with real annotationcosts. In Proceedings of the NIPS Workshop on Cost-Sensitive Learning, pages1–10, 2008a.

B. Settles, M. Craven, and S. Ray. Multiple-instance active learning. In Advancesin Neural Information Processing Systems (NIPS), volume 20, pages 1289–1296. MIT Press, 2008b.

H.S. Seung, M. Opper, and H. Sompolinsky. Query by committee. In Proceed-ings of the ACM Workshop on Computational Learning Theory, pages 287–294,1992.

C.E. Shannon. A mathematical theory of communication. Bell System TechnicalJournal, 27:379–423,623–656, 1948.

V.S. Sheng and C.X. Ling. Feature value acquisition in testing: A sequential batchtest algorithm. In Proceedings of the International Conference on MachineLearning (ICML), pages 809–816. ACM Press, 2006.

V.S. Sheng, F. Provost, and P.G. Ipeirotis. Get another label? improving dataquality and data mining using multiple, noisy labelers. In Proceedings of theInternational Conference on Knowledge Discovery and Data Mining (KDD).ACM Press, 2008.

V. Sindhwani, P. Melville, and R.D. Lawrence. Uncertainty sampling and trans-ductive experimental design for active dual supervision. In Proceedings of theInternational Conference on Machine Learning (ICML), pages 953–960. ACMPress, 2009.

R. Snow, B. O’Connor, D. Jurafsky, and A. Ng. Cheap and fast—but is it good?In Proceedings of the Conference on Empirical Methods in Natural LanguageProcessing (EMNLP), pages 254–263. ACM Press, 2008.

M. Steyvers and T. Griffiths. Probabilistic topic models. In T. Landauer, D.S.McNamara, S. Dennis, and W. Kintsch, editors, Handbook of Latent SemanticAnalysis. Erlbaum, 2007.

M. Sugiyama and N. Rubens. Active learning with model selection in linear re-gression. In Proceedings of the SIAM International Conference on Data Min-ing, pages 518–529. SIAM, 2008.

61

Page 64: Active Learning Literature Survey

R. Sutton and A.G. Barto. Reinforcement Learning: An Introduction. MIT Press,1998.

C.A. Thompson, M.E. Califf, and R.J. Mooney. Active learning for natural lan-guage parsing and information extraction. In Proceedings of the InternationalConference on Machine Learning (ICML), pages 406–414. Morgan Kaufmann,1999.

K. Tomanek and U. Hahn. Semi-supervised active learning for sequence labeling.In Proceedings of the Association for Computational Linguistics (ACL), pages1039–1047. ACL Press, 2009.

K. Tomanek and F. Olsson. A web survey on the use of active learning to supportannotation of text data. In Proceedings of the NAACL HLT Workshop on ActiveLearning for Natural Language Processing, pages 45–48. ACL Press, 2009.

K. Tomanek, J. Wermter, and U. Hahn. An approach to text corpus construc-tion which cuts annotation costs and maintains reusability of annotated data.In Proceedings of the Conference on Empirical Methods in Natural LanguageProcessing (EMNLP), pages 486–495. ACL Press, 2007.

S. Tong. Active Learning: Theory and Applications. PhD thesis, Stanford Uni-versity, 2001.

S. Tong and E. Chang. Support vector machine active learning for image re-trieval. In Proceedings of the ACM International Conference on Multimedia,pages 107–118. ACM Press, 2001.

S. Tong and D. Koller. Support vector machine active learning with applications totext classification. In Proceedings of the International Conference on MachineLearning (ICML), pages 999–1006. Morgan Kaufmann, 2000.

G. Tur, D. Hakkani-Tur, and R.E. Schapire. Combining active and semi-supervised learning for spoken language understanding. Speech Communica-tion, 45(2):171–186, 2005.

L.G. Valiant. A theory of the learnable. Communications of the ACM, 27(11):1134–1142, 1984.

V.N. Vapnik and A. Chervonenkis. On the uniform convergence of relative fre-quencies of events to their probabilities. Theory of Probability and Its Applica-tions, 16:264–280, 1971.

62

Page 65: Active Learning Literature Survey

S. Vijayanarasimhan and K. Grauman. What’s it going to cost you? Predictingeffort vs. informativeness for multi-label image annotations. In Proceedingsof the Conference on Computer Vision and Pattern Recognition (CVPR). IEEEPress, 2009a.

S. Vijayanarasimhan and K. Grauman. Multi-level active prediction of useful im-age annotations for recognition. In Advances in Neural Information ProcessingSystems (NIPS), volume 21, pages 1705–1712. MIT Press, 2009b.

A. Vlachos. A stopping criterion for active learning. Computer Speech and Lan-guage, 22(3):295–312, 2008.

Z. Xu, R. Akella, and Y. Zhang. Incorporating diversity and density in activelearning for relevance feedback. In Proceedings of the European Conferenceon IR Research (ECIR), pages 246–257. Springer-Verlag, 2007.

R. Yan, J. Yang, and A. Hauptmann. Automatically labeling video data usingmulti-class active learning. In Proceedings of the International Conference onComputer Vision, pages 516–523. IEEE Press, 2003.

D. Yarowsky. Unsupervised word sense disambiguation rivaling supervised meth-ods. In Proceedings of the Association for Computational Linguistics (ACL),pages 189–196. ACL Press, 1995.

H. Yu. SVM selective sampling for ranking with application to data retrieval.In Proceedings of the International Conference on Knowledge Discovery andData Mining (KDD), pages 354–363. ACM Press, 2005.

K. Yu, J. Bi, and V. Tresp. Active learning via transductive experimental design.In Proceedings of the International Conference on Machine Learning (ICML),pages 1081–1087. ACM Press, 2006.

C. Zhang and T. Chen. An active learning framework for content based informa-tion retrieval. IEEE Transactions on Multimedia, 4(2):260–268, 2002.

T. Zhang and F.J. Oles. A probability analysis on the value of unlabeled datafor classification problems. In Proceedings of the International Conference onMachine Learning (ICML), pages 1191–1198. Morgan Kaufmann, 2000.

Z. Zheng and B. Padmanabhan. On active learning for data acquisition. In Pro-ceedings of the IEEE Conference on Data Mining (ICDM), pages 562–569.IEEE Press, 2002.

63

Page 66: Active Learning Literature Survey

Z.H. Zhou, K.J. Chen, and Y. Jiang. Exploiting unlabeled data in content-basedimage retrieval. In Proceedings of the European Conference on Machine Learn-ing (ECML), pages 425–435. Springer, 2004.

X. Zhu. Semi-Supervised Learning with Graphs. PhD thesis, Carnegie MellonUniversity, 2005a.

X. Zhu. Semi-supervised learning literature survey. Computer Sciences TechnicalReport 1530, University of Wisconsin–Madison, 2005b.

X. Zhu, J. Lafferty, and Z. Ghahramani. Combining active learning and semi-supervised learning using Gaussian fields and harmonic functions. In Proceed-ings of the ICML Workshop on the Continuum from Labeled to Unlabeled Data,pages 58–65, 2003.

64

Page 67: Active Learning Literature Survey

Index0/1-loss, 19

active class selection, 33active classification, 32active clustering, 33active feature acquisition, 32active learning, 4active learning with feature labels, 41agnostic active learning, 29alternating selection, 42

batch-mode active learning, 35

classification, 4cost-sensitive active learning, 37

dual supervision, 42

entropy, 13equivalence queries, 47expected error reduction, 19expected gradient length (EGL), 18

Fisher information, 22

information density, 25information extraction, 4, 30

KL divergence, 17

learning curves, 7least-confident sampling, 12log-loss, 20

margin sampling, 13membership queries, 9model compression, 47model parroting, 47

multi-task active learning, 42multiple-instance active learning, 39

noisy oracles, 36

optimal experimental design, 22oracle, 4

PAC learning model, 28pool-based sampling, 5, 11

query, 4query strategy, 12query-by-committee (QBC), 15

rank combination, 42region of uncertainty, 10regression, 15, 18, 21reinforcement learning, 45return on investment (ROI), 38

selective sampling, 10semi-supervised learning, 44sequence labeling, 30speech recognition, 4stopping criteria, 44stream-based active learning, 10structured outputs, 30submodular functions, 46

tandem learning, 41

uncertainty sampling, 5, 12

variance reduction, 21VC dimension, 29version space, 10, 15

65