Margin-Based Algorithms for Information Filtering

Similar documents
Foundations of Machine Learning On-Line Learning. Mehryar Mohri Courant Institute and Google Research

Adaptive Sampling Under Low Noise Conditions 1

Online Passive-Aggressive Algorithms


Better Algorithms for Selective Sampling

Online Passive-Aggressive Algorithms

Worst-Case Analysis of the Perceptron and Exponentiated Update Algorithms

Learning noisy linear classifiers via adaptive and selective sampling

New bounds on the price of bandit feedback for mistake-bounded online multiclass learning

Linear Classification and Selective Sampling Under Low Noise Conditions

Online Learning. Jordan Boyd-Graber. University of Colorado Boulder LECTURE 21. Slides adapted from Mohri

Perceptron Mistake Bounds

Agnostic Online learnability

Linear & nonlinear classifiers

Ad Placement Strategies

Statistical Machine Learning Theory. From Multi-class Classification to Structured Output Prediction. Hisashi Kashima.

Statistical Machine Learning Theory. From Multi-class Classification to Structured Output Prediction. Hisashi Kashima.

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

Predicting Graph Labels using Perceptron. Shuang Song

Robust Selective Sampling from Single and Multiple Teachers

Tutorial: PART 1. Online Convex Optimization, A Game- Theoretic Approach to Learning.

Linear Discrimination Functions

Hybrid Machine Learning Algorithms

Online Learning, Mistake Bounds, Perceptron Algorithm

Discriminative Learning can Succeed where Generative Learning Fails

Adaptive Online Gradient Descent

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

1 Overview. 2 Learning from Experts. 2.1 Defining a meaningful benchmark. AM 221: Advanced Optimization Spring 2016

COMS 4771 Introduction to Machine Learning. Nakul Verma

Logistic Regression Logistic

The Perceptron Algorithm

Classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

ECE521 Lecture7. Logistic Regression

Learning theory. Ensemble methods. Boosting. Boosting: history

The Perceptron Algorithm 1

Linear & nonlinear classifiers

ECE521 week 3: 23/26 January 2017

Support vector machines Lecture 4

Click Prediction and Preference Ranking of RSS Feeds

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Foundations of Machine Learning

Evaluation. Andrea Passerini Machine Learning. Evaluation

Some Formal Analysis of Rocchio s Similarity-Based Relevance Feedback Algorithm

Evaluation requires to define performance measures to be optimized

Part of the slides are adapted from Ziko Kolter

Generalization bounds

STA 4273H: Statistical Machine Learning

NEAL: A Neurally Enhanced Approach to Linking Citation and Reference

From Binary to Multiclass Classification. CS 6961: Structured Prediction Spring 2018

The Decision List Machine

Machine Learning for NLP

Logistic Regression: Online, Lazy, Kernelized, Sequential, etc.

COMS 4771 Introduction to Machine Learning. Nakul Verma

Supervised Learning. George Konidaris

The No-Regret Framework for Online Learning

Learning Methods for Linear Detectors

CSE 417T: Introduction to Machine Learning. Lecture 11: Review. Henry Chai 10/02/18

On the Generalization Ability of Online Strongly Convex Programming Algorithms

Efficient and Principled Online Classification Algorithms for Lifelon

Discriminative Models

Beating the Hold-Out: Bounds for K-fold and Progressive Cross-Validation

Optimal and Adaptive Online Learning

Empirical Risk Minimization

1 A Support Vector Machine without Support Vectors

6.036 midterm review. Wednesday, March 18, 15

On Learning µ-perceptron Networks with Binary Weights

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

Discriminative Models

ECE521 Lectures 9 Fully Connected Neural Networks

Bregman Divergences for Data Mining Meta-Algorithms

Worst-Case Bounds for Gaussian Process Models

Multiclass Classification-1

Stephen Scott.

Online Convex Optimization. Gautam Goel, Milan Cvitkovic, and Ellen Feldman CS 159 4/5/2016

COMS 4771 Lecture Boosting 1 / 16

Linear Probability Forecasting

Predicting New Search-Query Cluster Volume

A PROJECT SUMMARY A 1 B TABLE OF CONTENTS B 1

Warm up: risk prediction with logistic regression

Using Additive Expert Ensembles to Cope with Concept Drift

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

Efficient Bandit Algorithms for Online Multiclass Prediction

Applied Machine Learning Lecture 5: Linear classifiers, continued. Richard Johansson

Neural Networks Learning the network: Backprop , Fall 2018 Lecture 4

An Introduction to Statistical Theory of Learning. Nakul Verma Janelia, HHMI

Midterm: CS 6375 Spring 2015 Solutions

CS 6375 Machine Learning

The Power of Selective Memory: Self-Bounded Learning of Prediction Suffix Trees

The Naïve Bayes Classifier. Machine Learning Fall 2017

Perceptron (Theory) + Linear Regression

Solving Classification Problems By Knowledge Sets

Linear Classifiers. Michael Collins. January 18, 2012

Classification and Pattern Recognition

Linear models: the perceptron and closest centroid algorithms. D = {(x i,y i )} n i=1. x i 2 R d 9/3/13. Preliminaries. Chapter 1, 7.

Discriminative Direction for Kernel Classifiers

ECE521 Lecture 7/8. Logistic Regression

Multiclass Classification with Bandit Feedback using Adaptive Regularization

Adaptive Subgradient Methods for Online Learning and Stochastic Optimization John Duchi, Elad Hanzan, Yoram Singer

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

Transcription:

MarginBased Algorithms for Information Filtering Nicolò CesaBianchi DTI University of Milan via Bramante 65 26013 Crema Italy cesabianchi@dti.unimi.it Alex Conconi DTI University of Milan via Bramante 65 26013 Crema Italy conconi@dti.unimi.it Claudio Gentile CRII Università dell Insubria Via Ravasi 2 21100 Varese Italy gentile@dsi.unimi.it Abstract In this work we study an information filtering model where the relevance labels associated to a sequence of feature vectors are realizations of an unknown probabilistic linear function. Building on the analysis of a restricted version of our model we derive a general filtering rule based on the margin of a ridge regression estimator. While our rule may observe the label of a vector only by classfying the vector as relevant experiments on a realworld document filtering problem show that the performance of our rule is close to that of the online classifier which is allowed to observe all labels. These empirical results are complemented by a theoretical analysis where we consider a randomized variant of our rule and prove that its expected number of mistakes is never much larger than that of the optimal filtering rule which knows the hidden linear model. 1 Introduction Systems able to filter out unwanted pieces of information are of crucial importance for several applications. Consider a stream of discrete data that are individually labelled as relevant or nonrelevant according to some fixed relevance criterion; for instance news about a certain topic emails that are not spam or fraud cases from logged data of user behavior. In all of these cases a filter can be used to drop uninteresting parts of the stream forwarding to the user only those data which are likely to fulfil the relevance criterion. From this point of view the filter may be viewed as a simple online binary classifier. However unlike standard online pattern classification tasks where the classifier observes the correct label after each prediction here the relevance of a data element is known only if the filter decides to forward that data element to the user. This learning protocol with partial feedback is known as adaptive filtering in the Information Retrieval community (see e.g. [14]). We formalize the filtering problem as follows. Each element of an arbitrary data sequence is characterized by a feature vector and an associated relevance label (say for relevant and for nonrelevant). At each time the filtering system observes the th feature vector and must decide whether or not to forward it. If the data is forwarded then its relevance label is revealed to the system The research was supported by the European Commission under the KerMIT Project No. IST 200125431.

which may use this information to adapt the filtering criterion. If the data is not forwarded its relevance label remains hidden. We call the th instance of the data sequence and the pair the th example. For simplicity we assume for all. There are two kinds of errors the filtering system can make in judging the relevance of a feature vector. We say that an example is a false positive if and is classified as relevant by the system; similarly we say that is a false negative if and is classified as nonrelevant by the system. Although false negatives remain unknown the filtering system is scored according to the overall number of wrong relevance judgements it makes. That is both false positives and false negatives are counted as mistakes. In this paper we study the filtering problem under the assumption that relevance judgements are generated using an unknown probabilistic linear function. We design filtering rules that maintain a linear hypothesis and use the margin information to decide whether to forward the next instance. Our performance measure is the regret; i.e. the number of wrong judgements made by a filtering rule over and above those made by the rule knowing the probabilistic function used to generate judgements. We show finitetime (nonasymptotical) bounds on the regret that hold for arbitrary sequences of instances. The only other results of this kind we are aware of are those proven in [9] for the apple tasting model. Since in the apple tasting model relevance judgements are chosen adversarially rather than probabilistically we cannot compare their bounds with ours. We report some preliminary experimental results which might suggest the superiority of our methods as opposed to the general transformations developed within the apple tasting framework. As a matter of fact these general transformations do not take margin information into account. In Section 2 we introduce our probabilistic relevance model and make some preliminary observations. In Section 3 we consider a restricted version of the model within which we prove a regret bound for a simple filtering rule called SIMPLEFIL. In Section 4 we generalize this filtering rule and show its good performance on the Reuters Corpus Volume 1. The algorithm employed which we call RIDGEFIL is a linear least squares algorithm inspired by [2]. In that section we also prove within the unrestricted probabilistic model a regret bound for the randomized variant PRIDGEFIL of the general filtering rule. Both RIDGEFIL and its randomized variant can be run with kernels [13] and adapted to the case when the unknown linear function drifts with time. 2 Learning model notational conventions and preliminaries The relevance of is given by a valued random variable (where means relevant ) such that there exists a fixed and unknown vector for which for all. Hence is relevant with probability. The random variables are assumed to be independent whereas we do not make any assumption on the way the sequence!! is generated. In this model we want to perform almost as well as the algorithm that knows and forwards if and only if " #. We consider linearthreshold filtering algorithms that predict the value of through SGN $&% (' ) where % is a dynamically updated weight vector which might be intended as the current approximation to and ' is a suitable timechanging confidence threshold. For any fixed sequence!! of instances we use + to denote the margin and + to denote the margin %. We define the expected regret of the linearthreshold filtering algorithm at time as.& + /' 0123.4 + 015. We observe that in the conditional 76 probability space where + 8' is given we have + 6 + + + + + + + + + 64 8' 092 8 & 0:5 ; 64 (' 0:2 ( 64 012< 5 8' + > ; 64 92 8 64 015? @ A' B+ C ED D 5 8' + >

where we use to denote the Bernoulli random variable which is 1 if and only if predicate is true. Integrating over all possible values of + 8' we obtain.& + (' 015 (4 + 015 D + D. + 8' + 5 (1) D + D.BD + 8' A+ D@ D + D + (' + (2) where the last inequality is Markov s. These (in)equalities will be used in Sections 3 and 4.2 for the analysis of SIMPLEFIL and PRIDGEFIL algorithms. 3 A simplified model We start by analyzing a restricted model where each data element has the same unknown probability of being relevant and we want to perform almost as well as the filtering rule that consistently does the optimal action (i.e. always forwards if : and never forwards otherwise). The analysis of this model is used in Section 4 to guide the design of good filtering rules for the unrestricted model. Let and let be the sample average of where is the number of forwarded data elements in the first time steps and is the fraction of true positives among the elements that have been forwarded. Obviously the optimal rule forwards if and only if. Consider instead the empirical rule that forwards if. To make and only if. This rule makes a mistake only when! the probability of this event go to zero with it suffices that BD CD D CD as which can only happen if increases quickly enough with. Hence data should be forwarded (irrespective to the sign of the estimate ) also when the confidence level for gets too small with respect to. A problem in this argument is that large deviation bounds require for making.d CD D CD small. But in our case is unknown. To fix this we use the condition. This looks dangerous as we use the empirical value of to control the large deviations of itself; however we will show that this approach indeed works. An algorithm which we call SIMPLEFIL implementing the above line of reasoning takes the form of the following simple rule: forward if and only if ' where ' "! # %$. The expected regret at time of SIMPLEFIL is defined as the probability that SIMPLEFIL makes a mistake at time minus the probability that the optimal filtering rule makes a mistake that is 4 on this regret. ' 0 2. & 0 5. The next result shows a logarithmic bound Theorem 1 The expected cumulative regret of SIMPLEFIL after any number ( steps is at most&2cd CD ' (! ). of time Proof sketch. We can bound the actual regret after time steps as follows. From (1) and the definition of the filtering rule we have.& 4 + 8' 015 D CD21 435 %$ 6 "! + 87 D CD ;6< A ; > + 4 /0:5 D CD + 43! +.4 9 %$ 8' 0/ 6 (! 875:

& 7 ; Without loss of generality assume. We now bound ; < and ;6> separately. Since %$ "! implies that %$ we have that ;6< implies for some. Hence we can write ; < 3 %$ 6 (! %$ + 9 %$ 3 6 "! + + 7 + %$ + %$ + + %$ + + (! D D D CD 9/ D D D CD / for "! D CD> D CD / Applying ChernoffHoeffding [11] bounds to which is a sum of @ valued independent random variables we obtain!; < + %$ + 2D CD> D CD! We now bound ;6> by adapting a technique from [8]. Let "# %$ '& "( )$ + )$ +/. $ +. We have ; > D + CD> " $ D CD / " + 10 $ $ 32 3 $ 9 %$ 6 (! + 87 %$ D CD@ " )$ D CD / 3 $. %$ 6 (! + +54 + 7 ;76 ;78 Applying ChernoffHoeffding bounds again we get ;96 + Finally one can easily verify that ;:8 result. 4 Linear least squares filtering C (!. Piecing everything together we get the desired In order to generalize SIMPLEFIL to the original (unrestricted) learning model described in Section 2 we need a lowvariance estimate of the target vector. Let < be the matrix whose columns are the forwarded feature vectors after the first time steps and let be the vector of corresponding observed relevance labels (the index will be momentarily dropped). Note that >?< holds. Consider the least squares estimator @<A<!BC<D of wheree<a< B is the pseudoinverse of <D<. For all belonging to the column space of < this is an unbiased estimator of that is GFB @<D<!BC<HJI EE<A< B%</> @<A< B%<A< To remove the assumption on we make <D< full rank by adding the identity K. This also allows us to replace the pseudoinverse with the standard inverse

1 RIDGEFULL RIDGEFIL FREQUENCY x 10 0.8 FMEASURE 0.6 0.4 0.2 0 34 CATEGORIES Figure 1: measure for each one of the 34 filtering tasks. The measure is defined by where is precision (fraction of relevant documents among the forwarded ones) and is recall (fraction of forwarded documents among the relevant ones). In the plot the filtering rule RIDGEFIL is compared with RIDGEFULL which sees the correct label after each classification. While precision and recall of RIDGEFULL are balanced RIDGEFIL s recall is higher than precision due to the need of forwarding more documents than believed relevant. This in order to make the confidence of the estimator converge to 1 fast enough. Note that in some cases this imbalance causes RIDGEFIL to achieve a slightly better measure than RIDGEFULL. obtaining @K <D< $ <D a sparse version of the ridge regression estimator [12] (the sparsity is due to the fact that we only store in < the forwarded instances i.e. those for which we have a relevance labels). To estimate directly the margin rather than we further modify along the lines of the techniques analyzed in [3 6 15] the sparse ridge regression estimator. More precisely we estimate with the quantity % where the % is defined by % K < %$ < %$ $ $ < %$ %$ (3) Using the ShermanMorrison formula we can then write out the expectation of % as F4% EI which holds for all and all matrices < %$. Let %$ be the number of forwarded instances among %$. In order to generalize to the estimator (3) the analysis of SIMPLEFIL we need to find a large deviation bound of the form 8$BD % A+ D' %$? ) > for all where > goes to zero sufficiently fast as 4. Though we have not been able to find such bounds we report some experimental results showing that algorithms based on (3) and inspired by the analysis of SIMPLEFIL do exhibit a good empirical behavior on realworld data. Moreover in Section 4.2 we prove a bound (not based on the analysis of SIMPLEFIL) on the expected regret of a randomized variant of the algorithm used in the experiments. For this variant we are able to prove a regret bound that scales essentially with the square root of (to be contrasted with the logarithmic regret of SIMPLEFIL). 4.1 Experimental results We ran our experiments using the filtering rule that forwards if SGN ;% 8' where % is the estimator (3) and ' "! # %$ Note that this rule which we call RIDGEFIL is a natural generalization of SIMPLEFIL to the unrestricted learning model; in particular SIMPLEFIL uses a relevance threshold ' of the very same form as RIDGEFIL although SIMPLEFIL s margin is defined differently. We tested our algorithm on a

Algorithm: PRIDGEFIL. Parameters: Real " : ; A. Initialization: % E4 B5 "K<. Loop for 1. Get and let + 9%. 2. If + 9 then forward get label and update as follows: % % + $ ; ; < < < < % E K@ $ <% where if D D % 3D D and 9 is such that D D % D D otherwise;. 3. Else forward with probability. If was forwarded then get label and do the same updates as in 2; otherwise do not make any update. Figure 2: Pseudocode for the filtering algorithm PRIDGEFIL. The performance of this algorithm is analyzed in Theorem 3. document filtering problem based on the first 70000 newswire stories from the Reuters Corpus Volume 1. We selected the 34 Reuters topics whose frequency in the set of 70000 documents was between 1% and 5% (a plausible range for filtering applications). For each topic we defined a filtering task whose relevance judgements were assigned based on whether the document was labelled with that topic or not. Documents were mapped to real vectors using the bagofwords representation. In particular after tokenization we lemmatized the tokens using a generalpurpose finitestate morphological English analyzer and then removed stopwords (we also replaced all digits with a single special character). Document vectors were built removing all words which did not occur at least three times in the corpus and using the TFIDF encoding in the form "! TF' (!7 DF where TF is the word frequency in the document DF is the number of documents containing the word and is the total number of documents (if TF the TFIDF coefficient was also set to ). Finally all document vectors were normalized to length 1. To measure how the choice of the threshold ' affects the filtering performance we ran RIDGEFIL with ' set to zero on the same dataset as a standard online binary classifier (i.e. receiving the correct label after every classification). We call this algorithm RIDGEFULL. Figure 1 illustrates the experimental results. The average measure of RIDGEFULL and RIDGEFIL are respectively and ; hence the threshold compensates pretty well the partial feedback in the filtering setup. On the other hand the standard Perceptron achieves here a measure of in the classification task hence inferior to that of RIDGEFULL. Finally we also tested the appletasting filtering rule (see [9 STAP transformation]) based on the binary classifier RIDGEFULL. This transformation which does not take into consideration the margin exhibited a poor performance and we did not include it in the plot. 4.2 Probabilistic ridge filtering In this section we introduce a probabilistic filtering algorithm derived from the (online) ridge regression algorithm for the class of linear probabilistic relevance functions. The and a algorithm called PRIDGEFIL is sketched in Figure 2. The algorithm takes " probability value as input parameters and maintains a linear hypothesis %. If % then is forwarded and % gets updated according to the following twosteps ridge regressionlike rule. First the intermediate vector % is computed via the standard online ridge regression algorithm using the inverse of matrix. Then the new vector % is obtained by projecting % onto the unit ball where the projection is taken w.r.t. the

distance function ; B% 4 (% ; (%. Note that D D % /D D implies % %. On the other hand if % 0 then is forwarded (and consequently % is updated) with some probability. The analysis of PRIDGEFIL is inspired by the analysis in [1] for a related but different problem and is based on relating the expected regret in a given trial to a measure of the progress of % towards. The following lemma will be useful. Lemma 2 Using the notation of Figure 2 let be the trial when the th update occurs. Then the following inequality holds: + : < < &+ : 8 (! ; % ; B% where D /D denotes the determinant of matrix and ; % 4 (%" ; 8%. Proof sketch. From Lemma 4.2 and Theorem 4.6 in [3] and the fact that D + D D D % D D it follows that + + 8 (! < < ; B% 4 B%. Now the function ; B% 4 % ; %" is a Bregman divergence (e.g. [4 10]) and it can be easily shown that % in Figure 2 is the projection of % onto the convex set D D D D w.r.t. ; i.e. %! %. By a projection property of Bregman divergences (see e.g. the appendix in [10]) it follows that 4 B% / 24 % for all such that D D D D. Putting together gives the desired inequality. ; Theorem 3 Let! D + D! D D. For all if algorithm PRIDGEFIL of Figure 2 is run with 1!! then its expected cumulative regret +.& + 012 +.& + 092 is at most!! " " (! $ ) # " Proof sketch. If is the trial when the th forward takes place we define the random variables $ % ; % & ; B% and ' 8 (! <. If no update occurs in trial we set $ ('. Let ) be the regret of PRIDGEFIL in trial and ) 6 be the regret of the update rule % % in trial. If + E then ) )6 and $ can be lower bounded via Lemma 2. If + 0 then! $ gets lower bounded via Lemma 2 only with probability while for the regret we can only use the trivial bound ). With probability 4 instead ) ) 6 and $. Let be a constant to be specified. We can write <? ) + $ ) + :C + $ + 1> +?! ) +09>! $ +09> (4) Now it is easy to verify that in the conditional space where + is given we have! &+ D + + and + D + + + +. Thus using Lemma 2 and Eq. (4) we can write?! ) + $. ) 6? + 1C + 5 ) 6 + 01C C 6 + A+ ( ' D4 + ;? + :C + 0:C + + ( ' D4 + ;? + 01C 1 This parametrization requires the knowledge of /. It turns out one can remove this assumption at the cost of a slightly more involved proof.

# # # " ) This can be further upper bounded by + A+ + + +? + 1C + + + + 09> ' 5D + & + 9>! ' 2D +& +.01C + + 01C (5) where in the inequality we have dropped from factor? and combined the resulting terms )6 + 01C and )6 + 9> into )6. In turn this term has been bounded as )6 + + " + + by virtue of (2) with '. At this point we work in the conditional space where + is given and distinguish the two cases + 9 and + 01. In the first case we have 1 + + D4 + 2 C! ' D4 + + > ' D4 + whereas in the second case we have 2 9 + A+ D + C ' D +& '? where in both cases we used + 3+. We set and sum over. Notice that + $ D D D D and that + ' 8 "! < < (! $ (e.g. [3] proof of Theorem 4.6 therein). After a few overapproximations (and taking the worst between the two cases + 1 and + 01 ) we obtain " # +? ) " "! $ ) thereby concluding the proof. ; References? C ' D + + ' [1] Abe N. and Long P.M. (1999). Associative reinforcement learning using linear probabilistic concepts. In Proc. ICML 99 Morgan Kaufmann. [2] Auer P. (2000). Using Upper Confidence Bounds for Online Learning. In Proc. FOCS 00 IEEE pages 270 279. [3] Azoury K. and Warmuth M.K. (2001). Relative loss bounds for online density estimation with the exponential family of distributions Machine Learning 43:211 246. [4] Censor Y. and Lent A. (1981). An iterative rowaction method for interval convex programming. Journal of Optimization Theory and Applications 34(3) 321 353. [5] CesaBianchi N. (1999). Analysis of two gradientbased algorithms for online regression. Journal of Computer and System Sciences 59(3):392 411. [6] CesaBianchi N. Conconi A. and Gentile C. (2002). A secondorder Perceptron algorithm. In Proc. COLT 02 pages 121 137. LNAI 2375 Springer. [7] CesaBianchi N. Long P.M. and Warmuth M.K. (1996). Worstcase quadratic loss bounds for prediction using linear functions and gradient descent. IEEE Trans. NN 7(3):604 619. [8] Gavaldà R. and Watanabe O. (2001). Sequential sampling algorithms: Unified analysis and lower bounds. In Proc. SAGA 01 pages 173 187. LNCS 2264 Springer. [9] Helmbold D.P. Littlestone N. and Long P.M. (2000). Apple tasting. Information and Computation 161(2):85 139. [10] Herbster M. and Warmuth M.K. (1998). Tracking the best regressor in Proc. COLT 98 ACM pages 24 31. [11] Hoeffding W. (1963). Probability inequalities for sums of bounded random variables. Journal of the American Statistical Association 58:13 30. [12] Hoerl A. and Kennard R. (1970). Ridge regression: biased estimation for nonorthogonal problems. Technometrics 12:55 67. [13] Vapnik V. (1998). Statistical learning theory. New York: J. Wiley & Sons. [14] Voorhees E. Harman D. (2001). The tenth Text REtrieval Conference. TR 500250 NIST. [15] Vovk V. (2001). Competitive online statistics. International Statistical Review 69:213 248.