Bloodgood, MichaelGrothendieck, JohnWithin the natural language processing (NLP) community, active learning has been widely investigated and applied in order to alleviate the annotation bottleneck faced by developers of new NLP systems and technologies. This paper presents the first theoretical analysis of stopping active learning based on stabilizing predictions (SP). The analysis has revealed three elements that are central to the success of the SP method: (1) bounds on Cohen’s Kappa agreement between successively trained models impose bounds on differences in F-measure performance of the models; (2) since the stop set does not have to be labeled, it can be made large in practice, helping to guarantee that the results transfer to previously unseen streams of examples at test/application time; and (3) good (low variance) sample estimates of Kappa between successive models can be obtained. Proofs of relationships between the level of Kappa agreement and the difference in performance between consecutive models are presented. Specifically, if the Kappa agreement between two models exceeds a threshold T (where T > 0), then the difference in F-measure performance between those models is bounded above by 4(1−T)/T in all cases. If precision of the positive conjunction of the models is assumed to be p, then the bound can be tightened to 4(1−T)/((p+1)T).en-UScomputer scienceartificial intelligencemachine learningactive learningselective samplingquery learningstopping criteriastopping methodsstabilizing predictionsstatistical analysisperformance boundsagreement statisticsagreement metricsannotation bottleneckKappa statisticCohen's KappaF-measureF-scorerelationship between Kappa and F-measurecontingency table analysisnatural language processingcomputational linguisticsAnalysis of Stopping Active Learning based on Stabilizing PredictionsArticle