Delayed labelling evaluation for data streams

Maciej Grzenda, Heitor Murilo Gomes, Albert Bifet

Research output: Contribution to journalArticlepeer-review

Abstract

A large portion of the stream mining studies on classification rely on the availability of true labels immediately after making predictions. This approach is well exemplified by the test-then-train evaluation, where predictions immediately precede true label arrival. However, in many real scenarios, labels arrive with non-negligible latency. This raises the question of how to evaluate classifiers trained in such circumstances. This question is of particular importance when stream mining models are expected to refine their predictions between acquiring instance data and receiving its true label. In this work, we propose a novel evaluation methodology for data streams when verification latency takes place, namely continuous re-evaluation. It is applied to reference data streams and it is used to differentiate between stream mining techniques in terms of their ability to refine predictions based on newly arriving instances. Our study points out, discusses and shows empirically the importance of considering the delay of instance labels when evaluating classifiers for data streams.

Original languageEnglish
Pages (from-to)1237-1266
Number of pages30
JournalData Mining and Knowledge Discovery
Volume34
Issue number5
DOIs
Publication statusPublished - 1 Sept 2020

Keywords

  • Classification
  • Delayed labels
  • Evaluation procedures
  • Stream mining

Fingerprint

Dive into the research topics of 'Delayed labelling evaluation for data streams'. Together they form a unique fingerprint.

Cite this