Gradient boosted bagging for evolving data stream regression

  • Nuwan Gunasekara
  • , Bernhard Pfahringer
  • , Heitor Murilo Gomes
  • , Albert Bifet

Research output: Contribution to journalArticlepeer-review

Abstract

Gradient boosting has been extensively studied in batch learning. Recently, its streaming adaptation, Streaming Gradient Boosted Trees (Sgbt), has surpassed existing state-of-the-art random subspace and random patches methods for streaming classification under various drift scenarios. However, its application in streaming regression remains unexplored. Vanilla Sgbt with squared loss exhibits high variance when applied to streaming regression problems. To address this, we utilize bagging streaming regressors in this work to create Streaming Gradient Boosted Regression (Sgbr). Bagging streaming regressors are employed in two ways: first, as base learners within the existing Sgbt framework, and second, as an ensemble method that aggregates multiple Sgbts. Our extensive experiments on 11 streaming regression datasets, encompassing multiple drift scenarios, demonstrate that the Sgb(Oza), a variant of the first Sgbr category, significantly outperforms current state-of-the-art streaming regression methods in terms of both predictive power and computational cost.

Original languageEnglish
Article number65
JournalData Mining and Knowledge Discovery
Volume39
Issue number5
DOIs
Publication statusPublished - 1 Sept 2025
Externally publishedYes

Keywords

  • Data stream boosting
  • Gradient boosted bagging
  • Gradient boosted regression
  • Gradient boosting
  • Stream learning
  • Streaming gradient boosted bagging
  • Streaming gradient boosted regression
  • Streaming gradient boosting
  • Streaming regression

Fingerprint

Dive into the research topics of 'Gradient boosted bagging for evolving data stream regression'. Together they form a unique fingerprint.

Cite this