Measuring the Shattering coefficient of Decision Tree models

Research output: Contribution to journalArticlepeer-review

Abstract

In spite of the relevance of Decision Trees (DTs), there is still a disconnection between their theoretical and practical results while selecting models to address specific learning tasks. A particular criterion is provided by the Shattering coefficient, a growth function formulated in the context of the Statistical Learning Theory (SLT), which measures the complexity of the algorithm bias as sample sizes increase. In attempt to establish the basis for a relative theoretical complexity analysis, this paper introduces a method to compute the Shattering coefficient of DT models using recurrence equations. Next, we assess the bias of models provided by DT algorithms while solving practical problems as well as their overall learning bounds in light of the SLT. As the main contribution, our results support other researchers to decide on the most adequate DT models to tackle specific supervised learning tasks.

Original languageEnglish
Pages (from-to)443-452
Number of pages10
JournalExpert Systems with Applications
Volume137
DOIs
Publication statusPublished - 15 Dec 2019
Externally publishedYes

Keywords

  • Decision Trees
  • Learning guarantees
  • Shattering coefficient
  • Statistical Learning Theory

Fingerprint

Dive into the research topics of 'Measuring the Shattering coefficient of Decision Tree models'. Together they form a unique fingerprint.

Cite this