Home > Research > Publications & Outputs > Decision Snippet Features

Links

Text available via DOI:

View graph of relations

Decision Snippet Features

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
Close
Publication date5/05/2021
Host publication2020 25th International Conference on Pattern Recognition (ICPR)
PublisherIEEE
Pages4260-4267
Number of pages8
ISBN (electronic)9781728188089
ISBN (print)9781728188096
<mark>Original language</mark>English

Abstract

Decision trees excel at interpretability of their prediction results. To achieve required prediction accuracies, however, often large ensembles of decision trees - random forests - are considered, reducing interpretability due to large size. Additionally, their size slows down inference on modern hardware and restricts their applicability in low-memory embedded devices. We introduce Decision Snippet Features, which are obtained from small subtrees that appear frequently in trained random forests. We subsequently show that linear models on top of these features achieve comparable and sometimes even better predictive performance than the original random forest, while reducing the model size by up to two orders of magnitude.

Bibliographic note

DBLP's bibliographic metadata records provided through http://dblp.org/search/publ/api are distributed under a Creative Commons CC0 1.0 Universal Public Domain Dedication. Although the bibliographic metadata records are provided consistent with CC0 1.0 Dedication, the content described by the metadata records is not. Content may be subject to copyright, rights of privacy, rights of publicity and other restrictions.