Skip to content Skip to navigation
University of Warwick
  • Study
  • |
  • Research
  • |
  • Business
  • |
  • Alumni
  • |
  • News
  • |
  • About

University of Warwick
Publications service & WRAP

Highlight your research

  • WRAP
    • Home
    • Search WRAP
    • Browse by Warwick Author
    • Browse WRAP by Year
    • Browse WRAP by Subject
    • Browse WRAP by Department
    • Browse WRAP by Funder
    • Browse Theses by Department
  • Publications Service
    • Home
    • Search Publications Service
    • Browse by Warwick Author
    • Browse Publications service by Year
    • Browse Publications service by Subject
    • Browse Publications service by Department
    • Browse Publications service by Funder
  • Help & Advice
University of Warwick

The Library

  • Login
  • Admin

Classifier cascades and trees for minimizing feature evaluation cost

Tools
- Tools
+ Tools

Xu, Zhixiang , Kusner, Matt, Weinberger, Kilian Q., Chen, Minmin and Chapelle, Olivier (2014) Classifier cascades and trees for minimizing feature evaluation cost. Journal of Machine Learning Research, 15 (1). pp. 2113-2144.

Research output not available from this repository, contact author.
Official URL: http://dl.acm.org/citation.cfm?id=2670319

Request Changes to record.

Abstract

Machine learning algorithms have successfully entered industry through many real-world applications (e.g., search engines and product recommendations). In these applications, the test-time CPU cost must be budgeted and accounted for. In this paper, we examine two main components of the test-time CPU cost, classifier evaluation cost and feature extraction cost, and show how to balance these costs with the classifier accuracy. Since the computation required for feature extraction dominates the test-time cost of a classifier in these settings, we develop two algorithms to efficiently balance the performance with the test-time cost. Our first contribution describes how to construct and optimize a tree of classifiers, through which test inputs traverse along individual paths. Each path extracts different features and is optimized for a specific sub-partition of the input space. Our second contribution is a natural reduction of the tree of classifiers into a cascade. The cascade is particularly useful for class-imbalanced data sets as the majority of instances can be early-exited out of the cascade when the algorithm is sufficiently confident in its prediction. Because both approaches only compute features for inputs that benefit from them the most, we find our trained classifiers lead to high accuracies at a small fraction of the computational cost.

Item Type: Journal Article
Divisions: Faculty of Science > Statistics
Journal or Publication Title: Journal of Machine Learning Research
Publisher: M I T Press
ISSN: 1532-4435
Official Date: 5 January 2014
Dates:
DateEvent
5 January 2014Published
Volume: 15
Number: 1
Page Range: pp. 2113-2144
Status: Peer Reviewed
Publication Status: Published
Access rights to Published version: Restricted or Subscription Access
Related URLs:
  • Publisher

Request changes or add full text files to a record

Repository staff actions (login required)

View Item View Item
twitter

Email us: wrap@warwick.ac.uk
Contact Details
About Us