Cost-effective ensemble models selection using deep reinforcement learning

Yoni Birman, Shaked Hindi, Gilad Katz, Asaf Shabtai

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Ensemble learning – the application of multiple learning models on the same task – is a common technique in multiple domains. While employing multiple models enables reaching higher classification accuracy, this process can be time consuming, costly, and make scaling more difficult. Given that each model may have different capabilities and costs, assigning the most cost-effective set of learners for each sample is challenging. We propose SPIREL, a novel method for cost-effective classification. Our method enables users to directly associate costs to correct/incorrect label assignment, computing resources and run-time, and then dynamically establishes a classification policy. For each analyzed sample, SPIREL dynamically assigns a different set of learning models, as well as its own classification threshold. Extensive evaluation on two large malware datasets – a domain in which the application of multiple analysis tools is common – demonstrates that SPIREL is highly cost-effective, enabling us to reduce running time by ∼80% while decreasing the accuracy and F1-score by only 0.5%. We also show that our approach is both highly transferable across different datasets and adaptable to changes in individual learning model performance.

Original languageEnglish
Pages (from-to)133-148
Number of pages16
JournalInformation Fusion
Volume77
DOIs
StatePublished - 1 Jan 2022

Keywords

  • Android package
  • Malware detection
  • Portable executable
  • Reinforcement learning
  • Transfer learning

Fingerprint

Dive into the research topics of 'Cost-effective ensemble models selection using deep reinforcement learning'. Together they form a unique fingerprint.

Cite this