Cost effective transfer of reinforcement learning policies

Research output: Contribution to journalArticlepeer-review

Abstract

Many challenging real-world problems require the deployment of ensembles – multiple complementary learning models – to reach acceptable performance levels. While effective, applying the entire ensemble to every sample is costly and often unnecessary. Deep Reinforcement Learning (DRL) offers a cost-effective alternative, where detectors are dynamically chosen based on the output of their predecessors, with their usefulness weighted against their computational cost. Despite their potential, DRL-based solutions are not widely used in ensemble management. This can be attributed to the difficulties in configuring the reward function for each new task, the unpredictable reactions of the DRL agent to changes in the data, and the inability to use common performance metrics (e.g., True and False-Positive Rates, TPR/FPR) to guide the DRL model in a multi-objective environment. In this study, we propose methods for fine-tuning and calibrating DRL-based policies to meet multiple performance goals. Moreover, we present a method for transferring effective security policies from one dataset to another. Finally, we demonstrate that our approach is highly robust against adversarial attacks.

Original languageEnglish
Article number121380
JournalExpert Systems with Applications
Volume237
DOIs
StatePublished - 1 Mar 2024

Keywords

  • Deep reinforcement learning
  • Ensemble learning
  • Machine learning

ASJC Scopus subject areas

  • General Engineering
  • Computer Science Applications
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Cost effective transfer of reinforcement learning policies'. Together they form a unique fingerprint.

Cite this