TY - GEN
T1 - Metric anomaly detection via asymmetric risk minimization
AU - Kontorovich, Aryeh
AU - Hendler, Danny
AU - Menahem, Eitan
PY - 2011/10/5
Y1 - 2011/10/5
N2 - We propose what appears to be the first anomaly detection framework that learns from positive examples only and is sensitive to substantial differences in the presentation and penalization of normal vs. anomalous points. Our framework introduces a novel type of asymmetry between how false alarms (misclassifications of a normal instance as an anomaly) and missed anomalies (misclassifications of an anomaly as normal) are penalized: whereas each false alarm incurs a unit cost, our model assumes that a high global cost is incurred if one or more anomalies are missed. We define a few natural notions of risk along with efficient minimization algorithms. Our framework is applicable to any metric space with a finite doubling dimension. We make minimalistic assumptions that naturally generalize notions such as margin in Euclidean spaces. We provide a theoretical analysis of the risk and show that under mild conditions, our classifier is asymptotically consistent. The learning algorithms we propose are computationally and statistically efficient and admit a further tradeoff between running time and precision. Some experimental results on real-world data are provided.
AB - We propose what appears to be the first anomaly detection framework that learns from positive examples only and is sensitive to substantial differences in the presentation and penalization of normal vs. anomalous points. Our framework introduces a novel type of asymmetry between how false alarms (misclassifications of a normal instance as an anomaly) and missed anomalies (misclassifications of an anomaly as normal) are penalized: whereas each false alarm incurs a unit cost, our model assumes that a high global cost is incurred if one or more anomalies are missed. We define a few natural notions of risk along with efficient minimization algorithms. Our framework is applicable to any metric space with a finite doubling dimension. We make minimalistic assumptions that naturally generalize notions such as margin in Euclidean spaces. We provide a theoretical analysis of the risk and show that under mild conditions, our classifier is asymptotically consistent. The learning algorithms we propose are computationally and statistically efficient and admit a further tradeoff between running time and precision. Some experimental results on real-world data are provided.
UR - http://www.scopus.com/inward/record.url?scp=80053380023&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-24471-1_2
DO - 10.1007/978-3-642-24471-1_2
M3 - Conference contribution
AN - SCOPUS:80053380023
SN - 9783642244704
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 17
EP - 30
BT - Similarity-Based Pattern Recognition - First International Workshop, SIMBAD 2011, Proceedings
T2 - 1st International Workshop on Similarity-Based Pattern Recognition, SIMBAD 2011
Y2 - 28 September 2011 through 30 September 2011
ER -