TY - JOUR
T1 - Labeling Network Intrusion Detection System (NIDS) Rules with MITRE ATT&CK Techniques
T2 - Machine Learning vs. Large Language Models
AU - Daniel, Nir
AU - Kaiser, Florian Klaus
AU - Giladi, Shay
AU - Sharabi, Sapir
AU - Moyal, Raz
AU - Shpolyansky, Shalev
AU - Murillo, Andres
AU - Elyashar, Aviad
AU - Puzis, Rami
N1 - Publisher Copyright:
© 2025 by the authors.
PY - 2025/2/1
Y1 - 2025/2/1
N2 - Analysts in Security Operations Centers (SOCs) are often occupied with time-consuming investigations of alerts from Network Intrusion Detection Systems (NIDSs). Many NIDS rules lack clear explanations and associations with attack techniques, complicating the alert triage and the generation of attack hypotheses. Large Language Models (LLMs) may be a promising technology to reduce the alert explainability gap by associating rules with attack techniques. In this paper, we investigate the ability of three prominent LLMs (ChatGPT, Claude, and Gemini) to reason about NIDS rules while labeling them with MITRE ATT&CK tactics and techniques. We discuss prompt design and present experiments performed with 973 Snort rules. Our results indicate that while LLMs provide explainable, scalable, and efficient initial mappings, traditional machine learning (ML) models consistently outperform them in accuracy, achieving higher precision, recall, and F1-scores. These results highlight the potential for hybrid LLM-ML approaches to enhance SOC operations and better address the evolving threat landscape. By utilizing automation, the presented methods will enhance the analysis efficiency of SOC alerts, and decrease workloads for analysts.
AB - Analysts in Security Operations Centers (SOCs) are often occupied with time-consuming investigations of alerts from Network Intrusion Detection Systems (NIDSs). Many NIDS rules lack clear explanations and associations with attack techniques, complicating the alert triage and the generation of attack hypotheses. Large Language Models (LLMs) may be a promising technology to reduce the alert explainability gap by associating rules with attack techniques. In this paper, we investigate the ability of three prominent LLMs (ChatGPT, Claude, and Gemini) to reason about NIDS rules while labeling them with MITRE ATT&CK tactics and techniques. We discuss prompt design and present experiments performed with 973 Snort rules. Our results indicate that while LLMs provide explainable, scalable, and efficient initial mappings, traditional machine learning (ML) models consistently outperform them in accuracy, achieving higher precision, recall, and F1-scores. These results highlight the potential for hybrid LLM-ML approaches to enhance SOC operations and better address the evolving threat landscape. By utilizing automation, the presented methods will enhance the analysis efficiency of SOC alerts, and decrease workloads for analysts.
KW - alerts investigation
KW - cyber threat intelligence
KW - natural language processing
UR - http://www.scopus.com/inward/record.url?scp=85218443220&partnerID=8YFLogxK
U2 - 10.3390/bdcc9020023
DO - 10.3390/bdcc9020023
M3 - Article
AN - SCOPUS:85218443220
SN - 2504-2289
VL - 9
JO - Big Data and Cognitive Computing
JF - Big Data and Cognitive Computing
IS - 2
M1 - 23
ER -