Recognition: no theorem link
AI-Driven Security Alert Screening and Alert Fatigue Mitigation in Security Operations Centers: A Comprehensive Survey
Pith reviewed 2026-05-12 00:46 UTC · model grok-4.3
The pith
AI methods for security alerts can be grouped into four workflow stages but lack realistic testing and resistance to attacks.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The paper claims that synthesizing the literature produces a four-stage workflow taxonomy—filtering, triage, correlation, and generative augmentation—that captures how AI is applied to alert screening, while simultaneously revealing consistent shortfalls in deployment realism, adversarial robustness, cross-environment validation, and evaluation practice across the reviewed studies.
What carries the argument
Four-stage workflow taxonomy that sequences filtering, triage, correlation, and generative augmentation to structure the alert-screening process.
If this is right
- Future AI systems for alert screening should incorporate explicit tests for adversarial robustness to remain effective against targeted attacks.
- Research must move beyond simulated data to include cross-environment validation before methods can be trusted in operational centers.
- Improved evaluation standards would allow direct comparison of filtering and correlation techniques across different security settings.
- Generative augmentation approaches need integration with human analyst workflows to realize the goal of cognitive security operations centers.
Where Pith is reading between the lines
- The taxonomy offers a ready checklist for practitioners evaluating commercial alert-screening tools.
- Persistent gaps in deployment realism suggest that close collaboration between researchers and operational centers could accelerate progress.
- Addressing evaluation shortfalls might enable standardized benchmarks that speed adoption of reliable AI components.
- The survey's call for trustworthy systems points to a need for ongoing monitoring of AI decision quality after initial deployment.
Load-bearing premise
The literature search captured a representative sample of work and the four-stage taxonomy usefully organizes both the existing research and its remaining gaps.
What would settle it
A broader or later literature search that identifies a large body of studies that cannot be placed into the four stages or that demonstrates consistent real-world deployment success contradicting the reported gaps.
Figures
read the original abstract
Security alert screening is the downstream task of filtering, prioritizing, correlating, and contextualizing alerts for analyst attention in Security Operations Centers. This survey reviews artificial-intelligence-driven alert screening and alert-fatigue mitigation from 2015 to 2026. We synthesize 119 records, including 87 core studies, into a four-stage workflow taxonomy covering filtering, triage, correlation, and generative augmentation. We find persistent gaps in deployment realism, adversarial robustness, cross-environment validation, and evaluation practice. The survey concludes with a research agenda toward trustworthy Cognitive Security Operations Centers.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper is a survey of AI-driven security alert screening and alert fatigue mitigation in Security Operations Centers. It reviews literature from 2015 to 2026, synthesizes 119 records (87 core studies) into a four-stage workflow taxonomy of filtering, triage, correlation, and generative augmentation, identifies persistent gaps in deployment realism, adversarial robustness, cross-environment validation, and evaluation practice, and concludes with a research agenda for trustworthy Cognitive Security Operations Centers.
Significance. If the literature synthesis proves representative, the work provides a useful organizing framework for the field of AI in SOC alert management. The explicit taxonomy and gap analysis could help direct research toward more realistic, robust, and well-evaluated systems, supporting progress in practical security operations.
major comments (2)
- [§2] §2 (Literature Search and Selection): The synthesis of 119 records and 87 core studies is presented without exact Boolean search strings, deduplication procedures, explicit inclusion/exclusion criteria, or a PRISMA-style flow diagram. Only high-level database names and year ranges are supplied. This directly undermines verification of sample representativeness and therefore the reliability of the four-stage taxonomy and the gap analysis that rests upon it.
- [§4] §4 (Taxonomy): The four-stage taxonomy is claimed to organize the field, yet no table or figure maps the 87 core studies to the stages or reports coverage counts per stage. Without this, it is impossible to assess whether the taxonomy is balanced or whether the identified gaps (e.g., adversarial robustness) are supported by the actual distribution of studies.
minor comments (2)
- [Abstract] Abstract: The review window is stated as '2015 to 2026'. Clarify whether this includes projected or forthcoming work or is intended to run only through the submission date.
- [Throughout] Notation and figures: Ensure every acronym is defined on first use and that all tables/figures are explicitly referenced in the body text.
Simulated Author's Rebuttal
We thank the referee for the constructive and detailed feedback. The comments identify important areas for improving the transparency and verifiability of our survey methodology and taxonomy. We address each point below and will revise the manuscript accordingly.
read point-by-point responses
-
Referee: [§2] §2 (Literature Search and Selection): The synthesis of 119 records and 87 core studies is presented without exact Boolean search strings, deduplication procedures, explicit inclusion/exclusion criteria, or a PRISMA-style flow diagram. Only high-level database names and year ranges are supplied. This directly undermines verification of sample representativeness and therefore the reliability of the four-stage taxonomy and the gap analysis that rests upon it.
Authors: We acknowledge that §2 provides only high-level information on the search process and lacks the specific details needed for full reproducibility. In the revised version, we will add the exact Boolean search strings employed for each database, describe the deduplication steps and tools used, state the explicit inclusion/exclusion criteria applied to arrive at the 119 records and 87 core studies, and include a PRISMA-style flow diagram. These changes will directly address the concern about verifying representativeness and thereby support the reliability of the taxonomy and gap analysis. revision: yes
-
Referee: [§4] §4 (Taxonomy): The four-stage taxonomy is claimed to organize the field, yet no table or figure maps the 87 core studies to the stages or reports coverage counts per stage. Without this, it is impossible to assess whether the taxonomy is balanced or whether the identified gaps (e.g., adversarial robustness) are supported by the actual distribution of studies.
Authors: We agree that the absence of an explicit mapping limits the ability to evaluate the taxonomy's balance and the evidentiary basis for the gaps. We will introduce a new table (or supplementary figure) in §4 that assigns each of the 87 core studies to one or more of the four stages (filtering, triage, correlation, generative augmentation), reports the count of studies per stage and major subcategory, and highlights the distribution of topics such as adversarial robustness. This addition will allow readers to assess coverage and confirm that the identified gaps reflect actual under-representation in the literature. revision: yes
Circularity Check
No circularity: survey synthesizes external studies without self-referential derivations
full rationale
This survey paper reviews and organizes 119 external records (87 core studies) into a four-stage taxonomy of filtering, triage, correlation, and generative augmentation. No equations, fitted parameters, predictions, or uniqueness theorems appear in the provided abstract or description. The central claims rest on synthesis of independent prior literature rather than any reduction of outputs to the paper's own inputs by construction. Self-citations, if present, are not load-bearing for the taxonomy or gap analysis, which derive from the reviewed studies themselves. The literature search process raises reproducibility questions but does not match any enumerated circularity pattern.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption The 119 selected records from 2015-2026 provide a representative view of AI-driven alert screening research.
Reference graph
Works this paper leans on
-
[1]
C. Sri Abhijit, Y. Annie Jerusha, S. P. Syed Ibrahim, and V. Varadharajan. 2025. Federated transfer learning for rare attack class detection in network intrusion detection systems.Scientific Reports15, 1 (2025), 33797. doi:10.1038/s41598-025-02068-x
-
[2]
E. Agyepong, Y. Cherdantseva, P. Reinecke, and P. Burnap. 2023. A systematic method for measuring the performance of a cyber security operations centre analyst.Comput. Secur.124 (2023), 102959. doi:10.1016/j.cose.2022.102959
-
[3]
A. Al-Shaer et al . 2024. CyberPal.AI: Empowering LLMs with Expert-Driven Cybersecurity Instructions. https://arxiv.org/abs/2408.18084 arXiv:2408.18084. Manuscript submitted to ACM AI-Driven Alert Screening in SOCs 29
-
[4]
Alahmadi, Louise Axon, and Ivan Martinovic
Bushra A. Alahmadi, Louise Axon, and Ivan Martinovic. 2022. 99% False Positives: A Qualitative Study of SOC Analysts’ Perspectives on Security Alarms. In31st USENIX Security Symposium (USENIX Security 22). USENIX Association, Boston, MA, 2783–2800. https://www.usenix.org/ conference/usenixsecurity22/presentation/alahmadi
work page 2022
- [5]
-
[6]
R. Alotaibi and M. A. Rassam. 2023. Adversarial Machine Learning Attacks against Intrusion Detection Systems: A Survey on Strategies and Defense.Future Internet15, 2 (2023), 62. doi:10.3390/fi15020062
-
[7]
Berkay Celik, Xiangyu Zhang, and Dongyan Xu
Abdulellah Alsaheel, Yuhong Nan, Shiqing Ma, Le Yu, Gregory Walkup, Z. Berkay Celik, Xiangyu Zhang, and Dongyan Xu. 2021. ATLAS: A Sequence-based Learning Approach for Attack Investigation. In30th USENIX Security Symposium (USENIX Security 21). USENIX Association, Berkeley, CA, 3005–3022. https://www.usenix.org/conference/usenixsecurity21/presentation/alsaheel
work page 2021
-
[8]
Amazon Web Services. 2025. Detect and Analyze. https://docs.aws.amazon.com/security-ir/latest/userguide/prepare.html AWS Security Incident Response User Guide, 2025
work page 2025
-
[9]
M. E. Aminanto, T. Ban, R. Isawa, T. Takahashi, and D. Inoue. 2020. Threat Alert Prioritization Using Isolation Forest and Stacked Autoencoder with Day-Forward-Chaining Analysis.IEEE Access8 (2020), 217977–217986. doi:10.1109/ACCESS.2020.3041282
-
[10]
Muhamad Erza Aminanto, Lei Zhu, Tao Ban, Ryoichi Isawa, Takeshi Takahashi, and Daisuke Inoue. 2019. Combating Threat-Alert Fatigue with Online Anomaly Detection Using Isolation Forest. InNeural Information Processing: 26th International Conference, ICONIP 2019, Sydney, NSW, Australia, December 12–15, 2019, Proceedings, Part I(Sydney, NSW, Australia). Spri...
-
[11]
R. W. Andrews, J. M. Lilly, D. Srivastava, and K. M. Feigh. 2023. The role of shared mental models in human-AI teams: a theoretical review.Theor. Issues Ergon. Sci.24, 2 (2023), 129–175. doi:10.1080/1463922X.2022.2061080
-
[12]
G. Apruzzese, P. Laskov, and J. Schneider. 2023. SoK: Pragmatic Assessment of Machine Learning for Network Intrusion Detection. doi:10.1109/ EuroSP57164.2023.00042 IEEE EuroS&P, 2023
-
[13]
Daniel Arp, Erwin Quiring, Feargus Pendlebury, Alexander Warnecke, Fabio Pierazzi, Christian Wressnegger, Lorenzo Cavallaro, and Konrad Rieck
-
[14]
In31st USENIX Security Symposium (USENIX Security 22)
Dos and don’ts of machine learning in computer security. In31st USENIX Security Symposium (USENIX Security 22). USENIX Association, Boston, MA, 3971–3988. https://www.usenix.org/conference/usenixsecurity22/presentation/arp
- [15]
-
[16]
Sikha S. Bagui, Dustin Mink, Subhash C. Bagui, Tirthankar Ghosh, Russel Plenkers, Tom McElroy, Stephan Dulaney, and Sajida Shabanali. 2023. Introducing UWF-ZeekData22: A Comprehensive Network Traffic Dataset Based on the MITRE ATT&CK Framework.Data8, 1 (2023), 1–18. doi:10.3390/data8010018
-
[17]
V. Balasubramanian et al. 2025. Generative AI for cyber threat intelligence: applications, challenges, and analysis of real-world case studies. doi:10.1007/s10462-025-11338-z Artif. Intell. Rev., 2025
-
[18]
Tao Ban, Ndichu Samuel, Takeshi Takahashi, and Daisuke Inoue. 2021. Combat Security Alert Fatigue with AI-Assisted Techniques. InProceedings of the 14th Cyber Security Experimentation and Test Workshop(Virtual, CA, USA)(CSET ’21). Association for Computing Machinery, New York, NY, USA, 9–16. doi:10.1145/3474718.3474723
-
[19]
T. Ban, T. Takahashi, S. Ndichu, and D. Inoue. 2023. Breaking Alert Fatigue: AI-Assisted SIEM Framework for Effective Incident Response.Appl. Sci. 13, 11 (2023), 6610. doi:10.3390/app13116610
-
[20]
Kritan Banstola, Faayed Al Faisal, and Xinming Ou. 2026. Experiences of Using Agentic AI to Fill Tooling Gaps in a Security Operations Center. doi:10.14722/wosoc.2026.23016 Workshop on Security Operations Center (WOSOC), 2026
-
[21]
N. Bhatt et al. 2024. CyberSecEval 2: A Wide-Ranging Cybersecurity Evaluation Suite for Large Language Models. https://arxiv.org/abs/2404.13161 Meta AI, 2024
-
[22]
S. Bhatt, P. K. Manadhata, and L. Zomlot. 2014. The Operational Role of Security Information and Event Management Systems.IEEE Secur. Privacy 12, 5 (2014), 35–41. doi:10.1109/MSP.2014.103
-
[23]
Tristan Bilot, Baoxiang Jiang, Zefeng Li, Nour El Madhoun, Khaldoun Al Agha, Anis Zouaoui, and Thomas Pasquier. 2025. Sometimes simpler is better: a comprehensive analysis of state-of-the-art provenance-based intrusion detection systems. InProceedings of the 34th USENIX Security Symposium(Seattle, WA, USA)(SEC ’25). USENIX Association, USA, Article 369, 2...
work page 2025
-
[24]
F. Binbeshr, M. Imam, M. Ghaleb, M. Hamdan, M. A. Rahim, and M. Hammoudeh. 2025. The Rise of Cognitive SOCs: A Systematic Literature Review on AI Approaches.IEEE Open Journal of the Computer Society6 (2025), 360–379. doi:10.1109/OJCS.2025.3536800
-
[25]
A. L. Buczak and E. Guven. 2016. A Survey of Data Mining and Machine Learning Methods for Cyber Security Intrusion Detection.IEEE Commun. Surv. Tutor.18, 2 (2016), 1153–1176. doi:10.1109/COMST.2015.2494502
-
[26]
Qizhi Cai, Lingzhi Wang, Yao Zhu, Zhipeng Chen, Xiangmin Shen, and Zhenyuan Li. 2026. Building Next-Generation Datasets for Provenance-Based Intrusion Detection. doi:10.14722/prism.2026.23021 Workshop on Predictive and Robust Intrusion System Modeling (PRISM), 2026
-
[27]
Lalitha Chavali, Tanay Gupta, and Paresh Saxena. 2022. SAC-AP: Soft Actor Critic based Deep Reinforcement Learning for Alert Prioritization. doi:10.1109/CEC55065.2022.9870423 2022 IEEE Congress on Evolutionary Computation (CEC), 2022
-
[28]
Zijun Cheng, Qiujian Lv, Jinyuan Liang, Yan Wang, Degang Sun, Thomas Pasquier, and Xueyuan Han. 2024. Kairos: Practical Intrusion Detection and Investigation using Whole-system Provenance. In2024 IEEE Symposium on Security and Privacy (SP). IEEE Computer Society, Los Alamitos, Manuscript submitted to ACM 30 Samuel Ndichu, Akira Yamada, Tao Ban, Seiichi Oz...
-
[29]
M. B. Chhetri, S. Tariq, R. Singh, F. Jalalvand, C. Paris, and S. Nepal. 2024. Towards Human-AI Teaming to Mitigate Alert Fatigue in Security Operations Centres.ACM Trans. Internet Technol.24, 3 (2024), 1–22. doi:10.1145/3670009
-
[30]
Frédéric Cuppens and Alexandre Miège. 2002. Alert Correlation in a Cooperative Intrusion Detection Framework. doi:10.1109/SECPRI.2002.1004372 IEEE Symposium on Security and Privacy, pp. 202–215, 2002. IEEE
-
[31]
D3 Security. 2025. The AI SOC Market Landscape 2025. https://d3security.com/resources/the-ai-soc-market-landscape-2025/ D3 Security report, 2025
work page 2025
-
[32]
Hervé Debar and Andreas Wespi. 2001. Aggregation and Correlation of Intrusion-Detection Alerts. doi:10.1007/3-540-45474-8_6 Recent Advances in Intrusion Detection (RAID), Lecture Notes in Computer Science, vol. 2212, pp. 85–103, 2001. Springer
-
[33]
Gelei Deng, Yi Liu, Víctor Mayoral-Vilches, Peng Liu, Yuekang Li, Yuan Xu, Tianwei Zhang, Yang Liu, Martin Pinzger, and Stefan Rass. 2024. PentestGPT: Evaluating and Harnessing Large Language Models for Automated Penetration Testing. In33rd USENIX Security Symposium (USENIX Security 24). USENIX Association, Philadelphia, PA, 847–864. https://www.usenix.or...
work page 2024
-
[34]
Min Du, Feifei Li, Guineng Zheng, and Vivek Srikumar. 2017. DeepLog: Anomaly Detection and Diagnosis from System Logs through Deep Learning. InProceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security(Dallas, Texas, USA)(CCS ’17). Association for Computing Machinery, New York, NY, USA, 1285–1298. doi:10.1145/3133956.3134015
-
[35]
Thijs van Ede, Hojjat Aghakhani, Noah Spahn, Riccardo Bortolameotti, Marco Cova, Andrea Continella, Maarten van Steen, Andreas Peter, Christopher Kruegel, and Giovanni Vigna. 2022. DEEPCASE: Semi-Supervised Contextual Analysis of Security Events. In2022 IEEE Symposium on Security and Privacy (SP). IEEE, Piscataway, NJ, USA, 522–539. doi:10.1109/SP46214.20...
-
[36]
B. Edelman, S. Peng, R. Rodriguez, and S. Ho. 2023. Randomized Controlled Trials for Microsoft Copilot for Security. doi:10.2139/ssrn.4648700 SSRN Working Paper 4648700; updated March 29, 2024
-
[37]
Elastic. 2025. How to reduce alert overload in defence SOCs. https://www.elastic.co/blog/reduce-alert-fatigue-with-ai-defence-soc Elastic blog, 2025
work page 2025
-
[38]
Pengcheng Fang, Peng Gao, Changlin Liu, Erman Ayday, Kangkook Jee, Ting Wang, Yanfang (Fanny) Ye, Zhuotao Liu, and Xusheng Xiao
-
[39]
In31st USENIX Security Symposium (USENIX Security 22)
Back-Propagating System Dependency Impact for Attack Investigation. In31st USENIX Security Symposium (USENIX Security 22). USENIX Association, Boston, MA, 2461–2478. https://www.usenix.org/conference/usenixsecurity22/presentation/fang
- [40]
-
[41]
M. A. Ferrag, A. Battah, N. Tihanyi, M. Debbah, T. Lestable, and L. C. Cordeiro. 2025. SecureFalcon: Are We There Yet in Automated Software Vulnerability Detection with LLMs?IEEE Trans. Softw. Eng.51, 4 (2025), 1248–1265. doi:10.1109/TSE.2025.3548168
-
[42]
J. Forsberg and T. Frantti. 2023. Technical performance metrics of a security operations center.Comput. Secur.135 (2023), 103529. doi:10.1016/J. COSE.2023.103529
work page doi:10.1016/j 2023
-
[43]
Scott Freitas, Jovan Kalajdjieski, Amir Gharib, and Robert McCann. 2024. AI-Driven Guided Response for Security Operation Centers with Microsoft Copilot for Security. arXiv preprint arXiv:2407.09017. https://arxiv.org/abs/2407.09017 Introduces GUIDE: 13M+ evidence items, 1.6M alerts, 1M+ triage-annotated incidents (TP/BP/FP) from 6,100+ organizations. CDL...
-
[44]
S. García, M. Grill, J. Stiborek, and A. Zunino. 2014. An empirical comparison of botnet detection methods.Comput. Secur.45 (Sept. 2014), 100–123. doi:10.1016/j.cose.2014.05.011
-
[45]
P. Garcia-Teodoro, J. Diaz-Verdejo, G. Macía-Fernández, and E. Vázquez. 2009. Anomaly-Based Network Intrusion Detection: Techniques, Systems and Challenges.Comput. Secur.28 (2009), 18–28. doi:10.1016/j.cose.2008.08.003
-
[46]
J. Ghadermazi, A. Shah, and S. Jajodia. 2024. A Machine Learning and Optimization Framework for Efficient Alert Management in a Cybersecurity Operations Center.Digital Threats: Research and Practice5, 2 (2024), 19:1–19:23. doi:10.1145/3644393
-
[47]
N. Goldenberg and A. Wool. 2013. Accurate Modeling of Modbus/TCP for Intrusion Detection in SCADA Systems.Int. J. Crit. Infrastruct. Prot.6, 2 (2013), 63–75. doi:10.1016/j.ijcip.2013.05.001
-
[48]
Akul Goyal, Gang Wang, and Adam Bates. 2024. R-CAID: Embedding Root Cause Analysis within Provenance-based Intrusion Detection. In2024 IEEE Symposium on Security and Privacy (SP). IEEE, Piscataway, NJ, USA, 3515–3532. doi:10.1109/SP54263.2024.00253
- [49]
-
[50]
Haixuan Guo, Shuhan Yuan, and Xintao Wu. 2021. LogBERT: Log Anomaly Detection via BERT. In2021 International Joint Conference on Neural Networks (IJCNN). IEEE, Piscataway, NJ, USA, 1–8. doi:10.1109/IJCNN52387.2021.9534113
-
[51]
A. Habibzadeh et al. 2025. Large Language Models for Security Operations Centers: A Comprehensive Survey. https://arxiv.org/abs/2509.10858 arXiv:2509.10858
-
[52]
Bardas, Michael Collins, Jaclyn Lauren Dudek, Daniel Lende, Xinming Ou, and S
Francis Hahn, Mohd Mamoon, Alexandru G. Bardas, Michael Collins, Jaclyn Lauren Dudek, Daniel Lende, Xinming Ou, and S. Raj Rajagopalan
-
[53]
doi:10.14722/wosoc.2026.23015 Workshop on Security Operations Center (WOSOC), 2026
Non-Disruptive Disruption: An Empirical Experience of Introducing LLMs in the SOC. doi:10.14722/wosoc.2026.23015 Workshop on Security Operations Center (WOSOC), 2026
-
[54]
X. Han, T. F. J.-M. Pasquier, A. Bates, J. Mickens, and M. Seltzer. 2020. UNICORN: Runtime Provenance-Based Detector for Advanced Persistent Threats. https://www.ndss-symposium.org/ndss-paper/unicorn-runtime-provenance-based-detector-for-advanced-persistent-threats/ NDSS, 2020.. Manuscript submitted to ACM AI-Driven Alert Screening in SOCs 31
work page 2020
-
[55]
Wajih Ul Hassan, Adam Bates, and Daniel Marino. 2020. Tactical Provenance Analysis for Endpoint Detection and Response Systems. In2020 IEEE Symposium on Security and Privacy (SP). IEEE, Piscataway, NJ, USA, 1172–1189. doi:10.1109/SP40000.2020.00096
-
[56]
W. U. Hassan, M. A. Noureddine, P. Datta, and A. Bates. 2019. NoDoze: Combatting Threat Alert Fatigue with Automated Provenance Triage. https://www.ndss-symposium.org/ndss-paper/nodoze-combatting-threat-alert-fatigue-with-automated-provenance-triage/ NDSS, 2019
work page 2019
-
[57]
W. U. Hassan, M. A. Noureddine, P. Datta, and A. Bates. 2020. OmegaLog: High-Fidelity Attack Investigation via Transparent Multi-Layer Log Analysis. doi:10.14722/ndss.2020.24270 NDSS, 2020
-
[58]
M. Hassanin and N. Moustafa. 2024. A Comprehensive Overview of Large Language Models (LLMs) for Cyber Defences: Opportunities and Directions. https://arxiv.org/abs/2405.14487 arXiv:2405.14487
-
[59]
Seth Hastings and Tyler Moore. 2025. Authentication-Event Processing for Enhanced SOC Investigations. doi:10.14722/wosoc.2025.23016 Workshop on Security Operations Center (WOSOC), 2025
-
[60]
B. Hawash, U. Asma’ Mokhtar, J. J. Jeong, S. B. Maynard, Z. Shukur, S. N. H. Sheikh Abdullah, R. Razali, J. S. Lim, and A. Ahmad. 2024. Cyber Situational Awareness in Security Operation Centres. https://aisel.aisnet.org/pacis2024/track07_secprivacy/track07_secprivacy/8/ Pacific Asia Conf. Information Systems (PACIS), 2024
work page 2024
-
[61]
Pinjia He, Jieming Zhu, Zibin Zheng, and Michael R. Lyu. 2017. Drain: An Online Log Parsing Approach with Fixed Depth Tree. In2017 IEEE International Conference on Web Services (ICWS). IEEE, Piscataway, NJ, USA, 33–40. doi:10.1109/ICWS.2017.13
-
[62]
Kate Highnam, Kai Arulkumaran, Zachary Hanif, and Nicholas R. Jennings. 2021. BETH Dataset: Real Cybersecurity Data for Anomaly Detection Research. doi:10.14469/hpc/9422 Imperial College London dataset record
-
[63]
Milajerdi, Junao Wang, Birhanu Eshete, Rigel Gjomemo, R
Md Nahid Hossain, Sadegh M. Milajerdi, Junao Wang, Birhanu Eshete, Rigel Gjomemo, R. Sekar, Scott Stoller, and V.N. Venkatakrishnan. 2017. SLEUTH: Real-time Attack Scenario Reconstruction from COTS Audit Data. In26th USENIX Security Symposium (USENIX Security 17). USENIX Association, Vancouver, BC, 487–504. https://www.usenix.org/conference/usenixsecurity...
work page 2017
-
[64]
Martin Husák, Martin Žádník, Václav Bartoš, and Pavol Sokol. 2020. Dataset of intrusion detection alerts from a sharing platform.Data in Brief33 (2020), 106530. doi:10.1016/j.dib.2020.106530
-
[65]
IBM Security. 2024. Cost of a Data Breach Report 2024. https://newsroom.ibm.com/2024-07-30-ibm-report-escalating-data-breach-disruption- pushes-costs-to-new-highs IBM, 2024
work page 2024
-
[66]
M. A. Inam et al. 2023. SoK: History is a Vast Early Warning System: Auditing the Provenance of System Intrusions. doi:10.1109/SP46215.2023. 10179405 IEEE S&P, 2023
-
[67]
R. Jaffal et al. 2025. Large Language Models in Cybersecurity: A Survey of Applications, Vulnerabilities, and Defense Techniques.AI6, 9 (2025),
work page 2025
-
[68]
doi:10.3390/ai6090216
-
[69]
F. Jalalvand, M. B. Chhetri, S. Nepal, and C. Paris. 2024. Alert Prioritisation in Security Operations Centres: A Systematic Survey on Criteria and Methods.ACM Comput. Surv.57, 2 (2024), 1–36. doi:10.1145/3695462
-
[70]
F. Jalalvand, E. Torkzadehmahani, G. Kesidis, and S. E. Kahou. 2025. Adaptive alert prioritisation in security operations centres via learning to defer with human feedback. https://arxiv.org/abs/2506.18462 arXiv:2506.18462
-
[71]
Woohyuk Jang, Hyunmin Kim, Hyungbin Seo, Minsong Kim, and Myungkeun Yoon. 2023. SELID: Selective Event Labeling for Intrusion Detection Datasets.Sensors23, 13, Article 6105 (2023), 11 pages. doi:10.3390/s23136105
-
[72]
Z. Jia, Q. Lin, Y. Shan, J. Liang, Z. Li, T. Wang, and Y. Liu. 2024. MAGIC: Detecting Advanced Persistent Threats via Masked Graph Representation Learning. https://www.usenix.org/conference/usenixsecurity24/presentation/jia-zian USENIX Security, 2024
work page 2024
-
[73]
Baoxiang Jiang, Tristan Bilot, Nour El Madhoun, Khaldoun Al Agha, Anis Zouaoui, Shahrear Iqbal, Xueyuan Han, and Thomas Pasquier. 2025. ORTHRUS: achieving high quality of attribution in provenance-based intrusion detection systems. InProceedings of the 34th USENIX Security Symposium(Seattle, WA, USA)(SEC ’25). USENIX Association, USA, Article 368, 20 page...
work page 2025
-
[74]
K. Julisch. 2003. Clustering Intrusion Detection Alarms to Support Root Cause Analysis.ACM Trans. Inf. Syst. Secur.6, 4 (2003), 443–471. doi:10.1145/950191.950192
-
[75]
Takeshi Kaneko, Hiroyuki Okada, Rashi Sharma, Tatsumi Oba, and Naoto Yanai. 2026. PAIEL: Protocol-Aware and Context-Integrated Protocol Explanation Using LLMs for SOCs. doi:10.14722/wosoc.2026.23013 Workshop on Security Operations Center Operations and Construction (WOSOC), 2026
-
[76]
Karunasingha, Mohan Baruwal Chhetri, Surya Nepal, Cécile Paris, and Salil S
Navodika M. Karunasingha, Mohan Baruwal Chhetri, Surya Nepal, Cécile Paris, and Salil S. Kanhere. 2025. SoK: AI Support for Analyst Situation Awareness in Security Operation Centres. In2025 European Symposium on Usable Security (EuroUSEC). ACM, New York, NY, USA, 151–163. doi:10.1145/3697926.3697939
-
[77]
Leon Kersten, Kim Beelen, Emmanuele Zambon, Chris Snijders, and Luca Allodi. 2025. A Field Study to Uncover and a Tool to Support the Alert Investigation Process of Tier-1 Analysts. doi:10.14722/usec.2025.23034 Workshop on Usable Security (USEC), 2025
-
[78]
Leon Kersten, Tom Mulders, Emmanuele Zambon, Chris Snijders, and Luca Allodi. 2023. ’Give Me Structure’: Synthesis and Evaluation of a (Network) Threat Analysis Process Supporting Tier 1 Investigations in a Security Operation Center. InNineteenth Symposium on Usable Privacy and Security (SOUPS 2023). USENIX Association, Anaheim, CA, 97–111. https://www.us...
work page 2023
-
[79]
M. Khayat, E. Barka, M. A. Serhani, F. M. Sallabi, K. Shuaib, and H. M. Khater. 2025. Empowering Security Operation Center With Artificial Intelligence and Machine Learning - A Systematic Literature Review.IEEE Access13 (2025), 19162–19197. doi:10.1109/ACCESS.2025.3532951 Manuscript submitted to ACM 32 Samuel Ndichu, Akira Yamada, Tao Ban, Seiichi Ozawa, ...
-
[80]
A. Khraisat, I. Gondal, P. Vamplew, and J. Kamruzzaman. 2019. Survey of Intrusion Detection Systems: Techniques, Datasets and Challenges. Cybersecurity2, 1 (2019), 1–22. doi:10.1186/s42400-019-0038-7
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.