Abstract
We present a new BERT model for the cybersecurity domain, CTI-BERT, which can improve the accuracy of cyber threat intelligence (CTI) extraction, enabling organizations to better defend against potential cyber threats. We provide detailed information about the domain corpus collection, the training methodology and its effectiveness for a variety of NLP tasks for the cybersecurity domain. The experiments show that CTI-BERT significantly outperforms several general-domain and security-domain models for these cybersecurity applications, indicating that the training data and methodology have a significant impact on the model performance.
Cite
CITATION STYLE
Park, Y., & You, W. (2023). A Pretrained Language Model for Cyber Threat Intelligence. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Industry Track (pp. 113–122). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-industry.12
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.