A Pretrained Language Model for Cyber Threat Intelligence

15Citations
Citations of this article
32Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a new BERT model for the cybersecurity domain, CTI-BERT, which can improve the accuracy of cyber threat intelligence (CTI) extraction, enabling organizations to better defend against potential cyber threats. We provide detailed information about the domain corpus collection, the training methodology and its effectiveness for a variety of NLP tasks for the cybersecurity domain. The experiments show that CTI-BERT significantly outperforms several general-domain and security-domain models for these cybersecurity applications, indicating that the training data and methodology have a significant impact on the model performance.

Cite

CITATION STYLE

APA

Park, Y., & You, W. (2023). A Pretrained Language Model for Cyber Threat Intelligence. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Industry Track (pp. 113–122). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-industry.12

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free