TY - JOUR
T1 - Ai-enabled automated common vulnerability scoring from common vulnerabilities and exposures descriptions
AU - Zhang, Zijing
AU - Kumar, Vimal
AU - Pfahringer, Bernhard
AU - Bifet, Albert
N1 - Publisher Copyright:
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2024.
PY - 2025/2/1
Y1 - 2025/2/1
N2 - With the sheer amount of vulnerabilities, manually evaluating the impact of them is challenging. This paper proposes employing artificial intelligence models as substitutes for humans or as aides to human experts in estimating vulnerabilities. We compare the precision, recall, and F1 score amongst the Universal Sentence Encoder, Generative Pre-trained Transformer, and Support Vector Machine, trained on 118,000 vulnerabilities and tested on 51,000 vulnerabilities, with human experts on mean estimation error and variance for each type of vulnerability from the state of the art work in estimating vulnerability severity scores. The Universal Sentence Encoder demonstrates superior performance with results (72/77 % accuracy on severity level prediction) that significantly outperform human experts in assessment tasks for various types of vulnerabilities with high efficiency for memory consumption and low running time. Additionally, we examine the efficacy of our models in predicting the components of vulnerability evaluation metrics and the severity level of vulnerabilities. The findings highlight the potential of artificial intelligence agents to assist cybersecurity experts in this task which in the current state of the art is entirely manual.
AB - With the sheer amount of vulnerabilities, manually evaluating the impact of them is challenging. This paper proposes employing artificial intelligence models as substitutes for humans or as aides to human experts in estimating vulnerabilities. We compare the precision, recall, and F1 score amongst the Universal Sentence Encoder, Generative Pre-trained Transformer, and Support Vector Machine, trained on 118,000 vulnerabilities and tested on 51,000 vulnerabilities, with human experts on mean estimation error and variance for each type of vulnerability from the state of the art work in estimating vulnerability severity scores. The Universal Sentence Encoder demonstrates superior performance with results (72/77 % accuracy on severity level prediction) that significantly outperform human experts in assessment tasks for various types of vulnerabilities with high efficiency for memory consumption and low running time. Additionally, we examine the efficacy of our models in predicting the components of vulnerability evaluation metrics and the severity level of vulnerabilities. The findings highlight the potential of artificial intelligence agents to assist cybersecurity experts in this task which in the current state of the art is entirely manual.
KW - CVSS
KW - Cybersecurity
KW - Large language model
KW - Machine learning
KW - Vulnerability assessment
U2 - 10.1007/s10207-024-00922-z
DO - 10.1007/s10207-024-00922-z
M3 - Article
AN - SCOPUS:85208939682
SN - 1615-5262
VL - 24
JO - International Journal of Information Security
JF - International Journal of Information Security
IS - 1
M1 - 16
ER -