TY - GEN
T1 - Improvement of DGA Long Tail Problem Based on Transfer Learning
AU - Fan, Baoyu
AU - Liu, Yue
AU - Cuthbert, Laurie
N1 - Publisher Copyright:
© 2023, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2023
Y1 - 2023
N2 - As the number of classes increases in traditional multiple classification and recognition tasks, there is often the problem of a long tail: the sample data is mainly distributed in a few classes. In the detection of domain names generating malware (DGA - domain generation algorithm), due to the variability of malware, the number of classes of DGA is also increasing and shows a long tail nature. However, in previous DGA detection research focused on the classes of a large amount of data so they do not address the long tail characteristics. We propose an effective knowledge transfer DGA detection model that transfers the knowledge learned in the previous stage of training to the next stage, and optimizes the impact of the long tail problem on the detection model. In order to inherit the continuity of the model, we propose a data balance review method, which can alleviate the catastrophic forgetting problem of transfer learning and detect new classes without retraining the whole model. Finally, the macro average F1 score of our model is 76.6%, 8.74% higher than ATT_BiLSTM and 6.34% higher than ATT_CNN_BiLSTM. So our model optimizes the long tail problem and better predicts all classes.
AB - As the number of classes increases in traditional multiple classification and recognition tasks, there is often the problem of a long tail: the sample data is mainly distributed in a few classes. In the detection of domain names generating malware (DGA - domain generation algorithm), due to the variability of malware, the number of classes of DGA is also increasing and shows a long tail nature. However, in previous DGA detection research focused on the classes of a large amount of data so they do not address the long tail characteristics. We propose an effective knowledge transfer DGA detection model that transfers the knowledge learned in the previous stage of training to the next stage, and optimizes the impact of the long tail problem on the detection model. In order to inherit the continuity of the model, we propose a data balance review method, which can alleviate the catastrophic forgetting problem of transfer learning and detect new classes without retraining the whole model. Finally, the macro average F1 score of our model is 76.6%, 8.74% higher than ATT_BiLSTM and 6.34% higher than ATT_CNN_BiLSTM. So our model optimizes the long tail problem and better predicts all classes.
KW - DGA
KW - Data balanced review
KW - Deep learning
KW - Long tail problem
KW - Transfer learning
UR - http://www.scopus.com/inward/record.url?scp=85144220245&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-12127-2_10
DO - 10.1007/978-3-031-12127-2_10
M3 - Conference contribution
AN - SCOPUS:85144220245
SN - 9783031121265
T3 - Studies in Computational Intelligence
SP - 139
EP - 152
BT - Computer and Information Science
A2 - Lee, Roger
PB - Springer Science and Business Media Deutschland GmbH
T2 - 22nd IEEE/ACIS International Conference on Computer and Information Science, ICIS 2022
Y2 - 26 June 2022 through 28 June 2022
ER -