This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
1. Introduction
With the commercialization and popularization of 5G, the IoT is coming closer to reality [1]. Meanwhile, with the scale expansion of connected terminals, data storage, and utilization, security issues are becoming more and more complex. As the methods of network crime are also constantly updated, the probability of network attack is greatly increased, which is not conducive to protecting personal privacy [2]. Therefore, in the 5G era, designing a model with good robustness is an important issue.
At present, malware attacks remain as one of the most urgent security issues users facing. In the last decade, deep neural network-based malware detection has fulfilled remarkable achievements [3]. A growing community of researchers is attempting to apply deep learning to malware detection and classification tasks [4–9]. Saxe and Berlin [10] extracted the binary features of PE files, which are portable executable ones under Windows operation systems and utilized a four-layer feed-forward neural network to detect malware. Kalash et al. [11] transformed malware binaries into greyscale images and classified malware by the use of a CNN. The DL-based malicious detection and classification models are now widely used.
However, much recent work indicates that adversarial attacks can cause serious damage to deep neural networks [12–15]. Adversarial examples in computer vision applications have been widely proven. In malware detection, adversarial modifications often need minor changes to malicious binaries. Different from language and images, codes are discrete sequences, which means that the generation technique of adversarial samples in images cannot be transferred to the malware detection field. Moreover, a minor change in the sequence may result in its functionality be changed completely. For example, in a binary file, changing a single byte may lead to a completely ineffective bytecode or distinct functionality. Therefore, it remains a great challenge to implement practical black-box attacks on malware binary-based deep learning models. Recently, a series of research works have been done in adversarial attacks. Hu and Tan [16] proposed adversarial modification of feature vectors. However, the malware binaries were not modified actually. This method cannot guarantee that the modified feature vector can be converted to actual binaries. Moreover, it destroyed the format or affected the functionality of the malware. Anderson et al. [17] directly modified malware binaries to perform adversarial attacks. In theory, these methods cannot disrupt the original functionality of the malware. However, in practice, we have found that these seemingly reliable methods also damage the malicious functionality. During an adversarial attack, if the malicious functionality of the original sample is destroyed, the adversarial attack is invalid. Therefore, it is necessary to perform malicious functionality detection on generated adversarial samples. However, most of the previous work did not address this issue. Besides, some of the previous work was done in a white-box adversarial model [18–20]. The white-box adversarial model requires knowing malware classifier architecture, making their methods impractical in real network environments. Therefore, while retaining the primary malicious functionality of the binaries, it is a great challenge to implement a practical black-box attack on the malware detection model based on deep learning.
This paper proposes an evolutionary algorithm-based adversarial sample generation method. In our approach, the generated samples by rewriting the file structure and adding adversarial information evade successfully the malware detection model, while preserving the original behavior of PE files. We test 1000 PE samples in four popular antivirus software on VirusTotal, showing that the method proposed can generate adversarial samples in binary format. The contributions of this paper are highlighted as follows.
(1) This paper proposes a new method of generating adversarial samples by the use of the evolutionary algorithm, which can automatically generate valid adversarial samples
(2) This paper uses a well-designed feature library as rewriting material in the evolutionary process, which helps generate modified samples with fewer attempts
(3) This paper applies the adversarial samples generated to attack DL-based malware detection engines on VirusTotal and obtains better experimental results than other attack methods
The rest of this study is organized as follows. Section 2 is a concise introduction of malware detection and adversarial attack methods. Section 3 proposes our attack framework in detail. Section 4 describes the experimental settings and main results and gives a deep analysis. Section 5 concludes this paper as well as the research directions.
2. Related Work
2.1. Machine Learning-Based Malware Detection Methods
In malware detection, machine learning (ML) is a popular approach. Moreover, in recent years, many ML-based malware detection methods have been put forward [21–24]. These methods are mainly categorized in static analysis [25, 26] and dynamic behavior analysis [27, 28]. Static analysis learns the statistical features of malware (e.g., API calls, OpCode), whereas dynamic behavior analysis detects abnormal (possibly malicious) behavior by observing deviations from the baseline of the system. Recently, malware detection efforts prefer to use raw software binaries as the input of DL models [29–31].
NVIDIA’s research group [32] proposed the MalConv network, which took the raw byte sequences of PE files as input directly, achieved 98.88% detection accuracy. Compared with the detection model that extracts only some features of PE files as input, MalConv links other discrete features. Therefore, it can detect samples with arbitrary size and avoid missing important features. This paper assesses the effectiveness and performance of our framework using the MalConv detection system.
2.2. Adversarial Attack against Malware Detection Model
DL-based malware detection approaches are susceptible to adversarial attacks [33–37]. Adversarial modifications by manipulating only a small fraction of raw binary data may lead to misclassification. Moreover, the raw binary contents of data are not changed in a nutshell; otherwise, its originally momentous functionality might lose.
Prior work has proposed various ways of adversarial attack against ML-based malware detection models. Through appending bytes at the end of a binary file while preserving its intrusive functionality, Kolosnjaji et al. [20] designed a gradient-based attack model. However, it is based on white-box attacks and cannot be applied to real scenarios. Kreuk et al. [38] proposed a modification method that injected a minor byte sequence into the originally binary file. It is also based on white-box attacks and is not efficient in real scenarios. Anderson et al. [17] designed an effective model which is based on a deep reinforcement learning method to attack static PE antimalware engines. In their work, the reward function and the environment of reinforcement learning were artificially defined. Later, Fang et al. [39] improved Anderson et al.’s work by autonomously generating the reward function according to the expert strategy. Numerous experiments showed that Fang et al.’s method [39] is more nimble and efficacious than Anderson et al.’s method [17]. Yuan et al. [40] proposed an adversarial sample generation model named GAPGAN. GAPGAN initially maps the discrete malware binaries into a contiguous space; the output is input to the generator of GAPGAN to generate adversarial payloads. Finally, the generated payloads are appended to the originally binary file to create an adversarial one. Because the valid part of the binary file was not changed, the original functionality of the binary file is preserved. GAPGAN can perform an efficient black-box attack. However, the modification action in GAPGAN involves only a simple action. The GAPGAN cannot perform complicated modifications similar to real malware writers. Song et al. [41] presented a framework for creating adversarial malware and evaluated the evasion capabilities in realistic scenarios. The authors firstly revealed the root causes that adversarial samples evade the malware detection method.
This study puts forward a novel binary manipulation-based attack framework, which generates adversarial samples with an evolutionary learning algorithm. Our method can adaptively simplify the actions of modifying binary samples and use an evolutionary algorithm to make adversarial samples more targeted. The generated adversarial samples by statically rewriting the PE file keep their dynamic behavior consistent and can evade the DL-based malware detection models. Experimental results verify the effectiveness of our method, which can efficiently and quickly generate adversarial samples without human intervention.
3. Adversarial Sample Generation Based on Evolutionary Algorithm
3.1. Problem Description
Our ultimate objective is to generate an antagonistic sample, which manipulates the classifier to classify malicious software as benign by mistake while still retaining the malicious function of the original sample.
Let us consider a classifier
In brief, for the malware sample
3.2. Rewriting Actions
PE file is a generic term for executable files in Windows operating system. A PE file consists of a header, section table, and section data. The MS-DOS header consists of three parts: a DOS header, the true PE header, and an optional header, and it includes some basic messages about the executable file. Section table describes the characteristics of each file section. The section table consists of a series of IMAGE_SECTION_HEADER structures arranged in a sequence. The structures and sections are arranged in a fixed order. Section data consists of 4 main parts: .text, .data, .rdata, and .idata, and the data part includes the practical contents relating to every section. The PE file format is shown in Figure 1.
[figure omitted; refer to PDF]
For a black-box model, we have no idea of the exact features of the classifier involved. However, by observing the chosen features in some open-sourced classifiers, we can make a wild guess at some of the common features in malware detection models. An adversarial sample is generated by modifying one or several features. The chosen actions of modifying the features should be easy to execute. Moreover, after the features are modified, the executability and functionality of malware should not be corrupted. In this paper, all actions applied to the PE file are shown as follows.
(1) Appending some bytes to the PE file
(2) Inserting an unused function to the import address table
(3) Appending some bytes to the untapped space in a section
(4) Adding a new section
(5) Changing a section name
(6) Packing the file
(7) Unpacking the file
The malicious binary file is modified through the following steps. Firstly, the original PE file is read, then the content in the specified location is added or deleted, and finally, the relative virtual address of the PE file is modified.
3.3. The Proposed Framework
The workflow of the framework includes three parts: the generation of the feature library, the generation of the modified samples, and the generation of the adversarial samples. Firstly, the feature library is generated using MalGAN. In the processing of generating the adversarial samples, the modified features are randomly selected from the feature library according to the rewriting actions. Then, modified samples by evolutionary algorithms are generated. Finally, the generated modified samples are tested whether the malicious functionality of the original samples remains or not. If a modified sample has the same malicious functionality as the original one, we save it as an adversarial sample. Figure 2 gives an overview of our framework. The details of the three parts are given in the following.
[figure omitted; refer to PDF]
The detailed evolutionary process is as follows.
(1) Step 1. Population initialization.
(2) Step 2. Binary modification. Firstly,
(3) Step 3. Fitness calculation. The modified samples are fed into the MalConv network, and the output of the MalConv is used as the fitness of the individual. A smaller output value of MalConv indicates a higher fitness. The higher the fitness, the higher the probability that a gene sequence will be selected for retention
(4) Step 4. Selecting the best offspring according to the fitness as the parent of the next generation
(5) Step 5. Performing genetic manipulation on the selected parents. New offspring through crossover and mutation are reproduced
(6) Step 6. The assessment of end condition. When the action sequence has reached a minimum value or the maximum number of iterations is reached, the evolution is ended. If the end condition is satisfied, the modified samples are output. Otherwise, skip to Step 2
The detailed process is described in Algorithm 1.
Algorithm 1: Malware sample evolution.
Input: malware samples
Output: modified samples
BEGIN
for
Initialize the population;
while current generation
Map binary sequences to action sequences;
Modify malware sample based on the action sequences;
Calculate fitness;
Select the best offspring;
Perform crossover;
Perform mutation;
Increase current generation;
end while
Append the optimal result
end for
Return
END
3.3.3. Generation of the Adversarial Samples
It should also be noted that the functionality of a malware sample may be corrupted during the modification process. In other words, its attacking characteristics may be damaged. We consider an adversarial sample without malicious functionality to be invalid. To detect whether the malicious functionality of the modified sample is retained, we use the sandbox to collect behaviors of the modified samples and original samples. If the behavior of the modified sample is the same as that of the original one, we think that it retains the malicious function of the original sample, and it is saved as an adversarial sample.
Suppose a behavior of the original sample
Because the modification operation is a direct manipulation on the original sample, it can inevitably alter the behaviors of the original samples. Therefore, we assume that if
4. Experimental Results and Analysis
This section firstly gives the setting of our experiments, including the datasets, evaluation metrics, and the target malware detection model. Then, we analyze the experiment results.
4.1. Experimental Settings and Evaluation Metrics
In the experiment, we construct a dataset with 1000 malware samples from VirusTotal. Moreover, we also produce some adversarial examples for PE binaries to evaluate the effectiveness of our proposed method.
To assess the effectiveness of adversarial samples, we measure some evaluation metrics in Table 1.
Table 1
Evaluation metrics.
Evaluation metrics | Formula |
The success rate of modified sample operation | |
The malicious function retention rate of modified samples | |
Evasion rate |
4.2. Experimental Results of Adversarial Attack
This section demonstrates some performances of our approach under antiattack scenes and compares the results with some methods available.
In our experiment, the attacked model is the MalConv proposed by Raff et al. [32]. We train attacked MalConv network using a dataset with 6230 malicious samples from VirusTotal and 5660 benign samples from web crawling and achieve 98.4% detection accuracy.
Furthermore, to explore the validity of the presented binary-based attack method against a deep learning-based detection system, we compare our approach with other byte-level attack methods, including the DQEAF method [39], which is based on reinforcement learning and Aut. method [41], which is based on code randomization and binary manipulation. The results are shown in Tables 2 and 3, respectively.
Table 2
The performance comparison of adversarial samples generated by different methods.
Attack methods | Evaluation metrics | ||
Average time | |||
Our method | 100% | 97.5% | 24.5 s |
DQEAF | 76% | 94.3% | 60 s |
Aut. | 68% | 95.6% | 42.5 s |
Table 3
The evasion rate of different attack methods against different detection engines.
Detector | Adversarial attack methods | ||
DQEAF | Aut. | Our method | |
ClamAV | 19.2% | 17.3% | 18.5% |
Cylance | 39.5% | 42.5% | 47.8% |
Endgame | 21.3% | 22.6% | 23.5% |
Trapmine | 20.6% | 18.8% | 19.2% |
Table 2 shows the performance comparison of adversarial samples generated by different methods. From the three evaluation metrics of generated samples, our approach outperforms other similar methods. The reason is that the selected action of modifying malicious samples is built on ensuring the execution of PE files. The specific modifying actions, such as inserting, changing, and adding actions, are taken from the generated feature library. The feature library has also been carefully designed to ensure the authenticity of all actions. Therefore, the modification does not involve invalid actions which damage the primary structure of malware or lose its original functionality. Meanwhile, our action set does not cover the irreversible actions, such as removing signatures, which makes our action set is more effective. Our approach can do a heuristic random search which simplifies the modification actions and does not require performing action sequence minimization and marking action weights and success content as the other two methods, which significantly raises the efficiency of our method.
To further test the effectiveness, we evaluate it using four representative malware detection engines on VirusTotal, including ClamAV, Cylance, Endgame, and Trapmine. Table 3 demonstrates different evasion rates of adversarial samples produced by some methods against different detection engines. From Table 3, we can see that the evasion rate of our method has a better performance compared with the other two in most cases. Meanwhile, to test the efficiency, we record the evasion rate with the generation increase, shown in Figure 4. From Figure 4, we can find that our method reaches a relatively stable evasion rate after 15 generations, which shows that our method is very efficient.
[figure omitted; refer to PDF]4.3. Defense against Adversarial Sample Attack
To defend against adversarial attacks, more and more defense countermeasures have been proposed. Among them, adversarial training is one of the most popular ways [42], in which adversarial samples are added to the training set; thus, DL models can adjust the decision strategies. Compared with other adversarial defense methods, adversarial training does not require modifying the detection model and is easy to implement.
In this paper, we use adversarial samples generated by evolutionary algorithms to test on MalConv network. Table 4 shows the performance evaluation on MalConv with and without adversarial training. The experimental results show that the detection accuracy increased from 80.2% to 90.5% after the adversarial training. It also illustrates that adversarial training can effectively improve the model robustness to adversarial attacks.
Table 4
Performance evaluation on MalConv with and without adversarial training.
Defense | Test dataset | Accuracy |
No defense | × | 98.4% |
√ | 80.2% | |
Adversarial training | √ | 90.5% |
× indicates no adversarial samples and √ denotes having adversarial ones in the test dataset.
5. Conclusion
To make DL-based IoT malware detection models more robust and effective, we propose a framework for generating adversarial samples and their defense. Our framework firstly adopts an evolutionary algorithm to generate modified samples, and then, the modified samples that retain the originally malicious functionality are saved as adversarial samples. This method does not need to obtain any information of the special detection models containing extracted features, internal parameters, etc. Moreover, our approach is entirely automated without human intervention. The experimental results demonstrate that our method can ensure the diversity of generated samples and greatly enhance the efficiency of adversarial sample generation. This paper also demonstrates that adversarial training is one of the effective methods to combat adversarial sample attacks.
The action space has a great influence on the diversity and versatility of evolutionary optimization algorithms. Defining more effective modification actions to expand the search space of evolutionary algorithms is our urgent task. Our future work also includes accelerating the convergence speed and improving the stability of the evolutionary algorithm. Moreover, we will also explore more methods of generating adversarial samples to defend against adversarial attacks on IoT.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (Grant No. 61572170), the Natural Science Foundation of Hebei Province of China (Grant Nos. F2019205163 and F2021205004), the Science and Technology Foundation Project of Hebei Normal University (Grant No. L2021K06), the Science Foundation of Returned Overseas of Hebei Province of China (Grant No. C2020342), the Science Foundation of Department of Human Resources and Social Security of Hebei Province (Grant Nos. 201901028 and ZD2021062), and the Natural Science Foundation of Hebei Normal University (Grant No. L072018Z10).
[1] J. Qiu, Z. Tian, C. du, Q. Zuo, S. Su, B. Fang, "A survey on access control in the age of Internet of Things," IEEE Internet of Things Journal, vol. 7 no. 6, pp. 4682-4696, DOI: 10.1109/JIOT.2020.2969326, 2020.
[2] Y. Xu, C. Zhang, Q. Zeng, G. Wang, J. Ren, Y. Zhang, "Blockchain-enabled accountability mechanism against information leakage in vertical industry services," IEEE Transactions on Network Science and Engineering, vol. 8 no. 2, pp. 1202-1213, DOI: 10.1109/TNSE.2020.2976697, 2021.
[3] X. Yan, Y. Xu, X. Xing, B. Cui, Z. Guo, T. Guo, "Trustworthy network anomaly detection based on an adaptive learning rate and momentum in IIoT," IEEE Transactions on Industrial Informatics, vol. 16 no. 9, pp. 6182-6192, DOI: 10.1109/TII.2020.2975227, 2020.
[4] A. Sharma, P. Malacaria, M. H. R. Khouzani, "Malware detection using 1-dimensional convolutional neural networks," 2019 IEEE European Symposium on Security and Privacy Workshops (EuroS and PW), pp. 247-256, DOI: 10.1109/EuroSPW.2019.00034, .
[5] X. Pei, L. Yu, S. Tian, "AMalNet: a deep learning framework based on graph convolutional networks for malware detection," Computers & Security, vol. 93,DOI: 10.1016/j.cose.2020.101792, 2020.
[6] J. Hemalatha, S. A. Roseline, S. Geetha, S. Kadry, R. Damaševičius, "An efficient densenet-based deep learning model for malware detection," Entropy, vol. 23 no. 3, pp. 344-367, DOI: 10.3390/e23030344, 2021.
[7] S. Yoo, S. Kim, S. Kim, B. B. Kang, "AI-HydRa: advanced hybrid approach using random forest and deep learning for malware classification," Information Sciences, vol. 546 no. 9, pp. 420-435, DOI: 10.1016/j.ins.2020.08.082, 2021.
[8] Y. Xu, J. Ren, Y. Zhang, C. Zhang, B. Shen, Y. Zhang, "Blockchain empowered arbitrable data auditing scheme for network storage as a service," IEEE Transactions on Services Computing, vol. 13 no. 2, pp. 289-300, DOI: 10.1109/TSC.2019.2953033, 2019.
[9] Y. Xu, Q. Zeng, G. Wang, C. Zhang, J. Ren, Y. Zhang, "An efficient privacy-enhanced attribute-based access control mechanism," Concurrency & Computation Practice & Experience, vol. 32 no. 5,DOI: 10.1002/cpe.5556, 2020.
[10] J. Saxe, K. Berlin, "Deep neural network based malware detection using two dimensional binary program features," 2015 10th International Conference on Malicious and Unwanted Software (MALWARE), pp. 11-20, DOI: 10.1109/MALWARE.2015.7413680, .
[11] M. Kalash, M. Rochan, N. Mohammed, N. D. B. Bruce, Y. Wang, F. Iqbal, "Malware classification with deep convolutional neural networks," 2018 9th IFIP International Conference on New Technologies, Mobility and Security (NTMS),DOI: 10.1109/NTMS.2018.8328749, .
[12] X. Liu, J. Zhang, Y. Lin, H. Li, "ATMPA: attacking machine learning-based malware visualization detection methods via adversarial examples," IEEE/ACM 27th International Symposium on Quality of Service (IWQoS),DOI: 10.1145/3326285.3329073, .
[13] H. Rathore, S. K. Sahay, P. Nikam, M. Sewak, "Robust Android malware detection system against adversarial attacks using q-learning," Information Systems Frontiers, vol. 22 no. 8,DOI: 10.1007/s10796-020-10083-8, 2020.
[14] C. Zhang, Y. Xu, Y. Hu, J. Wu, J. Ren, Y. Zhang, "A blockchain-based multi-cloud storage data auditing scheme to locate faults," IEEE Transactions on Cloud Computing,DOI: 10.1109/TCC.2021.3057771, 2021.
[15] D. Li, Q. Li, Y. Ye, S. Xu, "A framework for enhancing deep neural networks against adversarial malware," IEEE Transactions on Network Science and Engineering, vol. 8 no. 1, pp. 736-750, DOI: 10.1109/TNSE.2021.3051354, 2021.
[16] W. Hu, Y. Tan, "Generating adversarial malware examples for black-box attacks based on GAN," 2017. http://arxiv.org/abs/1702.05983
[17] H. S. Anderson, A. Kharkar, B. Filar, D. Evans, P. Roth, "Learning to evade static PE machine learning malware models via reinforcement learning," 2018. http://arxiv.org/abs/1801.08917
[18] K. Grosse, N. Papernot, P. Manoharan, M. Backes, P. McDaniel, "Adversarial perturbations against deep neural networks for malware classification," 2016. http://arxiv.org/abs/1606.04435
[19] O. Suciu, S. E. Coull, J. Johns, "Exploring adversarial examples in malware detection," 2019 IEEE Security and Privacy Workshops (SPW),DOI: 10.1109/SPW.2019.00015, .
[20] B. Kolosnjaji, A. Demontis, B. Biggio, D. Maiorca, G. Giacinto, C. Eckert, F. Roli, "Adversarial malware binaries: evading deep learning for malware detection in executables," 2018 26th European Signal Processing Conference (EUSIPCO), pp. 533-537, DOI: 10.23919/eusipco.2018.8553214, .
[21] M. Shafiq, Z. Tian, A. K. Bashir, X. du, M. Guizani, "CorrAUC: a malicious bot-IoT traffic detection method in IoT network using machine-learning techniques," IEEE Internet of Things Journal, vol. 8 no. 5, pp. 3242-3254, DOI: 10.1109/JIOT.2020.3002255, 2021.
[22] S. Su, Z. Tian, S. Liang, S. Li, S. du, N. Guizani, "A reputation management scheme for efficient malicious vehicle identification over 5G networks," IEEE Wireless Communications, vol. 27 no. 3, pp. 46-52, DOI: 10.1109/MWC.001.1900456, 2020.
[23] C. Luo, Z. Tan, G. Min, J. Gan, W. Shi, Z. Tian, "A novel web attack detection system for Internet of Things via ensemble classification," IEEE Transactions on Industrial Informatics, vol. 17 no. 8, pp. 5810-5818, DOI: 10.1109/TII.2020.3038761, 2021.
[24] M. Shafiq, Z. Tian, A. Bashir, A. Jolfaei, X. Yu, "Data mining and machine learning methods for sustainable smart cities traffic classification: a survey," Sustainable Cities and Society, vol. 60,DOI: 10.1016/j.scs.2020.102177, 2020.
[25] J. Yan, Y. Qi, Q. Rao, "Detecting malware with an ensemble method based on deep neural network," Security and Communication Networks, vol. 2018,DOI: 10.1155/2018/7247095, 2018.
[26] X. Liu, Y. Lin, H. Li, J. Zhang, "A novel method for malware detection on ML-based visualization technique," Computers & Security, vol. 89,DOI: 10.1016/j.cose.2019.101682, 2020.
[27] W. Huang, J. W. Stokes, "MtNet: a multi-task neural network for dynamic malware classification," Detection of Intrusions and Malware, and Vulnerability Assessment, pp. 399-418, DOI: 10.1007/978-3-319-40667-1_20, 2016.
[28] B. Kolosnjaji, A. Zarras, G. Webster, C. Eckert, "Deep learning for classification of malware system call sequences," Australasian Joint Conference on Artificial Intelligence, pp. 137-149, DOI: 10.1007/978-3-319-50127-7_11, 2016.
[29] R. Taheri, M. Ghahramani, R. Javidan, M. Shojafar, Z. Pooranian, M. Conti, "Similarity-based Android malware detection using Hamming distance of static binary features," Future Generation Computer Systems, vol. 105 no. 4, pp. 230-247, DOI: 10.1016/j.future.2019.11.034, 2020.
[30] H. Guo, S. Huang, C. Huang, M. Zhang, Z. Pan, F. Shi, H. Huang, D. Hu, X. Wang, "A lightweight cross-version binary code similarity detection based on similarity and correlation coefficient features," IEEE Access, vol. 8, pp. 120501-120512, DOI: 10.1109/ACCESS.2020.3004813, 2020.
[31] N. A. Azeez, O. E. Odufuwa, S. Misra, J. Oluranti, R. Damaševičius, "Windows PE malware detection using ensemble learning," Informatics, vol. 8 no. 1, pp. 10-20, DOI: 10.3390/informatics8010010, 2021.
[32] E. Raff, J. Barker, J. Sylvester, R. Brandon, B. Catanzaro, C. K. Nicholas, "Malware detection by eating a whole EXE," 2017. http://arxiv.org/abs/1710.09435
[33] J. Lin, L. Xu, Y. Liu, X. Zhang, "Black-box adversarial sample generation based on differential evolution," Journal of Systems and Software, vol. 170 no. 8, Article 110767, 2020.
[34] R. Taheri, R. Javidan, M. Shojafar, Z. Pooranian, A. Miri, M. Conti, "On defending against label flipping attacks on malware detection systems," Neural Computing and Applications, vol. 32 no. 18, pp. 14781-14800, DOI: 10.1007/s00521-020-04831-9, 2020.
[35] M. Li, Y. Sun, H. Lu, S. Maharjan, Z. Tian, "Deep reinforcement learning for partially observable data poisoning attack in crowdsensing systems," IEEE Internet of Things Journal, vol. 7 no. 7, pp. 6266-6278, DOI: 10.1109/JIOT.2019.2962914, 2020.
[36] R. Podschwadt, H. Takabi, "On effectiveness of adversarial examples and defenses for malware classification," International Conference on Security and Privacy in Communication Systems, pp. 380-393, DOI: 10.1007/978-3-030-37231-6_22, 2019.
[37] D. Maiorca, B. Biggio, G. Giacinto, "Towards adversarial malware detection," ACM Computing Surveys (CSUR), vol. 52 no. 4,DOI: 10.1145/3332184, 2019.
[38] F. Kreuk, A. Barak, S. Aviv-Reuven, M. Baruch, B. Pinkas, J. Keshet, "Deceiving end-to-end deep learning malware detectors using adversarial examples," 2018. http://arxiv.org/abs/1802.04528
[39] Z. Fang, J. Wang, B. Li, S. Wu, Y. Zhou, H. Huang, "Evading anti-malware engines with deep reinforcement learning," IEEE Access, vol. 7, pp. 48867-48879, DOI: 10.1109/ACCESS.2019.2908033, 2019.
[40] J. Yuan, S. Zhou, L. Lin, F. Wang, J. Cui, "Black-box adversarial attacks against deep learning based malware binaries detection with GAN," the 24th European Conference on Artificial Intelligence (ECAI 2020), pp. 2536-2542, DOI: 10.3233/FAIA200388, .
[41] W. Song, X. Li, S. Afroz, D. Garg, D. Kuznetsov, H. Yin, "Automatic generation of adversarial examples for interpreting malware classifiers," 2020. http://arxiv.org/abs/2003.03100
[42] I. J. Goodfellow, J. Shlens, C. Szegedy, "Explaining and harnessing adversarial examples," , 2015. http://arxiv.org/abs/1412.6572
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2021 Fangwei Wang et al. This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
5G is about to open Pandora’s box of security threats to the Internet of Things (IoT). Key technologies, such as network function virtualization and edge computing introduced by the 5G network, bring new security threats and risks to the Internet infrastructure. Therefore, higher detection and defense against malware are required. Nowadays, deep learning (DL) is widely used in malware detection. Recently, research has demonstrated that adversarial attacks have posed a hazard to DL-based models. The key issue of enhancing the antiattack performance of malware detection systems that are used to detect adversarial attacks is to generate effective adversarial samples. However, numerous existing methods to generate adversarial samples are manual feature extraction or using white-box models, which makes it not applicable in the actual scenarios. This paper presents an effective binary manipulation-based attack framework, which generates adversarial samples with an evolutionary learning algorithm. The framework chooses some appropriate action sequences to modify malicious samples. Thus, the modified malware can successfully circumvent the detection system. The evolutionary algorithm can adaptively simplify the modification actions and make the adversarial sample more targeted. Our approach can efficiently generate adversarial samples without human intervention. The generated adversarial samples can effectively combat DL-based malware detection models while preserving the consistency of the executable and malicious behavior of the original malware samples. We apply the generated adversarial samples to attack the detection engines of VirusTotal. Experimental results illustrate that the adversarial samples generated by our method reach an evasion success rate of 47.8%, which outperforms other attack methods. By adding adversarial samples in the training process, the MalConv network is retrained. We show that the detection accuracy is improved by 10.3%.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details




1 College of Computer & Cyber Security, Hebei Normal University, Shijiazhuang 050024, China; Key Lab of Network & Information Security of Hebei Province, Shijiazhuang 050024, China
2 College of Computer & Cyber Security, Hebei Normal University, Shijiazhuang 050024, China