Introduction
Mobile devices, such as smartphones, iPads, and computer tablets, have become everyday necessities to perform important tasks, including education, paying bills online, bank transactions, job information, and leisure. Based on the information from an online mobile device production website, Android is one of the popular operating systems (OS) used by manufacturers (Rayner, 2019; Jkielty, 2019). The open-source platform in Android has facilitated the smartphone manufacturers in producing Android devices of various sizes and types, such as smartphones, smartwatches, smart televisions, and smart glasses. In the most recent decades, the quantity of remarkable Android gadgets accessible worldwide has increased from 38 in 2009 to over 20,000 in 2016 (Android, 2019a). As a result of the demand for this Android OS, the recent statistics from Statista revealed that the number of Android malware increase to 26.6 million in March 2018 (Statista, 2019). Moreover, McAfee discovered a malware known as Grabos, which compromises the Android and breaches Google Play Store security (McAfee, 2019). It was also predicted that 17.5 million Android smartphones had downloaded this Grabos mobile malware before they were taken down.
Mobile malware is designed to disable a mobile device, allow malicious acts to remotely control the device, or steal personal information (Beal, 2013). Moreover, these malicious acts able to run stealthily and bypass permission if the Android kernel is compromised by mobile malware (Ma & Sharbaf, 2013; Aubrey-Derrick Schmidt et al., 2009b). In September 2019, a total of 172 malicious applications were detected on Google Play Store, with approximately 330 million installations. According to researchers, the malicious components were hidden inside the functional applications. When the applications are downloaded, it leads to the appearance of popup advertisements, which remain appear even when the application was closed (O’Donnell, 2019). To detect this malware, security practitioners conducting malware analysis, which aims to study the malware characteristics and behaviour. There are dynamic, static, and hybrid analysis.
Table 1 shows comparison for static, dynamic and hybrid analysis done from previous researches. Specifically, dynamic analysis is an analysis, which studies the execution and behaviour of the malware (Enck, 2011; Yaqoob et al., 2019). However, dynamic analysis is incapable of identifying several parts of the code operating outside the monitoring range. Besides, provided that the dynamic analysis is a high resource-consuming analysis with a high specification for hardware (Enck, 2011), static analysis is another alternative to detect malware. It is an analysis, which examines malware without executing or running the application. Additionally, this analysis able to identify malware more accurately, which would act under unusual conditions (Castillo, 2011). This is due to static analysis examine overall parts of a program including parts that excluded in dynamic analysis. Furthermore, static analysis is able to detect unknown malware just as dynamic analysis could (Yerima, Sezer & McWilliams, 2014) and requiring low resources.
Year | References | Analysis | Features |
---|---|---|---|
2020 | Fang et al. (2020) | Static | Texture, color, text |
2019 | Qiu et al. (2019) | Static | permissions, API calls, network addresses |
2019 | Zhang, Thing & Cheng (2019) | Static | Assembly, Dex, Xml, Apk |
2019 | Xu, Ren & Song (2019) | Static | CFG, DFG |
2019 | Omid Mirzaeiq et al. (2019) | Static | API calls |
2019 | Vega & Quintián (0000) | Static | Repackaging and standalone |
2019 | Vega et al. (2019) | Static | Root node, decision nodes, and leaf nodes |
2019 | Fasano et al. (2019) | Static | |
2019 | Blanc et al. (2019) | Static | Code metric |
2019 | Xie et al. (2019) | Static | Platform-based permissions, hard- ware components, and suspicious API calls |
2019 | Turker & Can (2019) | Static | Permissions and API calls |
2018 | Atzeni et al. (2018) | Hybrid | Manifest file (i.e., number of activities, permissions, receivers, filters), and the source code analysis |
2018 | Kim et al. (2019) | Hybrid | API call |
2018 | Ming Fan et al. (2018) | Static | Weighted-sensitive-API-call-based graph |
2018 | Sun et al. (2018, 2019) | Dynamic | Enabling the recording of parameters and return value of an API call |
2018 | Martín, Rodríguez-Fernández & Camacho (2018) | Dynamic | transitions probabilities, states frequencies, and aggregated state frequencies grouped |
2018 | Aktas & Sen (2018) | Hybrid | number of activities, services and receivers given in the Manifest file and the size of the APK file |
2018 | Garcia, Hammad & Malek (2018) | Static | API usage, reflection-based features, and features from native binaries of apps |
2018 | Calleja et al. (2018) | Static | API calls, intent actions and information flow |
2018 | Alswaina & Elleithy (2018) | Static | App’s permissions |
2017 | Massarelli et al. (2017) | Dynamic | Fingerprint |
2017 | Zhou et al. (2017) | Static | API call graphs |
2017 | Chakraborty, Pierazzi & Subrahmanian (2020) | Hybrid | API calls, code, Android Manifest, encryption or reflection |
2017 | Sedano et al. (2017b) | Static | Minimum-Redundancy Maximum- Relevance (MRMR) |
2016 | Battista et al. (2016) | Static | Java Bytecode |
2016 | Hsiao, Sun & Chen (2016) | Dynamic | API call |
2016 | González, Herrero & Corchado (2017) | Static | API call and the names of functions and methods |
2016 | Ming Fan et al. (2018) | Static | Subgraph |
2016 | Kang et al. (2016) | Static | n-opcode feature |
2016 | Malik & Khatter (2016) | Dynamic | System call |
2016 | Sedano et al. (2017a) | Static | Manifest file, apk file |
2016 | Feng et al. (2017) | Hybrid | Malware signatures |
2015 | Lee, Lee & Lee (2015) | Static | Signature extraction signature matching |
2015 | Aresu et al. (2016) | Dynamic | Fine-grained HTTP structural |
2015 | Li et al. (2015) | Static | API data dependency |
2014 | Deshotels, Notani & Lakhotia (2014a) | Static | API call, apk |
2014 | Kang et al. (2013) | Static | Bytecode frequency |
2014 | Suarez-Tangil et al. (2014) | Static | Code structures |
DOI: 10.7717/peerjcs.522/table-1
To integrate the characteristics of the static and dynamic method, three-layer detection model called SAMAdroid has been proposed by Saba Arshad et al. (2018) which combines static and dynamic characteristics. Mobile Sandbox by Spreitzenbarth et al. (2015) which proposed to use the results of static analysis to guide the dynamic analysis and finally realize classification. The hybrid analysis technique is great to help in improving the accuracy, but it also has a major drawback such as the waste of time and space for the huge number of malware samples to be detected and analyzed (Fang et al., 2020; Alswaina & Elleithy, 2020).
Table 2 presents the past review articles on Android, with Feizollah et al. (2015) specifically focusing on features, including static, dynamic, hybrid, and metadata. It summarizes the features preferred researchers in their analysis. Comparatively, this study placed more emphasis on features besides classification and obfuscation. Subsequent reviews, namely Sufatrio et al. (2015) and Schmeelk, Yang & Aho (2015), highlighted the survey, taxonomy, challenges, advantages, limitations in the existing research in the Android security area, and the technique in the static analysis research on Android. However, compared to the current review, the aforementioned reviews only presented a few features and information on static analysis. In the Android permission category (Fang, Han & Li, 2014), this article reviewed the issues and existed countermeasures for permission-based Android security. It also incorporated the necessary information related to Android permission, namely permission documentation, over-claim of permission, permission administration, and more.
References | Ma & Sharbaf (2013) | Fang, Han & Li (2014) | Feizollah et al. (2015) | Sufatrio et al. (2015) | Schmeelk, Yang & Aho (2015) | Pan et al. (2020) | This paper |
---|---|---|---|---|---|---|---|
Title | Investigation of Static and Dynamic Android Anti-virus Strategies | Permission-based Android Security: Issues and Countermeasures | A Review on Feature Selection in Mobile Malware Detection | Securing Android: A Survey, Taxonomy, and Challenges | Android Malware Static Analysis Techniques | A Systematic Literature Review of Android Malware Detection Using Static Analysis | Malware Detection using Static Analysis for Android: A Review and Open Research Issue |
Year | 2013 | 2014 | 2015 | 2015 | 2015 | 2020 | Current paper |
Citations | 9 | 132 | 172 | 146 | 21 | 1 | |
Dataset | |||||||
Reverse engineer tools | |||||||
All static features | |||||||
All classifications (Machine learning, deep learning, graph, and others) | |||||||
Obfuscation constraints and methods to overcome it |
DOI: 10.7717/peerjcs.522/table-2
Notably, our review article featured more aspects besides permission by putting an in-depth focus on static analysis. Meanwhile, the other paper, evaluates three anti-viruses, namely Stowaway, AASandbox, and Droidbox. To assist in the decision regarding the better anti-virus, the aforementioned anti-viruses were separated using static and dynamic analyses, followed by a comparison between one another (Ma & Sharbaf, 2013). Another related review (Pan et al., 2020), provided the common features only and mentioned less the deep learning part in static analysis. However, the information gained from past research, specifically on features, deep learning and obfuscation was still lacking. In comparison to the previous survey articles listed in the Table 2, the main contributions of the current article are as follows:
The remaining section of this article begins with section two, which presents the methodology of this study into four steps. Section three reviews the existing research on static analysis, which concludes the dataset, reverse engineer, features, and classification. Section four discusses the open research issues in the static analysis, followed by section five, which concludes the review.
Survey Methodology Methodology
This section describes the method to retrieve the articles related to malware detection using static analysis for Android. We used Web of Science to run the review, eligibility and exclusion criteria, steps of the review process (identification, screening, eligibility), and data analysis.
Identification
The review was performed based on the main journal database in the Web of Science (WoS). This database covers more than 256 disciplines with millions of journals regarding the subjects related to network security, computer system, development, and planning. It also stores over 100 years of comprehensive backfile and citation data established by Clarivate Analytics (CA), which are ranked through three separate measures, namely citations, papers, and citations per paper. The search strings in the CA database were “static analysis”, “malware”, and “Android”.
There were 430 records identified through database searching. These journals and conferences are mainly from Computer and Security and IEEE Access, which are listed in Table 3. Collections of the studies that are related to Android malware detection using static analysis in the reference section, where studies take up a small proportion in the primary studies. All the studies related to search terms are taken into account, and the searching range is from January 2009 to December 2019.
Category | Acronym | Full name |
---|---|---|
Journal | – | IEEE Access |
– | Computers and Security | |
PICECE | Palestinian International Conference on Electrical and Computer Engineering | |
– | Computer Virology | |
– | ||
ASME | Manufacturing Science and Engineering | |
KSII | Internet and Information Systems | |
– | ||
– | Current Bioinformatics | |
– | Neural Computation | |
– | Frontiers of Information Technology and Electronic Engineering | |
– | Neurocomputing | |
JISA | Information Security and Applications | |
– | Advances in Intelligent Systems and Computing | |
– | IEEE Transactions on Information Forensics and Security | |
Conference | ACM | Conference on Multimedia |
DSC | IEEE International Conference on Data Science in Cyberspace | |
FSE | ACM SIGSOFT International Symposium on Foundations of Software Engineering | |
TrustCom | IEEE International Conference on Trust, Security and Privacy in Computing and Communications | |
GLOBECOM | IEEE Global Communications Conference | |
SIN | International Conference on Security of Information and Networks | |
MALWARE | International Conference on Malicious and Unwanted Software | |
ICC | IEEE International Conference on Communications | |
ICISSP | International Conference on Information Systems Security and Privacy | |
PIMRC | IEEE International Symposium on Personal, Indoor and Mobile Radio Communications | |
IJCNN | International Joint Conference on Neural Networks | |
Big Data | IEEE International Conference on Big Data |
DOI: 10.7717/peerjcs.522/table-3
Screening
Experiment articles were identified in the static analysis, omitting other unrelated articles. Initially, the searching of articles was specified into a journal article and excluded review articles, books, and conference proceedings. To focus specifically on static analysis, the articles, which combined both static and dynamic analyses, were removed. Another criterion focused on the selection of the articles was the use of English, which therefore removed all non-English articles to avoid any difficulty in translating in the future. The selection of articles took place from 2009 to 2019, totaling the duration to 10 years. This duration was suitable for exploring the evolution of research in security areas. Apart from that, the Android platform was the focus of this study.
Eligibility
Figure 1 depicts the review that process involved four steps; identification, screening, eligibility, and analysis. The review was performed in mid of 2019. Based on previous studies, the process used similar keywords related to malware detection, static analysis, and security. After the identification process, we remove any duplicated articles. During the screening process, we discover 375 documents and remove a few articles and left 172 articles. This is because the articles were unrelated to the interested area. Lastly, we used 150 articles for review (Shaffril, Krauss & Samsuddin, 2018).
Figure 1: The flow diagram of the study. Each step indicates the number of articles taken from Web of Science based on identification, screening, eligibility, and analysis processes. DOI: 10.7717/peerjcs.522/fig-1
Data analysis included
Then we analyzed the remaining articles, extracted the abstract, and downloaded the full articles. This is to find the appropriate topic in-depth and to have a strong justification for the research. Then, this process organized the topic and subtopic accordingly based on the static analysis. Qualitative analysis was performed based on content analysis to identify issues related to this study.
Static analysis
Mobile malware compromises Android devices (smartphone, smartwatch, and smart television) for wealth, stealing data, and personal purposes. The examples of mobile malware include root exploit, botnets, worms, and Trojan. To detect malware, most of security practitioners perform two types of analysis; dynamic and static.
Specifically, dynamic analysis is an experiment, which detects malware by executing malware and benign applications to monitor and differentiate their behaviours. However, the monitoring of all behaviours is costly and requires high specifications in terms of device memory, CPU, and storage. Furthermore, the malware is inflicted on a device at a certain time or whenever the attacker decides on it. Accordingly, as the dynamic analysis only monitors behaviours at a certain range of time based on the research period, numerous malware activities outside the research period might be omitted (Feizollah et al., 2013; Yerima, Sezer & Muttik, 2015; Wei et al., 2017). Furthermore, dynamic analysis requires a separate and closed virtual environment to run a malware and observe its behaviour on the system. However, an isolated setup dynamic leads to an impractical analysis in the Android platform due to the increase in power and memory consumption. While power and memory are the most concerning constraints of Android devices, static analysis is the alternative for the dynamic analysis.
Static analysis is a category of analysis, which investigates the malware application code and examine full activities in an application within an unlimited range of time, by without executing the application (Chess & McGraw, 2004). The main step of static analysis procedure is the reverse engineer process, which retrieves the whole code and further scrutinises the structure and substance within the application (Sharif et al., 2008; Chang & Hwang, 2007; Aafer, Du & Yin, 2013). Therefore, this analysis can examine the overall code with low requirement for memory resources and minimum CPU processes. Additionally, the analysis process is prompt due to the absence of the application. With this analysis, unknown malware is also identified using enhanced detection accuracy through machine learning approaches (Narudin et al., 2014; Feizollah et al., 2013). Table 4 presents the advantages and disadvantages of dynamic and static analyses.
Dynamic | Static | Hybrid |
---|---|---|
Advantages | ||
Able to detect unknown malware | Able to detect unknown malware with the aid of machine learning | Able to detect unknown malware with combination of static and dynamic analysis |
Able to detect benign applications, which abruptly transform into malware during its execution | The application of reverse engineer takes a short amount of time | |
The examination on the overall code, followed by the identification of a possible action | ||
Low resources (e.g., CPU, memory, network, and storage). Therefore, this analysis is suitable for mobile device which equipped with low specifications. | ||
Limitations | ||
High resources (e.g., CPU, memory, network, and storage) | Inability to detect normal application, which promptly transforms the malware | Waste of time |
Higher time consumption to run the application for further analysis and exploration | Obfuscation | Require more spaces for huge number of malware samples |
Possibly omits the malware activities outside the analysis range | The investigation is continued to determine the minimal features (e.g., permission, a function call, and strings) to detect malware | |
Difficulty in detecting applications, which can hide malicious behaviour when it is operated | ||
The investigation is continued to determine the minimal features (e.g., traffic and memory) to detect malware |
DOI: 10.7717/peerjcs.522/table-4
A lot of researchers publish their works using static approaches for malware detection on the Android platform. Even in this static approach, in its turn, contains a number of approaches. For example, there are signature-based approach and other approach are depending on detection and classification of the source code. Signature-based detection utilizes its specification by having an information of malware signatures determined and arranged in advance inspection (Samra et al., 2019). However, signature-based approach are not able to detect unknown malware even though this approach is a set of features that uniquely differentiate the executable code (Gibert, Mateu & Planes, 2020).
Obfuscation is one of the obstacles in the static analysis, which is used by malware authors in their malicious software to evade the intrusion detection or antivirus system (Wei et al., 2017). The examples of the obfuscation methods are renaming the code, adding unnecessary codes, and encrypting the string. Therefore, security practitioners need to overcome obfuscation to increase their detection results. Accordingly, the alternatives performed by the security practitioners are presented in ‘Obfuscation’.
Table 4 shows that both static and dynamic analyses have similar limitations despite the selection of the ideal features in minimal amount. In detecting malware, features refer to the attributes or elements to differentiate an application, which may either be malware or benign. Security practitioners are faced with obstacles in investigating various features in all types of categories (e.g., permission, API, directory path, and code-based) and the need to simultaneously reduce these features. Notably, determining the ideal features in minimal amount is crucial to enhance the accuracy of the analyses (e.g., the accuracy of the predictive model) and reduce data and model complexity (Feizollah et al., 2015).
Figure 2: Malware detection using static analysis. The static analysis operation, which consisted of several steps. The steps included dataset collections, reverse engineer, features identification, and classification. DOI: 10.7717/peerjcs.522/fig-2
Figure 3: Dataset of Android samples. This dataset collected from five main dataset group samples. The datasets were used by the previous researchers for their works too. DOI: 10.7717/peerjcs.522/fig-3
Figure 2 illustrates the static analysis operation, which consisted of several steps. The first step was the acquirement of the benign and malware datasets in the Android application, each with the (.apk) filename extension. This was followed by the reverse engineering performed on these applications to retrieve the code by extracting a few folders from one .apk file, which consisted of nested files with codes (Java or smali). Furthermore, one .apk would comprise approximately a thousand lines of codes. Therefore, with a total of 1,000 applications in one dataset, the security practitioners were required to scrutinise millions of lines of code. With the completion of the reverse engineering, an analysis would be conducted, which involved features. Features consist of a series of application characteristics for the detection of malware, while classification is an approach used to differentiate between malware and benign (normal) application. The following section thoroughly discusses the static analysis, which specifically begins with a focus on the dataset.
Dataset
Figure 3 illustrates the Android malware dataset from different places. Notably, the majority of the datasets were obtained from universities. The datasets were in the form of an Android application package, which was followed by an .apk filename extension. Malgenome (Anonymous, 0000d) is the name of Android malware dataset, which was made to be publicly available with permission from their administrator. These malware samples, which were collected by North Carolina State University (NCSU) from August 2010 to October 2011, covered multiple families of malware consisting of botnet and root exploit. The characterization of the malware families was based on the method of the installation, the way the malware carried the malicious payloads, and its method of activation.
Androzoo (Allix et al., 2018; du Luxembourg, 2016) is another dataset consisting of approximately more than three million of Android applications (.apk). This dataset originates from the University of Luxembourg to contribute to the community for research purposes and further explore the notable development in the detection of malware, which damages the Android. Drebin (Technische Universität Braunschweig, 2016) dataset also presents Android malware publicly with strict requirements. A university from Germany (University in Braunschweig, Germany) collected 5,560 samples with 179 families. The time range provided for the malware was from August 2010 to October 2012. The university project, which was known as MobileSandbox, was an initiative for the acquirement of samples for academia and industry.
Android malware dataset (AMD) is a public Android malware dataset from the University of South Florida, which consists of 24,650 samples with 71 categorised families. To obtain this dataset, the user is required to acquire permission from the university and provide authentic information with evidence. The academia and the industry are allowed to use these samples for research purposes.
Contagio (MilaParkour, 2019) dataset presents the malware, which focuses on mobile malware, with a condition that the user should omit one sample to obtain another sample. It provides a dropbox for the user to share their mobile malware samples. According to their blogspot (MilaParkour, 2019), the name of the administrator of this dataset is Mila Parkour, who is reachable only through emails. Based on Table 5, which presents the research articles and the respective datasets, it could be seen that the dataset providers receive significant attention from other universities and the industry. It is hoped that this action would enhance the security of the Android device and its users from time to time.
Dataset | References of the articles with the use of the respective datasets |
---|---|
Malgenome Anonymous (0000d) | Yerima, Sezer & McWilliams (2014), Firdaus et al. (2017), Firdaus & Anuar (2015), Firdaus et al (2018) |
Drebin Anonymous (0000c) | Firdaus et al. (2017); Firdaus et al. (2018); Firdaus et al. (2018) |
Android malware dataset (AMD) | Badhani & Muttoo (2019) |
Contagio MilaParkour (2019) | Feldman, Stadther & Wang (2014); Islamic & Minna (2015) |
Androzoo Université du Luxembourg (2018) | Razak et al. (2019); Firdaus et al. (2017); Razak et al. (2018); Firdaus et al. (2017) |
DOI: 10.7717/peerjcs.522/table-5
Reverse engineer
Static analysis is an activity to investigate the code of an application without executing it. In order to investigate, security practitioners implement the reverse engineering method. This method reversed from the executable file to its source code (Dhaya & Poongodi, 2015). This reverse engineering process loads the executable into a disassembler to discover what the program does. Figure 4 illustrates the tools used to perform a reverse engineering method, which was also adopted by security practitioners to identify Android malware. Table 6 illustrates the tools adopted in the respective articles.
Figure 4: Reverse engineer tools for static analysis. This is the example of reverse engineer tools that have been used by the previous researchers to extract the code for malware. DOI: 10.7717/peerjcs.522/fig-4
Tools | References to the articles and the respective tools |
---|---|
apktool | Wiśniewski (2010); Deshotels, Notani & Lakhotia (2014b); Wu et al. (2012); Faruki et al. (2013); Luoshi et al. (2013) |
aapt | Android (2013); Sanz et al. (2013) |
androguard | Desnos (2012a); Suarez-Tangil et al. (2014); Sahs & Khan (2012); Aafer, Du & Yin (2013); Junaid, Liu & Kung (2016a) |
baksmali | Anonymous (2019b); Apvrille & Strazzere (2012); Huang, Tsai & Hsu (2012); Grace et al. (2012b); Zhou et al. (2013); Zhou et al. (2012); Grace et al. (2012a) |
dex2jar | Anonymous (2019d); Huang et al. (2014a); Sheen, Anitha & Natarajan (2015); Lee & Jin (2013); Luoshi et al. (2013) |
jadx | Skylot (2015); Firdaus & Anuar (2015) |
dedexer | Anonymous (0000b) |
smali | Kang et al. (2013); Anonymous (2019b) |
DOI: 10.7717/peerjcs.522/table-6
Features
Once the researchers reverse engineer the executable file using specific tools, they need to select features from the source code. Feature selection is important in order to increase the accuracy of the detection system (Feizollah et al., 2015; Chanda & Biswas, 2019; Klaib, Sara & Hasan, 2020) Figure 5 presents the taxonomy of multiple static features. The next sections are the details for each type of static feature.
Figure 5: Taxonomy of multiple static features. Each static feature was figure out from the various experiments done using the specific tools and methods. DOI: 10.7717/peerjcs.522/fig-5
Advertisement libraries.
Provided that most Android applications are available for free download, Android developers need to include advertisement libraries (ad libraries) in the free application for financial purposes. During the run-time of the application, the ad libraries would transfer the data regarding users’ activities. The developer would then receive an incentive based on certain metrics of the information. Adrisk (Grace et al., 2012a) scrutinised and measured the risk of the codes of the ad libraries to detect applications, which may harm the users.
Application programming interface.
Application program interface (API) is a set of code ready for certain functionalities. Android application developers use this API for their application. Usually, there is documentation ready to use the API. Nevertheless, the attacker uses certain API for their malware application. Accordingly, to detect malware, security practitioners inspect API features that regularly used by the attackers. The articles that involves API features are Droidlegacy (Deshotels, Notani & Lakhotia, 2014b), Droidapimiiner (Aafer, Du & Yin, 2013), Warning system (Lee & Jin, 2013; Chang & Hwang, 2007; Wu et al., 2012; Bartel et al., 2012; Grace et al., 2012b; Wu et al., 2012; Shuying Liang et al., 2013; Zhou et al., 2013; Feng et al., 2014; Huang et al., 2014; Steven Arzt et al., 2014; Arp et al., 2014; Luoshi et al., 2013; Yerima, Sezer & Muttik, 2014; Seo et al., 2014; Sheen, Anitha & Natarajan, 2015; Peiravian & Zhu, 2013).
Apk, dex and xml properties.
Several security practitioners adopted the features, which consist of .apk file properties. The authors of the malwares (Kang et al., 2015) and (Zhou et al., 2012) are examined in two experiments due to the significant number of Android malwares written by a similar person. Therefore, the features of the malwares include serial numbers of the author, author’s information, name, contact and organization information, developer certification, author’s ID, and public key fingerprints of the author. Other features highlighted in this section are the application name, category, package, description, rating values, rating counts, size, number of zip entries, and common folders (Samra, Kangbin & Ghanem, 2013; Shabtai, Fledel & Elovici, 2010).
Directory path.
Directory path allows access for a specific folder in the operating system (OS). It was found by security practitioners that the attacker incorporated a directory path for a sensitive folder in their malware. Meanwhile, several paths related to Android kernel directory were identified by another study (Firdaus & Anuar, 2015), such as ‘data/local/tmp/rootshell’, ‘/proc’, and ‘/system/bin/su’.
Commands.
Two types of commands are available, namely (1) root command and (2) botnet command. Specifically, several root commands were identified by (Firdaus & Anuar, 2015) in the Unix machine, such as ‘cp’, ‘cat’, ‘kill’, and ‘mount’. Normally, these commands were used by the administrators to execute higher privileged actions in the Unix machine. Provided that Android architecture was based on the Unix kernel, the attackers included root commands in their malware to control the victim’s Android devices. Therefore, the identification of root commands is crucial in investigating malwares.
The second type of command is a botnet command. Meanwhile, one type of malware, which is known as a mobile botnet, includes botnet commands in their malware codes, such as ‘note’, ‘push’, ‘soft’, ‘window’, ‘xbox’, and ‘mark’. The attacker used these commands to communicate with the command and control (C&C) server, while droidanalyzer (Seo et al., 2014) combines API, root command, and botnet command into a set of features to detect root exploit and mobile botnet.
Other than ad libraries, certain researchers inspect the Android Debug Bridge (adb) code. ADB (Android Developers, 2017) is a tool, which provides a command-line access facility for users or developers to communicate with Android mobile devices. This facility allows the installation of unwanted applications and execution of various Unix by the attacker in the victim’s device. Therefore, RODS (Firdaus et al., 2018) is a root exploit detection system for the detection of a root exploit malware with ADB features.
Function call.
In programming, a function call is a declaration, which consists of a name and is followed by an argument in parenthesis. The list of the argument may include any numbers of the name, which are either separated by commas or left empty. Another study by Aubrey-Derrick Schmidt et al. (2009a) involved the extraction of a function call through readelf, which was then used for the features in machine learning prediction. Meanwhile, Gascon et al. (2013) extracted the function calls in a graph to identify the nodes from the start to the end of the process.
Geographic location.
Geographic location is a feature, which identifies the origin of the application. The geographic detector was identified as one of the features in research by (Steven Arzt et al., 2014). Provided that 35% of the mobile malware families appeared to originate from China with 40% of the facilities originating from Russia, Ukraine, Belorus, Latvia, and Lithuania countries, it was crucial to consider geographic location as one of the features for the detection of Android malware. For this reason, researchers increased the risk signal for the applications originating from the aforementioned countries.
Manifest file.
Android application is built on the top of the application framework which provides an interface for the user. The program is based on the Android application package file in the (.apk) format, which is also used to install an application in android-based mobile devices. It consists of meta-inf, resource, assets and library directory, classes.dex, resources.arsc, and androidmanifest.xml file. One of the files, androidmanifest.xml (manifest file), is an essential file with contents of various features, such as permission, intent, hardware component, and components of the application (activities, services, broadcast receivers, and content providers) (Android, 2015).
(a) Permission
Permission is a unique security mechanism for Android devices. To enable the permission, the user needs to allow the application during the installation period. However, many users accidentally enable certain permissions, which leads to access to sensitive security-relevant resources. Therefore, permission features were examined in many studies. Based on the application of permission in several studies to measure the risk of the application, permission was further identified as malicious (Razak et al., 2018; Razak et al., 2019). Some other studies, such as (Hao Peng et al., 2012; Samra, Kangbin & Ghanem, 2013; Walenstein, Deshotels & Lakhotia, 2012; Huang, Tsai & Hsu, 2012; Sahs & Khan, 2012; Sanz et al., 2013; Talha, Alper & Aydin, 2015; Aung & Zaw, 2013), used the permission features as the inputs for machine learning prediction.
(b) Intent
The intent is coded in the manifest file and allows a component of the application to request certain functionality from another component from other application. For example, application A can use the component of application B for the management of photos in the device despite the exclusion of the component from application A. Provided that this feature enables malicious activities among the attackers, several experiments used intent (declared in the manifest file) as one of the features for the detection of malware, such as (Feizollah et al., 2017; Fazeen & Dantu, 2014).
(c) Application component
The manifest file declared application component, which consists of four types, namely (1) activities, (2) services, (3) broadcast receivers, and (4) content providers. Specifically, activity is represented as the user interface or interactive screen to the users, while service refers to an operation occurring in the backgrounds, which perform the long-service process. This is followed by broadcast receivers, which respond to system-wide broadcast announcements. On the other hand, content providers manage a structured set of application data. Overall, these four components follow a life cycle model during execution. Dexteroid (Junaid, Liu & Kung, 2016) proposed a framework, which systematically guides the event sequences through the reverse engineering/reconstruction of the life cycle models and the extraction of callback sequences from event sequences to detect malicious behaviours.
(d) Hardware component
The manifest file also incorporated hardware components in the Android application. To illustrate, the developer requested access to the camera of an Android device by declaring it in the manifest file to enable the use of the camera for the application. However, the attacker declared unrelated hardware components in their game application, such as camera and data. As a result, the security researchers were prompted to use hardware component as the features in their experiment (Arp et al., 2014) to detect malware (Allahham & Rahman, 2018).
Network address.
Access to the Internet is essential for attackers to retrieve private information of the victim, change the settings, or execute malicious commands. This process requires the incorporation of the Uniform Resource Locator (URL) or network address in the malware code. The examples of sensitive URLs include the Android Market on Google Play, Gmail, Google calendar, Google documents, and XML schemas. These features were used in Luoshi et al. (2013) and Apvrille & Strazzere (2012), Mohd Azwan Hamza & Ab Aziz (2019) for malware detection.
Code-based.
Code-based or code structure comprises a line or set of programming language codes in an application. Two studies applied code structures (code chunk grammar) as the features for malware detection, which is focused on the internal structure of the code units (Suarez-Tangil et al., 2014; Atici, Sagiroglu & Dogru, 2016). This feature enables the analysis and differentiation between malware and benign applications. Another study by Firdaus & Anuar (2015) identified several code-based strings, namely ‘.exec’, ‘forked’, ‘setptywindowsize’, and ‘createsubprocess’. In comparison with the normal application, it was found that the attacker frequently used these code-based features in the development of malware. Therefore, these features were also used in this study to detect malware.
Opcode (operation code) is another code-based feature. It is a part of the instruction to inform the CPU regarding the tasks to be fulfilled. Assembly language used this opcode to execute the instruction. Also referred to as bytecode, the examples of an opcode for Android included OP_ADD_DOUBLE, OP_ADD_FLOAT, OP_ADD_INT_2ADDR, and OP_SUB_LONG (Developer, 2020). Specifically, this feature was adopted in the studies by Zheng, Sun & Lui (2013), Medvet & Mercaldo (2016), Faruki et al. (2013) and Zhao et al. (2019) to detect Android malware in the static analysis. Further examples of the features in this section are method (Kim et al., 2018), opcode (Zhao et al., 2019), byte stream @ byte block (Faruki et al., 2013), Dalvik code (Gascon et al., 2013), and code involving encryption (Gu et al., 2018). The selection of the features by security practitioners is followed by classification. This process was performed to receive the features as input and differentiate between either the application malware or benign (normal).
Figure 6 depicts that researchers prefer to investigate permission and API features compare to others. However, the trend in permission features is decline from 2013 until 2018. However, API features takes place in previous experiments as it increased from six (2014) to 9 (2019). This indicates that the API trend would increase in following year in static detection.
Figure 6: Categories of features in total and years. (A) The graph shows the features identified using static analysis, depicts that researchers prefer to investigate permission and API features compare to others. (B) Total number of features based on years. DOI: 10.7717/peerjcs.522/fig-6
Classification
In the classification process for static analysis, many security analysts used two types of methods; (1) Machine learning (ML) and (2) Graph. The following section presents the ML studies with static features.
Machine learning (ML).
Machine learning is a scientific discipline, which is capable to predict future decisions based on the experience it has gained through past inputs (learning set), followed by a prediction of the outputs. Basing on a given dataset, the learning set makes intelligent decisions according to certain algorithms. One of the machine learning types is supervised based on the data for the training stage to create a function. Furthermore, each part of the training data contains input (features or characteristics) and output (class label-malware and benign). This is followed by the training stage, which calculates the approximate distance between the input and output examples to create a model. This training stage could classify unknown applications, such as malware or benign application. Four types of ML are present, such as (1) classical learning; (2) reinforcement learning, (3) neural network and deep learning, and (4) ensemble method. Figure 7 illustrates the ML taxonomy, which starts with classical learning.
Figure 7: Mining in static analysis. A few types of machine learning available, with each type, has its own classifier as listed in figure. DOI: 10.7717/peerjcs.522/fig-7
(a) Supervised Learning
Supervised learning (SL) is a process of learning from previous instances to predict future classes. Therefore, the prediction of the class label involves the construction of a concise model from previous experience. The machine learning classifier is then used to test the unknown class (Kotsiantis, 2007). To detect Android malware with static features, the SL method is widely used by security practitioners. Accordingly, the previous articles adopting this method are illustrated in Table 7.
Machine learning type | Classifier | Reference |
---|---|---|
Supervised | K-nearest neighbor | Anonymous (2019b); Aafer, Du & Yin (2013) |
Support vector machine (svm) | Aafer, Du & Yin (2013); Sahs & Khan (2012); Huang, Tsai & Hsu (2012); Arp et al. (2014); Anonymous (2019b); Anonymous (2019d); Arp et al. (2014) | |
CART | Aung & Zaw (2013) | |
Adaboost | Huang, Tsai & Hsu (2012); Sheen, Anitha & Natarajan (2015) | |
Bayes | Yerima, Sezer & McWilliams (2014); Lee, Lee & Lee (2015); Hao Peng et al. (2012); Sanz et al. (2013); Shabtai, Fledel & Elovici (2010) | |
Logistic Regression | Talha, Alper & Aydin (2015) | |
Prism (PART) | Aubrey-Derrick Schmidt et al. (2009a); Yerima, Sezer & Muttik (2014); Shabtai, Fledel & Elovici (2010) | |
Voting feature interval (vfi) | Shabtai, Fledel & Elovici (2010) | |
Random forest | Shabtai, Fledel & Elovici (2010); Sanz et al. (2013); Aafer, Du & Yin (2013); Huang, Tsai & Hsu (2012) | |
Sequential minimal optimisation (smo) | Sanz et al. (2013) | |
Instance-based learning with parameter k (ibk) | Sanz et al. (2013) | |
Simple logistic | Sanz et al. (2013) | |
Multilayer perceptron | Firdaus & Anuar (2015) | |
Unsupervised | K-means | Fan et al. (2019); Wu et al. (2012); Samra, Kangbin & Ghanem (2013); Aung & Zaw (2013) |
Normalised Compression distance (NCD) | Lu et al. (2012); Crussell, Gibler & Chen (2012) |
DOI: 10.7717/peerjcs.522/table-7
(b) Unsupervised Learning
Unsupervised learning is another type of learning involved in machine learning. It is a clustering technique where the data is unlabeled and has also been used in computer security areas, including malware detection and forensic (Beverly, Garfinkel & Cardwell, 2011). Clustering refers to the division of a large dataset into smaller data sets with several similarities. It classifies a given object set through a certain number of clusters (assume k clusters) to determine the k centroids assigned for each cluster. In this case, this algorithm selects the centroid by random from the applications set, extracts each application from a given dataset, and assigns it to the nearest centroid. Table 7 tabulates the previous articles, which adopted this method.
(c) Reinforcement learning
A reinforcement learning model consists of an agent (a set of actions A) and an environment (the state space S) (Anderson et al., 2018). Deep reinforcement learning was introduced by reinforcement agents as a framework to play Atari games, which often exceed human performance (Volodymyr Mnih et al., 2013; Volodymyr et al., 2015). The advances in deep learning may extract high-level features from raw sensory data, leading to breakthroughs in computer vision and speech recognition. In the case of deep learning, the agent would be required to learn a value function in an end-to-end way, which takes raw pixels as input and predicts the output rewards for each action.
The learned value function is called deep Q learning, in which Q function is learned and refined from over hundreds of games (Anderson, Filar & Roth, 2017). The Q-learning algorithm was trained in network (Volodymyr Mnih et al., 2013) with stochastic gradient descent to update the weights. Replay mechanism was used from random samples previous transitions to lead smooth training distribution over past behaviors to overcome the correlated data and non-stationary distributions problems. Anderson et al. (2018) propose a framework based on reinforcement learning (RL) for attacking static portable executable (PE) anti-malware engines. Meanwhile, a DQN-based mobile proposed by Wan et al. (2018) to enhance the malware detection performance. The results shown from simulation can increase the malware detection accuracy and reduce the detection delay as compared to a Q-learning based malware detection scheme.
(d) Neural Network and Deep Learning
The evolution of Neural Network (NN) has been associated with various challenges since the mid-20th century. McCulloch and Pitts obtained the first inspiration of NN in 1943 from biological neurons, which was followed by proposing a computational model for the development of hypothetical nets. Although this proposal was simulated by Nathaniel Rochester at IBM research laboratory, this attempt was unsuccessful at the end. Developed by Frank Rosenblatt at Cornell Aeronautical Laboratory, the perceptron became the first learning machine (Mahdavifar & Ghorbani, 2019).
Despite all the upgrades on NNs, Deep learning (DL) was developed in 2006 and has been used in almost every application. As a new variation of the classical Multilayer Perceptron (MLP), the DL aims to produce high-level and flexible features from the raw pixel data to assist in generalising the classification. Furthermore, DL also operates with complex applications containing millions of data, which require a large number of neurons and hidden layers. A few DL frameworks have been developed in the recent years, such as TensorFlow (Abadi, Agarwal & Barham, 2016), Caffe (Yangqing Jia et al., 2014), and Theano (Al-Rfou et al., 2016) to ensure an efficient implementation of Deep Network (DN) architectures and omit the unnecessary coding scratching (Mahdavifar & Ghorbani, 2019). Additionally, the DL method extracts the features based on the layer’s level, which could either be high or low.
Figure 8 depicts the differences between ML and DL. It shows that ML requires the security practitioners to extract the features manually and select the ML classifier, which is suitable for the selected features. However, DL involves automatic feature extraction part and malware classification. It trains the model end-to-end with the Android application package (.apk) file and their categories, each labelled as malware or benign. The DL gains and creates a prediction model through the automatic selection of the feature.
Figure 8: Differences between ML and DL. (A) ML extract the features manually and select the ML classifier. (B) DL involves automatic feature extraction and malware classification. DOI: 10.7717/peerjcs.522/fig-8
As one of the major models in deep learning, a convolutional neural network (CNN) has been widely used for image recognition (Li et al., 2018). It could be seen in the past few years that many studies have implemented Deep Neural Networks (DNN) to classify malware (Pascanu et al., 2015; Saxe & Berlin, 2016; Zhao et al., 2019). Additionally, although the recurrent neural networks have been explored since the 1980s, they have become uncommercial due to several issues (Pascanu et al., 2015). Several machine learning methods have addressed network or malware attacks on personal computers or mobile devices. Simultaneously, several techniques were proposed by researchers who applied DL algorithms to detect or categorize malware using static, dynamic, or hybrid approaches, detection of network intrusions and phishing/spam attacks, and inspection of website defacements (Venkatraman, Alazab & Vinayakumar, 2019).
(e) Ensemble method
Another technique in machine learning and pattern recognition is ensemble learning. The increase in the implementation of ensemble learning methods could be seen in the computational biology field due to the unique advantages in managing small sample size, high dimension, and complex data structures (Yang et al., 2010). The function of ensemble learning is to build a prediction model by combining the strengths of the collection of simpler base models (Hastie, Tibshirani & Friedman, 2009). A few approaches are applied in ensemble methods, such as bagging, boosting, and random forest. This method is also a simple device, which is popular especially in the predictive performance of a base procedure.
The bagging procedure appears to be a variance reduction scheme for some base procedure, while the boosting methods mainly reduce the bias of the base procedure. Therefore, the significant difference between bagging and boosting ensemble methods is indicated. Compared to bagging and boosting, the random forest approach is a highly distinguished ensemble method. The first proposal of the random forest was made by Amit & Geman (1997). While the performance of random forests is on the same level as boosting, it could exhibit better performance in the perspective of prediction.
Year | Ref | Machine Learning used ( ) | |||||
---|---|---|---|---|---|---|---|
Classical learning | Reinforcement learning | Neural network & deep learning | Ensemble method | Others | Metrics | ||
2019 | Yildiz & Doǧru (2019) | Support Vector Machine (SVM) | |||||
2019 | Singh, Jaidhar & Kumara (2019) | Linear Support Vector Machine (L-SVM) | |||||
2019 | Lei et al. (2019) | k-means | |||||
2018 | Firdaus et al. (2018) | Logitboost | |||||
2018 | SL et al. (2015) | Hotbooting-Q | |||||
2018 | Yangqing Jia et al. (2014) | Stochastic gradient descent (SGD) | |||||
2018 | Firdaus et al. (2017) | Multilayer Perceptron (MLP), Voted Perceptron (VP) and Radial Basis Function Network (RBFN). | |||||
2018 | Firdaus et al. (2018) | Naïve Bayes (NB), functional trees (FT), J48, random forest (RF), and multilayer perceptron (MLP) | |||||
2018 | Li et al. (2018) | Bayesian calibration | |||||
2017 | Anderson, Filar & Roth (2017) | Direct gradient-based, White-box, Binary black-box | |||||
2017 | Volodymyr Mnih et al. (2013) | Black-box | |||||
2016 | Junaid, Liu & Kung (2016) | Control flow graph | |||||
2016 | Wan et al. (2018) | FCNN, DistBelief | |||||
2016 | Abadi, Agarwal & Barham (2016) | FCNN | |||||
2015 | Pascanu et al. (2015) | Sequential Minimal Optimization (SMO), Support Vector Machine (SVM) | |||||
2015 | Kang et al. (2015) | Naïve Bayes | |||||
2015 | Talha, Alper & Aydin (2015) | Support Vector Machine (SVM) | |||||
2015 | Anderson et al. (2018) | Deep Q-network | |||||
2015 | Al-Rfou et al. (2016) | Echo state networks (ESN) | |||||
2014 | Deshotels, Notani & Lakhotia (2014a) | Class Dependence Graph (CDG) | |||||
2014 | Feng et al. (2014) | Inter-Component Call Graph | |||||
2014 | Huang et al. (2014) | Control flow graph (CFG) and call graph (CG) | |||||
2014 | Steven Arzt et al. (2014) | Inter-procedural control-flow graph (ICFG) | |||||
2014 | Mahdavifar & Ghorbani (2019) | FCNN, MNIST digit | |||||
2013 | Shuying Liang et al. (2013) | Entry-Point Saturation (EPS) | |||||
2013 | Zhou et al. (2013) | Vantage Point Tree (VPT) | |||||
2013 | Beverly, Garfinkel & Cardwell (2011) | Q-learning | |||||
2012 | Amit & Geman (1997) | Program dependence graphs (PDGs) | |||||
2010 | Saxe et al. (2015) | Jaccard set-based index |
DOI: 10.7717/peerjcs.522/table-8
Table 8 shows previous works done using different types of machine learnings as mentioned before. From the table, we can summarize classical learning is still valid to be used in experiment execution but there are a lot of works are using deep learning and graph method. The current trends show the demand using the deep learning technique to defend against an increasing number of sophisticated malware attacks where deep learning based have become a vital component of our economic and national security. Many recent studies on Android malware detection have leveraged graph analysis as mentioned in the next section.
Graph.
The use of a graph is another method in machine learning and pattern recognition, which is performed by investigating the data and control-flow analysis. It is also capable of identifying unknown malware through the examination on the flow of the code. This method is preferred by security analysts due to the uniform flow despite the changes made by the malware authors on the API calls to avoid intrusion detection systems. The types of analysis in graph method include call graph, inter-component call graph (ICCG), control-flow graph (CFG), and dependence graph, while Table 9 lists the previous works of research on static malware detection using the graph method.
A call graph (specifically known as flow graph) is a graph representing the control and data flow of the application, which investigates the exchange of information through the procedures. A node in the graph represents a procedure or function, as seen from the x and y symbols, which indicate that procedure x calls for procedure y. Apposcopy (Feng et al., 2014) presents its new form of call graph known as inter-component call graph (ICCG) to match malware signature. As a directed graph where nodes are known as components in an application, it establishes ICCG from a call graph and the results of the pointer analysis. The objective of apposcopy is to measure the inter-component communication (ICC), calls, and flow relations.
Another graph called a control flow graph (CFG) is also applied by many security analysts to investigate the malware programme. Woodpecker (Grace et al., 2012b) created the CFG start from each entry point (activity, service, receiver, content provider), which is defined in the permission stated in the androidmanifest.xml file. Furthermore, the public interface or services from an execution path is discovered through the flow graph. However, it would be considered by Woodpecker as a capability leak if it is not guarded by the permission requirement nor prevented from being invoked by another unrelated application. The same graph was applied in subsequent works of research, namely Flowdroid (Steven Arzt et al., 2014), Dendroid (Suarez-Tangil et al., 2014; Sahs & Khan, 2012), Asdroid (Huang et al., 2014a), Anadroid (Shuying Liang et al., 2013), Adrisk (Grace et al., 2012a), and Dexteroid (Junaid, Liu & Kung, 2016a).
Another graph is the dependency graph, which illustrates the dependencies of several objects on each other. An example could be seen in the dead code elimination case process, in which the graph identifies the dependencies between operation and variables. With the dependency of non-operation on certain variables, these variables would be considered dead and should be deleted. The studies, which adopted this type of graph are CHEX (Lu et al., 2012), Dnadroid (Crussell, Gibler & Chen, 2012), Droidlegacy (Deshotels, Notani & Lakhotia, 2014b; Zhou et al., 2013).
Type of graph | Reference |
---|---|
Call graph | Copes (Bartel et al., 2012), Leakminer (Yang & Yang, 2012), Riskranker (Grace et al., 2011), A3 (Luoshi et al., 2013) and (Gascon et al., 2013) |
Inter-component call graph (ICCG) | Feng et al. (2014) |
Control flow graph (CFG) | Woodpecker (Grace et al., 2012b), Flowdroid (Steven Arzt et al., 2014), Dendroid (Suarez-Tangil et al., 2014; Sahs & Khan, 2012), Asdroid (Huang et al., 2014a), Anadroid (Shuying Liang et al., 2013), Adrisk (Grace et al., 2012a), and Dexteroid (Junaid, Liu & Kung, 2016a) |
Dependency graph | CHEX (Lu et al., 2012), Dnadroid (Crussell, Gibler & Chen, 2012), Droidlegacy (Deshotels, Notani & Lakhotia, 2014b) and (Zhou et al., 2013) |
DOI: 10.7717/peerjcs.522/table-9
Others.
Besides machine learning and graph, several security practitioners adopted different methods, such as Normalized Compression Distance (NCD). Adopted in the studies by Desnos (2012b) and Paturi et al. (2013), this method can measure the similarities between the malwares and represent them in the form of a distance matrix. Despite the evolution of many malwares from time to time, some of their behaviour patterns are similar to each other. The calculation of the similarities using NCD would identify the malwares, which share the same distance.
A study known as DelDroid (Hammad, Bagheri & Malek, 2019) implemented a method called as Multiple-Domain Matrix (MDM). This method refers to a complex system, which calculates multiple domains and is based on the Design-Structure Matrix (DSM) model. Furthermore, MDM is formed by the connection of DSM models with each other. The study initialised multiple domains in the MDM to represent the architecture of an Android system for privilege analysis. To illustrate, the incorporation of certain definitions in the MDM representation in the architecture enables DelDroid to identify the communication of the application, which may result in an unauthorised malware attack.
Another previous static experiment was conducted on the MD5 signature of the application to detect malware (Seo et al., 2014). In the first process, the study assigned the application as level C (the lowest level of suspicion), followed by calculation and cross-reference in the database of signatures. The application would be recorded if the result was positive. However, it would be identified as malware if the result of the suspicion was R. The system examined the files inside the application to find any matched MD5 signature.
Androsimilar (Faruki et al., 2013) practised a method known as a statistical similarity digest hashing scheme, which inspects the similarity on the byte stream based on robust statistical malicious static features. It is also a foot-printing method, which identifies the regions or areas of statistical similarity with known malware. Following that, it generates variable-length signatures to detect unknown malware (zero-day).
The following study is DroidMOSS (Zhou et al., 2012), which identifies between the repackaged (modified) and original application. This function is important due to the content of malicious activities in many Android repackaged applications. This study used a fuzzy hashing technique, which generated fingerprint based on this technique to localise and detect any previously applied modifications to the original application. It then calculated the edited distance to measure the similarity between the applications. When the result of the similarity exceeds certain values, the application would be considered as a modified sample.
Under another static experiment, a study by Apvrille & Strazzere (2012) adopted a method known as a risk score weight, which was performed through the calculation of the risk score based on the selected features in the code. When the features were identified, the score increased according to certain risky patterns of properties. Particularly, the patterns were based on different likelihoods of the given situations between normal and malware samples. Lastly, the percentage of the likelihood was calculated. Figure 9 shows that both ML and graph were the popular methods among security practitioners in static analysis. The graph method was found to exceed the ML method in 2011, 2012, and 2014, although ML was more preferred compared to graph in other years. However, this situation reveals that graphs and ML are favourable options in the static experiment.
Figure 9: Popular methods among security practitioners in static analysis. Both ML and graph were the popular methods among security practitioners in static analysis. (A) ML was more preferred compared to graph. (B) The graph method was found to exceed the ML method in 2011, 2012, and 2014. DOI: 10.7717/peerjcs.522/fig-9
A study started to utilise DL (part of ML) in the static experiment in 2019, which also combined DL (Convolutional neural network—CNN) with Control flow graph (CFG). Notably, provided that API was the only feature utilised in this study, many future opportunities were available to combine different DL classifiers (Recurrent neural network—RNN, Generative* adversarial networks—GAN or Deep belief network*—DBN) with other features besides API and different types of graph. It is noteworthy that DL could also be combined with NCD and MDM.
Open research issues
This section explains the issues involved in the static analysis, which were derived from previous research articles. Specifically, a discussion is made on the advantages and disadvantages of the open-source operating system, which rely on the availability of the kernel application code. Another issue on static analysis is code obfuscation used by the malware developer to increase the difficulty to detect the malware and research done to counter obfuscation. Then, we review overall static analysis articles, how to detect unknown malware, the combination of static and dynamic, resource consumption, future work, and features.
Open source advantages and disadvantages
Provided that Android malware is an open-source operating system, there is a continuous increase in its amount. To illustrate, one of the open-source availabilities is the kernel application code. Accordingly, Samsung officially provides its kernel operating system to the public (Samsung, 2019) for kernel enhancement or private purposes. Furthermore, any person may download the link according to the mobile device version. The code is also available in Lenovo (Lenovo, 2021), LG (Official, 2019c), Sony (Sony, 2019e), Htc (HTC, 2019b), Asus (ASUS, 2019a), Motorola (Official, 2019d), and other mobile providers. Consequently, this code availability may allow malware writers to identify and manage the vulnerabilities of the kernel’s operating system.
Even though the availability of open-source contributes to its easy access, it is also available for security practitioners to research it. These practitioners may be the researchers of universities, the staff of the mobile providers, Android Google researchers, freelance programmers, and the Android community. They also invent various frameworks, algorithms, and suggestions to improve the security of the operating system. The version of the kernel is updated every year, while the mobile providers are informed to regarding the kernel updates. These studies, including static analysis, would increase the confidence of Android users worldwide.
Obfuscation
Static analysis involves reverse engineering, such as decompile and disassemble, while malware developer utilises the obfuscation method to increase the difficulty of the decompiling process and lead to confusion in it. Obfuscation is a technique, which increases the difficulty in understanding the programmes due to the failure of the lead security analysts to distinguish between malware and benign application. Notably, it is a well-known obstacle to be examined by static analysis. Figure 10 illustrates the types of obfuscation, which include encryption, oligomorphic, polymorphism, and metamorphism (Moser, Kruegel & Kirda, 2007; You & Yim, 2010).
Figure 10: Types of obfuscation. Types of obfuscation identified include encryption, oligomorphic, polymorphism, and metamorphism. DOI: 10.7717/peerjcs.522/fig-10
The encryption method was extensively practised by the malware writers. In this case, the important code or strings, which revealed the malware detector or security practitioner, should be identified. Accordingly, the code was encrypted and converted to the ciphertext. Furthermtore, various algorithms to encrypt the code are present, such as Caesar, Playfair, Data Encryption Standard (DES), Advanced Encryption Standard (AES), and Rivest-Shamir-Adelman (RSA). Therefore, for the security practitioner to understand the behaviour of the malware, the encrypted code should be decrypted using the correct decryptor (Wei et al., 2017).
Besides being a malware capable of mutating @ changing the decryptor, the oligomorphic is also able to generate multiple decryptors to hundreds of types (You & Yim, 2010). Consequently, the security practitioner would need to change different decryptor multiple times until the code is returned to the normal string. Nevertheless, this type of obfuscation does not affect the size or shape of the code. Another type of obfuscation is polymorphic. It is a descriptor, which affects the size or shape of the code. Compared to oligomorphic, it is more advanced due to the incorporation of code transposition, register reassignment, dead code @ nop insertion, and armoring. Meanwhile, metamorphism is an approach beyond the oligomorphic and polymorphic types due to the absence of decryptor in its mechanism. Therefore, its constant body could be hidden from memory and increase the difficulty of the static investigation to detect malware.
The following information is the obfuscation methods that regularly used by polymorphism (polimorphic) and metamorphism (metamorphic) obfuscation (You & Yim, 2010).
(a) Code transportation
Code transposition is a method, which restructures the orders of the original code without causing any effects on its conduct. This process is performed with two methods. The first method is the random restructure of the original code by including jumps or unconditional branches. However, security practitioners can detect obfuscation by removing those jumps or unconditional branches. The second method is the production of new generations by selecting and restructuring independent instructions without any impact on others. However, the adoption of these methods is challenging for the malware writer, while the security practitioners are faced with a difficulty to detect this method of obfuscation.
(b) Register reassignment
Register reassignment is another method of obfuscation, which shifts the registers of the code from one generation to another. This method is performed without changing the behaviour of the code while keeping the programme of the code similar to its original state.
(c) Dead-code/nop insertion
Known as nop insertion, dead-code is a method, which adds several unnecessary instructions in the code and simultaneously keeps the behaviour of the code similar to its original state. Nevertheless, in certain situations, security practitioners able to detect this obfuscation by removing the aforementioned code.
(d) Substitution of equivalent instruction sequences
The original code is changed through the substitution of several instructions. To illustrate, the SUB instruction is changed to XOR, while PUSH is changed to MOV.
Research to counter obfuscation.
To overcome obfuscation, many studies were conducted on different approaches. Study by Crussell, Gibler & Chen (2012) used program dependence graph (PDG) to prevent program transformations in obfuscation. Droidlegacy (Deshotels, Notani & Lakhotia, 2014b) use graph node to represents the java class in detecting light obfuscation. Droidanalytics (Zheng, Sun & Lui, 2013) and Drebin (Arp et al., 2014) extract the API calls while the codes running during execution time. In order to control the flow of obfuscation, Apposcopy use inter-component communication (ICC) to write the signature. Research by Firdaus (2017) uses jadx, one of reverse engineer tool to de-obfuscation the obfuscation codes. Summary of studies conducted to overcome obfuscation shown in Table 10.
References | Year | Solution for the obfuscation |
---|---|---|
DNADroid (Crussell, Gibler & Chen, 2012) | 2012 | Using programme dependence graphs (PDGs), DNADroid can prevent typical program transformations in obfuscation. |
Apvrille & Strazzere (2012) | 2012 | Detects encryption attempts as one of the obfuscation methods |
DroidAPIMiner (Aafer, Du & Yin, 2013) | 2013 | Includes features, which are regularly used for obfuscation, such as substring (), indexOf(), getBytes(), valueOf(), replaceAll(), Append(), getInstance(), doFinal(), and Crypto.spec.DESKeySpec |
Androsimilar (Faruki et al., 2013) | 2013 | Adopts statistically strong file features based on a normalised entropy |
Droidanalytics (Zheng, Sun & Lui, 2013) | 2013 | Only extracts the API calls in methods and classes, which will be executed in the run time. Additionally, the generated signature, which is based on the analyst-defined API, which have the ability to update flexibly. |
Apposcopy (Feng et al., 2014) | 2014 | Includes the predicate inter-component communication (ICC), which allows the writing of signatures, which are resilient to high-level control flow obfuscation. |
Drebin (Arp et al., 2014) | 2014 | DREBIN extracts API calls related to obfuscation and loading of code, such as DexClassLoader.loadClass() and Cipher.getInstance |
Dendroid (Suarez-Tangil et al., 2014) | 2014 | Concentrates on the internal structure of code units (methods) to resist obfuscation. |
Droidlegacy (Deshotels, Notani & Lakhotia, 2014b) | 2014 | Graph node represents the Java class, which detects light obfuscation. |
Firdaus (2017) | 2017 | Uses Jadx (a reverse engineering tool), which provides the de-obfuscation option. It is capable of de-obfuscating the obfuscation code in minimal error. |
DOI: 10.7717/peerjcs.522/table-10
Advantage of obfuscation.
Despite the adoption of the obfuscation method by the malware writers or the attackers to evade detection, obfuscation also serves the following advantages based on other points of views:
(a) Reduction of the size of the application
Google (Android, 2019b) encourages developers to enable shrinking in their release to build an application to remove any unused codes and resources. Furthermore, provided that obfuscation would shorten the names of the classes and members in the code, the developer will be able to reduce the size of the application. Notably, the size of the application is a significant concern in Android handheld devices (smartphones, smart glasses, and smartwatch) with limited storage and resources.
(b) The difficulty for the malware writer to understand the obfuscated normal application
To develop malware in certain situations, malware writers need to perform reverse engineering on the normal repackaged application. Therefore it is able to confuse them to steal private information and discover application vulnerabilities from that obfuscated normal @ benign application code (Diego et al., 2004).
(c) Security practitioners can detect malware easily
Obfuscation also facilitates the detection of malware by the researcher (Nissim et al., 2014). To illustrate, there are certain situations where malware regularly adopts similar obfuscation marks, which is impossible to exist in normal application. Therefore, security practitioners able to detect malware with the presence of these marks. Following all the advantages and drawbacks, continuous research on obfuscation is crucial to obtain better results from the detection of malware through the static analysis.
Year | Ref | Features (•) | Classification ( ) | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
AD | API | A | DP | C | F | G | M | N | CB | ML | DL | Graph | Other | ||
2019 | Hammad, Bagheri & Malek (2019) | • | Multiple-Domain Matrix (MDM) | ||||||||||||
2019 | Fan et al. (2019) | • | |||||||||||||
2019 | Badhani & Muttoo (2019) | • | • | ||||||||||||
2019 | Wu et al. (2019) | • | |||||||||||||
2019 | Liu et al. (2019) | • | • | ||||||||||||
2019 | Alotaibi (2019) | • | • | • | |||||||||||
2019 | Zhang, Tian & Duan (2019) | • | |||||||||||||
2019 | Alsoghyer & Almomani (2019) | • | |||||||||||||
2019 | Yildiz & Doǧru (2019) | • | • | ||||||||||||
2019 | Singh, Jaidhar & Kumara (2019) | • | • | ||||||||||||
2019 | Lei et al. (2019) | • | |||||||||||||
2018 | Firdaus et al. (2018) | • | • | • | |||||||||||
2018 | Zhao et al. (2019) | • | |||||||||||||
2018 | Gu et al. (2018) | • | |||||||||||||
2018 | Firdaus et al. (2017) | • | • | • | |||||||||||
2018 | Firdaus et al. (2018) | • | • | • | • | ||||||||||
2018 | Klieber et al. (2018) | • | |||||||||||||
2018 | Kumar, Kuppusamy & Aghila (2018) | • | |||||||||||||
2018 | Wang et al. (2018) | • | • | • | • | ||||||||||
2018 | Ma Zhao-hui et al. (2017) | • | |||||||||||||
2017 | Feizollah et al. (2017) | • | |||||||||||||
2017 | Kim et al. (2018) | • | |||||||||||||
2017 | Zhou et al. (2017) | • | |||||||||||||
2017 | Pooryousef & Amini (2017) | • | |||||||||||||
2017 | Wu et al. (2018) | • | |||||||||||||
2017 | Chang & De Wang (2017) | • | • | ||||||||||||
2016 | Junaid, Liu & Kung (2016a) | • | |||||||||||||
2016 | Atici, Sagiroglu & Dogru (2016) | • | |||||||||||||
2016 | Medvet & Mercaldo (2016) | • | |||||||||||||
2016 | Wu et al. (2016) | • | |||||||||||||
2016 | Nissim et al. (2016) | • | |||||||||||||
2015 | Sheen, Anitha & Natarajan (2015) | • | • | ||||||||||||
2015 | Kang et al. (2015) | • | • | • | |||||||||||
2015 | Kang et al. (2015) | • | • | • | |||||||||||
2015 | Firdaus & Anuar (2015) | • | • | • | |||||||||||
2015 | Talha, Alper & Aydin (2015) | • | |||||||||||||
2015 | Junaid, Liu & Kung (2016a) | • | • | ||||||||||||
2015 | Lee, Lee & Lee (2015) | • | |||||||||||||
2015 | Elish et al. (2015) | • | • | ||||||||||||
2015 | Gordon et al. (2015) | • | |||||||||||||
2014 | Yerima, Sezer & McWilliams (2014) | • | • | ||||||||||||
2014 | Arp et al. (2014) | • | • | • | |||||||||||
2014 | Deshotels, Notani & Lakhotia (2014b) | • | |||||||||||||
2014 | Feng et al. (2014) | • | |||||||||||||
2014 | Huang et al. (2014a) | • | • | ||||||||||||
2014 | Steven Arzt et al. (2014) | • | |||||||||||||
2014 | Yerima, Sezer & Muttik (2014) | • | • | • | |||||||||||
2014 | Seo et al. (2014) | • | • | MD5 signature | |||||||||||
2014 | Suarez-Tangil et al. (2014) | • | |||||||||||||
2013 | Aafer, Du & Yin (2013) | • | |||||||||||||
2013 | Lee & Jin (2013) | • | |||||||||||||
2013 | Shuying Liang et al. (2013) | • | • | ||||||||||||
2013 | Zhou et al. (2013) | • | • | ||||||||||||
2013 | Luoshi et al. (2013) | • | • | • | |||||||||||
2013 | Peiravian & Zhu (2013) | • | • | ||||||||||||
2013 | Samra, Kangbin & Ghanem (2013) | • | • | ||||||||||||
2013 | Gascon et al. (2013) | • | |||||||||||||
2013 | Huang, Tsai & Hsu (2012) | • | |||||||||||||
2013 | Aung & Zaw (2013) | • | |||||||||||||
2013 | Faruki et al. (2013) | • | Similarity digest hashing | ||||||||||||
2013 | Paturi et al. (2013) | • | Normalised Compression Distance (NCD) | ||||||||||||
2013 | Apvrille & Apvrille (2013) | • | |||||||||||||
2013 | Yerima et al. (2013) | • | • | ||||||||||||
2013 | Borja Sanz et al. (2013) | • | |||||||||||||
2012 | Grace et al. (2012a) | • | |||||||||||||
2012 | Wu et al. (2012) | • | • | ||||||||||||
2012 | Bartel et al. (2012) | • | • | ||||||||||||
2012 | Grace et al. (2012b) | • | • | ||||||||||||
2012 | Zhou et al. (2012) | • | Fuzzy hashing technique | ||||||||||||
2012 | Hao Peng et al. (2012) | • | |||||||||||||
2012 | Walenstein, Deshotels & Lakhotia (2012) | • | |||||||||||||
2012 | Sahs & Khan (2012) | • | |||||||||||||
2012 | Sanz et al. (2013) | • | |||||||||||||
2012 | Apvrille & Strazzere (2012) | • | • | • | Risk score weight | ||||||||||
2012 | Lu et al. (2012) | • | |||||||||||||
2012 | Crussell, Gibler & Chen (2012) | • | |||||||||||||
2012 | Desnos (2012b) | • | Normalised Compression Distance (NCD) | ||||||||||||
2012 | Sarma et al. (2012) | • | |||||||||||||
2012 | Yang & Yang (2012) | • | • | ||||||||||||
2011 | Grace et al. (2011) | • | • | ||||||||||||
2010 | Shabtai, Fledel & Elovici (2010) | • | • | • | |||||||||||
2009 | Aubrey-Derrick Schmidt et al. (2009a) | • |
DOI: 10.7717/peerjcs.522/table-11
Notes:
Legends:
AD
Advertisement libraries
API
API
A
apk, dex and XML properties
DP
Directory path
C
Commands
F
unction call
G
Geographic
M
Manifest file
N
Network address or URLs
CB
Code-based
The list of all articles in the detection of malware in static analysis
To identify the trends in the detection of malware through static analysis, this section presents a list of previous works of research, which cover all areas (year, features, and classification). Table 11 lists a study DroidARA (Fan et al., 2019) in 2019, which performed an experiment combined with DL and graph and differentiation between malware and normal application. It applied a call graph to extract the API features and convolutional neural network (CNN) for classification. At the time of writing this paper, this is a new trend in detecting Android malware. Therefore, in future research, it is possible to witness more research combination similar to this with different features.
From the lists, most of researchers used API and manifest file features in their experiments to detect the malware. It proofs that API features were the popular codes used by the malware developers to create the malware. The program is based on the Android application package file in the (.apk) format, which is also used to install an application in android-based mobile devices. Every app project must have an androidmanifest.xml file at the root of the project source set. This manifest file is regularly in a binary form inside the APK, however when chosen in the APK Analyzer, the xml form is restructured and produced. There are some changes of the androidmanifest.xml document from a library application depends on was converged into the last androidmanifest.xml record. Other package files fall down into apk, xml and dex properties feature.
Besides the combination of DL and graph, ML and graph were also combined in the studies by Atici, Sagiroglu & Dogru (2016) in 2016 and Sahs & Khan (2012) in 2012. These studies utilised a similar graph, which was the Control flow graph (CFG), indicating that the combination of ML and graph increased the detection result. Therefore, future work is suggested to test this combination in different static features. Other parts of classification (Multiple-Domain Matrix (MDM), MD5 signature, similarity digest hashing, normalized compression distance (NCD), and fuzzy hashing technique) were also useful in the detection of malware with static features. These classifications also contributed to the availability of future work combinations with ML, DL, and graph.
Detect unknown malware
Initially, static analysis is unable to discover new or unknown malware as it only examined the code of the application without executing it. To elaborate on this situation, certain malware only executes certain parts whenever the application runs. Provided the drawback of static analysis in the identification of unknown malware, many security practitioners started to adopt machine learning, such as (Lee, Lee & Lee, 2015; Yerima, Sezer & McWilliams, 2014), Drebin (Arp et al., 2014; Yerima, Sezer & Muttik, 2014), Droidapiminer (Aafer, Du & Yin, 2013; Apvrille & Apvrille, 2013), Androsimilar (Faruki et al., 2013; Lee & Jin, 2013; Yerima et al., 2013; Paturi et al., 2013; Shabtai, Fledel & Elovici, 2010), and (Firdaus & Anuar, 2015). Similarly, the graph approach was also a suitable approach for this identification, as shown in Elish et al. (2015), Riskranker (Grace et al., 2011), and Dendroid (Suarez-Tangil et al., 2014). The Elish et al. (2015) study utilised a data dependence graph (DDG), Riskranker, and Dendroid, which employed the control-flow graph (CFG).
Combination of static and dynamic analyses
It was proven in Moser, Kruegel & Kirda (2007) that static analysis was inadequate for the detection of malware as this analysis should be combined with dynamic analysis to detect the malware effectively. Compared to static analysis, the dynamic analysis can evade the obfuscation technique. Essentially, provided that each type of analysis (static and dynamic) has its advantages and drawbacks, the combination of static and dynamic analyses would increase the effectiveness of the countermeasure action on the malware.
Resource consumption in Android OS and deep learning
Deep learning (DL) is a subset of machine learning in artificial intelligence (AI), which is also known as a deep neural network or deep neural learning. Notably, with unlabeled and unstructured data, DL is capable of learning and predicting the output. It imitates the human brain in data processing, development of patterns from that data, and the implementation of decision making. It could be seen from the current trends that deep learning (DL) technique has a potential for further discovery. The implementation of this technique enables the DL to automatically determine the ideal features for prediction and classification. Currently, DL is widely used in almost every area, such as large scale image recognition tasks, automated driving, new cancer cell detection, hearing and speech translation, and aerospace area identification (Mathworks, 2020).
However, DL requires substantial computing power, which needs a high amount of the graphic processing unit (GPU) based on the data to be processed (Mathworks, 2020). This situation leads to an issue in the detection of malware, which attacks Android devices. Provided that Android mobile device is a device with small computing power, the adoption of DL becomes the main concern. However, the transfer of information from the Android device to the cloud provider is possible only for the execution of the DL process, which would then develop the device. Therefore, the large-scale adoption of DL is possible for future work in the static analysis.
Future work in static analysis
It could be seen from the review in the previous sections (‘Survey Methodology’) that many future opportunities for the static analysis to detect the malware, which attacks the Android. One of the opportunities is the combination of different DL classifier (Recurrent neural network—RNN, Generative* adversarial networks—GAN or Deep belief network*—DBN) with other features besides API, with different types of graph. However, ‘The list of all articles in the detection of malware in static analysis’ shows that only one experiment started the combination between DL and graph with one API feature in 2019. Therefore, the accuracy of detection results would increase, leading to the identification of a new family of malware. It is also noteworthy that other future alternatives are available for the combination of DL with NCD and MDM.
Popular features (API and manifest file)
‘The list of all articles in the detection of malware in static analysis’ shows that many static analysis researchers frequently applied the manifest file and API calls as the features in their experiments. To illustrate, these popular features had been examined by the researchers from 2010 until 2019 due to the official update for Android and the addition of new features from time to time. However, most of the malwares still utilised similar features within this timespan, while other malwares utilised the latest and updated features. Therefore, the researchers are required to place continuous focus on these popular features.
The manifest file is one file with numerous features in it, such as permission, intent, hardware component, and application component, while API is a ready code for the programmer to develop their application. Therefore, it is crucial for researchers to thoroughly scrutinise these two categories of features. Moreover, some researchers incorporated other features to combine the manifest and API for more effective malware detection.
Conclusions
Following the interest to explore the recent studies in the static analysis, a review was performed on the existing studies by past security investigators on Android malware detection, which was explained through phases (reverse engineer, features, and classification). Furthermore, this review covered the information within the ten years range (2009 to 2019). In this article, the features used in the static analysis were also reviewed. Within the aforementioned timespan, many security practitioners still preferred the API and manifest files, indicating the relevance of these two features. Moreover, the latest trends in classification were highlighted, which consists of machine learning, deep learning, graph, and other methods. These trends have proven the relevance of the graph method compared to machine learning. Static analysis researchers began the adoption of deep learning in their detection. This article also discussed the open research issues in the static analysis, including obfuscation as one of the weaknesses of static analysis and the methods of overcoming it. Many static analysis researchers implemented multiple methods to solve obfuscation and achieve a successful malware detection, such as concentrating the codes with the implementation of obfuscation (DexClassLoader.loadClass, Crypto.spec.DESKeySpec, and Cipher.getInstance) using a tool with the de-obfuscation option, including the adoption of graph node and program dependence graphs (PDGs).
Additional Information and Declarations
Competing Interests
The authors declare there are no competing interests.
Author Contributions
Rosmalissa Jusoh conceived and designed the experiments, performed the computation work, prepared figures and/or tables, and approved the final draft.
Ahmad Firdaus conceived and designed the experiments, performed the computation work, authored or reviewed drafts of the paper, and approved the final draft.
Shahid Anwar and Mohd Zamri Osman performed the experiments, authored or reviewed drafts of the paper, and approved the final draft.
Mohd Faaizie Darmawan analyzed the data, prepared figures and/or tables, and approved the final draft.
Mohd Faizal Ab Razak analyzed the data, prepared figures and/or tables, and approved the final draft.
Data Availability
The following information was supplied regarding data availability:
This is a review article.
Funding
This work was supported by the Ministry of Higher Education (MOHE) for Fundamental Research Grant Scheme (FRGS) with grant number RDU190190, FRGS/1/2018/ICT02/UMP/02/13, and Universiti Malaysia Pahang (UMP) internal grant with grant number RDU1803142. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Aafer Y, Du W, Yin H. 2013. DroidAPIMiner: mining API-Level Features for Robust Malware Detection in Android. In: Security and privacy in communication networks. Sydney, NSW, Australia. SecureComm. 86-103
Abadi M, Agarwal A, Barham P. 2016. TensorFlow: large-scale machine learning on heterogeneous distributed systems. http://arxiv.org/abs/1603.04467
Aktas K, Sen S. 2018. UpDroid: updated Android Malware and its familial classification. In: Gruschka N, ed. Secure IT Systems. NordSec 2018. Lecture Notes in Computer Science. Cham: Springer. Vol. 11252
Al-Rfou R, Alain G, Almahairi A, Angermueller C, Bahdanau D. 2016. Theano: A Python framework for fast computation of mathematical expressions. arXiv. . http://arxiv.org/abs/1605.02688
Allahham AA, Rahman MA. 2018. A smart monitoring system for campus using zigbee wireless sensor networks. International Journal of Software Engineering & Computer System 4(1):1-14
Allix K, Bissyandé TF, Klein J, Le Traon Y. 2018. AndroZoo: collecting millions of android apps for the research community. In: MSR ’16 proceedings of the 13th international conference on mining software repositories, Austin, Texas. 468-471
Alotaibi A. 2019. Identifying malicious software using deep residual long-short term memory. IEEE Access 7:163128-163137
Alsoghyer S, Almomani I. 2019. Ransomware detection system for android applications. Electron 8(8):1-36
Alswaina F, Elleithy K. 2018. Android malware permission-based multi-class classification using extremely randomized trees. IEEE Access 6:76217-76227
Alswaina F, Elleithy K. 2020. Android malware family classification and analysis: current status and future directions. Electron 9(6):1-20
Amit Y, Geman D. 1997. Shape quantization and recognition with randomized trees. Neural Computation 9(7):1545-1588
Anderson HS, Filar B, Roth P. 2017. Evading Machine Learning Malware Detection. In: BlackHat DC. 6 . https://www.blackhat.com/docs/us-17/thursday/us-17-Anderson-Bot-Vs-Bot-Evading-Machine-Learning-Malware-Detection-wp.pdf
Anderson HS, Kharkar A, Filar B, Evans D, Roth P. 2018. Learning to Evade Static PE Machine Learning Malware Models via Reinforcement Learning. ArXiv. . http://arxiv.org/abs/1801.08917
Android Developers. 2017. Android Debug Bridge (ADB) Google Developers.
Android. 2013. aapt.
Android. 2019a. What is Android. (accessed30 September 2019 )
Android. 2019b. ProGuard. (accessed02 October 2019 )
Android Developer. 2020. Opcodes. Google Developers.
Android Developer. 2018. Dedexer. (accessed02 October 2019 )
Android. 2015. App Manifest. (accessed02 October 2019 )
2019b. Baksmali & Smali. (accessed02 October 2019 )
Apvrille A, Strazzere T. 2012. Reducing the window of opportunity for Android malware Gotta catch ’em all. Journal of Computer Virology 8(1–2):61-71
Apvrille L, Apvrille A. 2013. Pre-filtering Mobile Malware with Heuristic Techniques. In: The 2nd international symposium on research in grey-hat hacking (GreHack). Grenoble, France. 1-16
Aresu M, Ariu D, Ahmadi M, Maiorca D, Giacinto G. 2016. Clustering android malware families by http traffic. In: 2015 10th Int. conf. malicious unwanted software, MALWARE 2015. 128-135
Arp D, Spreitzenbarth M, Malte H, Gascon H, Rieck K. 2014. DREBIN: effective and explainable detection of android malware in your pocket. In: 21th annual network and distributed system security symposium (NDSS). San Diego, CA. 1-15
ASUS. 2019a. Asus. (accessed02 October 2019 )
Atici MA, Sagiroglu S, Dogru IA. 2016. Android malware analysis approach based on control flow graphs and machine learning algorithms. In: IEEE 4th international symposium on digital iforensics and security, little rock, Arkansas. Piscataway. IEEE. 26-31
Atzeni A, Diaz F, Marcelli A, Sanchez A, Squillero G, Tonda A. 2018. Countering android malware: A scalable semi-supervised approach for family-signature generation. IEEE Access 6:59540-59556
Aubrey-Derrick Schmidt OK, Bye R, Schmidt H-G, Clausen J, Kamer SA, Yuksel A, Camtepe SA. 2009a. Static Analysis of Executables for Collaborative Malware Detection on Android. In: IEEE international conference on communications (ICC). Dresden, Germany. Piscataway. IEEE. 1-5
Aubrey-Derrick Schmidt SA, Schmidt H-G, Batyuk L, Clausen J, Camtepe SA. 2009b. Smartphone malware evolution revisited: android next target? In: 2009, 4th int. conf. malicious unwanted software, MALWARE 2009. 1-7
Aung Z, Zaw W. 2013. Permission-based android malware detection. International Journal of Scientific and Technology Research 2(3):228-234
Badhani S, Muttoo SK. 2019. CENDroid—A cluster-ensemble classifier for detecting malicious Android applications. Computers & Security 85:25-40
Bartel A, Klein J, Le Traon Y, Monperrus M. 2012. Automatically securing permission-based software by reducing the attack surface: an application to Android. In: Proceedings of the 27th IEEE/ACM international conference on automated software engineering (ASE). Essen, Germany. Piscataway. IEEE. 274-277
Battista P, Mercaldo F, Nardone V, Santone A, Visaggio CA. 2016. Identification of android malware families with model checking. In: Proceedings of the 2nd international conference on information systems security and privacy (ICISSP). 542-547
Beal V. 2013. Mobile malware.
Beverly R, Garfinkel S, Cardwell G. 2011. Forensic carving of network packets and associated data structures. Digital Investigation 8:S78-S89
Blanc W, Hashem LG, Elish KO, Hussain Almohri MJ. 2019. Identifying Android malware families using Android-oriented metrics. In: 2019 IEEE International Conference on Big Data (Big Data). Piscataway. IEEE. 4708-4713
Borja Sanz GÁ, Santos I, Laorden C, Ugarte-Pedrero X, Nieves J, Bringas PG. 2013. MAMA : manifest analysis for malware detection in android. Cybernetics and Systems
Calleja A, Martín A, Menéndez HD, Tapiador J, Clark D. 2018. Picking on the family: disrupting android malware triage by forcing misclassification. Expert Systems with Applications 95:113-126
Castillo CA. 2011. Android malware past, present, and future. McAfee white paper, Mobile Security Working Group.
Chakraborty T, Pierazzi F, Subrahmanian VS. 2020. EC2: ensemble clustering and classification for predicting android malware families. IEEE Transactions on Dependable and Secure Computing 17(2):262-277
Chanda M, Biswas M. 2019. Plant diseases classification using feature reduction, BPNN and PSO. International Journal of Software Engineering & Computer System 5(2):90-103
Chang YC, De Wang S. 2017. The concept of attack scenarios and its applications in android malware detection. In: 2016 IEEE 18th International Conference on High Performance Computing and Communications; IEEE 14th International Conference on Smart City; IEEE 2nd International Conference on Data Science and Systems (HPCC/SmartCity/DSS). Piscataway. IEEE. 1485-1492
Chang T-K, Hwang G-H. 2007. The design and implementation of an application program interface for securing XML documents. Journal of Systems and Software 80(8):1362-1374
Chess B, McGraw G. 2004. Static analysis for security. IEEE Security & Privacy Magazine 2(6):76-79
Crussell J, Gibler C, Chen H. 2012. Attack of the clones: detecting cloned applications on Android markets. In: Foresti S, Yung M, Martinelli F, eds. Computer Security – ESORICS 2012. ESORICS 2012. Lecture Notes in Computer Science. Berlin, Heidelberg: Springer. Vol. 7459
Deshotels L, Notani V, Lakhotia A. 2014a. DroidLegacy. 1-12
Deshotels L, Notani V, Lakhotia A. 2014b. DroidLegacy: automated Familial Classification of Android Malware. In: Proceedings of ACM SIGPLAN on program protection and reverse engineering workshop. New York. ACm.
Desnos A. 2012a. Androguard. GitHub.
Desnos A. 2012b. Android: static analysis using similarity distance. In: 2012 45th Hawaii international conference on system sciences. 5394-5403
Dhaya R, Poongodi M. 2015. Detecting software vulnerabilities in android using static analysis. In: 2014 IEEE International Conference on Advanced Communications, Control and Computing Technologies. Piscataway. IEEE. 915-918
Diego S, Kruegel C, Robertson W, Valeur F, Vigna G. 2004. Static disassembly of obfuscated binaries. In: Proceedings of the 13th USENIX Security Symposium. United States: USENIX Security Symposium 2004.
Elish KO, Shu X, Yao D, Ryder BG, Jiang X. 2015. Profiling user-trigger dependence for Android malware detection. Computers & Security 49(540):255-273
Enck W. 2011. Defending Users against Smartphone Apps: techniques and Future Directions. In: Proceedings of the 7th international conference on Information Systems Security, Kolkata, India. 49-70
Fan W, Chen Y, Liu Y, Wu F. 2019. DroidARA: android Application Automatic Categorization Based on API Relationship Analysis. IEEE Access 7(X):157987-157996
Fang Y, Gao Y, Jing FAN, Zhang LEI. 2020. Android balware familial classification based on dex file section features. In: IEEE Access. vol. 8:10614-10627
Fang Z, Han W, Li Y. 2014. Permission based Android security: issues and countermeasures. Computers & Security 43(0):205-218
Faruki P, Ganmoor V, Laxmi V, Gaur MS, Bharmal A. 2013. AndroSimilar: robust Statistical Feature Signature for Android Malware Detection. In: Proceedings of the 6th international conference on security of information and networks. Aksaray, Turkey. 152-159
Fasano F, Martinelli F, Mercaldo F, Santone A. 2019. Cascade learning for mobile malware families detection through quality and android metrics. In: 2019 International Joint Conference on Neural Networks (IJCNN). 1-10
Fazeen M, Dantu R. 2014. Another free app: does it have the right intentions? In: 2014 Twelfth Annual International Conference on Privacy, Security and Trust. Toronto, ON, Canada. 282-289
Feizollah A, Anuar NB, Salleh R, Amalina F, Ma’arof RR, Shamshirband S. 2013. A study of machine learning classifiers for anomaly-based mobile botnet detection. Malaysian Journal of Computer Science 26(4):251-265
Feizollah A, Anuar NB, Salleh R, Suarez-Tangil G, Furnell S. 2017. AndroDialysis: analysis of Android Intent Effectiveness in Malware Detection. Computers & Security 65:121-134
Feizollah A, Anuar NB, Salleh R, Wahab AWA. 2015. A review on feature selection in mobile malware detection. Digital Investigation 13:22-37
Feldman S, Stadther D, Wang B. 2014. Manilyzer: automated Android Malware Detection through Manifest Analysis. In: IEEE 11th international conference on mobile ad hoc and sensor systems, Philadelphia, PA, USA. Piscataway. IEEE. 767-772
Feng Y, Anand S, Dillig I, Aiken A. 2014. Apposcopy: semantics-based detection of Android malware through static analysis. In: FSE 2014: Proceedings of the 22nd ACM SIGSOFT International Symposium on Foundations of Software Engineering. New York. ACM. 576-587
Feng Y, Bastani O, Martins R, Dillig I, Anand S. 2017. Automated synthesis of semantic malware signatures using maximum satisfiability. In: NDSS Symposium 2017, California.
Firdaus A. 2017. Mobile malware anomaly-based detection systems using static analysis features. PhD thesis, University of Malaya, Kuala Lumpur, Malaysia
Firdaus A, Anuar NB. 2015. Root-exploit Malware Detection using Static Analysis and Machine Learning. In: Proceedings of the fourth international conference on computer science & computational mathematics (ICCSCM 2015). Langkawi, Malaysia. 177-183
Firdaus A, Anuar NB, Karim A, Razak MFA. 2018. Discovering optimal features using static analysis and a genetic search based method for Android malware detection. Frontiers of Information Technology & Electronic Engineering 19(6):712-736
Firdaus A, Anuar NB, Razak MFA, Hashem IAT, Bachok S, Sangaiah AK. 2018. Root exploit detection and features optimization: mobile device and blockchain based medical data management. Journal of Medical Systems 42(6)
Firdaus A, Anuar NB, Razak MFA, Sangaiah AK. 2017. Bio-inspired computational paradigm for feature investigation and malware detection: interactive analytics. Multimedia Tools and Applications 76(280):1-37
Garcia J, Hammad M, Malek S. 2018. Lightweight, obfuscation-Resilient detection and family identification of android malware. ACM Transactions on Software Engineering and Methodology 26(3):1-29
Gascon H, Yamaguchi F, Arp D, Rieck K. 2013. Structural detection of android malware using embedded call graphs. In: Proceedings of the 2013 ACM workshop on artificial intelligence and security. Berlin, Germany. New York. ACM. 45-54
Gibert D, Mateu C, Planes J. 2020. The rise of machine learning for detection and classification of malware: research developments, trends and challenges. J. Netw. Comput. Appl. 153(2019):102526
González A, Herrero Á, Corchado E. 2017. Neural visualization of android malware families. Advances in Intelligent Systems and Computing 527:576-583
Gordon MI, Kim D, Perkins J, Gilham L, Nguyen N, Rinard M. 2015. Information-Flow Analysis of Android Applications in DroidSafe. In: Network and distributed system security symposium (NDSS), San Diego, CA. 8-11
Grace MC, Zhou W, Jiang X, Sadeghi A-R. 2012a. Unsafe Exposure Analysis of Mobile In-App Advertisements. In: WISEC ‘12: Proceedings of the fifth ACM conference on Security and Privacy in Wireless and Mobile Networks. New York. ACM. 101-112
Grace M, Zhou Y, Wang Z, Jiang X. 2012b. Systematic Detection of Capability Leaks in Stock Android Smartphones. In: Proceedings of the 19th network and distributed system security symposium (NDSS), San Diego, CA. 1-15
Grace M, Zhou Y, Zhang Q, Zou S, Jiang X. 2011. RiskRanker: scalable and Accurate Zero-day Android Malware Detection. In: Proceedings of the 10th international conference on mobile systems, applications, and services. Low Wood Bay, Lake District, UK. 281-293
Gu J, Sun B, Du X, Wang J, Zhuang Y, Wang Z. 2018. Consortium blockchain-based malware detection in mobile devices. IEEE Access 6:12118-12128
Hammad M, Bagheri H, Malek S. 2019. DELDROID: an automated approach for determination and enforcement of least-privilege architecture in android. Journal of Systems and Software 149:83-100
Hao Peng CG, Bhaskar Sarma IM, Li N, Qi Y, Potharaju R, Nita-Rotaru C. 2012. Using Probabilistic Generative Models for Ranking Risks of Android Apps. In: ACM conference on computer and communications security, (CCS). Raleigh, North Carolina, USA. 241-252
Hastie T, Tibshirani R, Friedman J. 2009. Linear methods for classification.
HTC Malaysia. 2019b. HTC Malaysia. (accessed02 October 2019 )
Hsiao SW, Sun YS, Chen MC. 2016. Behavior grouping of Android malware family. In: 2016 IEEE International Conference on Communications (ICC). Piscataway. IEEE.
Huang C-Y, Tsai Y-T, Hsu C-H. 2012. Performance evaluation on permission-based detection for android malware. In: Proceedings of the international computer symposium ICS 2012, vol. 21. 111-120
Huang J, Zhang X, Tan L, Wang P, Liang B. 2014. AsDroid: detecting stealthy behaviors in android applications by user interface and program behavior contradiction. In: Proceeding ICSE 2014 Proceedings of the 36th International Conference on Software Engineering. Hyderabad, India. 1036-1046
Islamic I, Minna T. 2015. Android malware classification using static code analysis and apriori algorithm improved with particle swarm optimization. In: World information and communication technologies (WICT). Bandar Hilir, Malaysia. 123-128
Jkielty. 2019. Android v iOS market share. (accessed30 September 2019 )
Junaid M, Liu D, Kung D. 2016. Dexteroid: detecting malicious behaviors in android apps using reverse-engineered life cycle models. Computers & Security 59:92-117
Kang B, Kang BJ, Kim J, Im EG. 2013. Android malware classification method: dalvik bytecode frequency analysis. In: RACS ‘13: Proceedings of the 2013 Research in Adaptive and Convergent Systems. 349-350
Kang H, Jang J, Mohaisen A, Kim HK. 2015. Detecting and classifying android malware using static analysis along with creator information. International Journal of Distributed Sensor Networks 11(6):1-17
Kang BJ, Yerima SY, McLaughlin K, Sezer S. 2016. N-opcode analysis for android malware classification and categorization. In: 2016 International Conference On Cyber Security And Protection Of Digital Services (Cyber Security). 13-14
Kim H, Cho T, Ahn GJ, Hyun Yi J. 2018. Risk assessment of mobile applications based on machine learned malware dataset. Multimedia Tools and Applications 77(4):5027-5042
Kim HM, Seo JW, Song HM, Kim HK. 2019. Andro-Simnet: Android malware family classification using social network analysis.
Klaib MFJ, Sara MRA, Hasan M. 2020. D-GREEDY: greedy shortest superstring with delayed random choice. International Journal of Software Engineering & Computer System 6(1):8-17
Klieber W, Snavely W, Flynn L, Zheng M. 2018. Practical precise taint-flow static analysis for android app sets. In: ACM Int. Conf. Proceeding Ser.
Kotsiantis SB. 2007. Supervised machine learning: a review of classification techniques. Informatica 31:249-268
Kumar A, Kuppusamy KS, Aghila G. 2018. FAMOUS: forensic analysis of mobile devices using scoring of application permissions. Future Generation Computer Systems 83:158-172
Lee S-H, Jin S-H. 2013. Warning system for detecting malicious applications on android system. International Journal of Computer and Communication Engineering 2(3):324-327
Lee J, Lee S, Lee H. 2015. Screening smartphone applications using malware family signatures. Computers & Security 52:234-249
Lei T, Qin Z, Wang Z, Li Q, Ye D. 2019. EveDroid: event-aware Android Malware detection against model degrading for IoT devices. IEEE Internet of Things Journal 6(4):6668-6680
Lenovo. 2021. Lenovo official site.
Li Y, Shen T, Sun X, Pan X, Mao B. 2015. Detection, classification and characterization of android malware using API data dependency. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering 164:23-40
Li Y, Wang G, Nie L, Wang Q, Tan W. 2018. Distance metric optimization driven convolutional neural network for age invariant face recognition. Pattern Recognit. 75:51-62
Liu A, Guo C, Wang W, Qiu Y, Xu J. 2019. Static Back-Stack Transition Analysis for Android. IEEE Access 7:110781-110793
LG Electronics. 2019c. LG. (accessed02 October 2019 )
Lu L, Li Z, Wu Z, Lee W, Jiang G. 2012. CHEX: statically Vetting Android Apps for Component Hijacking Vulnerabilities. In: CCS Proceedings of the ACM Conference on Computer and Communications Security, Raleigh, North Carolina, USA. 229-240
Luoshi Z, Yan N, Xiao W, Wang Z, Xue Y. 2013. A3: automatic Analysis of Android Malware. In: International Workshop on Cloud Computing and Information Security (CCIS), Shanghai, China. 89-93
Ma Y, Sharbaf MS. 2013. Investigation of Static and Dynamic Android Anti-virus Strategies. In: 10th International Conference on Information Technology: new Generations (ITNG), Las Vegas, Nevada. 398-403
Ma Zhao-hui RX, Zi-hao C, Xin-ming W, Rui-hua N, Gan-sen Z, Jie-chao W. 2017. Shikra : a behavior-based Android malware detection framework.
Mahdavifar S, Ghorbani AA. 2019. Application of deep learning to cybersecurity: A survey. Neurocomputing 347:149-176
Malik S, Khatter K. 2016. System call analysis of Android Malware families. Indian Journal of Science and Technology 9(21):1-13
Martín A, Rodríguez-Fernández V, Camacho D. 2018. CANDYMAN: classifying Android malware families by modelling dynamic traces with Markov chains. Engineering Applications of Artificial Intelligence 74(June):121-133
Massarelli L, Aniello L, Ciccotelli C, Querzoni L, Ucci D, Baldoni R. 2017. Android malware family classification based on resource consumption over time (extended version)
Mathworks. 2020. What is deep learning?
McAfee. 2019. Mobile threat report.
Medvet E, Mercaldo F. 2016. Exploring the usage of topic modeling for android malware static analysis. In: 11th Int. Conf. Availability, Reliab. Secur. (ARES). Salzburg, Austria. 609-617
Ming Fan TL, Liu J, Luo X, Chen K, Tian Z, Zheng Q. 2018. Android malware familial classification and representative sample selection via frequent subgraph analysis. IEEE Transactions on Dependable and Secure Computing 13(8):1890-1905
MilaParkour. 2019. Contagio: Malware dump.
Mohd Azwan Hamza NO, Ab Aziz Mohd Juzaiddin. 2019. Sentence Similarity Measurement Based on Thematic Role and Semantic Network Techniques. International Journal of Software Engineering and Computer Systems (IJSECS) 301-316
Motorola. 2019d. Motorola homepage. (accessed02 October 2019 )
Moser A, Kruegel C, Kirda E. 2007. Limits of static analysis for malware detection. In: Twenty-Third Annual Computer Security Applications Conference (ACSAC 2007). 421-430
Narudin FA, Feizollah A, Anuar NB, Gani A. 2014. Evaluation of machine learning classifiers for mobile malware detection. Soft Computing 20(1):343-357
Nissim N, Moskovitch R, BarAd O, Rokach L, Elovici Y. 2016. ALDROID: efficient update of Android anti-virus software using designated active learning methods. Knowledge and Information Systems 49(3):795-833
Nissim N, Moskovitch R, Rokach L, Elovici Y. 2014. Novel active learning methods for enhanced PC malware detection in Windows OS. Expert Systems with Applications 41(13):5843-5857
O’Donnell L. 2019. Google Play malicious apps racked up 335M+ installs in September. (accessed2019 November 08 )
Omid Mirzaeiq GS, Suarez-Tangil G, de Fuentesq JM, Tapiadorq J. 2019. AndrEnsemble: leveraging API ensembles to characterize android malware families. In: Asia CCS ‘19: Proceedings of the 2019 ACM Asia Conference on Computer and Communications Security. New York. ACM. 307-314
Pan Y, Ge X, Fang C, Fan Y. 2020. A systematic literature review of android malware detection using static analysis. IEEE Access 8:116363-116379
Pascanu R, Stokes JW, Sanossian H, Marinescu M, Thomas A. 2015. Malware classification with recurrent networks. In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Piscataway. IEEE. 1916-1920
Paturi A, Cherukuri M, Donahue J, Mukkamala S. 2013. Mobile malware visual analytics and similarities of attack toolkits. In: Collaboration technologies and systems (CTS). San Diego, CA, USA. 149-154
Peiravian N, Zhu X. 2013. Machine learning for android malware detection using permission and API calls. In: International conference on tools with artificial intelligence (ICTAI). Herndon, VA, USA. 300-305
Pooryousef S, Amini M. 2017. Enhancing accuracy of android malware detection using intent instrumentation. In: Proceedings of the 3rd International Conference on Information Systems Security and Privacy - ICISSP. 380-388
Qiu J, Zhang JUN, Luo WEI, Pan LEI. 2019. A3CM : automatic Capability Annotation for Android Malware. IEEE Access 7:147156-147168
Rayner T. 2019. Android Power Rankings (2019): shake-up at the top, 2019. Android Authority. (accessed30 September 2019 )
Razak MFA, Anuar NB, Othman F, Firdaus A, Afifi F, Salleh R. 2018. Bio-inspired for features optimization and malware detection. Arabian Journal for Science and Engineering 43(12):6963-6979
Razak MFA, Anuar NB, Salleh R, Firdaus A, Faiz M, Alamri HS. 2019. ‘Less Give More’: evaluate and zoning Android applications. Measurement 133:396-411
Saba Arshad HY, Shah MA, Wahid A, Mehmood A, Song H. 2018. SAMADroid : a novel 3-level hybrid malware detection model for android operating system. IEEE Access 6:4321-4339
Sahs J, Khan L. 2012. A machine learning approach to android malware detection. In: 2012 European Intelligence and Security Informatics Conference. Piscataway. IEEE. 141-147
Samra AAA, Kangbin Y, Ghanem OA. 2013. Analysis of clustering technique in android malware detection. In: Seventh international conference on innovative mobile and internet services in ubiquitous computing (IMIS). Taichung, Taiwan. 729-733
Samra AAA, Qunoo HN, Al-Rubaie F, El-Talli H. 2019. A survey of Static Android Malware Detection Techniques. In: 2019 IEEE 7th Palestinian International Conference on Electrical and Computer Engineering (PICECE). Piscataway: IEEE.
Samsung. 2019. Mobile phone. (accessed02 October 2019 )
Sanz B, Santos I, Laorden C, Ugarte-Pedrero X, Bringas PG, Alvarez G. 2013. PUMA: permission usage to detect malware in Android. In: Advances in intelligent systems and computing. 289-298
Sarma B, Li N, Gates C, Potharaju R, Nita-rotaru C, Molloy I. 2012. Android Permissions: A Perspective Combining Risks and Benefits. In: SACMAT ’12 proceedings of the 17th ACM symposium on access control models and technologies. New York. ACM. 13-22
Saxe J, Berlin K. 2016. Deep neural network based malware detection using two dimensional binary program features. In: 2015 10th int. conf. malicious unwanted software, MALWARE 2015. 11-20
Saxe J, Labs I, Berlin K, Labs I. 2015. Deep Neural Network Based Malware Detection Using Two Dimensional Binary Program Features. In: 2015 10th int. conf. malicious unwanted softw. 11-20
Schmeelk S, Yang J, Aho A. 2015. Android Malware Static Analysis Techniques. In: CISR ’15 Proceedings of the 10th Annual Cyber and Information Security Research Conference. 1-8
Sedano J, Chira C, González S, Herrero Á, Corchado E, Villar JR. 2017a. Characterization of Android malware families by a reduced set of static features. In: Graña M, López-Guede J, Etxaniz O, Herrero Á, Quintián H, Corchado E, eds. International Joint Conference SOCO’16-CISIS’16-ICEUTE’16. SOCO 2016, CISIS 2016, ICEUTE 2016. Advances in Intelligent Systems and Computing. Cham: Springer. vol. 527
Sedano J, González S, Chira C, Herrero AL, Corchado E, Villar JR. 2017b. Key features for the characterization of Android malware families. Logic Journal of the IGPL 25(1):54-66
Seo S-H, Gupta A, Mohamed Sallam A, Bertino E, Yim K. 2014. Detecting mobile malware threats to homeland security through static analysis. Journal of Network and Computer Applications 38:43-53
Shabtai A, Fledel Y, Elovici Y. 2010. Automated Static Code Analysis for Classifying Android Applications Using Machine Learning. In: Ninth international conference on computational intelligence and security, Nanning, Guangxi Zhuang Autonomous Region China. 329-333
Shaffril HAM, Krauss SE, Samsuddin SF. 2018. A systematic review on Asian’s farmers’ adaptation practices towards climate change. Science of the Total Environment 644:683-695
Sharif M, Yegneswaran V, Saidi H, Porras P, Lee W. 2008. Eureka: a framework for enabling static malware analysis. In: Lecture Notes in Computer Science. Vol. 5283:481-500
Sheen S, Anitha R, Natarajan V. 2015. Android based malware detection using a multifeature collaborative decision fusion approach. Neurocomputing 151:905-912
Shuying Liang DVH, Keep AW, Might M, Gilray T, Lyde S, Aldous P. 2013. Sound and Precise Malware Analysis for Android via Pushdown Reachability and Entry-Point Saturation. In: ACM workshop on Security and privacy in smartphones and mobile devices. Berlin, Germany. New York. ACM. 21-32
Singh AK, Jaidhar CD, Kumara MAA. 2019. Experimental analysis of Android malware detection based on combinations of permissions and API-calls. Journal of Computer Virology and Hacking Techniques 15(3):209-218
Skylot. 2015. Jadx. GitHub (accessed02 October 2019 )
Sony. 2019e. Sony Malaysia. (accessed02 October 2019 )
Sourceforge. 2019d. Dex2jar. (accessed02 October 2019 )
Spreitzenbarth M, Schreck T, Echtler F, Arp D, Hoffmann J. 2015. Mobile-Sandbox : combining static and dynamic analysis with machine-learning techniques. 141-153
Statista. 2019. Development of new Android malware worldwide from 2011 to 2018. (accessed30 September 2019 )
Steven Arzt SR, Christian Fritz JK, Bodden E, Bartel A, PM, Traon YL, Octeau D. 2014. FlowDroid: precise context, flow, field, object-sensitive and lifecycle-aware taint analysis for android apps. In: Proceedings of the 35th ACM SIGPLAN conference on programming language design and implementation. Edinburgh, United Kingdom. New York. ACM. 259-269
Suarez-Tangil G, Tapiador JE, Peris-Lopez P, Blasco J. 2014. Dendroid: a text mining approach to analyzing and classifying code structures in Android malware families. Expert Systems with Applications 41(4 PART 1):1104-1117
Sufatrio, Tan DJJ, Chua T, Thing VLL. 2015. Securing android: a survey, taxonomy, and challenges. ACM Computing Surveys 47(4):1-45
Sun YS, Chen CC, Hsiao SW, Chen MC. 2018. ANTSdroid: Automatic Malware Family Behaviour Generation and Analysis for Android Apps. In: Susilo W, Yang G, eds. Information Security and Privacy. ACISP 2018. Lecture Notes in Computer Science. Cham: Springer. vol. 10946
Talha KA, Alper DI, Aydin C. 2015. APK Auditor: permission-based Android malware detection system. Digital Investigation 13:1-14
Technische Universität Braunschweig. 2016. Drebin dataset.
Traceflight. 2019. AMD dataset. GitHub.
Turker S, Can AB. 2019. AndMFC: android malware family classification framework. In: 2019 IEEE 30th int. symp. pers. indoor mob. radio commun. PIMRC work. 2019. Piscataway. IEEE. 3-8
Université du Luxembourg. 2016. Androzoo.
Vega RV, Quintián H. 0000. Gaining deep knowledge of Android malware families through dimensionality. 00
Vega RV, Quintián H, Cambra C, Basurto N, Herrero Á, Calvo-rolle JL. 2019. Delving into android malware families with a novel neural projection method, vol. 2019. United Kingdom: Hindawi.
Venkatraman S, Alazab M, Vinayakumar R. 2019. A hybrid deep learning image-based analysis for effective malware detection. Journal of Information Security and Applications 47:377-389
Volodymyr Mnih MR, Kavukcuoglu K, Silver D, Graves A, Antonoglou I, Wierstra D. 2013. Playing atari with deep reinforcement learning.
Volodymyr M, Kavukcuoglu K, Silver D, Rusu AA, Veness J, Bellemare MG, Graves A, Riedmiller M, Fidjeland AK, Ostrovski G, Petersen S, Beattie C, Sadik A, Antonoglou I, King H, Kumaran D, LeggS W, Hassabis D+8 more. 2015. Human-level control through deep reinforcement learning. Nature 518(7540):529-533
Walenstein A, Deshotels L, Lakhotia A. 2012. Program structure-based feature selection for android malware analysis. In: Schmidt AU, Russello G, Krontiris I, Lian S, eds. Security and Privacy in Mobile Information and Communication Systems. MobiSec 2012. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering. Berlin, Heidelberg: Springer. Vol. 107:51-52
Wan X, Sheng G, Li Y, Xiao L, Du X. 2018. Reinforcement learning based mobile offloading for cloud-based malware detection. In: 2017 IEEE Glob. Commun. Conf. GLOBECOM 2017. Piscataway. IEEE. 1-6
Wang W, Gao Z, Zhao M, Li Y, Liu J, Zhang X. 2018. DroidEnsemble: detecting Android malicious applications with ensemble of string and structural static features. IEEE Access 6:31798-31807
Wei F, Li Y, Roy S, Ou X, Zhou W. 2017. Deep ground truth analysis of current android malware. In: Polychronakis M, Meier M, eds. Detection of Intrusions and Malware, and Vulnerability Assessment. DIMVA 2017. Lecture Notes in Computer Science. Cham: Springer. vol. 10327:252-276
Wiśniewski R. 2010. apktool.
Wu D, Liu J, Sui Y, Chen S, Xue J. 2019. Precise static happens-before analysis for detecting UAF order violations in android. In: IEEE 12th Int. Conf. Softw. Testing, Verif. Validation, ICST. Piscataway. IEEE. 276-287
Wu D-J, Mao C-H, Wei T-E, Lee H-M, Wu K-P. 2012. DroidMat: android Malware Detection through Manifest and API Calls Tracing. In: Seventh Asia Joint Conference on Information Security, Tokyo, Japan. 62-69
Wu S, Wang P, Li X, Zhang Y. 2016. Effective detection of android malware based on the usage of data flow APIs and machine learning. Information and Software Technology 75:17-25
Wu S, Zhang Y, Jin B, Cao W. 2018. Practical static analysis of detecting intent-based permission leakage in Android application. In: 2017 IEEE 17th International Conference on Communication Technology (ICCT). Piscataway. IEEE. 1953-1957
Xie N, Wang X, Wang W, Liu J. 2019. Fingerprinting Android malware families. Frontiers of Computer Science 13(3):637-646
Xu Z, Ren K, Song F. 2019. Android Malware Family Classification and Characterization Using CFG and DFG. In: 2019 International Symposium on Theoretical Aspects of Software Engineering (TASE). 49-56
Yang P, Hwa Yang Y, Zhou BB, Zomaya AY. 2010. A review of ensemble methods in bioinformatics. Current Bioinformatics 5(4):296-308
Yang Z, Yang M. 2012. Leak miner: detect information leakage on android with static taint analysis. In: Third World Congress on Software Engineering (WCSE). Wuhan, China. 101-104
Yangqing Jia SK, Shelhamer E, Donahue J, Jonathan Long TD, Girshick R, Guadarrama S. 2014. Caffe: convolutional architecture for fast feature embedding. In: MM ’14: Proceedings of the 22nd ACM international conference on Multimedia. New York. ACM. 675-678
Yajin Zhou XJ. 2012. Malgenome Project.
Yaqoob I, Abaker I, Hashem T, Ahmed A, Kazmi SMA. 2019. Internet of things forensics: recent advances, taxonomy, requirements, and open challenges. Future Generation Computer Systems 92(2018):265-275
Yerima SY, Sezer S, McWilliams G. 2014. Analysis of Bayesian classification-based approaches for Android malware detection. IET Information Security 8(1):25-36
Yerima SY, Sezer S, McWilliams G, Muttik I. 2013. A New Android malware detection approach using Bayesian classification. In: IEEE 27th international conference on advanced information networking and applications (AINA). Barcelona, Spain. Piscataway. IEEE. 121-128
Yerima SY, Sezer S, Muttik I. 2014. Android malware detection using parallel machine learning classifiers. In: Eight international conference on next generation mobile apps, services and technologies, (NGMAST). UK. St. Anthony’s College of the University of Oxford. 37-42
Yerima SY, Sezer S, Muttik I. 2015. High accuracy android malware detection using ensemble learning. IET Information Security 9(6):313-320
Yildiz O, Doǧru IA. 2019. Permission-based Android malware detection system using feature selection with genetic algorithm. International Journal of Software Engineering and Knowledge Engineering 29(2):245-262
You I, Yim K. 2010. Malware obfuscation techniques: A brief survey. In: 2010 International Conference on Broadband, Wireless Computing, Communication and Applications. 297-300
Zhang J, Tian C, Duan Z. 2019. FastDroid: efficient taint analysis for android applications. In: IEEE/ACM 41st Int. Conf. Softw. Eng. Companion. Piscataway. IEEE. 236-237
Zhang L, Thing VLL, Cheng Y. 2019. A scalable and extensible framework for android malware detection and family attribution. Computers & Security 80:120-133
Zhao L, Li D, Zheng G, Shi W. 2019. Deep neural network based on android mobile malware detection system using opcode sequences. In: 2019 IEEE/ACM 41st International Conference on Software Engineering: Companion Proceedings (ICSE-Companion). 1141-1147
Zheng M, Sun M, Lui JCS. 2013. Droid analytics: A signature based analytic system to collect, extract, analyze and associate android malware. In: 12th IEEE international conference on trust, security and privacy in computing and communications, (TrustCom). Piscataway. IEEE. 163-171
Zhou H, Zhang W, Wei F, Chen Y. 2017. Analysis of Android Malware Family Characteristic Based on Isomorphism of Sensitive API Call Graph. In: 2017 IEEE Second International Conference on Data Science in Cyberspace (DSC). Piscataway. IEEE. 319-327
Zhou W, Zhou Y, Grace M, Jiang X, Zou S. 2013. Fast, scalable detection of ‘Piggybacked’ mobile applications. In: CODASPY ’13 proceedings of the second ACM conference on data and application security and privacy. San Antonio, Texas, USA. 185-195
Zhou W, Zhou Y, Jiang X, Ning P. 2012. Detecting Repackaged Smartphone Applications in Third-Party Android Marketplaces. In: CODASPY ’12 Proceedings of the second ACM conference on Data and Application Security and Privacy. Piscataway. IEEE. 317-326
Rosmalissa Jusoh1, Ahmad Firdaus1, Shahid Anwar2, Mohd Zamri Osman1, Mohd Faaizie Darmawan3, Mohd Faizal Ab Razak1
1 Faculty of Computing, College of Computing and Applied Sciences, Universiti Malaysia Pahang, Pekan, Pahang, Malaysia
2 Department of Information Engineering Technology, National Skills University, Islamabad, Pakistan
3 Faculty of Computer & Mathematical Sciences, Universiti Teknologi Mara, Tapah, Perak, Malaysia
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2021 Jusoh et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: https://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Android is a free open-source operating system (OS), which allows an in-depth understanding of its architecture. Therefore, many manufacturers are utilizing this OS to produce mobile devices (smartphones, smartwatch, and smart glasses) in different brands, including Google Pixel, Motorola, Samsung, and Sony. Notably, the employment of OS leads to a rapid increase in the number of Android users. However, unethical authors tend to develop malware in the devices for wealth, fame, or private purposes. Although practitioners conduct intrusion detection analyses, such as static analysis, there is an inadequate number of review articles discussing the research efforts on this type of analysis. Therefore, this study discusses the articles published from 2009 until 2019 and analyses the steps in the static analysis (reverse engineer, features, and classification) with taxonomy. Following that, the research issue in static analysis is also highlighted. Overall, this study serves as the guidance for novice security practitioners and expert researchers in the proposal of novel research to detect malware through static analysis.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer