1. Introduction
Recently, machine learning has been fueling active discussions among clinicians and health researchers, particularly for decision making in e-diagnosis, disease detection and medical image analysis [1,2,3,4,5]. A few common questions are “can machine learning replace conventional statistics?”, “are they the same?” and “how statistics be integrated with machine learning?”. This review focuses on the concept of conventional statistics and machine learning in health research and the explanation, comparison and examples may answer the aforementioned questions.
It is seen from various research that conventional statistics have dominated health research [6,7,8,9,10,11,12,13,14,15]; however, machine learning, since its inception, is widely being used by data scientists in various fields [16,17,18,19,20,21,22,23,24,25,26,27]. Examples of common conventional statistical analyses are hypothesis testing (t-test, ANOVA), probability distributions (regression) and sample size calculation (hazard ratio), whereas in machine learning, the common concepts are model evaluation, variable importance, decision tree, classification and prediction analysis. While many industries are capturing the vast potential of machine learning, healthcare is still slow in attaining the optimum level to make sense of newer technologies and computational methods. This could be due to uncertain reliability and trust in machines to analyze big data and make timely decisions on patients’ health. The production of large amounts of healthcare data [28,29], such as administrative data, patients’ medical history, clinical data, lifestyle data and other personal data, makes their analyses unmanageable using basic statistical software tools, which subsequently leads to the need of technologically advanced applications with cost-efficient high-end computational power. More importantly, the applications must meet the target or be designed with user-friendly interfaces and data protection utilities to aid the end users, who are the biostatisticians, researchers and clinicians. The experts need automated systems to optimize diagnoses and treatments, enhance prognostications, predict health risks and develop long-term care plans. In line with this, we need to answer the important question of whether machine learning is completely different from conventional statistics or they are correlated with each other.
To begin with, conventional statistics have a history of over 50 years, beginning in the early 17th and 18th centuries, when mathematical theories were introduced by various scientists. In the 18th century, the importance of advanced statistics in medicine was a prominent topic, where more theories were integrated to invent inferential statistical models. Later, the use of computational power in statistical analysis was given priority, hence advanced software tools were developed. Machine learning was introduced in 1952, and recently it has advanced into deep learning and is used as the basis of artificial intelligence [30,31,32]. Figure 1 describes the evolution of conventional statistics and machine learning in health research.
1.1. Past Reviews, Rationale for the Review and Intended Audience
There are recent reviews on the comparison between conventional statistics and machine learning [33,34,35,36,37]. These reviews have presented the definitions of the two terms and the advantages and disadvantages over each other. However, to the best of our knowledge, the conceptual integration between these two fields using examples in health research have not been discussed previously. Moreover, the similarities between these two fields have not been investigated thoroughly. This review may clarify the confusion among clinicians as to whether machine learning can be integrated with conventional statistics in health research. Clarifying the integration between conventional statistics and machine learning may be able to convince health researchers to explore this approach in the future. The intended audience of this review is not only healthcare researchers, but statisticians and data scientists as well.
1.2. Review Content
This review contains five sections: (i) concepts in conventional statistics and machine learning, (ii) advantages and disadvantages of conventional statistics and machine learning, (iii) a case study of breast cancer survival analysis using a few techniques comparing conventional statistics and machine learning, (iv) simplified machine learning algorithms and their relationship with conventional statistics and (v) a discussion explaining the integration of conventional statistics with machine learning and the significance of machine learning, derived from fundamental conventional statistics. Section (iii) is explained using a proven breast cancer prediction model [38], which has attracted a broad range of readers both from the medical domain and computer science. The terms conventional statistics (CS) and machine learning (ML) are used throughout the review.
2. Survey Methodology
The review was conducted using published works related to:
i.. history of conventional statistics and machine learning in medicine
ii.. comparison between conventional statistics and machine learning
iii.. use of machine learning in various fields
iv.. analysis of medical data using conventional statistics and
v.. use of machine learning and artificial intelligence in medical analysis.
The digital libraries and search engines used to extract the literature are Google Scholar, Web of Science and PubMed. The literature search was followed by selecting relevant literature using inclusion and exclusion criteria as listed below.
2.1. Inclusion Criteria
-
(i). all papers with year of publication between 2015 to 2022
-
(ii). all open access papers that are freely available
-
(iii). the keywords used for the search are conventional statistics, machine learning, medical data, comparison and health research. The entries by using these keywords were from various medical domains, machine learning analyses and statistics in healthcare research, not focusing only on one type of disease.
2.2. Exclusion Criteria
-
(i). all papers not relevant to our topic
-
(ii). all papers that are not freely accessible
-
(iii). all papers with year of publication before 2015
The initial total number of papers is 511. The selection process is explained in Figure 2. The final total number of selected literature is 102.
3. Results
3.1. Concepts in Conventional Statistics
3.1.1. Hypothesis Testing and Statistical Inference for Classification
Hypothesis testing is the interpretation of results by making assumptions (hypotheses) based on experimental data. The statistical tests (e.g., t-test, ANOVA) are used to interpret the results based on measures such as p-value (significant difference). Biostatisticians and medical scientists perform statistical analysis using conventional software tools [39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56] because healthcare providers’ main objective is to focus on analysis based on hypothesis testing in the context of patient care to check if treatments or drugs yield positive outcomes or how to control certain risk factors for a particular disease. Therefore, they barely explore or pay attention to the use of advanced computer science applications and automated predictive tools such as Predict, CancerMath and Adjuvant. [57]. The basic concepts of hypothesis testing are explained in Table 1.
In conventional statistics, the approach used is the conclusion or “inference” in the form of mathematical equations and measures to make predictions. For instance, an inferential work to assess unknown evidence from observed data could be achieved via a hypothesis-testing framework. The aim of hypothesis testing is to reject the null hypothesis if the evidence found is true and clinically significant. For example, in deciding which surgical treatment, “does breast-conserving therapy or mastectomy promote better survival among breast cancer patients?” is an inferential question and the answer is unobservable. In this scenario, patients are considered the observation, whereas the treatment types and survival data are the independent variables, which decide the inference [34]. The results of the analysis classify the dependent variable (surgical treatment) based on the patterns of independent variables.
3.1.2. Regression
Regression analysis is a set of statistical methods to estimate the relationship between a dependent variable and a set of independent variables. Regression has been widely used in healthcare research to analyze and make predictions on various diseases. Selection of a particular type of regression depends on the type of dependent variable, such as continuous and categorical. Linear regression is used to determine the relationship between a continuous dependent variable and a set of independent variables. This analysis estimates the model by minimizing the sum of squared errors (SSE). Moreover, nonlinear regression also requires a continuous dependent variable, but this is considered advanced as it uses an iterative algorithm rather than the linear approach of direct matrix equations [58].
A categorical dependent variable is analyzed using logistic regression. This analysis transforms the dependent variables which have values of distinct groups based on specific categories and uses Maximum Likelihood Estimation to estimate the parameters. Logistic regression is further divided into binary, ordinal and nominal categories. A binary variable has only two values, such as survival status (alive or dead), an ordinal variable has at least three values in order, such as cancer stage (stage 1, stage 2, stage 3), whereas a nominal variable has at least three values which are not categorized in any order, such as treatment (chemotherapy, radiotherapy, surgery) [38].
3.2. Concepts in Machine Learning (ML)
3.2.1. Predictive Analytics
Prediction works with the concept of modeling using machine learning algorithms. This prediction model requires a reliable relationship between the observations (patients) and variables (independent variables). Prediction models generate accuracy measures to determine the quality of data and predict the final outcome using the observations (patients), input data (independent variables) and output data (dependent variable). The basic concepts of predictive analysis are explained in Table 2.
3.2.2. Representation Learning
Representation learning is the process of training machine learning algorithms to discover representations which are interpretable. Different representations can entangle various explanatory factors in a specific dataset. The outcome of representation learning should ease the subsequent task in the decision-making process. For example, representation learning handles and groups very large amounts of unlabeled training data in unsupervised or semi-supervised learning. The grouping of the unlabeled data is used for the corresponding task, such as feature selection and decision tree, to predict outcomes. The challenging factor of representation learning is that it has to preserve as much information as the input data contains in order to attain accurate predictions. Healthcare research utilizes representation learning mostly in image recognitions, such as biomedical imaging-based predictions [59].
3.2.3. Reinforcement Learning
Reinforcement learning (RL) trains machine learning models to make a sequence of decisions, unlike supervised learning, which relies on a one-shot or single dependent factor. Its main objective is to endow an individual’s skills to make predictions through experience with the environment around them and develop evaluative feedback. This unique feature of reinforcement learning helps in providing prevailing solutions in various healthcare diagnosis and treatment regimens which are usually characterized by a prolonged and sequential procedure. Reinforcement learning follows a few techniques for sequential decision making, namely, efficient techniques such as experience-level, model-level and task-level and representational techniques such as representation for value function, reward function and task or models). Applications of RL in different healthcare domains, such as chronic diseases and critical care, especially sepsis and anesthesia, are explained in detail in [60,61].
3.2.4. Causal Inference/Generative Models
Causal inferencing plays a vital role in understanding the mechanisms of variables to find a generative model and predict outcomes which the variables are subjected to. The characteristic of causal inference is to find answers for questions about the mechanisms by which the variables come to take on values. For example, epidemiologists gather dietary-related data and find the factors affecting life expectancy to predict the effects of guiding people to change their diet. Examples of causal models are models with free parameters (fixed structure and free parameters) and models with fixed parameters (free parameters with values). A large number of variables, small sample size and missing values are considered serious impediments to proper data analysis and production of accurate decision making in the medical domain. Causal inferencing is used in healthcare research mainly for clinical risk prediction and improving accuracy of medical diagnosis, despite the issues with data [62].
3.3. Advantages and Disadvantages of Conventional Statistics and Machine Learning
3.3.1. Data Management
Conventional statistics are more suitable for simpler datasets, whereas machine learning can manage complex datasets. In general, conventional statistical analyses are performed if the research has prior literature about the topic of interest, the number of variables involved in the study is relatively small and the number of observations (samples) is bigger than the number of variables. This may assist the scientist in understanding the topic conveniently by selecting the important variables from prior knowledge, as well by applying appropriate analytical models to check the association between variables (independent) and outcomes (dependent). The conventional statistical approach gives more priority to the type of dataset, for example, those including a cohort study, which follow a specific hypothesis [35]. On the other hand, prediction analysis based on machine learning algorithms learn from data without relying on rules-based programming, which does not make any prior assumption, but is rather based on the original data provided. Machine learning algorithms can handle multi-dimensional big data, but conventional statistics can handle only one specific format of data at a time. Furthermore, machine learning algorithms can handle data from different data sources such as external databases or online data repositories.
3.3.2. Computational Power, Interpretation/Explainability and Visualization of Results
The analytical strategy of biostatisticians depends on their pathophysiological knowledge and experience where data-driven prediction analysis challenges this paradigm of thought, and the increasing computational power may unmask the associations not realized by the human mind.
In conventional statistical analysis, scientists use basic software tools, which lack the capability to handle big data and visualization of results. Machine learning black box algorithms have the ability to uncover subtle hidden patterns in multi-model data [63]. However, interpretability is domain-specific, hence the visualization techniques play a vital role in explaining the results to higher stakeholders. Conventional statistical software tools produce basic visualization, whereas the advanced data analytics tools produce domain-specific, customized, inherently interpretable models and results [63]. Machine learning is often very complex and difficult to be interpreted by clinicians because it uses computational programming and not a user-friendly tool such as SPSS. Conventional statistics are easily interpretable and have lower capacity, thus present a smaller risk of failing to generalize non-causal effects.
Conventional statistics are considered more computationally efficient and more readily acceptable in the medical domain. Contrarily, the results and visualizations produced by ML algorithms are different from the CS methods, and no proper guideline is available on the ways to explain the graphs for interpretation of final results. Machine learning requires high computational power in terms of processing power and storage. Moreover, ML algorithms are updated regularly into newer versions, which requires updates in coding. Furthermore, ML models have the ability to over-predict (overfitting), where the predicted model is closely related to the provided dataset. This could constrain the possibility to generalize the model in different datasets to produce better accuracy. Therefore, validity is required in both cases to finalize decisions. ML algorithms are able to provide required results and decisions automatically from precise training data based on their built-in functions from the programming tools. Nevertheless, when dealing with large amount of data, more hybrid models can be designed to resolve the issues arising in data science for knowledge extraction, especially in healthcare. There is a myriad of algorithms and software for ML techniques to build prediction models on diseases. In medical informatics, R, Matlab, Waikato Environment for Knowledge Analysis (WEKA) toolkit and Python [64,65,66,67,68,69,70] are a few of the widely used programming languages and software in conducting prediction analysis. In the need of viable decisions and interpretation, healthcare providers and researchers can consider leveraging explainable ML models, instead of focusing only on the results. If the interpretability in a domain-specific notion can be followed by researchers, the level of trust on black boxes among clinicians could be improved.
3.3.3. Dimensionality Reduction
Dimensionality reduction involves reduction of either dimension of the observation vectors (input variables) into smaller representations [71]. Technically, dimensionality reduction transforms original dataset A of dimensionality N into a new dataset B of dimensionality n [72]. Machine learning models follow various dimensionality reduction techniques based on the types of data in a specific research analysis. The larger the number of input variables, the greater the complication in the predictive models; thus, dimensionality reduction helps to select the best input variables to predict the models. A few methods of reduction are Principal Component Analysis (PCA), Kernel PCA (KPCA), t-distributed Stochastic Neighbor Embedding (t-SNE) and UMAP. Dimensionality reduction techniques remove irrelevant input variables from the dataset, which could increase the accuracy of machine learning models. It also helps to eliminate multi-collinearity, which enhances the way of interpreting the variables. In line with this, the dataset with the relevant input variables saves storage space, and less computing power is needed to analyze the data [73].
3.3.4. Frequently Used Models or Methods for Data Assessment
The most frequently used models for the association study in conventional statistics is logistic regression or Cox regression models for binary outcomes, linear regression for continuous outcomes and more extensive models such as generalized linear models based on the distribution of data. This scenario is typically popular in studies addressing public health significance, especially when the analysis involves a population study [74,75,76,77]. Statisticians believe that, in order to draw a firm conclusion or inference, the number of observations in an association study plays an important role [34]. This is a direct approach in hypothesis testing.
Machine learning models are able to capture high-capacity relationships and they are amenable to more operational tasks rather than direct research questions; thus, more research gaps could be solved through the one-stop analysis [38]. Various medical data analyses used a machine learning approach to make decisions [78,79,80,81,82,83,84,85,86,87,88]. Biostatisticians are in a need of an updated methodology that uses a machine learning approach to conduct analysis on a variety of medical data [89]. In this case, the similar concepts in CS and ML need to be emphasized. Machine learning algorithms serve as alternatives to the conventional statistics for common analyses, such as determining effect size, significant factors, survival analysis and imputations. While conceptually they are similar, they are distinct in terms of methods. The core differences between CS and ML concepts are described in Table 3.
3.4. Case Study to Compare Conventional Statistics and Machine Learning
A breast cancer dataset from the University Malaya Medical Centre (UMMC), n = 8066, diagnosed between 1993 and 2017, was used to perform prediction analysis using both conventional statistics and machine learning. Written informed consent was obtained from the participants included in this study. This dataset was extracted from the cancer registry within the electronic medical record system of UMMC called iPesakit. A total of 23 independent variables and survival status (dependent variable) were used to determine the most important prognostic factors of breast cancer survival. The data description is provided in Table 3. The methods and results from three different types of analysis are compared. SPSS was used to perform conventional statistics and R was used to perform machine learning. The R codes used for machine learning analysis stated in the case study of this paper are deposited on GitHub [90].
3.4.1. Imputation and Data Pre-Processing
Imputation applies both to conventional statistics and machine learning during data cleaning. Single or multiple imputations can be performed using conventional statistical software and programming tools such as R. In this case study, imputation was performed on the dataset to fill the missing values only for conventional statistical analysis. This is because the machine learning algorithms are able to handle the data with missing values. The dataset was split into testing (30%) and training (70%) for machine learning.
3.4.2. Significant Factors (CS) and Variable Importance (ML)
The objective of this analysis was to compare conventional statistics and machine learning (variable importance) to determine the similarities and differences in the results using the same dataset. Table 4 shows the results using significant factor analysis in SPSS. The results from the chi squared test (categorical variable) and Mann–Whitney U test (continuous variables) show that all the independent variables are statistically significant (p-value < 0.05).
Figure 3 shows the variable importance plot using random forest VSURF and randomForestExplainer packages in R [72]. The variables are ranked based on variable importance mean from highest to lowest. A threshold was set up to 0.01 and six variables were selected as the most important prognostic factors of breast cancer survival.
The difference between significant testing and variable importance is that the order of the importance is determined in variable importance, but only the status of significance (statistically significant or not) could be determined using the significant factor analysis.
3.4.3. Survival Analysis
Survival analysis in machine learning follows exactly the same concept as the conventional statistics, which is the Kaplan–Meier (KM) estimator. The time series data, date of diagnosis, date of death and date of last follow-up are used to calculate the overall survival rate. The methods used are different; in machine learning, the KM estimator is encapsulated into a single package called survival in R. Programming codes are used to plot the survival curve directly by specifying the variables. In contrast to conventional statistics, it is not an algorithm, but a type of data analysis where the time series data are selected to plot survival curves with a life table and hazard ratio. Both conventional statistics and machine learning follow the same rules to predict survival rate. The survival curves are shown in Table 5. Survival curves are created for three variables: tumor size, cancer stage and positive lymph nodes. The survival curves from SPSS and R produced quite similar results in terms of survival rate for various categories in each variable, but with differences in numerical values (survival percentages).
3.5. Simplified Machine Learning Algorithms and Their Relationship with Conventional Statistics
The mathematical equations in conventional statistics are encapsulated to form algorithms in machine learning. These algorithms are used to perform predictions using supervised and unsupervised machine learning. The integration between the mathematics behind conventional statistics and machine learning are explained using the techniques, model evaluation (supervised learning), variable importance (supervised learning) and hierarchical clustering (unsupervised learning). A proven breast cancer prediction model [38] has been used to explain the concepts of the algorithms.
Model evaluation in machine learning is similar to power analysis in conventional statistics for assessing the quality of data. It is the key step in machine learning, as the ability of the model to make predictions on unseen or future samples enhances the trust on the model to be used in a particular dataset. The measurement for model evaluation is the accuracy in percentage (estimate of generalization of a model on prospective data). Six different supervised machine learning algorithms (decision tree, random forest, extreme gradient boosting, logistic regression, support vector machine, artificial neural networks) are simplified. These algorithms have been widely used in medical informatics [65,67,91,92,93].
3.5.1. Decision Tree
Decision tree has been widely used in medical informatics [63,71], as it is the basic concept used by other algorithms such as random forest and gradient boosting, but with certain differences in the processes to predict the final output.
The decision tree algorithm follows the model of a tree structure, where it has a root node, decision node and terminal node. The root node starts with the most important independent variable followed by decision nodes (other independent variables). The terminal node indicates the dependent variable, which is the final predicted output.
The processes in the decision tree are summarized into three steps: (i) choosing features, (ii) setting conditions to split and (iii) stopping the splitting process to produce a final output. A tree structure is built based on the observation falling in each region and the mean value of prediction.
The splitting process is continued until a user-defined stopping criteria (the number of observations per node) is reached. In the case of more than two variables, the regions cover all the variables with multiple axes.
3.5.2. Random Forest
Random forest is an ensemble learning algorithm, which is derived from decision tree. It follows the rule of decision tree, but constructs a multitude of decision trees at training time and outputs the class with the maximum vote. Random forest is the state-of-the-art algorithm in medical informatics, as it has the ability to manage multivariate data [72].
Random forest is known as an improved version of decision tree, as it constructs more than one tree to select the best output, whereas decision tree constructs only one tree. The number of trees constructed during the training process is not default, as the users can specify it based on the number of samples. The number of trees is directly proportional to the number of samples.
3.5.3. Extreme Gradient Boosting
Gradient boosting follows the principle of random forest but with an added interpretation to predict the final output. This algorithm also constructs multiple trees called boosted trees. A prediction score is assigned for each leaf in the boosted trees (gradients), whereas random forest only contains the final decision value for one tree. Several studies have used the gradient boosting approach to analyze medical data [73,92].
This algorithm also considers the weak and strong prediction values during training before making the final decision, unlike decision tree and random forest, which only select the tree with the best class, without considering the other classes. This method in gradient boosting is known as the impurity measure. The scores of all the leaves in the trees are summed up to produce the gradient values, and the final prediction is made based on the mean value, called gradient boosting.
3.5.4. Logistic Regression
Most studies use regression for prediction analytics in medicine [61]. Logistic regression predicts categorical output, for example, the survival status (alive or dead). The predictions are made based on the probabilities shown by a curve. This process is repeated for all the samples. The curve is shifted to calculate new likelihoods of the samples falling on that line. Finally, the likelihood of the data is calculated by multiplying all the likelihoods together and the maximum likelihood is selected as the final result.
3.5.5. Support Vector Machine
Just like all other algorithms, support vector machine (SVM) segregates data into different classes, but it involves discovery of hyperplanes. The hyperplane divides the data into two groups (classes). The points closer to the decision boundary or hyperplane are called support vectors. The final prediction is made based on the values of independent variables and the support vectors corresponding to the hyperplane. The number of hyperplanes depends on the number of independent variables. The SVM structure is complicated, with more than three features, but its ability to process multiple variables with multiple hyperplanes at a time to predict the final outcome is one of the advantages of this algorithm.
3.5.6. Artificial Neural Networks
Neural networks are an artificial representation of the human nervous system. It can be explained using the structure of neurons and how they work. The dendrites collect information from other neurons in the form of electrical impulses (input). The cell body generates inferences based on the inputs and decides the actions to be taken. The outputs are transmitted through exon terminals as electrical impulses to other neurons.
The same concept is implied in artificial neural networks (ANN). The inputs refer to the independent variables and samples provided to the algorithm. The inputs are multiplied by weights to calculate the summation function. The higher the weight an input has, the more significant the input is to predict the final output. The activation function predicts the probabilities from the training data and generates a final outcome. This is known as a single-layer perceptron. There are three types of layers in ANN, which are input layer, hidden layer and output layer.
Model evaluation is followed by variable importance in machine learning. Variable importance (importance score) is an alternative to identifying the significant factors (p-value) in conventional statistics using confidence interval measure and hypothesis testing. After performing model evaluation, the elements (variables) of the input data need to be explored further in regard to how they contribute to the accuracy measure. Hence, machine learning algorithms are built-in with a technique called variable importance or feature importance to analyze the variables or features in the input data. The distribution of these variables contributes to the prediction of the final outcome using machine learning models.
4. Discussion
4.1. Integration of Conventional Statistics with Machine Learning
Statistics is a branch of mathematics that consists of a combination of mathematical techniques to analyze and visualize data 90. On the other hand, machine learning is a branch of artificial intelligence that is composed of algorithms performing supervised and unsupervised learning. The comparison or integration between conventional statistics and machine learning has gained momentum over the last few years [94,95]. It is plausible that data integrity with protection is the most challenging task in healthcare analytics [96,97,98]. Hence, from this review, it is found that the integration between these two fields could unlock and outline the key challenges in healthcare research, especially in handling the valuable asset called data. Individuals should not be subject to a final decision based solely on automated processing or machine learning using algorithms, but integration of statistics and human decision making is essential at an equal rate. The integration between statistics and machine learning is shown in Figure 4.
4.2. Significance of Machine Learning to Healthcare, Education and Society
The review on the integration between conventional statistics and machine learning is the key factor to convince clinicians and researchers that machine learning algorithms are based on core conventional statistical ideas; thus, they could be used to supplement data analysis using conventional statistics. From this review, we believe that machine learning, which follows the fundamentals of conventional statistics, has a positive impact on healthcare. The significance of machine learning to healthcare is explained (Figure 5).
Prior to the emergence of the data deluge, healthcare providers made clinical decisions based on formal education and their experience over time in practice. Decision analysis in healthcare has been criticized because the experience and knowledge of the decision makers (clinicians) on patient characteristics are not the same or standardized. The linear process of the decision-making model involves four steps, which are data gathering, hypothesis generation, data interpretation and hypothesis evaluation. All four steps require data from different departments, clinicians from different expertise and various data analytical methods to make the final decision. Experienced clinicians may not deliberately go through each step of the process and may use intuition to make decisions, instead of facing obstacles handling several hypotheses with different personnel. As this is applied to experienced clinicians, a novice clinician would have to understand and rely on the analytical principles and theory behind a decision analysis process in a particular situation handling a patient. In this case, the healthcare sector is in need of clinical decision support tools to enhance and standardize clinical decision making.
Machine learning algorithms are widely used to develop clinical decision support tools. These algorithms compile the four steps (data gathering, hypothesis generation, data interpretation and hypothesis evaluation) of traditional decision making into one. The advantages of machine learning algorithms in medical informatics depend on the objectives of the research and the types of data used. ML algorithms such as decision tree, random forest, gradient boosting, regression, support vector machine and artificial neural networks are suitable for medical informatics, as they are able to handle big data, a combination of numerical and categorical data and missing values. Moreover, these algorithms generate visualizations, which could be transformed automatically (integrated into tools) to be used by the clinicians as guidelines for patients.
In any machine learning analysis, domain experts are still required to enhance the reliability of the machine and make sense of the results. In medical informatics specifically, the decision of clinicians on a particular patient’s health condition plays an important role in giving suggestions to the patient. The automated decision support tools may help clinicians in decision making to save time and costs, and to follow a standard procedure to prevent conflict in final decisions.
The field of medicine relies heavily on knowledge discovery and understanding of diseases associated with the growth in information (data). Diagnosis, prognosis and drug development are the challenging key principles in medicine, especially in complex diseases, such as cancer [91]. Based on the principal of evidence-based medicine, decision making based on data and validation should be more agile and flexible to better translate the basic knowledge of complexities into growing advances. The integration of conventional statistics and machine learning to clinical applications should be carefully adopted with a collaborative efforts that includes all major stakeholders for the positive influence of machine learning in medicine [91].
Comparison between the basic workflow of conventional statistics and machine learning is explained in Figure 6.
4.3. Automation of Machine Learning in Healthcare Research
The machine learning approach could be transformed into an updated guideline for academicians and researchers. The medical academic sector may use the methodologies for teaching and learning programs to educate medical students on the importance of machine learning. Moreover, researchers in the same field can follow the techniques and machine learning models to conduct research and cohort studies in any healthcare domain.
Biostatisticians may consider using machine learning techniques and automated tools [92,93] together with conventional statistics in order to improve the performance of analytics and reliability of results. The integration between statistics and machine learning may assist biostatisticians to provide novel research outcomes.
The automation of machine learning in healthcare analysis has been applied in a recent study by our research group [99]. The automated tools may assist biostatisticians to provide novel research outcomes.
A guideline to transform statistics and machine learning (derived from the fundamental mathematics of conventional statistics) into an automated decision support tool is illustrated in Figure 7.
This guideline could be a standardized pipeline for the data science community to analyze medical data and to develop artificial intelligence-enabled decision support tools for clinicians and researchers.
In a decision support tool, (i) the data gathering is replaced by the automated data capture from electronic medical records (EMR) or databases from multiple heterogeneous sources, concomitantly; (ii) hypothesis generation is the specification of input variables (independent and target) and the final outcome based on the research question or a question for clinical decision (output); (iii) data interpretation is done using algorithms such as random forest, support vector machine and neural networks, which have their specific formulas to read the data, clean the data, capture the required variables, analyze the data based on the specified requirements and perform comparative analytics automatically using different algorithms; (iv) finally, hypothesis evaluation is done by producing interactive charts to visualize the final outcomes to make decisions efficiently. All these steps are performed in a streamlined environment often referred to as automated clinical decision making, which saves the effort of engaging different experts and analytical platforms. The experience which clinicians traditionally use to make decisions is replaced by the legacy data the algorithms leverage to make decisions.
Most important of all, the data management or completeness of data for automated decision making plays an essential role when it comes to statistics and machine learning. The integration between statistics and machine learning can be used to train automated models with imputed missing values in the data, which would improve the generalizability and robustness of models [100].
The integration between statistics and machine learning does not only contribute to data augmentation, but also to medical diagnostics using multi-model data. In the future, this approach together with deep learning methods is suggested to be used in bioinformatics analysis using genomic data or a combination of genomic and clinical data to enhance the automated decision-making process. Deep learning, being one of the unprecedented technical advances in healthcare research, assists clinicians in understanding the role of artificial intelligence in clinical decision making. Hence, deep learning could serve as a vehicle for the translation of modern biomedical data, including electronic health records, imaging, omics, sensor data and text, which are complex, heterogeneous, poorly annotated and generally unstructured, to bridge clinical research and human interpretability [101,102].
5. Conclusions
Conventional statistics are the fundamentals of machine learning, as the mathematical concepts are encapsulated into simplified algorithms executed using computer programming to make decisions. Machine learning has the added benefit of automated analysis, which can be translated into decision support tools, providing user-friendly interfaces based on interactive visualizations and customization of data values. Such tools could assist clinicians in looking at data in different perspectives, which could help them make better decisions. Despite the debate between conventional statistics and machine learning, the integration between the two accelerates decision-making time, provides automated decision making and enhances explainability. This review suggests that clinicians could consider integrating machine learning with conventional statistics for added benefits. Both machine learning and conventional statistics are best integrated to build powerful automated decision-making tools, not limited to clinical data, but also for bioinformatics analyses.
Conceptualization, S.K.D.; data curation, M.D.G.; formal analysis, S.K.D. Methodology, M.D.G., S.K.D., S.S., P.L. and N.A.T.; Project administration, S.K.D. and N.A.T.; Resources, S.K.D.; Software, M.D.G. and S.K.D.; Supervision, S.K.D.; Validation, M.D.G.; Visualization, M.D.G. and S.S.; Writing—original draft, S.K.D., M.D.G. and S.S.; Writing—review and editing, M.D.G., S.K.D., S.S., P.L. and N.A.T. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
The authors thank the breast cancer research team from University Malaya Medical Centre for contributing the case study data.
The authors declare no conflict of interest.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figure 1. The evolution of conventional statistics and machine learning in health research.
Figure 3. Results of variable importance using VSURF and randomForestExplainer R packages to determine the important factors affecting breast cancer survival.
Figure 5. Significance of machine learning to healthcare, education and research and society.
Figure 7. Pipeline to develop automated health research platform with machine learning.
Concepts in hypothesis testing.
Approach | Concept | Procedure |
---|---|---|
Hypothesis testing |
Research question: |
Step 1: Identify predictors from related literature |
Concepts in predictive analysis (classification).
Approach | Concept | Procedure |
---|---|---|
Classification | Research question: |
Step 1: Split dataset into training and testing datasets |
Machine learning alternatives to conventional statistics.
Analysis | Conventional Statistics | Machine Learning |
---|---|---|
1 | Imputations | Imputations |
Objective | To impute missing value based on pattern of missing data. For example, missing by random | To impute missing values to maintain the quality of data |
Method | Missing values are identified on the percentage of missing data, with an acceptance range between (10–20%) | 1. Determine the missing values in the data |
Result | Imputed data/clean data | Imputed data/clean data |
2 | Effect Size | Model Evaluation |
Objective | To determine if the data explain the variability in data. Often called residual error, the residual should be as minimized as possible | To determine the quality of data to be used for further analysis |
Method | Residual analysis, if necessary, standardized residual error is performed using linear regression | 1. Split data into training and testing sets |
Results | Measurable R2 |
Accuracy, sensitivity, specificity, precision, Matthew Correlation Coefficient, area under the receiver operating curve (AUROC) |
3 | Significant Factors | Variable Importance |
Objective | To select important independent variables, which affect the target variable (dependent variable) | To select important independent variables, which affect the target variable (dependent variable) |
Method | 1 Run the analysis using all data |
1. Run variable importance using the best model, which fit the data from model evaluation |
Results | p-value, 95%CI |
Variable importance score/mean (numerical) and variable importance plot (visualization) |
4 | Survival Analysis | Survival Analysis |
Objective | To determine survival rate using time series data | To determine survival rate (%) using time series data |
Method | Similar to machine learning, just the software is different | 1. Specify independent variable and target (survival years) |
Results | Survival rate in percentage, |
Survival rate in percentage, |
Results from significant factor analysis in SPSS.
Variables (Independent) | Total, n (%) | Survival Status (Dependent) | p-Value 1 | |
---|---|---|---|---|
Alive, n (%) | Death, n (%) | |||
Age (years), median | 51 (42, 61) | 51 (43, 60) | 53 (42, 63) | 0.001 |
Marital status | 0.001 | |||
Married | 6397 (81.6) | 4554 (82.5) | 1843 (79.3) | |
Not married | 1443 (18.4) | 963 (17.5) | 480 (20.7) | |
Menopausal status | 0.000 | |||
Natural menopause | 3984 (50.8) | 2675 (48.5) | 1309 (56.3) | |
Pre-menopause | 3347 (42.7) | 2459 (44.6) | 888 (38.2) | |
Surgical menopause | 509 (6.5) | 383 (6.9) | 126 (5.4) | |
Presence of family history | 0.000 | |||
No | 6357 (81.1) | 4378 (79.4) | 1979 (85.2) | |
Yes | 1483 (18.9) | 1139 (20.6) | 344 (14.8) | |
Race | 0.000 | |||
Chinese | 5394 (68.8) | 4041 (73.2) | 1353 (58.2) | |
Indian | 921 (11.7) | 608 (11.0) | 313 (13.5) | |
Malay | 1525 (19.5) | 868 (15.7) | 657 (28.3) | |
Method of diagnosis | 0.000 | |||
Excision | 1617 (20.6) | 1294 (23.5) | 323 (13.9) | |
FNAC | 1886 (24.1) | 1013 (18.4) | 873 (37.6) | |
Imaging only | 35 (0.4) | 31 (0.6) | 4 (0.2) | |
Trucut | 4302 (54.9) | 3179 (57.6) | 1123 (48.3) | |
Classification of breast cancer | 0.000 | |||
Insitu | 366 (4.7) | 348 (6.3) | 18 (0.8) | |
Invasive | 7474 (95.3) | 5169 (93.7) | 2305 (99.2) | |
Laterality | 0.000 | |||
Bilateral | 97 (1.2) | 26 (0.5) | 71 (3.1) | |
Left | 3553 (45.3) | 2464 (44.7) | 1089 (46.9) | |
Right | 3895 (49.7) | 2830 (51.3) | 1065 (45.8) | |
Unilateral side unknown | 295 (3.8) | 197 (3.6) | 98 (4.2) | |
Cancer stage | 0.000 | |||
Pre-cancer | 365 (4.7) | 347 (6.3) | 18 (0.8) | |
Curable cancer | 6624 (84.5) | 4956 (89.8) | 1668 (71.8) | |
Metastatic | 851 (10.9) | 214 (3.9) | 637 (27.4) | |
Tumour size (cm), median | 2.7 (1.6, 4.5) | 2.3 (1.5, 3.5) | 4.00 (2.5, 8.0) | 0.000 |
Total axillary lymph nodes removed, median | 11 (4, 16) | 12 (6, 17) | 9 (0,16) | 0.000 |
Number of positive lymph nodes, median | 0 (0, 2) | 0 (0, 1) | 0 (0,4) | 0.000 |
Grade of differentiation in tumour | 0.000 | |||
Good | 2548 (32.5) | 1631 (29.6) | 917 (39.5) | |
Moderate | 2936 (37.4) | 2259 (40.9) | 677 (29.1) | |
Poor | 2356 (30.1) | 1627 (29.5) | 729 (31.4) | |
Estrogen status | 0.000 | |||
Negative | 3198 (40.8) | 1936 (35.1) | 1262 (54.3) | |
Positive | 4642 (59.2) | 3581 (64.9) | 1061 (45.7) | |
Progesterone status | 0.000 | |||
Negative | 4157 (53.0) | 2603 (47.2) | 1554 (66.9) | |
Positive | 3683 (47.0) | 2914 (52.8) | 769 (33.1) | |
c-er-b2 status | 0.000 | |||
Positive | 1862 (23.8) | 1245 (22.6) | 617 (26.6) | |
Negative | 5148 (65.7) | 3666 (66.4) | 1482 (63.8) | |
Equivocal | 830 (10.6) | 606 (11.0) | 224 (9.6) | |
Primary treatment type | 0.000 | |||
Chemotherapy | 976 (12.4) | 438 (7.9) | 538 (23.3) | |
Hormone therapy | 270 (3.4) | 100 (1.8) | 170 (7.3) | |
Surgery | 6140 (78.3) | 4812 (87.2) | 1328 (57.2) | |
None | 454 (5.8) | 167 (3.0) | 287 (12.4) | |
Surgery status | 0.000 | |||
Surgery done | 6740 (86.0) | 5121 (92.8) | 1619 (69.7) | |
No surgery | 1100 (14.0) | 396 (7.2) | 704 (30.3) | |
Type of surgery | 0.000 | |||
Breast-conserving surgery | 1916 (24.4) | 1661 (30.1) | 255 (11.0) | |
Mastectomy | 4821 (61.5) | 3456 (62.6) | 1365 (58.8) | |
No surgery | 1103 (14.1) | 400 (7.3) | 703 (30.3) | |
Method of axillary lymph node dissection | 0.000 | |||
Yes | 5553 (70.8) | 4048 (73.4) | 1505 (64.8) | |
SLNB (sentinel lymph node biopsy) | 540 (6.9) | 531 (9.6) | 9 (0.4) |
1p-value reported based on chi square test for all categorical variables and Mann–Whitney U test for continuous variables.
Survival analysis comparison using SPSS and R.
Comparison | SPSS | R | |||||
---|---|---|---|---|---|---|---|
Method | 1. Under survival, life table is used to plot the survival curve |
1. Package survival is loaded |
|||||
Results | a. Tumor size | a. Tumor size | |||||
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | ||||||
b. Cancer stage | b. Cancer stage | ||||||
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | ||||||
c. Positive lymph nodes | c. Positive lymph nodes | ||||||
[Image omitted. Please see PDF.] | [Image omitted. Please see PDF.] | ||||||
Interpretation | p-value < 0.05 means it is statistically significant | The survival percentages are extracted from the survival curve to estimate the survival rate of patients. | |||||
Log Rank (Mantel-Cox) | Chi-square | df | Sig. | ||||
Tumor size | 1105.407 | 2 | 0.000 | ||||
Cancer stage | 1721.517 | 2 | 0.000 | ||||
Positive lymph nodes | 234.999 | 3 | 0.000 |
References
1. Vercio, L.L.; Amador, K.; Bannister, J.J.; Crites, S.; Gutierrez, A.; Macdonald, M.E.; Moore, J.; Mouches, P.; Rajashekar, D.; Schimert, S. et al. Supervised machine learning tools: A tutorial for clinicians. J. Neural Eng.; 2020; 17, 062001. [DOI: https://dx.doi.org/10.1088/1741-2552/abbff2]
2. Tonekaboni, S.; Joshi, S.; McCradden, M.D.; Goldenberg, A. What clinicians want: Contextualizing explainable machine learning for clinical end use. arXiv; 2019; arXiv:1905.05134.
3. Rowe, M. An introduction to machine learning for clinicians. Acad. Med.; 2019; 94, pp. 1433-1436. [DOI: https://dx.doi.org/10.1097/ACM.0000000000002792] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31094727]
4. Faes, L.; Liu, X.; Wagner, S.K.; Fu, D.J.; Balaskas, K.; Sim, D.A.; Bachmann, L.M.; Keane, P.A.; Denniston, A.K. A clinician’s guide to artificial intelligence: How to critically appraise machine learning studies. Transl. Vis. Sci. Technol.; 2020; 9, pp. 3-5. [DOI: https://dx.doi.org/10.1167/tvst.9.2.7] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32704413]
5. Wei, J.-X.; Wang, J.; Zhu, Y.-X.; Sun, J.; Xu, H.-M.; Li, M. Traditional Chinese medicine pharmacovigilance in signal detection: Decision tree-based data classification. BMC Med. Inform. Decis. Mak.; 2018; 18, 19. [DOI: https://dx.doi.org/10.1186/s12911-018-0599-5] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/29523131]
6. Cro, S.; Morris, T.P.; Kenward, M.G.; Carpenter, J.R. Sensitivity analysis for clinical trials with missing continuous outcome data using controlled multiple imputation: A practical guide. Stat. Med.; 2020; 39, pp. 2815-2842. [DOI: https://dx.doi.org/10.1002/sim.8569]
7. Austin, P.C.; Fine, J.P. Accounting for competing risks in randomized controlled trials: A review and recommendations for improvement. Stat. Med.; 2017; 36, pp. 1203-1209. [DOI: https://dx.doi.org/10.1002/sim.7215]
8. Austin, P.C.; Cafri, G. Variance estimation when using propensity-score matching with replacement with survival or time-to-event outcomes. Stat. Med.; 2020; 39, pp. 1623-1640. [DOI: https://dx.doi.org/10.1002/sim.8502]
9. Bowden, J.; Del Greco, M.F.; Minelli, C.; Davey Smith, G.; Sheehan, N.; Thompson, J. A framework for the investigation of pleiotropy in two-sample summary data Mendelian randomization. Stat. Med.; 2017; 36, pp. 1783-1802. [DOI: https://dx.doi.org/10.1002/sim.7221]
10. Martin, G.P.; Sperrin, M.; Snell, K.I.E.; Buchan, I.; Riley, R.D. Clinical prediction models to predict the risk of multiple binary outcomes: A comparison of approaches. Stat. Med.; 2021; 40, pp. 498-517. [DOI: https://dx.doi.org/10.1002/sim.8787]
11. Loh, W.W.; Vansteelandt, S. Confounder selection strategies targeting stable treatment effect estimators. Stat. Med.; 2021; 40, pp. 607-630. [DOI: https://dx.doi.org/10.1002/sim.8792] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33150645]
12. Ray, E.L.; Sakrejda, K.; Lauer, S.A.; Johansson, M.; Reich, N.G. Infectious disease prediction with kernel conditional density estimation. Stat. Med.; 2017; 36, pp. 4908-4929. [DOI: https://dx.doi.org/10.1002/sim.7488] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/28905403]
13. Irimata, K.M.; Broatch, J.; Wilson, J.R. Partitioned GMM logistic regression models for longitudinal data. Stat. Med.; 2019; 38, pp. 2171-2183. [DOI: https://dx.doi.org/10.1002/sim.8099]
14. Wu, Z.Y.; Kim, H.J.; Lee, J.W.; Chung, I.Y.; Kim, J.S.; Lee, S.B.; Son, B.-H.; Eom, J.-S.; Kim, S.-B.; Jung, K.H. et al. Long-term Oncologic Outcomes of Immediate Breast Reconstruction vs. Conventional Mastectomy Alone for Breast Cancer in the Setting of Neoadjuvant Chemotherapy. JAMA Surg.; 2020; 155, pp. 1142-1150. [DOI: https://dx.doi.org/10.1001/jamasurg.2020.4132] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33052412]
15. Im, S.-A.; Lu, Y.-S.; Bardia, A.; Harbeck, N.; Colleoni, M.; Franke, F.; Chow, L.; Sohn, J.; Lee, K.-S.; Campos-Gomez, S. et al. Overall Survival with Ribociclib plus Endocrine Therapy in Breast Cancer. N. Engl. J. Med.; 2019; 381, pp. 307-316. [DOI: https://dx.doi.org/10.1056/NEJMoa1903765]
16. Romeo, L.; Loncarski, J.; Paolanti, M.; Bocchini, G.; Mancini, A.; Frontoni, E. Machine learning-based design support system for the prediction of heterogeneous machine parameters in industry 4.0. Expert Syst. Appl.; 2020; 140, 112869. [DOI: https://dx.doi.org/10.1016/j.eswa.2019.112869]
17. Çinar, Z.M.; Nuhu, A.A.; Zeeshan, Q.; Korhan, O.; Asmael, M.; Safaei, B. Machine learning in predictive maintenance towards sustainable smart manufacturing in industry 4.0. Sustainability; 2020; 12, 8211. [DOI: https://dx.doi.org/10.3390/su12198211]
18. Fiebrink, R. Machine learning education for artists, musicians, and other creative practitioners. ACM Trans. Comput. Educ.; 2019; 19, pp. 1-32. [DOI: https://dx.doi.org/10.1145/3294008]
19. Villegas-Ch, W.; Román-Cañizares, M.; Palacios-Pacheco, X. Improvement of an online education model with the integration of machine learning and data analysis in an LMS. Appl. Sci.; 2020; 10, 5371. [DOI: https://dx.doi.org/10.3390/app10155371]
20. Sekeroglu, B.; Dimililer, K.; Tuncal, K. Student performance prediction and classification using machine learning algorithms. Proceedings of the 2019 8th International Conference on Educational and Information Technology; Online, 2 March 2019; pp. 7-11.
21. Leblanc, E.; Washington, P.; Varma, M.; Dunlap, K.; Penev, Y.; Kline, A.; Wall, D.P. Feature replacement methods enable reliable home video analysis for machine learning detection of autism. Sci. Rep.; 2020; 10, 21245. [DOI: https://dx.doi.org/10.1038/s41598-020-76874-w]
22. Odabaşı, Ç.; Yıldırım, R. Machine learning analysis on stability of perovskite solar cells. Sol. Energy Mater. Sol. Cells; 2020; 205, 110284. [DOI: https://dx.doi.org/10.1016/j.solmat.2019.110284]
23. De Felice, F.; Polimeni, A. Coronavirus disease (COVID-19): A machine learning bibliometric analysis. Vivo; 2020; 34, pp. 1613-1617. [DOI: https://dx.doi.org/10.21873/invivo.11951] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32503819]
24. Agne, N.A.; Tisott, C.G.; Ballester, P.; Passos, I.C.; Ferrão, Y.A. Predictors of suicide attempt in patients with obsessive-compulsive disorder: An exploratory study with machine learning analysis. Psychol. Med.; 2020; 52, pp. 715-725. [DOI: https://dx.doi.org/10.1017/S0033291720002329] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32669156]
25. Punn, N.S.; Sonbhadra, S.K.; Agarwal, S. COVID-19 epidemic analysis using machine learning and deep learning algorithms. medRxiv; 2020; pp. 1-10. [DOI: https://dx.doi.org/10.1101/2020.04.08.20057679]
26. Min, B.; Kim, M.; Lee, J.; Byun, J.-I.; Chu, K.; Jung, K.-Y.; Lee, S.K.; Kwon, J.S. Prediction of individual responses to electroconvulsive therapy in patients with schizophrenia: Machine learning analysis of resting-state electroencephalography. Schizophr. Res.; 2020; 216, pp. 147-153. [DOI: https://dx.doi.org/10.1016/j.schres.2019.12.012]
27. Nabipour, M.; Nayyeri, P.; Jabani, H.; Shahab, S.; Mosavi, A. Predicting Stock Market Trends Using Machine Learning and Deep Learning Algorithms Via Continuous and Binary Data; A Comparative Analysis. IEEE Access; 2020; 8, pp. 150199-1501212. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3015966]
28. Clare, S.E.; Shaw, P.L. “Big Data” for breast cancer: Where to look and what you will find. NPJ Breast Cancer; 2016; 2, 16031. [DOI: https://dx.doi.org/10.1038/npjbcancer.2016.31]
29. Schaeffer, C.; Booton, L.; Halleck, J.; Studeny, J.; Coustasse, A. Big Data Management in US Hospitals. Health Care Manag.; 2017; 36, pp. 87-95. [DOI: https://dx.doi.org/10.1097/HCM.0000000000000139]
30. Chen, T.T. History of statistical thinking in medicine. Adv. Med. Stat.; 2015; pp. 3-19. [DOI: https://dx.doi.org/10.1142/9789814583312_0001]
31. Jiang, F.; Jiang, Y.; Zhi, H.; Dong, Y.; Li, H.; Ma, S.; Wang, Y.; Dong, Q.; Shen, H.; Wang, Y. Artificial intelligence in healthcare: Past, present and future. Stroke Vasc. Neurol.; 2017; 2, pp. 230-243. [DOI: https://dx.doi.org/10.1136/svn-2017-000101]
32. ÖĞÜŞ, E. To be Together Medicine and Biostatistics in History: Review. Turkiye Klin. J. Biostat.; 2017; 9, pp. 74-83. [DOI: https://dx.doi.org/10.5336/biostatic.2016-52843]
33. Paramasivam, V.; Yee, T.S.; Dhillon, S.K.; Sidhu, A. A methodological review of data mining techniques in predictive medicine: An application in hemodynamic prediction for abdominal aortic aneurysm disease. Biocybern. Biomed. Eng.; 2014; 34, pp. 139-145. [DOI: https://dx.doi.org/10.1016/j.bbe.2014.03.003]
34. Li, J.J.; Tong, X. Statistical Hypothesis Testing versus Machine Learning Binary Classification: Distinctions and Guidelines. Patterns; 2020; 1, 100115. [DOI: https://dx.doi.org/10.1016/j.patter.2020.100115] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33073257]
35. Rajula, H.; Verlato, G.; Manchia, M.; Antonucci, N.; Fanos, V. Comparison of conventional statistical methods with machine learning in medicine: Diagnosis, drug development, and treatment. Medicina; 2020; 56, 455. [DOI: https://dx.doi.org/10.3390/medicina56090455]
36. Feng, J.-Z.; Wang, Y.; Peng, J.; Sun, M.-W.; Zeng, J.; Jiang, H. Comparison between logistic regression and machine learning algorithms on survival prediction of traumatic brain injuries. J. Crit. Care; 2019; 54, pp. 110-116. [DOI: https://dx.doi.org/10.1016/j.jcrc.2019.08.010]
37. Shameer, K.; Johnson, K.W.; Glicksberg, B.S.; Dudley, J.T.; Sengupta, P.P. Machine learning in cardiovascular medicine: Are we there yet?. Heart; 2018; 104, pp. 1156-1164. [DOI: https://dx.doi.org/10.1136/heartjnl-2017-311198]
38. Ganggayah, M.D.; Taib, N.A.; Har, Y.C.; Lio, P.; Dhillon, S.K. Predicting factors for survival of breast cancer patients using machine learning techniques. BMC Med Informatics Decis. Mak.; 2019; 19, 48. [DOI: https://dx.doi.org/10.1186/s12911-019-0801-4]
39. Bhoo-Pathy, N.; Verkooijen, H.M.; Tan, E.-Y.; Miao, H.; Taib, N.A.M.; Brand, J.S.; Dent, R.A.; See, M.H.; Subramaniam, S.; Chan, P. et al. Trends in presentation, management and survival of patients with de novo metastatic breast cancer in a Southeast Asian setting. Sci. Rep.; 2015; 5, 16252. [DOI: https://dx.doi.org/10.1038/srep16252]
40. Kummerow, K.L.; Du, L.; Penson, D.F.; Shyr, Y.; Hooks, M.A. Nationwide trends in mastectomy for early-stage breast cancer. JAMA Surg.; 2015; 150, pp. 9-16. [DOI: https://dx.doi.org/10.1001/jamasurg.2014.2895]
41. Zhang, B.-L.; Sivasubramaniam, P.G.; Zhang, Q.; Wang, J.; Zhang, B.; Gao, J.-D.; Tang, Z.-H.; Chen, G.-J.; Xie, X.-M.; Wang, Z. et al. Trends in Radical Surgical Treatment Methods for Breast Malignancies in China: A Multicenter 10-Year Retrospective Study. Oncologist; 2015; 20, pp. 1036-1043. [DOI: https://dx.doi.org/10.1634/theoncologist.2014-0281]
42. Sinnadurai, S.; Kwong, A.; Hartman, M.; Tan, E.Y.; Bhoo-Pathy, N.T.; Dahlui, M.; See, M.H.; Yip, C.H.; Taib, N.A. Breast-conserving surgery versus mastectomy in young women with breast cancer in Asian settings. BJS Open; 2019; 3, pp. 48-55. [DOI: https://dx.doi.org/10.1002/bjs5.50111] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/30734015]
43. Vila, J.; Gandini, S.; Gentilini, O. Overall survival according to type of surgery in young (≤40 years) early breast cancer patients: A systematic meta-analysis comparing breast-conserving surgery versus mastectomy. Breast; 2015; 24, pp. 175-181. [DOI: https://dx.doi.org/10.1016/j.breast.2015.02.002] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/25728282]
44. Guo, T.; Fan, Y.; Chen, M.; Wu, X.; Zhang, L.; He, T.; Wang, H.; Wan, J.; Wang, X.; Lu, Z. Cardiovascular Implications of Fatal Outcomes of Patients with Coronavirus Disease 2019 (COVID-19). JAMA Cardiol.; 2020; 5, pp. 811-818. [DOI: https://dx.doi.org/10.1001/jamacardio.2020.1017]
45. Islam, T.; Musthaffa, S.; Hoong, S.M.; Filza, J.; Jamaris, S.; Cheng, M.L.; Harun, F.; Din, N.A.; Rahman, Z.A.; Mohamed, K.N. et al. Development and evaluation of a sustainable video health education program for newly diagnosed breast cancer patients in Malaysia. Support. Care Cancer; 2020; 29, pp. 2631-2638. [DOI: https://dx.doi.org/10.1007/s00520-020-05776-4]
46. Kong, Y.C.; Bhoo-Pathy, N.; O’Rorke, M.; Subramaniam, S.; Bhoo-Pathy, N.T.; See, M.H.; Jamaris, S.; Teoh, K.-H.; Bustam, A.Z.; Looi, L.-M. et al. The association between methods of biopsy and survival following breast cancer: A hospital registry based cohort study. Medicine; 2020; 99, e19093. [DOI: https://dx.doi.org/10.1097/MD.0000000000019093] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32028433]
47. Lim, Y.C.; Hoe, V.C.W.; Darus, A.; Bhoo-Pathy, N. Association between night-shift work, sleep quality and metabolic syndrome. Occup. Environ. Med.; 2018; 75, pp. 716-723. [DOI: https://dx.doi.org/10.1136/oemed-2018-105104]
48. Sinnadurai, S.; Okabayashi, S.; Kawamura, T.; Mori, M.; Bhoo-Pathy, N.; Taib, N.A.; Ukawa, S.; Tamakoshi, A. The JACC Study Group. Intake of common alcoholic and non-alcoholic beverages and breast cancer risk among Japanese women: Findings from the Japan collaborative cohort study. Asian Pac. J. Cancer Prev.; 2020; 21, pp. 1701-1707. [DOI: https://dx.doi.org/10.31557/APJCP.2020.21.6.1701] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32592367]
49. Balakrishnan, N.; Teo, S.-H.; Sinnadurai, S.; Pathy, N.T.B.; See, M.-H.; Taib, N.A.; Yip, C.-H.; Pathy, N.B. Impact of Time Since Last Childbirth on Survival of Women with Premenopausal and Postmenopausal Breast Cancers. World J. Surg.; 2017; 41, pp. 2735-2745. [DOI: https://dx.doi.org/10.1007/s00268-017-4081-9] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/28653143]
50. Birkeland, K.I.; Jørgensen, M.E.; Carstensen, B.; Persson, F.; Gulseth, H.L.; Thuresson, M.; Fenici, P.; Nathanson, D.; Nyström, T.; Eriksson, J.W. et al. Cardiovascular mortality and morbidity in patients with type 2 diabetes following initiation of sodium-glucose co-transporter-2 inhibitors versus other glucose-lowering drugs (CVD-REAL Nordic): A multinational observational analysis. Lancet Diabetes Endocrinol.; 2017; 5, pp. 709-717. [DOI: https://dx.doi.org/10.1016/S2213-8587(17)30258-9]
51. Cheng, M.L.; See, M.H.; Sinnadurai, S.; Islam, T.; Alip, A.; Ng, C.G.; Taib, N.A. The MyBCC Study Group. Adherence rate and the factors contribute toward the surgical adherence of breast cancer in Malaysia. Breast J.; 2020; 26, pp. 568-570. [DOI: https://dx.doi.org/10.1111/tbj.13617]
52. Hedayati, E.; Papakonstantinou, A.; Gernaat, S.A.M.; Altena, R.; Brand, J.S.; Alfredsson, J.; Bhoo-Pathy, N.; Herrmann, J.; Linde, C.; Dahlstrom, U. et al. Outcome and presentation of heart failure in breast cancer patients: Findings from a Swedish register-based study. Eur. Hear. J. Qual. Care Clin. Outcomes; 2020; 6, pp. 147-155. [DOI: https://dx.doi.org/10.1093/ehjqcco/qcz039]
53. Kamada, M.; Shiroma, E.J.; Buring, J.E.; Miyachi, M.; Lee, I. Strength training and all-cause, cardiovascular disease, and cancer mortality in older women: A cohort study. J. Am. Heart Assoc.; 2017; 6, e007677. [DOI: https://dx.doi.org/10.1161/JAHA.117.007677] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/29089346]
54. Lee, D.-C.; Pate, R.R.; Lavie, C.J.; Sui, X.; Church, T.S.; Blair, S.N. Leisure-time running reduces all-cause and cardiovascular mortality risk. J. Am. Coll. Cardiol.; 2014; 64, pp. 472-481. [DOI: https://dx.doi.org/10.1016/j.jacc.2014.04.058] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/25082581]
55. Herrmann, J. From trends to transformation: Where cardio-oncology is to make a difference. Eur. Heart J.; 2019; 40, pp. 3898-3900. [DOI: https://dx.doi.org/10.1093/eurheartj/ehz781] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31761931]
56. Roos-Hesselink, J.; Baris, L.; Johnson, M.; De Backer, J.; Otto, C.; Marelli, A.; Jondeau, G.; Budts, W.; Grewal, J.; Sliwa, K. et al. Pregnancy outcomes in women with cardiovascular disease: Evolving trends over 10 years in the ESC Registry of Pregnancy and Cardiac disease (ROPAC). Eur. Heart J.; 2019; 40, pp. 3848-3855. [DOI: https://dx.doi.org/10.1093/eurheartj/ehz136]
57. Phung, M.T.; Tin, S.T.; Elwood, J.M. Prognostic models for breast cancer: A systematic review. BMC Cancer; 2019; 19, 230. [DOI: https://dx.doi.org/10.1186/s12885-019-5442-6]
58. Chachi, J.; Taheri, S.M.; D’Urso, P. Fuzzy regression analysis based on M-estimates. Expert Syst. Appl.; 2022; 187, 115891. [DOI: https://dx.doi.org/10.1016/j.eswa.2021.115891]
59. Yi, H.-C.; You, Z.-H.; Huang, D.-S.; Kwoh, C.K. Graph representation learning in bioinformatics: Trends, methods and applications. Brief. Bioinform.; 2022; 23, bbab340. [DOI: https://dx.doi.org/10.1093/bib/bbab340]
60. Yu, C.; Liu, J.; Nemati, S.; Yin, G. Reinforcement learning in healthcare: A survey. ACM Comput. Surv.; 2021; 55, pp. 1-36. [DOI: https://dx.doi.org/10.1145/3477600]
61. Recht, B. A tour of reinforcement learning: The view from continuous control. Annu. Rev. Control. Robot. Auton.; 2019; 2, pp. 253-279. [DOI: https://dx.doi.org/10.1146/annurev-control-053018-023825]
62. Prosperi, M.; Guo, Y.; Sperrin, M.; Koopman, J.S.; Min, J.S.; He, X.; Rich, S.; Wang, M.; Buchan, I.E.; Bian, J. Causal inference and counterfactual prediction in machine learning for actionable healthcare. Nat. Mach. Intell.; 2020; 2, pp. 369-375. [DOI: https://dx.doi.org/10.1038/s42256-020-0197-y]
63. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell.; 2019; 1, pp. 206-215. [DOI: https://dx.doi.org/10.1038/s42256-019-0048-x] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/35603010]
64. Raschka, S.; Mirjalili, V. Python Machine Learning: Machine Learning and Deep Learning with Python, Scikit-Learn, and TensorFlow; 2nd ed. Packt Publishing Ltd.: Birmingham, UK, 2017.
65. Talukdar, J.; Kalita, S.K. Detection of Breast Cancer using Data Mining Tool (WEKA). Int. J. Sci. Eng. Res.; 2015; 6, pp. 1124-1128.
66. Yeulkar, K. R Analysis of SEER Breast Cancer Dataset Using Naive Bayes and C4.5 Algorithm. Int. J. Comput. Sci. Telecommun.; 2017; 8491, pp. 43-45.
67. Al-Salihy, N.K.; Ibrikci, T. Classifying breast cancer by using decision tree algorithms. Proceedings of the 6th International Conference on Software and Computer Applications; Bangkok, Thailand, 26 February 2017; pp. 144-148. [DOI: https://dx.doi.org/10.1145/3056662.3056716]
68. Huang, B.F.; Boutros, P.C. The parameter sensitivity of random forests. BMC Bioinform.; 2016; 17, 331. [DOI: https://dx.doi.org/10.1186/s12859-016-1228-x] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/27586051]
69. Jamil, L.S. Data Analysis Based on Data Mining Algorithms Using Weka. Int. J. Eng. Sci. Res. Technol.; 2016; 5, pp. 262-267.
70. Rashmi, G.D.; Lekha, A.; Bawane, N. Analysis of efficiency of classification and prediction algorithms (Naïve Bayes) for Breast Cancer dataset. Proceedings of the 2015 International Conference on Emerging Research in Electronics, Computer Science and Technology (ICERECT); Mandya, India, 17–19 December 2015; pp. 108-113.
71. Reddy, G.T.; Reddy, M.P.K.; Lakshmanna, K.; Kaluri, R.; Rajput, D.S.; Srivastava, G.; Baker, T. Analysis of dimensionality reduction techniques on big data. IEEE Access; 2020; 8, pp. 54776-54788. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2980942]
72. Colgan, R.E.; Gutierrez, D.E.; Sundram, J.; Tenali, G.B. Analysis of Medical Data Using Dimensionality Reduction Techniques. Przegląd Elektrotechniczny; 2013; 89, pp. 279-281.
73. Zebari, R.; AbdulAzeez, A.; Zeebaree, D.; Zebari, D.; Saeed, J. A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction. J. Appl. Sci. Technol. Trends; 2020; 1, pp. 56-70. [DOI: https://dx.doi.org/10.38094/jastt1224]
74. Esmaeily, H.; Dolat, E.; Heidarian Miri, H.; Taji-Heravi, A.; Kiani, O. Reference Values for Serum Total Cholesterol Concentrations Using Percentile Regression Model: A Population Study in Mashhad. Iran. J. Health Sci.; 2019; 7, pp. 26-35. [DOI: https://dx.doi.org/10.18502/jhs.v7i1.1022]
75. Mostafaei, S.; Kabir, K.; Kazemnejad, A.; Feizi, A.; Mansourian, M.; Hassanzadeh Keshteli, A.; Afshar, H.; Arzaghi, S.M.; Rasekhi Dehkordi, S.; Adibi, P. et al. Explanation of somatic symptoms by mental health and personality traits: Application of Bayesian regularized quantile regression in a large population study. BMC Psychiatry; 2019; 19, 207.pp. 1-8. [DOI: https://dx.doi.org/10.1186/s12888-019-2189-1]
76. Bujang, M.A.; Sa’At, N.; Sidik, T.M.I.T.A.B.; Joo, L.C. Sample size guidelines for logistic regression from observational studies with large population: Emphasis on the accuracy between statistics and parameters based on real life clinical data. Malays. J. Med. Sci.; 2018; 25, pp. 122-130. [DOI: https://dx.doi.org/10.21315/mjms2018.25.4.12]
77. Huang, C.; Lv, X.-W.; Xu, T.; Ni, M.-M.; Xia, J.-L.; Cai, S.-P.; Zhou, Q.; Li, X.; Yang, Y.; Zhang, L. et al. Alcohol use in Hefei in relation to alcoholic liver disease: A multivariate logistic regression analysis. Alcohol; 2018; 71, pp. 1-4. [DOI: https://dx.doi.org/10.1016/j.alcohol.2017.08.001] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/29864674]
78. Sinha, P.; Delucchi, K.L.; McAuley, D.F.; O’Kane, C.M.; A Matthay, M.; Calfee, C.S. Development and validation of parsimonious algorithms to classify acute respiratory distress syndrome phenotypes: A secondary analysis of randomised controlled trials. Lancet Respir. Med.; 2020; 8, pp. 247-257. [DOI: https://dx.doi.org/10.1016/S2213-2600(19)30369-8]
79. Sonabend, R.; Király, F.J.; Bender, A.; Bischl, B.; Lang, M. mlr3proba: Machine learning survival analysis in R. arXiv; 2020; 30, pp. 2019-2021. [DOI: https://dx.doi.org/10.1093/bioinformatics/btab039] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33523131]
80. Nemesure, M.D.; Heinz, M.V.; Huang, R.; Jacobson, N.C. Predictive modeling of depression and anxiety using electronic health records and a novel machine learning approach with artificial intelligence. Sci. Rep.; 2021; 11, 1980. [DOI: https://dx.doi.org/10.1038/s41598-021-81368-4]
81. Rehm, G.B.; Cortés-Puch, I.; Kuhn, B.T.; Nguyen, J.; Fazio, S.A.; Johnson, M.A.; Anderson, N.R.; Chuah, C.-N.; Adams, J.Y. Use of Machine Learning to Screen for Acute Respiratory Distress Syndrome Using Raw Ventilator Waveform Data. Crit. Care Explor.; 2021; 3, [DOI: https://dx.doi.org/10.1097/CCE.0000000000000313]
82. Christodoulou, E.; Ma, J.; Collins, G.S.; Steyerberg, E.W.; Verbakel, J.Y.; Van Calster, B. A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models. J. Clin. Epidemiol.; 2019; 110, pp. 12-22. [DOI: https://dx.doi.org/10.1016/j.jclinepi.2019.02.004]
83. Ford, E.; Sheppard, J.; Oliver, S.; Rooney, P.; Banerjee, S.; Cassell, J.A. Automated detection of patients with dementia whose symptoms have been identified in primary care but have no formal diagnosis: A retrospective case-control study using electronic primary care records. BMJ Open; 2021; 11, e039248. [DOI: https://dx.doi.org/10.1136/bmjopen-2020-039248]
84. Afsari, B.; Kuo, A.; Zhang, Y.; Li, L.; Lahouel, K.; Danilova, L.; Favorov, A.; Rosenquist, T.A.; Grollman, A.P.; Kinzler, K.W. et al. Supervised mutational signatures for obesity and other tissue-specific etiological factors in cancer. Elife; 2021; 10, e61082. [DOI: https://dx.doi.org/10.7554/eLife.61082]
85. Matheny, M.E.; Ricket, I.; Goodrich, C.A.; Shah, R.U.; Stabler, M.E.; Perkins, A.M.; Dorn, C.; Denton, J.; Bray, B.E.; Gouripeddi, R. et al. Development of Electronic Health Record—Based Prediction Models for 30-Day Readmission Risk Among Patients Hospitalized for Acute Myocardial Infarction. JAMA Netw. Open; 2021; 4, e2035782. [DOI: https://dx.doi.org/10.1001/jamanetworkopen.2020.35782] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33512518]
86. Roimi, M.; Gutman, R.; Somer, J.; Ben Arie, A.; Calman, I.; Bar-Lavie, Y.; Gelbshtein, U.; Liverant-Taub, S.; Ziv, A.; Eytan, D. et al. Development and validation of a machine learning model for predicting illness trajectory and hospital resource utilization of COVID-19 hospitalized patients—A nationwide study. J. Am. Med. Inform. Assoc.; 2021; 28, pp. 1188-1196. [DOI: https://dx.doi.org/10.1093/jamia/ocab005] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33479727]
87. Pepic, I.; Feldt, R.; Ljungström, L.; Torkar, R.; Dalevi, D.; Söderholm, H.M.; Andersson, L.-M.; Axelson-Fisk, M.; Bohm, K.; Sjöqvist, B.A. et al. Early detection of sepsis using artificial intelligence: A scoping review protocol. Syst. Rev.; 2021; 10, 28. [DOI: https://dx.doi.org/10.1186/s13643-020-01561-w]
88. Sun, Y.; Rashedi, N.; Vaze, V.; Shah, P.; Halter, R.; Elliott, J.T.; A Paradis, N. Predicting Future Occurrence of Acute Hypotensive Episodes Using Noninvasive and Invasive Features. Mil. Med.; 2021; 186, pp. 445-451. [DOI: https://dx.doi.org/10.1093/milmed/usaa418] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33499528]
89. Noorbakhsh, J.; Chandok, H.; Karuturi, R.K.M.; George, J. Machine Learning in Biology and Medicine. Adv. Mol. Pathol.; 2019; 2, pp. 143-152. [DOI: https://dx.doi.org/10.1016/j.yamp.2019.07.010]
90. Ganggayah, M.D. Machine learning on breast cancer prediction. Available online: https://github.com/MoganaD/Machine-Learning-on-Breast-Cancer-Survival-Prediction (accessed on 1 November 2020).
91. Chen, C. Ascent of machine learning in medicine. Nat. Mater.; 2019; 18, 407.
92. De Glas, N.A.; Bastiaannet, E.; Engels, C.C.; De Craen, A.J.M.; Putter, H.; Van De Velde, C.J.H.; Hurria, A.; Liefers, G.J.; Portielje, J.E.A. Validity of the online PREDICT tool in older patients with breast cancer: A population-based study. Br. J. Cancer; 2016; 114, pp. 395-400. [DOI: https://dx.doi.org/10.1038/bjc.2015.466]
93. Hoveling, L.A.; van Maaren, M.C.; Hueting, T.; Strobbe, L.J.A.; Hendriks, M.P.; Sonke, G.S.; Siesling, S. Validation of the online prediction model CancerMath in the Dutch breast cancer population. Breast Cancer Res. Treat.; 2019; 178, pp. 665-681. [DOI: https://dx.doi.org/10.1007/s10549-019-05399-2] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31471837]
94. Islam, T.; Bhoo-Pathy, N.; Su, T.T.; Majid, H.A.; Nahar, A.M.; Ng, C.G.; Dahlui, M.; Hussain, S.; Cantwell, M.; Murray, L. et al. The Malaysian breast Cancer survivorship cohort (MyBCC): A study protocol. BMJ Open; 2015; 5, e008643. [DOI: https://dx.doi.org/10.1136/bmjopen-2015-008643]
95. Pan, I.; Mason, L.R.; Matar, O.K. Data-centric Engineering: Integrating simulation, machine learning and statistics. Challenges and opportunities. Chem. Eng. Sci.; 2022; 249, 117271. [DOI: https://dx.doi.org/10.1016/j.ces.2021.117271]
96. Van Calster, B.; Verbakel, J.; Christodoulou, E.; Steyerberg, E.W.; Collins, G. Statistics versus machine learning: Definitions are interesting (but understanding, methodology, and reporting are more important). J. Clin. Epidemiol.; 2019; 116, pp. 137-138. [DOI: https://dx.doi.org/10.1016/j.jclinepi.2019.08.002] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31425736]
97. Sra, S. Directional statistics in machine learning: A brief review. Appl. Dir. Stat. Mod. Methods Case Stud.; 2018; 225, 6.
98. Eloranta, S.; Smedby, K.E.; Dickman, P.W.; Andersson, T.M. Cancer survival statistics for patients and healthcare professionals–a tutorial of real-world data analysis. J. Intern. Med.; 2021; 289, pp. 12-28. [DOI: https://dx.doi.org/10.1111/joim.13139] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32656940]
99. Pandey, A.K.; Khan, A.I.; Abushark, Y.B.; Alam, M.M.; Agrawal, A.; Kumar, R.; Khan, R.A. Key issues in healthcare data integrity: Analysis and recommendations. IEEE Access; 2020; 8, pp. 40612-40628. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2976687]
100. Shadbahr, T.; Roberts, M.; Stanczuk, J.; Gilbey, J.; Teare, P.; Dittmer, S.; Thorpe, M.; Torne, R.V.; Sala, E.; Lio, P. et al. Classification of datasets with imputed missing values: Does imputation quality matter?. arXiv; 2022; arXiv:2206.08478. arXiv: 2206.08478
101. Jin, D.; Sergeeva, E.; Weng, W.H.; Chauhan, G.; Szolovits, P. Explainable deep learning in healthcare: A methodological survey from an attribution view. WIREs Mech. Dis.; 2022; 14, e1548. [DOI: https://dx.doi.org/10.1002/wsbm.1548]
102. Riccardo, M.; Wang, F.; Wang, S.; Jiang, X.; Dudley, J. Deep learning for healthcare: Review, opportunities and challenges. Brief. Bioinform.; 2018; 19, pp. 1236-1246. [DOI: https://dx.doi.org/10.1093/bib/bbx044]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The practice of medical decision making is changing rapidly with the development of innovative computing technologies. The growing interest of data analysis with improvements in big data computer processing methods raises the question of whether machine learning can be integrated with conventional statistics in health research. To help address this knowledge gap, this paper presents a review on the conceptual integration between conventional statistics and machine learning, focusing on the health research. The similarities and differences between the two are compared using mathematical concepts and algorithms. The comparison between conventional statistics and machine learning methods indicates that conventional statistics are the fundamental basis of machine learning, where the black box algorithms are derived from basic mathematics, but are advanced in terms of automated analysis, handling big data and providing interactive visualizations. While the nature of both these methods are different, they are conceptually similar. Based on our review, we conclude that conventional statistics and machine learning are best to be integrated to develop automated data analysis tools. We also strongly believe that machine learning could be explored by health researchers to enhance conventional statistics in decision making for added reliable validation measures.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 Data Science & Bioinformatics Laboratory, Institute of Biological Sciences, Faculty of Science, Universiti Malaya, Kuala Lumpur 50603, Malaysia
2 Department of Econometrics and Business Statistics, School of Business, Monash University Malaysia, Kuala Lumpur 47500, Malaysia
3 Department of Population Medicine and Lifestyle Disease Prevention, Medical University of Bialystok, 15-269 Bialystok, Poland
4 Department of Computer Science and Technology, University of Cambridge, 15 JJ Thomson Avenue, Cambridge CB3 0FD, UK
5 Department of Surgery, Faculty of Medicine, Universiti Malaya, Kuala Lumpur 50603, Malaysia