1. Introduction
The use of mobile devices while doing daily activities is increasing [1]. These devices have different types of sensors that allow the acquisition of several data related to the user, including the accelerometer, magnetometer, gyroscope, Global Positioning System (GPS) receiver, and microphone [2,3]. These sensors allow the creation of intelligent systems to improve the quality of life. The monitoring of older adults or people with chronic diseases is one of the critical purposes. Furthermore, it can be useful to support sports activities and stimulate the practice of physical activity in teenagers [4]. The development of these systems is included in the research of Ambient Assisted Living (AAL) systems and Enhanced Living Environments (ELE) [5,6,7,8,9,10].
The automatic recognition of ADL is widely researched [11,12,13,14,15,16], where the previously proposed framework [2,17,18,19,20,21,22,23,24,25] was tested and validated with different types of Artificial Neural Networks (ANN) [26,27,28], verifying that the best results were achieved with Deep Neural Networks (DNN). The proposed framework allows the recognition of eight ADL, i.e., walking, running, standing, going upstairs, going downstairs, watching television, sleeping, driving, and other activities without motion, and nine environments, i.e., bar, classroom, gym, hall, kitchen, library, street, bedroom, and living room. This framework uses sensors available in mobile devices [29,30], reporting different accuracies. The proposed architecture is composed of data acquisition, data processing, data fusion, and data classification. The classification module is divided into three small stages, including the recognition of simple ADL, i.e., running, standing, walking, going upstairs, going downstairs, and other activities without motion, with accelerometer, gyroscope, and magnetometer sensors, the recognition of environments, i.e., bar, classroom, gym, hall, kitchen, library, street, bedroom, and living room, with the microphone data, and the recognition of activities without motion, i.e., sleeping, watching television, driving, and other activities without movement.
This research is based on the creation of a framework for the recognition of ADL and its environments. Still, its main goal is related to the testing of ensemble learning methods to further improve the obtained accuracy in the recognition.
The main contribution of this paper is the implementation of different machine learning methods with the same dataset used for the creation of the framework [31], including AdaBoost [32,33] and Instance Based k-nearest neighbour (IBk) [34], using different Java based frameworks, including Weka [35] and Smile [36]. Finally, the results obtained with the different methods should be compared to decide the best method for implementation using the ADL and environment recognition framework.
The results show that the application of the IBk method implemented with Weka software reported better results than others, reporting results with around 77.68% accuracy in recognition of ADL, 41.43% accuracy in recognition of environments, and 99.73% accuracy in recognition of activities without motion. However, AdaBoost applied with Smile also gave important results, reporting results between 85.44% (going upstairs) and 99.98% (driving).
Section 2 gives the presentation of the different methods implemented. The results and the comparative study of this paper are presented in Section 3. Finally, the discussion and conclusions are presented in Section 4.
2. Methods
2.1. Study Design
This study consisted of the use of the same structure and data acquired by the research presented in [18,21,22,24,25] to implement a comparative study between three types of studies. The tests were conducted with the dataset available in [24], which included data related to the eight ADL and nine environments. The information was acquired from the accelerometer, magnetometer, gyroscope, microphone, and GPS receiver available in the mobile device.
As presented in [21], an Android application was used for the acquisition of the data related to the different sensors. This mobile application is responsible for data acquisition and data processing using built-in smartphone sensors such as the accelerometer, magnetometer, gyroscope, sound, and GPS data. The software was responsible for managing five seconds of data every five minutes. It was installed in a smartphone, and it was placed in the front pocket of the pants of 25 subjects with different lifestyles, aged between 16 and 60 years old. For ADL and environment identification, a minimum of 2000 samples with five seconds of data acquired from the different sensors was available in the dataset used for this research. Different environments were used in the performed tests and were strictly related to specific activities. The volunteers had to select the ADL that would be performed using the mobile application before the start of the test. By default, the mobile application did not save any data without user input. However, the proposed method had limitations related to battery consumption and the processing power needed to perform the tests. Currently, the majority of the smartphones available on the market incorporate high performance processing units that can be used to perform the tests, and the main problem is related to power consumption. However, most people usually recharge their mobile phones daily. Therefore, the proposed method can be used in real-life scenarios.
2.2. Overview of the Framework for the Recognition of the Activities of Daily Living and Environments
Based on the previously proposed framework [20], Figure 1 shows a framework composed of four stages, including data acquisition, data processing, data fusion, and data classification. The data processing consisted of several phases, including data cleaning and feature extraction. The data classification was divided into three stages, the recognition of simple ADL (Stage 1), the identification of environments (Stage 2), and the activities without motion (Stage 3). Stage 1 included the use of the data acquired from the accelerometer, magnetometer, and gyroscope sensors. The data received from the microphone were processed in Stage 2. Finally, Stage 3 increased the number of sensors, combining the data acquired from the accelerometer, magnetometer, and gyroscope sensors with the data obtained from the GPS receiver and the environment previously recognised.
Mobile devices are composed of several sensors, which are capable of acquiring different types of data. The framework proposed was capable of acquiring and analysing 5 seconds of data and identifying the current ADL executed and the current environment frequented. The next stage consisted of the processing of the data acquired from the sensors for a further fusion of the different data acquired from the sensors. The final module of the framework consisted of the classification of the data, which started to process all features extracted from the sensors available in the mobile device and identified if the ADL executed was available in the set of ADL proposed. In the affirmative case, the ADL performed was presented to the user. Next, the environment frequented was recognised in the next stage, and it was presented to the user. If no ADL was recognised or the ADL recognized was standing, the identification of a standing ADL would be executed, trying to discover the activity performed by the user.
2.2.1. Data Acquisition
This study was based on the same dataset used in [21], which is publicly available in [31]. This dataset was composed of small sets of data (five seconds every five minutes) captured by the sensors available in the off-the-shelf mobile phones, i.e., accelerometer, magnetometer, gyroscope, microphone, and GPS receiver, and stored in the cloud. The dataset used in the presented study was created using an Android mobile application for data collection. On the one hand, the running and walking data were collected in outdoor environments. On the other hand, standing and going down and upstairs were performed inside buildings.
Moreover, the tests were conducted at different times of the day. In total, thirty-six hours of data were collected, which corresponded to 2000 samples with five seconds of raw sensor data each. Before data acquisition, the user had to use the smartphone to select the ADL that would be conducted and the time needed.
2.2.2. Data Cleaning
Data cleaning is a step performed during data processing. It is mainly used to minimise the effects of the environmental noise acquired during the acquisition of the data from the sensors. Data cleaning methods depend on the type of data acquired and the sensors used. On the one hand, a low pass filter was applied to the data obtained from the accelerometer, magnetometer, and gyroscope sensors [37]. On the other hand, the Fast Fourier Transform (FFT) [38] was used to extract the relevant information from the data collected from the microphone. There were no methods needed to clean the received data from the other types of sensors.
2.2.3. Feature Extraction
After the cleaning of the data, we extracted the features. Table 1 presents the extracted features from the selected sensors, which consisted mainly of statistical features. In Stage 1, the statistical features were mainly used, i.e., standard deviation, mean, maximum and minimum value, variance, and median, of the raw data and the peaks of the motion and magnetic sensors. It also included the calculation of the five greatest distances between calculated peaks. Stage 2 was composed of the feature acquired from the microphone, including the statistical features, i.e., standard deviation, mean, maximum and minimum value, variance, and median, of the raw data, and the calculation of 25 Mel frequency cepstrum coefficients with the microphone. Finally, Stage 3 included also the distance travelled calculated from the Global Positioning (GPS) receiver data and the environment recognised in Stage 2.
2.2.4. Data Fusion and Classification
Data fusion and classification were included in the last stage of the ADL and environment recognition framework. The previous studies reported that the best accuracies were achieved with the DNN method [18,21,22,24,25], and all the features are presented in Table 1. This study presents the results of the test and validation of different methods, including IBk, AdaBoost with the decision stump, and AdaBoost with the decision tree, implemented in the Java programming language for compatibility with Android based devices. The configurations used were different for the different methods implemented. Firstly, the DNN method was implemented with an activation function named sigmoid, which is a function that has the sigmoid curve, widely used as an activation function for neural networks [39]. Several learning rates were previously studied, and it was verified that we obtained better results with a value equal to 0.1. For this method, the maximum number of training iterations was established as 4 × 106. The method was implemented without distance weighting, with three hidden layers, a seed value of six, and backpropagation. The Xavier function [40] was used as an initialization function, implementing L2 regularization [41]. Secondly, the IBk method was implemented with a batch size of 100, a k value of 1, and the linear nearest neighbour search algorithm [42]. Finally, in the last two methods implemented, the main difference was the weak classifier used in combination with the AdaBoost method as the decision stump classifier [43], for the first one, and the decision tree classifier [44], for the second one. Other differences were revealed, where the combination of the AdaBoost method with the decision stump classifier was implemented with a maximum number of training iterations as 10, a seed value of 1, a batch size of 100, a weight threshold of 100, and without resampling. Thus, the combination of the AdaBoost method with the decision tree classifier was implemented with a seed value of 2, a batch size of 10, a number of maximum nodes equal to 4, and 200 as the number of trees.
Initially, we started with the identification of simple ADL, i.e., walking running, standing, going upstairs, and going downstairs, which was performed with the data acquired from the accelerometer, magnetometer, and gyroscope sensors. Secondly, the recognition of environments, i.e., bar, classroom, gym, library, street, hall, living room, kitchen, and bedroom, was performed with the data retrieved from the microphone. Finally, the recognition of activities without motion, i.e., driving, sleeping, and watching television, was performed with the data collected by the accelerometer, magnetometer, gyroscope, and GPS receiver with the inclusion of the environment recognised. Thus, the framework provided the recognition of eight ADL and nine environments.
For the implementation of the methods, the following technologies and frameworks were used:
DNN: DeepLearning4j framework [45];
IBk: Weka software [35];
AdaBoost with the decision stump: Weka software [35];
AdaBoost with the decision tree: Smile (Statistical Machine Intelligence and Learning Engine) framework [36].
3. Results
3.1. Recognition of Simple ADL
The results of simple ADL recognition with the IBk method presented around 80% accuracy using the different combinations of motion and magnetic sensors, as presented in Table 2.
AdaBoost is a binary classifier that uses a weak classier to improve the recognition of different events. The implementation of this algorithm was performed with the identification of each ADL. The results of simple ADL identification with the AdaBoost with the decision stump method implemented with Weka software are presented in Table 3, verifying that all of the ADL were recognised with an accuracy between 25.61% (going downstairs recognised with the accelerometer and magnetometer sensors) and 98.44% (standing recognised with the accelerometer, magnetometer, and gyroscope sensors).
In addition, Table 4 presents the clarification of the values obtained in Table 3, presenting the True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) values. As this recognition was performed as binary recognition, i.e., the comparisons were performed by comparing the correct value with all records, we verified that the values of TP and TN were higher than others, proving the reliability of the method.
Moreover, the results on the recognition of simple ADL with AdaBoost with the decision tree method implemented with the Smile framework are presented in Table 5, verifying that all of the ADL presented an accuracy between 83.79% and 99.55% using the different combinations of motion and magnetic sensors.
Additionally, Table 6 presents the clarification of the values obtained in Table 5, presenting the True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) values. As this recognition was performed as binary recognition, i.e., the comparisons were performed by comparing the correct value with all records, we verified that the sum of the values of TP and TN was 2000. This was the value of the number of samples equal to each activity, but the method reported a high number of FP.
Finally, the results previously obtained with the implementation of the recognition of simple ADL with the DNN method implemented with the Deeplearning4j framework are presented in Table 7, verifying that all of the ADL showed an accuracy between 66.70% and 99.35% using the different combinations of motion and magnetic sensors.
3.2. Recognition of Environments
The use of the IBk method for the recognition of environments using the microphone data reported an average accuracy of 41.43%, as presented in Table 8. The remaining results presented in Table 9 showed that the AdaBoost with the decision stump method implemented with Weka software had an accuracy between 10.36% and 91.78%. Next, the AdaBoost with the decision tree implemented with the SMILE framework reported an accuracy between 88.74% and 99.08%. Finally, the DNN method implemented with the Deeplearning4j framework presented an accuracy between 19.90% and 98.00%.
In addition, Table 10 presents the clarification of the values obtained in Table 9, presenting the True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) values. As this recognition was performed as binary recognition, i.e., the comparisons were performed by comparing the correct value with all records, we verified that the values of TP were higher in the recognition of bar, library, hall, and street. However, in the remaining classes, the values of TN were correctly recognised.
Furthermore, Table 11 presents the clarification of the values obtained in Table 5, presenting the True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) values. As this recognition was performed as binary recognition, i.e., the comparisons were performed comparing the correct value with all records, we verified that the values of TP were higher in the recognition of bar, library, hall, and street. However, in the remaining classes, the values of TN were also correctly recognised.
3.3. Recognition of Activities without Motion
Initially, we presented, in Table 12, the results on the recognition of activities without motion with the IBk method reporting an accuracy between 99.27% and 100% using the data acquired from the accelerometer, magnetometer, gyroscope, GPS receiver, and the environment previously identified.
Furthermore, the results of the implementation of the recognition of activities without motion with the AdaBoost with the decision stump method implemented with Weka software are presented in Table 13 and Table 14, verifying that the events were recognised with an accuracy between 98.32% and 100% using the data acquired from the accelerometer, magnetometer, gyroscope, GPS receiver, and the environment previously identified.
Additionally, Table 15 and Table 16 present the clarification of the values obtained in Table 13 and Table 14, presenting the True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) values. As this recognition was performed as binary recognition, i.e., the comparisons were performed by comparing the correct value with all records, we verified that the values of TP and TN were higher than others, proving the reliability of the method.
Additionally, the results on the recognition of activities without motion with the AdaBoost with the decision tree implemented with the SMILE framework are presented in Table 17 and Table 18, verifying that the events were recognised with an accuracy between 98.50% and 100% using the data acquired from the accelerometer, magnetometer, gyroscope, GPS receiver, and the environment previously identified.
Table 19 and Table 20 present the clarification of the values obtained in Table 17 and Table 18, presenting the True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) values. As this recognition was performed as binary recognition, i.e., the comparisons were performed comparing the correct value with all records, we verified that the values of TP and TN were higher than others, proving the reliability of the method.
Finally, the results of the activity recognition without motion using the DNN method implemented with the DeepLearning4j framework are presented in Table 21 and Table 22, verifying that the events were recognised with an accuracy between 79.55% and 98.50% using the data acquired from the accelerometer, magnetometer, gyroscope, GPS receiver, and the environment previously identified.
Based on the results reported, Table 23 presents the average of the results obtained with the different algorithms implemented. As shown, the best results were achieved with the IBk method (99.68%) and AdaBoost with the decision tree as a weak classifier (94.05%).
The training stage was faster with IBk and AdaBoost with the decision tree than the DNN method previously implemented. These methods were less complicated to implement than the DNN method and were more efficient.
Based on the limitations of mobile devices, these methods should be implemented in the ADL and environment recognition framework to improve the results provided to the user. The results showed that the recognition of ADL and its environments was possible with the implementation of the AdaBoost, IBk, and DNN methods. It allows opportunities to create a personal digital life coach and monitor the different lifestyles. It is important for all people, because mobile devices are widely used. They exploit the possibilities to improve the quality of life.
4. Discussion and Conclusions
The implementations of DNN, IBk, AdaBoost with the decision stump, and AdaBoost with the decision tree were performed with success with the dataset previously acquired, which was based on the data received from the accelerometer, magnetometer, gyroscope, GPS receiver, and microphone. The framework was composed of data acquisition, data processing, data cleaning, feature extraction, data fusion, and data classification, to recognise eight ADL and nine environments.
In general, the overall accuracies of the methods depended on the number of sensors and resources available during data acquisition. The framework should be a function of the number of sensors available in mobile devices. The methods with an accuracy higher than 90% were the IBk method and AdaBoost with the decision tree as the weak classifier.
The AdaBoost and IBk methods reported the best results because these methods were not susceptible to overfitting in comparison with the DNN method. Notably, one of the reasons for this conclusion was the use of a weak classifier by AdaBoost that handled the discrimination of some results.
According to the previously proposed structure of a framework for the recognition of ADL and environments [2,17,18,19,20,21,22,23,24,25], the main focus of this study was related to the data classification module, taking into account the implementations of the other modules performed in previous studies. Previously, the DNN method was implemented, and it reported reliable results. Still, for the recognition of the environments with acoustic data, the results obtained were below the expectations, because it took many resources from the processing unit. For the validation of the different implemented methods, we performed cross-validation with 10 folds.
Following the tests of the different methods for the recognition of simple ADL, the best results were achieved with AdaBoost with the decision tree implemented with the SMILE framework, reporting an overall accuracy of 91.33% with all combinations of sensors. Still, there was a high number of FP. In the case of the recognition of environments, the best method was also AdaBoost with the decision tree implemented with the SMILE framework, reporting an overall accuracy of 99.87%. Still, it did not recognise correctly two environments. However, the AdaBoost with the decision stump method implemented with Weka software did not recognise five environments correctly, reporting an overall accuracy of 32.04%. Finally, in the recognition of activities without motion, the results obtained with AdaBoost with the decision tree implemented with the SMILE framework were the same as the results obtained with the DNN method (99.87%).
As future work, the methods should be implemented during the development of the framework for the identification of ADL and its environments, adapting the approach to all the sensors available on mobile devices.
Author Contributions
Conceptualization, methodology, software, validation, formal analysis, investigation, writing, original draft preparation, and writing, review and editing: J.M.F., I.M.P., G.M., N.M.G., E.Z., P.L., F.F.-R., S.S., and L.X. All authors have read and agreed to the published version of the manuscript.
Funding
This work is funded by FCT/MCTES through national funds and when applicable co-funded EU funds under the project UIDB/EEA/50008/2020 (Este trabalho é financiado pela FCT/MCTES através de fundos nacionais e quando aplicável cofinanciado por fundos comunitários no âmbito do projeto UIDB/EEA/50008/2020).
Acknowledgments
This work is funded by FCT/MCTES through national funds and when applicable co-funded EU funds under the project UIDB/EEA/50008/2020 (Este trabalho é financiado pela FCT/MCTES através de fundos nacionais e quando aplicável cofinanciado por fundos comunitários no âmbito do projeto UIDB/EEA/50008/2020). This article is based on work from COST Action IC1303 - AAPELE - Architectures, Algorithms and Protocols for Enhanced Living Environments, and COST Action CA16226 - SHELD-ON- Indoor living space improvement: Smart Habitat for the Elderly, supported by COST (European Cooperation in Science and Technology). More information at
Conflicts of Interest
The authors declare no conflicts of interest.
Figure and Tables
Figure 1. Flowchart of the ADL and environment recognition framework implemented in this study.
Features extracted.
Sensor | Type of Data | Features |
---|---|---|
Accelerometer Magnetometer Gyroscope | Raw data | standard deviation, mean, maximum and minimum value, variance, and median |
Peaks | five greatest distances between peaks, mean, standard deviation, variance, and median | |
Microphone | Raw data | 26 MFCC, standard deviation, mean, maximum value, minimum value, variance, and median |
GPS receiver | Raw data | distance travelled |
ADL recognition using the Instance Based k-nearest neighbour (IBk) method implemented with Weka software.
Sensors | Correlation Coefficient | Mean Absolute Error | Root Mean Squared Error | Relative Absolute Error | Root Relative Squared Error | Accuracy |
---|---|---|---|---|---|---|
Accelerometer | 0.8335 | 0.261 | 0.817 | 21.8138% | 57.7675% | 73.9% |
Accelerometer and Magnetometer | 0.8771 | 0.2076 | 0.7011 | 17.2911% | 49.5751% | 79.23% |
Accelerometer, Magnetometer, and Gyroscope | 0.8781 | 0.2009 | 0.6991 | 16.733% | 49.4287% | 79.91% |
Accuracies of ADL recognition using the AdaBoost with the decision stump method implemented with Weka software.
ADL | Accelerometer | Accelerometer and Magnetometer | Accelerometer, Magnetometer, and Gyroscope |
---|---|---|---|
Going downstairs | 26.24% | 25.61% | 37.79% |
Going upstairs | 31.73% | 32.64% | 32.91% |
Running | 93.13% | 93.00% | 92.26% |
Standing | 96.35% | 96.58% | 98.44% |
Walking | 37.51% | 51.23% | 50.87% |
Confusion matrix values of ADL recognition using the AdaBoost with the decision stump method implemented with Weka software (TP = True Positive; TN = True Negative; FP = False Positive; FN = False Negative).
ADL | Accelerometer | Accelerometer and Magnetometer | Accelerometer, Magnetometer and Gyroscope | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
TN | FP | FN | TP | TN | FP | FN | TP | TN | FP | FN | TP | |
Going downstairs | 7469 | 1061 | 531 | 939 | 7467 | 1073 | 533 | 927 | 7606 | 1017 | 394 | 983 |
Going upstairs | 7075 | 630 | 925 | 1370 | 7379 | 967 | 621 | 1033 | 7627 | 1498 | 373 | 502 |
Running | 7919 | 81 | 81 | 1919 | 7914 | 82 | 86 | 1918 | 7917 | 97 | 83 | 1903 |
Standing | 7938 | 26 | 62 | 1974 | 7933 | 33 | 67 | 1967 | 7977 | 23 | 23 | 1977 |
Walking | 7472 | 552 | 528 | 1448 | 7629 | 632 | 371 | 1368 | 7609 | 546 | 391 | 1454 |
Accuracies of ADL identification using AdaBoost with the decision tree implemented with the SMILE framework.
ADL | Accelerometer | Accelerometer and Magnetometer | Accelerometer, Magnetometer, and Gyroscope |
---|---|---|---|
Going downstairs | 83.79% | 84.21% | 86.07% |
Going upstairs | 85.29% | 84.70% | 85.44% |
Running | 98.49% | 98.47% | 98.43% |
Standing | 99.04% | 99.01% | 99.55% |
Walking | 86.90% | 89.53% | 91.13% |
Confusion matrix values of ADL identification using AdaBoost with the decision tree implemented with the SMILE framework (TP = True Positive; TN = True Negative; FP = False Positive; FN = False Negative).
ADL | Accelerometer | Accelerometer and Magnetometer | Accelerometer, Magnetometer, and Gyroscope | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
TP | TN | FP | FN | TP | TN | FP | FN | TP | TN | FP | FN | |
Going downstairs | 1017 | 983 | 7362 | 638 | 972 | 1028 | 7449 | 551 | 974 | 1026 | 7633 | 367 |
Going upstairs | 1086 | 914 | 7443 | 557 | 940 | 1060 | 7530 | 470 | 1083 | 917 | 7461 | 539 |
Running | 1917 | 83 | 7932 | 68 | 1917 | 83 | 7930 | 70 | 1908 | 92 | 7935 | 65 |
Standing | 1965 | 35 | 7939 | 61 | 1963 | 37 | 7938 | 62 | 1976 | 24 | 7979 | 21 |
Walking | 1060 | 940 | 7620 | 380 | 1317 | 683 | 7636 | 364 | 1494 | 506 | 7619 | 381 |
Accuracies of ADL identification using the DNN method.
ADL | Accelerometer | Accelerometer and Magnetometer | Accelerometer, Magnetometer, and Gyroscope |
---|---|---|---|
Going downstairs | 66.70% | 67.95% | 77.25% |
Going upstairs | 84.45% | 81.55% | 82.40% |
Running | 95.45% | 95.70% | 95.85% |
Standing | 99.25% | 99.20% | 99.35% |
Walking | 86.10% | 88.05% | 90.09% |
Recognition of environments using the IBk method implemented with Weka software.
Sensors | Sound |
Correlation coefficient | 0.8171 |
Mean absolute error | 0.5857 |
Root mean squared error | 1.5574 |
Relative absolute error | 26.3488% |
Root relative squared error | 60.3156% |
Accuracy | 41.43% |
Accuracies of recognition of environments using the AdaBoost and DNN methods.
Environments | AdaBoost with the Decision Stump | AdaBoost with the Decision Tree | DNN |
---|---|---|---|
Bar | 91.78% | 99.08% | 22.05% |
Classroom | 20.67% | 88.74% | 37.95% |
Gym | 10.36% | 88.87% | 87.85% |
Hall | 40.36% | 92.38% | 34.80% |
Kitchen | 16.11% | 88.89% | 51.35% |
Library | 34.01% | 91.59% | 19.90% |
Street | 38.38% | 90.92% | 25.35% |
Bedroom | 17.88% | 88.88% | 98.60% |
Living room | 18.82% | 89.20% | 33.50% |
Confusion matrix values of the recognition of environments using AdaBoost with the decision stump implemented with Weka software (TP = True Positive; TN = True Negative; FP = False Positive; FN = False Negative).
ADL | Sound | |||
---|---|---|---|---|
TN | FP | FN | TP | |
Bar | 15,961 | 146 | 39 | 1854 |
Library | 15,791 | 1183 | 209 | 817 |
Hall | 15,119 | 645 | 881 | 1355 |
Kitchen | 16,000 | 1999 | 0 | 1 |
Bedroom | 16,000 | 1999 | 0 | 1 |
Street | 15,517 | 1180 | 483 | 820 |
Classroom | 16,000 | 1999 | 0 | 1 |
Living room | 16,000 | 1999 | 0 | 1 |
Gym | 16,000 | 1999 | 0 | 1 |
Confusion matrix values of the recognition of environments using AdaBoost with the decision tree implemented with the SMILE framework (TP = True Positive; TN = True Negative; FP = False Positive; FN = False Negative).
ADL | Sound | |||
---|---|---|---|---|
TP | FP | TN | FN | |
Bar | 1917 | 83 | 15,918 | 82 |
Library | 720 | 1280 | 15,767 | 233 |
Hall | 1419 | 581 | 15,210 | 790 |
Kitchen | 1 | 1999 | 16,000 | 0 |
Bedroom | 14 | 1986 | 15,984 | 16 |
Street | 787 | 1213 | 15,579 | 421 |
Classroom | 148 | 1852 | 15,825 | 175 |
Living room | 168 | 1832 | 15,888 | 112 |
Gym | 1 | 1999 | 15,995 | 5 |
Accuracies of the recognition of activities without motion using the IBk method implemented with Weka software.
Sensors | Correlation Coefficient | Mean Absolute Error | Root Mean Squared Error | Relative Absolute Error | Root Relative Squared Error | Accuracy |
---|---|---|---|---|---|---|
Accelerometer and environment | 1 | 0 | 0 | 0 | 0 | 100% |
Accelerometer, Magnetometer, and Environment | 1 | 0 | 0 | 0 | 0 | 100% |
Accelerometer, Magnetometer, Gyroscope, and Environment | 1 | 0 | 0 | 0 | 0 | 100% |
Accelerometer, Distance, and Environment | 0.9969 | 0.0042 | 0.0645 | 0.6235% | 7.903% | 99.58% |
Accelerometer, Magnetometer, Distance, and Environment | 0.9964 | 0.0045 | 0.0695 | 0.6734% | 8.5118% | 99.55% |
Accelerometer, Magnetometer, Gyroscope, Distance, and Environment | 0.9943 | 0.0073 | 0.0876 | 1.0974% | 10.7201% | 99.27% |
Accuracies of the activities’ recognition without motion using the AdaBoost with the decision stump method implemented with Weka software for motion and magnetic sensors after the recognition of the environment.
Accelerometer and Environment | Accelerometer, Magnetometer, and Environment | Accelerometer, Magnetometer, Gyroscope, and Environment | |
---|---|---|---|
Watching television | 100% | 100% | 100% |
Sleeping | 100% | 100% | 100% |
Accuracies of the activities’ recognition without motion using the AdaBoost with the decision stump method implemented with Weka software for motion, magnetic, and location sensors after the recognition of the environment.
Accelerometer, Distance, and Environment | Accelerometer, Magnetometer, Distance, and Environment | Accelerometer, Magnetometer, Gyroscope, Distance, and Environment | |
---|---|---|---|
Watching television | 98.58% | 98.98% | 98.98% |
Driving | 100% | 100% | 100% |
Sleeping | 98.32% | 98.32% | 98.32% |
Confusion matrix values of the recognition of activities without motion using the AdaBoost with the decision stump method implemented with Weka software for motion and magnetic sensors after the recognition of the environment (TP = True Positive; TN = True Negative; FP = False Positive; FN = False Negative).
ADL | Accelerometer and Environment | Accelerometer, Magnetometer, and Environment | Accelerometer, Magnetometer, Gyroscope, and Environment | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
TN | FP | FN | TP | TN | FP | FN | TP | TN | FP | FN | TP | |
Watching television | 2000 | 0 | 0 | 2000 | 2000 | 0 | 0 | 2000 | 2000 | 0 | 0 | 2000 |
Sleeping | 2000 | 0 | 0 | 2000 | 2000 | 0 | 0 | 2000 | 2000 | 0 | 0 | 2000 |
Confusion matrix values of the recognition of activities without motion using the AdaBoost with the decision stump method implemented with Weka software for motion, magnetic, and location sensors after the recognition of the environment (TP = True Positive; TN = True Negative; FP = False Positive; FN = False Negative).
ADL | Accelerometer, Distance, and Environment | Accelerometer, Magnetometer, Distance, and Environment | Accelerometer, Magnetometer, Gyroscope, Distance, and Environment | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
TN | FP | FN | TP | TN | FP | FN | TP | TN | FP | FN | TP | |
Watching television | 3979 | 0 | 21 | 2000 | 3998 | 13 | 2 | 1987 | 3998 | 13 | 2 | 1987 |
Driving | 4000 | 1 | 0 | 1999 | 4000 | 1 | 0 | 1999 | 4000 | 1 | 0 | 1999 |
Sleeping | 3974 | 0 | 26 | 2000 | 3974 | 0 | 26 | 2000 | 3974 | 0 | 26 | 2000 |
Accuracies of the activities’ recognition without motion using the AdaBoost with the decision tree implemented with the SMILE framework for motion and magnetic sensors after the recognition of the environment.
Accelerometer and Environment | Accelerometer, Magnetometer, and Environment | Accelerometer, Magnetometer, Gyroscope, and Environment | |
---|---|---|---|
Watching television | 100% | 100% | 100% |
Sleeping | 100% | 100% | 100% |
Accuracies of the activities’ recognition without motion using the AdaBoost with the decision tree implemented with the SMILE framework for motion, magnetic, and location sensors after the recognition of the environment.
Accelerometer, Distance, and Environment | Accelerometer, Magnetometer, Distance, and Environment | Accelerometer, Magnetometer, Gyroscope, Distance, and Environment | |
---|---|---|---|
Watching television | 99.67% | 99.97% | 99.97% |
Driving | 99.98% | 99.98% | 99.98% |
Sleeping | 99.52% | 99.52% | 99.50% |
Confusion matrix values of the recognition of activities without motion using the AdaBoost with the decision tree implemented with the SMILE framework for motion and magnetic sensors after the recognition of the environment (TP = True Positive; TN = True Negative; FP = False Positive; FN = False Negative).
ADL | Accelerometer and Environment | Accelerometer, Magnetometer, and Environment | Accelerometer, Magnetometer, Gyroscope, and Environment | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
TP | FP | TN | FN | TP | FP | TN | FN | TP | FP | TN | FN | |
Watching television | 2000 | 0 | 2000 | 0 | 2000 | 0 | 2000 | 0 | 2000 | 0 | 2000 | 0 |
Sleeping | 2000 | 0 | 2000 | 0 | 2000 | 0 | 2000 | 0 | 2000 | 0 | 2000 | 0 |
Confusion matrix values of the recognition of activities without motion using the AdaBoost with the decision tree implemented with the SMILE framework for motion, magnetic, and location sensors after the recognition of the environment (TP = True Positive; TN = True Negative; FP = False Positive; FN = False Negative).
ADL | Accelerometer, Distance, and Environment | Accelerometer, Magnetometer, Distance, and Environment | Accelerometer, Magnetometer, Gyroscope, Distance, and Environment | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
TP | FP | TN | FN | TP | FP | TN | FN | TP | FP | TN | FN | |
Watching television | 2000 | 0 | 3980 | 20 | 2000 | 0 | 3998 | 2 | 2000 | 0 | 3998 | 2 |
Driving | 1999 | 1 | 4000 | 0 | 1999 | 1 | 4000 | 0 | 1999 | 1 | 4000 | 0 |
Sleeping | 1998 | 2 | 3973 | 27 | 1998 | 2 | 3973 | 27 | 1998 | 2 | 3972 | 28 |
Accuracies of the activities’ recognition without motion using the DNN method for motion and magnetic sensors after the recognition of the environment.
Accelerometer and Environment | Accelerometer, Magnetometer, and Environment | Accelerometer, Magnetometer, Gyroscope, and Environment | |
---|---|---|---|
Watching television | 94.05% | 94.00% | 94.15% |
Sleeping | 97.90% | 97.85% | 98.00% |
Accuracies of the activities’ recognition without motion using using the DNN method for motion, magnetic, and location sensors after the recognition of the environment.
Accelerometer, Distance, and Environment | Accelerometer, Magnetometer, Distance, and Environment | Accelerometer, Magnetometer, Gyroscope, Distance, and Environment | |
---|---|---|---|
Watching television | 94.15% | 94.25% | 94.35% |
Driving | 80.65% | 79.55% | 84.15% |
Sleeping | 98.50% | 98.30% | 98.15% |
Average of the accuracy of each implemented method.
Stages | DNN | IBk | AdaBoost with the Decision Stump | AdaBoost with the Decision Tree |
---|---|---|---|---|
Stage 1 | 87.29% | 77.68% | 59.75% | 91.33% |
Stage 2 | 45.71% | 41.43% | 32.04% | 90.95% |
Stage 3 | 99.87% | 99.73% | 92.83% | 99.87% |
Overall | 77.62% | 72.95% | 61.54% | 94.05% |
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2020 by the authors.
Abstract
The recognition of Activities of Daily Living (ADL) using the sensors available in off-the-shelf mobile devices with high accuracy is significant for the development of their framework. Previously, a framework that comprehends data acquisition, data processing, data cleaning, feature extraction, data fusion, and data classification was proposed. However, the results may be improved with the implementation of other methods. Similar to the initial proposal of the framework, this paper proposes the recognition of eight ADL, e.g., walking, running, standing, going upstairs, going downstairs, driving, sleeping, and watching television, and nine environments, e.g., bar, hall, kitchen, library, street, bedroom, living room, gym, and classroom, but using the Instance Based k-nearest neighbour (IBk) and AdaBoost methods as well. The primary purpose of this paper is to find the best machine learning method for ADL and environment recognition. The results obtained show that IBk and AdaBoost reported better results, with complex data than the deep neural network methods.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details






1 Computer Science Department, University of Beira Interior, 6200-001 Covilha, Portugal;
2 Institute of Telecommunications, University of Beira Interior, 6200-001 Covilha, Portugal;
3 Institute of Telecommunications, University of Beira Interior, 6200-001 Covilha, Portugal;
4 Faculty of Computer Science and Engineering, University Ss Cyril and Methodius, 1000 Skopje, Macedonia;
5 Department of Computing Technology, University of Alicante, P.O. Box 99, E-03080 Alicante, Spain;
6 Department of Information Engineering, Marche Polytechnic University, 60131 Ancona, Italy;
7 School of Computer Science, University College Dublin, Dublin 4, Ireland;