Human Activity Recognition Based On Hybrid Learning Algorithm For Wearable Sensor Data - Elsevier Enhanced Reader

You might also like

You are on page 1of 14
Measurement: Sensors 24 (2022) 100512 Contents lists available at ScienceDirect Measurement: Sensors ELSEVIER Journal homepage: worn. sciencedirect.com/journal/measurement-sensors 5 Human activity recognition based on hybrid learning algorithm for wearable sensor data Ravi Kumar Athota’, D. Sumathi”*” "PRD. Reach Scar, Schoo of Computer Sco end gine VIFAP Unies, Vier AP, nda " asocte Profesor Gree, School of Comper Selous and Brgeeig, VIEAP Url; Wyawods, AP, na LE INFO AgstRact ower uaa activity reson yb Teaming orth Convio nea newark ‘onvelton moray fsion agri ‘onvotion gated sion gor Human Activity Recognition (HAR, bsed on sensor devies and the Internet of Things (oD, atracted many researchers since ithasdiversied application in heal sectors, smart envionment, and entertainment HAR has emerged as one of the important health monitoring sppllestions and Wt necessitates the constant usage of smartphones, smartwatches, and wearable devices to capture patients’ daily activities. To predict multiple human ctvitis, deep leaing (DL)-based methods have been sucessfully applied to time-series data that ae generated by smartphones and wearable sensors. although DL-based approaches were deployed in activity cognition, they ll have encountered afew lstes when working with ine-seres data, Thowe less could be ‘managed with the proposed methodology. This work proposed a couple of Hybrid Learning Algorithms (HLA) to buld comprehensive clasiieation methods for HAR using wearble sensor daa. The aim ofthis work is to make tse ofthe Convolution Memory Fusion Algorthn(CMEA) and Convolution ated Fusion Algocthm(CGFA) that, ‘model learns both local features and long-term and gated-erm dependencies In Sequential data. Feature ‘uration ha Ben enhanced with the deployment of various ltr ses, They are used to capture diferent eel temporal dependences, and thus the enkancement is implemented. This Amalgam Learing Model has been Aeployed on the WISDM dataset, andthe proposed models have achleved 97.76, 94.98% for smarowatch and smartphone of CMFA, 96.91%, 64.35% for smartwatch and smartphone of CGFA. Experimental eau show that ‘hese models demonstrated greater accuracy than other existing deep neural network frameworks. 1. Introduetion both online and offline using sensory observation data It can be used to learn more about an individual, such as their gender, age, ete, and the ‘Human activity recognition (HAR) plays significant part in people's daily lives since it can learn advanced information about human actions from raw sensor data [1] Since the rise of human-computer interaction (HCY applications, HAR technology has become a prominent research fa both domestically and internationally. Automated classification ‘and extraction of human body information from daily activities could lay the groundwork for future intelligent applications based on machine Jeaming and natural language processing. This technology was formerly uulized extensively in applications related to ambient aided living (AAL) [0], suchas intelligent aetivity monitoring systems meant fr elderly and disabled persons in healtheare systems [1], automatic interpretation of hhand gestures in sports [, user identity verification for security sys- tems utilizing galt characteristic [5], and human-robot interactions via gesture recognition [6], tis possible to discover what a person ora group i curently doing, * Corresponding autor context in Which observed interactions occurred. HAR systems can be categorized according to the modality of sensory data usd [7,5] since the sort of functionality, algorithms, designs, and methodologies used for research are greatly impacted by the type of sensory data used. Different forms of HAR system research are based on different types of| Inertial sensors, such as IMUs, and signal strength derived from com- modity networks in the surrounding region [2]. For example, HAR systems based on visual information (images and videos) are one type of HAR system study, Sensor-based human aciity recognition [10] is the subject of current study. An sutomated HAR design collects data from sensors that are affected by the motion characterstes of various body Joints. Asa result of these measures, a number of attributes are collected and utilized to train activity models, which are then applied in the process of identifying these activities [11,12]. Dally actions including eating walking, washing, and geting dressed are good examples ofthis FBmall adéreser:aotaravikumar gail com,eavikumar 20ph47109@ivitp. sn (RAC ABRs), sumsth eesaeeh28¢@pmal com (D, Samat) nups//doL.or/10.1016/)measen. 2022 100512 Received 14 July 2022; Accepted 5 October 2022 Availabe online 14 October 2022 2665-9174/0 2022 The Authors. Published by Elsevier Ltd. This san open access arile under the CC BY-NCND license (hi: /ereatvecommons org/lienses/y end/4.00, 1. Approaches for human activity reogniion. type of activity [13]. Sensory measurements are used in a variety of ‘ways to identity these behaviors, which are all based on different ap roaches [14]. Innovations that emerges exclusively inthe field of wearable sensors, ‘and IoT devices have made the life of people more comfortable. A va ety of attempts must be taken to ensure that the population un derstands what they are doing and how a technique can assist them in ‘completing their assignment, but this can be achieved, Smartwatches, smartphones, and ftnes trackers that have multiple sensors can be used for HAR, according to research (15].Ithas been observed that HAR has @ _great potential o play a significant pat in strengthening the progress of ‘the human quality of life. HAR based on sensors focuses on predicting the activities based on low-level sensor data, Our study estimates the participant's activity using the WISDM dataset (16] and the signal in- formation obtained from these sensors. In general, HAR can be thought ‘of as a traditional Pattern Recognition (PR) problem, a common IoT application that can be used for good observational multiple body ‘movement for human action recognition. The importance of HAR based on intelligent sensors is evident, because it is used to detect and track numerous everyday activities such a eating, drinking, cleaning teeth, and detecting bedtime anomalies, ‘which are not limited to aerobic activities, Simple and complex behavior ‘can be found in HAR. Studies on the recognition of complex human activities, such as brushing teeth or moving away from a ball, are rare [17], Accelerometers are used to deteet moving objects by measuring ‘eceleratione along a reference axis. Tis allows the accelerometers to Identify the velocity ofthe objects, Running and resting are particularly ‘well-suited to tracking by these devices because they include a lot of ‘body movement [15,19]. The aeceleromerer's measurements can be utilized to detect abrupt changes in motion. The gyroscope, which em. ploys gravity to compute orientation, i another sensor that has proven popular in smart-wearable devices. The signal data from the gyroscope ‘can be studied to identify the device's placement and alignment (20) ‘There has been a significant amount of research on SHA recognition, but ‘relatively litle on CHA recognition [21]. To be viable, both domains ‘must address a wide range of essential challenges (identity accuracy, privacy, energy consumption, computing cost, mobility), ‘Machine learning's HAR with annotated data isa diferent kind of time series categorization and supervised training than other ap: proaches, Several studies have been conducted to explore the task of recognition and classification utilizing different algorithms such as XGBoost, Random Forest, SVM, and others and quasi-deep learning ‘methods such as LSTM, CNN, ANN, RNN, and others. Existing ap- proaches have the drawback of requiring a substantial amount of time ‘consuming extraction of features and manual feature engineering. On the other hand, deep learning techniques can immediately recognize features from data and are better suited to trying to identity complex ‘human activities Meanremen Semon 4 (2022) 100512 111. Contributions ofthis work ‘The ambition ofthis research isto develop a high-accuracy hybrid learning algorithm capable of categorizing 18 human activities that are well recognized and relatively straightforward into two groups using data from the WISDM dataset: 2) Genera activities that donot focus on the ands, such as walking, jogging, and other similar activities. b) ‘Activities that focus on the hands in general, such as typing, texting, eating cereal or ramen noodles, et. It is derived from lower computa tional series sensor data acquired from triaxial gyroseopes and acceler- ometers in smartphones and smartwatches that could capture the motion and angular momentum of human body actions. 1. A wearable sensor-based hybrid learning algorithm has been pro: posed for HAR. This HLA model was deployed in the process of analyzing and classifying data pertaining to simple and complicated human activities, 2, Sequence data is processed through proposed CMPA and CGFA ‘models, each layer ofthe model witha diferent filter size to capture local and spatial features and long-term dependencies, 3, An accelerometer-gyroscope-based smartwatch and smartphone were texted to find the most effective modal fr activity recognition 4, We investigated the influence ofa variety of factors onthe evaluation ‘outcomes (bidirectional method, filter sizes, and integration with convolutional layers) Convolutional Neural Networks (CNN) can be combined with BISTM, used for long-term dependencies in sequencing data, to detect simple and complex human activities with high precision and accuracy. The remaining portion of the paper is arranged as follows: The sec- tion that follows conducts a systematic review of ether the algorithms and techniques used in sensor-based activity recognition, including the proposed methodologies objectives. Section ® describes the proposed frameworks for activity recognition of humans used inthis work, aswell as the deployment, methodologies, and testing procedure. Section 4 summarizes and examines the findings by comparing the outputs of the various models in terms of performance measures. Finally, Section 5 summarizes the completed work as well asthe future steps. 2, Literature review Various methodologies have been adopted for determining human activities, and iis still a challenging task for researchers to analyze the complex and multitasks thet humans perform. This section briefs the ‘gap analysis, methodologies, and the shortcomings along withthe per- formance asessments, 2.1, Human aetviy Every day, people engage in a wide range of activities, There are two sorts of physical activity that ean be done. The simple motion would be the frst activity, and the complex activity would be the second. The ull. body motor activity incorporates human body movements, such as walking, resting, diving, star climbing, or waiting, The second eate- sory, complex activity, is relevant human functions such as reading, using a computer, watching television, playing sports, preparing food, or cleaning the house. There was a lot of research on simple human action recognition but not much on complex human action recognition. 122. Approaches for Human Activity Recognition In HAR, three types of strategies are used. As shown in Fs. 1, these tare computer vision-based, environmental sensor-based, and time geography-based. Sensor-based approaches are further clasified into three types: wearable [23], object tagged, and dense sensing (241. The HAR is implemented using, computer vision from a series of Meanremen Semon 4 (2022) 100512 ‘Table 1 {Comparative Analysis of Human Activity Recognition Dataset and Algorithm based on Previous Studies, ference Casier Devices Sensors ‘Accuracy Contbuton [Random Smarwatch At 77 ‘ier Hencieaon Based on wring stv performed In ale [GANSTA Smaphone Ace2Gr01 9457 ser identaton aed on yoamic ction eet by srniphone senor (walking pei, aig doa and wag) baw Aectore1 9075 [enfant wer tee th aping aceon the sre ofthe aie device sing te mtion sensor date 5) Random Smartphone Acct Gyo 927 [detieaton of wes thro cal opiate ol Random Siphone ced God 938 Uses an bien as nth wang habits thnks othe we esembe machine Foret lene [oo] Rapdom——Smarvatch cet Gyoa 9 dentton ofthe er sed om static characteris and characteris reaing to Foret human bebavor extrac rom enor dts To] Dilon——Smatvatch Acct Goa 9868 ence fhe er aed onthe war's ambultry ete ough the we of machine Tree Map lene fe aw Saighooe Acel wo 6841 ‘ser entiation based on walling ates sing dese conven Map lckwoek INN ‘photos or videos in which one or more participants execute the activities [25,26]. When it comes t0 computer vision applications, HAR has become one of the most exciting new development. It necessitates the projection of eameras to a target region to capture images and videos (27), Utlizing visual sensors has some drawbacks and limitations. The primary concern is confidentiality, as cameras may not beable to instal, ‘everywhere, and also due to laws and regulatory requirements (24. Tals approach has some drawbacks, including confidentiality, flexibility, and ‘computation cost. The recording of video or the capture of images may infringe on the privacy of the user. The approach lacks mobility and pervasiveness due to the cameras fixed place. Recognizing objects and ‘activites from photographs or a series of images requires computer ‘power, which is computationally expensive. Environmental sensors are also referred to as dense sensing sensors ‘and devie-free sensors. Afloor-mounted sound sensor, for example, or a oom-mounted light sensor, RFID (Radio Frequency Wentifeation Technology) as an entrance tag, and the signal quality of wireless de- vices (WII, ZigBee, Bluetooth, Radio, FM & Microwave-based) are some sensors and signals used to detect various activities [25-1]. Such device‘ree sensors are placed in the surroundings and record the Ihumans’ activities, movements, and inteaetions. RFID-based sensors the most common of these types of sensors. In wearable sensor approaches, motion detectors (accelerometers, ‘and gyroscopes) are connected to various parts of the body, which can bee used to locate and identify human actions. Such sensors could be Integrated with bands, wearabes, garments, and smartphones or known to wear on the body. Magnetometers, Accelerometers, Gyroscopes, Electrocardiography (EO), Electromyography (EMG), and a multitude fof loT-based sensing applications that are secure and can also be worn ‘on the body to record motions are available on the market (32). The placement of sensors at a key position directly impacts a person's ac- tivity data. Sensors are most typically found near the waist, lower back, ‘and breasts. The closer the devices are tothe center of mass, the more accurate the gesture representation In this situation, smartwatchies and “smartphones with gyroscopes and accelerometers come in handy. When ‘esmartwatch or smartphone is worn on the dominant hand, the gadget, may eapture the small movements of daily actions such as cleaning one's teeth or slurping soup. Because of temporal geography techniques (9,34), human behav: lors can be detected using time and place information separately. Hlagerstrand [35] invented the term “time geography” to characterize this concept after belng one of the first to postulate that human activities are limited not only by place but also by time. Its crucial to recall that, ‘ashe acknowledged and emphasized [3°], the individual is significant research component in human activity. Humans were accustomed to and liked to repeat routines based on multiple cycles. The circadian clock, ‘weekly plans, festive season, and yearly holidays are examples of such cycles, When a person’s location and time [2,37] of the day are ‘well-documented, i is possible to estimate their actions based on their particular timetables, The deduction becomes more specific as the granularity of the time/activity data Increases. Environmental health [11,¢2), animal monitoring, and automobile [28.0] systems design are jus few ofthe industries that are nour utilizing this technology. Unlike location data, which is accessible to any compute, time data is aces- sible to any computing device. tis easier to keep track of things outside than indoors, where accuracy and dependability are missing. By locating the exact location, improved activity recognition can be accomplished. 23. Algorithms for human activity recognition Recently, ithas been proven that DL frameworks can perform well in a wide range of areas, including object detection, lasification, natural language processing, and more, Wit the help of Di-based algorithms, it is possible to learn features from large datasets with no human inter- ‘vention automatically. ‘The publishers of [19] used an accelerometer to coleet motion data (arm motions) from four distinet human subjects for only one activity Making a cup of tes. Different hand motions were predicted using a CNN method that they developed and achieved 99.8%, While the authors of [44] proposed an improved method for realtime activity detection using CNN architecture, they found that the accuracy was lower than we expected. According to researchers, data from time series analysis can be saved worldwide by combining CNN and stitial features. ‘A new attention-based HAR method was developed by the re- Searchers in Ref. [15] to cope with the weekly labelled activity data collected from wearable sensor devices, Convolutional and fully con nected layer convalutional and global feature compatibility wae deter mined using the typical CNN pipeline and atention submodules ‘An activity recognizing DL architecture has been proposed in Ref, [46] that uses focal los instead of global loss. The sensor-based activity recognition framework’s memory requirements were signifi cantly educed by this method. The researchers in Ref [47] devised a similar approach to acquire the loca interconnections and scale partially invariant of mobile sensor data by using Convolution and they conclude that using more activity data to conduct an experiment to achieve more performance. The authors [48] created stacked LSTM architecture based on smartphone sensor data to ereate effective classifiers for human actv- ities, they did not incorporate the temporal attention in the LSTM rbwork. In [1, the authors suggested a smartphone sensor-based BI-LSTM, eh Meanremen Semon 4 (2022) 100512 Activity Classification => Fig. 2 Overall pipeline ofthe work ‘table Summary ofthe dataset ‘Sibets| 3 Sensor poling ete 2am Soriponeatlaed ‘Google Neus 55x or Samsung olny 55 ‘table 3 Detaled information on the activty- based datasets, Damset—Gatpory Ay aw Sasor Peete Despont oo. HARS Olea Gating enn 167686 5.45 a Dating 1918s 5.5 (Gasket bln Desking from 19701775 cp ting Paste 188609 S31 Trine wis Sa ng chips 192085558 ing Sandwich 190191553 Nolin Gower 191555557 ‘at FldingCotes 19857362 Downstale sing 195050867 Siancng 19010980 network, which is more accurate for simple activities. [As clescribed in Ref. [50], HAR's deep residual LSTM could benefit, from a bidirectional connection between forwarding and backward states. Improvements in the spatially and temporally dimensions were shown, as well as an improved recognition rate, as result ofthis work. By combining CNN and LSTM, the authors in Ret. [51] were able to improve the forecast performance of human activities using raw sensor data in their Spatiotemporal sequence forecasting. [52] An LSTM Convolutional neural network comprising. LSTM layers and convolutional layers was used to construct HAR. An average ‘pooling layer was used to reduce the model parameters. The concept of broad hybrid CNN/LSTM-based HAR can benefit multi-modality wear ‘ble sensors. Deep learning authors were motivated to use ensemble earning methods because they saw the advantages of machine leaning. Using CNN, LSTM, and attention models, the authors in Ref. (5°] proposed an RNN-based attention network (RAN). When using wearable sensors to collect activity samples this model helped to locate diferent {ypes of multiaetivty. Benchmark datasets for detecting human activity in images eapcured, by uncrowed serial vehicles are presented in (UAV) [54]. Directed transformer [3D models were also used to identify actions, 24, Research gap analysis According ta what we found inthe review ofthe relevant literature in “Inble 1, most researchers have more success with the more specific activities involving the data repository. On the other hand, fewer re- searchers have attempted the more complex activities, but they have not been successful. AS a result, the authors now have the motivation to distinguish complicated activities with high accuracy in their pre dictions. Keeping the set of variables a low as feasible was one of the key objectives in constructing the model. The workflow is depicted in Fig. 2 asa series of stages, each of whichis based on raw accelerometer and gyroscope values obtained from the embedded sensor in the smartphone. 25. inferences In order to recognize complex human behaviors such a eating, drinking, brushing teeth, dribbling, and so on, this approach must be able to discem both geographical and temporal elements from raw sensor data. Convolution layers and Bi-directional Long Short-Term Memory are combined with caution to achieve this. 3. Materials and methods 3.1, Description of WISDM-HARB dataset ‘The data for this research has been obtained and available publicly 2s the WISDM-HARR Smartphone and Smartwatch Activity and Biometrics Dataset [22] therein UCI Machine Learning Repository. This dataset is uunigue in thatthe actions to be categorized include eating pasta, bur- gets, crumpling clothes, and cleaning up, making it adaptable in areas where other data sources that are restricted to exercise oF mobility do not. Activities include standing, walking, kicking, stairwells, writing, applauding, brushing teeth, jogging, folding clothes, moving the ball, laying to catch up, typing, eating pizza, eating chips, swigging, eating ramen noodles, and slurping soup. Sensor data will be saved in four distinct data folders for sensor type (gyroscope, accelerometer) and each gadget. Each folder contains S1 documents, each of the SL attendees, Which should be merged using suitable frames. The predictors are the gyroscope and accelerometer sensor readings inside the x,y, and z as pects, and the variable seems to be the action code. The dataset’s uniqueness lies in the fact that the actions to be Fig. 8. Proposed convolution memory fasion algorithm (CMA). ‘categorized are not restricted to only motion. Other datasets that focus ‘on exercise or motion do not have categories ike “eating pasta,” “eating ‘2 hamburger,” “folding laundry,” and “brushing tecth.” 51 people participated in this dataset, all of whom had a Google Nexus or Samsung Galaxy $5 phone in their pocket and wore a smartwatch on their left Wrist. Their right pocket was stuffed wih the phone upright, and the Meanremen Semon 4 (2022) 100512 screen turned away from them. It specifies that for each ofthe eighteen activities that the attendees perform, 20 data points have been regis tered per second for 3 min, Table 2 summarizes the data’s general characteristics as well asthe methods used to collect them, Describes the properties of both actvity-based datasets in Tie 3 32, Preprocesing and wansformation of data [As an Initial step in the pre-processing process, the datasets have bbeen combined into a single fat file. In this work, two data frames for each device (smanwatch, smartphone) have been created and used. The data from the accelerometer and gyroscope sensors will also be com- bined in every frame. Participants’ data were fist acctmilated into SL frames based on the date and time fr each ofthe sensors (accelerometer and gyroscope). Late, it is considered individually for smartphone and smartwatch. “Max pooling layer as been introduced between the layers of hybrid ‘models to mitigate noise inthe dataset that may be atrbuted to human activity, Pre-processng functionalities like removing missing values and other iregular values have been removed from the datasets accordingly Data imbalance has been managed by configuring all 18 human activ ites that could be predicted by a smartwatch, smartphone, and gesture dataset are evenly distributed among the data points. To segment, the data, a sliding window of 10 s, a step size of fity percent, and an overlap of fifty percent are utilized. This is the initial stage in all deep learning models that have tobe completed. To train and validate the deep learning methods used for this task, as well as to optimize the hyperparameters, the datast is partitioned into three ses: 2 training dataset, 2 validation dataset, and 2 testing dataset. These datasets will then be evaluated on a final test dataset to produce the finished research results, ‘Loss function: The categorical erostentropy los funetion(L) isthe loss function employed inthis study. Calculated fss foreach class label per observation, as shown in equation (1), implies that the results are added. The variable N denotes how many scaler values there are in the final output, p(x) is the probability estimate of the model that the i-th part ofthe sensor data is L=-31 / Nog: n) a Optimizer: Because there isa requirement for the optimization of the loss function, the Adam optimizer is utilized. This optimizer is wel- known for its quick and efficient performance with minimal memory sage. 23, Systeme models ‘A) Convolution Memory Fusion Algorithm (CMEA) Extinction studies and deep learning-based hybrid models are used in, this research to monitor and recognize human activities in an indoor environment. In this section, the suggested work's intemal structure and comparative analysis are described briefly. Pig. 2 depicts the network architecture of the CMFA, which can be found. There are a total of| thirteen layers to it In the beginning, the data that has been pre- processed are introduced into a 1D CNN layer. This layer is respon- sible forthe extraction of spatial features It has three 1D CNN layers, ‘max pooling layers, and filters of each convolution layer 128, 64, and 32. Besides, two layers of BLSTM are used for extracting temporal fea- tures and thas iter size of 128 each, The model's las stage consists of| 1 dense layer, which is immediately followed by layer normalization GN), Finally, the classification layer’s output (Softmax) is used to extract some information from the model, which ereates a probability distribution over the clases. CCNN’s prominence has grown due tits capacity to learn distintive Meanremen Semon 4 (2022) 100512 | Dense, Kernel (6x6) Bais (6), SoftMax_ Fig. 4. Proposed convolution gated fasion algottho (CGFA) ‘representations from images or voices, The convolutlon, which uses convolutional ites to convolute the inputs, is perhaps the essential unit in CNN. As shown below, itserves as filter prior to Being triggered by @ nonlinear ativation function, (5, m= em an+8) ® ‘Where (iar the neurons that corresponding activation Is Ay, Wan I the weight matric (mn) of the convolution kere, X;. nj +n denotes ‘the neuron that connected aetivation ofthe upper neuron, Bi the bias, ‘and isa non-linear function. The 1D convolution layer employed witha ‘gaussian error linear unit (GeLU) to calibrate the feature maps is also a ‘non-linear function defined below. eee . Sever / 9) (081.U()=054(1-+1nh(VE /a-+00un15)) ” Ai ‘The convolutional filter is used in dhree convolution layers. The ‘numerous filter sizes that are employed inthe suggested model are able ‘to describe a variety of temporal local dependencies, which allows for the temporal locality characteristics to be captured, italy, the inputs fed to the ist convolution 1D (Conv1D) layer, as shown in Fig. 5, t0 extract the locality features from the sequential data. By utilizing local receptive fields, CNN can obtain local features fom, data. To train the CNN, the data is fed in the form of time series with dimensions (MN), where M indicates the data’ length and N indicates its numberof features, The Time Distributed Conv layer accept input data, and it has four types of magnitudes: segments, sequences, steps, and features, Sequences represent the total numberof subsequences in a ‘indo, steps represent the numberof steps in each subsequence, and features represent how many features are in the datasets that were used for the segmentation. There are four subssequences of length 64 in the WISDM dataset, ‘being used in this study, with each sample being 256 bytes long. To Introduce non Jneartis in Conv1D layers, a GeL.U activation function is used, In order to prevent overfttng, an extra dropout layer with a value of 0. is added after the third Conv1D layer. This layer is then used to reduce the numberof dimensions by using a 1D Max Pooling layer with pool size 2. Long-term dependencies in input sequences can be captured by using BLSTM As shown in Pg. 3, the results ofthe BLSTM layers are fed into @ dense layer. As a result of the training procedure, which involves continuously modifying weight values in the top layer, the input data Aistribution ofeach layer would vary frequently. Asa result the weight Meanremen Semon 4 (2022) 100512 input Inpattayer eater Papa [eNone, Noo 50,6) | [None None 50, 6) ¥ ‘TineDisuibued(Conv1D) (one, Nove, 50,6) (one, Nove, 48,128) J TimeDisuibuted(MaxPooling!D) a (None, None, 42, 128) | (None, None, 24, 128) : se ‘TumeDistibuted(MaxPooling1D) | “P| one, None, 2, 64) | (None, None, 11, 64) come ‘TimeDistibuted(Coav1D) aa ‘None, Nowe, 11, 64) | (None, None, 9, 32) i om oan | Ti ta 9.38 one A992 ‘TineDisuibute(MaxPooing1©) |" | qyone, Nowe, 9,32) | (one, Nowe, 4, 32) = ' ‘TmeDisuibute(Flate) oo (one, None, 4,32) (one, None, 128) le = ipa Bidirectional STM) oo Crm | 255 | inpat up Bidiecsonal. STM) (one, None, 128) | (None, 256) inpee ouput [smc (ion 23 86) (None, 256) | (None, 512) J a (on ue 1,512) | (one, 128) | ayeNormalzation (one, 128) | (one, 128) | inpet uur Dense oo (a ne, 128) | None, 3) Fig 5. CMEA Configuration, parameters must be changed to adapt to this different distribution, ‘which complicates the training process and slows convergence time. In ‘order to solve this problem, th layer normalization (LN) step is added ‘afte the dense layer in order to hasten the convergence ofthe model. In ‘order to improve accuracy and training speed, the LN layer first rere ‘ates the input data on training samples before normalizing them. This is ddone to preserve the integrity ofthe output ofthe layer that came before it. The final layer, known as the softmax function layer, uses the normalized output ofthe layer that came before it, which was dense, to Predict the input class, The Adam optimizers wed to train the proposed framework to decrease classification cros-entropy loss. ‘martwatchCNN-LST Confusion Mati ener No Hand need aera ant Ore Cantal Hand need ig. 6. Confusion matrix for smartwatch of MFA Model Accuracy & Loss CNN-BiLSTM decry ont Fig. 7. Accuracy and los for Smartwatch of CMEA. Sma Phone CHNBALSTH Confusion Matrix Fig. 8. Confusion matrix on smartphone of CMFA 'B Convolution Gated Fusion Algorithm (CGFA) In this paper, we propose and implement the CGFA model for HAR, using the Keras API and Tensor Flo's backend, Fis. shows the pro- ‘posed model's flow diagram, which includes various convolution ft ‘izes used to capture various temporal local dependencies. It consists of twelve layers Initially, che pre-processed data isfedto the three 1D CNN layers and filters ofeach convolution layer 32,64, and 128. which are shown in Fig. 4, These 1D CNN outputs are fed to another three 1D CNN layers and 128 filters of each convolution. The sequential data is pro- ‘cessed using convolutional 1D (Conv1D) layers, which extract local Meanremen Semon 4 (2022) 100512 Model Accuracy & Loss CNN-BiLSTM —————— Fig. 9. Accuracy and loss on Smartphone of CMA. features, And concatenate these layers followed by dropout layer also two BGRU layers with filter sie is 128 for each layer Ultimately, the ‘uu ofthe model isa probability distribution across classes obtained from the classification layer’ output (Sofa). ‘There are four sub-sequences of length 64 inthe WISDM dataset used ln this work, with each sample is 256 bytes long. Nonlinearities are introduced into Conv layers using the ReLU activation function. After the third ConvID layer, an extra dropout ayer (0.1) isadded to prevent overfiting. ‘An activation function known as SoftMax is used to classify input data, and its normalized results are fed into a clasification layer. The ‘Adam optimizer is used to each the proposed architecture how to decrease categorical ross-ntropy. 24, Evaluation analysis Various evaluation metrics [49] are used to evaluate the proposed, (MEA and CGPA in this work. ‘Accuracy: It isthe proportion of accurately predicted samples to the total number of predicted samples, where TP stands for true positives, EN stands for false negatives, IN stands fr true negatives, and FP stands for false positives. Te 41 Aecuracy= ap EH © Precision: Precision is the correctly classified positive instances to the total number of instances classified as postive, © Recall: The proportion of correctly anticipated postive observations in relation to the total number of actual observations made in the class referred to as recall 1 Recall = ” Fl-score: The Fl-score isthe weighted mean of recall and precision, 2x Precision x Recall FT Score" Precision + Recall . Confusion Matrix: The confusion Matrix displays the relationship between predieted and actual classes to show how well a classifier performs in a table format 4, Experimentation and result ‘This section analyses and discusses the results of human activity, recognition. In this study, there are numerous ways 10 assess the Meanremen Semon 4 (2022) 100512 rire | R81 ee nota las yn.40.9] [tadebatceny | pons] 29) | toate i. fon ans | I ' 1 1 a mo EE ; mcr ||] abc |=] fen) abt] 1 a coc ne ne 9] ae ia 9) a 6 paca a | on) po) i 2) m2) re in 0 po Fig. 10. CGFA configuration. Smartatch CNN-IGRU Confusion Matrix al 1000 Gener nd rete 11, Confusion matrix on smaratch of CGFA effectiveness of Deep Learing-based methodologies such as CMFA and, (CGFA. Precision (recall), FI-Score (accuracy), and misclasifiction rate ‘are among the evaluation metrics. The model results can be viewed using the model configurations mentioned in te preceding section, Model Accuracy & Loss CNN-BGRU Fig. 12. Accuracy and loss on Smartwatch of CGFA 4.1. Experimentation sep Tensortlow 2.4.0, Python 3.6.9, Pandas 1.1.5, Numpy 1.19.4, and Keras 2.43, are used in an experimental implementation. An evaluation ofthe hybrid learning methods employed inthis study was carried out fn Google Colab, which has a 12.72 GB RAM, 68.40 GB HDD, and @ ‘Tesla T4, OptKeras 0.0.7, based on Optuna 0.14.0, was used for hyper- ‘mart Phone CNN-GRU Confusion Matrix Geert nets pa Geverl Non Hand Gnented Genera! Hand rene Fig. 13, Confsion matrix on smariphone of FA Accu a Accuracy and loss on Smariphone of CGA. 3 (a) MAS © Meanremen Semon 4 (2022) 100512 parameter tweaking. 4.2, Results 4.2.1. Experiment A: result on smarwatch of CMFA tn WISDM.HARB daraset The CMFA configuration described in Fig. 5 was used for perfor: mance evaluation, including the hyper-parameters listed in the figure. ‘The datasets were split into three portions: For training, validation, and testing, use 60:20:20. Smartwatch datasets have a confusion matrix depicted in Fig. 6. The top model achleved 97.76% accuracy. The smartwatch datasets confusion matrix shows that many non-hand- oriented behaviours were misclassified as hanet-oriented. Raw sensor data from smartwatches allowed them to discriminate between hand.oriented activities and those that did not involve the use ofthe hands. The loss and accuracy curves forthe smartwatch are shown in Fig, 7. The framework training and validation loss both steadily reduce over the course of 25 epochs, at which point they remain con: stant. In addition, the accuracy of model validation and training are parallel to one another and comparable to one another, and there is na evidence of overfitting, 4.2.2. Result on smartphone of CMA in WISDM-HARB dacaset| ‘The dataset used for CMPA model performance evaluation on smartphones was split as per the rato of 60:20:20 for training, valida- tion, and testing. The smartphone dataset isthe most accurate model, and it has a precision of 94.98%. The confusion matrix of the smart. Phone dataset is depited in Fig. 8. According tothe smartphone data- set's confusion matrix, several nonhand-oriented behaviors were misclassified as hand-oriented. The proposed model was able to distinguish between general non hhand-oriented activities and general hand-oriented activities from the raw sensor data. There isa plot of the smartwatch dataset model acc racy and loss curves in Fs. 9, According tothe graph, losses for model training and validation steadily decrease after 25 iterations. There are no signs of verfittng in the model's training and validation accuracy. (b) —— @) Fig. 15, Plot of aecuracy and les obtained by GMA Smartwatch with diferent ke ies Meanremen Semon 4 (2022) 100512 a Soc = (b) @ ¥ Esa f«. = © ) Fig. 16. Plot of accurtey snd lot obtained by CMFA Smartphone with diferent fer sas, i. 16() Plot of accuracy and los obtaned by CMEA Smartphone with Alter sie is 8, Fi, (Plo of accuracy and loss obtained by CMFA Smartphone with fer size i 5, rip. 1°(2) Plat af accuracy and lass obtained by CMA ‘Smartphone with iter ie is 7, and Fs. 1G) Proposes securacy and los plot obtained by CMEA Smartphone with combining ites ‘Table 4 Performance comparisons of bth sizes of CMA model in terms ofFI-Seote and accuracy. ‘Table Comparisons of the proposed made’s performance agains baseline hybrid models. ‘Model atch WISDM Datser ‘wisbM Dataset ey Smartphone ow co oH oo ‘Tables Performance comparisons of batch ies of GFA model in terms of FL-Seore and accuracy. ‘Model atch WISDN Datsee ‘WisbM Dataset = iatwah Smartphone ow oo. oH oo 4.2.3, Experiment Bi result on smartwatch of CGFA in WISDM-HARB davaset ‘The CGFA configuration shown in Fig. 10 is used for performance ‘evaluation, Training, validation, and testing datasets were divided (60:20:20. The most appropriate prediction for the smartwatch dataset ‘was 96.91% accurate. ig. 11 displays the CGFA confusion matrix. The smartwatch data sets confusion matrix suggests that many activities that are not speci leally hand-oriented were misclassified as hand-oriented activities. Reece Daas Used nee OO (e) sont ‘ached Soh 9813 [ol OproRTuwrry espe 9857 fer vara wisn ous erase o575, for UaraR spre ote (os) Swat dase wor PaMAP2 acne sos eae Seat ‘Proposed WSO core. 94:98 for Smartphone Smarowateh ‘Proposed WSO ora 8438 or Smartphone Because the smartwatch is attached to the wearer's wrist, the model could distinguish between activities involving the general non-hand oriented and general hand-oriented activities. Fg, 12 shows the Smarcwatch accuracy and lors curves. According tothe graph, losses for model traning and validation steadily decrease after 25 iterations. “There are no signs of overfiting in the model's traning and validation ‘Table? {A difeentiation ofthe proposed models performance with that ofa hybrid ‘ework consisting of RNNs and other CNN versions with regard othe accarecy ‘of both smartwatches and smariphones. Model Sartwatch Accuracy in Smartphone (Aecray fo % * GwoRU Ta 9654 one Inception Tine (70) 95:79 sso DeepceowisTa 87.65 ral vo accuracy. 4.24, Result on smartphone of CGEA in WISDM-HARB dataset ‘The dataset utilized for evaluating CGFA performance on smart- phones was divided into the following proportions: 60:20:20 for training, validation, and testing. For the smartwatch dataset, the best ‘sccurate model was 84.35% accurate. ig, 18 shows the CGA smartphone datase's confusion matrix, The smartphone datasets confusion matrix suggests that many activities that ‘are not specifically hand-oriented were misclassified as hanct-oriented ‘etivities, ‘The model was able to identify between general activities that did not include the use of hands and general activites that did involve the use of hands based on raw sensor data. Datasets from smartphones are shown in Fig. 14 together with their corresponding accuracy and loss ‘curves. According tothe graph, losses for model training and validation steadily decrease after 25 iterations. There are no signs of overfiting in the model's taining and validation accuracy. 4.2.5. A comparison ofthe overall performance of several models with a variety of filter sizes [A igher number of important features will be extracted during the learning process using the proposed approach. A comparison is made between the different filter sizes in the proposed CMA. Diflerent filter sizes are employed, which results in slight differences in performance. ‘The CMFA use filter sizes of 3,5, and 7, respectively, (or ts training, Due to its training using the same parameters asthe developed frame- ‘work, it has this property. Figs. 15 and 16 show how the proposed approach affects the accuracy and loss ofthe WISDM dataset. From the loss and accuracy curves it is observed thatthe filter size is set as 7, ‘which gives 98.34% accuracy for Smartwatch and 94.98% accuracy for smartphones, then the filter is set as 5 and 3, which gives 97.38% for a smartwatch, 94.86% for a smartphone, 95.11% for a smartwatch and 188.7196 for smartphone, respectively. As a consequence of this, the ‘lasification results that were produced were ideal forthe vast majority ‘of the activities when the proposed model was applied to all three flter sizes simultaneously (thats, 3,7, and 11), Table present the results of ‘2 comparison of the proposed model performance versus that of several ‘atch sizes of CMFA, According to the results, there isa clear advantage to using mulple filter sizes rather than a single iter size ina model. “Tale 5 presents the results of the proposed CGFA. In this CGFA model, the authors didnot use different iter sees, and they used only one filter Ge,3). “ig. 15(@) Plot of accuracy and loss obtained by CMA Smartwateh, with fer size is 8, Fig. 15(b) Plot of accuracy and loss obtained by (CMFA Smartwatch with filter size i S, Fi. 15(€) Plot of accuracy and loss plot obtained by CMFA Smartwatch with fer size Is 7, and ig, 15 (@) Proposes accuracy and los plot obtained by CMEA Smartwatch with ‘combining fers. 42.6. Model performance comparisons with various hybrid models ‘A comparison shows the performance of the proposed work with various other models, and Table 6 is a comparative analysis of other Meanremen Semon 4 (2022) 100512 baseline fusion models with our proposed fusion models. 42.7. Comparative evaluation of the proposed model's achievement with the RN andl CNN variants The proposed models, CMFA and CGFA, were also examined against DeepConvLSTM and CNN-GRU modeling techniques, which are also mixtures of CNN and RNN varieties. The DeepConvLSTM and Hybeid ‘models like CNN-GRU share the same hybrid architecture as the pro- Posed HLA models, with the exception that the LSTM and GRU layers replace the BLSTM, and BGRU layers. Both the proposed model and existing baseline models undergo training with the same group of hyperparameters. The BLSTM makes use Of information from both the past and the future in eircumstanees in Which ihas access to the entre sequence of time series data. In contrast, the LSTM and GRU models rely ole on data from the past. The CMFA network can make better predictions than LSTM and GRU because ofthe additional context it provides. Table 7 confirms this, as does the data presented there, DeepConv1STM and GNN-GRU models have similar performance metres, (On the other hand, in comparison tothe previous CNN-RNN varia- tons, the performance of the suggested model in terms of recognition is ‘much improved. The dataset that was employed in the proposed model was obtained using a number of different types of sensors that were Positioned in discrete areas of the human body, and the data were captured ata polling rate of 208z, The suggested models can identify complex and simple activities ike walking, jogging, stairs, sitting, standing, typing, drinking soup, eating chips and pasta, drinking, kick: ing dribbling, writing, clapping, folding ete, with decent accuracy. We proposed models presented inthis study appear to generalise well across the dataset use. 5. Conclusion This study ereates a hybrid learning algorithm forthe classification of human activities. Tis investigation utilized unprocessed sensor data from the WISDM collection. The original dataset was split into two diferent datasets for smartphones and smartwatches. During. pre processing, the sliding window approach was used to transform the data, This research did not include any manual feature engineering. ‘According to this work's findings, smartwatches outperform smart- phones in accurately detecting a wide range of human activities, and the results were further evaluated using the validation, test, and train datasets, When extracting spatialtemporal characteristic from rave sensor data, our findings show that a hybrid learning strategy is faster and more efficient than previous deep learning strategies, which had a more advanced architectural model. In addition to HLA, the eurent research Investigates more complex models of neural networks as part of its investigation. Deep Convolution Transformer models may also be used ln future work to classify human activity time series from the other datasets. To grasp the relationships between raw sensor data inputs, neural networks equipped with selFaitention processes might be use “More activities and participants can be added to the WISDM dataset in the future, which can be used to classify the data even further. Declaration of competing interest ‘The authors declare that they have no known competing financial, interests or personal relationships that could have appeared t influence the work reported in this paper References 1 2 ig he aX ng Da aig ert city (2) B Damatvian Vier, J Slevi, M Wosna, Haman ety reconin In AL cvronmers using random projections, Comput Nah RA Athan. Suc Matos Md. 2016 2016, 4073564, ot or/1.1155/2016/9073555, 1 pe. Manes, Damen, 8 Segal A reve of nee of tings technolo or anbven aed vig enone, Fre inert 1 (12) ‘Go19) 255. 4 engl V. Radon Masala, R. Danaevss, Recoglon of Tel rece gral rom de wing itera feed rans (HOS) ‘Sh sopor vate machie (SM), roa Compt So. 150 (018) 955950 Damar, Makan, A Venlausan A Weak Sarthe oe ey veienton using galt characters, Syanetry 8 (10) (201) 100. tan, Resta ening based CN Tor gesture ronan In foot neat, afoot Proce Set 17 (2) (2021) 395-258. {Cdohanpay, J Bava Doo Haman ety recopitona avy, road ‘Canpat Sa 135 G19) 638-703, TEAC ang KH Wang AP, CH Lee, Moa, Somor tal ad Seni ham ach ception combine, an ees {evice te aman counting ug WH IEE Seno 1.19 CE) (209) 9821-9528 {Tyebin Sealy, ih Orange, Homan avy recogion wither ‘enor ting dep lensing approach, 2016 IEEE Sess, 2016, Ook, peisinee Seuss, A span, stm network sing smartphone at for ‘oso aed aman sty recount hres, Snare 2 (5) (2021) 5. elsavnich A Jpatanal, loner ser Wonton aston huan fet ecopation ng warble sess an experimen ig Sep earning tmoal, ec 10 () (202) 508. S'Rat BAdacson, MLW. Jue AS Utell, CE Tuk, Mula ste Fes in age persons: compan std of rehabiated and tonrehliate pent with fare the Np, J hon. Di 15 10 (1982) bt i, 8, Gar, N. Pombo, F. Pre Reva, Fm dts sequin to tits faa congrehenive rowan roadmap fr he Mectscton of fetes o aly ving wing mob devs, Sensors 16) 8) 188 ‘Ngacy, Reade hunan acy reco rom eeermete data wag Goneluional Neural Networs, ADL So amp 62 208) 915-02, Gh Weis, K Yoneda, Hayne, Smarghone and arth based omer sing sctisso aly ving I Act 7 (2019) 153190-135202, MM Vagte, @Siko, LA Kaladi A review of aman actly ecogaion Imex, rs lp Rootes and At 22015) 28 GGL Sans, PT Endo, KHLDC Mont, EDS. Roch, L Sta, T. Lyn, ‘ecelrometer bbe human fal detection eg conoltona! eral betwee Semmes 19 (7) (201) Beat, SMtrusavaich A Jeptankal,P. Youn. Yop, Eahacnd hand fein act gain ase on smth wor ang Symacy 12.6) (2030) 1570. Yah, shir, Kees, Human cy recon fom eieonmcat Ieckgrond sods for ries senor nctwork EE) Trance, Informa. Stam 190 (4) O10) 365-572 (Lar, ALA Labrador, Amare on haan acti secon ing wearsle Sensor IEEE Cosmet. surveys Tati 15 (3) (2012) 192-1208. (6. Wes, UCI Machine ening Rept, Unies of Caer, chet of Invrmation and Compute Snes ving CA, USA, 2019, Ava eine: epee essed anne WISDM| Sargon nada ‘Nien and ome bet (Aged 1 Ag 2022) aces on 1 Cen. Zhang. Yo, B Ga, ZY, ¥ tr Deepening for sensor based hima sty recogston: rena, calenges and opportunites, AM Camp sur” 54 (4 (2021) 1-40 "asin, Sheng WE. Zhang Dene Approaches for Human Ay Recogtion: A Survey, 2019 ar prep art 1906 08076, 125) JX: Angra, Q. cl Huan maton aap a even, Compo Vis image Understn. 73 (3) (1999) 428-440, ohana, Vs perception of Bicol motion nd model fr its nly Przepe Pychopbye14(2)(1973) 20-211 P. Tura, fe Cheleppe, V5. Subramanian, 0, Us, Machine recognition of Tunanecdvis a surey IEEE Tan Cle. Sst. Video Tech 18 (11) 2008) Eb, Tapa SS. nile KLaron, Aci reopen i the home wig sil $d chigtonr seo, Intemational Confrence on Pov Compug Springs, ein, Hedley, 208, App 18-175. MLC Mout, The neural network house an eoveniet hat adap ts Inaba in Poe AAAI Spy Srp. Baronets ol 58,198, uch Hi Abdeone, M.Youne, A Haas, Wipe bigutour wise gesture ‘copton sion 2015 BEE Concrene on Computer Communications (axrocon, 2015, App, 1672-480. TREE Sang Zhu, Arve on ado based sty ecgan, Dig CGnmaicar Netw 1 1) (2015) 20-29. 1 Drammond,WNearblesomor ior materia he pyle sad chemical Poi af hemlet, NP geal mediine 2) (019) Li (55) J. Cote, Torsten hageseand, tine geography, CIS Class (200) poset. og elem 27568, a I i 1 1 no} on nay na ns) na os) ro vo en co pa ps a ps v9) in oy a (38) (38) in (ss) 39 eo ca wa co co es 8) va coy col ‘80 ou 1s) ‘ss a ‘ss (58) in (58) 1s (oo (on) wa Mearemene Senor 24 (202) 100512 Athol. Sarr sng GPS to ler guia locos and predict ‘orem cou lil ses, PeoalUsguusCompt 7 () 02) “ igrsvand, What aout People in Region Sees, vl 24, repos scene tioetato 1970. IC eigen, RA Feske, 3.6, Ys, 6. Palo, Tue ction anys for spose scent sro cheng novel slob! poston sem Iseramet tnirn: Heals Pespect 111) Gon 8-122 Iho an hdden atk mda. Network. 4 (10) (2009) 976-984 BX. Dewangan, Se, Poet poole deceton for utonomoas veh ‘pete ing omeltiona ner artwork, Eto, Ltt 57 (2) (2021) 5386, BX. Dewangan, SF. Sahu, Deep ening based seed bp tection sel for Ineigen vile sem sing rspenyP TEEE Ser. 21 (3) 020) DK. Devangan, SF. Sahu, RO ond elasfation convolutional nel ‘eto for neliget vb system, tliat Sve Rabe 12) (2021) rs Gana, C Sh Borges, Gta, Annee apc for mec omy detection ing Sep patch conolton! nel newer Vi Compe. 36 (0) (2020) 1865-182, ‘Vs Smale Mend, CC Nand Robust and acute Feature seton for mand ph ceovery and clit: asp ering approach, Nesta CGmpue Appl 282) 019 65-574. Panag Si Dy Ke Pras, Das, A. Acarys, K: Maar, ©. Fat, CY base approach or acy recognition usa wrist wor eset, es 2077 ath Anal nena Cojo OF te EEE agai Mein fd SeeyCENDC), EE, 7,2 Swan Q Xu Tong, Z a, Dep leaming models or vessime aman feat econ th satphnes, Moble Network. App 2 C2) (2020) Wang Je, Zhan, Aten tase convlational eral neta for weakly [Shela man sce secpaion with wena sensor, IE Senor. 1 (G7) coi) 7998-7608, (Teng, K Wang Zhang. J: He, The layer wise raining convolutional neural eo lots seas ed Duman ait ecgnon, IEEE Sens, 520 1) (020) 7265-7274 i tong, LT. Nye, B.Yu,01, Mensol, JZ, P. Ws. Ban, Corellon neural newer fr uta svi epi tng able Secor I th ineroational Confrence on Mable Compu Apts nl ‘Sevies, 2014, Nove pp 197-205. ERE. 1M Ula Ha, SD Kn, FA Cet, Saad st network or amas tty mcogaton wing smartphone daa 2019 th Eaopean Worksop ‘Vinal Iron Praag (EUV), EEE, 201, October, 175-180 Ferner, LF Sate, Vilma ML Alive Haman cy recogation ot Srphoes wig leona LST sew, 019 2 Speposian on Image, Sigel Peay and Arial Von (STSVA), IEE, 209, Apel. T-. YP, A Yang, 6. Cheater, X Xe 2 tans, Deep rail iL STM oe {human sty nga nog wena ses, Math rol Bg. 2018 (2018), asa sige angi ea st 8 page ee I Mtge DS Han, A CXNSTM apeoach human tiny renga, in 2120 iterations Confrence on Afi! Ielgece Ie eration And immer CAI) EE, 2020, Peery. pp. 362 Ss, CX Haar, Wang LSTALCNN acter fo uma actiy cognition, Te Acne nao) 8 Se 1 Wang. Hel Zhang, Sequel weakly abled mls leaiacon and ‘reopen oneal snr wing recent leon ets EL Tras Hum acioe sys 51 (0) (021) 35508 TUL We Zng, YM. ang. Usama: arg bench fe hua beaver undsaning with unmaaped aerial veils, Postings of the IEEE‘CV Conereneeon Compu Vison and aterm Recgnton, 202, ‘A Pule, C Carmona Dare, MA. Fee, A. Maell, 20 9130 online weer entfeaton: seomparive st). Intermairal Confrence on Bacar! ‘nats And Recaption, Spige, Cham, 2021, Speer, pp. 307-32. Sherwani, A Jipatanal, let see Wenteaon son human fed recogni sng wearable esr: experiment wl Sep ering ‘Rodel, error 10 (3) (202) 308. Benepe lone, Coma cura etek or wer etfcaton| ‘Rendon nian som epee te tapes HEE Acer © 220), ‘Goel, Yoneda. Hayate, Snarghoo and snatch bsed Dometic sing acies of daily ving, IEE Acs 7 (2019) 153190-138202, ‘Ahogistng ML Breed M. inte 5 Gagne M. Vag entices of leroy ing marion sora experience Teaming IEE Aas 8 (2020) 2745-27907. Mule, Bak, Wer, 8 Wed, BJ. Chol, You walk we ent: Ughivelht series auhenteton bse o ata wearable Ist Sens IEEE Aces 7 (2019) 3788-37995 Bond MLA Alps, am AK, asa Chaban, M. Mazar, Dian, Sarath aed Lepinte User entiteaon or Coad sed Secure Servis, Male Information ste, 2018, 018 NNewrov,C Wall. Lace. Pima, D. Chane, B Rebel, Tye, Taming hun enim en patter IEEE acs (2016) 1610 1850. 16S) ola, bh, D-Kan, A Chit, Stace a sever or haa eit esa sng smartphone dats 2019 6s European Worksop ‘Vi Infernaon Prong (2UVP), IEEE 2019, ober, 175180, oA) ¥ oe, .Yg. 6 Cl, XZ, zhang, Dep resale STM fo ‘tansy eomnton sng weal ese, Wath, Probl Eng. 2018 2018), 7340964 ip loi 0 1180 20877516954 18 pas I pres 68) Rtgs DS tan, A CANSTI apo to human actiry rena, 2000 International Goafrence on Arua! neler In fran And. (Ginmancatn (CAI), IE. 200, Foner, pp. 362 365, {ee} Forte, D. Roxen, Dep conlunal td Kun creat nel pewooks foe multimsl wenable acti ecogaton, Sensors 16 1) (2018) 115, Mearemene Senor 24 (202) 100512 W501 ang He ae een wey bed iin at nt Taman on Hume che Syate GoRD 26 364 (68) Tn Zang, MW, Wang, Uehara: ag benchmark for Tuna beter undastandig wih unmanved eel veil, Pocoling of the IEEE‘CV Conereneon Compu Vso and Pater Recognon, 202, (69) HE Damien R Kaos, LD. Tepe, ARC acy rcogaion ‘eh caps i 090 19 IEE lneratoal Goes a Mace Ling ‘and Apleans ICMLA), TEE, 2020, Decne, ph. 1982-1388, 170) 5cupe, Dep leaning based man ecvy econo (HAR) sng wearable ‘Scoun datnteratinal oural of lfrmaton Managers Des ait 1 (2) (Goat, tose

You might also like