Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Classifying tasks performed by electrical line workers using a wrist-worn sensor: A data analytic approach

  • Saeb Ragani Lamooki ,

    Contributed equally to this work with: Saeb Ragani Lamooki, Sahand Hajifar

    Roles Data curation, Formal analysis, Methodology, Software, Visualization, Writing – original draft

    Affiliation Mechanical and Aerospace Engineering, University at Buffalo, Buffalo, NY, United States of America

  • Sahand Hajifar ,

    Contributed equally to this work with: Saeb Ragani Lamooki, Sahand Hajifar

    Roles Data curation, Formal analysis, Software, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Industrial and Systems Engineering, University at Buffalo, Buffalo, NY, United States of America

  • Jacqueline Hannan,

    Roles Data curation, Visualization

    Affiliation Biomedical Engineering, University at Buffalo, Buffalo, NY, United States of America

  • Hongyue Sun,

    Roles Conceptualization, Funding acquisition, Methodology, Supervision, Writing – review & editing

    Affiliation Industrial and Systems Engineering, University at Buffalo, Buffalo, NY, United States of America

  • Fadel Megahed,

    Roles Conceptualization, Formal analysis, Funding acquisition, Project administration, Writing – review & editing

    Affiliation Farmer School of Business, Miami University, Oxford, OH, United States of America

  • Lora Cavuoto

    Roles Conceptualization, Funding acquisition, Methodology, Project administration, Supervision, Writing – review & editing

    loracavu@buffalo.edu

    Affiliation Industrial and Systems Engineering, University at Buffalo, Buffalo, NY, United States of America

Abstract

Electrical line workers (ELWs) experience harsh environments, characterized by long shifts, remote operations, and potentially risky tasks. Wearables present an opportunity for unobtrusive monitoring of productivity and safety. A prerequisite to monitoring is the automated identification of the tasks being performed. Human activity recognition has been widely used for classification for activities of daily living. However, the literature is limited for electrical line maintenance/repair tasks due to task variety and complexity. We investigated how features can be engineered from a single wrist-worn accelerometer for the purpose of classifying ELW tasks. Specifically, three classifiers were investigated across three feature sets (time, frequency, and time-frequency) and two window lengths (4 and 10 seconds) to identify ten common ELW tasks. Based on data from 37 participants in a lab environment, two application scenarios were evaluated: (a) intra-subject, where individualized models were trained and deployed for each worker; and (b) inter-subject, where data was pooled to train a general model that can be deployed for new workers. Accuracies ≥ 93% were achieved for both scenarios, and increased to ≥96% with 10-second windows. Overall and class-specific feature importance were computed, and the impact of those features on the obtained predictions were explained. This work will contribute to the future risk mitigation of ELWs using wearables.

1 Introduction

The reliable delivery of energy is a crucial factor of economic development and a key metric to assess a developed nation’s infrastructure [1]. Electric power reliability demands a high level of maintenance, which is performed by Electric Power Industry Workers (EPIWs). Among the EPIW, Electrical Line Workers (ELWs) have the third highest injury rate after meter readers and welders [2]. For an overall injury rate of 3.2 per 100 employee-years for the EPIWs from 1995 to 2013, the ELWs injury rate was 10.37 based on the data collected from 18 electric power companies in the US [3]. Among the tasks that ELWs perform, walking up and down ladders and stairs; climbing down poles and transmission towers; entering, stepping out of, or approaching the utility trucks, bucket, or vaults; and performing repetitive work/overtime are found to be associated with the highest risks of injury [4]. Thus, timely monitoring and analysis of ELWs activities can provide significant information about the type, duration, and frequency of risky tasks and can be used for prevention of work-related injuries [5].

In the past decade, the Internet of Things (IoT) has enabled the collection of data to elucidate system operations and facilitate smart operation decisions to enhance several key performance measures [68]. From a safety perspective, wearable sensors and IoT-based monitoring can be used to estimate the risk of injury and support risk reduction [913]. In addition to risk assessment, IoT data analysis can support worker training, work assignment, and predictive scheduling, and the outcomes can ultimately be fed back to the worker in the form of feedback and safety alerts. In the current application, we are interested in leveraging IoT to collect kinematic data from ELWs. As shown in Fig 1, there are work domain characteristics for electrical work that support the use of IoT for safety enhancement and that can be captured through kinematic analysis, including the nature of the work tasks (e.g., often performed at height and with heavy loads), variability in tasks across days, and risk in remote/lone worker situations. Worker kinematics allows for monitoring which tasks are performed, how many times, and for what duration, all of which are critical factors for risk assessment.

With the kinematics data, one needs to first identify the activities that ELWs performed, to track the time and duration of these activities and to guarantee the worker safety and prevent economic loss. Therefore, a robust activity recognition method that can be effectively integrated to the ELWs work settings for detection of relevant activities is of interest. In the literature, many studies have investigated the application of human activity recognition (HAR) methods in activities of daily living (ADL), sports, and healthcare [1416]. However, the activities involved in industrial occupations are often more complex and the complexity further varies between different industries. The application of HAR methods in occupational activities has received attention in certain industries, such as construction [5, 17, 18] and assembly line works [1921]. Considering the uniqueness of the tasks in different industries, the effectiveness of HAR algorithms needs to be investigated independently in each industry. Particularly, we are interested in classifying common tasks among ELWs, such as climbing ladders, hoisting tools, and working on electrical panels. Moreover, in HAR applications different hardware and software challenges need to be addressed based on the limitations in the industry. For ELWs due to the long shifts in remote locations, the hardware must be lightweight and able to operate long hours on a single charge, while preserving a high recognition accuracy. Towards this goal, we investigated the feasibility of a single wrist-worn accelerometer for activity recognition among ELWs.

In HAR applications an inherent trade-off exists between the number and location of the sensors/signals and the feasibility of the method. The accuracy of the classification methods generally increases with more sensors/signals at different body locations. However, this can disrupt the underlying task in real-world settings. In occupational applications, it is critical to attach the sensor set to the worker with minimum obtrusiveness to allow maximum flexibility for complex motions [21]. The performance of HAR algorithms depends on the location of the sensors on the body, since the acceleration signals vary at different body locations while performing the same activity [22]. Since in many occupational tasks the upper limb motions and particularly the hand kinematics contain significant information about the underlying activities [22], we have investigated the effectiveness of HAR methods to classify the ELWs activities using a single accelerometer on the dominant wrist.

The application of individual sensors on the upper limbs for activity recognition has been investigated under certain circumstances. Nath et al. [5] attached a single smartphone to the upper arm and reported the highest accuracy of 90.2% for a 3-class classification of construction tasks. They used three measurements from the smart phone’s built-in accelerometer, linear accelerometer, and gyroscope. Koskimaki et al. [20] used an inertial measurement unit (IMU) on the dominant wrist to classify 5 activities of assembly line workers in automotive industry using the accelerometer and gyroscope data and reported 90% accuracy. Using a single wrist-worn accelerometer, Ryu et al. [18] classified 4 construction activities and reported the best accuracy of 88.1%. Also, Yang et al. [23] utilized a single accelerometer on the dominant wrist to classify 8 domestic activities and reported 95% accuracy. Additionally, by the use of a single wrist-worn accelerometer, Chernbumroong [24] classified 5 ADLs and reported the best accuracy of 94.13%. Classification of occupational tasks are in general more challenging than the ADLs due to the complexity of the motions and therefore lower accuracy is reported for them in the literature especially when fewer sensors/signals are used. We have summarized the activities, classification methods, number of participants, sensor type and location, and the reported results from a number of studies that used a single sensor on the upper limbs in Table 1.

thumbnail
Table 1. Activity recognition studies using a single sensor worn on upper limb.

https://doi.org/10.1371/journal.pone.0261765.t001

Choosing sensor attachment location(s) is a crucial decision when using wearable sensors [30], which depends on the application and types of the tasks. Here, we review the sensor attachment locations and number of the activities in related occupational studies as depicted in Fig 2. The number of sensors utilized per study ranges from 1 to 24, with a median of 1. Fig 2 illustrates the total number of times a wearable sensor was placed at a certain body location (shown by circles), with 71 sensor attachments across the 20 studies surveyed. Those studies with multiple sensor placements are indicated in the upper left and lower left of Fig 2. In 40 out of 71 sensor placements, the sensors were placed bilaterally (∼ 56%). In case a sensor was placed unilaterally, placement on either the dominant side or the right side was more common (18 out of 20) compared to non-dominant side or left side (2 out of 20). The wrist was the most often used location for sensor placement (6 out of 20 studies), followed by waist, upper arm and foot (5 of 20 studies each).

thumbnail
Fig 2. Body locations of the wearable sensors across the reviewed studies that involved occupational tasks (20 studies).

https://doi.org/10.1371/journal.pone.0261765.g002

Moreover, we categorize the studies into two groups based on the number of tasks/classes and present it in Fig 2 where fewer than 6 tasks are shown by blue circles and 6 or more tasks are shown by white circles. The number of tasks per study ranged from 3 to 23, with a median of 5.5. The studies that included more than 5 tasks used multiple sensors at multiple locations more often (7 out of 9). For instance, [31] used four wrist-mounted sensors to classify 23 bicycle maintenance tasks and, sometimes, the classification involved distinguishing between the tasks as similar as fastening and unfastening a screw (both of them require rotational motion in both directions). On the other hand, the studies considering fewer tasks were more likely to use a single sensor (8 out of 10). However, [32, 33] tried to classify large number of classes using a single wearable sensor to reduce the obtrusiveness. From Fig 2, it is clear that using a single wrist-mounted sensor to classify large number of classes was not explored across the 20 studies that investigated the occupational tasks, which is the focus of our research [5, 18, 20, 24, 27, 3146]. Besides being less obtrusive, wristband sensors are commercially available and easily adoptable by the consumers [47, 48].

Based on a single wrist-worn sensor, this paper aims at proposing a data-driven methodology for accurate activity recognition for ELWs. To achieve this goal, we have addressed three research questions: (1) What is the predictive performance obtained from using a single wearable sensor to classify different ELWs tasks? (2) How well does the approach in research question 1 generalize to individuals not used in the training of the models? and (3) What is the relative importance and impact of the features extracted from the wearable sensor in order to distinguish the performed tasks?

To address these questions, we will achieve automated human activity recognition for ELWs, capitalizing on existing feature extraction and statistical learning techniques. Specifically, we designed an in-lab experiment to examine:

  1. Intra-subject Classification: We perform the activity recognition for 10 ELWs’ tasks using k-nearest neighbors (k-NN), support vector machine (SVM), and random forest (RF) models with the data from a wearable sensor for each subject.
  2. Inter-subject Classification: We further test the extrapolation performance of the activity recognition for new participants, to determine whether our models are robust to the heterogeneity across participants.
  3. Interpretation of Predictions: In addition to the overall feature importance, we analyze the task-specific feature importance, since the contribution of different features varies for prediction of different classes, using local interpretable model-agnostic explantions (LIME) method [49].

Addressing these questions paves the way for automatic and remote monitoring of the duration and frequency of the hazardous tasks performed by ELWs.

2 Methods

2.1 Experiment design and data collection

We recruited 37 participants (demographic and anthropometric data summarized in Table 2) to perform experimental sessions that included 10 simulated common activities of ELWs. The study was approved by the University at Buffalo Institutional Review Board and all participants provided signed informed consent prior to participating. At the start of the session, participants were instrumented with an Empatica E4 wristband (Empatica, Boston, United States) on their dominant wrist to collect the acceleration in the X, Y, and Z directions (with respect to the local coordinates of the accelerometer) at a sampling rate of 32 Hz. Participants then completed the series of 10 activities as presented in Table 3. The participants were asked to carry a smartphone in their pocket for real-time collection of data from the wristband via the E4 real-time app and Bluetooth connection. The data were uploaded to the cloud automatically by the app at the end of the data collection session. The start and end time of each activity was recorded by an annotator during the session.

thumbnail
Table 2. Summary of anthropometric data for the participants.

The values in columns 3–5 are the average (± standard deviation).

https://doi.org/10.1371/journal.pone.0261765.t002

thumbnail
Table 3. The experimental tasks, which are performed in a columnwise order by row, i.e., first column → second column → third column.

https://doi.org/10.1371/journal.pone.0261765.t003

2.2 Workflow

Our analysis workflow is presented in Fig 3. After the data was collected and stored locally, the tasks were annotated and the signals within each task were segmented to windows. Next, three types of features in the time, frequency, and time-frequency domains were generated within each window. The features were used as the inputs to three classifiers: k-NN, SVM, and RF. The classifiers were trained and tested for intra-subject and inter-subject scenarios. The best set of window length, feature set, and classifier were selected based on the inter-subject analysis for potential deployment.

thumbnail
Fig 3. An overview of activity recognition framework for acquiring, preprocessing, and analyzing acceleration data.

https://doi.org/10.1371/journal.pone.0261765.g003

2.3 Windowing

We labeled the acceleration signals and segmented the annotated signals into equal length windows. Multiple windowing strategies have been used in the HAR literature, including non-overlapping sliding windows, overlapping sliding windows, event-defined windows, and activity-defined windows [50]. In our work, we segmented the labeled acceleration data into non-overlapping fixed-sized windows, as this is the most widely employed windowing technique, due to simplicity and efficient pre-processing [51]. Considering the complexity of the activities and the use of a single accelerometer, we selected two different window lengths of 4 and 10 seconds, equivalent to 128 and 320 data points in the acceleration signal (sampling rate of 32 Hz). While smaller window sizes can result in faster activity recognition, larger windows are better suited for recognition of complex activities. Banos et al. reported the optimal window size of 0.5 to 6.75 seconds in sport/wellness applications [51]. In Fig 4 we visualized the three acceleration signals in X, Y, and Z directions without any filtering within a segmented window of 10 seconds for the 10 activities of a single participant.

thumbnail
Fig 4. Acceleration signals in X, Y, and Z directions in a 10 seconds window presenting the 10 activities for a sample subject.

https://doi.org/10.1371/journal.pone.0261765.g004

2.4 Feature extraction

We have focused on using the shallow features in our study since the deep features from deep learning algorithms require larger data sets to train, and are difficult to interpret, due to their black box nature [52, 53]. In this study we examined the performance of features from 3 domains: time, frequency, and time-frequency. Time domain features include information regarding the central tendency, average power, dispersion, and the range of the signal. The frequency domain features capture the central tendency and dispersion of the signal by the FFT coefficients at lower and higher frequencies, respectively. While the frequency domain features include more detailed information about the frequency components of the signal as compared to the time domain features, they are better suited for stationary signals. Given the complexity of certain tasks in our data set, windows / observations with non-stationary signals are present in our data. Therefore, we also evaluated the performance of time-frequency (wavelet) domain features as they perform better for non-stationary signals.

Time domain features are commonly statistical measures that are simple to compute [54, 55]. However, these features are unable to describe the signal’s functional shape [56], and in spite of their good performance for distinguishing between the static and dynamic tasks, they often do not perform well in differentiating between dynamic activities [55]. To extract the time domain features in our study, we first removed the high frequency noise by applying a 4th order low pass Butterworth filter with a cutoff frequency of 10 Hz on the three acceleration signals within each window. The cut-off frequency of 10 Hz was selected since the acceleration signals associated with the human movements usually do not exceed 10 Hz [57, 58]. Next, we calculated 9 statistics for each of the three filtered signals within the window. These statistics included mean to capture the central tendency of the signal, root mean square (RMS) to represent the average power of the signal, mean average deviation (MAD) and standard deviation (σx) to represent the signal dispersion, and minimum, maximum, median, 25th percentile, and 75th percentile as the summary statistics of the signal. The 9 statistics per signal summed up to a total of 27 time domain features for each window/observation.

The frequency domain features provide information about the intensity of different frequency components in the raw signal. Advantages of frequency domain features are robustness to noise and interpretability [58, 59]. On the other hand, since the frequency domain features do not provide information about the timing of frequency components [60], their performance is limited in the case of non-stationary signals [61]. We transformed the three acceleration signals to the frequency domain using fast Fourier transform (FFT). To remove the data associated with high frequency noise, we only preserved the FFT coefficients at up to 10 Hz. This resulted in feature vectors of length 120 and 300 for the 4 and 10 second windows, respectively.

The time-frequency domain features preserve the frequency information at each time point. We used discrete wavelet transform (DWT) for extraction of the time-frequency features. The DWT features are more reliable for modelling non-stationary acceleration signals [62]. Moreover, the DWT features can represent sharp peaks and abrupt changes [63], enabling the detection of transitions between postures [64]. However, compared to FFT coefficients, the DWT coefficients are less interpretable. For the DWT transformation, we used a Daubechies-4 mother wavelet (db 4) with the filter bank approach to decompose the original signals to approximation (cA) and detail (cD) coefficients. The decomposition was applied up to the maximum possible level on the three acceleration signals (4 and 5 levels for the cases of 4 and 10 second windows, respectively). To reduce the dimension of the wavelet coefficeints, entropy, number of zero- and mean-crossings, and the 9 statistics used for time features were calculated for the last approximation coefficients and all detail coefficients and were concatenated to generate the time-frequency features. This resulted in features with the lengths of 180 and 216 for the 4 and 10 second windows, respectively.

The utility of different feature sets depends on the classification application. Selecting the best set of features depends on the type of movements associated with different activities. It is noteworthy to mention that, even though using the acceleration magnitude instead of the 3 acceleration signals could decrease the number of features by two thirds and thereby decrease the computational cost, we included the 3 individual acceleration signals since they contain valuable information in terms of the accelerometer orientation and allow for more accurate classification. For instance, in Fig 4, it is evident that in sitting and standing tasks the gravity as the sole existing acceleration, appears as different components in the three Cartesian coordinates due to the different wrist orientation, whereas the acceleration magnitude is the same for the two tasks and cannot be used to distinguish them.

2.5 Classification and performance

We performed the activity recognition of the 10 activities in two scenarios of intra-subject and inter-subject settings. In each scenario, we investigated k-NN, SVM, and RF classifiers in combination with the two window lengths and the three feature sets discussed in the previous sections. We avoided adoption of deep learning since for the intra-subject analysis the number of observations are not enough for training such models. Moreover, the large computational cost and the lack of interpretability of deep learning models led to prioritization of conventional classifiers.

The training and testing splits have been performed differently for the intra- and inter-subjects cases. In both designs we tuned the classifier’s hyper-parameters using 10-fold cross validation and grid search. For the k-NN classifier, the grid search was performed for the number of neighbors (from 1 to 28 by increments of 3), the distance metric (Euclidean and Manhattan), and the weight (uniform and distance). For the SVM classifier the grid search was performed on the regularization parameter C (from 0.01 to 50), and the kernel (linear, polynomial, RBF, and Sigmoid). Also, for the RF classifier we performed the grid search on the maximum depth of the trees (5, 10, 20, 30, 40), and the number of features to consider for best split (20, 30, 40, 50, 60).

For the intra-subject classification we randomly split 80% of observations as training set and 20% of observations as testing set from the same subject. We used stratified sampling of each activity, to ensure balanced distribution of labels. Next, we used 10-fold cross-validation to select the optimum hyper-parameters as described above. We performed the hyper-parameter tuning for each participant independently and reported the mean testing accuracy, precision, recall, and F1 score across all participants.

In the inter-subject analysis, we split the subjects into five folds and repeated the training and testing five times, using each fold as the test data in one iteration and the remaining subjects as the training data. The best model hyper-parameters were chosen using 10-fold cross validation in the training data. We calculated the average overall accuracy of the test sets as well as the average precision, recall, and F1 score for individual activities in order to evaluate the classifier’s performance for different labels. In order to come to a conclusion regarding model performance, we performed a two-way analysis of variance (ANOVA) to evaluate the main and interactive effects of feature set and classifier on the overall accuracy. Post hoc comparisons for significant effects were performed using Tukey’s HSD. The significance level was set at p < 0.05.

2.6 Feature importance and interpretation

In this study, by finding the features that contributed the most to the predictions we attempted to investigate the mechanism based on which the classifiers performed. To that end, we obtained the overall feature importance from the RF classifier. In the case of multi-class classification, feature contribution differs by class, therefore, we further investigated the importance of features for prediction of individual classes. In the latter case, we used LIME to explain the predictions by perturbing the input data locally around a selected sample via an interpretable model such as a linear model [49]. Thus, we can better understand the feature importance during the prediction in each sample, and make the classification decisions more transparent.

3 Results

3.1 Intra-subject classification performance

Given the large number of participants, we present the average test accuracy, precision, recall, and F1 score over all participants in Table 4. The results are presented for all combinations of window length, feature set, and classifier. In the intra-subject scenario, the longer window length always resulted in better accuracy for any combination of the classifiers and feature sets. For the window length of 4 seconds, the time and time-frequency domain feature sets resulted in high accuracy, ranging between 94% and 96%, with SVM slightly outperforming the others. For the 10 second windows, the accuracy increased overall and the time and time-frequency domain features outperformed the frequency features. High accuracy of 99% was achieved for the k-NN and SVM classifiers with the time and time-frequency domain features.

thumbnail
Table 4. Intra-subject classification performance.

Abbreviations: WL: window length, T: time, F: frequency EP: electrical panel, H: hoisting, Ld: ladder, Lf: lifting; OH: overhead; P: pushing; St: sitting; Sd: standing; Tp: typing; W: walking.

https://doi.org/10.1371/journal.pone.0261765.t004

3.2 Inter-subject classification performance

Table 5 shows the average classification results for the 5 repetitions. Better classification performance is observed for the 10 second windows as compared to the 4 second. Highest average accuracy’s of 91% and 97% are achieved with the time-frequency features in conjunction with k-NN or SVM for the 4- and 10-second windows, respectively. For both window lengths we observe that the combination of the SVM with any feature set and window length result in higher / equal accuracy as compared to the other classifiers.

thumbnail
Table 5. Inter-subject classification performance.

Abbreviations: WL: window length, T: time, F: frequency EP: electrical panel, H: hoisting, Ld: ladder, Lf: lifting; OH: overhead; P: pushing; St: sitting; Sd: standing; Tp: typing; W: walking.

https://doi.org/10.1371/journal.pone.0261765.t005

For both window lengths, the results of the ANOVA test showed significant main effects of feature set and classifier on model performance, with no significant interaction effect (Table 6). Post hoc analysis showed significant differences between the performance of the time-frequency domain features and both of the separate time domain and frequency domain feature sets. There were no significant pairwise differences between the classifiers (Table 6). We present the overall accuracies of each feature set in Fig 5 using separate box plots for the two window lengths.

thumbnail
Fig 5. Inter-subject overall accuracy’s for the feature sets, presented separately for each window length.

The * indicates statistically significant difference (p < 0.05) between pairs.

https://doi.org/10.1371/journal.pone.0261765.g005

thumbnail
Table 6. Two-way ANOVA for main and interaction effects of classifier and feature on inter-subject overall accuracy and Tukey’s HSD for significant effects (p < 0.05).

Analyses are performed separately for each window length.

https://doi.org/10.1371/journal.pone.0261765.t006

3.3 Overall feature importance

In Fig 6 we present the most important features, along with their contributions, obtained from the RF classifier for the case of the 10 second windows. For the time domain, since only 27 features were used, the importance of all features are presented. However, for the frequency and time-frequency domain sets, given the large number of features the top 30 and 15 important features, respectively, are included. An important observation from the importance plots is that the most important features are from the Y acceleration signal, implying that the Y acceleration is, in general, more dissimilar across different tasks, and therefore more distinctive. While this is an informative observation, the importance plots do not provide further details about the contribution of each feature for individual activities.

3.4 Class-specific feature importance using the LIME method

We show the results for the methods mentioned in Section 2.6 here, focusing on the features from the frequency and time-frequency domains. As an illustration, we identified the important features for prediction of two activities, i.e., working at an electrical panel and hoisting a weighted bucket. For the case of the 10 second windows, we identified the 30 most important frequency features and used the inverse Fourier transform to reconstruct the three acceleration signals. For time-frequency domain (wavelet) features, after finding the important features, we used the coefficients from which the statistics were derived to reconstruct the signal using the inverse wavelet transform. Therefore, we used the 15 most important features for the reconstruction. Since the LIME method is model agnostic, we performed the analysis with only one classifier, i.e., k-NN. Moreover, since LIME is a local interpretation approach, we applied the method on 30 observations that were correctly classified for the two tasks. The feature contributions were averaged over the 30 observations.

3.4.1 Electrical panel.

The top panel in Fig 7 presents the feature importance for the electrical panel task from LIME method for the frequency and time-frequency features. From this figure, the features from Z acceleration are the most important in both feature sets. Also, from the time-frequency domain features, the second and third most important features are selected from the X acceleration signals. These observations are in contrast with the overall feature importance from the RF classifier, and reaffirm that feature importance should be investigated for individual classes. We selected a sample observation out of the 30 observations used to calculate the feature importance for electrical panel and reconstructed the signal using the important features. The reconstructions from the frequency and time-frequency features are presented in the bottom panels of Fig 7.

thumbnail
Fig 7. LIME interpretation: Class specific feature importance and reconstruction for electric panel.

https://doi.org/10.1371/journal.pone.0261765.g007

For the frequency domain features, only two (out of 30) were selected from the X signal. The reconstruction shows a simple periodic pattern. From the time-frequency domain features, three (out of 15) features were selected from the X signal and the reconstruction from the time-frequency domain features set showed an emphasis on the abrupt changes from the X signal to classify the task.

The important frequency features for the Y signal were mainly selected from higher frequencies and therefore the reconstruction does not capture the dominant pattern in the signal. This could be due to similar dominant pattern in the Y signal among different tasks and more variation in the Y signal for the electrical panel. However, the majority of important time-frequency domain features were selected from the last approximation coefficient. As can be seen in the figure, this helped capture the 1st two peaks that were present in the signal, and the more steady pattern present towards the end. Moreover, two other features were selected from the 1st two detail coefficients of the Y signal, which appear as more noise in the reconstruction.

Comparing the Z acceleration signal against the other two highlights the dominance of the Z acceleration signal for this task. In frequency domain, the first six important features were from the Z acceleration. In the time-frequency domain, the most important feature was also selected from the Z axis.

3.4.2 Hoisting.

Fig 8 presents the LIME feature importance and reconstruction results for the hoisting task. From the original signals in the bottom panels it can be observed that, unlike the electrical panel, there was no one acceleration component that dominated in terms of magnitude or pattern.

thumbnail
Fig 8. LIME interpretation: Class specific feature importance and reconstruction for Hoisting.

https://doi.org/10.1371/journal.pone.0261765.g008

Frequency features from all three acceleration signals were among the most important features, with similar levels of contribution. However, the majority of the selected features from the X signal were from frequencies higher than 3 Hz The reconstructions show that for the X signal, the high frequency values were selected as the distinctive characteristics of the activity. Comparing the original Y and Z signals against the X signal, it is evident that the X acceleration values include larger variation, and this can be a valid distinctive factor between hoisting and other activities.

For the time-frequency domain, the four most important features had higher contributions compared to the remaining selected features. Only two features from the Y acceleration were selected, and this is reflected in the reconstructions for the X and Z signals, where the shapes of their original counterparts were captured more accurately.

4 Discussion and conclusions

4.1 Summary of the main contributions

In this paper we examined the utility of a single accelerometer, worn on the dominant wrist, for the classification of common work activities among ELWs. We evaluated three main research questions that pertain to (1) the utility/accuracy of individualized models for classifying ELW tasks, (2) the feasibility of constructing a generalized model that does not need to be tuned for new workers, and (3) communicating the relative importance and impact of our proposed/engineered features on the performance of the classification models. An experimental lab study was designed to assess the research questions, where 37 participants were recruited to perform 10 simulated ELW tasks.

Based on our proposed modeling approach and the experimental study, this paper makes five main contributions. First, we have shown that the features captured from a single wrist-worn sensor can be used to correctly classify complex and dynamic ELW tasks. Second, our classification accuracies were similar for both the individualized and generalized models. This is a practically significant finding since it presents researchers/practitioners with flexibility to train models based on a limited number of workers without the need to have ELWs perform all tasks (i.e., the ELWs’ data can be used even if they do not complete all tasks as long as they are expected to follow the prescribed work instructions). Third, the feature importance and the LIME method indicate that certain tasks show a dominant signal in one of the three Cartesian coordinates. While this can be expected from a biomechanics perspective, our ability to show and quantify these effects is important in providing justification of the suitability of our engineered features and communicating model findings in a “clear box” manner. Fourth, we showed that 10-second windows present slightly higher classification accuracies; however, their classification performance is not substantially better than the shorter 4-second window (which is the minimum time unit to guarantee a full task cycle based on our most complex task). The implication is that near-real time models for task classification can be deployed. Fifth, time-frequency features resulted in a relatively stable testing accuracy across the intra- and inter-subject deployment scenarios for three investigated machine learning models.

4.2 Relevance to human performance monitoring research/ practice

It is important to contextualize our work in the context of the existing human activity recognition (HAR) literature. The results from our experimental study advances the HAR literature in two major ways. First, our inter-subject modeling scenario is more conservative and complex than the studies highlighted in Table 1 since we (1) did not include data for any of the test subjects in our training set; and (2) had a relatively large number of tasks, 10, when compared to those studies (ranging from 3 to 11 with a median of 5). Hence, our reported accuracies of ≥ 91% for the 4-second window, which increases to ≥ 97% with 10-second windows, is excellent when compared to those studies where the accuracy ranges from 66.8% to 98.0% with a median of 90.1%. Second, we captured how features engineered from the sensor contribute to an obtained classification result. The acceleration in certain activities showed a dominant signal in one of the three Cartesian coordinates. In some cases this is due to a unique movement that associates with the relevant task and does not appear in other activities (e.g., the Z acceleration signal was the dominant signal for classifying “electrical panel” task). Moreover, if higher dispersion is a characteristic of certain activities, the features are expected to be selected from higher frequencies for that activity. The hoisting, as one of the tasks that induced the largest load on the upper body, caused some degree of quivering. As expected, the signal reconstructions for the hoisting task resulted in more noisy signals.

4.3 Limitations and suggestions for future research

There are three limitations that may impact the generalizability of our results. First, a laboratory experiment is a tightly controlled simulation of the real workplace and lacks the heterogeneities inherent in a field study. Our work was based on 37 subjects and this large number of subjects was unprecedented in previous studies. Although this can mitigate the heterogeneity issue, a field study can be better generalized to other cases. Second, our study was designed such that the repetitive tasks were performed for ten or twenty repetitions and the time-based tasks continued for three minutes. However, in the real world applications different tasks can be performed intermittently and for fewer repetitions / shorter duration. This scenario, which we call a mixed task scenario, is more challenging in terms of annotation and windowing, and may require automatic change point detection for creating the windows. Third, we did not study the impact of muscle fatigue on the performance of the proposed activity recognition framework, while fatigue presence may cause changes in the pattern of later repetitions of each task or the tasks that are performed near to the end of simulation.

Finally, we provide three recommendations for future research. First, studies should investigate how active learning can complement our framework. While our current classification accuracy is excellent for both intra-subject and inter-subject scenario, it will suffer from a limited number of training labels. To alleviate this and maintain a good accuracy, one can use active learning, where the learning algorithm is allowed to query a subject to label the task that he/she is performing. Meanwhile, a penalty for the request label action should be considered, because excessive queries can complicate the performance of the subjects. Also, as mentioned above, change point detection is a prerequisite for the mixed task scenario. Towards this end, future research should be devoted to the development of a real-time segmentation approach for detecting the task transition times. Moreover, the data between consecutive change points can be further segmented into cycles for repetitive tasks. These segments can be used to investigate the effect of fatigue on the patterns of task performance. Finally, the real-time implementation of activity recognition is of great importance in industrial applications for worker monitoring purposes. We suggest that the future research should investigate and address the challenges in real-time implementation of the proposed algorithms.

References

  1. 1. Balducci PJ, Roop JM, Schienbein LA, DeSteese JG, Weimar MR. Electric power interruption cost estimates for individual industries, sectors, and US economy; 2002. U.S. Department of Energy https://www.osti.gov/biblio/926127.
  2. 2. Kelsh MA, Lu ET, Ramachandran K, Jesser C, Fordyce T, Yager JW. Occupational injury surveillance among electric utility employees. Journal of Occupational and Environmental Medicine. 2004;46(9):974–984. pmid:15354064
  3. 3. Volberg V, Fordyce T, Leonhard M, Mezei G, Vergara X, Krishen L. Injuries among electric power industry workers, 1995–2013. Journal of Safety Research. 2017;60:9–16. pmid:28160819
  4. 4. Vergara X, Bhatnagar M, Fordyce T. Exploratory narrative text analysis to characterize tasks associated with injuries among electric utility line workers: EPRI Occupational Health and Safety Database 1995–2013. American Journal of Industrial Medicine. 2020;64(3):198–207. pmid:33373058
  5. 5. Nath ND, Chaspari T, Behzadan AH. Automated ergonomic risk monitoring using body-mounted sensors and machine learning. Advanced Engineering Informatics. 2018;38:514–526.
  6. 6. Graham M, Haarstad H. Transparency and development: ethical consumption through Web 2.0 and the internet of things. Open Development: Networked Innovations in International Development. 2014;79.
  7. 7. Ullah I, Sohail Khan M, Kim D. IoT services and virtual objects management in hyperconnected things network. Mobile Information Systems. 2018;2018.
  8. 8. Moynihan GP. Introductory chapter: background and current trends in industrial engineering. In: Concepts, applications and emerging opportunities in industrial engineering. IntechOpen; 2021.
  9. 9. Paksoy T, Kochan CG, Ali SS. Logistics 4.0: Digital transformation of supply chain management. CRC Press; 2020.
  10. 10. Maman ZS, Yazdi MAA, Cavuoto LA, Megahed FM. A data-driven approach to modeling physical fatigue in the workplace using wearable sensors. Applied Ergonomics. 2017;65:515–529.
  11. 11. Baghdadi A, Cavuoto LA, Jones-Farmer A, Rigdon SE, Esfahani ET, Megahed FM. Monitoring worker fatigue using wearable devices: A case study to detect changes in gait parameters. Journal of Quality Technology. 2021;53(1):47–71.
  12. 12. Hajifar S, Sun H, Megahed FM, Jones-Farmer LA, Rashedi E, Cavuoto LA. A forecasting framework for predicting perceived fatigue: Using time series methods to forecast ratings of perceived exertion with features from wearable sensors. Applied Ergonomics. 2021;90:103262. pmid:32927403
  13. 13. Lamooki SR, Kang J, Cavuoto LA, Megahed FM, Jones-Farmer LA. Challenges and Opportunities for Statistical Monitoring of Gait Cycle Acceleration Observed from IMU Data for Fatigue Detection. In: 2020 8th IEEE RAS/EMBS International Conference for Biomedical Robotics and Biomechatronics (BioRob). IEEE; 2020. p. 593–598.
  14. 14. Li K, Habre R, Deng H, Urman R, Morrison J, Gilliland FD, et al. Applying multivariate segmentation methods to human activity recognition from wearable sensors’ data. JMIR mHealth and uHealth. 2019;7(2):e11201. pmid:30730297
  15. 15. Zhuang Z, Xue Y. Sport-related human activity detection and recognition using a smartwatch. Sensors. 2019;19(22):5001. pmid:31744127
  16. 16. Torres-Huitzil C, Alvarez-Landero A. Accelerometer-based human activity recognition in smartphones for healthcare services. In: Mobile Health. Springer; 2015. p. 147–169.
  17. 17. Sherafat B, Ahn CR, Akhavian R, Behzadan AH, Golparvar-Fard M, Kim H, et al. Automated methods for activity recognition of construction workers and equipment: State-of-the-art review. Journal of Construction Engineering and Management. 2020;146(6):03120002.
  18. 18. Ryu J, Seo J, Jebelli H, Lee S. Automated action recognition using an accelerometer-embedded wristband-type activity tracker. Journal of Construction Engineering and Management. 2019;145(1):04018114.
  19. 19. Zappi P, Stiefmeier T, Farella E, Roggen D, Benini L, Troster G. Activity recognition from on-body sensors by classifier fusion: sensor scalability and robustness. In: 2007 3rd International Conference on Intelligent Sensors, Sensor Networks and Information. IEEE; 2007. p. 281–286.
  20. 20. Koskimaki H, Huikari V, Siirtola P, Laurinen P, Roning J. Activity recognition using a wrist-worn inertial measurement unit: A case study for industrial assembly lines. In: 2009 17th Mediterranean Conference on Control and Automation. IEEE; 2009. p. 401–405.
  21. 21. Stiefmeier T, Roggen D, Ogris G, Lukowicz P, Tröster G. Wearable activity tracking in car manufacturing. IEEE Pervasive Computing. 2008;7(2):42–50.
  22. 22. Bao L, Intille SS. Activity recognition from user-annotated acceleration data. In: International Conference on Pervasive Computing. Springer; 2004. p. 1–17.
  23. 23. Yang JY, Wang JS, Chen YP. Using acceleration measurements for activity recognition: An effective learning algorithm for constructing neural classifiers. Pattern Recognition Letters. 2008;29(16):2213–2220.
  24. 24. Chernbumroong S, Atkins AS, Yu H. Activity classification using a single wrist-worn accelerometer. In: 2011 5th International Conference on Software, Knowledge Information, Industrial Management and Applications (SKIMA) Proceedings. IEEE; 2011. p. 1–6.
  25. 25. Muehlbauer M, Bahle G, Lukowicz P. What can an arm holster worn smart phone do for activity recognition? In: 2011 15th Annual International Symposium on Wearable Computers. IEEE; 2011. p. 79–82.
  26. 26. Gjoreski H, Bizjak J, Gjoreski M, Gams M. Comparing deep and classical machine learning methods for human activity recognition using wrist accelerometer. In: Proceedings of the IJCAI 2016 Workshop on Deep Learning for Artificial Intelligence, New York, NY, USA. vol. 10; 2016. p. 970.
  27. 27. Akhavian R, Behzadan AH. Smartphone-based construction workers’ activity recognition and classification. Automation in Construction. 2016;71:198–209.
  28. 28. Tao W, Lai ZH, Leu MC, Yin Z. Worker activity recognition in smart manufacturing using IMU and sEMG signals with convolutional neural networks. Procedia Manufacturing. 2018;26:1159–1166.
  29. 29. Frank AE, Kubota A, Riek LD. Wearable activity recognition for robust human-robot teaming in safety-critical environments via hybrid neural networks. In: IEEE/RSJ International Conference on Intelligent Robots and Systems; 2019.
  30. 30. Lim S, D’Souza C. A narrative review on contemporary and emerging uses of inertial sensing in occupational ergonomics. International Journal of Industrial Ergonomics. 2020;76:102937. pmid:33762793
  31. 31. Ogris G, Lukowicz P, Stiefmeier T, Tröster G. Continuous activity recognition in a maintenance scenario: combining motion sensors and ultrasonic hands tracking. Pattern Analysis and Applications. 2012;15(1):87–111.
  32. 32. Khan SH, Sohail M. Activity monitoring of workers using single wearable inertial sensor. In: 2013 International Conference on Open Source Systems and Technologies. IEEE; 2013. p. 60–67.
  33. 33. Lin F, Song C, Xu X, Cavuoto L, Xu W. Sensing from the bottom: Smart insole enabled patient handling activity recognition through manifold learning. In: 2016 IEEE First International Conference on Connected Health: Applications, Systems and Engineering Technologies (CHASE). IEEE; 2016. p. 254–263.
  34. 34. Joshua L, Varghese K. Selection of accelerometer location on bricklayers using decision trees. Computer-Aided Civil and Infrastructure Engineering. 2013;28(5):372–388.
  35. 35. Malaisé A, Maurice P, Colas F, Charpillet F, Ivaldi S. Activity recognition with multiple wearable sensors for industrial applications. In: The Eleventh International Conference on Advances in Computer-Human Interactions; 2018.
  36. 36. Malaisé A, Maurice P, Colas F, Ivaldi S. Activity recognition for ergonomics assessment of industrial tasks with automatic feature selection. IEEE Robotics and Automation Letters. 2019;4(2):1132–1139.
  37. 37. Zhang M, Chen S, Zhao X, Yang Z. Research on construction workers’ activity recognition based on smartphone. Sensors. 2018;18(8):2667. pmid:30110892
  38. 38. Yang Z, Yuan Y, Zhang M, Zhao X, Tian B. Assessment of construction workers’ labor intensity based on wearable smartphone system. Journal of Construction Engineering and Management. 2019;145(7):04019039.
  39. 39. Akhavian R, Behzadan AH. Coupling human activity recognition and wearable sensors for data-driven construction simulation. Journal of Information Technology in Construction. 2018;23:1–15.
  40. 40. Cezar G. Activity recognition in construction sites using 3D accelerometer and gyrometer; 2012. Stanford University https://stanford.io/2Ryfwrs.
  41. 41. Lim TK, Park SM, Lee HC, Lee DE. Artificial neural network–based slip-trip classifier using smart sensor for construction workplace. Journal of Construction Engineering and Management. 2016;142(2):04015065.
  42. 42. Fridolfsson J, Arvidsson D, Doerks F, Kreidler TJ, Grau S. Workplace activity classification from shoe-based movement sensors. BMC Biomedical Engineering. 2020;2(1):1–8. pmid:32903356
  43. 43. Joshua L, Varghese K. Construction activity classification using accelerometers. In: Construction Research Congress 2010: Innovation for Reshaping Construction Practice; 2010. p. 61–70.
  44. 44. Joshua L, Varghese K. Accelerometer-based activity recognition in construction. Journal of Computing in Civil Engineering. 2011;25(5):370–379.
  45. 45. Yang K, Aria S, Ahn CR, Stentz TL. Automated detection of near-miss fall incidents in iron workers using inertial measurement units. In: Construction Research Congress 2014: Construction in a Global Network; 2014. p. 935–944.
  46. 46. Antwi-Afari MF, Li H, Umer W, Yu Y, Xing X. Construction activity recognition and ergonomic risk assessment using a wearable insole pressure system. Journal of Construction Engineering and Management. 2020;146(7):04020077.
  47. 47. Berglund ME, Duvall J, Dunne LE. A survey of the historical scope and current trends of wearable technology applications. In: Proceedings of the 2016 ACM International Symposium on Wearable Computers; 2016. p. 40–43.
  48. 48. Al-Eidan RM, Al-Khalifa H, Al-Salman AM. A review of wrist-worn wearable: Sensors, models, and challenges. Journal of Sensors. 2018;2018.
  49. 49. Ribeiro MT, Singh S, Guestrin C. “Why should I trust you?” Explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; 2016. p. 1135–1144.
  50. 50. Preece SJ, Goulermas JY, Kenney LP, Howard D, Meijer K, Crompton R. Activity identification using body-mounted sensors—a review of classification techniques. Physiological Measurement. 2009;30(4):R1. pmid:19342767
  51. 51. Banos O, Galvez JM, Damas M, Pomares H, Rojas I. Window size impact in human activity recognition. Sensors. 2014;14(4):6474–6499. pmid:24721766
  52. 52. Webb S. Deep learning for biology. Nature. 2018;554 (7693).
  53. 53. Rudin C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence. 2019;1(5):206–215. pmid:35603010
  54. 54. Maurer U, Smailagic A, Siewiorek DP, Deisher M. Activity recognition and monitoring using multiple sensors on different body positions. In: International Workshop on Wearable and Implantable Body Sensor Networks (BSN’06). IEEE; 2006. p. 4–pp.
  55. 55. Vanrell SR, Milone DH, Rufiner HL. Assessment of homomorphic analysis for human activity recognition from acceleration signals. IEEE Journal of Biomedical and Health Informatics. 2017;22(4):1001–1010. pmid:28682268
  56. 56. Lara OD, Labrador MA. A survey on human activity recognition using wearable sensors. IEEE Communications Surveys & Tutorials. 2012;15(3):1192–1209.
  57. 57. Antonsson EK, Mann RW. The frequency content of gait. Journal of Biomechanics. 1985;18(1):39–47. pmid:3980487
  58. 58. Bhat G, Deb R, Chaurasia VV, Shill H, Ogras UY. Online human activity recognition using low-power wearable devices. In: 2018 IEEE/ACM International Conference on Computer-Aided Design (ICCAD). IEEE; 2018. p. 1–8.
  59. 59. Khatun A, Hossain S, Sarowar G. A fourier domain feature approach for human activity recognition & fall detection. arXiv preprint arXiv:200305209. 2020;.
  60. 60. Zhang S, Rowlands A, Murray P, Hurst T. Physical activity classification using the GENEA wrist-worn accelerometer. Medicine & Science in Sports & Exercise. 2012;44(4):742–8. pmid:21988935
  61. 61. Sifuzzaman M, Islam MR, Ali M. Application of wavelet transform and its advantages compared to Fourier transform. Journal of Physical Science. 2009;13:121–134.
  62. 62. Tian Y, Zhang J, Wang J, Geng Y, Wang X. Robust human activity recognition using single accelerometer via wavelet energy spectrum features and ensemble feature selection. Systems Science & Control Engineering. 2020;8(1):83–96.
  63. 63. Avci A, Bosch S, Marin-Perianu M, Marin-Perianu R, Havinga P. Activity recognition using inertial sensing for healthcare, wellbeing and sports applications: A survey. In: 23th International Conference on Architecture of Computing Systems 2010. VDE; 2010. p. 1–10.
  64. 64. Najafi B, Aminian K, Paraschiv-Ionescu A, Loew F, Bula CJ, Robert P. Ambulatory system for human motion analysis using a kinematic sensor: monitoring of daily physical activity in the elderly. IEEE Transactions on Biomedical Engineering. 2003;50(6):711–723. pmid:12814238