Figures
Abstract
Activity recognition modelling using smartphone Inertial Measurement Units (IMUs) is an underutilized resource defining and assessing work efficiency for a wide range of natural resource management tasks. This study focused on the initial development and validation of a smartphone-based activity recognition system for excavator-based mastication equipment working in Ponderosa pine (Pinus ponderosa) plantations in North Idaho, USA. During mastication treatments, sensor data from smartphone gyroscopes, accelerometers, and sound pressure meters (decibel meters) were collected at three sampling frequencies (10, 20, and 50 hertz (Hz)). These data were then separated into 9 time domain features using 4 sliding window widths (1, 5, 7.5 and 10 seconds) and two levels of window overlap (50% and 90%). Random forest machine learning algorithms were trained and evaluated for 40 combinations of model parameters to determine the best combination of parameters. 5 work elements (masticate, clear, move, travel, and delay) were classified with the performance metrics for individual elements of the best model (50 Hz, 10 second window, 90% window overlap) falling within the following ranges: area under the curve (AUC) (95.0% - 99.9%); sensitivity (74.9% - 95.6%); specificity (90.8% - 99.9%); precision (81.1% - 98.3%); F1-score (81.9% - 96.9%); balanced accuracy (87.4% - 97.7%). Smartphone sensors effectively characterized individual work elements of mechanical fuel treatments. This study is the first example of developing a smartphone-based activity recognition model for ground-based forest equipment. The continued development and dissemination of smartphone-based activity recognition models may assist land managers and operators with ubiquitous, manufacturer-independent systems for continuous and automated time study and production analysis for mechanized forest operations.
Citation: Becker RM, Keefe RF (2022) A novel smartphone-based activity recognition modeling method for tracked equipment in forest operations. PLoS ONE 17(4): e0266568. https://doi.org/10.1371/journal.pone.0266568
Editor: Peng Chen, Second Institute of Oceanography Ministry of Natural Resources, CHINA
Received: January 4, 2022; Accepted: March 22, 2022; Published: April 6, 2022
Copyright: © 2022 Becker, Keefe. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: The data that support the findings of this study are openly available at https://doi.org/10.7923/1VQA-M812.
Funding: This research was funded, in part, through the Idaho Forest Utilization Research (FUR-UIEF-FY20) program. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Time and motion studies are used widely in forest operations and provide insight for improving efficiency and reducing delays in forest operations through the observation, analysis, and quantification of productive cycles and work elements [1–11]. Traditionally, harvest cycle times and individual elements are estimated using stopwatches, video recorders and handheld data loggers [12]. These studies are often labor intensive and cost prohibitive due to the manual, observational sampling they require. The application of production functions and treatment cost simulation models developed from time and motion studies is often limited in scope due these constraints [13, 14]. The methods by which productive work is defined in forestry and natural resource management are rapidly evolving, where traditional sampling techniques are increasingly replaced by less labor-intensive methods resulting from advances in real-time spatial data collection, activity recognition, remote sensing, and wearable and mobile technologies [15].
In forest operations, global navigation satellite systems (GNSS) have been successfully used in various forestry applications to automate the sampling and production analysis of equipment, monitor the movements of ground workers, and assess equipment and ground-worker interactions for both safety and production purposes in lieu of or in support of traditional time and motion methods [16, 17]. GNSS-supported approaches for production analysis have been developed for timber harvesting equipment including skidders [18–22], forwarders [14], and cable logging carriages [23]. GNSS devices work particularly well for equipment such as forwarders and skidders due to the large distances they cover and the relationship of their movements, speed, and location with their productive activities [14]. In harvester studies, Standard for Forest Data (StandForD) and CAN bus data have been used in combination with GNSS devices to perform production analysis [12, 24–26]. Further, Carter et al. successfully used GNSS to monitor traffic intensity and soil impacts resulting from timber harvesting [27].
One limitation of many basic GNSS devices is the inability to share and process data in real-time independently of additional hardware, limiting their usefulness for production analysis where immediate feedback and processing may be desired. GNSS-RF uses radio frequency to transmit and share data in real-time independent of cellular and satellite infrastructure, enabling data sharing in remote work environments and off-the-grid scenarios. The use of these technologies for improving safety and situational awareness in forest management and logging scenarios has received significant interest [28–31]. Further, real-time communication between multiple GNSS-RF devices has successfully been used to assess independently moving components of a single machine to characterize productive work elements [32, 33]. Classification of component work elements using consumer-grade GNSS-RF is limited by the spatial resolution and accuracy at which data can be recorded [32, 33]. GNSS accuracy degradation due to multipath error from dense forest canopies or topography further limits the extent spatially explicit data can be used for production analysis in forest operations [14, 16, 19, 34]. This may prohibit the exclusive use of GNSS-based devices for data collection and production representation where high-accuracy position, navigation, and timing (PNT) data are required to ensure worker safety and effective work assessment [35], although these studies predated recent advances in RTK system availability for mobile devices. When GNSS error is unacceptably large, additional data inputs may be necessary to increase the robustness of production models and ensure their improved accuracy and utility across varying sites and applications. Sensor-based data, including those derived from triaxial accelerometers and gyroscopes, provide a rich, high-resolution, and highly descriptive data source for characterizing equipment activities. Wearable and mobile devices, including smartphones and smartwatches, are regularly equipped with inertial measurement units (IMUs), GNSS chipsets, and other sensors that provide detailed information regarding the environment in which the device is located [36, 37]. Inertial measurement units (IMU) contain both a triaxial accelerometer and gyroscope and enables the measurement of acceleration and angular velocity along three orthogonal axes and are regularly used in activity recognition modelling [38].
Activity recognition refers to the prediction of an individual’s actions based on sensor-derived data capturing their movements or other biophysical characteristics and was first deployed in the lifecare and wellness industries in the 1980’s [39]. Smartphone and smartwatch ubiquity has led to rapid expansion in their use for human activity recognition modeling techniques characterizing physical activities. Data collected from smart device-based accelerometers, gyroscopes and sound pressure meters have been used successfully in various human activity recognition applications [36, 37, 40, 41]. This includes uses in healthcare, behavior modelling, commerce, traffic condition monitoring, transportation identification, intelligent environments, fall detection, and to assess the wellness and fitness of athletes [37, 41–49]. Zhang et al. [50] and Akhavian and Behzadan [51] studied the use of smartphones for activity recognition and classification in the construction industry, showing the ability of smartphones to detect and define human-performed work activities. While sensor-based activity recognition has proven effective, model development and evaluation in the natural resource management field provides new opportunities to quantify and log both human and equipment-based work activities simultaneously within a common framework. Research on the development of consumer accessible smartphone and smartwatch-based activity recognition solutions for natural resource industries is lacking.
Previous activity recognition modelling using accelerometers, GNSS devices, and gyroscopes in natural resource industries has examined applications for motor-manual felling, motor-manual willow coppice operations, bandsaw mills, and carriages in cable logging operations [11, 52–55]. Pierzchala et al. [56] used a combination of smartphone derived IMU and GNSS data, video recording, and tethered onboard computers to perform automatic work phase recognition and subsequent quantification for a cable logging carriage with activity segmentation success of 78%. Keefe et al. [40] was the first successful use of stand-alone smartphone-based data collection for the development of human activity recognition models in forestry by assessing manual fallers in industrial forest operations. Using a smartphone secured to the belt of a timber faller, the activity recognition model characterized manual felling work elements and delay with accuracies between 65.9% and 99.6%. More recently, smartwatch sensors have been successfully used to develop activity recognition models for rigging crew workers setting and disconnecting log chokers on cable logging operations [57]. These new smart device-based activity recognition applications show promise for further integration into forest industries. There is a significant need for continued development of models for real-time production analysis and logistics. For new automated technologies to be used for time study research purposes, McDonald and Fulton [20] suggested several requirements need to be met: 1.) It must be unobtrusive, simple, and easy to implement; 2.) It must be applicable for a wide range of machinery and site conditions without need for major reconfiguring; 3.) It must be durable and robust; and 4.) It must produce data of comparable accuracies to that collected by a skilled field crew. Smartphones and smartwatches have been shown to provide a technological solution capable of meeting these evaluation metrics.
There is no currently published research on the development of stand-alone smartphone-based activity recognition models to predict work cycle elements and production rates of heavy equipment in forestry. Smartphones are useful for activity recognition model development due to their ubiquity, unobtrusiveness, ease of use, portability, high storage capacity, powerful computing, powerful embedded sensors, low energy consumption, extensive developer support and programmable capabilities [58–60]. Integrated production and logistic monitoring systems for collecting and processing data regarding machine and operator condition and performance are found in new forest equipment and exist for harvesters, forwarders, and other equipment [24, 25, 56]. However, in most instances in the United States, these are typically machine specific, proprietary hardware and software systems that are limited to a single machine, application, and manufacturer. Mobile activity recognition tools using smartphone-based IMUs present a standalone, mobile solution enabling integration, utilization, and customization for a wide range of mechanized equipment and work activities through a single device, similar to fitness tracking applications. Forest equipment movements entail unique, but significant, actions including lateral movements, swings, hydraulic boom manipulation, and hydraulic cutting or processing head operations. The distinct work elements and the cycles they comprise may enable the development of activity recognition-based applications based on equipment movements within their environments.
Accurately capturing, recording, and characterizing the elements of equipment is key to successful activity recognition development and opens the door to correlating equipment productivity analysis to environmental and site characteristics. The production and costs of mechanized felling, skidding, and processing equipment are directly impacted by forest stand metrics (tree size, stand density, undergrowth) and geophysical attributes (slope, rocks, debris, soils) [6, 8, 61]. Representing geographic and forest stand characteristics and their resulting impacts on equipment trafficability, tractability, and productivity is important for developing accurate, reliable, and robust activity recognition models for automating work activity characterization.
Remotely sensed data may support activity recognition model development by providing descriptive information related to site and stand conditions within the work environment. Light detection and ranging (lidar) is used extensively in forest, range, and fire management to quantify vertical forest structure, succession, and other forest attributes, as well as terrain morphology [62–70]. Use of lidar to conduct inventories and to characterize geophysical and ecological conditions across temporal and spatial scales can reduce inventory costs and improve the resolution of data for land management agencies and the forest industry [71]. In forest operations, lidar has been used to locate skid roads, log landings [64], detect disturbance related to harvest [72–74], quantify harvesting productivity [8], delineate harvest units and cable corridors [75, 76], and perform harvest system classification [77]. Additionally, Maltamo et al. (2019) assessed the accuracy of stand stem distributions derived using single-tree data from a harvester with an integrated positioning system and lidar metrics [78].
Area-based and individual tree-based approaches for deriving forest information from lidar data have been studied and compared extensively [79–83]. With improvements in lidar technologies and increased pulse densities, individual tree-based analysis has rapidly increased [84]. Pulse densities of 5–10 per square meter enable the efficient detection and segmentation of individual trees, and many current lidar acquisitions are easily within or exceed this pulse density [79]. Single tree detection and delineation is becoming a common tool for forestry applications including characterizing individual tree attribute distributions at the population level, modeling individual tree growth, classifying wildlife habitat, estimating carbon sequestration and other ecosystem services, modeling fire behavior, and precisely estimating biomass for inventory and operational applications [85–87]. Individual tree metrics including height, diameter at breast height (DBH), canopy base height, crown condition, crown area, biomass, stem volume, and specie can be derived from segmented tree data [64, 71, 79, 88]. The University of Idaho Experimental Forest has incorporated lidar-based single tree inventory into management operations for applied research, demonstration and teaching purposes since late 2020. However, to effectively use these high-resolution, remotely sensed data for operational forestry applications, it is also necessary to accurately position mobile heavy equipment during regular work conditions in the forest. Lidar-derived single tree products can have high geographic precision and accuracy, but positional data derived from GNSS-based devices can return high positional error. The inaccurate positioning of equipment or attachments in relation to remotely sensed products may lead to invalid assumptions regarding the relationship between these two data types.
The integration of smart devices monitoring equipment movements with lidar-derived single-tree information is critical to advance precision forest operations. Precision forestry, which commonly utilizes remotely sensed data sources and analysis, is the forest management technique emphasizing data-intensive and innovative practices, technologies, and processes to increase productivity, reduce costs, reduce negative site impacts resulting from management, and increase overall forest management success [89, 90]. These concepts are used in forest inventory and forest planning and have many opportunities for advancing industrial forest operations, as described above, while also allowing forest management planning and implementation at microsite levels [18]. Advances in precision forestry could provide greater insight into operational trends, processes, production, and machine interactions during logging operations for both production and safety analysis. In turn, this may lead to more sustainable, efficient, and safe forest operations. The expansion of smartphone and mobile-based internet of things (IoT) in forested environments through activity recognition will help evolve the definition, monitoring, and assessment of work in the woods.
This study focuses on the development and validation of a smartphone-based activity recognition modelling framework for forestry equipment. Our first objective is to develop equipment specific, smartphone-based activity recognition models for an excavator-based masticator using random forest machine learning algorithm. Our second objective is to evaluate the effects of sliding window sizes, window overlap proportion, and sampling frequency on activity recognition model performance. Lastly, we evaluated the accuracy of matching equipment body and attachment locations derived from consumer-grade GNSS-RF devices to lidar-derived individual tree locations during operational mastication work.
If successful, the development, validation, and future implementation of smartphone-based activity recognition models could assist land managers and operators by providing a ubiquitous system for continuous and unattended analysis of mastication treatments. Additionally, successful integration of lidar-derived single tree inventories and location-enabled activity recognition may facilitate the enhanced evaluation of forest and operational factors affecting treatment production and cost analytics in real-time. This will simultaneously provide data which can be used in planning and implementation of future forest operations and management. If successful, identifying masticated trees within a single-tree stem map could also foster accurate prediction of residual stand density, residual fuel characteristics, and other prescription features to inform and prioritize cost-effective long-term management planning scenarios.
Materials and methods
Study site
Three replicated stands on the University of Idaho Experimental Forest (UIEF) were sampled as part of a larger fuels management operation with treatment units located in three UIEF management units: Flat Creek; West Hatter; and East Hatter. The UIEF is located approximately 20 kilometers northeast of Moscow, Idaho in the Palouse Range. All sampled stands were approximately 27-year-old ponderosa pine (Pinus ponderosa) plantations established following a clearcut timber harvest and subsequent prescribed burn. The primary fuel treatment prescription was the reduction of stand density to a tree spacing of 5 x 5 meters using mastication. This pre-commercial treatment focused on reduction of stand density, treatment of western pine beetle damage, and hazardous fuels reduction for the residual stand.
Mastication sampling and field procedures
A time and motion study was performed within the three stands to facilitate the observation of mastication work cycles with corresponding and simultaneously collected smartphone sensor data. Treatments were implemented with a tracked excavator (Kobelco ED150) running a hydraulic disk mastication head (Promac Brush Cutter). Due to hazardous conditions for observers during mastication and the density of the stands, observational sampling was supplemented with video recording captured by two Garmin Dash Cam 45 (Garmin®, Olathe, KA, USA) devices placed within the machine cab. These devices provided time-stamped video and audio recordings, enabling productive work cycle classification while increasing observer safety and reducing sampling errors associated with poor visibility, observer inattentiveness or element misclassification. GNSS-RF multi-transponder devices with 2.5 second sampling intervals were secured outside the cab of the equipment to record movements and treatment progression throughout sampling. Data recorded from these devices was shared in real-time via radio transmission to a handheld receiver.
Google Pixel 2 smartphones (Google, Mountain View, CA, USA) collected the sensor data necessary for activity recognition modelling. These devices were placed within the machine cab oriented with the rear of the phones facing the front windshield. The AndroSensor application [91] was used to record the IMU and other smartphone sensor measurements. Within the AndroSensor application, a unique data sampling rate frequency was selected for each of the three phones used: 10, 20, and 50 hertz (Hz). The 10, 20, and 50 Hz sampling rates recorded information every 0.1, 0.05, and 0.02 seconds respectively. Device placement on the equipment used in field sampling is shown in Fig 1.
In human activity recognition, sampling frequencies vary greatly from 1 Hz [92] to 200 Hz [93, 94]. 50 Hz is often used to capture dynamic activities [58, 95], though 20 Hz has been shown to provide acceptable recognition accuracy while also preserving smartphone battery life [37, 45, 46, 96]. While decreasing sampling rate can increase the battery life of devices, reducing frequencies too drastically can limit the potential for achieving accurate activity recognition [92]. Collecting data at a high sampling rate (50 Hz) enables data to be downsampled to lower frequencies for additional analysis [60, 97, 98]. Therefore, the 50 Hz sensor data was downsampled to 20 Hz and 10 Hz to determine if downsampled data returned similar activity recognition classification results as natively sampled signals of the same frequencies.
Sensor data for acceleration, linear acceleration, orientation, and sound level was measured during field sampling using the Google Pixel 2. The IMU in the smartphone is a combination accelerometer and gyroscope that detects triaxial acceleration measurements in the x, y, and z directions in meters per second squared (m s-2), triaxial linear acceleration measurements in the x, y, and z directions in meters per second squared (m s-2), and triaxial orientation or tilt using the gyroscope in radians per second (rad s-1). Sound pressure levels were measured in decibels (dB) using an integrated sensor. Each recorded data point is assigned a date and time stamp indicating the year, month, and day and the associated hour, minute, second, and millisecond (Y-M-D h:m:s:SSS).
Productive work cycles and cycle elements were defined using the audio and video recordings obtained from the dash cameras used during sampling and were subsequently merged with smartphone-derived sensor data. Cycle elements for the masticator are described in Table 1 below.
Activity recognition modelling and assessment
Smartphone-derived sensor data processing.
Activity recognition modelling using random forests and other shallow learning algorithms primarily consist of four main phases: data collection; segmentation; feature extraction and selection; and classification [37, 59, 99]. Once data is collected, extracted from the smartphones, and work elements for the entirety of the data set are manually added to the data based on corresponding time stamps, data was imported into the R statistical programming environment where all remaining analysis and modelling occurred [100]. The segmentation phase of activity recognition entailed dividing the full data set into smaller time segments to simplify data retrieval. This process consisted of two steps: signal preprocessing and windowing [37, 59]. The Google Pixel 2 accelerometer and gyroscope report raw data in three axes (x, y, and z) enable the detection of orientation changes. To reduce the impact of orientation on activity recognition performance, the vector magnitudes of the triaxial sensors were calculated using Eq 1 [99, 101, 102]. This resulted in orientation-independent measures for acceleration, linear acceleration, and orientation/tilt.
In activity recognition, sliding windows are often used to segment data into set time series where defined time domain features are calculated [37, 38, 43, 45, 46, 95, 103]. Studies have examined and compared model accuracies using multiple window sizes ranging from 1 to 13 seconds [40, 43, 45, 57, 92, 95, 103–110]. Activity recognition accuracy is greatly influenced by window size, with windows that are too short not covering the extent of the activity being classified and windows that are too long overlapping multiple unrelated activities [46]. Based on our study objectives and prior research, four different windows (1, 5, 7.5 and 10 seconds) were assessed to balance the ability of short windows to recognize simple activities and the ability of longer windows to capture less repetitive activities effectively [37, 40, 111]. Assigned window sizes are passed through the entire data set incrementally, with either overlapping or non-overlapping windows [112]. The strength of the impact that window size has on classification is dependent on the amount of overlap of subsequent windows [46]. While overlap enables the classification of activities in smaller time intervals than the window, it does require greater computational capacity than non-overlapping windows [104]. 50% window overlap is commonly used in activity recognition modelling [38, 43, 51, 95, 110, 111], though other window overlaps, including 90%, have been used successfully [40, 57, 105, 109]. Window overlaps of 50% and 90% were selected for this study.
Within each sliding window, time domain features were derived using mathematical and statistical techniques [113]. Nine time domain features were calculated: including the mean, maximum, minimum, kurtosis, variance, standard deviation, skewness, root mean square (RMS) and signal vector magnitude. These selected features were comparable with those used in Zhang et al. [50], Keefe et al. [40], Shoaib et al. [111], and Zimbelman & Keefe [57]. Mean is a versatile metric for various types of sensors for discriminating between rest and activity in multi-activity recognition and assists in preprocessing data by removing random spikes and noise [46, 113]. Similarly, RMS, signal vector magnitude and standard deviation are all effective statistics for activity classification and distinguishing between periods of rest and activity [46]. Skewness is used to measure “the degree of asymmetry of a probability density curve compared to the average” [50]. The purpose of sampling and data collection of various sensors is not to restore the raw signals of activities, but to detect different activities according to their statistical properties using the aforementioned time domain features [92].
Random forest.
Following windowing, the data were separated into training and testing sets with a 70/30 split, respectively, to provide an unseen partition of data for final model assessment. Stratified random sampling was applied to each cycle element class to preserve element distribution ratios between testing and training data. Random forest (RF) was selected to develop machine learning algorithms for the productive cycle element classification. RF is an ensemble classifier that produces multiple decision trees using bagging and randomly selected subsets of training samples and variables to provide a majority vote from which a prediction is made [114–118]. Excellent classification results, speed of processing, and the ability to reduce bias, correlation, and overfitting compared to other classification and regression trees (CART) models make RF one of the most widely used machine learning methods available [118–121]. In activity recognition modelling, random forest has been widely tested with high levels of resulting classification success [40, 43, 47, 50, 57, 102, 103, 105, 108]. Performance was found to exceed 90% in instances when using multiple smartphone-based sensors and RF classifiers [36, 50].
Forty unique RF algorithms were developed to account for all combinations of sampling rate (10 Hz, 20 Hz, 50 Hz, 10 Hz downsampled, 20 Hz downsampled), window overlap (50% and 90%), and sliding window length (1 second, 5 seconds, 7.5 seconds, 10 seconds). The workflow for all data processing and model development is found in Fig 2. The R package caret (version 6.0–88) was used for all random forest training and development [122].
Hyperparameters, specifically mtry and ntree, are adjustable components of random forest which can increase the performance of algorithms when properly tuned. The mtry parameter determines the number of variables split at each node of the decision trees and ntree is the number of trees used in the forest [118]. Repeated k-fold cross validation with 10 folds and 3 repeats was used on training datasets to train and validate mtry and ntree values for the machine learning algorithms. Mtry values (1–25) and ntree values (1–550) were assessed during this process. The final selected value for mtry was chosen based on the maximum accuracy achieved. For ntree, it is understood that prediction accuracy increases at a decreasing rate as trees are added and larger values for ntree increase computational demands [114]. To balance processing time and model performance, ntree was selected by determining the point at which out of box error (OOB) and accuracy (%) of the model stabilized [123]. These hyperparameter values were then applied to the final training of the random forest algorithms which then predicted the test data.
The multiclass area under the curve (AUC) of the receiver characteristic operator (ROC) was used to evaluate the individual element characterization and overall activity recognition algorithm performance using the multiclass.roc function of the R pROC package (version 1.17.0.1) [124]. This measure of performance is commonly used for binary and multiclass classification algorithms [125–128] and performs well when working with imbalanced datasets [57, 126, 129–131]. The AUC for ROC represents the probability of a classifier ranking a randomly drawn positive before a randomly drawn negative [132]. In general, AUC provides an aggregate measure of classifier performance across all thresholds. This value falls between 0 and 1, with higher values indicating better performance [133]. AUCs between 0.5 and 0.7 are considered low accuracy; 0.7 and 0.9 are moderate accuracy; and accuracy is considered high when AUC exceeds 0.9 [125]. These values were converted to percentages. The AUC of all 40 trained random forest models were compared to determine which combination of window size, window overlap, and sampling frequency returned the highest AUC. The highest value indicated the best performing model overall.
Final model performance was also assessed using the confusionMatrix function in the caret package [122]. Five metrics quantifying the classification success of the activity recognition models per cycle element were provided by this function: sensitivity, specificity, precision, F1, and balanced accuracy. These accuracy metrics determined the impact of sliding window, window overlap, and sampling frequency on activity recognition performance. Sensitivity, also referred to as recall, is the true positive rate and represents the percentage of instances where an element was correctly classified [126]. This metric is calculated by dividing the number of true positives by the sum of the true positives and false negatives. Specificity is the true negative rate and is the percentage of negative instances correctly classified [126]. To calculate specificity, the number of true negatives is divided by the sum of the true negatives and false positives. Precision is used to capture the percentage of classified elements that represent an actual occurrence of that element and is calculated by dividing the number of true positives by the sum of the true positives and false positives [126]. In a basic sense, precision represents how sure one can be of their true positives, while sensitivity represents how sure one can be that no positive classifications are missed for a particular class. F1-score, or F-measure, combines the sensitivity and precision into a single performance metric and represents the harmonic mean of these two metrics [133–135]. This is the most common performance metric used in unbalanced classification problems [136–138]. The value of the F1-score ranges from 0 to 1, with values close to one indicating high classification performance [133], though they can easily be converted to a percentage scale and were in this study. Finally, the balanced accuracy was found for each model which combines the sensitivity and specificity metrics to find their average and works well with balanced and imbalanced data [133]. The equations for F1-score and balanced accuracy are found in Eqs 2 and 3 where F1 is the F1-score, and BA is the balanced accuracy.
Lidar point cloud processing procedures
Airborne laser scanning (ALS) is a commonly used lidar acquisition method for natural resource applications and derived products include terrain elevation, vegetation heights, stand volume, tree density, and tree species identification [68, 69]. The ALS data used in this study were collected using an Optech Galaxy Prime Lidar system mounted on a fixed-wing aircraft. This was part of a 9,363 km2 (3,615 mile2), multi-agency, state-wide, lidar acquisition performed by Atlantic LLC in October 2019. Nominal point spacing of 0.33 meters and an average nominal pulse density of 9.11 pulses m-2 was achieved. Individual tree detection and segmentation derived from the lidar point cloud was performed by Northwest Management Incorporated (NMI) using a proprietary software and processing workflow associated with NMI’s ForestView™ software [139]. This point-based lidar processing resulted in a full coverage ITD stem map for the four main University of Idaho Experimental Forest (UIEF) management units. In addition to individual stem coordinates, each predicted information associated with each tree included species, height (m), crown height (m), diameter (cm), crown diameter (m), and gross volume (MBF). It should be noted that the 9.11 pulses m-2 point cloud density used to develop the individual stem map using NMI’s ForestView™ software was an experimental evaluation using lower resolution pulse densities than are typically used to develop this commercial remote sensing product. Point densities of 16–20 pulses m-2 are typically used.
Cycle element and individual stem location matching
Quantifying intensive equipment production that accounts for individual tree characteristics requires matching individual mastication events with single tree locations. The machine’s location for each mastication element was determined by aligning time series data for the GNSS-RF transponders and observed time and motion data. Once the equipment location for each mastication element was determined, it was then necessary to locate feasibly treated stems within the treatment zone of the masticator at that point in time. The equipment used had a maximum boom reach of approximately 9 meters. A circular buffer of 7 meters placed around the cab’s locations during mastication elements enabled the filtering of trees to include only those within the equipment’s reach. 7 meters was used because the mastication of standing trees does not enable full boom extension, thus limiting the overall reach of the machine.
Potential target trees were further filtered by accounting for the orientation of the machine cab within the stand in relation to the stems in the treatment zone. The cab and boom GNSS-RF devices enabled the derivation of the cab’s orientation as an azimuth. The azimuths between the equipment cab and each target tree within the associated buffer was then quantified. Cab-to-boom and cab-to-tree azimuths were compared, and it was assumed that cab-to-tree azimuths within +/- 10 degrees of cab-to-boom within the treatment zone indicated a feasibly masticated tree. All spatial analysis was performed using the sf packages in R [140]. Visual representation of the process used for identifying potentially treated trees using the stem map and GNSS-RF positions is shown in Fig 3.
Process for identifying potential target trees for mastication using GNSS-RF-based equipment locations, equipment orientation, equipment to tree orientation, and individual tree locations.
Results
Work sampling and cycle analysis
Summarized results from the mastication work sampling data are found in Table 2. A total of 88,829 seconds (~24.7 hours) of mechanical fuel treatments were observed across the three treatment units, with units 1, 2, and 3 accounting for approximately 41%, 29%, and 30% of the sampling time, respectively. The equipment performed the move element most often during work sampling, accounting for 32.2% of the overall sampling time, which corresponds directly with the most individual occurrences of a single element, 1,348. Delay and travel occurred significantly fewer times than the other elements, accounted for the smallest percentages of total duration of sampled time, but also had the highest mean element duration, largest range of element durations, and the largest standard deviations in element durations. The delay element had the largest coefficient of variation (223.8%) indicating a large variation of values around the mean. This is, in part, a direct result of the large range of observed element durations with a total of 2,244 seconds (37.4 minutes) separating the longest and shortest elements. Total sampling duration for the clear, masticate and move elements were relatively equal and accounted for 26.0, 28.7 and 32.2% of the total work sampling time respectively.
Random forest and activity recognition
Following the repeated 10-fold cross validation training and evaluation for the machine learning algorithms, 40 final random forest algorithms were evaluated using each combination of data sampling rate (Hz), sliding window size (seconds), and sliding window overlap (%). A ntree value of 150 was used for all random forests, with model training indicating a stabilization of error and accuracy at this number. This was found across all models, with training results for the 50% and 90% window overlap models for the 50 Hz sampling rate and 10 second sliding window algorithms shown in Fig 4. While larger ntree values provided slight increases in model performance, added benefits were outweighed by the significant increases in processing time. Increased processing time is a concern because of the delays caused in potential future real-time model applications. A ntree value of 150 provided an acceptable balance between model performance and processing time.
Overall activity recognition performance was quantified using the area under the curve (AUC) with results for all models (5 sampling rates, 4 window sizes, 2 window overlaps) displayed in Fig 5. In every instance, models with 90% window overlaps significantly outperformed models with identical sampling rates and window sizes but 50% window overlaps. Increasing sampling rate led to increases in AUC in most cases and similar trends were found when accounting for sliding window size. These trends were exaggerated in models with 50% sliding window overlaps. Final model selection was based on overall model AUC, with the best overall performance resulting from the 50 Hz sampling rate, 10 second sliding window, and 90% sliding window overlap combination with an AUC of 97.8%. AUC values for individual elements are found in Table 3.
Comparison of activity recognition model area under the curve (AUC) percentages for 50% and 90% sliding window overlaps across the four sliding window lengths (1, 5, 7.5, 10 seconds) for each work element (clear, delay, masticate, move, travel) and for full models (All).
The influence of sampling rate, window overlap, and window size on activity recognition sensitivity, specificity, precision, F1-score, and balanced accuracy are found in Figs 6–10, respectively. Like AUC, window overlap had the greatest impact on model performance for all metrics, with 90% models performing better than 50% overlap models. Increases in sampling rate provided marginal increases in model performance, with 10 and 20 Hz data downsampled from the 50 Hz data performing better than the natively sampled 10 and 20 Hz data for classification. Positive trends between increasing model performance and increasing sliding window size were also seen, with 10 second windows returning the highest performance metrics in most instances. Select work elements in some models returned higher performance metrics at 7.5 second windows than 10 second windows, but this was not consistent for all elements in these classification algorithms. Performance metrics for the best performing activity recognition model (50 Hz sampling rate, 10 second sliding window, 90% sliding window overlap) are found in Table 3.
Comparison of activity recognition model sensitivity percentages for 50% and 90% sliding window overlaps across the four sliding window lengths (1, 5, 7.5, 10 seconds) for each work element (clear, delay, masticate, move, travel).
Comparison of activity recognition model specificity percentages for 50% and 90% sliding window overlaps across the four sliding window lengths (1, 5, 7.5, 10 seconds) for each work element (clear, delay, masticate, move, travel).
Comparison of activity recognition model precision percentages for 50% and 90% sliding window overlaps across the four sliding window lengths (1, 5, 7.5, 10 seconds) for each work element (clear, delay, masticate, move, travel).
Comparison of activity recognition model F1-score percentages for 50% and 90% sliding window overlaps across the four sliding window lengths (1, 5, 7.5, 10 seconds) for each work element (clear, delay, masticate, move, travel).
Comparison of activity recognition model balanced accuracy percentages for 50% and 90% sliding window overlaps across the four sliding window lengths (1, 5, 7.5, 10 seconds) for each work element (clear, delay, masticate, move, travel).
The ranges of performance metrics used to assess the classification of the 5 work elements (masticate, move, clear, delay, travel) for the best model are as follows: AUC (95.0–99.9%); sensitivity (74.9–95.6%); specificity (90.8–99.9%); precision (81.1–98.3%); F1-score (81.9–96.9%); and balanced accuracy (87.4–97.7%). The delay element was correctly classified with the highest percentage of all elements and had sensitivity of 95.6%. Travel had the lowest sensitivity at 74.9%, but also achieved the highest specificity at 99.9%. The very high specificity means other work elements were rarely incorrectly classified as travel, though lower sensitivity shows actual travel events were incorrectly predicted as other elements relatively often, approximately 25% of the time. Travel element misclassification was most often associated with the move element, with 250 travel events incorrectly classified as move (Table 4). The move element had the lowest specificity at 90.8%, with 787 clear, 66 delay, 574 masticate, and 250 travel elements incorrectly classified as move. This shows the activity recognition model had the lowest ability to discern between false positives and true negatives with the move element.
Confusion matrix for the best performing activity recognition model (50 Hz sampling rate, 10 second window, 90% window overlap) indicating true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN).
The confusion matrix in Table 4 provides a detailed summary of activity recognition classification, with green shaded cells indicating correctly classified instances for each element. It is clear from the confusion matrix that the greatest misclassification occurred between the move, masticate and clear elements. The precision measures for all elements exceeded 80%, meaning over 80% of all elements predicted as a particular element were correctly classified. In all instances, the balanced accuracy exceeded 85%, showing the ability of the model to account for both positive and negative classification outcomes for elements as an average of specificity and sensitivity. Areas under the curve (AUC) exceeding 95% for all elements shows excellent overall classification accuracy and activity recognition performance for this model. Each individual event shown in the confusion matrix represents a single 10-second window used for activity classification. Models with smaller window sizes would contain larger numbers of instances within each cell because more windows were necessary to elapse the entire sampling period. The importance of the top 20 time domain features used in the final model are described by mean decrease in Gini and are found in Fig 11. Features from the gyroscope and sound derived features were consistently some of the most influential features in the models for all combinations of factors, consistent with the findings of Keefe et al. [40].
Variable importance of the top 20 time domain features used during model development for the best performing activity recognition algorithm in terms of mean decrease in Gini.
Stem and cycle element matching
A total of 11,758 individual trees were detected within the treatment units, though work sampling did not encompass the entirety of the planned treatment area due to time constraints (Table 5). During the work sampling period, 1,140 individual trees fell within the 7m buffer of the equipment cab while masticating. Only 221 of these trees were within a range of azimuths from the cab to constitute a feasibly treated tree based on the GNSS-RF coordinates derived from the cab and boom transponders. In all, 921 mastication elements were observed during work sampling, though only 890 of these elements occurred with a lidar detected tree falling within the 7m cab buffer. 315 masticate elements could also be matched with an individual tree based on machine orientation at the time of mastication. More masticate elements (315) were matched with an individual tree than there were feasible unique trees (221) because masticate elements were matched with the same unique tree on multiple occasions. It can be assumed that once a tree is masticated, it can no longer be masticated again. Therefore, a maximum of 24% of all masticate elements (221) could be matched with a unique individual tree based on data derived from lidar-based individual tree locations and GNSS-RF-derived machine locations.
This table displays the number of lidar-derived individual trees based on the ForestView™ software: within the overall treatment units; within the 7m treatment buffer of the machine cab during mastication elements; and within the cab buffer while corresponding with the machine orientation. The total number of mastication elements and the number of elements where an individual tree fell within the machine buffer and matched the machine orientation are also shown.
Discussion
Class imbalance has been shown to impact and bias classification results toward the majority classes [141]. The most common sampling methods to fix imbalanced data are over and under sampling [142]. Over sampling approaches resolve data imbalance by duplicating cases of the minority class but at the expense of computational intensity required for machine learning, and increased risk of model overfitting [142–144]. This leads to computationally demanding models. Under sampling entails the reduction of majority classes to meet minority class sample sizes but can lead to a loss of valuable data trends and information necessary for effective classification and was therefore not used in this study [137,143]. A third approach, synthetic minority oversampling technique (SMOTE), generates synthetic examples as opposed to data replication used in traditional oversampling methods, but are limited to binary classification models [143]. Ultimately, the data imbalance did not appear to have significant impacts on activity recognition model performance, even in the delay and travel minority classes, as seen in the performance metrics found in Table 3. Therefore, sampling procedures to address class imbalance in the data were forgone. This decision was further supported by Kamei et al. who found sampling techniques provided no performance benefit for classification tree models which encompasses random forest algorithms [145] and Dittman et al. [146] who concluded sampling procedures are not always a necessary step for classification because random forest classifiers’ robustness to imbalanced data.
While imbalance likely contributed to the slightly lower sensitivity of the travel element, this was not a concern, as the F1-score and balanced accuracy remained relatively high at 84.9 and 87.4%, respectively. The travel element is operationally similar in function to the move element, apart from element duration. These similarities in the move and travel elements may account for the comparatively poor sensitivity of the travel element. Here, data imbalance may have favored the majority class (move) over the minority class (travel) during classification. During classification, travel was misclassified as move more than any other element. The confusion matrix (Table 4) showed actual travel elements misclassified as move 250 times, which account for approximately 20% of the actual travel element occurrences and 77% of the total travel element misclassifications. In future work, the travel and move elements could be combined into a single element to simplify classification given their operational similarities. Primary productive cycle elements: masticate, clear, and move, returned sensitivities exceeding 80% and balanced accuracies exceeding 85% showing the strength of the model in classifying these work elements (Table 3).
Despite being a minority class, delay returned the highest performance metrics for AUC, sensitivity, precision, F1-score, and balanced accuracy among all work elements and the second highest specificity only behind the travel element. The machine was predominantly stationary during delay, simplifying the classification of this element and attributing to the comparatively high performance metrics. The limited impact of data imbalance identified in this study may vary with other datasets and should be examined further in future studies. The sensitivity (74.9–95.6%), specificity (90.8–99.9%), precision (81.1–98.3%), F1-score (81.9–96.9%), and balanced accuracy (87.4–97.7%) for element classification of the best performing activity recognition model were similar to those in other studies using random forest algorithms, though all other studies focused on human activity recognition [41, 47].
All time domain features were retained for the four sensors during modelling. To reduce the time and complexity of model processing for future iterations, the number of time domain features could be reduced by excluding less influential predictors. For example, the limited presence of linear acceleration in the top tier of time domain feature importance suggests this sensor data could potentially be excluded in future work without significant impacts to overall model performance. The five time domain features with the lowest importance were all derived from linear acceleration. The high importance ranking of the sound pressure meter derived time domain features is consistent with the findings of Keefe et al. [40]. Sound is a valuable measure for activity recognition for equipment as delay, lateral machine movements, clearing of light materials, and the mastication of standing stems all expectedly produce distinct levels of sound and should be included in future modelling efforts.
Window overlap percentage was the most influential factor associated with improved model performance for all combinations of sliding window length and sampling rate. The 50% window overlap was outperformed by the 90% window overlap by significant margins in all instances. This is consistent with findings from other studies [57, 109]. For example, the 50 Hz, 10-second sliding window model returned an overall model accuracy of 68.8% for the 50% window overlap and 84.8% for the 90% window overlap. One disadvantage of a greater window overlap is the added computational load of the model, with 90% overlap models taking longer to process than 50% overlap models. However, the increased performance metrics for activity recognition necessitates this computational cost to ensure effective and accurate models. Despite longer processing times, 90% window overlap can support real-time activity recognition by increasing the frequency at which predictions are made on the data and users receive activity characterizations [105].
Marginal improvements in performance metrics were obtained by increasing from 1, 5, and 7.5, to 10-second sliding windows for the 90% window overlap models. For example, the overall model AUC for 50 Hz, 90% window overlap models for the 1, 5, 7.5, and 10 second windows were 97.1, 97.7, 97.8, and 97.8, respectively (Fig 5). Window size impacted model performance more significantly for 50% overlap models than the 90% overlap models, which is supported by Khusainov et al. [46]. Overall AUCs for 50 Hz, 50% window overlap models for 1, 5, 7.5, and 10 second windows were 86.4, 89.6, 90.6, and 91.2, respectively. While 1, 5, 7.5, and 10 second windows can be used with similar effectiveness when using 90% window overlaps, the increased processing time required by the shorter window sizes should also be considered. Shorter sliding windows require high computational overhead because the recognition algorithm must classify more total windows than a larger window length would require over the same period of time [112]. The 10-second window provides strong model performance while minimizing this computational cost, which is an important consideration in real-time activity recognition. Increasing window length has also been shown to improve the performance of activity recognition models [40, 105, 111, 138]. This is contrary to findings of Zimbelman and Keefe who found better model performance with shorter window sizes when modeling work activities for rigging-crew workers on cable logging operations [57]. These findings may differ for other forest equipment where work element duration may favor shorter window lengths due to faster transitions between subsequent elements and the inability of longer window lengths to differentiate between short actions.
While limited, there were model performance improvements when moving from the 10 Hz and 20 Hz sampling rates to the 50 Hz rate. Highly active or “agitated” movements have been shown to be more accurately represented by higher sampling rates [37]. These improvements were more significant with the 50% window overlap models compared to the models with 90% overlap. A similar relationship was found with sliding window length. When accounting for only the models using a 90% window overlap, all models regardless of sampling rate returned strong values for AUC exceeding 90%, indicating strong classification success. The strongest classification models resulted from the 50 Hz rates. The increased performance of downsampled 10 and 20 Hz models over natively sampled 10 and 20 Hz models was likely a direct result of these data being a subsample of the 50 Hz data. The performance trends in relation to cycle element, window overlap, and window size mimicked those seen with the 50 Hz models, albeit with slight performance reductions.
Reduced battery consumption associated with sampling at lower frequencies [45, 96] was not found to be a necessary consideration in this study and computational resources needed for modelling were not impacted by sampling rate. During field sampling, all smartphones experienced similar battery draw when removed from auxiliary power, though longer periods of sampling and additional computational requirements associated with real-time activity recognition may show greater discrepancies. The use of equipment auxiliary electricity to power devices while sampling removes any concerns about sensor sampling frequency or other factors impacting battery life of devices during use. This is a distinguishing consideration when using smart devices for activity recognition on equipment versus human activity recognition where auxiliary power may be unavailable or cumbersome.
Integrating GNSS-RF derived machine positions and lidar-derived individual tree detection and segmentation to provide equipment positioning in addition to the smartphone activity recognition model was ultimately ineffective for our data. The ~25% success rate achieved matching mastication elements and individual stems is not operationally useful. The poor matching performance limits confidence in the stems which were matched, and further work should explore more effective and accurate means for this process. The exact reason for this poor performance is unclear and a combination of multiple factors may have led to the outcomes encountered. First, the accuracy of the individual tree detection (ITD) product used in this study has not been assessed. However, ITD accuracy of the ForestView™ algorithm was quantified in a separate study on the University of Idaho Experimental Forest using lidar data with an average pulse density of 20 pulses m-2 [139]. Recall of ITD ranged from 0.36 to 0.67 for varying canopy cover conditions and is comparable to other ITD algorithms [139]. It is assumed ITD recall for this study is likely less than that determined by Sparks and Smith [139] due to the lower pulse density of the lidar data used. It is possible that suppressed or intermediate trees were occluded or omitted during lidar-processing which would have reduced the number of detected stems, especially in areas of high stand density [147–150]. The mastication treatments primary focused on stand density reduction and stand improvement. Stems that were masticated were often smaller than those left in the residual stand. The use of terrestrial-based lidar may prove a more effective means of producing precise and accurate stems maps used for this application in future work. Mobile terrestrial laser scanning systems help reduce occlusion and omission and enable easier acquisition of multi-temporal lidar data sets [151, 152].
Increasing the precision and accuracy of the GNSS positioning of the equipment during the operation is another area for improvement in future work. Improving equipment positioning relative to individual stems detected from lidar would likely increase the accuracy of detecting removed stems. It is well established that forest canopies impact the accuracy of GNSS devices. However, the increased availability of consumer technologies to improve the GNSS accuracy of mobile devices provide new opportunities to enhance real-time positioning and logistics in forested environments. External smartphone antennas, raw GNSS data collection, availability of consumer-grade devices with differential correction and improved dual-frequency GNSS chipsets available for the newest generations of smartphones provide opportunities to greatly increase GNSS accuracy [153–159]. These technological advances may reduce the need for dedicated GNSS-RF devices and streamline activity recognition and positioning integration. This could in turn reduce variability associated with fusing sensor data from multiple devices.
Conclusion
We have shown that models developed from the smartphone-based sensors placed in machine cabs can be used to characterize activities of mastication treatments to reduce fuels in forested stands with balanced accuracies for elements between 87.4% and 97.7%. As the first known smartphone-based activity recognition study for mechanical forestry equipment, it was necessary to examine multiple factors impacting recognition accuracy and set the foundation for future equipment activity recognition modeling research. By demonstrating the application of this technology for automated productivity analysis, we have advanced research supporting ubiquitous smartphone-based activity recognition solutions for forestry equipment. This work can be improved by further refining field sampling and modelling procedures and integrating new mobile technologies enabling the collection of sensor and accurate positional data from a single device. Additional field sampling with varying topographic and stand conditions, operators, and equipment types should also be completed to help quantify the extent to which these factors may impact activity recognition performance given the limited scope of conditions encountered in this study. Activity recognition of forest equipment using consumer available devices is an important advancement in the field of precision forestry and forest operations management. By identifying effective sampling and modelling processes, future work can focus on refining and optimizing data processing and model development and continue advancing toward real-time data support in smartphone applications. The current era of forest digitalization and smart and precision operational forestry necessitates land managers, operators, and researchers are equipped with the latest resources and technologies to successfully address complex forest management challenges. Accurate and efficient assessment of equipment productivity and work practices through activity recognition is valuable for advancing safer, more cost effective, and sustainable operational forestry supply chains.
Acknowledgments
The authors would like to thank University of Idaho Experimental Forest undergraduate forestry staff who planned and administered these stand treatments and Triple B Excavation for allowing their equipment and operators to participate in the research.
References
- 1.
Miyata ES. Determining fixed and operating costs of logging equipment. St. Paul, Minnesota: USDA-Forest Service, North Central Forest Experiment Station; 1980 p. 1–16. Report No.: NC-55.
- 2. Kellogg LD, Milota GV, Miller Jr. M. A comparison of skyline harvesting costs for alternative commercial thinning prescriptions. J For Eng. 1996;7–23.
- 3.
Klepac J, Rummer B. Evaluation of a shovel logging system in the gulf coastal plain. In: A Global Perspective. COFE 2002: Proceedings of the Council on Forest Engineering; 2002 Jun 16–20; Aubrun, AL, USA; 2002. p. 1–5.
- 4. Wang J, McNeel J, Baumgras J. A computer-based time study system for timber harvesting operations. For Prod J. 2003;53(3):47–53.
- 5. Wang J, Long C, McNeel J, John Baumgras. Productivity and cost of manual felling and cable skidding in central Appalachian hardwood forests. For Prod J. 2004;54(12):45–51.
- 6. Acuna MA, Kellogg LD. Evaluation of alternative cut-to-length harvesting technology for native forest thinning in Australia. Int J For Eng. 2009;20(2):17–25.
- 7. Acuna M, Skinnell J, Evanson T, Mitchell R. Bunching with a self-levelling feller-buncher on steep terrain for efficient yarder extraction. Croat J For Eng. 2011;32(2):521–31.
- 8. Alam M, Acuna M, Brown M. Self-levelling feller-buncher productivity based on lidar-derived slope. Croat J For Eng. 2013;34(2):273–81.
- 9. Strandgard M, Alam M, Mitchell R. Impact of slope on productivity of a self-levelling processor. Croat J For Eng. 2014;35(2):193–200.
- 10. Talbot B, Aalmo GO, Stampfer K. productivity analysis of an un-guyed integrated yarder-processor with running skyline. Croat J For Eng. 2014;35(2):201–10.
- 11. Borz SA, Talagai N, Cheta M, Gavilanes Montoya AV, Castillo Vizuete DD. Automating data collection in motor-manual time and motion studies implemented in a willow short rotation coppice. Bioresources. 2018;13(2):3236–49.
- 12. Strandgard M, Walsh D, Mauricio Acuna. Estimating harvester productivity in pinus radiata plantations using stanford stem files. Scand J For Res. 2013;28(1):73–80.
- 13. Bell CK, Keefe RF, Fried JS. Validation of the OpCost logging cost model using contractor surveys. Int J For Eng. 2017;28(2):73–84.
- 14. Strandgard M, Mitchell R. Automated time study of forwarders using GPS and a vibration sensor. Croat J For Eng. 2015;36(2):175–84.
- 15. Keefe RF, M.Wempe A, Becker RM, Zimbelman EG, Nagler ES, Gilbert SL, et al. Positioning methods and the use of location and activity data in forests. Forests. 2019;10(458):1–46.
- 16.
Taylor SE, McDonald TP, Veal MW, Grift TE. Using GPS to evaluate productivity and performance of forest machine systems. In: Proceedings of the First International Precision Forestry Symposium; 2001 June 17–20; Seattle, WA, USA; 2001. p. 1–20.
- 17.
Keefe RF, Eitel JUH, Smith AMS, Wade T. Tinkham. Applications of multi transmitter GPS-VHF in forest operations. In: Proceedings of the 47th International Symposium on Forestry Mechanization and the 5th International Forest Engineering Conference; 2014 Jan; Gerardmer, France; 2014.
- 18.
McDonald TP, Taylor SE, Rummer RB. Deriving forest harvesting machine productivity from positional data. In: 2000 ASAE Annual International Meeting; 2000 Jul 9–12; Milwaukee, Wisconsin, USA; 2000. p. 1–8.
- 19. Veal MW, Taylor SE, McDonald TP, McLemore DK, Dunn MR. Accuracy of tracking forest machines with GPS. Trans Am Soc Agric Eng. 2001;44:1903–11.
- 20. McDonald TP, Fulton JP. Automated time study of skidders using global positioning system data. Comput Electron Agric. 2005;48:19–37.
- 21.
de Hoop CF, Dupré RH. Using GPS to document skidder motions- a comparison with manual data collection. In: Working Globally–Sharing Forest Engineering Challenges and Technologies Around the World. Coeur d’Alene, ID, USA; 2006. p. 1–11.
- 22. Hejazian M, Hosseini S, Lotfalian M, Ahmadikoolaei P. Possibility of global positioning system (GPS) application for time studies in forest machinery. Eur J Exp Biol. 2013;3(4):93–8.
- 23. Gallo R, Grigolato S, Cavalli R, Mazzetto F. GNSS-based operational monitoring devices for forest logging operation chains. J Agric Eng. 2013;44(2):140–5.
- 24. Palander T, Nuutinen Y, Kariniemi A, Vaatainen K. Automatic time study method for recording work phase times of timber harvesting. For Sci. 2013;59(4):472–83.
- 25. Olivera A, Visser R. Using the harvester on-board computer capability to move towards precision forestry. N Z J For. 2016;60(4):3–7.
- 26. Olivera A, Visser R, Acuna M, Morgenroth J. Automatic GNSS-enabled harvester data collection as a tool to evaluate factors affecting harvester productivity in a eucalyptus spp. harvesting operation in Uruguay. Int J For Eng. 2016;27(1):15–28.
- 27.
Carter EA, McDonald TP, Torbert JL. Application of GPS technology to monitor traffic intensity and soil impacts in a forest harvest operation. In: Tenth Biennial Southern Silvicultural Research Conference; 1999 Feb 16–18; Shreveport, LA, USA; 1999. p. 609–13.
- 28. Grayson LM, Keefe RF, Tinkham WT, Eitel JUH, Saralecos JD, Smith AMS, et al. Accuracy of WAAS-enabled GPS-RF warning signals when crossing a terrestrial geofence. Sensors. 2016;16(912):1–8.
- 29. Wempe AM, Keefe RF. Characterizing rigging crew proximity to hazards on cable logging operations using GNSS-RF: effect of GNSS positioning error on worker safety status. Forests. 2017;8(357):1–17. pmid:30976381
- 30. Zimbelman EG, Keefe RF, Strand EK, Kolden CA, Wempe AM. Hazards in motion: development of mobile geofences for use in logging safety. Sensors. 2017;17(882):1–16. pmid:28394303
- 31. Newman SM, Keefe RF, Brooks RH, Ahonen EQ, Wempe AM. Human factors affecting logging injury incidents in Idaho and the potential for real time location-sharing technology to improve safety. Safety. 2018;4(4):1–28. pmid:30515383
- 32. Becker RM, Keefe RF, Anderson NM. Use of real-time GNSS-RF data to characterize the swing movements of forestry equipment. Forests. 2017;8(44):1–15.
- 33. Becker RM, Keefe RF. Development of activity recognition models for mechanical fuel treatments using consumer-grade GNSS-RF devices and lidar. Forestry. 2022;1–13.
- 34. Zimbelman EG, Keefe RF. Real-time positioning in logging: effects of forest stand characteristics, topography, and line-of-sight obstructions on GNSS-RF transponder accuracy and radio signal propagation. PLoS ONE. 2018;13(1):1–17.
- 35. Noordermeer L, Sørngård E, Astrup R, Næsset E, Gobakken T. Coupling a differential global navigation satellite system to a cut-to-length harvester operating system enables precise positioning of harvested trees. Int J For Eng. 2021;1–9.
- 36. Shoaib M, Bosch S, Incel OD, Scholten H, Havinga PJM. Fusion of smartphone motion sensors for physical activity recognition. Sensors. 2014;14:10146–76. pmid:24919015
- 37. Lima WS, Souto E, El-Khatib K, Jalali R, Gama J. Human activity recognition using inertial sensors in a smartphone: an overview. Sensors. 2019;19(3213):1–29.
- 38. del Rosario MB, Redmond SJ, Lovell NH. Tracking the evolution of smartphone sensing for monitoring human movement. Sensors. 2015;15:18901–33. pmid:26263998
- 39. Hur T, Bang J, Kim D, Banos O, Lee S. Smartphone location-independent physical activity recognition based on transportation natural vibration analysis. Sensors. 2017;17(931):1–21. pmid:28441743
- 40. Keefe RF, Zimbelman EG, Wempe AM. Use of smartphone sensors to quantify the productive cycle elements of hand fallers on industrial cable logging operations. Int J For Eng. 2019;1–12.
- 41. Bayat A, Pomplun M, Tran DA. A study on human activity recognition using accelerometer data from smartphones. Procedia Comput Sci. 2014;34:450–7.
- 42. Reddy S, Mun M, Burke J, Estrin D, Hansen M. Using mobile phones to determine transportation modes. ACM Trans Sens Netw. 2010;6(2):1–27.
- 43.
Casale P, Pujol O, Radeva P. Human activity recognition from accelerometer data using a wearable device. In: IbPRIA 2011; 2011. p. 289–96.
- 44. Chen L, Hoey J, Nugent CD, Cook DJ, Yu Z. Sensor-based activity recognition. IEEE Trans Syst Man Cybern. 2012;42(6):790–808.
- 45. Khan AM, Siddiqi MH, Lee S-W. Exploratory data analysis of acceleration signals to select light-weight and accurate features for real-time activity recognition on smartphones. Sensors. 2013;13:13099–122. pmid:24084108
- 46. Khusainov R, Azzi D, Achumba IE, Bersch SD. Real-time human ambulation, activity, and physiological monitoring: taxonomy of issues, techniques, applications, challenges and limitations. Sensors. 2013;13:12852–902. pmid:24072027
- 47. Attal F, Mohammed S, Dedabrishvili M, Chamroukhi F, Oukhellou L, Amirat Y. Physical human activity recognition using wearable sensors. Sensors. 2015;15:31314–38. pmid:26690450
- 48. Shoaib M, Bosch S, Incel OD, Scholten H, Havinga PJM. A survey of online activity recognition using mobile phones. Sensors. 2015;15:2059–85. pmid:25608213
- 49. Gjoreski M, Gjoreski H, Luštrek M, Gams M. How accurately can your wrist device recognize daily activities and detect falls? Sensors. 2016;16(800):1–21. pmid:27258282
- 50. Zhang M, Chen S, Zhao X, Yang Z. Research on construction workers’ activity recognition based on smartphone. Sensors. 2018;18(2667):1–18. pmid:30110892
- 51. Akhavian R, Behzadan AH. Smartphone-based construction workers’ activity recognition and classification. Autom Constr. 2016;71:198–209.
- 52. Gallo R, Visser R, Mazzetto F. Developing an automated monitoring system for cable yarding systems. Croat J For Eng. 2021;42(2):213–25.
- 53. Borz SA, Paun M. Integrating offline object tracking, signal processing, and artificial intelligence to classify relevant events in sawmilling operations. Forests. 2020;11(1333):1–19.
- 54. Cheţa M, Marcu MV, Iordache E, Borz SA. Testing the capability of low-cost tools and artificial intelligence techniques to automatically detect operations done by a small-sized manually driven bandsaw. Forests. 2020;11(739):1–13.
- 55. Cheta M, Marcu MV, Borz SA. Effect of training parameters on the ability of artificial neural networks to learn: a simulation on accelerometer data for task recognition in motor-manual felling and processing. Bull Transilv Univ Braşov. 2020;13(62):1–18.
- 56. Pierzchała M, Kvaal K, Stampfer K, Talbot B. automatic recognition of work phases in cable yarding supported by sensor fusion. Int J For Eng. 2018;29(1):12–20.
- 57. Zimbelman EG, Keefe RF. Development and validation of smartwatch-based activity recognition models for rigging crew workers on cable logging operations. PLoS ONE. 2021;16(5):1–25. pmid:33979355
- 58. Fahim M, Fatima I, Lee S, Park Y-T. EFM: Evolutionary fuzzy model for dynamic activities recognition using a smartphone accelerometer. Appl Intell. 2013;39:475–88.
- 59. Incel OD, Kose M, Ersoy C. A review and taxonomy of activity recognition on mobile phones. BioNano Sci. 2013;3:145–71.
- 60. San-Segundo R, Blunck H, Moreno-Pimentel J, Stisen A, Gil-Martín M. Robust human activity recognition using smartwatches and smartphones. Eng Appl Artif Intell. 2018;72:190–202.
- 61. Hiesel P, Benjamin JG. Applicability of international harvesting equipment productivity studies in Maine, USA: a literature review. Forests. 2013;4:898–921.
- 62. Reutebuch SE, Andersen H-E, McGaughey RJ. Light detection and ranging (lidar): an emerging tool for multiple resource inventory. J For. 2005;September:286–92.
- 63. Falkowski MJ, Gessler PE, Morgan P, Hudak AT, Smith AMS. Characterizing and mapping forest fire fuels using ASTER imagery and gradient modeling. For Ecol Manag. 2005;217:129–46.
- 64. Falkowski MJ, Evans JS, Martinuzzi S, Gessler PE, Hudak AT. Characterizing forest succession with lidar data: an evaluation for the inland northwest, USA. Remote Sens Environ. 2009;113:946–56.
- 65. Akay AE, Oğuz H, Karas IR, Aruga K. Using lidar technology in forestry activities. Environ Monit Assess. 2009;151:117–25. pmid:18365761
- 66. Hudak AT, Strand EK, Vierling LA, Byrne JC, Eitel JUH, Martinuzzi S, et al. Quantifying aboveground forest carbon pools and fluxes from repeat lidar surveys. Remote Sens Environ. 2012;123:25–40.
- 67. Wulder MA, White JC, Nelson RF, Næsset E, Ørka HO, Coops NC, et al. Lidar sampling for large-area forest characterization: a review. Remote Sens Environ. 2012;121:196–209.
- 68. Man Q, Dong P, Guo H, Liu G, Shi R. Light detection and ranging and hyperspectral data for estimation of forest biomass: a review. J Appl Remote Sens. 2014;8:1–22.
- 69. Silva CA, Klauberg C, Hudak AT, Vierling LA, Jaafar WSM, Mohan M, et al. Predicting stem total and assortment volumes in an industrial pinus taeda l. forest plantation using airborne laser scanning data and random forest. Forests. 2017;8(254):1–17.
- 70. Silva CA, Klauberg C, Hudak AT, Vierling LA, Fennema SJ, Corte APD. modeling and mapping basal area of pinus taeda l. plantation using airborne lidar data. Ann Braz Acad Sci. 2017;89(3):1895–905. pmid:28813098
- 71. Falkowski MJ, Smith AMS, Gessler PE, Hudak AT, Vierling LA, Evans JS. the influence of conifer forest canopy cover on the accuracy of two individual tree measurement algorithms using lidar data. Can J Remote Sens. 2008;34:338–50.
- 72. d’Oliveira MVN, Reutebuch SE, McGaughey R j., Andersen H-E. Estimating forest biomass and identifying low-intensity logging areas using airborne scanning lidar in Antimary State Forest, Acre State, Western Brazilian Amazon. Remote Sens Environ. 2012;124:479–91.
- 73. Ellis P, Griscom B, Walker W, Gonçalves F, Tina Cormier. Mapping selective logging impacts in Borneo with GPS and airborne lidar. For Ecol Manag. 2016;365:184–96.
- 74. Slesak RA, Kaebisch T. Using lidar to assess impacts of forest harvest landings on vegetation height by harvest season and the potential for recovery over time. Can J For Res. 2016;46:869–75.
- 75. Heinimann HR, Breschan J. Pre-harvest assessment based on lidar data. Croat J For Eng. 2012;33(2):169–80.
- 76.
Kato A, Schiess P. Lidar derived tree parameters for optimal cable logging system design. In: Proceedings of the International Mountain Logging and 13th Pacific Northwest Skyline Symposium; 2007 Apr; Corvallis, OR, USA; 2007. p. 173–9.
- 77. Becker RM, Keefe RF, Anderson NM, Eitel JUH. Use of lidar-derived landscape parameters to characterize alternative harvest system options in the inland northwest. Int J For Eng. 2018;29(3):179–91.
- 78. Maltamo M, Hauglin M, Næsset E, Gobakken T. Estimating stand level stem diameter distribution utilizing harvester data and airborne laser scanning. Silva Fenn. 2019;53(3):1–19.
- 79. Yu X, Hyyppä J, Holopainen M, Vastaranta M. Comparison of area-based and individual tree-based methods for predicting plot-level forest attributes. Remote Sens. 2010;2:1481–95.
- 80. Yao W, Krzystek P, Heurich M. Tree species classification and estimation of stem volume and dbh based on single tree extraction by exploiting airborne full-waveform lidar data. Remote Sens Environ. 2012;123:368–80.
- 81. Hyyppä J, Yu X, Hyyppä H, Vastaranta M, Holopainen M, Kukko A, et al. Advances in forest inventory using airborne laser scanning. Remote Sens. 2012;4:1190–207.
- 82. Melville G, Stone C, Turner R. Application of lidar data to maximise the efficiency of inventory plots in softwood plantations. N Z J For Sci. 2015;45(9):1–16.
- 83. Jeronimo SMA, Kane VR, Churchill DJ, McGaughey RJ, Franklin JF. Applying lidar individual tree detection to management of structurally diverse forest landscapes. J For. 2018;116(4):336–46.
- 84.
Rahman MZA, Gorte BGH, Bucksch AK. A new method for individual tree measurement from airborne lidar. In: Proceedings of Silvilaser 2009; 2009 Oct 14–16; College Station, Texas, USA; 2009. p. 1–10.
- 85. Jakubowski MK, Li W, Guo Q, Kelly M. Delineating individual trees from lidar data: a comparison of vector- and raster-based segmentation approaches. Remote Sens. 2013;5:4163–86.
- 86. Kandare K, Ørka HO, Chan JC-W, Dalponte M. Effects of forest structure and airborne laser scanning point cloud density on 3d delineation of individual tree crowns. Eur J Remote Sens. 2016;49:337–59.
- 87. Luo L, Zhai Q, Su Y, Ma Q, Kelly M, Guo Q. Simple method for direct crown base height estimation of individual conifer trees using airborne lidar data. Opt Express. 2018;26(10):1–17. pmid:29801269
- 88. Silva CA, Hudak AT, Vierling LA, Loudermilk EL, O’Brien JJ, Hiers JK, et al. Imputation of individual longleaf pine (pinus palustris mill.) tree attributes from field and lidar data. Can J Remote Sens. 2016;42(5):554–73.
- 89.
Taylor S, Veal M, Grift T, McDonald T, Corley F. Precision forestry: operational tactics for today and tomorrow. In: Proceedings of the 25th Annual Meeting of the Council of Forest Engineers; 2002. Available from: www.eng.auburn.edu/files/file169.pdf
- 90. Kovacsova P, Antalova M. Precision forestry—definition and technologies. Sumar List. 2010;134(11):603–11.
- 91.
Asim F. AndroSensor [Internet]. 2015 [cited 2018 Sep 15]. Available from: https://play.google.com/store/apps/details?id=com.fivasim.androsensor&hl=en&gl=US
- 92. Zheng L, Wu D, Ruan X, Weng S, Peng A, Tang B, et al. A novel energy-efficient approach for human activity recognition. Sensors. 2017;17(2064):1–21. pmid:28885560
- 93. Ermes M, P¨arkk¨a J, M¨antyj¨arvi J, Ilkka Korhonen. Detection of daily activities and sports with wearable sensors in controlled and uncontrolled conditions. IEEE Trans Inf Technol Biomed. 2008;12(1):20–7. pmid:18270033
- 94. Koping L, Shirahama K, Grzegorzek M. A general framework for sensor-based human activity recognition. Comput Biol Med. 2018;95:248–60. pmid:29361267
- 95. Hassan MM, Uddin MdZ, Mohamed A, Almogren A. A robust human activity recognition system using smartphone sensors and deep learning. Future Gener Comput Syst. 2018;81:307–13.
- 96.
Krause A, Ihmig M, Rankin E, Leong D, Gupta S, Siewiorek D, et al. Trading off prediction accuracy and power consumption for context-aware wearable computing. In: Proceedings of the 9th IEEE International Symposium on Wearable Computers; 2005. p. 1–7.
- 97.
Maurer U, Smailagic A, Siewiorek DP, Deisher M. Activity recognition and monitoring using multiple sensors on different body positions. In: Proceedings of the International Workshop on Wearable and Implantable Body Sensor Networks; 2006. p. 1–4.
- 98.
Siirtola P, Laurinen P, Roning J. Efficient accelerometer-based swimming exercise tracking. In: Proceedings of the 2011 IEEE Symposium on the Computational Intelligence and Data Mining; 2011 Apr 11–15; Paris, France; 2011. p. 1–6.
- 99.
Siirtola P, Roning J. Ready-to-use activity recognition for smartphones. In: Proceedings of the 2013 IEEE Symposium on Computational Intelligence and Data Mining; 2013 Apr 16–19; Singapore; 2013. p. 1–6.
- 100.
R Core Team. R: A Language and Environment for Statistical Computing [Internet]. Vienna, Austria: R Foundation for Statistical Computing; R Core Team; 2020 [cited 2020 Feb 12]. Available from: https://www.R-project.org
- 101. Bianchi F, Redmond SJ, Narayanan MR, Cerutti S, Lovell NH. Barometric pressure and triaxial accelerometry- based falls event detection. IEEE Trans Neural Syst Rehabil Eng. 2010;18(6):619–27. pmid:20805056
- 102. Ellis K, Kerr J, Godbole S, Lanckriet G, Wing D, Marshall S. A random forest classifier for the prediction of energy expenditure and type of physical activity from wrist and hip accelerometers. Physiol Meas. 2014;35(11):2191–203. pmid:25340969
- 103.
Lara OD, Labrador MA. A mobile platform for real-time human activity recognition. In: Proceedings of the 6th IEEE International Workshop on Personalized Networks; 2012. p. 667–71.
- 104. Morales J, Akopian D. Physical activity recognition by smartphones, a survey. Biocybern Biomed Eng. 2017;37:388–400.
- 105. Mehrang S, Pietilä J, Korhonen I. An activity recognition framework deploying the random forest classifier and a single optical heart rate monitoring and triaxial accelerometer wrist-band. Sensors. 2018;18(613):1–13.
- 106.
Sun L, Zhang D, Li B, Guo B, Li S. Activity recognition on an accelerometer embedded mobile phone with varying positions and orientations. In: Proceedings of the International Conference on Ubiquitous Intelligence and Computing; 2010. p. 548–62.
- 107. Trost SG, Zheng Y, Wong W-K. Machine learning for activity recognition: hip versus wrist data. Physiol Meas. 2014;35:2183–9. pmid:25340887
- 108.
Weiss GM, Timko JL, Gallagher CM, Yoneda K, Schreiber AJ. Smartwatch-based activity recognition: a machine learning approach. In: Proceedings of the 2016 IEEE-EMBS International Conference on Biomedical and Health Informatics; 2016 Feb 24–27; Las Vegas, Nevada; 2016. p. 426–9.
- 109.
Achumba IE, Bersch S, Khusainov R, Azzi D, Kamalu U. On time series sensor data segmentation for fall and activity classification. In: Proceedings of the IEEE 14th International Conference on e-Health Networking, Applications and Services; 2012 Oct 10–13; Beijing, China; 2012. p. 427–30.
- 110. Wu W, Dasgupta S, Ramirez EE, Peterson C, Norman GJ. Classification accuracies of physical activities using smartphone motion sensors. J Med Internet Res. 2012;14(5):1–9. pmid:23041431
- 111. Shoaib M, Bosch S, Incel OD, Scholten H, Havinga PJM. Complex human activity recognition using smartphone and wrist-worn motion sensors. Sensors. 2016;16(426):1–24. pmid:27023543
- 112. Lara OD, Labrador MA. A survey on human activity recognition using wearable sensors. IEEE Commun Surv Tutor. 2013;15(3):1192–209.
- 113. Figo D, Diniz PC, Ferreira DR, Cardoso JMP. Preprocessing techniques for context recognition from accelerometer data. Pers Ubiquit Comput. 2010;14:645–62.
- 114. Breiman L. Random Forests. Mach Learn. 2001;45:5–32.
- 115. Liaw A, Wiener M. Classification and regression by randomForest. R News. 2002;2(3):18–22.
- 116. Strobl C, Boulesteix A-L, Kneib T, Augustin T, Zeileis A. Conditional variable importance for random forests. BMC Bioinformatics. 208AD;9(307):1–11.
- 117. Genuer R, Poggi J-M, Tuleau-Malot C. Variable selection using random forests. Pattern Recognit Lett. 2010;31(14):2225–36.
- 118. Belgiu M, Dragut L. Random forest in remote sensing: a review of applications and future directions. ISPRS J Photogramm Remote Sens. 2016;114:24–31.
- 119.
Leo Breiman. Manual on setting up, using, and understanding random forests v3.1 [Internet]. Statistics Department University of California Berkeley, CA, USA; 2002 [cited 2020 Feb 13]. Available from: https://www.stat.berkeley.edu/~breiman/RandomForests/cc_manual.htm
- 120. Brosofske KD, Froese RE, Falkowski MJ, Banskota A. A review of methods for mapping and prediction of inventory attributes for operational forest management. For Sci. 2014;60(4):733–56.
- 121. Cinaroglu S. Comparison of performance of decision tree algorithms and random forest: an application on OECD countries health expenditures. Int J Comput Appl. 2016;138(1):37–41.
- 122. Kuhn M, Wing J, Weston S, Williams A, Keefer C, Engelhardt A, et al. caret: classification and regression training [Internet]. 2021 [cited 2020 Feb 13]. Available from: https://cran.r-project.org/web/packages/caret/index.html
- 123. Svetnik V, Liaw A, Tong C, Culberson JC, Sheridan RP, Feuston BP. Random forest: a classification and regression tool for compound classification and QSAR modeling. J Chem Inf Comput Sci. 2003;43(6):1947–58. pmid:14632445
- 124. Robin X, Turck N, Hainard A, Tiberti N, Lisacek F, Sanchez J-C, et al. pROC: an open-source package for R and S+ to analyze and compare ROC curves. BMC Bioinformatics [Internet]. 2011;12(77). Available from: http://expasy.org/tools/pROC/ pmid:21414208
- 125. Streiner DL, Cairney J. What’s under the roc? an introduction to receiver operating characteristics curves. Can J Psychiatry. 2007;52(2):121–8. pmid:17375868
- 126. Sokolova M, Japkowicz N, Szpakowicz S. Beyond accuracy, F-score and ROC: a family of discriminant measures for performance evaluation. Lect Notes Comput Sci. 2006;
- 127. Hajian-Tilaki K. Receiver operating characteristic (ROC) curve analysis for medical diagnostic test evaluation. Casp Jounral Int Med. 2013;4(2):627–35. pmid:24009950
- 128. Hand DJ, Till RJ. A simple generalisation of the area under the ROC curve for multiple class classification problems. Mach Learn. 2001;45:171–86.
- 129. Luque A, Carrasco A, Martín A, de las Heras A. The impact of class imbalance in classification performance metrics based on the binary confusion matrix. Pattern Recognit. 2019;91:216–31.
- 130. Boughorbel S, Jarray F, El-Anbari M. Optimal classifier for imbalanced data using matthews correlation coefficient metric. PLoS ONE. 2017;12(6):1–17. pmid:28574989
- 131. Ling C, Li C. Data mining for direct marketing: problems and solutions. KDD. 1998;1–7.
- 132. Hanley JA, Barbara J. McNeil. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology. 1982;143(1):29–36. pmid:7063747
- 133. Tharwat A. Classification assessment methods. Appl Comput Inform. 2021;17(1):168–92.
- 134.
Minnen D, Westeyn T, Starner T, Ward JA, Lukowicz P. Performance metrics and evaluation issues for continuous activity recognition. In: Proceedings of the Performance Metrics for Intelligent Systems Workshop; 2006 Aug 21–23; Gaithersburg, MD, USA: NIST; 2006. p. 141–8.
- 135. Chicco D, Jurman G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genomics. 2020;21(6):1–13. pmid:31898477
- 136. Banos O, Galvez J-M, Damas M, Pomares H, Rojas I. Window size impact in human activity recognition. Sensors. 2014;14:6474–99. pmid:24721766
- 137.
Sarada B, Dandu M, Tarun SK. A comparison of tree based classification and neural network based classification. In: Proceedings of the 21st International Arab Conference on Information Technology; 2020 Nov 28–30; Al Ain University of Science and Technology, Al Ain, UAE: IEEE; 2020. p. 1–3.
- 138. Bonomi AG, Goris AHC, Yin B, Westerterp KR. Detection of type, duration, and intensity of physical activity using an accelerometer. Med Sci Sports Exerc. 2009;1770–7. pmid:19657292
- 139. Sparks AM, Smith AMS. Accuracy of a lidar-based individual tree detection and attribute measurement algorithm developed to inform forest products supply chain and resource management. Forests 2022;13(3):1–15.
- 140. Pebesma E, Bivand R, Racine E, Sumner M, Cook I, Keitt T, et al. sf: simple features for R [Internet]. 2021. Available from: https://CRAN.R-project.org/package=sf
- 141.
Prusa J, Khoshgoftaar TM, Dittman DJ, Napolitano A. Using random undersampling to alleviate class imbalance on tweet sentiment data. In: Proceedings of the IEEE 16th International Conference on Information Reuse and Integration; 2015 Aug 13–15; San Francisco, CA, USA: IEEE; 2015.
- 142. Goyal A, Rathore L, Sharma A. SMO-RF: a machine learning approach by random forest for predicting class imbalancing followed by SMOTE. Mater Today Proc. 2021;
- 143. Elrahman SMA, Abraham A. A review of class imbalance problem. J Netw Innov Comput. 2013;1:332–40.
- 144. Westra BL, Dey S, Fang G, Steinbach M, Kumar V, Oancea C, et al. Interpretable predictive models for knowledge discovery from home-care electronic health records. J Healthc Eng. 2011;2(1):55–74.
- 145.
Kamei Y, Monden A, Matsumoto S, Kakimoto T, Matsumoto K. The effects of over and under sampling on fault-prone module detection. In: Proceedings of the First International Symposium on Empirical Software Engineering and Measurement; 2007 Sept; IEEE; 2007. p. 196–204.
- 146.
Dittman DJ, Khoshgoftaar TM, Napolitano A. The effect of data sampling when using random forest on imbalanced bioinformatics data. In: Proceedings of the IEEE 16th International Conference on Information Reuse and Integration; 2015 Aug 13–15; San Francisco, CA, USA: IEEE; 2015. p. 457–63.
- 147. Li W, Guo Q, Jakubowski MK, Kelly M. A new method for segmenting individual trees from the lidar point cloud. Photogramm Eng Remote Sens. 2012;78(1):75–84.
- 148. Balsi M, Esposito S, Fallavollita P, Nardinocchi C. Single-tree detection in high-density lidar data from UAV-based survey. Eur J Remote Sens. 2018;51(1):679–92.
- 149. Eysn L, Hollaus M, Lindberg E, Berger F, Monnet J-M, Dalponte M, et al. A benchmark of lidar-based single tree detection methods using heterogeneous forest data from the alpine space. Forests. 2015;6:1721–47.
- 150. Dong T, Zhou Q, Gao S, Shen Y. Automatic detection of single trees in airborne laser scanning data through gradient orientation clustering. Forests. 2018;9(219):1–19.
- 151. Aijazi AK, Checchin P, Malaterre L, Laurent Trassoudaine. Automatic detection and parameter estimation of trees for forest inventory applications using 3D terrestrial lidar. Remote Sens. 2017;9(946):1–24.
- 152. White JC, Coops NC, Wulder MA, Vastaranta M, Hilker T, Tompalski P. Remote sensing technologies for enhancing forest inventories: a review. Can J Remote Sens. 2016;42(5):619–41.
- 153. Paziewski J. Recent advances and perspectives for positioning and applications with smartphone GNSS observations. Meas Sci Technol. 2020;31:1–13.
- 154. Paziewski J, Fortunato M, Mazzoni A, Odolinski R. An analysis of multi-GNSS observations tracked by recent android smartphones and smartphone-only relative positioning results. Meas Sci Technol. 2021;175(109162):1–16.
- 155. Tomastík J, Chudá J, Tunák D, Chudý F, Kardos M. Advances in smartphone positioning in forests: dual-frequency receivers and raw GNSS data. Forestry. 2021;94:292–310.
- 156.
Darugna F, Wübbena J, Ito A, Wübbena T, Wübbena G, Schmitz M. RTK and PPP-RTK using smartphones: from short-baseline to long-baseline applications. In: Proceedings of the 32nd International Technical Meeting of the Satellite Division of the Institute of Navigation; Miami, FL, USA; 2019. p. 3932–45.
- 157. Robustelli U, Paziewski J, Pugliano G. Observation quality assessment and performance of gnss standalone positioning with code pseudoranges of dual-frequency android smartphones. Sensors. 2021;21(2125):1–19. pmid:33803768
- 158. Dabove P, DiPietra V, Piras M. GNSS positioning using mobile devices with the android operating system. Int J Geo-Inf. 2020;9(220):1–13.
- 159. Dabove P, DiPietra V. Towards high accuracy GNSS real-time positioning with smartphones. Adv Space Res. 2019;63:94–102.