Figures
Abstract
The electrical penetration graph (EPG) is a well-known technique that provides insights into the feeding behavior of insects with piercing-sucking mouthparts, mostly hemipterans. Since its inception in the 1960s, EPG has become indispensable in studying plant-insect interactions, revealing critical information about host plant selection, plant resistance, virus transmission, and responses to environmental factors. By integrating the plant and insect into an electrical circuit, EPG allows researchers to identify specific feeding behaviors based on their distinctive waveform patterns. However, the traditional manual analysis of EPG waveform data is time-consuming and labor-intensive, limiting research throughput. This study presents a novel Machine Learning (ML) approach to automate the annotation of EPG signals. We rigorously evaluated six diverse ML models, including neural networks, tree-based models, and logistic regression, using an extensive dataset from multiple aphid feeding experiments. Our results demonstrate that a Residual Network (ResNet) architecture achieved the highest overall waveform classification accuracy of 96.8% and highest segmentation overlap rate of 84.4%, highlighting the potential of ML for accurate and efficient EPG analysis. This automated approach promises to accelerate research in this field significantly and broaden insights into insect-plant interactions, showcasing the power of computational techniques for insect biological research. The source code for all experiments conducted within this study is publicly available at https://github.com/HySonLab/ML4Insects.
Citation: Dinh QD, Kunk D, Son Hy T, Nalam V, Dao PD (2025) Machine learning for automated electrical penetration graph analysis of aphid feeding behavior: Accelerating research on insect-plant interactions. PLoS ONE 20(4): e0319484. https://doi.org/10.1371/journal.pone.0319484
Editor: Sean Michael Prager
Received: July 10, 2024; Accepted: February 2, 2025; Published: April 3, 2025
Copyright: © 2025 Dinh et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All code written in support of this publication is publicly available at https://github.com/HySonLab/ML4Insects. All data used in the publication is from previously published manuscripts and the relevant references have been provided in the text.
Funding: The work was supported by start-up funds provided to Drs. Nalam and Dao from Colorado State University.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Aphids (order Hemiptera) are major agricultural pests that feed on plant phloem sap using their needle-like mouthparts. Their rapid asexual reproduction and high population densities can lead to significant crop damage. Copious feeding by aphids depletes essential nutrients from plants and introduces potentially toxic saliva and plant viruses, further compromising crop quality and yield. In severe cases, aphid infestations can result in hundreds of millions of dollars in economic losses, underscoring the importance of understanding and managing these pests [1]. The electrical penetration graph (EPG) technique, introduced in 1964 by Donald McLean and Marvin Kinsey, provided researchers with a tool to better understand aphid feeding behavior. The initial EPG technique involved creating an electrical circuit by wiring the aphid and the plant, then recording the resulting voltage changes once the aphid closes the circuit by penetrating the plant tissue [2]. Throughout the 20th century, entomologists enhanced this technique by analyzing the behavioral mechanisms behind EPG waveforms. This was achieved by performing stylectomy followed by dissection of the plant tissue to determine the approximate location of the stylet at the time the behavior was recorded [3]. This research further allowed for the understanding of the underlying electrical component of each behavioral waveform, and through the invention of the direct current (DC) EPG system, it allowed researchers to record behaviors originating from both plant-insect resistance (Ri) and those originating from the electromotive force (emf) [4,5]. Today, the EPG technique stands as a leading method for studying hemipteran feeding behavior [6,7] .
Currently, commercially available DC-EPG systems provide the necessary hardware and software for acquiring EPG recordings [8]. To set up and create the circuit, aphids are restrained, and a gold wire is glued to their abdomen using conductive paint. The gold wire is typically glued to a copper wire that is soldered onto a brass nail, which can be inserted into a specially designed housing attached to the EPG circuit. For the plant, a brass electrode is inserted into the soil and connected to the EPG circuit. The analog signal is converted to a digital signal in GIGA systems and recorded on a computer using the Stylet+ data acquisition software (.DAQ file and .DX extension for hour-by-hour recordings). After recording feeding behaviors, the user manually annotates the waveforms by visual inspection, creating a file with columns for the start time, the specific behavior, and the voltage value at the start of the behavior. Various parameters associated with feeding behaviors are then manually calculated from these annotation files or using publicly available Excel workbooks [9].
However, the throughput of EPG analysis is greatly constrained because it demands a time-consuming manual review process and the expertise of an experienced observer to annotate the signals accurately. During each annotation session, the observer analyzes data on a second-by-second basis, which can take up to 30 minutes (for an 8 h recording), depending on the complexity of the observed patterns and the length of the recording. Additionally, recordings with ambiguous waveforms, excessive noise, or numerous individual events can significantly prolong the annotation process. Thus, an automated annotation pipeline will greatly streamline data analysis. Previous attempts at automation, such as Aphid Auto EPG [10], had limited classification ability due to its basic discrimination rules based on waveform amplitudes. Similarly, Adasme-Carreño et al. (2015) developed the Assisted Analysis of EPG (A2EPG) program to automate the annotation of EPG signals [11]. However, the criteria for distinguishing waveforms was overly simplistic, leading to the misclassification of waveforms.
Researchers have explored ML methods for EPG signals due to their versatility with diverse data types and reliable predictive performance. In 2018, Wu et al. [12] (2018) used Decision Tree (DT) for classifying samples of EPG waveform obtained from the green peach aphid (Myzus persicae) feeding on tobaco host plants (Zhongyan No. 1 variety), reporting a mean classification accuracy of 91.43% over 4 experiments. On the same dataset, Xing et al. (2023) [13] proposed using Extreme Kernel Machine (EKM) instead of DT improviing the prediction accuracy to 94.47%. A Random Forest model was introduced to characterize the behaviors of Asian citrus psyllids (Diaporina citri) on nine citrus genotypes [14], achieving an overall accuracy score of 97.4%, with the Hidden Markov Model further revealing previously unknown feeding patterns. However, these studies were limited in scale, necessitating further validation of larger and more diverse datasets to assess the broader applicability of these methods. In recent years, Convolutional Neural Networks (CNNs) have shown great promise in insect behavior monitoring [15–19], with their ability to automatically learn informative features from raw data eliminating the need for manual feature engineering. Furthermore, CNN variants have proven effective in segmenting biological time series [20–23], showcasing their potential in EPG analysis.
In this study, we aim to 1) evaluate the performance of well-established ML models for characterizing and annotating EPG signals, 2) evaluate different feature extraction methods on model performance, and 3) develop a ready-to-use pipeline for automatic waveform annotation for users who may not be familiar with ML tools. Our study builds on previous work by evaluating various ML models, including Fully Convolutional Network (1DCNN), 2D Convolutional Neural Network (2DCNN), Residual Network (ResNet), Extreme Gradient Boosting classifier (XGB), Random Forest (RF), and Logistic Regression (LR). A Python package was developed based on well-established libraries such as Scikit-learn and PyTorch, providing an end-to-end automatic annotation process for reproducibility and widespread use.
Materials and methods
EPG signal annotation and characterization
Aphid’s feeding behavior.
Waveforms within the EPG signals are classified into seven categories: non-penetrating (NP), pathway (C), potential drop (pd), phloem salivation (E1), phloem-feeding (E2), xylem feeding (G), and derailed stylet phase (F), depending on the pattern observed from both the time and frequency domain. During the NP phase, the insect has yet to insert its mouthpart into the host plant, so the waveform is usually almost a straight line. Pathway phase (C) is when the stylet is inserted into the plant tissue and is present in the epidermal and mesophyll tissues. During C, brief intracellular punctures by stylet tips occur, referred to as pd or potential drops. It is during C that aphids make decisions regarding host acceptance or rejection. The E1 and E2 phases are sometimes combined and called the sieve element phase (SEP). During SEP, the stylet tips are within phloem sieve elements. During E1, the aphid actively salivates while maintaining stylet position within the sieve elements. The E2 phase signifies active phloem sap ingestion, with the stylet remaining in the sieve element. In contrast, the G waveform indicates xylem sap ingestion, characterized by the stylet positioned in the xylem. Lastly, the F waveform represents derailed stylets, where the aphid encounters mechanical difficulties while probing host tissue. In this phase, despite being within host tissue, the aphid does not engage in feeding behaviors due to stylet malfunctions. It is important to note that these behaviors are sequential. For instance, the aphid must enter C before engaging in phloem or xylem feeding (E1, E2, or G), and the pd pattern is only observed during C. Examples of waveforms of these behaviors are shown in (Fig 1).
Workflow.
We propose an end-to-end pipeline following a sliding window approach to automate the EPG signal annotation process. We attempt to create a complete partition of an input recording by making predictions on the component chunks of the signal. Our pipeline consists of three main steps: initial segmentation, waveform sample classification, and label aggregation (Fig 2). Specifically, the input signal is initially divided into consecutive segments of fixed-length d with a step size s. By default, we set d = 1024 (corresponding to a 10.24s segment) and s = d to obtain a non-overlap segment. Then, the segments of signals are transformed using Fourier or Wavelet transformation into valuable features that characterize the signal’s behavior. The Data Processing section discusses this step in greater detail. In the second step, trained ML models will classify the segments using the transformed data as input by generating a probability distribution illustrating the likelihood that a segment belongs to a specific waveform. Each segment will be assigned a label with the highest probability. The following sections discuss the details of the models studied in this work. Lastly, the predicted labels will be concatenated to create a complete segmentation of the input recording. To do this, each time step is given a predicted label according to that of the segment containing it; consequently, each segment yields d labels for d time steps. A complete predicted waveform is thus composed by concatenating such sequences, from which one may determine the predicted locations of each waveform. For example, assume that an EPG recording is 8 hours long, recorded at a sampling frequency of 100Hz (i.e., 100 time steps per second). A waveform sample located between 0s and 10.23s, assigned with label NP, will generate a sequence of 1024 identical labels (NP, NP, ..., NP). Concatenating these sequences of predictions will thus result in 2,880,000 labels for the input recording.
Classification models.
Random Forest (RF): A well-known ML algorithm used for both classification and regression tasks. An RF consists of multiple Decision Trees (DTs) trained on subdatasets created by bootstrapping from random subsets of features. The time series adaptation of RF [24,25] involves segmenting the input time series into intervals and extracting various features from these segments. Experimental studies show that using RF with simple features such as mean, standard deviation, and slope is computationally efficient and outperforms strong competitors such as 1-NN classifiers with dynamic time-warping distance.
Extreme Gradient Boosting classifier (XGB): A very fast and accurate state-of-the-art tree-based model built under the Gradient Boosting framework. The algorithm operates by sequentially adding weak learners to the ensemble, with each new learner focusing on correcting the errors made by the existing ones while using a gradient descent optimization technique to minimize a predefined loss function during training. In [26], four state-of-the-art gradient boosting-based algorithms, including XGBoost, LGBM, GBM, and CatBoost, are benchmarked based on 12 datasets of different sizes and contexts. The results showed that all models achieved state-of-the-art performance, with LGBM ranking first. However, it is crucial to note that extensive fine-tuning was essential to attain this level of performance.
Logistic Regression (LogReg): A popular and widely used classification algorithm owing to its simplicity, interpretability, and effectiveness for binary classification tasks. Moreover, this method can be extended to multi-class classification problems through one-vs-all or softmax regression methods.
Convolutional neural networks (CNNs): A type of neural network algorithm that learns characteristics of the input data by optimizing the kernels. In CNNs, learning units are arranged in layers, and the heart of CNNs is the matrix convolution operator used at the convolutional layers. At each convolutional layer, the input convolves with the kernels in order to produce feature maps. The kernel allows the network to concentrate on specific parts of the input sequentially, thereby enabling CNNs to minimize the parameters per layer and promote a more extensive network architecture. This feature makes CNNs particularly effective at tasks where focus on local patterns is required, as in the case of image and time-series data.
Data
The ML models were benchmarked on datasets collected from four aphid species: Rhopalosiphum padi (bird cherry-oat aphid; BCOA1, BCOA2) Myzus persicae (green peach aphid; GPA), Phorodon cannabis (cannabis aphid; CA) and Aphis glycines (Soybean aphid; SA) that were obtained from the Stylet+ application [8]. The datasets include a set of signal data in ASCII format and a text file that provides corresponding ground-truth annotations. Each EPG recording consists of a single signal channel that spans 8 hours at a sampling rate of 100Hz. The components of each dataset are described in (Fig 3). Typically, the patterns from the pathwave phase (C) and sieve element phase (E1, E2) account for a major proportion of the total recording time. In two datasets, GPA and BCOA2, there were virtually no F and G waveforms, which partly explains the difficulty in distinguishing between these waveforms and the others. The potential drop (pd) waveform tends to appear frequently but over a short time, so they account for a negligible proportion of the data distribution.
BCOA1. This dataset is derived from an experiment on the bird cherry-oat aphid (R. padi) feeding on wheat (Triticum aestivum var. Chinese Spring). Aphid feeding was recorded for 8 h over a 24-hour period during the morning, afternoon, and night under two different light conditions: a photoperiod of 16 h light/8 h dark and continuous 24-hour light. The aim of the experiment is to understand how the time of day and light-dark cycles impact feeding behavior. Aphids were allowed to feed from the 2nd leaf of seedlings at Zadoks stage Z1.2 [27]. The characteristics of their feeding waveforms exhibited minimal variation across the different experimental conditions.
BCOA2 [28]. This dataset is derived from an experiment conducted on the bird cherry-oat aphid (R. padi) feeding on Wheat (Triticum aestivum var. Coker 9553, AgriPro®). Aphid feeding was recorded at the beginning of the day and the beginning of the night to understand the influence of the time of day on aphid feeding behavior. Aphids and experimental plants were reared in a 16 h light and 8 h dark photoperiod at 24 ± 1 °C. Experiments were conducted on seedlings at Zadoks stage Z1.2.
CA [29]. This dataset comes from an experiment conducted on the cannabis aphid (P. cannabis) feeding either on its primary host, Hemp (Cannabis sativa var. Elite, New West Genetics) or a secondary host, potato (Solanum tuberosum var. CO07015-4RU). The aim of this experiment was to understand the differences in the feeding behavior of viruliferous and non-viruliferous aphids on hemp and potato to determine the potential of the cannabis aphid to vector potato virus Y to these species. Experimental plants were reared in a greenhouse in a 16 h light/8 h dark photoperiod at 21 ± 1 °C/16 ± 1 °C, and two three-week-old hemp and potato plants were used in the experiments. Aphids performed poorly on potatoes, probably due to the difference in host plant physiology and cellular anatomy, and waveform characteristics between aphids feeding on potatoes and hemp plants are likely not identical.
SA [30]. This dataset is from the soybean aphid (A. glycines) feeding on two genotypes of Glycine max (soybean). The experiment compared aphid feeding behavior on the susceptible cultivar, Williams 82, and the resistant cultivar PI 567301B, which contains the Rag5 gene, both on whole plants and detached leaves. The aim was to investigate the origins of Rag5-mediated resistance. Aphids and plants were reared under a 16 h light/8 h dark photoperiod at 24 ± 1 °C, with plants at the V1 growth stage used for all recordings. Since experiments were conducted on both whole plants and detached leaves, waveform characteristics may differ across treatments.
GPA [31]. This dataset is from experiments with green peach aphid (M. persicae) feeding on Arabidopsis. The aim of the experiment is to understand the relationship between plant autophagy and insect feeding behavior and performance. Aphid colonies and experimental plants were reared in a 16 h light and 8 h dark photoperiod at 22 ± 3 °C. Aphids were allowed to feed on one-month-old wildtype Col-0 plants or autophagy mutants atg5.1 and atg7.2. Aphids showed reduced feeding on mutant atg7.2; however, waveform characteristics remained consistent across all experimental treatments.
These datasets are summarized in Table 1 for reference.
Data processing
Preprocessing.
Data splitting. For each dataset, 90% the total number of recordings (referred to as Subset 1) were used to train ML models and evaluate their classification performance. Subset 1 recordings were divided into fixed-length segments, creating a database of waveform samples, each with length d. This database was then shuffled and split into three parts: 70% for training, 10% for validation, and 20% for testing. The remaining subset (Subset 2) containing 10% the total number of recordings were reserved for the ultimate task of predicting waveform locations within complete recordings. Unlike Subset 1, recordings of Subset 2 were processed individually to maintain their completeness and avoid shuffling. Detailed data splitting information is provided in Table 2.
Next, we describe the data preprocessing techniques. In the following, we denote by a signal of L time steps and d channels.
Normalization. Data scaling is not mandatory but usually helps prevent numerical instability in calculations. To normalize each input recording x, we scaled the amplitude range between [ 0 , 1 ] using the following formula:
Augmentation. Recall that in the initial segmentation stage, the input was divided into fixed-length segments w (default: 10.24s or 1024 time steps). Waveform samples shorter than the predefined value d, were padded to ensure uniform size. Doing so, we implicitly allows one padded waveform sample to contain the transition between waveforms. This is reasonable because our algorithm works by classifying each consecutive segment with fixed endpoints, e.g. from 0s to 10.23s, 10.24s to 20.47s, 20.48s to 30.71s and so on. For all samples except those of pd, a waveform sample x=(xi,...,xj) (where i − j = n < d), was extended symmetrically from its midpoint using the following formula:
For samples of pd, whose cardinality was relatively small compare to others, we also performed oversampling in addition to padding to compensate for the issue. Specifically, padding was done with bigger pad size using
Then, we oversampled from by taking overlapping segments with a reduced step size s (default: 1.28s or 128 time steps), resulting in a set of 2d ∕ s + 1 samples for pd:
Features extraction.
Fourier transform (FT) (S1 Fig): A technique that converts signals from the time domain (amplitude/time representation) to the frequency domain (amplitude/frequency representation) by assuming that the input signal is a linear combination of sine and cosine waves at different frequencies. FT identifies which frequencies are present in the input signal by determining the coefficients of these component waves. This method has proven to be highly effective in many time series classification frameworks because it captures the characteristics of time series data well in their frequency domain representation [32–35]. Here, we fixed the hop length at 14 and the window size at 128 for generating spectrogram features.
Wavelet transform (WT) (S2 Fig): A technique in signal processing that assumes the input signal is a linear combination of a family of special functions called wavelets, which are obtained by translating and scaling a mother wavelet. For our study, the Morlet wavelet family with geometric scales of length 64 and the Symlet4 wavelet family with 3 levels of decomposition were used for the continuous WT and discrete WT, respectively. Both versions work based on the same fundamental principles, but for simplicity we focus on the discrete version. The discrete WT decomposes the original signal into several phases. In each phase, the signal is transformed into two complementary sets of wavelet coefficients: high frequency (detail, or "D") and low frequency (approximation, or "A") components. This is achieved by translating a wavelet along the time axis and convolving it with the input signal. The input signal is initially decomposed into "A1" and "D1". In each subsequent phase, the mother wavelet is scaled by a factor of 2, and the same decomposition process is applied to the approximation coefficients from the previous stage using this scaled wavelet. Wavelet transform has proven effective in various signal processing tasks, including analyzing human biological signals [36,37].
Gramian Angular Field (S3 Fig): A technique that converts a time-series into image format by representing the rescaling time-series, transforming it into polar coordinates, and finally representing it as a matrix comprising of the trigonometric sum (Gramian Angular Summation Field - GASF) or difference (Gramian Angular Difference Field - GADF) [38]. This representation allows us to apply 2D convolutional networks to time-series classification. In our study, GASF is used.
Statistical features: Mean, standard deviation, skewness, and quantiles are examples of these features, offering insights into data distribution. These features can capture the central tendency, dispersion, shape, and outliers of the data distribution, making them valuable for identifying changes or anomalies. Essentially, we can view these calculations as a method for dimensionality reduction because they can describe the data through a much smaller set of numbers. The specific statistical features used in this study include mean, root mean square, standard deviation, variance, skewness, the quantiles at level 0.05, 0.25, 0.5, 0.75, 0.95, and the zero crossing rates of the consecutive segments.
The complexity measures: Entropy measurements provide valuable insights into a time series’s regularity, long-range autocorrelation, and order structure. Shannon Entropy (SE) [39] was originally defined in classical information theory for quantifying the uncertainty of information directly through the probability of different situations. Meanwhile, Permutation Entropy (PE) [40] is a particularly robust tool for analyzing time series, as it quantifies the complexity of a dynamic system by examining the order relations between values and deriving a probability distribution of ordinal patterns. In addition to the substantial information provided, the calculation of SE and PE are relatively straightforward and quick to compute. The two entropy measurements and the statistical features are either calculated on the raw signal, the Fourier coefficients, or the discrete WT coefficients.
Model accuracy assessment
The performance of the ML models was first assessed based on their capability to classify individual waveform samples. Since our main interest is their effectiveness in localizing the waveform, we propose an additional evaluation step using the overlap rate between the predicted and the ground-truth complete annotation. We call these (Evaluation) Task 1 and (Evaluation) Task 2, respectively. The evaluations were done using a 10-fold cross-validation scheme to avoid biases, and the mean of these metrics is reported. In the following, we explicitly present the metrics used for each evaluation task.
For a classification problem, the following metrics are commonly used:
More specifically, the acronyms TP, TN, FP, and FN refer to True Positive, True Negative, False Positive, and False Negative, respectively. Assume that for a classification problem, we need to distinguish among L labels/classes (L = 7 in our case). Consider a label L∗; the observations/samples whose label is the same as L∗ are said to be Positive and Negative otherwise. True (False) Positive (Negative) refers to the number of observations that were correctly (incorrectly) classified as Positive (Negative). These quantities are often displayed in a confusion matrix of size L × L as illustrated in Table 4. The rows and columns in a confusion matrix correspond to the true and predicted labels, respectively. Hence, the diagonal elements represent correct classifications, while off-diagonal elements indicate errors.
For Task 1, as we are in a multiclass classification scenario, we decided to use the per-class accuracy scores, which measure the accuracies class-wise across the predictions on the test set. We record these values in a confusion matrix, which is normalized with respect to the number of predictions for each class. In addition, the overall accuracy (OA), which demonstrates the proportion of correct predictions across all classes, was considered. The average of the per-class F1 Score (avg-F1), which is a balanced measure of precision and recall of class-wise prediction, was also used.
For Task 2, we measure the degree of accuracy between the predicted and the manually annotated waveform. For an 8-hour recording consisting of 2,880,000 data points, 2,880,000 predicted labels are produced. The accuracy score between the predicted and the ground-truth labels (called the overlap rate, or OR) is calculated, showing the extent to which the predicted labels agree with the ground-truth labels.
Experimental settings
Setups.
For each ML model, we performed 3 groups of experiments with respect to 3 different techniques to preprocess its inputs, namely no transformation, Fourier transform, and Wavelet transform (See Table 3). Each group consists of 5 experiments on individual datasets having different settings.
One of the main goals of our study is to develop a transferable model applicable to datasets derived from various aphid species, host plants, and experimental conditions. To enhance the analysis, we further evaluated the performance of two representative models using their best input feature found in the previous experiments, 1DCNN and XGB, on the combined dataset created by merging data from BCOA1, BCOA2, CA, SA, and GPA. We selected 1DCNN to train on raw data over the other candidates because of its computational simplicity and performance comparable to the other two methods. The XGB model with WT features was selected as it is the best model in terms of both training speed and performance metrics out of the non-deep learning models.
Models configurations.
The 1DCNN model consists of three convolutional layers that output 64, 128, and 64 channels, respectively. Each layer uses a dilation of 3 and a stride of 2, followed by ReLU activation and batch normalization. After these convolutional layers, there is a max pooling layer with a size of 3. The feature maps are then flattened, subjected to dropout with a probability of 0.5, and passed through a fully-connected layer.
The ResNet model adopts the idea of skip-connection. This model comprises three skip-connection blocks, each containing three consecutive convolutional layers with d1, d2, and d3 output channels, respectively, followed by ReLU activation, a skip connection, and finally, normalization. We set the number of output channels for each block to 64, 128, and 128. After each convolution block, max pooling with a kernel size of 3 and a stride of 3 is performed to reduce computational complexity. The outputs of the convolutional layers were flattened out, and a dropout with a probability of 0.5 was applied for normalization before being passed through a final fully connected layer.
The 2DCNN model includes two convolutional layers, each outputting 16 channels and using a 3 × 3 kernel with a stride of 1. Each convolutional layer is followed by ReLU activation and max pooling with a size of 2 and a stride of 2. The results are then flattened and passed through two fully connected layers with dropout to obtain the probability distribution for the 7 classes. The dropout rates are set to 0.5 for both fully-connected layers, with kernel size and stride for the max pooling layers set to 2.
The Cross-Entropy loss function is used for training the DL models, which is typical for multilabel classification tasks. The baseline hyperparameters are set to {256, 0.0001, 9, 100}, representing the batch size, learning rate, kernel size, and number of training epochs, respectively.
For traditional ML algorithms, we configured XGB with 100 trees (n = 100), a learning rate of 0.3 (η = 0.3), and a maximum tree depth of 6 (dmax = 6). For RF, n is set to 100, but dmax is not specified. For Logistic Regression, we used the default parameters provided by scikit-learn.
Hyperparameter tuning.
Hyperparameter tuning for optimal parameter selection is critical in ML model training. However, our work only employed hyperparameter tuning on the combined dataset (i.e., on two representative models, 1DCNN and XGB) due to the high demand for computational resources. We must note that although the hyperparameter tuning process is done completely automatically, the selection of hyperparameters’ ranges is heuristic, and there is no concrete procedure for this. We started with a set of heuristically chosen hyperparameters. Once their performance was validated, the values of the hyperparameters were then varied in a range including the initial value. In our study, we only perform hyperparameter tuning on the combined dataset and the XGB and 1DCNN models due to the immense computational requirements if we were to also conduct it on individual datasets.
The Grid Search (GS) technique was chosen to fine-tune the XGB model. We conducted an exhaustive search to identify the best combination by manually exploring all possible scenarios using the whole set of 52 features. The hyperparameters of interest were number of trees, learning rate, and maximum tree depth drawn from {50, 100, 200, 300}, {0.01, 0.1, 0.2, 0.3} and {3, 4, 5, 6}, respectively.
For tuning 1DCNN, we employed a more aggressive strategy due to the computational complexity, which involves manually adjusting hyperparameters in a fixed sequence: batch size, learning rate, kernel size, and then the number of epochs. The hyperparameters are selected from the following ranges: batch size {128, 256, 512, 1024, 2048, 4096}, learning rate {10−2, 10−3, 10−4, 10−5}, kernel size {5, 9, 15}, and number of training epochs {20, 50, 100}.
Hardware and software.
Our experiments were conducted on a personal laptop GIGABYTE G7 equipped with a 12-core Intel i5-10500H CPU (2.5GHz) processor and an NVIDIA GeForce RTX 3060 Laptop GPU (6GB memory). Training times ranged from half an hour to two hours, depending on the data volume, while the segmentation task was typically completed in minutes. All scripts of the DL models are implemented using the Pytorch framework [41], while the traditional ML models are implemented using scikit-learn. For signal processing tasks, we use librosa and PyWavelet to perform Fourier transform and Wavelet transform, respectively.
Results
Training on individual datasets
Deep learning models.
The 1DCNN and ResNet models excelled and yielded superior outcomes when trained with different types of inputs, showing the strong learning capability of DL models. These models achieved near-perfect OA and average F1 scores across the three groups of experiments. The highest OAs for ResNet were 96.8%, 93.9%, and 95.3% with raw, FT, and WT inputs, respectively. Likewise, those for 1DCNN were 92.8%, 91.6% and 93.8%, respectively. On average, the OAs of ResNet were 2.6%, 2.5%, and 3.7% higher than those of 1DCNN, while ResNet also demonstrates significantly better performance in terms of the avg-F1 scores, with 4.0%, 2.6% and 6.6% higher on average. In addition, ResNet perform annotation noticeably well with simple raw input, achieving an average of 84.5% OR over the 5 datasets, with the highest of 94.8% observed in the BCOA1 dataset. However, using transformed inputs with FT and WT decreased the generalizability of ResNet by reducing the average OR of ResNet to 81.4% and 81.8%, respectively. This means that raw data were more informative to this model, as it is robust to capture relevant features from this input type. The waveforms of aphids are characterized only by certain frequency ranges. Therefore, using all Fourier coefficients for training the model might be redundant. In contrast, this additional information was useful for 1DCNN, whose architecture is much simpler than ResNet, as it increases the mean OR of 1DCNN to 80.8% using FT inputs. The simplicity in the design of 1DCNN made it less vulnerable to overfitting. Thus the more complex features were more beneficial for this model. The per-class accuracy scores given by the confusion matrix in (Fig 4) reveal that ResNet was slightly less efficient than 1DCNN when trained on a small dataset such as SA, indicating potential overfitting in the model training.
The 2DCNN model differs from 1DCNN and ResNet as it requires image inputs. To implement this model, we first generated input images of segments using three different techniques: GASF, spectrogram, and scalogram. The model yielded the best performance when trained with scalogram inputs, reaching the highest OA and avg-F1 of 95.4% and 93.5% with the BCOA1 dataset. While the spectrogram inputs may produce similar classification results, model performance with scalogram as input could achieve better generalization as there was an 8% increase in mean OR, from 75.7% with spectrogram to 83.7% with scalogram. We further investigated the confusion matrices of the 2DCNN model with scalogram input, as this approach yields the best performance. On all five datasets, we observed that 2DCNN could misclassify between pd and C waveforms, which is understandable as C is a complex waveform and pd is a part of C. The scalograms of pd may have captured redundant frequency information. With the SA dataset, 2DCNN reached perfect accuracy for E1, while ResNet and 1DCNN struggled to identify this waveform (see Fig 4). Despite experiencing fast learning after the initial 20 epochs (see Table 5), the test scores of 2DCNN show that it can generalize quite well even after 100 epochs of training.
In (Fig 5), we visualize the training process of ResNet, 1DCNN, and 2DCNN through the loss and accuracy curve when training with raw inputs. A similar pattern can be observed when training with the two other types of inputs. The training process of 1DCNN experienced a spiking phenomenon, a common phenomenon in stochastic gradient descent, which should only raise concern if training or validation accuracy declines. Conversely, ResNet exhibited a much smoother loss curve and a tendency to overfit rapidly, especially after just 20 epochs. This underscores the importance of robust regularization techniques to enhance its generalization capabilities across diverse datasets.
Traditional ML models.
Regarding the non-deep learning models, the Extreme Gradient Boosting classifier (XGB) performed fairly well and achieved results that were comparable to the best case of the DL algorithms, especially to the best metrics reported by ResNet, and outperformed Logistic Regression in all cases. Features extracted from WT at various resolutions demonstrated the effectiveness in characterizing the EPG waveforms, which was also the most effective of the three signal processing techniques for XGB. Indeed, WT provides the models with richer information about the frequency domain of the input signal, with a total of 4 resolutions instead of 1 in FT. Moreover, the transformation principle of WT helps localize the waveform behavior regarding its frequency, which is crucial for characterizing EPG waveforms as they are defined based on the observed patterns across a certain range of frequencies. For classification tasks, the OAs range from 94.2% to 96.4%, while the avg-F1 ranges from 90.9% to 95.8%. The OR reported by XGB demonstrates a robust annotation performance, with an average OR of 84.5% across 5 datasets. Meanwhile, Random Forest (RF) with WT features showed comparative performance, and its accuracy metrics closely follow XGB’s in all experiments. The highest mean OA, avg-F1, and OR metrics obtained by RF when trained on 5 datasets are 93.4%, 89.5%, and 81.8%, respectively. This is unsurprising as RF and XGB are built on the same basis of aggregating the results from a set of weaker learners - decision trees. Details of the classification results are reported in Table 6. Overall, the two models demonstrated impressive performance across global scores and individual per-class scores, often exceeding 90% with only a few exceptions that were slightly lower (see (Fig 6)).
Training on the combined dataset
Deep learning model.
The 1DCNN was selected out the three DL models experimented based on its competitive prediction accuracy as well as its simple architecture design. For 1DCNN, 16 experiments for hyperparameter tuning were conducted, and the outcomes are detailed in Table 7. Ultimately, we determined that a batch size of 256, a learning rate 0.0001, and a minimum kernel size 9 yield the best outcomes. The left subfigure of (Fig 7) shows the training curves of the 1DCNN models with a batch size of 256, a learning rate of 10−4, and a kernel size of 9. As seen, a number of training epochs from 20 to 50 was generally sufficient for the model to achieve robust learning. The middle and right figures show the final metrics for both tasks of the 1DCNN models at an epoch of 100. The 1DCNN model achieves high classification scores at 90.2% OA, and 90.2% avg-F1, respectively. The per-class accuracies showed that the highest misclassification is frequently observed between C and pd throughout our studies, primarily due to pd always being nested between the C waveform. Meanwhile, the OR rates across the 5 proposed datasets were 85.5%, 81.5%, 65.3%, 83%, and 63.1% for BCOA1, BCOA2, CA, GPA, and SA, respectively. We observe low ORs in the datasets that contain non-uniform samples of aphids’ behavior, such as CA and SA, which indicates that the uniformness of the recorded EPG signal can affect the overall quality of the ML models.
Traditional ML model.
Similar to 1DCNN, we conduct a training of a general model with Extreme Gradient Boosting Classifier (XGB) as this model demonstrate great performance in terms of accuracy and speed. For XGB, the inputs include 52 statistical and non-linear complexity features. Therefore, it is beneficial to perform feature selection (FS) as an attempt to reduce the overall computational complexity. Furthermore, FS is also pivotal to discerning correlations and determining the best training features. Employing Pearson’s correlation score, we identified features with correlations exceeding 0.8 as highly correlated. Fig 8 shows the correlations matrix before and after FS, in which we note that the feature group containing the k-quantiles {0.05, 0.25, 0.5, 0.75, 0.95} at all three WT detail resolutions are highly correlated, so we only keep the median (or the 0.5 quantile) features and discard the others. Moreover, the other feature groups consisting of standard deviation, variance, and root-mean-square also show strong correlations, as they are very similar in terms of calculation. We remove variance and root mean square features and only keep the standard deviation as representative for the three features for each resolution. The remaining 24 features include the means, standard deviations, skewnesses, zero-crossing rates, and the Shannon and permutation entropies of four resolutions A3, D3, D2, and D1. We note that FS dramatically improves the feature engineering and extraction step speed, alleviating the burdensome time requirements associated with these processes.
Regarding hyperparameter tuning, the optimal hyperparameters found are {300, 0.3, 6} for a number of estimators, learning rate, and maximum depth. Table 8 shows the results of different training settings. Using a smaller set of features (FS), XGB attained a high classification accuracy for all the waveforms and only experienced a slight reduction in OA and avg-F1 metrics from 92% to 89.1% and from 92% to 88.9%, respectively. These results can be slightly improved by using the optimal set of hyperparameters obtained from Grid Search (FS + GS), which consequently yields 90.9% OA and 90.8% avg-F1. The most accurate prediction results were achieved using identical hyperparameters on the original dataset with 52 features (GS). (Fig 9) display the evaluation metrics of XGB in more detail.
Discussion
This study presents the benchmark results of six ML models in EPG signal annotation for characterizing aphid feeding behavior. The annotation procedure was implemented through three main steps: initial segmentation, waveform sample classification, and label aggregation. Two evaluation steps were conducted to thoroughly evaluate each model regarding individual waveform sample classification (Task 1) and overall annotation quality assessment (Task 2). For Task 1, the prediction accuracy of individual segments was evaluated, including the classification accuracy calculation and the F1 score. Meanwhile, the overlap rate between the predicted and the ground-truth annotated waveforms was estimated for Task 2. The results show that ML models, in general, are robust enough to annotate and characterize EPG signals and aphid-feeding behavior. Two representative models, 1DCNN and XGB, were selected based on their prediction accuracy and computational complexity to be additionally evaluated on a large aggregated dataset. The results show that our generalized models can be transferred to other datasets collected from different experimental conditions and host plant species.
Next, we shall discuss an issue regarding the waveform localization errors of our approach. In (Fig 10), the predicted annotation from the 1DCNN model almost perfectly matches the ground-truth version, reaching a commendable OR of 95%. Often, long and dominant waveforms such as G, F, and E2 are correctly identified, although instances of non-overlap are visible, which is primarily due to the offset between the predicted and ground-truth waveforms. This issue is; however, an unavoidable issue with the sliding approach since we classify the time steps every consecutive 1024 time steps, e.g., from 0 to 1024, from 1024 to 2048, from 2048 to 3072, until we reach the end of the input recording. On the other hand, the natural behavior of aphids is not fixed in such a way, leading to the misalignment of the waveform endpoints. The phenomenon is particularly profound with short waveforms such as pd, which ML models often overlook. Thus, users should conduct a post-prediction evaluation to mitigate these inaccuracies.
Throughout the experiments, we noticed that the consistency and uniformness of aphid’s behavior significantly affect the accuracy metrics used to assess the performance of the ML models, especially the OR metrics. Therefore, we present (Fig 11), which demonstrates a situation in a recording from the SA dataset where the 2DCNN model attempts to annotate a long section of the first half of the recording differently than the ground-truth version. To be explicit, the upper panel displays the ground-truth annotation containing a long green section labeled as the C waveform. In the lower panel, different labels were given to this section, which are E1, E2, and NP (in light blue, pink, and light brown, respectively). The annotation provided by the ML model in this example is reasonable, as it reflects behaviors that align with the theoretical behavior observed in other feeding stages, particularly the phloem-feeding stage. While some false positives were noted, ML models appear robust and consistent with the learned pattern. The experimental results also support this claim, as we often observe the highest accuracy metrics in experiments where aphid behaviors experience minimal variation, such as in the BCOA1 and GPA datasets. In the BCOA2 and SA datasets, where the aphid’s waveform pattern is not identical throughout the experiments, the predicted waveform often has a low overlap rate with the ground-truth version. Hence, providing the ML models with well-annotated data for training is essential to ensure the best performance.
Compared to the previous works on applying ML on characterizing EPG signal, our works showed several advantages. First and foremost, our datasets is not only extensive in volume, but also diverse in biological background. It is composed of a total of 2656 hours (332 8-hour recordings) of data, compared to 470 hours [14] and only 7000s [13], while includes experiments on 4 aphid species and 5 different host plant. To our understanding, our approach is more similar to that of Xing et al. [13] in the sense that we classify waveform samples of uniform length (our default length was 10.24s compared to theirs at 10s). Regarding the classification results, the highest mean classification accuracy was achieved by ResNet at 94.43%, very close to 94.47%. Although Willett et al. [14] reported 97.4% accuracy with RF, it was difficult to make direct comparison to ours since the length and specific data processing techniques were not mentioned. The advtanage of our work is also presented in the final step of our pipeline, which was the label aggregation step that returns a complete analysis file specifying the predicted locations of each waveform. Furthermore, we evaluated this aggregated result by using the OR metric, which was not done in previous work.
Nevertheless, our study demonstrates several limitations that require further investigation. As waveform samples were fixed-length and were treated independently, the models was not exceptionally good at annotating short waveforms such as pd that exist within C, a major waveform. The length of the waveform sample also affects the level of offset between the endpoints of the predicted and ground-truth waveform, which we are willing to tolerate. A possible solution is to adopt a more flexible algorithm that can both learn these boundary positions and classify the waveform contained within. An analogous problem to EPG signal annotation is image segmentation. Various DL models were designed to effectively tackle this problem such as Mask R-CNN, YOLO and SAM, which have greatly outperformed traditional algorithms. For time-series adaptation, U-Time [23] is a CNN-based model that has shown effectiveness for sleep-staging, a similar problem to segmenting an EPG recording. We believe that by adopting this idea, the reported results of our study can be greatly advanced.
In conclusion, we have shown that the ML methods presented can achieve great accuracy and have the potential to greatly increase the productivity of EPG research. For example, the average EPG experiment can have a sample size ranging anywhere from 15–60 recordings, depending on the number of treatments. With the average recording taking anywhere from 10-30 minutes to annotate, this can lead to hours of annotation for a given experiment. Due to the cumbersome task of manual annotation, often this task is distributed amongst multiple researchers, which can lead to potential operator error and variation in the annotated waveforms. Thus, having accurate ML tools widely available to EPG researchers will allow for higher accuracy amongst waveform calls and allow for the users to increase sample sizes without risking the addition of large analysis timelines or high operator variability. In addition, an open-source Python package available to the public will enable EPG researchers to fine-tune and customize models that can be used to better fit their research needs. For example, a research group could create a custom model based on their labs recording data for a species of interest and pick and choose which models and model parameters work best in that particular system. The availability for EPG users to have access to ML tools allows for the field of EPG to gain the rigor associated with using computational tools rather than relying solely on the human eye. While further work needs to be done to allow even the most inexperienced users to utilize these tools, potentially via easily accessible web applications, this manuscript and the associated software packages provide a significant step in the introduction of ML tools for use in EPG waveform annotation. In conclusion, using ML tools in EPG research has the ability to enhance the accuracy of waveform classification across the field and significantly reduce the time and effort required for annotation.
Supporting information
S1 Fig. Fourier transform and spectrogram. (A) An example of a 10.24s segment extracted from an E1 waveform. (B) The coefficients of positive frequency signal obtained by performing Fourier transform on the segment. This is often called the frequency domain representation of the segment. (C) Representative spectrograms of segments extracted from 7 waveforms, which are the results of short-time Fourier trasnsform.
https://doi.org/10.1371/journal.pone.0319484.s001
(TIF)
S2 Fig. Wavelet transform and scalogram. (A) The left panel displays the hierachical structure of the discrete wavelet transform decomposition. The entire signal is initially decomposed into approximation and detail coefficient, corresponding to low and high frequency, respectively. In each subsequent stage, the approximation coefficient are further decomposed in a similar manner. On the right hand side, one observes an example of decomposing a fixed-length segment using wavelet transform with symlet4 wavelet. (B) The scalograms of segments extracted from 7 waveforms which were obtained using Morlet wavelet and continuous wavelet transform.
https://doi.org/10.1371/journal.pone.0319484.s002
(TIF)
S3 Fig. Gramian angular field transformation. (A) The process of generating a Gramian Angular Summation Field from an input time-series is presented from left to right. The time-series are transformed into polar coordinate, then convert into a matrix consisting of trigonometrics values between pairs of angles. (B) Represntative GASFs of segments extracted from 7 waveforms.
https://doi.org/10.1371/journal.pone.0319484.s003
(TIF)
References
- 1. Zapata S, Dudensing R, Sekula D, Esparza-Díaz G, Villanueva R. Economic impact of the sugarcane aphid outbreak in South Texas. J Agric Appl Econ. 2018;50(1):1–18.
- 2. McLean DL, Kinsey MG. A technique for electronically recording aphid feeding and salivation. Nature 1964;202(4939):1358–9.
- 3. Tjallingii WF, Esch THH. Fine structure of aphid stylet routes in plant tissues in correlation with EPG signals. Physiol Entomol 1993;18(3):317–28.
- 4. Schaefers GA. The use of direct current for electronically recording aphid feeding and salivation. Ann Entomol Soc Am. 1966;59(5):1022–4.
- 5. Tjallingii WF. Electronic recording of penetration behaviour by aphids. Entomol Exp Appl. 1978;24(3):721–30.
- 6. Backus EA. History, development, and applications of the AC electronic monitoring system for insect feeding. History, development, and application of AC electronic insect feeding monitors. 1994. p. 1.
- 7. Tayyab U, Arif M, Gogi M, Akhtar S, Abdullah M, Ali F. Tracking the feeding mechanism of sap-sucking insect-pests through Electropenetrography (EPG). J Insect Behav. p. 1–24 (2024)
- 8. Salvador VR, Tjallingii F. A new application of the electrical penetration graph (EPG) for acquiring and measuring electrical signals in phloem sieve elements. J Visualiz Exp. 2015;106:e53250.
- 9. Sarria E, Cid M, Garzo E, Fereres A. Excel workbook for automatic parameter calculation of EPG data. Comput Electron Agricult. 2009;67(1–2):35–42.
- 10. Prüfer T, Thieme T, Tjallingii W. Aphid-AutoEPG software for analysing electrically monitored feeding behaviour of aphids. Eur J Environ Sci. 2014;4(1):53–9.
- 11. Adasme F, Muñoz C, Salinas-Cornejo J, Ramírez C. A new software for the analysis of electrical penetration graphs to study plant probing behavior of hemipteran insects. Comput Electron Agricult. 2015;113:128–35.
- 12. Wu L, Jia S, Xing Y, Lu S, Pan J, Yan F. Machine identification of electrical penetration graphic waveforms of aphid based on fractal dimension and Hilbert-Huang transform. Trans Chinese Soc Agricult Eng. 2018;34(8):175–83.
- 13. Xing Y, Li B, Wu L, Yan F. Waveforms eavesdropping prevention framework: the case of classification of EPG waveforms of aphid utilizing wavelet kernel extreme learning machine. Appl Artif Intell. 2023;37(1):79–100.
- 14. Willett DS, George J, Willett NS, Stelinski LL, Lapointe SL. Machine learning for characterization of insect vector feeding. PLoS Comput Biol 2016;12(11):e1005158. pmid:27832081
- 15. Mendoza QA, Pordesimo L, Neilsen M, Armstrong P, Campbell J, Mendoza PT. Application of machine learning for insect monitoring in grain facilities. AI 2023;4(1):348–60.
- 16. Gerovichev A, Sadeh A, Winter V, Bar-Massada A, Keasar T, Keasar C. High throughput data acquisition and deep learning for insect ecoinformatics. Front Ecol Evol. 2021;9:727056.
- 17. Tannous M, Stefanini C, Romano D. A deep-learning-based detection approach for the identification of insect species of economic importance. Insects 2023;14(2):148. pmid:36835717
- 18. Butera L, Ferrante A, Jermini M, Prevostini M, Alippi C. Precise agriculture: effective deep learning strategies to detect pest insects. IEEE/CAA J Autom Sinica 2022;9(2):246–58.
- 19. Hong S, Nam I, Kim S, Lee J, Park Y, Lee S. Automatic pest counting from pheromone trap images using deep learning object detectors for Matsucoccus thunbergianae monitoring. IEEE/CAA J Autom Sinica. 2021:8(4);342.
- 20. Morokuma S, Hayashi T, Kanegae M, Tanabe Y, Kawahara T, Nagai K. Deep learning-based sleep stage classification with cardiorespiratory and body movement activities in individuals with suspected sleep disorders. Sci Rep. 2023;13:1984.
- 21. Yildirim O, Baloglu UB, Acharya UR. A deep learning model for automated sleep stages classification using PSG signals. Int J Environ Res Public Health 2019;16(4):599. pmid:30791379
- 22. Lawhern V, Solon A, Waytowich N, Gordon S, Hung C, Lance B. EEGNet: a compact convolutional network for EEG-based brain-computer interfaces. J Neural Eng. 2016;15.
- 23. Perslev M, Jensen M, Darkner S, Jennum PJ, Igel C. U-Time: a fully convolutional network for time series segmentation applied to sleep staging. Adv Neural Inf Process Syst. 2019;32:4417–28.
- 24. Deng H, Runger G, Tuv E, Vladimir M. A time series forest for classification and feature extraction. Inf Sci. 2013;239:142–53.
- 25. Baydogan MG, Runger G, Tuv E. A bag-of-features framework to classify time series. IEEE Trans Pattern Anal Mach Intell 2013;35(11):2796–802. pmid:24051736
- 26. Florek P, Zagdański A. Benchmarking state-of-the-art gradient boosting algorithms for classification. ArXiv. 2023.
- 27. Zadoks J, Chang T, Konzak C. History, development, and applications of the AC electronic monitoring system for insect feeding. Weed Res. 1974;14(6):415–21.
- 28. Nalam V, Isaacs T, Moh S, Kansman J, Finke D, Albrecht T, et al. Diurnal feeding as a potential mechanism of osmoregulation in aphids. Insect Sci 2021;28(2):521–32. pmid:32240579
- 29. Pitt WJ, Kairy L, Villa E, Nalam VJ, Nachappa P. Virus infection and host plant suitability affect feeding behaviors of cannabis aphid (Hemiptera: Aphididae), a newly described vector of potato virus Y. Environ Entomol 2022;51(2):322–31. pmid:35243512
- 30. Joshi K, Baumgardner JL, MacPhail M, Acharya SR, Blotevogel E, Dayan FE, et al. The source of Rag5-mediated resistance to soybean aphids is located in the stem. Front Plant Sci. 2021;12:689986. pmid:34335657
- 31. Hao L, Dangol A, Shavit R, et al. Self-eating while being eaten: elucidating the relationship between aphid feeding and the plant autophagy machinery in Arabidopsis leaves. bioRxiv. 2023.
- 32. Kumar M A, Chakrapani A. Classification of ECG signal using FFT based improved Alexnet classifier. PLoS One 2022;17(9):e0274225. pmid:36166430
- 33. Li M, Chen W. FFT-based deep feature learning method for EEG classification. Biomed Signal Process Control. 2021;66:102492.
- 34. Xu J, Hugelier S, Zhu H, Gowen A. Deep learning for classification of time series spectral images using combined multi-temporal and spectral features. Anal Chim Acta. 2020;1143.
- 35. Fadaei S, Rashno A, Hamidi A. Speaker recognition using convolutional neural network and neutrosophic. J Model Eng. 2023;21(75):1–18.
- 36. Subasi A. EEG signal classification using wavelet feature extraction and a mixture of expert model. Exp Syst Appl 2007;32(4):1084–93.
- 37. Malik F, Barin S, Yuksel M. Accurate classification of heart sound signals for cardiovascular disease diagnosis by wavelet analysis and convolutional neural network: preliminary results. In: Proceedings of the 28th Signal Processing and Communications Applications Conference (SIU). 2020.
- 38. Wang Z, Oates T. Imaging time-series to improve classification and imputation. In: Proceedings of the 24th International Conference on Artificial Intelligence. 2015. p. 3939–45.
- 39. Shannon CE. A mathematical theory of communication. Bell Syst Tech J 1948;27(3):379–423.
- 40. Bandt C, Pompe B. Permutation entropy: a natural complexity measure for time series. Phys Rev Lett 2002;88(17):174102. pmid:12005759
- 41. Paszke A, Gross S, Massa F, Lerer A, Jamieson K, Liu J, et al. PyTorch: an imperative style, high-performance deep learning library. Adv Neural Inf Process Syst. 2019;32:8024–35.
- 42. Wang Z, Yan W, Oates T. Time series classification from scratch with deep neural networks: a strong baseline. Int Jt Conf Neural Netw. 2017;21(75):1578–85.
- 43. Ismail Fawaz H, Forestier G, Weber J, Idoumghar L, Muller P-A. Deep learning for time series classification: a review. Data Min Knowl Disc 2019;33(4):917–63.
- 44. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. IEEE Conf Comput Vis Pattern Recognit (CVPR). 2016. p. 770–8.