Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

HybridoNet-Adapt: A domain-adapted framework for accurate lithium-ion battery RUL prediction

  • Khoa Tran,

    Roles Formal analysis, Investigation, Methodology, Resources, Writing – original draft, Writing – review & editing

    Affiliation AIWARE Limited Company, Da Nang, Vietnam

  • Bao Huynh,

    Roles Investigation, Methodology, Visualization, Writing – original draft, Writing – review & editing

    Affiliation AIWARE Limited Company, Da Nang, Vietnam

  • Tri Le,

    Roles Investigation, Methodology, Validation, Writing – original draft, Writing – review & editing

    Affiliation AIWARE Limited Company, Da Nang, Vietnam

  • Lam Pham,

    Roles Data curation, Validation, Visualization

    Affiliation AIT Austrian Institute of Technology GmbH, Vienna, Austria

  • Vy-Rin Nguyen,

    Roles Investigation, Methodology, Software, Writing – original draft, Writing – review & editing

    Affiliation Software Engineering Department, FPT University, Da Nang, Vietnam

  • Duong Tran Anh,

    Roles Data curation, Validation, Visualization

    Affiliations Laboratory of Environmental Sciences and Climate Change, Institute for Computational Science and Artificial Intelligence, Van Lang University, Ho Chi Minh City, Vietnam, Faculty of Environment, Van Lang School of Technology, Van Lang University, Ho Chi Minh City, Vietnam

  • Hung-Cuong Trinh

    Roles Conceptualization, Funding acquisition, Project administration, Supervision, Writing – original draft, Writing – review & editing

    trinhhungcuong@tdtu.edu.vn

    Affiliation Natural Language Processing and Knowledge Discovery Research Group, Faculty of Information Technology, Ton Duc Thang University, Ho Chi Minh City, Vietnam

Abstract

Accurate prediction of the Remaining Useful Life (RUL) of lithium-ion batteries is critical for safe, reliable Battery Health Management in diverse operating conditions. Existing RUL models often fail to generalize when test data diverge from the training distribution. To address this, we introduce HybridoNet-Adapt, a domain-adaptive RUL prediction framework that explicitly bridges the gap between labeled source and unlabeled target domains. During training, we minimize the Maximum Mean Discrepancy (MMD) between feature distributions to learn domain-invariant representations. Simultaneously, we employ two parallel predictors—one tailored to the source domain and one to the target domain—and balance their outputs via two learnable trade-off parameters, enabling the model to dynamically weight domain-specific insights. Our architecture couples this adaptation strategy with LSTM, multi-head attention, and Neural ODE blocks for deep temporal feature extraction, but its core novelty lies in the MMD-based alignment and hybrid prediction mechanism. On two large, publicly available battery datasets, HybridoNet-Adapt consistently outperforms non-adaptive baselines (Structural Pruning, Multi-Time Scale Feature Extraction Hybrid model, XGBoost, Elastic Net), archiving an RMSE reduction of up to 152 cycles under domain shifts. These results demonstrate that incorporating domain adaptation into RUL modeling substantially enhances robustness and real-world applicability.

1 Introduction

1.1 Motivations

Lithium-ion batteries (LIBs) [1], renowned for their affordability and high energy density, are extensively utilized [25] in electric vehicles (EVs), portable devices, and energy storage stations. The global lithium-ion battery (LIB) market is projected to surpass 170 billion dollars by 2030 [6]. With the wide-ranging adoption of LIBs, interest in battery health management (BHM) has surged within both academia and industry in recent years. In a BHM system, several common and essential techniques are employed, including thermal management [7,8], fault diagnosis/detection [9], state of charge (SOC) [1012] and state of health (SOH) [13] estimation, remaining useful life (RUL) prediction [14,15], and cycle life early prediction [1619]. Among these, RUL prediction plays a crucial role in ensuring the proactive maintenance, minimizing downtime, and enhancing the operational efficiency of LIBs over their lifespan. RUL can be assessed based on the number of remaining cycles the battery can undergo before reaching its EOL. RUL prediction falls into three categories: model-based, data-driven, and hybrid approaches.

Model-based approaches often utilize physics-based degradation models, such as the Double Exponential Model (DEM) [20], two-phase degradation models [21], and Markov Models [22], constructed using early-cycle data (200-500 cycles) to forecast the entire battery’s capacity degradation curve. However, relying solely on maximum discharge capacity degradation during early cycles often leads to inaccuracies due to the influence of various factors (current, voltage, temperature, time) and sudden changes in degradation trends [20,21].

Data-driven models predict the RUL of LIBs by analyzing data during current cycles. Techniques like dual-input Deep Neural Networks (DNN) [23], 1D Convolutional Neural Networks (1DCNN) [24], Dense layers [25], Long Short-Term Memory (LSTM) networks [26], and Echo State Networks (ESN) [27] have shown superior performance.

Hybrid approaches combine model-based and data-driven methods to improve RUL prediction. For instance, a hybrid model using the Double Exponential Degradation Model (DEDM) and Gated Recurrent Unit (GRU) network fused with a Bayesian neural network (BNN) offers enhanced predictions [28].

Accurately predicting the RUL of LIBs remains a significant challenge due to the complex, non-linear, and cycle-dependent degradation behavior of battery features. This variability necessitates a highly adaptive prediction model capable of tracking and learning degradation patterns over time throughout charge–discharge cycles. The specific challenges associated with this task will be discussed in the following section.

1.2 Problem statement and potential

As summarized in Table 1, current state-of-the-art studies categorize RUL prediction methods into two primary approaches: historical data-independent methods, which estimate the current RUL based on current and preceding few cycles, and historical data-dependent methods, which leverage early-cycle data to predict the battery’s full lifespan.

thumbnail
Table 1. Overview of RUL prediction methods in LIB research.

https://doi.org/10.1371/journal.pone.0335066.t001

Historical data-dependent methods estimate the future capacity trajectory. The EOL cycle index is then determined, typically defined as 70% [33] or 80% [20] of the nominal capacity, and the RUL is calculated as the difference between the current cycle index and the EOL cycle index. These approaches primarily rely on early-cycle data. While historical data-dependent methods can achieve reasonable accuracy in benchmark evaluations [28,38], they face practical limitations such as the unavailability of early-cycle records, varying operational conditions [17] throughout a battery’s lifespan, and challenges in battery repurposing [43]. Therefore, historical data-independent approaches are considered more suitable for real-world scenarios.

Small datasets, like the Oxford Battery dataset (13 cells) and NASA battery datasets (4–34 cells) limit model robustness in real-world failure prediction. In contrast, large datasets such as the TRI dataset (124 cells, fast-charging) and the LHP dataset (77 cells, diverse discharge) provide extensive charge-discharge scenarios, making them well-suited for both training and testing of data-independent models.

Signal preprocessing can be limited by high dimensionality, especially with variational decomposition methods like EMD and VMD, which preserve or expand the original signal size. In contrast, statistical feature extraction methods—such as mean and standard deviation—offer a low-dimensional, efficient alternative that captures essential characteristics, making them ideal for real-time industrial applications.

Model-based and hybrid approaches typically rely on early-cycle data for RUL prediction, yet each battery exhibits unique degradation patterns over its lifespan requiring adaptive data-driven strategies. In data-driven approaches, domain adaptation (DA) techniques such as domain-adversarial neural networks (DANN)[44,45] and generative adversarial networks (GANs)[46] offer effective solutions for transferring degradation patterns from a source domain to improve prediction in the target domain. [31] proposed a transferable RUL prediction method using DA that enforces cycle-consistency of degradation trends across batteries, aligning feature representations to mitigate domain shifts. Their approach improves cross-battery generalization, it relies on comparable degradation levels between source and target domains. This work demonstrates that domain adaptation techniques benefit RUL prediction in the battery domain, highlighting an direction for further exploration. To address these mentioned challenges and leverage the identified potential, our proposed approach is introduced in the next section.

1.3 Main contribution

The main contributions of this work are summarized as follows:

  • We propose a historical data-independent RUL prediction framework for lithium-ion batteries that relies solely on current and recent cycling data, eliminating the need for early-cycle information. The prediction model integrates advanced deep learning components—including Long Short-Term Memory (LSTM), Multihead Attention, and Neural Ordinary Differential Equations (NODE) blocks—as a powerful feature extractor, along with linear layers in the predictors. Furthermore, a domain adaptation strategy is employed, combining two predictors with trainable trade-off parameters and an MMD-based loss to learn domain-invariant features, thereby enhancing transferability from source to target domains. This strategy improves the generalizability of the prediction model.
  • The framework includes a lightweight yet robust preprocessing pipeline-noise reduction, statistical feature extraction (e.g., mean, standard deviation), and normalization—to improve signal quality and reduce dimensionality for efficient, real-time prediction.
  • Extensive evaluations on the two largest publicly available datasets of A123 APR18650M1A cells [16,23], covering diverse charging and discharging conditions, validate the superior performance of our approach in real-world battery health management.

The remainder of this paper is organized as follows. Sect 2 introduces the preliminaries. Sect 3 presents the proposed method. Sect 4 describes the experiments and discussion. Finally, Sect 5 concludes the paper and outlines future work.

2 Preliminaries

This section presents an overview of the key components of the prediction model architecture, including LSTM, Multihead Attention, and NODE blocks.

2.1 LSTM

The LSTM network [26] is a recurrent architecture designed to mitigate the vanishing gradient problem by introducing gating mechanisms. Its operations at time step t are defined as:

(1)(2)(3)(4)(5)(6)

where is the input vector at time step t. and are the hidden and cell states, respectively. denote the input, forget, and output gates. is the candidate cell state. , , and are trainable weight matrices and bias vectors. is the sigmoid activation, is the hyperbolic tangent, and denotes element-wise multiplication.

2.2 Multihead attention

Multihead Attention [47] is a critical mechanism in Transformer models, enabling the network to attend jointly to information from different subspaces. The basic building block is the scaled dot-product attention:

(7)

where , , and denote the query, key, and value matrices, respectively, and dk is the dimensionality of the keys. In a multihead setting, multiple attention heads are computed as

(8)

where , , and are trainable projection matrices. Finally, the outputs of all heads are concatenated and linearly transformed:

(9)

where is a trainable output projection matrix, and h is the number of attention heads.

2.3 NODE

NODE is a framework that extends deep learning architecture by modeling continuous-time dynamics instead of discrete transformations between layers. In NODE, the evolution of a hidden state h(t) is governed by an ordinary differential equation (ODE):

(10)

where f is a neural network parameterized by θ. The final state is the hidden state at time t, obtained by solving this ODE over a time interval, which provides a flexible and memory-efficient representation.

3 Proposed method

3.1 Overall architecture

Fig 1 illustrates the RUL prediction process for Lithium-ion battery cells. In the data collection phase, Lithium Iron Phosphate (LFP)/graphite cells are monitored to capture voltage, current, and capacity signals for each individual charge-discharge cycle.

thumbnail
Fig 1. The overall RUL prediction process for Lithium-ion battery cells.

https://doi.org/10.1371/journal.pone.0335066.g001

Regarding cycle life degradation, the cycle life of a battery is defined as the total number of charge-discharge cycles from the Beginning of Life (BOL) to the End of Life (EOL). The EOL is typically identified when the battery’s maximum capacity in a charge-discharge cycle degrades to 70% [33] or 80% [20] of its nominal capacity. The RUL, expressed in terms of the remaining number of cycles, is computed as:

(11)

where is the total cycle life of the battery, is the number of aging cycles already completed.

In the signal preprocessing phase, the raw signals—voltage, current, and capacity from the most recent charge-discharge cycles—are passed through a noise-reduction filter named median filter [48] to smooth out sudden peaks. The filtered signals are then processed using feature extraction methods, including mean, standard deviation (Std), minimum (Min), maximum (Max), variance (Var), and median (Med) [49,50]. The extracted features for each cycle are represented as

(12)

where the 3 correspond to the three signal types (voltage, current, and capacity), and the 6 represent the extracted features (Mean, Std, Min, Max, Var, Med). Each input sample to the prediction model consists of 10 selected cycles, uniformly sampled from a 30-cycle window (i.e., one cycle every three cycles) [29]. The input sample is represented as

(13)

Thus, the shape of the total target input data after the feature extraction step becomes

(14)

where N denotes the number of samples. During normalization, a MinMaxScaler is fitted and applied to scale each feature across all time steps and samples between 0 and 1.

In the prediction phase, the prediction model, named HybridoNet-Adapt, maps the target input to the predicted RUL . The details of the proposed RUL prediction model are presented in the following section.

3.2 HybridoNet-Adapt: A proposed RUL prediction model with novel domain adaptation

As shown in Fig 2, HybridoNet-Adapt is composed of four key components: the source predictor , the target predictor , and the feature extractor GF, which is equipped with a DA technique to bridge the distribution gap between the source and target domains.

thumbnail
Fig 2. Architecture of the proposed HybridoNet-Adapt model during training process with domain adaptation.

https://doi.org/10.1371/journal.pone.0335066.g002

The feature extractor integrates a LSTM (Sect 2.1), a Multihead Attention mechanism (Sect 2.2), and a NODE block (Sect 2.3). The NODE block models the hidden state h(t), which evolves continuously over time according to the following ODE: where h(t) denotes the hidden state at time t, f is a trainable function parameterized by , and t represents the continuous time variable. In our implementation, f is a single linear layer to strike a balance between performance and computational efficiency. The initial condition for the NODE block is given by h(t0), and the final transformed state h(t1) is obtained by solving the ODE over the time interval :

(15)

In our experiments, the time bounds are set to t0 = 0 and t1 = 1, based on empirical results (see Fig 9). The function h(t) thus represents the dynamic trajectory of the hidden state under continuous transformation, enabling the model to capture nuanced temporal dependencies.

The target and source predictions in HybridoNet-Adapt are respectively computed as follows:

(16)(17)

where and are learnable trade-off parameters that balance the contributions from the source and target predictors. The outputs and denote the target and source predictions, respectively. The source predictor and the feature extractor GF are trained using both source and target data, enabling the model to transfer domain-invariant features from the source domain to the target domain. This promotes robust prediction performance in the target domain, especially in scenarios where the model would otherwise underperform if trained solely on target data.

The hyperparameters of the proposed HybridoNet-Adapt are summarized in Table 2. LayerNorm denotes layer normalization [51], and Dropout refers to a dropout layer [52] with a rate of 0.1. FC represents a fully connected layer [53], while in the NODE block, h(t) is parameterized by an FC layer. The learnable trade-off parameters and are initialized at 0.5 and updated during training. ReLU refers to the rectified linear unit activation function [54], Sigmoid refers to the sigmoid activation function [55], and BN denotes batch normalization [56]. The columns Input Dim. and Output Dim. specify the dimensionality of the inputs and outputs of each module, respectively. These values are determined empirically, as discussed in Sect 4.7. Note: The output from the Multihead Attention block is taken from the last step along the time dimension.

To optimize the model, we employ a domain adaptation strategy that combines two loss functions: the mean squared error (MSE) loss, , used for regression targets and ; and the maximum mean discrepancy (MMD) [57] loss, , which encourages alignment between the feature distributions extracted from the source and target domains. The total loss function is defined as:

(18)

where and are the input samples from the source and target domains, respectively, and and are their corresponding RUL labels. The hyperparameter controls the weight of the MMD loss in the overall objective. The MMD loss quantifies the discrepancy between the distributions of source and target feature embeddings. Given extracted feature sets from the source domain and from the target domain , it is defined as:

(19)

where is a kernel function, commonly chosen as the Gaussian kernel: with σ as the kernel bandwidth parameter. n and m denote the number of training samples from source and target domains, respectively. The MSE loss is used to optimize the regression outputs by penalizing the squared differences between predicted and ground truth values. It is defined as:

(20)

where denotes the predicted value, and Yi is the corresponding label for the RUL. u is the number of training samples.

In the following section, a series of experiments are conducted to identify the optimal configuration of HybridoNet-Adapt and to demonstrate its superiority over state-of-the-art methods.

For validating the proposed domain adaptation used in HybridoNet-Adapt, we construct a supervised learning model named HybridoNet, consisting of the target predictor and the feature extractor GF. This model is trained solely on labeled target data using the MSE loss function. By comparing between HybridoNet (without domain adaptation) and HybridoNet-Adapt (with domain adaptation), we highlight the performance improvements achieved through the incorporation of our proposed Domain Adaptation technique.

4 Experiments and discussion

4.1 Experimental setup

Our proposed RUL model is implemented using the PyTorch framework and optimized using the AdamW algorithm [58] to minimize the respective loss functions. All experiments are conducted on an NVIDIA A100 GPU with 80GB of memory. Each experiment is trained for 10 epochs with a batch size of 128 and a fixed learning rate of 0.0005. To reduce variability in the training process, each experiment is repeated 10 times, and the final prediction is computed as the average of these runs. The training data is divided into 90% for training and 10% for validation, with the model selected based on the lowest RMSE on the validation set (see Sect 4.3). The weighting factor λ in Eq 18 is dynamically adjusted during training using the following schedule [59]:

(21)

where denotes the current training epoch, and is the total number of training epochs.

4.2 Datasets

4.2.1 First dataset: Varied fast-charging conditions, with consistent discharging conditions.

The first dataset, referred to as the TRI dataset [16], encompasses a detailed study of 124 LFP/graphite lithium-ion batteries. Each LIB in the dataset has a nominal capacity of 1.1 Ah and a nominal voltage of 3.3 V. The cycle life span of these batteries ranges from 150 to 2,300 cycles, showcasing a wide spectrum of longevity. In terms of operational conditions, all LIBs were subjected to uniform discharge protocols. Specifically, they were discharged at a constant current rate of 4 C until the voltage dropped to 2 V, followed by a constant voltage discharge at 2 V until the current diminished to C/50. The LIBs were charged at rates between 3.6 C and 6 C, under a controlled temperature of 30°C within an environmental chamber. The dataset contains approximately 96,700 cycles, making it one of the largest datasets to consider various fast-charging protocols. The dataset is divided into three distinct parts: a training set with 41 LIBs, a primary test set with 43 LIBs, and a secondary test set comprising 40 LIBs.

4.2.2 Second dataset: Varied discharge conditions, with consistent fast-charging conditions.

The second dataset, referred as the LHP dataset [29], was developed through a battery degradation experiment involving 77 cells (LFP/graphite A123 APR18650M1A) with a nominal capacity of 1.1 Ah and a nominal voltage of 3.3 V. Each of the 77 cells was subjected to a unique multi-stage discharge protocol, while maintaining an identical fast-charging protocol for all cells. The experiment was conducted in two thermostatic chambers at a controlled temperature of 30°C. The dataset encompasses a total of 146,122 discharge cycles, making it one of the largest datasets to consider various discharge protocols. The cells exhibit a cycle life ranging from 1,100 to 2,700 cycles, with an average of 1,898 cycles and a standard deviation of 387 cycles. The discharge capacity as a function of cycle number reveals a wide distribution of cycle lives. The dataset is divided into two distinct parts: a training set with 55 LIBs, and a test set with 22 LIBs.

4.3 Evaluation metrics

To evaluate RUL prediction, we use Root Mean Square Error (RMSE) [60], R-squared (R2) [38,61], and Mean Absolute Percentage Error (MAPE) [62]. These are calculated as follows:

(22)(23)(24)

Where yi and are the observed and predicted RUL, respectively. y is cycle life.The smaller the RMSE and MAPE, and the larger the R2, the better the performance.

4.4 Signal analysis

Fig 3a and 3b analyze battery cycle life. Fig 3a tracks an individual cell’s charge and discharge capacities, marking EOL when the maximum capacity degrades to 80% of nominal capacity. Fig 3b compares cycle life across cells, revealing significant variation in discharge capacity. This variability challenges prediction models for RUL, emphasizing the need for accurate and adaptable RUL predictions for BHM systems.

thumbnail
Fig 3. Comparison of maximum charge and discharge capacities over cycle life for battery cells.

(a) Maximum charge and discharge capacities over charge-discharge cycles for a single battery cell. (b) Maximum discharge capacities over charge-discharge cycles for many battery cells.

https://doi.org/10.1371/journal.pone.0335066.g003

4.5 Signal preprocessing

Before feature extraction step in the signal preprocessing phase (as mentioned in Sect 3), the raw signals exhibit sudden peaks and fluctuations, resembling noise. Smoothing the time-series data can help reduce noise and enhance the key characteristics of the signal. To achieve this, a median filtering method is applied to eliminate abrupt peaks in the signals before feature extraction. As a result, the application of median filtering improves overall model performance. The filtered data leads to better RMSE, R2, and MAPE (%) values compared to the unfiltered data, as illustrated in Fig 4.

thumbnail
Fig 4. Comparison of RUL prediction performance with and without median filtering.

https://doi.org/10.1371/journal.pone.0335066.g004

Fig 5a presents the feature importance ranking derived from XGBoost [63] trained and evaluated on the second dataset. The analysis considers 19 common feature extraction methods: 75th, 90th, 50th, 25th, and 10th percentiles; Maximum; Range; Energy; Mean; Interquartile Range (IQR); Median; Skewness; Standard Deviation (Std); Kurtosis; Root Mean Square (RMS); Minimum; Variance; Zero-Crossing Rate; and Autocorrelation (implemented using NumPy [64] and SciPy [65]). Among these, the 75th, 90th, and 10th percentiles demonstrate the highest contribution. We group the feature extraction methods into five categories: three based on high importance (Groups 1–3), one consisting of fundamental statistics (Group 4), and one hybrid group (Group 5):

  • Group 1: Percentiles 75th, 90th, 10th, Range, IQR.
  • Group 2: Group 1 + Maximum, Percentiles 25th, Skewness, Kurtosis, Standard Deviation (Std).
  • Group 3: Group 2 + Energy, Median, RMS, Mean, Variance, Minimum.
  • Group 4: Mean, Std, Min, Max, Variance, Median.
  • Group 5: Group 4 + Percentiles 75th, 90th, 10th.

Fig 5b shows that although Group 1 and 2 contain features with the highest importance scores, they do not yield strong overall prediction performance. Group 4 (fundamental statistics), which is selected as the feature extraction step during the Signal Preprocessing phase of the proposed framework, combines both high- and low-importance features but achieves the best performance, with an RMSE of 181.45, significantly outperforming others.

thumbnail
Fig 5. Feature contribution analysis: (a) feature importance ranking, and (b) RMSE comparison across feature extraction groups.

https://doi.org/10.1371/journal.pone.0335066.g005

4.6 Feature extractor

The feature extractor is progressively developed, starting with an LSTM architecture and sequentially integrating Multihead Attention (MA) and a NODE block. To evaluate the effectiveness of each component, we assess the performance of HybridoNet-Adapt at different stages. With each addition as shown in Fig 6, the model’s predictive capability improves. Ultimately, HybridoNet-Adapt achieves an RMSE of 166.33, an R2 score of 0.86, and a MAPE of 7.44%, demonstrating its superior performance.

thumbnail
Fig 6. Performance comparison of LSTM-based blocks for feature extractor.

Experiment on the testing data of the second dataset.

https://doi.org/10.1371/journal.pone.0335066.g006

4.7 HybridoNet-Adapt with domain adaption

HybridoNet-Adapt is evaluated with various feature loss functions, including CORAL Loss, Domain Loss [44], MMD, as well as combinations such as MMD with Domain Loss, and MMD with Domain Loss and CORAL Loss, as shown in Fig 7. The results indicate that using only MMD as the feature loss function yields the best performance, achieving an RMSE of 160.05.

thumbnail
Fig 7. Comparison of different feature loss methods.

Experiment on the testing data of the second dataset.

https://doi.org/10.1371/journal.pone.0335066.g007

To determine the optimal hyperparameters, including hidden dimension of all layers, the number of recurrent LSTM layers, and the dropout rate, 27 experiments were conducted. The results are presented in Fig 8. In the graph, L represents the number of recurrent layers, H denotes the hidden dimension size. Based on RMSE score, the best performance is achieved with 2 recurrent LSTM layers, a hidden dimension of 64, and a dropout rate of 0.1.

thumbnail
Fig 8. Comparison of RMSE across different number of LSTM layers, hidden dimensions, and dropout configurations.

Experiment on the testing data of the second dataset.

https://doi.org/10.1371/journal.pone.0335066.g008

To identify the optimal time step in the sequence dimension for both Multihead Attention and NODE outputs, a comprehensive evaluation was conducted. Various NODE output time steps ranging from 2 to 6 were tested, along with different Multihead Attention output time step selections, including the last, the second-to-last, and the mean time step. As shown in Fig 9, the best performance was achieved when using the second-to-last time step of the Multihead Attention output and a NODE output time step of 2.

thumbnail
Fig 9. Comparison of RMSE for different NODE discrete time steps (t) and Multihead Attention output time step selections.

Experiment on the testing data of the second dataset.

https://doi.org/10.1371/journal.pone.0335066.g009

The proposed HybridoNet-Adapt model is systematically evaluated under various scenarios by experimenting with four different target sets, each derived from the training data of the second dataset. The source data is the training data from the first dataset. Below are four target groups of battery cells selected from the training data of the second dataset. These groups are carefully formed to ensure each set represents a diverse range of battery performance. For instance, Group 1 includes both high-cycle cells (e.g., 2-2 with 2,651 cycles) and low-cycle cells (e.g., 1-6 with 1,143 cycles), ensuring a comprehensive representation of aging behaviors.

  • Group 1: 1-3 (1,858 cycles), 1-6 (1,143 cycles), 2-2 (2,651 cycles), 2-6 (1,572 cycles), 3-2 (2,283 cycles), 3-6 (2,491 cycles), 4-3 (1,142 cycles), 5-4 (1,962 cycles)
  • Group 2: 1-5 (1,971 cycles), 1-8 (2,285 cycles), 2-4 (1,499 cycles), 2-7 (2,202 cycles), 3-3 (1,649 cycles), 3-7 (2,479 cycles), 4-4 (1,491 cycles), 5-5 (1,583 cycles)
  • Group 3: 2-8 (1,481 cycles), 3-4 (1,766 cycles), 3-8 (2,342 cycles), 4-1 (2,217 cycles), 4-7 (2,216 cycles), 5-1 (2,507 cycles), 5-6 (2,460 cycles), 6-3 (1,804 cycles)
  • Group 4: 4-8 (1,706 cycles), 5-2 (1,926 cycles), 5-7 (1,448 cycles), 6-4 (1,717 cycles), 6-5 (2,178 cycles), 7-2 (2,030 cycles), 7-7 (1,685 cycles), 8-2 (2,041 cycles)
  • All: All battery cells from the training set of the second dataset.

Table 3 shows that HybridoNet-Adapt outperforms both HybridoNet (without DA) and DANN (with DA) across all groups. It achieves the lowest RMSE and MAPE while maintaining the highest R2, demonstrating better generalization. For instance, in Group 1, HybridoNet-Adapt reduces RMSE from 368.99 to 356.46 and improves R2 from 0.21 to 0.30. On the full dataset, it achieves the best RMSE of 153.24 and R2 of 0.88, significantly outperforming DANN, which shows degraded performance (RMSE = 835.35, R2 = −1.37). DANN struggles with large variations in battery aging, while HybridoNet-Adapt effectively adapts to different distributions, leading to consistently better predictions.

thumbnail
Table 3. Comparison of HybridoNet, DANN, and HybridoNet-Adapt across four target data groups from the testing data of the second dataset.

https://doi.org/10.1371/journal.pone.0335066.t003

Table 4 presents the evaluation metrics for RUL prediction on the test data from the second dataset, comparing Elastic Net, A1, A2 of paper [29] (see Table S4), with our HybridoNet, and HybridoNet-Adapt methods. The results indicate that HybridoNet-Adapt achieves competitive RMSE values, particularly in cases where Elastic Net exhibits high errors. The R2 values show that HybridoNet-Adapt generally improves predictive accuracy compared to the baseline methods. Additionally, MAPE results suggest that HybridoNet-Adapt provides more stable and reliable predictions, especially in challenging scenarios. Overall, these findings demonstrate the potential of HybridoNet-Adapt for enhanced RUL estimation.

thumbnail
Table 4. Evaluation metrics for RUL prediction performance using existing Elastic Net (Ela), A1, A2 results of paper [29] (see Table S4), along with HybridoNet (H), and our proposed HybridoNet-Adapt (H-Adapt).

Experiment on the testing data of the second dataset.

https://doi.org/10.1371/journal.pone.0335066.t004

Fig 10 illustrates the RUL predictions of XGBoost, HybridoNet, HybridoNet-Adapt, and DANN, compared to the true (observed) RUL for Cell 4-5 and Cell 3-1 in the testing set of the second dataset. Among all methods, HybridoNet-Adapt demonstrates the closest alignment with the observed RUL, highlighting its superior predictive accuracy. This improvement is attributed to HybridoNet-Adapt’s ability to align feature representations from the source domain to the target domain, as shown in Fig 11. By effectively increasing the amount of target-relevant data through our domain adaptation technique, HybridoNet-Adapt enhances robustness, making it more adaptable to diverse real-world battery degradation scenarios.

thumbnail
Fig 10. Comparison of model predictions with observed RUL for Cells 4-5 and 3-1 from the testing data of the second dataset.

https://doi.org/10.1371/journal.pone.0335066.g010

thumbnail
Fig 11. PCA-based comparison of embedding features between HybridoNet-Adapt and DANN from Cell 3-1 the testing data of the second dataset.

https://doi.org/10.1371/journal.pone.0335066.g011

4.8 Comparison with state-of-the-art methods

Fig 12a presents a performance comparison of different models on the secondary testing data from the first dataset. The Multi-Time Scale Feature Extraction Hybrid (MSFEH) model [42], XGBoost, and HybridoNet were trained using the training data from the first dataset. HybridoNet-Adapt, in contrast, was trained with the training data of the second dataset as the source input and the training data of the first dataset as the target input. HybridoNet-Adapt achieves the best results, with the lowest RMSE (146.52), demonstrating its superior predictive accuracy through domain adaptation. Moreover, HybridoNet outperforms both XGBoost and MSFEH, highlighting the effectiveness of deep learning–based approaches. The additional improvements achieved by HybridoNet-Adapt further validate the benefits of domain adaptation in enhancing RUL prediction performance. It should be noted that in the MSFEH paper [42], the MAPE formula differs from the one used in our work; therefore, we report only the RMSE comparison with MSFEH.

thumbnail
Fig 12. Comparison of our proposed models with existing state-of-the-art methods.

(a) Results on the secondary testing data of the first dataset. (b) Results on the testing data of the second dataset. (a) The first dataset. (b) The second dataset.

https://doi.org/10.1371/journal.pone.0335066.g012

Fig 12b presents a comparison of our HybridoNet and HybridoNet-Adapt models with state-of-the-art methods, including Elastic Net [29], [29], [29], Ridge Linear [23], Random Forest [23], and Structural Pruning [66], evaluated on the testing data from the second dataset. HybridoNet-Adapt, was trained with the training data of the first dataset as the source input and the training data of the second dataset as the target input, whereas the other methods were trained solely on the training data from the second dataset. The results demonstrate that HybridoNet-Adapt achieves the lowest RMSE (153.24), outperforming all other approaches. This highlights the effectiveness of our proposed method in enhancing predictive performance. Overall, HybridoNet-Adapt consistently outperforms across large datasets with diverse charging and discharging profiles.

In future work, we plan to explore Physics-Informed Neural Networks (PINNs) as a approach to enhance interpretability and physical consistency. By embedding governing equations and domain-specific constraints directly into the learning process, PINNs can reduce reliance on purely data-driven correlations while improving extrapolation to unseen conditions. Such a framework would allow the model to not only achieve strong predictive performance but also provide insights grounded in physical principles, thereby addressing the gap between predictive accuracy and practical interpretability.

5 Conclusion

In this paper, we proposed HybridoNet-Adapt, a novel domain-adaptive framework for accurate Remaining Useful Life (RUL) prediction of lithium-ion batteries. Our approach addresses the challenge of distribution shift between training and testing data by leveraging domain adaptation techniques, specifically Maximum Mean Discrepancy (MMD), to align feature representations between source and target domains. By integrating a hybrid prediction mechanism with trainable trade-off parameters, the model effectively balances contributions from both domain-specific predictors. The proposed architecture combines LSTM, Multi-head Attention, and NODE blocks within a feature extractor, enabling the model to capture both temporal and dynamic characteristics of battery degradation. Extensive experiments on two large-scale benchmark datasets demonstrate that HybridoNet-Adapt consistently outperforms state-of-the-art baselines, such as XGBoost and Elastic Net, as well as state-of-the-art deep learning models like Structural Pruning and Multi-Time Scale Feature Extraction Hybrid (MSFEH) model, affirming its effectiveness in the RUL prediction task. For future work, we plan to enhance model generalization through Physics-Informed Neural Networks, and explore multi-modal data integration to improve scalability and robustness across diverse BHM applications. We will investigate the integration of multi-modal data to enhance scalability and robustness across a wide range of BHM applications.

Preprint availability

A preprint of this manuscript is available at: https://arxiv.org/pdf/2503.21392.

Contact information

For access to the code and further information about this proposed system, please contact AIWARE Limited Company at: https://aiware.website/Contact.

References

  1. 1. Goodenough JB, Park K-S. The Li-ion rechargeable battery: a perspective. J Am Chem Soc. 2013;135(4):1167–76. pmid:23294028
  2. 2. Guo R, Wang F, Akbar Rhamdhani M, Xu Y, Shen W. Managing the surge: a comprehensive review of the entire disposal framework for retired lithium-ion batteries from electric vehicles. Journal of Energy Chemistry. 2024;92:648–80.
  3. 3. Dunn B, Kamath H, Tarascon J-M. Electrical energy storage for the grid: a battery of choices. Science. 2011;334(6058):928–35. pmid:22096188
  4. 4. Zhang J, Fan T, Yuan S, Chang C, Wang K, Song Z, et al. Patent-based technological developments and surfactants application of lithium-ion batteries fire-extinguishing agent. Journal of Energy Chemistry. 2024;88:39–63.
  5. 5. Larcher D, Tarascon J-M. Towards greener and more sustainable batteries for electrical energy storage. Nat Chem. 2015;7(1):19–29. pmid:25515886
  6. 6. Pillot C. The rechargeable battery market and main trends 2018 –2030. In: 36th annual international battery seminar & exhibit. 2019. p. 35.
  7. 7. Karimi G, Li X. Thermal management of lithium-ion batteries for electric vehicles. Int J Energy Res. 2012;37(1):13–24.
  8. 8. Zhang G, Wei X, Chen S, Wei G, Zhu J, Wang X, et al. Research on the impact of high-temperature aging on the thermal safety of lithium-ion batteries. Journal of Energy Chemistry. 2023;87:378–89.
  9. 9. Du X, Meng J, Amirat Y, Gao F, Benbouzid M. Exploring impedance spectrum for lithium-ion batteries diagnosis and prognosis: a comprehensive review. Journal of Energy Chemistry. 2024;95:464–83.
  10. 10. Ma L, Hu C, Cheng F. State of charge and state of energy estimation for lithium-ion batteries based on a long short-term memory neural network. Journal of Energy Storage. 2021;37:102440.
  11. 11. Awadallah MA, Venkatesh B. Accuracy improvement of SOC estimation in lithium-ion batteries. Journal of Energy Storage. 2016;6:95–104.
  12. 12. Hong J, Wang Z, Chen W, Wang L-Y, Qu C. Online joint-prediction of multi-forward-step battery SOC using LSTM neural networks and multiple linear regression for real-world electric vehicles. Journal of Energy Storage. 2020;30:101459.
  13. 13. Luo K, Chen X, Zheng H, Shi Z. A review of deep learning approach to predicting the state of health and state of charge of lithium-ion batteries. Journal of Energy Chemistry. 2022;74:159–73.
  14. 14. Li X, Yu D, Søren Byg V, Daniel Ioan S. The development of machine learning-based remaining useful life prediction for lithium-ion batteries. Journal of Energy Chemistry. 2023;82:103–21.
  15. 15. Ma L, Tian J, Zhang T, Guo Q, Hu C. Accurate and efficient remaining useful life prediction of batteries enabled by physics-informed machine learning. Journal of Energy Chemistry. 2024;91:512–21.
  16. 16. Severson KA, Attia PM, Jin N, Perkins N, Jiang B, Yang Z, et al. Data-driven prediction of battery cycle life before capacity degradation. Nat Energy. 2019;4(5):383–91.
  17. 17. Ibraheem R, Strange C, dos Reis G. Capacity and internal resistance of lithium-ion batteries: full degradation curve prediction from Voltage response at constant Current at discharge. Journal of Power Sources. 2023;556:232477.
  18. 18. Lin D, Zhang Y, Zhao X, Tang Y, Dai Z, Li Z, et al. Early prediction of remaining useful life for grid-scale battery energy storage system. J Energy Eng. 2021;147(6).
  19. 19. Zhao S, Zhang C, Wang Y. Lithium-ion battery capacity and remaining useful life prediction using board learning system and long short-term memory neural network. Journal of Energy Storage. 2022;52:104901.
  20. 20. Ma G, Wang Z, Liu W, Fang J, Zhang Y, Ding H, et al. A two-stage integrated method for early prediction of remaining useful life of lithium-ion batteries. Knowledge-Based Systems. 2023;259:110012.
  21. 21. Wang R, Zhu M, Zhang X, Pham H. Lithium-ion battery remaining useful life prediction using a two-phase degradation model with a dynamic change point. Journal of Energy Storage. 2023;59:106457.
  22. 22. Zhang Y, Feng F, Wang S, Meng J, Xie J, Ling R, et al. Joint nonlinear-drift-driven Wiener process-Markov chain degradation switching model for adaptive online predicting lithium-ion battery remaining useful life. Applied Energy. 2023;341:121043.
  23. 23. Xia J, Shi Q, Li H, Zhou M, Wang W, Wang K, et al. Historical data-independent remaining useful life prediction method based on dual-input deep learning neural network. Journal of Energy Storage. 2023;72:108427.
  24. 24. Kiranyaz S, Avci O, Abdeljaber O, Ince T, Gabbouj M, Inman DJ. 1D convolutional neural networks and applications: a survey. Mechanical Systems and Signal Processing. 2021;151:107398.
  25. 25. Javid AM, Das S, Skoglund M, Chatterjee S. A ReLU dense layer to improve the performance of neural networks. In: ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 2021. p. 2810–4. https://doi.org/10.1109/icassp39728.2021.9414269
  26. 26. Graves A, Graves A. Long short-term memory. Supervised sequence labelling with recurrent neural networks. 2012. p. 37–45.
  27. 27. Jaeger H. Echo state network. Scholarpedia. 2007;2(9):2330.
  28. 28. Liang J, Liu H, Xiao N-C. A hybrid approach based on deep neural network and double exponential model for remaining useful life prediction. Expert Systems with Applications. 2024;249:123563.
  29. 29. Ma G, Xu S, Jiang B, Cheng C, Yang X, Shen Y, et al. Real-time personalized health status prediction of lithium-ion batteries using deep transfer learning. Energy Environ Sci. 2022;15(10):4083–94.
  30. 30. Zhu J, Wang Y, Huang Y, Bhushan Gopaluni R, Cao Y, Heere M, et al. Data-driven capacity estimation of commercial lithium-ion batteries from voltage relaxation. Nat Commun. 2022;13(1):2261. pmid:35477711
  31. 31. Wang F, Zhao Z, Ren J, Zhai Z, Wang S, Chen X. A transferable lithium-ion battery remaining useful life prediction method from cycle-consistency of degradation trend. Journal of Power Sources. 2022;521:230975.
  32. 32. Attia PM, Grover A, Jin N, Severson KA, Markov TM, Liao Y-H, et al. Closed-loop optimization of fast-charging protocols for batteries with machine learning. Nature. 2020;578(7795):397–402. pmid:32076218
  33. 33. Qu J, Liu F, Ma Y, Fan J. A neural-network-based method for RUL prediction and SOH monitoring of lithium-ion battery. IEEE Access. 2019;7:87178–91.
  34. 34. Zhou Y, Huang M. Lithium-ion batteries remaining useful life prediction based on a mixture of empirical mode decomposition and ARIMA model. Microelectronics Reliability. 2016;65:265–73.
  35. 35. Saha B, Goebel K. Battery data set. NASA AMES prognostics data repository. 2007.
  36. 36. Liu Y, Hou B, Ahmed M, Mao Z, Feng J, Chen Z. A hybrid deep learning approach for remaining useful life prediction of lithium-ion batteries based on discharging fragments. Applied Energy. 2024;358:122555.
  37. 37. Vasan ASS, Mahadeo DM, Doraiswami R, Huang Y, Pecht M. Frontiers in bioscience. 2013;5:39–71.
  38. 38. Wang S, Ma H, Zhang Y, Li S, He W. Remaining useful life prediction method of lithium-ion batteries is based on variational modal decomposition and deep learning integrated approach. Energy. 2023;282:128984.
  39. 39. Bole B, S. Kulkarni C, Daigle M. Adaptation of an electrochemistry-based li-ion battery model to account for deterioration observed under randomized use. PHM_CONF. 2014;6(1).
  40. 40. Ansari S, Ayob A, Lipu MSH, Hussain A, Abdolrasol MGM, Zainuri MAAM, et al. Optimized data-driven approach for remaining useful life prediction of Lithium-ion batteries based on sliding window and systematic sampling. Journal of Energy Storage. 2023;73:109198.
  41. 41. NASA. Li-ion Battery Aging Datasets. https://data.nasa.gov/dataset/Li-ion-Battery-Aging-Datasets/uj5r-zjdb/about_data
  42. 42. Yu Q, Wang F, Zhai Z, Zheng S, Liu B, Zhao Z, et al. Multi-time scale feature extraction for early prediction of battery RUL and knee point using a hybrid deep learning approach. Journal of Energy Storage. 2025;117:116024.
  43. 43. Sanz-Gorrachategui I, Pastor-Flores P, Pajovic M, Wang Y, Orlik PV, Bernal-Ruiz C, et al. Remaining useful life estimation for LFP cells in second-life applications. IEEE Trans Instrum Meas. 2021;70:1–10.
  44. 44. Ganin Y, Ustinova E, Ajakan H, Germain P, Larochelle H, Laviolette F. Domain-adversarial training of neural networks. Journal of Machine Learning Research. 2016;17(59):1–35.
  45. 45. Ye Z, Yu J. State-of-health estimation for lithium-ion batteries using domain adversarial transfer learning. IEEE Trans Power Electron. 2022;37(3):3528–43.
  46. 46. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S. Generative adversarial nets. Advances in Neural Information Processing Systems. 2014;27.
  47. 47. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. Advances in neural information processing systems. 2017;30.
  48. 48. Justusson B. Median filtering: statistical properties. Two-dimensional digital signal processing II: transforms and median filters. 2006. p. 161–96.
  49. 49. Elangovan M, Devasenapati SB, Sakthivel NR, Ramachandran KI. Evaluation of expert system for condition monitoring of a single point cutting tool using principle component analysis and decision tree algorithm. Expert Systems with Applications. 2011;38(4):4450–9.
  50. 50. Sugumaran V, Ramachandran KI. Effect of number of features on classification of roller bearing faults using SVM and PSVM. Expert Systems with Applications. 2011;38(4):4088–96.
  51. 51. Ba JL, Kiros JR, Hinton GE. Layer normalization. arXiv preprint 2016.
  52. 52. Baldi P, Sadowski PJ. Understanding dropout. Advances in neural information processing systems. 2013;26.
  53. 53. Ma W, Lu J. An equivalence of fully connected layer and convolutional layer. arXiv preprint arXiv:171201252. 2017.
  54. 54. Agarap AF. Deep learning using rectified linear units (relu). arXiv preprint 2018.
  55. 55. Yin X, Goudriaan J, Lantinga EA, Vos J, Spiertz HJ. A flexible sigmoid function of determinate growth. Ann Bot. 2003;91(3):361–71. pmid:12547689
  56. 56. Bjorck N, Gomes CP, Selman B, Weinberger KQ. Understanding batch normalization. Advances in neural information processing systems. 2018;31.
  57. 57. Borgwardt KM, Gretton A, Rasch MJ, Kriegel H-P, Schölkopf B, Smola AJ. Integrating structured biological data by Kernel Maximum Mean Discrepancy. Bioinformatics. 2006;22(14):e49-57. pmid:16873512
  58. 58. Loshchilov I, Hutter F. Decoupled weight decay regularization. arXiv preprint 2017. https://arxiv.org/abs/1711.05101
  59. 59. Ganin Y, Lempitsky V. In: International conference on machine learning, 2015. 1180–9.
  60. 60. Willmott C, Matsuura K. Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Clim Res. 2005;30:79–82.
  61. 61. Colin Cameron A, Windmeijer FAG. An R-squared measure of goodness of fit for some common nonlinear regression models. Journal of Econometrics. 1997;77(2):329–42.
  62. 62. de Myttenaere A, Golden B, Le Grand B, Rossi F. Mean absolute percentage error for regression models. Neurocomputing. 2016;192:38–48.
  63. 63. Chen T, Guestrin C. Xgboost: a scalable tree boosting system. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining; 2016. p. 785–94.
  64. 64. Harris CR, Millman KJ, van der Walt SJ, Gommers R, Virtanen P, Cournapeau D, et al. Array programming with NumPy. Nature. 2020;585(7825):357–62. pmid:32939066
  65. 65. Virtanen P, Gommers R, Oliphant TE, Haberland M, Reddy T, Cournapeau D, et al. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nat Methods. 2020;17(3):261–72. pmid:32015543
  66. 66. Ge Y, Ma J, Sun G. A structural pruning method for lithium-ion batteries remaining useful life prediction model with multi-head attention mechanism. Journal of Energy Storage. 2024;86:111396.