Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Using the TSA-LSTM two-stage model to predict cancer incidence and mortality

Abstract

Cancer, the second-leading cause of mortality, kills 16% of people worldwide. Unhealthy lifestyles, smoking, alcohol abuse, obesity, and a lack of exercise have been linked to cancer incidence and mortality. However, it is hard. Cancer and lifestyle correlation analysis and cancer incidence and mortality prediction in the next several years are used to guide people’s healthy lives and target medical financial resources. Two key research areas of this paper are Data preprocessing and sample expansion design Using experimental analysis and comparison, this study chooses the best cubic spline interpolation technology on the original data from 32 entry points to 420 entry points and converts annual data into monthly data to solve the problem of insufficient correlation analysis and prediction. Factor analysis is possible because data sources indicate changing factors. TSA-LSTM Two-stage attention design a popular tool with advanced visualization functions, Tableau, simplifies this paper’s study. Tableau’s testing findings indicate it cannot analyze and predict this paper’s time series data. LSTM is utilized by the TSA-LSTM optimization model. By commencing with input feature attention, this model attention technique guarantees that the model encoder converges to a subset of input sequence features during the prediction of output sequence features. As a result, the model’s natural learning trend and prediction quality are enhanced. The second step, time performance attention, maintains We can choose network features and improve forecasts based on real-time performance. Validating the data source with factor correlation analysis and trend prediction using the TSA-LSTM model Most cancers have overlapping risk factors, and excessive drinking, lack of exercise, and obesity can cause breast, colorectal, and colon cancer. A poor lifestyle directly promotes lung, laryngeal, and oral cancers, according to visual tests. Cancer incidence is expected to climb 18–21% between 2020 and 2025, according to 2021. Long-term projection accuracy is 98.96 percent, and smoking and obesity may be the main cancer causes.

1. Introduction

Cancer is a group of disorders in which abnormal cells divide and spread to surrounding tissues, according to the National Cancer Institute (NCI). Thus, cancer can travel throughout the body via the blood. Globally, cancer kills second only to cardiovascular disease. Cancer killed the sixth person in the globe, second only to cardiovascular disease. Nearly 18.9 million new cancer infections from 2018 caused 9 million deaths [1]. Due to life expectancy increases, an aging population, population growth, and the lack of efficient health programs like large-scale screening, cancer incidence is likely to climb. Globalization, lifestyle changes, expansion, and the eventual control of infectious diseases boost cancer rates [2]. Due to population age structure changes, elderly persons are more likely to have cancer and other non-communicable diseases. This study’s TSA-LSTM model attention technique ensures that the model encoder obtains an agreement on a subset of input sequence characteristics before predicting output sequence features [3]. This enhances the quality of the model’s predictions and its natural learning trend. For enhanced sequence modeling, the Temporal Self-Attention Long Short-Term Memory (TSA-LSTM) model is an innovative deep-learning architecture that blends temporal self-attention processes with LSTM cells [4]. In cancer incidence and mortality prediction, the TSA-LSTM model has demonstrated substantial improvements over existing models with regard to precision, dependability, and computing economy [5]. The TSA-LSTM model outperforms conventional models when it comes to accurately capturing intricate temporal patterns and dependencies in the data, resulting in more accurate predictions [6]. This ability to handle long-term connections predicts cancer incidence and mortality. Research on cancer spans decades. Researchers screen for pre-symptomatic cancer. We also developed early cancer therapy outcome prediction tools [7]. The medical research community can access a lot of cancer data when new information emerges [8]. However, surgeons struggle most with illness outcome prediction. Traditional statistics and computer algorithms are used in data mining to find new vital knowledge from enormous amounts of data. It is a healthy computing method. By illustrating cancer incidence and death disparities by area, race, gender, and socioeconomic characteristics, it estimates population health needs and cancer burden. Data mining can reveal hidden links between patient files, cancer therapy, and monitoring [9,10]. Therefore, extracting medical data is a desirable task. Due to this, data mining and healthcare services have developed reliable early detection systems and other healthcare systems based on cancer-related clinical and diagnostic data to anticipate cancer stages and help clinicians make clinical decisions [11].

To avoid inaccurate cancer incidence estimates, forecasting can predict future peak occurrences. A tripartite prediction is timely, clear, and better than an advanced forecast. Medical researchers employed many prediction methodologies to address the health sector’s predictive issues. Increased forecasting knowledge has allowed for huge accuracy improvements, yet most physicians don’t use it [12]. Difficulties in changing forecasting tools, personnel not knowing the modern procedure, and stakeholders not using forecasted data in decisions may be to blame [13,14]. Studies have divided forecasting methods into three categories. Existing studies were classified as artificial intelligence, statistical, or hybrid. In statistics, algorithms like ARMA, regression, ARIMA, exponential smoothing (ES), and Holt-Winters are used (HW). Smoothing methods are widely used in time series analysis and forecasting [15]. Unlike regression, statistical models can forecast univariate time series with a single variable of interest (dependent variable) available for different periods. Recently popular statistical models include exponential smoothing (ES). Smoothing exponentially is weighted moving averages. SES is the simplest exponential smoothing method [16]. No seasonality in SES projections. The Holt-Winters smoothing includes SES seasonality or trend. The popular data visualization and forecasting tool Tableau uses exponential smoothing. Accounting, libraries, and others utilize Tableau. Tableau lets users drag-and-drop charts, maps, dashboards, and stories to visualize data [17]. Deep learning/machine learning algorithm forecasting methods like Artificial neural networks (ANN) and Recurrent neural networks (RNN) have garnered attention in recent years. Recent studies use deep learning with intelligence and flexibility to improve prediction models [18]. Deep learning approaches match model parameters from observed data and are best for discovering behavioral patterns in random data series. This makes deep learning systems ideal for anomaly detection. ANN models have performed well for real-time estimations in deep learning [19]. Anticipating future cancer incidence and mortality trends is challenging due to the complex interplay of risk factor profiles and the extended latency period that might elapse between exposure to certain factors and the development of certain malignancies [20]. To establish whether the exponential model or the machine learning method performs better on cancer data, we evaluate the current state of artificial neural networks. Existing studies use several ANN models for predictions: gated recurring units. sequence-to-sequence Gateway Recurrent Unit. large short-term memory [21]. First, we test if the machine learning approach outperforms exponential smoothing on cancer data, and then we design a novel forecasting algorithm [22,23]. We upgrade previous models with a Two-Stage attention LSTM [24,25]. While maintaining deep learning’s flexibility and versatility, attention models may give interpretability. Regarding dependability, the TSA-LSTM model’s attention mechanisms direct attention to relevant temporal information, reducing the impact of unnecessary or noisy data points on predictions. This improves forecast reliability, which policymakers and doctors can use to make decisions [25]. The efficiency of training on large datasets has been optimized with respect to the computational efficiency of the TSA-LSTM model’s architecture [26]. The model may efficiently acquire knowledge from both immediate and enduring dependencies in the data without necessitating an overabundance of computational resources by utilizing self-attention methods [3,27]. For healthcare applications, where precise and timely predictions are critical, this characteristic renders it exceptionally suitable [5]. The TSA-LSTM model excels in accuracy, reliability, and computing efficiency when compared to other cancer prediction models. The advanced design and functionality of this tool make it ideal for oncology researchers and practitioners seeking to enhance decision-making.

Public health has been substantially impacted by lifestyle patterns. Lifestyle decisions impact health and death, explain Tuffour et al. Smoking, inactivity, alcohol, and obesity-causing diets are recognized cardiovascular disease and cancer risk factors. The second leading cause of mortality in the US, cancer has long been a public health issue [19,28]. Therefore, precise predictions of future cancer incidence and fatalities are crucial because they affect not only the population but also the economy due to the huge sums of money lost fighting it. Current research divides forecasting approaches into three categories: artificial intelligence, statistical, and hybrid. Recent interest in statistical models has focused on exponential smoothing (ES) and double exponential smoothing (DES). ES, DES are weighted moving averages [29]. Tableau, a popular data visualization and forecasting application, uses exponential smoothing. These approaches can reliably anticipate short-term data. Second, deep learning/machine learning algorithm forecasting techniques use intelligence and flexibility to improve prediction models. Artificial Neural Network (ANN) models are ideal for real-time estimations in deep learning [30,31]. The third algorithm combines the two [32]. Different algorithms have been utilized in the health sector based on study goals, however, the optimum algorithm for cancer data accuracy has not been studied. The goal of this study is to find the most accurate cancer data prediction program[33]. An efficient and effective deep learning system for cancer data forecasting is proposed and implemented in the study.

The study’s contributions and research needs are predicated on two indistinguishable elements: first, the performance of a DES algorithm, and second, an analysis of the existing deep learning algorithm’s impact on cancer prediction [34,35]. In this regard, the results of the study indicate that as time passes, the DES model deviates from the cancer data. The lack of seasonality in the DES is responsible for this. This suggests that while the DES method may be advantageous for short-term forecasts (1–6 months), it is not suitable for long-term forecasting [36]. This study analyzes eleven algorithms for cancer incidence prediction accuracy. Double Exponential Smoothing (DES) ranks tenth. Thus, DES cannot predict long-term cancer and produces imprecise data. We conclude, therefore, that the DES method is unsuitable for long-term cancer projections in order to facilitate effective policy formulation [37,38]. Second, among the eleven algorithms utilized in our research, the method is deemed most suitable for cancer prediction [32]. Compared to previous models, our suggested model has the highest and most precise accuracy at 97.76 percent. The attention sequence utilized by our approach to sequence the LSTM model consisted of two stages. Attention models maintain the adaptability and versatility of deep learning approaches while giving the potential for interpretability [39]. We use the prior encoder concealed state to develop an input attention approach in the early stage to adaptively extract relevant driving series at each time step. A temporal attention approach selects relevant encoder hidden states over all time steps in the second stage [40,41]. Our model demonstrates effective prediction capabilities and is also readily interpretable due to the implementation of this dual-stage attention technique. The model we have put forward possesses the capability to not only forecast time series but also function as a versatile feature learning instrument in the domain of computer vision [42]. Our cancer data forecasting model outperforms others. Our proposed strategy may be accurate due to the model’s attentions, which are interpretable and adjustable.

Furthermore, the research distinguishes itself by establishing a methodology for predicting cancer incidence by 2025. This is achieved through the implementation of an innovative two-stage attention model, which utilizes 2017 as its reference year to forecast the rate of increase in cancer incidence. The anticipated rise in cancer rates can be attributed, in part, to an increase in unhealthy lifestyle choices, including alcohol and tobacco use, physical inactivity, and poor diet, which contribute to the prevalence of obesity in the United States. In order to mitigate the escalating incidence of cancer in the United States, we propose the following measures: first augmenting public safety education regarding the detrimental consequences of adopting an unhealthy lifestyle, both inside households and in the workplace. Second, it is recommended that government money be directed towards research endeavors that aim to eliminate or reduce the deleterious consequences of lifestyle choices and cancer in its entirety. Third, similar strategies for deterring such detrimental lifestyle decisions should be applied provincially. For instance, reforms ought to target the imposition of a daily maximum on the quantity of tobacco or alcohol that may be sold to an individual.

The main goal of this paper is to expand the application of visual analysis tools in the field of health care, introduce artificial neural network modeling technology, improve the accuracy of data analysis and pattern recognition ability, and remind people to give up bad life habits and develop good life habits with quantitative data such as correlation analysis between bad lifestyle and cancer occurrence, analysis of future cancer development trend. This data helps the government create medical reform and economic allocation initiatives. Searching big data and healthcare data analytics for the latest studies starts the literature review. This study chose R for data preprocessing and Tableau for analysis of healthcare big data utilizing a theoretical and empirical approach. This work uses theoretical knowledge from Tableau and R white papers, books, articles, and conference proceedings on information visualization, cancer analysis, double exponential smoothing modeling, artificial neural networks, long short-term memory models, and deep learning algorithms. Tableau software, which requires proficiency and deep learning models, provides the empirical basis for the idea. This research predicts high-risk cancer data using a revolutionary deep-learning Two-Way Attention algorithm. Model of LSTM. This study uses long short-term memory and recurrent neural network deep learning architectures.

Six sections make up this manuscript. The first section provides the study’s background, rationale, research objectives and questions, problem statement, significance, and innovation. In Section 2, an overview of the pertinent literature is presented. Big data and healthcare data analytics are explained in Section 3. It discusses data visualization with Tableau. The exponential smoothing model and interpolation method in data expansion are examined again in the chapter. Next, artificial neural networks and long short-term memory are discussed. Our study used raw data cleaning and sample expansion (section 4). Discussed are CDC and American Cancer Society datasets, how variables are constructed, R cleaning and analysis, and Tableau visualization. This section discusses data manipulation for visualization step-by-step. Finally, our data interpolation method for maximum results is verified and discussed. Our proposed LSTM model and the Double Exponential Smoothing algorithm’s data processing and analysis modeling are discussed in Section 5. This section discusses TSA-LSTM implementation again. Study experiments, results, and analysis are covered in this chapter. Interpretations from visualizations provide insight. As the study concludes, Conclusions, policy implications, and suggestions for further research are summarized in Section 6.

2. Literature review

The literature review focuses on the theoretical basis of the advent of big data analytics and healthcare, tableau as a data visualization tool, exponential smoothing model, the common data interpolation methods used in data processing, introduction To ANN’s, and two ANNs: convolutional neural network (CNN) and recurrent neural network (RNN).

2.1 Big data analytics and healthcare

Data collected and stored digitally is growing rapidly. Data storage and analysis are in high demand due to the internet and the digital economy’s rapid growth. Thus, data management and analysis are improving to help firms turn this massive supply of data into information and knowledge that helps them succeed [4345]. Big data comes from massive datasets that typical database management solutions cannot handle. The majority of software and storage systems cannot quickly store, manage, and process massive data [46]. Organizations and health institutions keep and collect more data than ever since businesses and lives depend on it [1]. Accessing petabytes and zettabytes of data is essential. It shows how to avoid losing important data, ergo big data analytics in healthcare. Big data analytics finds invisible patterns and connections in data that current approaches cannot manage [30,45]. Health institutions are slow to adopt and study big data analytics, despite its value [47,48]. Data analytics can process and analyze patient data to improve diagnosis and treatment outcomes, relieving the overburdened healthcare system [4951]. Data on new studies, emergencies, and disease outbreaks shows evolving trends. As electronic medical record systems progress, healthcare systems and personnel improve their ability to document patient experience and visit details [52]. Monitoring public health involves data analysis. This option is meaningless without connecting, trending, and patterning data. Outside of data analysis, healthcare record standards and the human condition may affect patients’ symptoms and indications [53]. Costs, chronic sickness, and aging are forcing healthcare industry changes. Healthcare data analytics may treat old health conditions, and episodic approaches are unhelpful [54,55]. Healthcare payers and providers use data and analytics to improve service and lower costs [56]. Health researchers need useful and intelligent tools to aid in their work [57,58]. Systemic quality and cost improvements in health reforms require data analytics. Due to the widespread use of electronic medical records and health outcome data, academics can study health analytic methods to find and evaluate effective interventions and anticipate patient health outcomes and costs [46].

2.2 Tableau as a data visualization tool

Tableau provides data insights through dynamic, interactive reporting and visualization [59]. Tableau exposes latent correlations within datasets. Graphs are employed to underscore relationship principles and facilitate code-free data exploration. Tableau efficiently extracts meaningful insights from data [60]. We used double exponential smoothing for our predictions in this investigation. Analytics software streamlines the procedure, provides visual feedback, and responds to complex questions. This makes data more important [61]. Tableau Desktop supports Amazon Redshift, MySQL, Oracle, Google Analytics, Microsoft SQL, OLAP databases [62]. Tableau supports hundreds of texts, PDF, Excel, geographic, statistical, and database data connections. Tableau supports numerous data types, making it convenient. Tableau’s data-modified images let customers quickly spot visual trends [63,64]. In cancer research, RNNs and LSTMs operate as powerful microscopes, peering into intricate datasets to reveal previously unknown associations between risk factors [65]. Such models are masters of analysis, but Tableau provides the magnifying glass by simplifying their complex outputs into comprehensible visuals. Bridge the gap between AI and human comprehension with Tableau [66], which cleans and prepares data for the models before showing their predictions [67]. Together, they bring researchers a clearer view of cancer risk, which in turn leads to more effective preventative efforts and better dissemination of these vital results [68].

2.3 Common data interpolation methods

Linear spline interpolation provides straight lines between data points in table gaps, usually astronomical data. It is criticized for its rapid trend change from one segment to the next, which does not occur naturally [69]. However, it is a good way to generate accurate lookup tables with quick look-ups for smooth functions without many table entries [70]. This method is a valuable tool used across fields to approximate values between known data points. Linear interpolation, which assumes a straight line between two neighboring points, is common [71]. Interpolating curves with polynomials and data points. Spline interpolation fits a polynomial while segmenting the dataset for smoother results. The nearest neighbor interpolation replaces missing or interpolated data points with their closest known values [72]. Inverse distance weighting uses the weighted average of neighboring sites to compute interpolated values, unlike kriging interpolation, which accounts for spatial correlations in geostatistical data [73]. The linear spline is mathematically; Pi(x) = aix+bi with two parameters ai and bi can only satisfy the following two equations required S(x) to be continuous:

Linear spline: Pi(x) = aix+bi with two parameters ai and bi can only satisfy the following two equations required S(x) to be continuous: (Eq 1) or, equivalently, Pi(x) has to pass the two endpoints (xi−1,yi−1) and (xi,yi) (Eq 2)

Solving either of the two problems above, we get: which is represented in the form of Pi = aix+bi the first expression, or a linear interpolation of the two endpoints yi−1 = f(xi−1) and yi = f(xi) in the second expression. As Pi(xi) = Pi+1(xi) = yi, the linear spline S(x) is continuous at xi. But as in general (Eq 3) S(x) is not smooth, i.e., k = 0.

Quadratic spline: with three parameters ai, bi and ci can satisfy the following three equations required for S(x) to be smooth (k = 1) as well as continuous: (Eq 4)

To obtain the three parameters ai, bi and ci in Qi(x), we consider Qi(x) = 2aix+bi, which, as a linear function, can be linearly fit by the two end points f′(xi−1) = Di−1 and f′(xi) = Di: (Eq 5)

Integrating, we get As Qi(xi) = yi, we have, and then solving this for ci and substituting it back into the expression of Qi(x), we get, as Qi(xi−1) = yi−1 we have (Eq 6)

Given f′(x0) = D0, we can get iteratively all subsequent D1,…,Dn, and thereby Qi(x) Alternatively, given f′(xn) = Dn, we can also get iteratively all previous Dn−1,⋯,D0. It is obvious that with only three free parameters, the quadratic polynomials cannot satisfy both boundary conditions and . The third area cubic splines excel in is smooth interpolation modeling. Cubic spline interpolation uses a continuous formula to compute the first and second derivatives inside intervals and at interpolating nodes. A smoother interpolating function will result. First derivative continuity implies graph y = Z(x) will not have sharp corners. The continuity of the second derivative means that the radius of the curve is defined at each point. Given the n data points (x1, y1),…,(xn, yn) where x1 are distinct and in increasing order. A cubic spline Z(x) through the data points (x1, y1),…,(xn, yn) is a set of cubic polynomials: Given the n data points (x1, y1),…,(xn, yn) where x1 are distinct and in increasing order. A cubic spline Z(x) through the data points (x1, y1),…,(xn, yn) is a set of cubic polynomials: (Eq 7) (Eq 8) (Eq 9) With the following conditions (known as properties) (Eq 10)

This property guarantees that the spline S(x) interpolates the data points.

(Eq 11)

Z′(x) is continuous on the interval [x1, xn]; this property forces the slopes of the neighboring parts to agree when they meet.

(Eq 12)

Z′′(x) is continuous on the interval [x1, xn], which also forces the neighboring spline to have the same curvature, to guarantee smoothness.

2.4 Artificial neural networks ANN

Imitate cognitive and theoretical knowledge of how the human brain learns, including neural interconnections in the network. ANN is a fully linked multi-neuron network having an input layer that communicates with the following layer [7476]. Artificial Neural Networks (ANN) try to mimic the human brain. Humans may be slower and less precise than neural networks [77]. An artificial neural network (ANN) is a fully linked multi-neural network that consists of an input layer, a layer that communicates with the layer below it (hidden layers), a layer that processes additional weighted connections and links to the output layer, and finally, an output layer of nodes that provide a variable for output [78,79]. Neural network nodes output real numbers from inputs. Nodes’ outputs are calculated by applying a function (ψ) to the previous layer nodes’ outputs [80]. Before that, the input layer output (o(0)) equals the input. We calculate the output layer output by computing layer outputs sequentially. This is an inferential procedure. Neural networks are structured according to the number of layers and the functions that govern the outputs of each layer. (Eq 13) Where o{k) = the output vector representing the outputs of nodes in the layer lk, = the general output of the node lk, = function to determine the outputs of nodes and xn ≕ nth input data. In order for two consecutive layers to be fully connected, all nodes in the previous layer must be connected to the next. Fully linked neural network topology is shown in Fig 1.

thumbnail
Fig 1. The graph representation of a fully connected neural network.

https://doi.org/10.1371/journal.pone.0317148.g001

Neural network predictions are closer to actual standards since the model may be changed iteratively to reduce inaccuracies. Handwriting recognition, medicine, forecasting, and many other fields use ANNs because they outperform conventional technologies [81,82]. Convolutional neural networks (CNNs) are derived from visual cortical neurobiology. It interprets visuals as RGB layers of color and perceives them as three-dimensional entities. It then applies filters on convolutions of inputs in order to extract useful features [83]. In recent times, CNN’s achievements in computer vision tasks including as object identification, real-time face recognition, and image categorization and recognition have garnered significant attention [20]. Recurrent Neural Networks (RNN) are feed-forward networks with feedback edges that provide context [84]. ANNs’ hidden layers have duplicate cells whose states are modified by past and current feedback input. The ANN’s t-1 decisions affect its t-time judgment [50,85]. Cell-internal connection differences determine network design capability. Backpropagate over time with RNN memory, which stores all concealed states. Our networks learned short-range dependencies but not intermediate-range ones. Other scientists have addressed this restriction by allowing RNN to acquire mid-term dependencies Recurrent neural networks may feature a bidirectional, gated, or long-short-term memory architecture (LSTM). The challenge arises when the RNN must store in memory long-term dependencies. Long short-term memory networks solved this problem (LSTM) [86]. The LSTM uses truncation to efficiently flow a range of constant faults in special unit internal states.

2.5 Deep learning approach

Deep learning algorithms have outperformed computer vision in recent years, but their time series models are still shallow. Data scientists in most sectors utilize autoregressive models instead of deep learning. Common reasons for using these methods include interpretability, constrained data, convenience of use, and training cost [87]. It typically improves both performance (compared to other vanilla LSTMs/RNNs) and interpretability [88]. Their speed surpasses that of RNN/LSTM in a few of cases [89]. In addition, attention networks include categorical inference in deep neural networks [90]. These brain networks control consciousness, sensory orientation, and competing reactions. Many sequence transduction models use complex recurrent or convolutional neural networks in an encoder-decoder arrangement, but they only work locally and lack global information [91]. This constraint prompted attention networks to record long-range interactions. Attention networks are popular due to their parallel computation and relationship modeling adaptability. A simple residual network model that replaced all spatial convolutions with self-attention showed attention networks’ impact. This fully self-attentional model surpassed the baseline with less categorization parameters [92]. Additionally, subsequent tests on attention in general deep neural networks show continuous progress in categorization and detection with varied models. This influenced this study’s use of attention networks [93].

2.6 RNN & LSTM

Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) models exhibit remarkable capabilities in processing sequential data and uncover complex relationships that are concealed within time series. Long-sufficiency-term (LSTM) models, classified as a subtype of RNNs, exhibit exceptional efficacy in managing data dependencies that endure for an extended duration. Furthermore, the greatest distinction between RNN and other networks is that it shows the relationship between current and historical information, although gradient dispersion limits its ability to learn relevant content quickly. LSTM (Long short-term memory) was introduced in 1997 to address the vanishing gradient issue [94]. Unlike the standard RNN, the LSTM has the same repeated modules, but its underlying structure is three "Gates" instead of a single neural network layer, making it better for processing and predicting critical events with lengthy intervals and delays in time series [95]. RNNs are crucial to deep learning algorithms that operate on time-series data. Remembering historical data allows these neural networks to predict current events. [96]. Due to fading and expanding gradients, conventional RNNs struggle to manage long-term dependencies in extended time series. After developing the LSTM network, critical state information was extracted from historical time-series data [97]. This event reinforces the link between past and future data. LSTM has been effective in welding quality monitoring and penetration prediction [98].

Another major perk is the TSA-LSTM model’s ability to process massive amounts of time-series data quickly without sacrificing forecast accuracy [99]. Cancer researchers and practitioners can benefit greatly from its sophisticated design and capacity to simulate intricate temporal patterns, which enhances their predicting abilities and allows them to deliver superior patient treatment [100]. However, unlike RNN, LSTM allows just one memory superposition formula. Show two network diagrams. Data analysis and modeling are done using the LSTM neural network for long-term time series data [101]. Long Short-Term Memory RNNs are efficient in deep learning. LSTM temporal model emphasizes keyhole progression during welding. The LSTM network learns temporal feature information from image sequences, demonstrating keyhole dynamics [102]. LSTMs use memory cells instead of artificial neurons in the hidden layer, unlike feedforward neural networks [103105]. Also, systems can rapidly correlate memories and input remotely with memory cells to understand and predict data structure over time. Internal indirect state units are LSTM long-term or short-term memory. A gate to the conventional recurrent cell improves retention for “long-term dependencies” in the LSTM cell. Researchers used this to produce LSTM topologies with and without forget gates [106,107]. In the LSTM network, indirect state units function as either short-term or long-term memory. Unit status regulates the generation of an LSTM network. This is not only the last entry; it is a fundamental quality when a network forecast must be influenced by input history. It processes video, audio, and image data [108].

In the end, the TSA-LSTM is an advanced deep-learning architecture used in predictive analytics to predict cancer incidence rates. This approach exploits LSTM networks’ ability to capture long-term associations in sequential data. The model utilizes LSTM networks, known for their proficiency in processing sequential data such as cancer development, enabling the model to grasp the evolving characteristics of the disease. Furthermore, the two-stage architecture has the potential to learn intricate correlations among various elements in the data. The initial phase extracts fundamental patterns, which are then utilized in the subsequent phase for prediction or classification purposes. At various levels of prediction, attention processes focus on key traits [109,110]. This study used TSLM to predict cancer, which has two stages: Feature extraction first. Use LSTM layers to uncover temporal correlations and sequential patterns from cancer incidence data [111]. After two steps, the model accurately forecasts cancer incidence rates by capturing their complex dynamics. Previous research and articles on TSA-LSTM models for cancer incidence prediction may inform technical implementation [112,113]. The following section describes this model’s cancer prediction benefits. We provide historical cancer incidence data, which contains demographics, environmental variables, cancer diagnosis rate time series, and other essential information, unlike other studies. Normalize, fix missing values, and create LSTM-friendly data windows [114]. The TSA-LSTM model captures sequential data long-term dependencies with several LSTM cell layers. LSTM prediction uses attention to highlight relevant traits. Identify and refine important qualities for successful forecasting in two steps. Compare predicted values against real-world findings to assess the model’s accuracy [115]. Assess model estimates using epidemiology, historical patterns, and cancer risk indicators. Attention processes and LSTM networks in a dual-stage framework assess cancer incidence and capture its complex temporal patterns in TSA-LSTM [116]. Through enhanced deep learning, researchers and data scientists can improve cancer risk assessment, resource distribution, and public health policy [117]. This section has comprehensively examined the emergence of big data analytics within enterprises, as well as its application in the healthcare industry [118]. This article examines cancer analytics data visualization. The varied statistical models, deep learning models, and architectures used to anticipate data throughout time are also examined, along with the obstacles that spurred our research.

3. Empirical methodology

3.1 Data source and experimental manipulation

The dataset came from the Centers for Disease Control and prevention (CDC) and the American Cancer Society (S1 Appendix). The data includes cancer incidence rates, risk factors such as tobacco smoking, alcohol consumption, physical inactivity, obesity/excess body weight, and cancers caused by these risk factors. Our 1990–2021 data spans 32 years. Moreover, the study sets itself apart by developing a methodology to forecast the occurrence of cancer by the year 2025. The deployment of a novel two-stage attention model is used to forecast the rate of growth in cancer incidence, with 2017 serving as the reference year. It has been established that the prevalence of cancer is projected to increase by 9 percent by December 2020 and 12 percent by December 2021. There is an expected increase of 14 percent and 16 percent in newly diagnosed cancer cases by the end of 2022 and 2023, respectively. Furthermore, it is estimated that the number of cancer cases will rise by 19 percent and 21 percent in the years 2024 and 2025, respectively. The expected increase in cancer rates can be ascribed, partially, to a rise in detrimental lifestyle choices, such as the consumption of alcohol and tobacco, lack of physical activity, and a subpar diet. These factors contribute to the high occurrence of obesity in the United States. The accuracy and generalizability of a methodology are significantly influenced by the quality and completeness of the training data, in the case of data dependence. Inconsistencies, anomalies, and missing data points can all hinder a model’s ability to discover accurate patterns. During the feature selection phase for the TSA-LSTM model, the most pertinent variables or characteristics that are probable to exert a substantial influence on the prediction of cancer incidence and mortality rates are identified. By eliminating extraneous complexity and noise and concentrating on the most informative data points, this procedure is vital for optimizing the model’s performance. Furthermore, the complexity of the model. While simpler models may offer greater interpretability, they may fail to encompass the complete intricacy of the data. Extremely complex models, on the other hand, are susceptible to overfitting and have difficulty generalizing to unobserved data. Moreover, with regard to generalizability, it is frequently constructed and evaluated on particular datasets. The TSA-LSTM model selects features by extracting relevant variables from big datasets and using domain expertise to find those most likely to be associated with cancer outcomes. Environmental factors (pollution, carcinogen exposure), genetic markers, lifestyle factors (smoking, nutrition), demographic data (age, gender, ethnicity), and prior cancer incidence rates are commonly considered for feature selection [31,119]. It also protects Americans from international and domestic health, safety, and security risks with special reaction teams to fight diseases [120,121]. It oversees public health trend research.

The American Cancer Society’s mission is to save and celebrate lives and fight for a world without cancer. The organization estimates the present cancer burden by projecting the number of new cancer cases and deaths expected each year as cancer incidence and mortality data are three to four years behind [122,123]. Researching and pursuing public health trends is its responsibility. The American Cancer Society is a nonprofit organization dedicated to eliminating cancer and promoting a future without it. Additionally, they save and celebrate lives. Since cancer incidence and mortality data are three to four years behind the current year, the organization estimates the current cancer burden by projecting future cancer cases and deaths. The TSA-LSTM model consistently shows better predictive performance metrics, such as accuracy, precision, recall, and F1 score, compared to traditional models like logistic regression, decision trees, random forests, and simpler LSTM architectures. The TSA-LSTM model excels in cancer prediction tasks by effectively capturing long-range dependencies and temporal linkages, leading to superior accuracy and reliability compared to other models. Accurate cancer incidence and mortality rate estimates are crucial for healthcare planning and patient outcomes due to the significant impact of early diagnosis and exact forecasting [124,125]. In this regard, we wish to illustrate the correlation between cancer incidence and risk factors including obesity, smoking, drinking, and inactivity. In order to achieve our visualization objective, four preprocessing procedures are necessary. First, we verify each lifestyle choice and its annual cancer incidence in Excel for missing data. In R, we use mean imputation to fill missing values. Missing values are imputed with the sample mean. Third, we delete duplicate observations using Excel. Fourth, we classify lifestyle choices by age, sex, or type of cancer. Finally, for correlation analysis, we use R to sum the average classification data.

The study aims to create a more accurate forecasting algorithm. Our initial cancer data comprises just 32 yearly entries (February 1990 to March 2021), which is not enough to improve model training and prediction. To improve outcomes, the study interpolates data into numerous points. Use interpolation to discover the optimum data enhancement method. Interpolation is estimating an unknown quantity between two known variables or inferring missing information from one. Building a function that intersects a discrete set of known data points is the core of this strategy. When missing data is available and cycles are long, this technique works [126,127]. A variable’s predictive strength, relevance to the result, and multicollinearity concerns usually determine its inclusion. When variables are highly correlated with the target variable and can illuminate patterns, they are more likely to be included in the model. Superfluous variables that don’t affect predictions or add noise may be deleted to simplify the model. Through domain knowledge-based pruning, correlation analysis, feature importance ranking, and mutual information analysis, TSA-LSTM model feature selection variables can be identified. The feature selection method is critical for ensuring that the TSA-LSTM model can accurately capture the complex correlations in cancer data and to optimize its performance. Constantly, it is assumed that the performance of the model will effectively extrapolate to unobserved data that possesses comparable attributes. Nevertheless, such an outcome cannot always be ensured, particularly in cases where the unobserved data diverges substantially from the training data. The model’s ability to generate more reliable estimates of cancer incidence and mortality rates is due to its rigorous selection process that eliminates extraneous or redundant components.

A frequent interpolation method is spline. Spline interpolation uses a piecewise polynomial as the interpolant. Many prefer spline interpolation over polynomial interpolation because the inaccuracy can be minimized even with low-degree polynomials. Spline interpolation eliminates Runge’s phenomenon, which causes oscillation between locations when using high-degree polynomials [28,123,128]. As in 2.4, spline interpolation can be linear, quadratic, or cubic. We purposely construct missing values in 32 data points to decide the study’s interpolation strategy (Linear, Quadratic, or Cubic). We keep only certain years’ data (1990, 1994, 1997, 2001, 2004, 2007, 2010, 2012,2014,2017,2019 and 2021). We now utilize each interpolation method to predict and fill in missing data for other years (1991, 1992, 1993, 1995, 1996, 1998, 1999, 2000, 2002, 2003, 2005, 2006, 2008, 2009, 2011 2013,2015,2017,2019, and 2021) Interpolation, which anticipates missing figures more accurately, is our choice. Each method’s missing value-filling results are in (S2 Appendix) The table shows that all three models predicted but did not match. Therefore, we augmented our data using the cubic spline technique which upgraded and increased the data into hundreds of entries [41,113]. This folded the data from yearly to monthly entries, using some for-model testing and the rest for training. Adding synthetic data points to our data would improve model training for accurate deployment. The data is augmented using the cubic spline. Focusing on our raw data (from 1990–2021), we have 32 data points for the cubic spine for 420 data points.

3.2 Variable construction and correlation design

Tableau will be used to highlight the association between lifestyle and cancer incidence. Our dependent variable new cases, fatalities, and cancers are linked to lifestyle choices. Smoking, alcohol, nutrition, inactivity/lack of exercise, and bad nutrition cause obesity. In this study, cancer data shows a dependent variable, where the total number of new cancer incidences per year. Tobacco smoking, alcohol consumption, physical inactivity, and obesity show the independent variables.

3.3 Tableau data visualization

Effectively using people and data is as important as the method. Data must be reviewed, cleaned, processed, and modeled to find usable information, draw conclusions, and aid decision-making [31,129].Tableau is a powerful intelligence platform that uses analytics to harness data. It streamlines data viewing, processing, and comprehension. New data modeling tools in Tableau make sophisticated data analysis easier without database knowledge. We can seamlessly mix data utilizing table relationships to maintain the data model adaptable for analysis [130,131]. We describe forecasting, analysis modeling, statistical measures, and smoothing coefficients dataset processing. For comprehension, Tableau visualizes data. Launch. Tableau’s launch page links to databases, text files, and other data. Tableau analysis and source access begin. Once connected and imported, Tableau instantly displays all associated data sets on the data source display page. Without unnecessary data, the data translator cleans and visualizes loaded data on this page. Create Tableau visualizations and spreadsheets using all loaded data. Create dashboard and story visualizations by dragging and dropping measures and dimensions in Tableau’s spreadsheet. Enhancing perspectives Dashboards should make sense from data visualization. Dashboards with many stories allow users to make crucial decisions. In Tableau dashboards, data can be dragged and dropped into stories. [31,125,127]. In the results and analysis, many visualizations are discussed. LSTMs are neural network architectures, not models, therefore they can be changed for specific operations. A rudimentary LSTM model may perform poorly due to weak operations. The basic LSTM model presented to handle gradient issues (Exploring Gradient Problems and Vanishing Gradient Problems) contained simply an input and output gate and maintained the RNN’s cell state [113]. Besides gradient difficulties, the RNN had input and output weight conflicts that the simple LSTM solved. Though better than the RNN, this technique does not fully solve its problems [113,132]. A forget gate reduced the number of sequences truncated in the original LSTM Model by deleting valuable material from the hidden state and forgetting unneeded ones. Forget-gate biases impact LSTM performance. Many academics have presented Vanilla, Stacked, CNN, Bidirectional, and other LSTM designs. Evolutionary approaches were preferred over gradient descent by Shalev-Shwartz, Shamir, and Shammah (2017). Attention emphasizes certain data [23,133]. Studies have shown that attention networks are needed to exhibit data parts. This study proposes a two-stage attention model to improve the LSTM model’s attentiveness and prediction accuracy [28,123]. Our work visualizes lifestyle patterns and specific malignancies and estimates future cancer incidences using raw data cleanup and sample enlargement. Interpretive research and an innovative way to examine complicated lifestyle datasets with Tableau dashboard visual analytics were also discussed. The chapter explained our data preparation materials and new model design. This part also checks our data interpolation strategy for optimal results.

4. Results and analysis of modelling based on LSTM

4.1 Double exponential smoothing data analysis modeling and forecasting

Here, we examine and describe the data analytic modeling used to anticipate cancer data using Tableau’s double exponential smoothing approach. Alpha and beta values of 0.5 produced the lowest error margin in LSTM-based data processing and analysis models. In testing, these coefficients were employed. Table 1 shows forecast alternatives based on actual yearly time series entries from 1990 to 2021 and measured by age-standardized measure per 100,000.

thumbnail
Table 1. Summary of forecast options and model used in DES.

https://doi.org/10.1371/journal.pone.0317148.t001

The forecast is extended to three years without seasonal pattern, although the model seeks a recurrent seasonal pattern in every period [134]. The forecast model shows multiplicative levels, trends, and no seasonality. Root mean squared error (RMSE), Akaike information criterion (AIC), mean absolute scaled error (MASE), mean absolute percentage error (MAPE), and mean absolute error (MAE) are used to ensure forecast quality in DES [135,136]. Each quality indicator provides statistical model quality data. Level, trend, and seasonal smoothing coefficients are alpha, beta, and gamma. Level and trend smoothing coefficients alpha and beta are 0.50 and gamma is 0.00 for the forecast. Forecasts are more naïve because to smoothing coefficients in our investigation. In this study when assessing the effectiveness of a prediction technique, it is essential to consider more than just one measurement. Two important approaches, Root Mean Squared Error (RMSE) and R-squared (R2), provide essential information’s primarily measures the size of mistakes, giving the average proximity between predictions and actual values (lower RMSE indicates better accuracy). The coefficient of determination, R2, quantifies the degree to which the model’s predictions align with the general pattern observed in the data (a greater R2 indicates a stronger fit). By utilizing both measurements, one can acquire a fuller comprehension of the model’s performance—its average accuracy and its ability to effectively capture the overall patterns. Furthermore, the use of boxplots to visualize the distribution of errors is quite useful. Boxplots can expose outliers or biases in the predictions, offering a more comprehensive examination beyond the average error measured by RMSE. Overall, the performance of a prediction approach can be thoroughly assessed by using a combination of RMSE, R2 (Coefficient of Determination), and boxplots.

4.2 TSA-LSTM model design using LSTM

The paper’s model is based on time series lifestyle and cancer incidence data from the Centers for Disease Control and Prevention and the American Cancer Association from 1990–2021. Thus, it is better for LSTM network applications since it solves the complexity and timeliness issues produced by the three gates. First and foremost, the TSA-LSTM model attention mechanism prioritizes the input characteristics, specifically their significance, to guarantee that the model encoder converges to specific characteristics of the input sequence when predicting the specific characteristics of the output sequence. This enhances the model’s natural learning trend and prediction quality. The quality of predictions is enhanced by selecting network features and maintaining model performance in real-time. Our model employs a Long Short-Term Memory (LSTM) Recurrent Neural Network that is capable of training long-term dependencies. LSTMs avoid long-term reliance, making long-term recall possible [23,113]. We exploited attention mechanisms and their strong operations to retain major cancer trend features for a decent forecasting model for active cancer research. The network uses attention processes to focus on one part of a complex input until classification. To simplify difficult tasks, process smaller regions sequentially. Like the human mind breaks down a new difficulty into smaller tasks and addresses them separately [41,127]. Input attention, temporal attention, and repetition help our model outperform baseline models in studies. Based on the model idea, we propose the Two-Stage Attention LSTM (TSA-LSTM) model for cancer data prediction in this research [137]. Sensitivity analysis helps users understand the model’s sensitivity to a wide range of scenarios and potential sources of uncertainty. This improves user decision-making. These goals are achieved by perturbing input features and analyzing the results. Healthcare practitioners can develop trust in the model and understand its restrictions by providing model-specific documentation on its structure, training methods, input characteristics, and output assessments. These techniques increase the TSA-LSTM model’s interpretability, making it accessible to healthcare practitioners without machine learning experience [23,35].

The following sections of this section provide a detailed explanation of the model’s components and provide essential insights into its testing configuration. In this work, we suggest a time series model for the accurate evaluation and long-term prediction of cancer incidences. Given n and τ as input sequence and window size length respectively; the yearly incidence movement is denoted . Then, the moving sequence of length τ is denoted an, d a vector of n moving input sequence at time t as . Again, given the previous values of the target sequence as (y1,y2,y3,…,yτ−1) and yt∈ℝ, and the current and past values of the n moving sequence as (x1,x2,x3,…,xτ);xt∈ℝn, non-linear autoregressive exogenous models aim to optimize nonlinear mappings to the current value of a target sequence yτ using with F(•) as the non-linear mapping function to determine via optimization. In the suggested model, we used LSTM units as encoding units [127,138]. We chose an LSTM encoding unit because its cell states sustain states across time, avoiding vanishing gradients and improving long-term information retention in our time series model. Formally, taking each time t, an input vector xt, a memory state vector ct, and ht as the output of ct, then the LSTM units can be formulated as: where, W and U are the weight matrices. b is a bias vector. it, ft, ot are the input, forget, and output gating units and σ is our activation function for the gating units. For our time series prediction at the time t, given the input sequence X = (x1,x2,x3,…,xτ)|xt∈ℝn, we apply our LSTM encoder to learn the mapping optimum mapping from xt to ht using ht = f1(ht−1,xt) ht∈ℝs, where ht is the hidden state of the encoder at the time t f1 is our LSTM non-linear activation function s is the size of the hidden state ht.

The existing models are compared in (Fig 2) based on their average accuracy ratings in testing and forecasting the cancer data. We trained and evaluated each model ten times in the same environment. Each model overwrote the training losses of the previously implemented models. The initial testing of the Gated Recurrent Unit model yielded an average accuracy of 94.27% for 128 units. The first tested network, the GRU network, showed a clear intent with an accuracy of 93.158. The second indicator is the BIGRU network (92.099), which displays 94.14%. Three, the LSTM network (94.101) exhibits a 79.23% percentage. Following this, the BILSTM network (91.686) exhibits a 54.87% increase. Fifth, the GRU seq-2 network (78.132) displays a 93.78% accuracy. The sixth network, the BIGRU seq-2 network (51.555), displays a 93.43%. Seventh, the LSTM seq-2 network (92.212) exhibits a 93.38% rating. Eighth, the BILSTM seq-2 network (91.799) boasts a 95.98% efficiency. On the ninth position, the LSTM seq-2-seq network with input intention (95.959) exhibits an accuracy of 87.98%. The LSTM network with two stages of attention (97.763) demonstrates a 98.87% accuracy the tenth time. Finally, the testing set executes the proposed TSA-LSTM model. Each model overwrote the training losses of the previously implemented models. Results indicate an average accuracy of 87.96%.

4.3 Two stage attention mechanism

Our LSTM encoder reads and encodes input sequences into fixed-length vectors for two-stage attention. Sequences encoded as fixed-length vectors prohibit models from memorizing longer ones (Eqs 1418). We augment our model’s encoding unit to boost prediction accuracy. The model’s LSTM encoder’s convergence on input and output sequence features is examined first. Learning improved model prediction. We construct a second attention mechanism, temporal attention, to choose critical network attributes at all times while decoding to preserve model performance as sequence length rises [125,139]. Real-time model performance picks network features and improves prediction. Assuming a-th moving sequence input, , we create an input attention mechanism with a multilayer perceptron by applying the LSTM cell state ct−1 and the previous hidden state ht−1 in the equation: (Eq 14) Where

  • is the attention weight that holds the relevance of the a-th input feature at time t.
  • , We∈ℝτ×2s, Ue∈ℝτ×τ are learnable parameters.

Lastly, we sum all the attention weights to 1 by applying a soft auction to . In combining the attention mechanism with the LSTM encoder, a resulting sequence, , the hidden state of the encoder is updated as and thus, allows the LSTM encoding unit of the model can selectively focus on unique characteristics of the input sequence. We construct an LSTM decoding unit that uses the model’s internal representation to predict output . We construct an LSTM decoding unit that uses the model’s internal representation to predict output a. In order to maintain the performance of the model through the entire network with as the sequence length increases, we construct a temporal attention mechanism that chooses vital features of the network at all times while decoding. Given that at time t, the attention weight of an encoder’s hidden state is computed as: (Eq 15) with (Eq 16) where

  1. dt−1∈ℝp is the hidden state of a previous decoder.
  2. is the cell state of the LSTM unit.
  3. are learnable parameters.

Similar to the encoding unit’s input attention mechanism, the temporal attention mechanism holds the relevant features, but this time, of the i-th encoder hidden state for prediction using attention weight at the decoding level [140]. Considering the encoder’s hidden state hi is mapped to the input’s temporal component, the attention mechanism evaluates the context vector ct (unique at each time) as the weighted sum of all the hidden states {h1,h2,h3,…,hτ} of the encoding units: (Eq 17)

Hence, the combination of the target sequence and context vectors arrives at: (Eq 18) where and maps [yt−1;ct−1] to the decoder’s size. is then used to update the decoder’s hidden state at time t. We created and implemented nine neural baseline models to compare our model to others using neural networks. Our hierarchical neural model has customizable window and encoder/decoder hidden state sizes. We chose the two-stage attention encoder-decoder model since it obtained the best results among state-of-the-art models. The TSA-LSTM model pseudocode is in S3 Appendix. The analysis and correlational data have been refined with the inclusion of six subfactors. Priority is given to the following associations: alcohol use and cancer, tobacco use and cancer, obesity and cancer, physical inactivity and cancer, and cancer itself. The sixth proportion of malignancies in the United States that are ascribed to risk factors exhibits temporal patterns of incidence. Female breast cancer is one of the most common alcohol-related cancers in the US, with 138.3 per 100,000 cases [41,141]. Women who drink excessively are at risk of breast cancer. Alcohol intake is also linked to 65.8 colon and rectum cancers, 21.7 lip cancers, 8.1 liver cancers, 5.8 esophageal cancers, and 3.3 larynx cancers per 100,000 persons This visual uses a horizontal polygon map to quickly visualize data patterns. Second, (Fig 3) depicts how tobacco smoking promotes cancer. Tracheal, lung, bronchus, colon rectal, and urinary bladder cancers are the most common tobacco-related malignancies. Smoking causes 58.80 cases of tracheal, lung, and bronchus cancer per 100,000 persons, making it the most common cancer caused by smoking. Tobacco smoking causes 49.70 colon and rectal cancers, 30.10 urinary bladder cancers, and 36.60 kidney/renal pelvis cancers. Additionally, tobacco smoking causes 14.20 cases of acute myeloid leukemia per 100,000 people. Additionally, the cancer is the most terrifying outcome of smoking, which casts a dark cloud over people’s health. And smoke increases the chance for a terrifying array of malignancies due to its over 7,000 compounds, including 70 proven carcinogens, which rampage through the body [142]. There is extensive harm, including but not limited to lung cancer (the leading cause of cancer-related deaths, accounting for 80% of all cases) and malignancies of the esophagus, larynx, lip, mouth, pharynx, bladder, and kidneys [143].

Third, the bubble chart in (Fig 3) illustrates that obesity is a major cause of cancer. Breast cancer after menopause is the most common cancer-related obesity, followed by colon, rectum, uterine, thyroid, pancreatic, ovary, and kidney malignancies. Obesity is linked to 439 postmenopausal female breast cancer, 43.9 colon/rectum cancer, 46.7 corpus/uterine cancer, and 31.4 thyroid cancers per 100,000 cases. Due to obesity, 21.2 new pancreatic, 31.2 ovarian, and 40.9 kidney malignancies are diagnosed. Obesity is linked to liver, gallbladder, esophageal, and meninges cancers [123,144]. The smoke can reach every part of the body, including the digestive tract. It can alter the balance of good bacteria in the gut, which can lead to colon cancer, as well as other malignancies in the liver and stomach. Smoking raises the risk of Acute Myeloid Leukemia, while the link is less clear [145]. Cancer risk drops considerably after the body’s extraordinary healing mechanisms take effect. When people stop smoking, they reclaim agency, enhance air quality, and dramatically reduce their cancer risk. Color and size of circles demonstrate how fat impacts cancer [146]. Attribution increases with circle size. The bubble chart is visually appealing and makes it easier to identify categories and compare obesity attribution data by bubble size. The aspect of interpretability holds significant importance, particularly within healthcare environments where the outcomes of patient outcomes can be profoundly impacted by decisions predicated on predictive models [18,127]. The TSA-LSTM model is difficult, however healthcare professionals without machine learning model knowledge can apply many approaches to simplify its conclusions. Essentiality of the feature Visualisations or quick summaries of the model’s essential characteristics may help healthcare providers[20]. During the process of prediction, the visualization of the attention mechanism that is utilized by the TSA-LSTM model might provide information regarding which particular segments of the input data are given priority by the model. In the process of making judgments, this can be of assistance to users in understanding the manner in which the model assigns weights to different time steps or qualities [22]. Explaining forecasts may help healthcare providers understand the model’s logic for predicting a patient or case.

Fourth, physical inactivity and cancer show (Fig 4) that physical inactivity is the leading cause of US female breast, colon, and corpus/uterine cancers. Physical inactivity causes 438.2 postmenopausal female breast cancers per 100,000 persons. Physical inactivity also causes 38.6 colon, rectum, and 35.7 corpus/uterine cancers in men and women. If female youngsters exercise more, the worrying number of female breast cancers may decrease [147]. It is possible that the present graph, which attempts to demonstrate a correlation between obesity and cancer incidence, is deceiving. It probably shows the total number of cancer cases, but it doesn’t show how much of an increase in risk there is because of fat [148]. This is due to the fact that, similar to breast cancer, a high total incidence does not always indicate that a large percentage of those instances are attributed to obesity. The ideal representation would center on the percentage of cancer cases linked to obesity for each kind in order to offer a more accurate picture. Alternatives include stacked bars, proportionate bars, and heatmaps. A heatmap would use color intensity to show attributable risk, proportionate bars to show obesity’s percentage of cases, and stacked bars to divide the total incidence into obesity, other risk factors, and unknown causes. Consider age and sex for a clearer image[149]. Menopause may affect breast cancer obesity risk factors. The graphic should be accompanied by the number of obesity-related cancer cases for clarity. Focusing on these regions helps researchers understand obesity’s particular impact in cancer risk.

Fig 5 bar chart simplifies comparing malignancies caused by physical inactivity and speeds up decision-making. Fifth, US cancer incidence statistics from 1990 to 2021 reveal that cancer incidence for all ages is rising, indicating a bigger number of cancers. New cancer cases soared to approximately 23 million. Cancer rates worsened and increased with time [150]. The line chart in this study simplifies these perspectives and trends. Sixth, Fig 6’s bar chart reveals that health risk factors directly cause all cervical cancers and 4.4% of prostate cancers and deaths. Risk factors cause 84.1 percent of tracheal, bronchus, and lung cancer, 71.8 percent of laryngeal cancer, 78.94 percent of lip and oral cavity, 69.8 percent of esophageal, and 64 percent of nasopharynx cancer cases and deaths. Health risk factors did not affect brain and nervous system cancer, pulmonary sarcoidosis, lymphatic filariasis, malignant skin melanoma, non-melanoma skin cancer, or testicular cancer. Health risk factors are linked to > 50% of 7 of 26 cancer types. Health risks cause 43.2% of colon and rectum cancers, 41.5 percent of liver cancer, 36.5 percent of uterine cancer, 42.3 percent of kidney cancer, 52 percent of bladder cancer, 27.5 percent of pancreatic cancer, 34.9% of breast cancer, 21.78% of stomach cancer, and 35% of gallbladder and biliary tract cancers. The bar’s height and color indicate how much health risk factors contribute to cancer incidence and death. Risk factors cause 7.9% of thyroid cancer and 7.8% of Non-Lymphoma Hodgkin’s incidence and mortality. The existing graph appears to have an issue with the presentation of the cervical cancer risk variables. The shade and bar height provide the wrong idea, even though they are meant to indicate the risk associated with obesity. The Centers for Disease Control and Prevention (CDC) reports that the Human Papillomavirus (HPV) is the causative agent of cervical cancer in around 90% of cases [151]. However, being overweight is not a primary cause. It isn’t the main cause of the condition, but it could weaken the immune system and impact HPV progression in some circumstances. The existing layout of the graph, which uses color shading and bar height to portray the same data, incorrectly suggests that obesity significantly increases the risk of cervical cancer [152,153].This might be because of an error in the data mapping process or because the graph is lacking a data point that would indicate the risk of HPV infection. Having two separate data points for cervical cancer would be excellent for a clearer picture[154]. A large value, maybe the entire height of the bar or a dark shade, would indicate a high risk of HPV infection, whereas a little value, maybe a minimal height of the bar or a light shade, would indicate a low risk of obesity. It is essential to have labels that clearly explain the meaning of the colors and heights in order to prevent any misunderstandings. These changes will make the graph a better tool for learning about the real cervical cancer risk factors[155].

thumbnail
Fig 5. Share of cancer deaths by type attributable to risk factors in the USA.

https://doi.org/10.1371/journal.pone.0317148.g005

4.4 Forecasting results

All models were trained and tested 10 times in the same environment. We test on a GPU cluster of eight NVIDIA 1080Ti 32GB in simultaneously. Mini batching (128-person) was employed during training. A joint Stochastic Gradient Descent and Adam optimizer with 0.001 learning rate were used. We use 1000 training epochs and an early stopping strategy to cease when training losses do not decrease after 20 epochs. Our goal function was Mean Squared Error. Dropping 20% on all layers of each model we tested prevents overfitting. LSTM models are susceptible to overfitting, particularly when data is sparse. This can happen when the model grasps patterns that are unique to the training data but do not apply effectively to new data. Fig 6 presents the experimental results from the testing of cancer data on the Double Exponential Smoothing method. It is observed that the Double Exponential Smoothing model diverges as the years increase, even after testing data points of two years, unlike the neural networks models which undulate evenly and smoothly along the timeline [127,144,156]. This could be due to the absence of seasonality in the Double Exponential Smoothing, whereas the neural networks approach and factor the seasonality. After the model testing, it is realized that the Double Exponential Smoothing has an average accuracy of 88.50% in the prediction of cancer data.

This implies that the Double Exponential Smoothing method may help in short-term predictions (1–6 months), however, this method is comparatively low/weak in accuracy and not good for a long-term cancer projection for effective policy formulation. To test our model’s network, we first conduct experiments on nine existing prediction state-of-the-art model networks, including: A Gated Recurrent Unit (GRU), A Bidirectional Gated Recurrent Unit (BiGRU), A Sequence-to-Sequence Gated Recurrent Unit (Seq-2-Seq GRU), A Bidirectional Sequence-to-Sequence Gated Recurrent Unit (Seq-2-Seq BiGRU), A Long short-term memory (LSTM), A Bidirectional Long short-term memory (BiLSTM), A Sequence-to-Sequence Long short-term memory (Seq-2-Seq LSTM), A Sequence-to-Sequence Bidirectional Long short-term memory (Seq-2-Seq BiLSTM), A Sequence-to-Sequence Long short-term memory with Input Attention (Seq-2-Seq LSTM + Input Attention). The average cancer data testing and predicting accuracy scores of existing models are compared [150,156]. The models underwent 10 rounds of training and evaluation in a consistent context. Each subsequent model replaced the training loss of the previous models. The initial testing findings of the Gated Recurrent Unit (GRU) model, as demonstrated in Fig 6, indicate an accuracy of 98.26%. Subsequently, the performance of the Bidirectional Gated Recurrent Unit model, consisting of 128 units, was evaluated, yielding an average accuracy of 97.1 percent. After the completion of the third round of testing, the Gated Recurrent Unit Sequence to Sequence network consisting of 128 units demonstrated an accuracy of 88.23%. The Bidirectional Sequence-to-Sequence Gated Recurrent Unit that was displayed was subjected to four tests. The Long short-term memory model, which incorporated post-training data training, had an average accuracy of 95.34 percent, whereas this model achieved 64.36 percent. Subsequently, the performance of the Bidirectional Long short-term memory Network, consisting of 128 units, was evaluated and yielded an average accuracy of 94.31 percent. Sequence-to-Sequence Long short-term memory, Bidirectional Long short-term memory, and long short-term memory with Input Attention were directly tested and achieved average accuracies of 89.21, 94.30, and 86.86. Finally, the set of testing is implemented on our proposed TSA-LSTM model. Just like the other State of the Art models, we trained ours and evaluated it ten times in the same environment. Each model overwrote the training loss of the previously implemented models. The result shows an average accuracy of 98.23%.

4.5 Quantitative analysis of models

Using real forecasts is crucial for accuracy evaluation. Thus, residual size does not predict genuine forecast errors. Only by testing a model on new data not utilized when fitting it can forecast accuracy be verified. Data is usually split into training and test data when picking models, with the former used to estimate forecasting technique parameters and the latter to assess accuracy. Since the test data is not used to forecast, it should give a good indicator of how well the model will forecast fresh data. In this study, training lasted from February 1990 to March 2021 and testing was from We also increased the test size to the maximum prediction horizon. We presume that a model that matches the training data well will not forecast well when creating our new model [157]. A model with enough parameters produces a perfect match every time. Overfitting a model is as harmful as not finding a pattern in data. Some publications call the test set the “hold-out set” since these data are “held out” of fitting data. This study uses “training data” and “test data” instead of “in-sample data” and “out-of-sample data” as references do.

We defined forecast error as the difference between an actual value and its forecast in this investigation. Thus, an error is the unpredictable component of an observation, not a mistake. Write it as: (Eq 19) Where the training data is given by {y1,…,yT} and the test data is given by {yT+1, yT+2,}.

Forecast errors differ from residuals in two ways. Remainders are calculated on the training set and forecast mistakes on the test set. Second, residuals use one-step forecasts whereas forecast errors use multi-step forecasts. We use scaled-dependent errors to quantify forecast accuracy. This study’s forecast mistakes are on the same scale as the data hence the models’ accuracy metrics are scaled-dependent as the cancer data is in numerous time series with the same units (date and cancer incidence). Different forecasting models are compared by their error and accuracy rates. Forecasting models vary in accuracy. Table 2 compares double exponential smoothing, deep learning techniques, and our model. As mentioned earlier, accuracy is a key indicator of predictive algorithm success [158].

Our algorithm has the maximum accuracy of 98.96 percent and the lowest error rate of 3.24 percent. Double exponential smoothing predicts cancer incidence 9th among deep learning methods. We employed one double exponential smoothing approach, but the alpha-beta values with the fewest mistakes were used during implementation. Thus, the amount of double exponential methods utilized will not affect performance but may decrease it. Again, deep learning systems outperform double exponential smoothing in cancer data prediction accuracy. Comparatively, our model is the most accurate. Our algorithm uses two-stage attention. Attention models might provide interpretability while keeping deep learning’s flexibility and versatility. Comprehending the prediction process of the model might be challenging because of the intricate structure of LSTMs. In cancer research, the inability to grasp the model’s reasoning can be a drawback for creating novel therapy approaches. We initially use an input attention technique to adaptively extract important driving series at each time step by referencing the encoder’s hidden state. We pick relevant encoder hidden states across all time steps in the second stage using temporal attention. Our model can predict and interpret well with this dual-stage attention method [123,159,160]. Time series prediction and feature learning in computer vision tasks are possible with our model. This implies that our model forecasts cancer data better than others. Our algorithm may have a low margin error since the model’s attentions provide interpretability and flexibility.

4.6 Qualitative algorithm description

This section describes each algorithm based on three essential time series observations on our predicting outcomes. 1). predict smoothness 2) movement distance and orientation; 3) training average time. These observations may help explain algorithm behavior on cancer data. We discuss the neural network and DES algorithm training procedures to elucidate the observations. The first section gives four rudimentary forecasts and the second four refined ones. Primitive forecasts are training’s first projections without error margin. Refined forecasts are error margin-influenced. Both crude and sophisticated forecasts improve accuracy. The last two training instances just increased the error margin [161,162]. The two training instances and error margin are meant to help refined forecast training instances match the genuine trend [163]. The predicting results are supported by the error margin and two training instances using three fundamental time series data. The training procedures are exclusive to DES. However, as DES is not a neural network, the training instance only constituted a piece of the process. We assessed the effectiveness of each method in improving forecast accuracy by applying filtering techniques. Smoothing helps to clarify the trends in forecasting [35,164] Smoothing describes time series trend components. Smoothing can be linear, undulating, or non-undulating [165]. The optimal neural network algorithm increases undulating trends and minimizes non-undulating.

According to Table 3, the best neural network has a whole undulating pattern for forecast results and full distances between forecast results and true trends. Such a neural network needs to be trained for over 1700 seconds. Several experiments led to this conclusion. We trained our data with basic LSTM and GRU. We found that the simple LSTM model outperforms the GRU. This supports our decision to build on the basic LSTM with an attention network [166]. We fed the basic LSTM algorithm a one-stage attention network to improve performance. Our next hypothesis is that a two-stage attention network as an input to the basic LSTM would improve performance. We found that the two-stage attention network improves cancer data performance but requires many hours of training.

thumbnail
Table 3. Observations that influence the accuracy of the algorithm.

https://doi.org/10.1371/journal.pone.0317148.t003

4.7 Prediction of cancer incidences using the TSA-LSTM model

Fig 6 shows our unique TSA-LSTM model’s experimental cancer incidence forecasts. Predictions for new cancer cases are until 2025. The new cancer incidence estimates are increasing linearly, indicating high cancer rates until 2025. Our actual numbers ending 2021 suggest cancer cases may rise 9% by December 2020 and 18% by December 2021. Additionally, 2022 and 2023 are predicted to see 18% and 21% more cancer cases. Cancer incidences are expected to grow by 20% and 22% in 2024 and 2025 [167,168]. The forecasts are reasonable and incremental, with close prediction intervals reflecting previous data fluctuation. Table 4 summarizes the total projected cancer incidence in December each year. Additionally, it has been determined that the eating of particular foods is likely to elevate the risk of developing cancer. For example, consuming meals that induce hyperglycemia has been associated with an elevated risk of developing colorectal, stomach, and breast malignancies [31,62]. High-temperature cooking techniques, including grilling, frying, and barbecuing, can generate heterocyclic amines (HA), which have been linked to the development of cancer and various other ailments [62]. A high dairy intake is associated with an increased risk of getting prostate cancer. According to [169] an increased consumption of whole milk significantly heightens the likelihood of illness progression and mortality. Furthermore, it is widely acknowledged that physical inactivity is the leading cause of breast, colon, and corpus/uterine malignancies among women in the United States. Our projection shows rising US cancer rates by 2025. As shown in sections 4–20, lifestyle decisions and other risk factors contribute to the rising cancer rate [170]. Lifestyle choices including alcohol, smoke, inactivity, and poor diet have boosted obesity in the US, and this trend is expected to continue. Previous studies have connected poor lifestyle choices to cancer. The International Agency for Research on Cancer showed increased colon, breast, and endometrial cancer risk. Smoking causes breast and lung cancer, research shows [24,25,171].

Our results reveal that the Double Exponential Smoothing model diverges on cancer data over time. but bad for long-term forecasting. One of our 11 study algorithms. Comparatively, our model has a maximum accuracy of 98.96 percent. LSTM model sequenced by our technique used a two-stage attention sequence. We forecast US cancer incidence growth using our two-stage attention model with 2021 as our baseline (Table 4). Furthermore, our research findings suggest that the Double Exponential Smoothing model exhibits a divergence from the observed outcomes over time. This result is attributable to the Double Exponential Smoothing model’s absence of seasonal variation. This indicates that while the Double Exponential Smoothing method may be advantageous for short-term forecasts (1–6 months), it is not considered suitable for long-term forecasting. One of eleven algorithms that were implemented in our research. In comparison to alternative models, our proposed model exhibits the maximum magnitude of accuracy at 97.76 percent. The attention sequence to sequence LSTM (TSA-LSTM) model was utilized by our method in two stages. The rate of growth in cancer incidence in the United States is predicted using our two-stage attention model, with 2017 serving as the baseline year. By December of 2020, 2021, 2022, 2023, 2024, and 2025, the incidence of cancer is projected to grow by 9 percent, 12 percent, 14 percent, 16 percent, 19 percent, and 21 percent, respectively.

5. Discussion

In this study the exponential smoothing (ES) method can be considered as weighted moving averages after Brown (1959,1962) and Holt (1960) studied stock control system forecasting, it was introduced. This is the weighted average of historical observations [172]. Past observations are weighted exponentially decreasingly, while recent observations are weighted most heavily. Since it adapts well to short-term forecasting, it makes good daily predictions. SES, the simplest exponential smoothing approach, forecasts data without trend or seasonality [173]. SES is decreasing past observation weights exponentially. SES estimates future values using the preceding period’s forecast and forecast error. The Holt-Winters smoothing approach emphasizes seasonality or trend, unlike most ES methods. Holt (1957) and Winters (1960) introduced smoothing factors, trend (a slope), and seasonality (cyclical, recurring pattern) to this method to capture seasonality and trend. Triple exponential smoothing comprises the value, trend, and seasonality of the time series. The Holt-Winters’ adaptive model allows the level, t, rend, and seasonality patterns to change corresponding to time [171]. Past observations are getting exponentially less SES weights. The forecast for the preceding period and the forecast error are used to estimate future values in SES. Holt-Winters smoothing considers seasonality or trend, unlike most ES methods. By incorporating smoothing variables, trend (a slope), and seasonality (cyclical, recurring pattern), Holt (1957) and Winters (1960) improved this method to incorporate seasonality and trend. Time series value, trend, and seasonality are called triple exponential smoothing. Recent observations are weighted greater in exponential smoothing. These models extrapolate data trends and seasonality [174]. Exponential smoothing iteratively forecasts from future values of a time series form weighted averages of past values [175]. The use of the TSA-LSTM model in clinical settings is accompanied by several specific limitations that must be surmounted in order to fully exploit its capabilities. Error-prone and fragmented clinical data and the need for large amounts of labeled data for efficient training may reduce model precision and dependability. Healthcare professionals have difficulty deciphering the model as a result of its complexity. To foster confidence and promote adoption, offer straightforward explanations for predictions and showcase the value of the features being proposed. Furthermore, healthcare organizations that lack expertise and infrastructure in machine learning may encounter difficulties in training and deploying deep learning models such as TSA-LSTM as a result of the computational resource demands[176]. To ensure the model’s ethical clinical deployment, patient privacy, permission, bias mitigation, and healthcare data rules must be carefully considered. The TSA-LSTM model must also be smoothly linked into clinical procedures and electronic health record systems for practical adoption and user acceptance. This highlights the need for data scientists, healthcare providers, legislators, and regulators to work together to address these problems and use predictive analytics in healthcare.

The deep learning algorithms have outperformed computer vision in recent years, but their time series models are still shallow. Many data scientists in most industries use autoregressive models instead of deep learning. Interpretability, restricted data, ease of use, and training cost are common reasons for choosing these methods [177]. There is no single solution to these difficulties, but deep models with care make a strong case. They often increase performance (other vanilla LSTMs/RNNs) and interpretability. In many circumstances, they are faster than RNN/LSTM. Attention networks are effective for embedding categorical inference in deep neural networks [178]. Brain networks assist in awareness, sensory orientation, and conflict resolution. In most sequence transduction models, large recurrent or convolutional neural networks encode and decode locally without global information. Cancer and predictive healthcare analytics demonstrate the TSA-LSTM model’s utility and limits. The algorithm predicts cancer using historical data, demographics, environment, and genetic markers. It can assist governments and healthcare practitioners find patterns and cure cancer. It predicts cancer and chronic illness mortality over time using patient data and risk indicators, which may help clinicians improve therapy. This algorithm predicts therapeutic reactions using clinical data and illness development, helping clinicians optimize treatment regimens and select the best drugs. This tool analyzes patient data for temporal trends to diagnose cancer and other diseases early. This technological advancement empowers medical professionals to proactively enhance treatment outcomes through the identification of gradual fluctuations in health indicators that could potentially signify the start of an illness. Healthcare systems can use the TSA-LSTM model to predict future healthcare needs, such as cancer case volume, and allocate resources strategically. Oncology departments and facilities may improve resource use, capacity planning, and healthcare delivery. In general, the TSA-LSTM model improves prognostics, personalized medicines, and evidence-based decision-making in oncology and other healthcare sectors. The utilization of the model’s sophisticated temporal relationships and sequencing functions has the potential to enhance health outcomes, optimize resource allocation, and improve patient care. Critical healthcare applications exist for TSA-LSTM. The prediction of early cancer diagnosis, therapy responses, and cancer incidence and mortality. Beyond the realm of oncology, medical practitioners have the ability to optimize resource allocation, customize patient treatment, and enhance overall patient outcomes by utilizing the model’s comprehensive temporal sequences and patterns. This limitation led to the development of attention networks to capture long-distance interactions. Attention networks are increasingly utilized due to its fully parallelized computation and flexibility in modeling dependencies [179]. According to [180]. The best models use attention to link the encoder and decoder. Recurrent and convolutional models based only on attention mechanisms perform better, are more parallelizable, and take less time to train on two machine translation tasks. The work of (Ramachandran et al., 2019 [181]; and Sand uddle & Bashir, 2022 [182]) A simple ResNet model that replaced all spatial convolutions with self-attention showed attention networks’ impact. This fully self-attentional model surpassed the baseline with less categorization parameters. Later investigations on attention in generic deep neural networks show ongoing categorization and detection with different models.

Based on qualitative algorithm description, neural network techniques should have equally undulating or non-undulating forecasting instances in both training parts. We describe full or partial undulating and non-undulating filtering. We have two fully undulating training instances with equal tendencies. Fully non-undulating occurs when both training instances are non-undulating. Partial means undulating and non-undulating training instances are not equal [35,136]. In contrast, the DES trend is largely linear. Distance describes the intervals and slopes between the underlying trend and each anticipated result [183]. A neural network method functions best when primitive and refined training cases predict the true trend. Additionally, a neural network performs best when anticipated slopes match true trends. We call these lengths full or partial[184]. When both training instances give forecasting results closer to genuine trends, full distance is reached. Also, partial distance occurs when only one training instance forecasts trends closer to reality. We observe that the simple LSTM model has a full distance for most anticipated results since the underlying trend is in the middle. However, BILSTM outcomes have a partial distance because one anticipated result is far from the true trend [185]. Based on the TSA-LSTM Model connection between alcohol taking and cancers have been found. It has been found that cancer deaths attributable to alcohol consumption are 3.5% in the USA [186]. The consumption of certain foods has also been found to likely increase the probability of developing cancer. For instance, foods that cause blood glucose levels to rise are connected with an increased risk of breast, stomach, and colorectal cancers Moreover, Physical inactivity is seen as a principal cause of female breast, colon, and corpus/uterine cancers in the USA [68,186].

6. Conclusion and policy implication

There are various subtypes of cancer. Cancer ranks second in global prevalence, accounting for one in every six fatalities. The incidence of cancer is projected to increase as a result of factors such as lifestyle changes, population growth, and extended life expectancy. Thus, an accurate cancer forecast will enable health professionals, government agencies, and the public to prepare for its detrimental repercussions. The search for a good forecasting tool has been difficult for stakeholders. A solid forecasting approach offers accurate and understandable outcomes, not complex methodology. Current research divides forecasting approaches into three categories: artificial intelligence, statistical, and hybrid. Recent interest in statistical models has focused on Exponential Smoothing (ES) and (DES). ES, DES are weighted moving averages. Tableau, a popular data visualization and forecasting application, uses exponential smoothing. These approaches can reliably anticipate short-term data. Second, deep learning/machine learning algorithm forecasting techniques use intelligence and flexibility to improve prediction models. In deep learning, Artificial Neural Network (ANN) models are optimal for real-time estimations. The two are combined in the third algorithm. The health sector has employed various algorithms in pursuit of research objectives; nevertheless, the optimal algorithm for ensuring the accuracy of cancer data has yet to be investigated. This research endeavors to identify the most precise cancer data prediction software. The study does this by Testing double exponential smoothing on cancer data and comparing it to other deep-learning techniques. Second, creating a cancer data-accurate deep learning algorithm. Third, predict cancer incidence rise by 2025 using our innovative approach.

We contributed three parts to the study. First, Cancer Prediction using Double Exponential Smoothing and Deep Learning Algorithms Our results reveal that the Double Exponential Smoothing model diverges on cancer data over time. This is because Double Exponential Smoothing lacks seasonality. Thus, Double Exponential Smoothing may be useful for short-term predictions (1–6 months) but not long-term predictions. This study examined 11 methods, and double exponential smoothing predicted cancer incidence 9th best. Thus, DES has poor long-term cancer prognosis and data accuracy. Double Exponential Smoothing is unsuitable for long-term cancer projection in policy planning. Our model has a maximum accuracy of 98.96%. Our two-stage LSTM model sequencing method. While maintaining deep learning’s flexibility and versatility, attention models may give interpretability. We reference the encoder’s hidden state to adaptively extract significant driving series at each time step using input attention. We pick relevant encoder hidden states across all time steps in the second stage using temporal attention. Our model can predict and interpret well with this dual-stage attention method. Time series prediction and feature learning in computer vision tasks are possible with our model. This implies that our model forecasts cancer data better than others. Our approach may have great accuracy since the model’s attentions provide interpretability and flexibility. Third, 2025 cancer incidence forecast Using 2021 as a baseline, our innovative two-stage attention model predicts cancer incidence growth. The stipulated increase in cancer incidences is 14% by December 2020 and 12% by December 2021. Additionally, 2022 and 2023 are predicted to have 14% and 16% more cancer cases. Cancer incidences are expected to grow by 19% and 21% in 2024 and 2025. Lifestyle decisions including alcohol drinking, tobacco smoking, physical inactivity, and poor eating practices that lead to obesity in the US are projected to raise cancer rates. We advocate (i) widespread safety education on the harmful impacts of unhealthy lifestyle choices at home at work to lower cancer rates in the US. (ii) Lifestyle and cancer research should receive government financing. (iii). Provincial governments should also try new approaches to discourage unhealthy lifestyles. Reforms should limit the amount of tobacco or alcohol sold per day.

According to future prediction, we created a novel two-stage attention model technique that can forecast cancer incidence monthly with better accuracy than existing algorithms. Next, we examine engineering, health, economy, and social research directions. The engineering community can improve cancer prediction. Researchers can use our model as a baseline because long-term prediction demands great accuracy. Our unique approach has 98.96 percent accuracy. Next steps should boost accuracy and shorten training. This study is ongoing; therefore, we hope to build a model that can detect cancer cases weekly and daily instead of monthly. This strategy will emphasize cancer case management—early detection and death. We hope to achieve this through network optimization. Increased cancer cases may impact economic indicators. More cancer cases equal more treatment costs. This represents a national fiscal burden because cancer treatment requires large amounts. Cancer rise may affect GDP and inflation in future studies. Society views cancer survivors as having psychological and emotional disorders following treatment. This hypothesis can be tested by future research.

Acknowledgments

The author is profoundly grateful to the instructors for their encouragement and guidance, and will respond to their insightful advice. Fuzhou University of Technology’s educational opportunities are sincerely acknowledged. China’s National Social Science Foundation (22BGL007). GY-H-24176: Research on marine environmental protection policies and low-carbon ship emission control. The “Research on the Impact of Low Carbon Strategies on the Rural Revitalization Strategy in Fujian Province” (GY-S20014) project has been initiated by Fujian University of Technology.

References

  1. 1. Wieder R., Adam N.: Drug repositioning for cancer in the era of AI, big omics, and real-world data. Crit. Rev. Oncol. Hematol. 175, 103730 (2022). pmid:35654244
  2. 2. Gupta P., Jindal A., Ahuja G., Jayadeva, Sengupta, D.: A new deep learning technique reveals the exclusive functional contributions of individual cancer mutations. J. Biol. Chem. 298, 102177 (2022). https://doi.org/10.1016/j.jbc.2022.102177
  3. 3. Yang Y., McMahan C.S., Wang Y.-B., Ouyang Y.: Estimation of l0 norm penalized models: A statistical treatment. Comput. Stat. Data Anal. 192, 107902 (2024). pmid:38222104
  4. 4. Wang K., Diao H., Wei W., Xiao T., Chen Y., Zhou B.: Power system transient security assessment based on multi-channel time series data mining. Energy Reports. 8, 843–851 (2022). https://doi.org/10.1016/j.egyr.2022.08.153
  5. 5. Mostafa R.R., Khedr A.M., Al Aghbari Z., Afyouni I., Kamel I., Ahmed N.: An adaptive hybrid mutated differential evolution feature selection method for low and high-dimensional medical datasets. Knowledge-Based Syst. 283, 111218 (2024). https://doi.org/10.1016/j.knosys.2023.111218
  6. 6. Hu Y., Wang H., Zhang Y., Wen B.: Frequency prediction model combining ISFR model and LSTM network. Int. J. Electr. Power Energy Syst. 139, 108001 (2022). https://doi.org/10.1016/j.ijepes.2022.108001
  7. 7. Bilgili M., Pinar E.: Gross electricity consumption forecasting using LSTM and SARIMA approaches: A case study of Türkiye. Energy. 284, 128575 (2023). https://doi.org/10.1016/j.energy.2023.128575
  8. 8. Chan H.-L., Ouyang Y., Huang P.-J., Li H.-T., Chang C.-W., Chang B.-L., et al: Deep neural networks for the detection of temporal-lobe epileptiform discharges from scalp electroencephalograms. Biomed. Signal Process. Control. 84, 104698 (2023). https://doi.org/10.1016/j.bspc.2023.104698
  9. 9. Ahmad B., Achek A., Farooq M., Choi S.: Accelerated NLRP3 inflammasome-inhibitory peptide design using a recurrent neural network model and molecular dynamics simulations. Comput. Struct. Biotechnol. J. 21, 4825–4835 (2023). pmid:37854633
  10. 10. Jiang L., Sun X., Mercaldo F., Santone A.: DECAB-LSTM: Deep Contextualized Attentional Bidirectional LSTM for cancer hallmark classification. Knowledge-Based Syst. 210, 106486 (2020). https://doi.org/10.1016/j.knosys.2020.106486
  11. 11. Liu X., Zhou J.: Short-term wind power forecasting based on multivariate/multi-step LSTM with temporal feature attention mechanism. Appl. Soft Comput. 150, 111050 (2024). https://doi.org/10.1016/j.asoc.2023.111050
  12. 12. Meena G., Mohbey K.K., Kumar S., Lokesh K.: A hybrid deep learning approach for detecting sentiment polarities and knowledge graph representation on monkeypox tweets. Decis. Anal. J. 7, 100243 (2023). https://doi.org/10.1016/j.dajour.2023.100243
  13. 13. Wang Y., Kung L.A., Byrd T.A.: Big data analytics: Understanding its capabilities and potential benefits for healthcare organizations. Technol. Forecast. Soc. Change. (2018). https://doi.org/10.1016/j.techfore.2015.12.019
  14. 14. Korayim D., Chotia V., Jain G., Hassan S., Paolone F.: How big data analytics can create competitive advantage in high-stake decision forecasting? The mediating role of organizational innovation. Technol. Forecast. Soc. Change. 199, 123040 (2024). https://doi.org/10.1016/j.techfore.2023.123040
  15. 15. Bas E., Egrioglu E., Cansu T.: Robust training of median dendritic artificial neural networks for time series forecasting. Expert Syst. Appl. 238, 122080 (2024). https://doi.org/10.1016/j.eswa.2023.122080
  16. 16. Kim K., Lee K., Cho S., Kang D.U., Park S., Kang Y., et al: PAIP 2020: Microsatellite instability prediction in colorectal cancer. Med. Image Anal. 89, 102886 (2023). pmid:37494811
  17. 17. Men L., Ilk N., Tang X., Liu Y.: Multi-disease prediction using LSTM recurrent neural networks. Expert Syst. Appl. 177, 114905 (2021). https://doi.org/10.1016/j.eswa.2021.114905
  18. 18. Soori M., Arezoo B., Dastres R.: Artificial neural networks in supply chain management, a review. J. Econ. Technol. 1, 179–196 (2023). https://doi.org/10.1016/j.ject.2023.11.002
  19. 19. Wu X., Wang H.-Y., Shi P., Sun R., Wang X., Luo Z., et al: Long short-term memory model–A deep learning approach for medical data with irregularity in cancer predication with tumor markers. Comput. Biol. Med. 144, 105362 (2022). pmid:35299045
  20. 20. Atta H.A.: Artificial neural network [ANN] modeling for Tetracycline Adsorption on Rice husk using continuous system. Desalin. Water Treat. 100026 (2024). https://doi.org/10.1016/j.dwt.2024.100026
  21. 21. Smith M.B., Sparks H., Almagro J., Chaigne A., Behrens A., Dunsby C., Salbreux G.: Active mesh and neural network pipeline for cell aggregate segmentation. Biophys. J. 122, 1586–1599 (2023). pmid:37002604
  22. 22. Shahzad H., Sadiq M.N., Li Z., Algarni S., Alqahtani T., Irshad K.: Scientific computing of radiative heat transfer with thermal slip effects near stagnation point by artificial neural network. Case Stud. Therm. Eng. 54, 104024 (2024). https://doi.org/10.1016/j.csite.2024.104024
  23. 23. Najmuddin Omaid, Li Zhihui, Khan Rabnawaz, and W.Z.: Valuation of Land-Use/Land-Cover-Based Ecosystem Services in Afghanistan—An Assessment of the Past and Future. Land. 11, (2022)
  24. 24. Wang R., Tang Z., Wu Z., Xiao Y., Li J., Zhu J., et al: Construction and validation of nomograms to reduce completion thyroidectomy by predicting lymph node metastasis in low-risk papillary thyroid carcinoma. Eur. J. Surg. Oncol. 49, 1395–1404 (2023). pmid:37061404
  25. 25. Huang R., Ying Q., Lin Z., Zheng Z., Tan L., Tang G., et al: Extracting keyframes of breast ultrasound video using deep reinforcement learning. Med. Image Anal. 80, 102490 (2022). pmid:35717873
  26. 26. Jhala K., Byrne S.C., Hammer M.M.: Interpreting Lung Cancer Screening CTs: Practical Approach to Lung Cancer Screening and Application of Lung-RADS. Clin. Chest Med. (2023). https://doi.org/10.1016/j.ccm.2023.08.014
  27. 27. Khan, R.: The impact of agricultural intensification on carbon dioxide emissions and energy consumption: a comparative study of developing and developed nations.
  28. 28. Khan R.: Catch-up growth with alpha and beta decoupling and their relationships between CO2 emissions by GDP, population, energy production, and consumption. Heliyon. e31470 (2024). pmid:38845995
  29. 29. SabbaghGol H., Saadatfar H., Khazaiepoor M.: Evolution of the random subset feature selection algorithm for classification problem. Knowledge-Based Syst. 285, 111352 (2024). https://doi.org/10.1016/j.knosys.2023.111352
  30. 30. Catelli R., Casola V., De Pietro G., Fujita H., Esposito M.: Combining contextualized word representation and sub-document level analysis through Bi-LSTM+CRF architecture for clinical de-identification. Knowledge-Based Syst. 213, 106649 (2021). https://doi.org/10.1016/j.knosys.2020.106649
  31. 31. Rabnawaz W.J.K.: Renewable energy and CO2 emissions in developing and developed nations: a panel estimate approach. Front. Environ. Sci. 2024, (2024). https://doi.org/10.3389/fenvs.2024.1405001
  32. 32. Ma Z., Pan T., Tian J.: Deep reinforcement learning optimized double exponentially weighted moving average controller for chemical mechanical polishing processes. Chem. Eng. Res. Des. 197, 419–433 (2023). https://doi.org/10.1016/j.cherd.2023.07.049
  33. 33. Rashid T.A., Fattah P., Awla D.K.: Using Accuracy Measure for Improving the Training of LSTM with Metaheuristic Algorithms. Procedia Comput. Sci. 140, 324–333 (2018). https://doi.org/10.1016/j.procs.2018.10.307
  34. 34. Miao H., Kuruoğlu E.E., Xu T.: Non-homogeneous Poisson and renewal processes as spatial models for cancer mutation. Comput. Biol. Chem. 106, 107922 (2023). pmid:37499435
  35. 35. Khan R., Zhuang W., Najumddin O., Butt R.S., Ahmad I., Al-Faryan M.A.S.: The impact of agricultural intensification on carbon dioxide emissions and energy consumption: A comparative study of developing and developed nations. Front. Environ. Sci. 10, (2022). https://doi.org/10.3389/fenvs.2022.1036300
  36. 36. Golmankhaneh A.K., Tunç S., Schlichtinger A.M., Asanza D.M., Golmankhaneh A.K.: Modeling tumor growth using fractal calculus: Insights into tumor dynamics. Biosystems. 235, 105071 (2024). pmid:37944632
  37. 37. Niu X., Yan J., Wang X., Yang H., Ke L., Chen J., et al: Evaluation of energy correction algorithm for signals of PET in heavy-ion cancer therapy device. Nucl. Eng. Technol. 52, 101–108 (2020). https://doi.org/10.1016/j.net.2019.07.008
  38. 38. Bhuvaneshwari K.S., Rama Parvathy L., Chatrapathy K., Krishna Reddy C.V.: An internet of health things-driven skin cancer classification using progressive cyclical convolutional neural network with ResNexT50 optimized by exponential particle swarm optimization. Biomed. Signal Process. Control. 91, 105878 (2024). https://doi.org/10.1016/j.bspc.2023.105878
  39. 39. Ma G., Duan Y., Li M., Xie Z., Zhu J.: A probability smoothing Bi-RRT path planning algorithm for indoor robot. Futur. Gener. Comput. Syst. 143, 349–360 (2023). https://doi.org/10.1016/j.future.2023.02.004
  40. 40. Svetunkov I., Chen H., Boylan J.E.: A new taxonomy for vector exponential smoothing and its application to seasonal time series. Eur. J. Oper. Res. 304, 964–980 (2023). https://doi.org/10.1016/j.ejor.2022.04.040
  41. 41. Rabnawaz Khan Y.K.: Effects of Energy Consumption on GDP: New Evidence of 24 Countries on Their Natural Resources and Production of Electricity. Ekonomika. 99, 26–49 (2020). https://doi.org/10.15388/Ekon.2020.1.2
  42. 42. Hu Y., Sharma N.: Ergodic estimators of double exponential Ornstein–Uhlenbeck processes. J. Comput. Appl. Math. 434, 115329 (2023). https://doi.org/10.1016/j.cam.2023.115329
  43. 43. Garapati S.L., Garapati S.: Application of Big Data Analytics: An Innovation in Health Care, (2018)
  44. 44. Belavadi S. V, Rajagopal S., R, R., Mohan R.: Air Quality Forecasting using LSTM RNN and Wireless Sensor Networks. Procedia Comput. Sci. 170, 241–248 (2020). https://doi.org/10.1016/j.procs.2020.03.036
  45. 45. Singha S., Arha H., Kar A.K.: Healthcare analytics: A techno-functional perspective. Technol. Forecast. Soc. Change. 197, 122908 (2023). https://doi.org/10.1016/j.techfore.2023.122908
  46. 46. Chaudhury S., Sau K.: A blockchain-enabled internet of medical things system for breast cancer detection in healthcare. Healthc. Anal. 4, 100221 (2023). https://doi.org/10.1016/j.health.2023.100221
  47. 47. Zubair Hassan: The Impact of Using Big Data in Industries Today
  48. 48. Schuhmacher A., Haefner N., Honsberg K., Goldhahn J., Gassmann O.: The dominant logic of Big Tech in healthcare and pharma. Drug Discov. Today. 28, 103457 (2023). pmid:36427777
  49. 49. Sanjeev Agrawal: Why Is Data Analytics Important in Healthcare?
  50. 50. Dou T., Clasie B., Depauw N., Shen T., Brett R., Lu H.-M., et al: A deep LSTM autoencoder-based framework for predictive maintenance of a proton radiotherapy delivery system. Artif. Intell. Med. 132, 102387 (2022). pmid:36207077
  51. 51. Selmy H.A., Mohamed H.K., Medhat W.: Big data analytics deep learning techniques and applications: A survey. Inf. Syst. 120, 102318 (2024). https://doi.org/10.1016/j.is.2023.102318
  52. 52. Patra G.K., Bhimala K.R., Marndi A., Chowdhury S., Rahaman J., Nandi S., et al: Chapter 6—Deep learning methods for scientific and industrial research. In: Govindaraju V., Srinivasa Rao A.S.R., and Rao C.R.B.T.-H. of S. (eds.) Deep Learning. pp. 107–168. Elsevier (2023)
  53. 53. Hou Y., Zhang W., Cheng R., Zhang G., Guo Y., Hao Y., et al: Meta-adaptive-weighting-based bilateral multi-dimensional refined space feature attention network for imbalanced breast cancer histopathological image classification. Comput. Biol. Med. 164, 107300 (2023). pmid:37557055
  54. 54. IBM Watson Health: The importance of healthcare data analytics
  55. 55. Schary W., Brockmann F., Simantzik J., Paskali F., Kohl M.: Chapter 5—Big Data and Health Analytics explained. Presented at the (2024)
  56. 56. IBM White Papers and Reports: Data-driven healthcare organizations use big data analytics for big gains
  57. 57. Karatas M., Eriskin L., Deveci M., Pamucar D., Garg H.: Big Data for Healthcare Industry 4.0: Applications, challenges and future perspectives. Expert Syst. Appl. 200, 116912 (2022). https://doi.org/10.1016/j.eswa.2022.116912
  58. 58. Pandey M.K., Subbiah K.: A Deep Learning Approach for Classification of Medicare Beneficiaries Based on Gender and being Affected with Cancer. Procedia Comput. Sci. 218, 532–541 (2023). https://doi.org/10.1016/j.procs.2023.01.035
  59. 59. Clausen P.: Towards the Industry 4.0 agenda: Practitioners’ reasons why a digital transition of shop floor management visualization boards is warranted. Digit. Bus. 3, 100063 (2023). https://doi.org/10.1016/j.digbus.2023.100063
  60. 60. Garas S., Wright S.L.: A data analytics case study analyzing IRS SOI migration data using no code, low code technologies. J. Account. Educ. 66, 100885 (2024). https://doi.org/10.1016/j.jaccedu.2024.100885
  61. 61. Garberis I., Andre F., Lacroix-Triki M.: L’intelligence artificielle pourrait-elle intervenir dans l’aide au diagnostic des cancers du sein?–L’exemple de HER2: Could artificial intelligence play a role in breast cancer diagnosis?–The example of HER2. Bull. Cancer. 108, 11S35–11S45 (2021). https://doi.org/10.1016/S0007-4551(21)00635-4
  62. 62. Lai F.-L., Gao F.: LSA-ac4C: A hybrid neural network incorporating double-layer LSTM and self-attention mechanism for the prediction of N4-acetylcytidine sites in human mRNA. Int. J. Biol. Macromol. 253, 126837 (2023). pmid:37709212
  63. 63. Scheible R., Rusch S., Maccari M.E., Seidel M.G., Ehl S., Nieters A., et al: Json2Xlsx: Extraction and visualization of nested data in a sparse spreadsheet. Softw. Impacts. 18, 100588 (2023). https://doi.org/10.1016/j.simpa.2023.100588
  64. 64. Chung M.-H. Miles , Yang Y. Alisha , Wang L., Cento G., Jerath K., Taank P., et al: Enhancing cybersecurity situation awareness through visualization: A USB data exfiltration case study. Heliyon. 9, e13025 (2023). pmid:36820176
  65. 65. Khalifa M., Albadawy M.: Using artificial intelligence in academic writing and research: An essential productivity tool. Comput. Methods Programs Biomed. Updat. 5, 100145 (2024). https://doi.org/10.1016/j.cmpbup.2024.100145
  66. 66. <Wettengel, J.: Germany’s dependence on imported fossil fuels, https://www.cleanenergywire.org/factsheets/germanys-dependence-imported-fossil-fuels#:~:text=Germany’s largest source of domestic fossil fuel is,tonnes in 2018—and imports very little.
  67. 67. Al-Rami Al-Ghamdi B.A.M.: Analyzing the impact of data visualization applications for diagnosing the health conditions through hesitant fuzzy-based hybrid medical expert system. Ain Shams Eng. J. 15, 102705 (2024). https://doi.org/10.1016/j.asej.2024.102705
  68. 68. Shouman M.A., El-Fiky A., Hamada S., El-Sayed A., Karar M.E.: Computer-assisted lung diseases detection from pediatric chest radiography using long short-term memory networks. Comput. Electr. Eng. 103, 108402 (2022). https://doi.org/10.1016/j.compeleceng.2022.108402
  69. 69. Xu F., Cheng L., Chen J., Zhu Q.: The Nesterov accelerated gradient algorithm for Auto-Regressive Exogenous models with random lost measurements: Interpolation method and auxiliary model method. Inf. Sci. (Ny). 659, 120055 (2024). https://doi.org/10.1016/j.ins.2023.120055
  70. 70. Senogles A., Olsen M.J., Leshchinsky B.: LADI: Landslide displacement interpolation through a spatial-temporal Kalman filter. Comput. Geosci. 180, 105451 (2023). https://doi.org/10.1016/j.cageo.2023.105451
  71. 71. Calvo R., Taragan R., Rosenzweig R.: Influence of geostatistical interpolation method and cell-size resolution on CO2 storage capacity estimates: Application to saline aquifers with sparse data availability in Israel. Int. J. Greenh. Gas Control. 130, 104013 (2023). https://doi.org/10.1016/j.ijggc.2023.104013
  72. 72. Cavoretto R., De Rossi A., Dell’Accio F., Di Tommaso F., Siar N., Sommariva A., et al: Numerical cubature on scattered data by adaptive interpolation. J. Comput. Appl. Math. 444, 115793 (2024). https://doi.org/10.1016/j.cam.2024.115793
  73. 73. Zhao L., Wang Z., Chen T., Lv S., Yuan C., Shen X., et al: Missing interpolation model for wind power data based on the improved CEEMDAN method and generative adversarial interpolation network. Glob. Energy Interconnect. 6, 517–529 (2023). https://doi.org/10.1016/j.gloei.2023.10.001
  74. 74. Mehrotra K., Mohan C., Ranka S.: Elements of Artificial Neural Networks. (2019)
  75. 75. Ameri R., Hsu C.-C., Band S.S., Zamani M., Shu C.-M., Khorsandroo S.: Forecasting PM 2.5 concentration based on integrating of CEEMDAN decomposition method with SVM and LSTM. Ecotoxicol. Environ. Saf. 266, 115572 (2023). pmid:37837695
  76. 76. Acharjee P.K., Souliman M.I., Khalifah R., Elwardany M.: Frequency- and temperature-dependent dynamic shear modulus and phase angle prediction models based on existing asphalt binder viscosity data using Artificial Neural Network (ANN). Constr. Build. Mater. 414, 134772 (2024). https://doi.org/10.1016/j.conbuildmat.2023.134772
  77. 77. Abut S., Okut H., Kallail K.J.: Paradigm shift from Artificial Neural Networks (ANNs) to deep Convolutional Neural Networks (DCNNs) in the field of medical image processing. Expert Syst. Appl. 244, 122983 (2024). https://doi.org/10.1016/j.eswa.2023.122983
  78. 78. Buscema P.M., Massini G., Breda M., Lodwick W.A., Newman F., Asadi-Zeydabadi M.: Artificial neural networks. In: Studies in Systems, Decision and Control (2018)
  79. 79. Alfwzan W.F., Selim M.M., Althobaiti S., Hussin A.M.: Application of Bi-LSTM method for groundwater quality assessment through water quality indices. J. Water Process Eng. 53, 103889 (2023). https://doi.org/10.1016/j.jwpe.2023.103889
  80. 80. Rehman S., Algarni S., Imtiaz M., Alqahtani T., ElSeabee F.A.A., Jamshed W., et al: Inclined magnetic force impact on cross nanoliquid flowing with widening shallow and heat generating by using artificial neural network (ANN). Case Stud. Therm. Eng. 52, 103690 (2023). https://doi.org/10.1016/j.csite.2023.103690
  81. 81. Tealab A.: Time series forecasting using artificial neural networks methodologies: A systematic review. Futur. Comput. Informatics J. (2018). https://doi.org/10.1016/j.fcij.2018.10.003
  82. 82. Lin H., Zhang S., Li Q., Li Y., Li J., Yang Y.: A new method for heart rate prediction based on LSTM-BiLSTM-Att. Measurement. 207, 112384 (2023). https://doi.org/10.1016/j.measurement.2022.112384
  83. 83. Huang X., You Y., Zeng X., Liu Q., Dong H., Qian M., et al: Back propagation artificial neural network (BP-ANN) for prediction of the quality of gamma-irradiated smoked bacon. Food Chem. 437, 137806 (2024). pmid:37871425
  84. 84. Caterini A.L., Chang D.E.: Recurrent neural networks. In: SpringerBriefs in Computer Science (2018)
  85. 85. Rojc M., Mlakar I.: An LSTM-based model for the compression of acoustic inventories for corpus-based text-to-speech synthesis systems. Comput. Electr. Eng. 100, 107942 (2022). https://doi.org/10.1016/j.compeleceng.2022.107942
  86. 86. Nandhini K., Tamilpavai G.: Hybrid CNN-LSTM and modified wild horse herd Model-based prediction of genome sequences for genetic disorders. Biomed. Signal Process. Control. 78, 103840 (2022). https://doi.org/10.1016/j.bspc.2022.103840
  87. 87. Benkirane F.E., Crombez N., Hilaire V., Ruichek Y.: Hybrid AI for panoptic segmentation: An informed deep learning approach with integration of prior spatial relationships knowledge. Comput. Vis. Image Underst. 240, 103909 (2024). https://doi.org/10.1016/j.cviu.2023.103909
  88. 88. Ali M.A., Hossain M.S., Hossain M.K., Sikder S.S., Khushbu S.A., Islam M.: AMDNet23: Hybrid CNN-LSTM deep learning approach with enhanced preprocessing for age-related macular degeneration (AMD) detection. Intell. Syst. with Appl. 21, 200334 (2024). https://doi.org/10.1016/j.iswa.2024.200334
  89. 89. Zhang D., Shi K., Wang W., Wang X., Zhang Y., Qin B., et al: An optical mechanism-based deep learning approach for deriving water trophic state of China’s lakes from Landsat images. Water Res. 252, 121181 (2024). pmid:38301525
  90. 90. Yousif A.J., Al-Jammas M.H.: Exploring deep learning approaches for video captioning: A comprehensive review. e-Prime—Adv. Electr. Eng. Electron. Energy. 6, 100372 (2023). https://doi.org/10.1016/j.prime.2023.100372
  91. 91. Aarthy S.T., Mazher Iqbal J.L.: A novel deep learning approach for early detection of cardiovascular diseases from ECG signals. Med. Eng. Phys. 125, 104111 (2024). pmid:38508789
  92. 92. Ameri R., Hsu C.-C., Band S.S.: A systematic review of deep learning approaches for surface defect detection in industrial applications. Eng. Appl. Artif. Intell. 130, 107717 (2024). https://doi.org/10.1016/j.engappai.2023.107717
  93. 93. Mumtaz R., Maqsood M.H., Haq I.ul, Shafi U., Mahmood Z., Mumtaz M.: Integrated digital image processing techniques and deep learning approaches for wheat stripe rust disease detection and grading. Decis. Anal. J. 8, 100305 (2023). https://doi.org/10.1016/j.dajour.2023.100305
  94. 94. Chen J., Zhang Y., Wu J., Cheng W., Zhu Q.: SOC estimation for lithium-ion battery using the LSTM-RNN with extended input and constrained output. Energy. 262, 125375 (2023). https://doi.org/10.1016/j.energy.2022.125375
  95. 95. Islam M.J., Datta R., Iqbal A.: Actual rating calculation of the zoom cloud meetings app using user reviews on google play store with sentiment annotation of BERT and hybridization of RNN and LSTM. Expert Syst. Appl. 223, 119919 (2023). pmid:36969371
  96. 96. Peng F.-L., Qiao Y.-K., Yang C.: A LSTM-RNN based intelligent control approach for temperature and humidity environment of urban utility tunnels. Heliyon. 9, e13182 (2023). pmid:36798772
  97. 97. Wu Z., Guo J.: MFPD-LSTM: A steganalysis method based on multiple features of pitch delay using RNN-LSTM. J. Inf. Secur. Appl. 74, 103469 (2023). https://doi.org/10.1016/j.jisa.2023.103469
  98. 98. Kłosowski G., Hoła A., Rymarczyk T., Mazurek M., Niderla K., Rzemieniak M.: Use of the double-stage LSTM network in electrical tomography for 3D wall moisture imaging. Measurement. 213, 112741 (2023). https://doi.org/10.1016/j.measurement.2023.112741
  99. 99. Li Z., Bai J., Jiang M., Feng J., Chen X., Wei R., et al: Continuous monitoring of tissue oxygen metabolism based on multi-wavelength diffuse correlation spectroscopy using LSTM-based RNN model. Opt. Laser Technol. 171, 110384 (2024).
  100. 100. Vijay E., Aparna K.: RNN-BIRNN-LSTM based spectrum sensing for proficient data transmission in cognitive radio. e-Prime—Adv. Electr. Eng. Electron. Energy. 6, 100378 (2023). https://doi.org/10.1016/j.prime.2023.100378
  101. 101. Jiang Z., Zhang Z., He X., Li Y., Yuan H.: Efficient and accurate TEC modeling and prediction approach with random forest and Bi-LSTM for large-scale region. Adv. Sp. Res. 73, 650–662 (2024). https://doi.org/10.1016/j.asr.2023.09.003
  102. 102. Al-azazi F.A., Ghurab M.: ANN-LSTM: A deep learning model for early student performance prediction in MOOC. Heliyon. 9, e15382 (2023). pmid:37089306
  103. 103. Sherstinsky A.: Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) network. Phys. D Nonlinear Phenom. (2020). https://doi.org/10.1016/j.physd.2019.132306
  104. 104. Budak Ü., Cömert Z., Rashid Z.N., Şengür A., Çıbuk M.: Computer-aided diagnosis system combining FCN and Bi-LSTM model for efficient breast cancer detection from histopathological images. Appl. Soft Comput. 85, 105765 (2019). https://doi.org/10.1016/j.asoc.2019.105765
  105. 105. Liu H., Yang J., Chang C.-H., Wang W., Zheng H.-T., Jiang Y., et al: AOG-LSTM: An adaptive attention neural network for visual storytelling. Neurocomputing. 552, 126486 (2023). https://doi.org/10.1016/j.neucom.2023.126486
  106. 106. Gers F.A., Schmidhuber J., Cummins F.: Learning to forget: Continual prediction with LSTM. Neural Comput. (2000). pmid:11032042
  107. 107. Akila Agnes S., Anitha J., Arun Solomon A.: Two-stage lung nodule detection framework using enhanced UNet and convolutional LSTM networks in CT images. Comput. Biol. Med. 149, 106059 (2022). pmid:36087510
  108. 108. Song S., Chen J., Ma L., Zhang L., He S., Du G., et al: Research on a working face gas concentration prediction model based on LASSO-RNN time series data. Heliyon. 9, e14864 (2023). pmid:37025771
  109. 109. Cao X., Wu K., Geng X., Guan Q.: Field detection of indoor fire threat situation based on LSTM-Kriging network. J. Build. Eng. 84, 108686 (2024). https://doi.org/10.1016/j.jobe.2024.108686
  110. 110. Bao S., Tang S., Sun P., Wang T.: LSTM-based energy management algorithm for a vehicle power-split hybrid powertrain. Energy. 284, 129267 (2023). https://doi.org/10.1016/j.energy.2023.129267
  111. 111. Shaohu L., Yuandeng W., Rui H.: Prediction of drilling plug operation parameters based on incremental learning and CNN-LSTM. Geoenergy Sci. Eng. 234, 212631 (2024). https://doi.org/10.1016/j.geoen.2024.212631
  112. 112. Situ Z., Wang Q., Teng S., Feng W., Chen G., Zhou Q., et al: Improving urban flood prediction using LSTM-DeepLabv3+ and Bayesian optimization with spatiotemporal feature fusion. J. Hydrol. 630, 130743 (2024). https://doi.org/10.1016/j.jhydrol.2024.130743
  113. 113. Khan R.: Comparing the effects of agricultural intensification on CO2 emissions and energy consumption in developing and developed countries. Front. Environ. Sci. https://doi.org/10.3389/fenvs.2022.1065634
  114. 114. Sabareesh S.U., Aravind K.S.N., Chowdary K.B., Syama S., Devi V S K.: LSTM Based 24 hours Ahead Forecasting of Solar PV System for Standalone Household System. Procedia Comput. Sci. 218, 1304–1313 (2023). https://doi.org/10.1016/j.procs.2023.01.109
  115. 115. Muñoz-Organero M., Callejo P., Hombrados-Herrera M.Á.: A new RNN based machine learning model to forecast COVID-19 incidence, enhanced by the use of mobility data from the bike-sharing service in Madrid. Heliyon. 9, e17625 (2023). https://doi.org/10.1016/j.heliyon.2023.e17625
  116. 116. Elmezughi M.K., Salih O., Afullo T.J., Duffy K.J.: Path loss modeling based on neural networks and ensemble method for future wireless networks. Heliyon. 9, e19685 (2023). pmid:37809436
  117. 117. Shi H., Wei A., Xu X., Zhu Y., Hu H., Tang S.: A CNN-LSTM based deep learning model with high accuracy and robustness for carbon price forecasting: A case of Shenzhen’s carbon market in China. J. Environ. Manage. 352, 120131 (2024). pmid:38266520
  118. 118. Boda S., Mahadevappa M., Kumar Dutta P.: An automated patient-specific ECG beat classification using LSTM-based recurrent neural networks. Biomed. Signal Process. Control. 84, 104756 (2023). https://doi.org/10.1016/j.bspc.2023.104756
  119. 119. William C. Shiel Jr., MD, FACPF Medical Definition of CDC
  120. 120. Centres for Disease Control and Prevention: Mission, Role and Pledge
  121. 121. John C., Mathew O.K., Sahoo J.: CNN-LSTM based classification of polo like kinase family of Proteins: An emerging cancer drug target. Mater. Today Proc. 58, 445–450 (2022). https://doi.org/10.1016/j.matpr.2022.02.395
  122. 122. American Cancer Society: Cancer Facts and Statistics
  123. 123. Khan R.: The impact of a new techno-nationalism era on eco-economic decoupling. Resour. Policy. 82, 103452 (2023). https://doi.org/10.1016/j.resourpol.2023.103452
  124. 124. Johnson A.E., Birru Talabi M., Bonifacino E., Culyba A.J., Jonassaint N.L., Nance M.A., et al: Considerations for Racial Diversity in the Cardiology Workforce in the United States of America. J. Am. Coll. Cardiol. 77, 1934–1937 (2021). https://doi.org/10.1016/j.jacc.2021.02.043
  125. 125. Rabnawaz K.: Beta decoupling relationship between CO2 emissions by GDP, energy consumption, electricity production, value-added industries, and population in China. PLoS One. 16, (2021)
  126. 126. Lombardo E., Rabe M., Xiong Y., Nierer L., Cusumano D., Placidi L., et al: Evaluation of real-time tumor contour prediction using LSTM networks for MR-guided radiotherapy. Radiother. Oncol. 182, 109555 (2023). pmid:36813166
  127. 127. Khan S.Z.R.: Performance evaluation of e-commerce firms in China: Using three-stage data envelopment analysis and the Malmquist productivity index. PLoS One. 11, (2021). https://doi.org/10.1371/journal.pone.0255851
  128. 128. Khan R.: Agricultural production and CO2 emissions causes in the developing and developed countries: New insights from quantile regression and decomposition analysis. bioRxiv. (2020)
  129. 129. Sarvari A.V.P., Sridevi K.: An optimized EBRSA-Bi LSTM model for highly undersampled rapid CT image reconstruction. Biomed. Signal Process. Control. 83, 104637 (2023). pmid:36776947
  130. 130. Tableau: New Data Modeling Capabilities
  131. 131. Öztürk Ş., Özkaya U.: Residual LSTM layered CNN for classification of gastrointestinal tract diseases. J. Biomed. Inform. 113, 103638 (2021). pmid:33271341
  132. 132. Poddar A.K., Rani D.R.: Hybrid Architecture using CNN and LSTM for Image Captioning in Hindi Language. Procedia Comput. Sci. 218, 686–696 (2023). https://doi.org/10.1016/j.procs.2023.01.049
  133. 133. Tang X., Zheng P., Li X., Wu H., Wei D.-Q., Liu Y., et al: Deep6mAPred: A CNN and Bi-LSTM-based deep learning method for predicting DNA N6-methyladenosine sites across plant species. Methods. 204, 142–150 (2022). pmid:35477057
  134. 134. Dong X., Wong R., Lyu W., Abell-Hart K., Deng J., Liu Y., et al: An integrated LSTM-HeteroRGNN model for interpretable opioid overdose risk prediction. Artif. Intell. Med. 135, 102439 (2023). pmid:36628797
  135. 135. Vankdothu R., Hameed M.A., Fatima H.: A Brain Tumor Identification and Classification Using Deep Learning based on CNN-LSTM Method. Comput. Electr. Eng. 101, 107960 (2022). https://doi.org/10.1016/j.compeleceng.2022.107960
  136. 136. Khan R.: Beta decoupling relationship between CO2 emissions by GDP, energy consumption, electricity production, value-added industries, and population in China. PLoS One. 16, 1–22 (2021). pmid:33793639
  137. 137. Sivamurugan J., Sureshkumar G.: Applying dual models on optimized LSTM with U-net segmentation for breast cancer diagnosis using mammogram images. Artif. Intell. Med. 143, 102626 (2023). pmid:37673584
  138. 138. Asim M.N., Ibrahim M.A., Malik M.I., Zehe C., Cloarec O., Trygg J., et al: EL-RMLocNet: An explainable LSTM network for RNA-associated multi-compartment localization prediction. Comput. Struct. Biotechnol. J. 20, 3986–4002 (2022). pmid:35983235
  139. 139. Neeraj Singhal, Mathew V, J., Behera, R.K.: Detection of alcoholism using EEG signals and a CNN-LSTM-ATTN network. Comput. Biol. Med. 138, 104940 (2021). pmid:34656864
  140. 140. Zhao L., Xu W., Gao S., Guo J.: Cross-sentence N-ary relation classification using LSTMs on graph and sequence structures. Knowledge-Based Syst. 207, 106266 (2020). https://doi.org/10.1016/j.knosys.2020.106266
  141. 141. Sakalle A., Tomar P., Bhardwaj H., Acharya D., Bhardwaj A.: A LSTM based deep learning network for recognizing emotions using wireless brainwave driven system. Expert Syst. Appl. 173, 114516 (2021). https://doi.org/10.1016/j.eswa.2020.114516
  142. 142. Lai H., Liu Q., Ye Q., Liang Z., Long Z., Hu Y., et al: Impact of smoking cessation duration on lung cancer mortality: A systematic review and meta-analysis. Crit. Rev. Oncol. Hematol. 196, 104323 (2024). pmid:38462148
  143. 143. Wu J., Gao F., Meng R., Li H., Mao Z., Xiao Y., et al: Single-cell and multi-omics analyses highlight cancer-associated fibroblasts-induced immune evasion and epithelial mesenchymal transition for smoking bladder cancer. Toxicology. 504, 153782 (2024). pmid:38493947
  144. 144. Zhang X., Li H., Sekar M., Elgendi M., Krishnamoorthy N.R., Xia C., et al: Machine learning algorithms for a diesel engine fuelled with biodiesel blends and hydrogen using LSTM networks. Fuel. 333, 126292 (2023). https://doi.org/10.1016/j.fuel.2022.126292
  145. 145. Rashidian H., Hadji M., Ansari-Moghaddam A., Bakhshi M., Nejatizadeh A., Marzban M., et al: Association between waterpipe smoking and lung cancer: a multicentre case–control study in Iran. Public Health. 231, 166–172 (2024). pmid:38701659
  146. 146. de Nijs K., ten Haaf K., van der Aalst C., de Koning H.J.: Projected effectiveness of lung cancer screening and concurrent smoking cessation support in the Netherlands. eClinicalMedicine. 71, 102570 (2024). pmid:38813448
  147. 147. Shukla N., Shah K., Rathore D., Soni K., Shah J., Vora H., et al: Androgen receptor: Structure, signaling, function and potential drug discovery biomarker in different breast cancer subtypes. Life Sci. 348, 122697 (2024). pmid:38710280
  148. 148. Jang B.-S., Chun S.-J., Choi H.S., Chang J.H., Shin K.H.: Estimating the risk and benefit of radiation therapy in (y)pN1 stage breast cancer patients: A Bayesian network model incorporating expert knowledge (KROG 22–13). Comput. Methods Programs Biomed. 245, 108049 (2024). https://doi.org/10.1016/j.cmpb.2024.108049
  149. 149. Sasada S., Nagura N., Shimo A., Ogiya A., Saiga M., Seki H., et al: Impact of radiation therapy for breast cancer with involved surgical margin after immediate breast reconstruction: A multi-institutional observational study. Eur. J. Surg. Oncol. 50, 108360 (2024). pmid:38669780
  150. 150. Mursalim M.K.N., Mengko T.L.E.R., Hertadi R., Purwarianti A., Susanty M.: BiCaps-DBP: Predicting DNA-binding proteins from protein sequences using Bi-LSTM and a 1D-capsule network. Comput. Biol. Med. 163, 107241 (2023). pmid:37437362
  151. 151. Ortmann N., Haddad Y.K., Beck L.: Special Report from the CDC: Provider knowledge and practices around driving safety and fall prevention screening and recommendations for their older adult patients, DocStyles 2019. J. Safety Res. 86, 401–408 (2023). pmid:37718068
  152. 152. Ayenew L.G., Hoelscher M.A., Emshoff J.G., Kidder D.P., Ellis B.A.: Evaluation of the public health achievements made by projects supported by a federal contract mechanism at the Centers for Disease Control and Prevention (CDC), USA. Eval. Program Plann. 88, 101949 (2021). https://doi.org/10.1016/j.evalprogplan.2021.101949
  153. 153. Jie W.; Khan R.: Breaking the CO2 Gridlock: Can Renewables Lead the Way for the OECD? Energies. 17, (2024). https://doi.org/10.3390/en17174511
  154. 154. Gohar A., Ali A.A., Elkhatib W.F., El-Sayyad G.S., Elfadil D., Noreddin A.M.: Combination therapy between prophylactic and therapeutic human papillomavirus (HPV) vaccines with special emphasis on implementation of nanotechnology. Microb. Pathog. 171, 105747 (2022). pmid:36064102
  155. 155. Terada M., Shimazu T., Saito J., Odawara M., Otsuki A., Yaguchi-Saito A., et al: Age, gender and socioeconomic disparities in human papillomavirus (HPV) awareness and knowledge among Japanese adults after a 7-year suspension of proactive recommendation for the HPV vaccine: A nationally representative cross-sectional survey. Vaccine. 41, 7147–7158 (2023). pmid:37866996
  156. 156. Nawaz K.: The implications of internet-based Chinese language courses on online classes. Front. Psychol. 14, (2024). https://doi.org/10.3389/fpsyg.2023.1203136
  157. 157. Lu Y., Aslani S., Zhao A., Shahin A., Barber D., Emberton M., et al: A hybrid CNN-RNN approach for survival analysis in a Lung Cancer Screening study. Heliyon. 9, e18695 (2023). pmid:37600411
  158. 158. Rauniyar S.K., Hashizume M., Yoneoka D., Nomura S.: Projection of morbidity and mortality due to breast cancer between 2020 and 2050 across 42 low- and middle-income countries. Heliyon. 9, e16427 (2023). pmid:37274661
  159. 159. Escorcia-Gutierrez J., Gamarra M., Ariza-Colpas P.P., Roncallo G.B., Leal N., Soto-Diaz R., et al: Galactic swarm optimization with deep transfer learning driven colorectal cancer classification for image guided intervention. Comput. Electr. Eng. 104, 108462 (2022). https://doi.org/10.1016/j.compeleceng.2022.108462
  160. 160. Khan R.: Data set-GANN Principle
  161. 161. Langhi Prata LGP, Tchkonia T.K.J, .: Cell senescence, the senescence-associated secretory phenotype, and cancers. PLoS Biol. 21, (2023). https://doi.org/10.1371/journal.pbio.3002326
  162. 162. Hussain J., Khan A., Zhou K.: The impact of natural resource depletion on energy use and CO2 emission in Belt & Road Initiative countries: A cross-country analysis. Energy. 199, 117409 (2020). https://doi.org/10.1016/j.energy.2020.117409
  163. 163. Sujatha K., Bhavani N.P.G., Srividhya V., Karthikeyan V., Jayachitra N.: Chapter 6—Soft sensor with shape descriptors for flame quality prediction based on LSTM regression. In: Das H., Dey N., and Emilia Balas V.B.T.-R.-T.D.A. for L.S.S.D (eds.) Advances in Ubiquitous Sensing Applications for Healthcare. pp. 115–138. Academic Press (2020)
  164. 164. Chen L., Liao H., Zhu J., Li K., Bai Y., Yue X., et al: Increases in ozone-related mortality in China over 2013–2030 attributed to historical ozone deterioration and future population aging. Sci. Total Environ. 858, 159972 (2023). pmid:36356763
  165. 165. Han Y., Lam J.C.K., Li V.O.K., Reiner D.: A Bayesian LSTM model to evaluate the effects of air pollution control regulations in Beijing, China. Environ. Sci. Policy. 115, 26–34 (2021). https://doi.org/10.1016/j.envsci.2020.10.004
  166. 166. Ju B., Liu Y., Hu X., Zhao D., Jiang L.: A novel cell contour-based instance segmentation model and its applications in HER2 breast cancer discrimination. Biomed. Signal Process. Control. 85, 104941 (2023). https://doi.org/10.1016/j.bspc.2023.104941
  167. 167. Atrey K., Singh B.K., Bodhey N.K., Bilas Pachori R.: Mammography and ultrasound based dual modality classification of breast cancer using a hybrid deep learning approach. Biomed. Signal Process. Control. 86, 104919 (2023). https://doi.org/10.1016/j.bspc.2023.104919
  168. 168. Bi Z.; Guo R.; Khan R.: Renewable Adoption, Energy Reliance, and CO2 Emissions: A Comparison of Developed and Developing Economies. Energies. 17, (2024). https://doi.org/10.3390/en17133111
  169. 169. Nguyen K.H., Smith A.K., Kim B.T., Browne A.W.: A case report of cystoid macular edema, uveitis and vitreomacular traction in a patient taking Anastrozole. Am. J. Ophthalmol. Case Reports. 25, 101339 (2022). https://doi.org/10.1016/j.ajoc.2022.101339
  170. 170. Wankhade S., S., V.: A novel hybrid deep learning method for early detection of lung cancer using neural networks. Healthc. Anal. 3, 100195 (2023). https://doi.org/10.1016/j.health.2023.100195
  171. 171. Sharma S., Gupta V., Mudgal D., Srivastava V.: Predicting biomechanical properties of additively manufactured polydopamine coated poly lactic acid bone plates using deep learning. Eng. Appl. Artif. Intell. 124, 106587 (2023). https://doi.org/10.1016/j.engappai.2023.106587
  172. 172. Xu P., Gao Q., Zhang Z., Zhao K.: Multi-source data based anomaly detection through temporal and spatial characteristics. Expert Syst. Appl. 237, 121675 (2024). https://doi.org/10.1016/j.eswa.2023.121675
  173. 173. Ramprasad M.V.S., Nagesh S.S. V, Sahith V., Lankalapalli R.K.: Hierarchical agglomerative clustering-based skin lesion detection with region based neural networks classification. Meas. Sensors. 29, 100865 (2023). https://doi.org/10.1016/j.measen.2023.100865
  174. 174. Tutorial Point: Tableau—Forecasting
  175. 175. Aksangür İ., Eren B., Erden C.: Evaluation of data preprocessing and feature selection process for prediction of hourly PM10 concentration using long short-term memory models. Environ. Pollut. 311, 119973 (2022). pmid:35985430
  176. 176. Lipton Z.C., Kale D.C., Elkan C., Wetzel R.: Learning to diagnose with LSTM recurrent neural networks. In: 4th International Conference on Learning Representations, ICLR 2016—Conference Track Proceedings (2016)
  177. 177. Sabique P. V, Pasupathy G., Ramachandran S., Shanmugasundar G.: Investigating the influence of dimensionality reduction on force estimation in robotic-assisted surgery using recurrent and convolutional networks. Eng. Appl. Artif. Intell. 126, 107045 (2023). https://doi.org/10.1016/j.engappai.2023.107045
  178. 178. Zhang S., Zhang X., Du J., Wang W., Pi X.: Multi-target meridians classification based on the topological structure of anti-cancer phytochemicals using deep learning. J. Ethnopharmacol. 319, 117244 (2024). pmid:37777031
  179. 179. Bello I., Zoph B., Le Q., Vaswani A., Shlens J.: Attention augmented convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision (2019)
  180. 180. Eren B., Aksangür İ., Erden C.: Predicting next hour fine particulate matter (PM2.5) in the Istanbul Metropolitan City using deep learning algorithms with time windowing strategy. Urban Clim. 48, 101418 (2023). https://doi.org/10.1016/j.uclim.2023.101418
  181. 181. Ramachandran P., Parmar N., Vaswani A., Bello I., Levskaya A., Shlens J.: Stand-Alone Self-Attention in Vision Models. In: 33rd Conference on Neural Information Processing Systems. pp. 1–13., Vancouver, Canada (2019)
  182. 182. Suddle M.K., Bashir M.: Metaheuristics based long short term memory optimization for sentiment analysis. Appl. Soft Comput. 131, 109794 (2022). https://doi.org/10.1016/j.asoc.2022.109794
  183. 183. Neog H., Dutta P.E., Medhi N.: Health condition prediction and covid risk detection using healthcare 4.0 techniques. Smart Heal. 26, 100322 (2022). https://doi.org/10.1016/j.smhl.2022.100322
  184. 184. Jamil I., Lucheng H., Iqbal S., Aurangzaib M., Jamil R., Kotb H., et al: Predictive evaluation of solar energy variables for a large-scale solar power plant based on triple deep learning forecast models. Alexandria Eng. J. 76, 51–73 (2023). https://doi.org/10.1016/j.aej.2023.06.023
  185. 185. Zhang R., He X., Liu J., Xiong J.: VOC transport in an occupied residence: Measurements and predictions via deep learning. Sci. Total Environ. 892, 164559 (2023). pmid:37263430
  186. 186. Yang D., Wang J., Yan X., Liu H.: Subway air quality modeling using improved deep learning framework. Process Saf. Environ. Prot. 163, 487–497 (2022). https://doi.org/10.1016/j.psep.2022.05.055