Retraction
The PLOS One Editors retract this article [1] because it was identified as one of a series of submissions for which we have concerns about peer review integrity and potential manipulation of the publication process. These concerns call into question the validity and provenance of the reported results. We regret that the issues were not identified prior to the article’s publication.
All authors did not agree with the retraction.
9 Feb 2026: The PLOS One Editors (2026) Retraction: N-Beats architecture for explainable forecasting of multi-dimensional poultry data. PLOS ONE 21(2): e0342381. https://doi.org/10.1371/journal.pone.0342381 View retraction
Figures
Abstract
The agricultural economy heavily relies on poultry production, making accurate forecasting of poultry data crucial for optimizing revenue, streamlining resource utilization, and maximizing productivity. This research introduces a novel application of the N-BEATS architecture for multi-dimensional poultry data forecasting with enhanced interpretability through an integrated Explainable AI (XAI) framework. Leveraging its advanced capabilities in time series modeling, N-BEATS is applied to predict multiple facets of poultry disease diagnostics using a multivariate dataset comprising key environmental parameters. The methodology empowers decision-making in poultry farm management by providing transparent and interpretable forecasts. Experimental results demonstrate that N-BEATS outperforms conventional deep learning models, including LSTM, GRU, RNN, and CNN, across various error metrics, achieving MAE of 0.172, RMSE of 0.313, MSLE of 0.042, R-squared of 0.034, and RMSLE of 0.204. The positive R-squared value indicates the model’s robustness against underfitting and overfitting, surpassing the performance of other models with negative R-squared values. This study establishes N-BEATS as a superior and interpretable solution for complex, multi-dimensional forecasting challenges in poultry production, with significant implications for enhancing predictive analytics in agriculture.
Citation: Kaur B, Rakhra M, Sharma N, Prashar D, Mrsic L, Khan AA, et al. (2025) RETRACTED: N-Beats architecture for explainable forecasting of multi-dimensional poultry data. PLoS ONE 20(4): e0320979. https://doi.org/10.1371/journal.pone.0320979
Editor: Ayman A. Swelum, King Saud University /Zagazig University, EGYPT
Received: October 3, 2024; Accepted: February 27, 2025; Published: April 24, 2025
Copyright: © 2025 Kaur et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: Data Availability Statement: The dataset used for this work is publicly available: https://github.com/ravi95/Dataset/blob/main/poultry_data%20(1).csv.
Funding: The author(s) received no specific funding for this work.
Competing interests: The authors have declared that no competing interests exist.
I. Introduction
This Poultry farming plays a significant role in bolstering a country’s agricultural economy. Further, accurate prediction of poultry data holds immense significance, where poultry farming assumes a substantial position within the agricultural economy. Accurate forecasting plays a pivotal role in the optimization of production processes, effective resource management, and the maximization of revenue. The necessity for accurate forecasting within the poultry industry can be supported by a multitude of significant factors and empirical data [1].
Based on the findings reported by the Punjab (Northern State in India) State Poultry Farmers Association, it has been determined that the poultry industry in the state of Punjab exhibits a significant economic impact, with an estimated annual revenue exceeding INR 10,000 crore, equivalent to approximately USD 1.3 billion. In the context of Punjab, it is noteworthy that the consumption of broilers stands at an estimated 120 lakh (12 million) per month [2]. This substantial figure underscores the necessity for accurate forecasting methodologies to effectively cater to consumer demands. Accurate forecasting plays a crucial role in facilitating the strategic planning of poultry feed procurement and other vital resources. The poultry industry in Punjab exhibits a substantial daily consumption of feed, surpassing 25,000 metric tons [3]. This figure underscores the criticality of implementing effective resource management strategies within the sector. Further, Punjab region experiences recurring episodes of avian influenza, which highlights the need for a strong and reliable forecasting system to effectively mitigate the adverse effects associated with these outbreaks. Lastly, accurate forecasts play a crucial role in optimizing operational efficiency as they facilitate the streamlining of production schedules, workforce allocation, and logistics. This can be validated from the data of Punjab, where an extensive number of approximately 25,000 poultry farms exist, necessitating meticulous attention to detail in terms of strategic organization and operational oversight [4].
Certain diseases impact the bird population more severely than others and have more dire repercussions. The first of them, pullorum-typhoid illness, was so serious that the National Poultry Improvement Plan (NPIP) had to be developed in response. As a result of the NPIP’s extremely effective proactive attempts to curb the disease, it has almost completely disappeared. Even if pullorum-typhoid illness is no longer a major concern, other diseases continue to pose a threat to the world’s chicken population. Among them are two: exotic Newcastle disease and avian influenza. Knowledge of these illnesses can help flock owners safeguard their birds and advance improved animal health [5].
The N-BEATS (Neural Basis Expansion Analysis for Time Series Forecasting) architecture is a state-of-the-art framework in time series forecasting, with specific use cases that emphasize the assessment and forecasting of Early-Age Incidence (EAI) metrics [6]. Due to the complex temporal connections and dynamic nature of the underlying processes, time series forecasting is a considerable challenge, particularly in the context of poultry data. Through the use of stacked, fully connected blocks that can be adjusted to various EAI forecasting scenarios, N-BEATS addresses this issue and improves predicted accuracy and robustness. The forecasting of EAI indicators, which include crucial elements like mortality rates, growth patterns, and disease outbreaks, temperature, humidity among others, is extremely important to the poultry business. The optimization of production techniques, resource distribution, and protecting the general health and welfare of the poultry population all depend on accurate forecasts in this area. By utilizing deep neural networks to recognize complex patterns and dependencies within time series data for the poultry industry, N-BEATS provides a strong foundation to meet these forecasting requirements [7]. This research delves into the N-BEATS architecture, specifically focusing on its application in forecasting poultry EAI data. The study involves a thorough examination of the various components that make up the N-BEATS architecture, such as the stacked FC-Blocks, considerations regarding width and horizon, and the ensemble learning capabilities. This exposition endeavors to offer a comprehensive understanding to stakeholders in the poultry industry, data scientists, and researchers regarding the utilization of N-BEATS as an advanced solution to improve the accuracy and dependability of EAI forecasts in the poultry sector. The manuscript presents the related work in the subsequent section. Further, the architecture is detailed in the methodology section. Results and discussion are presented next. Lastly, the conclusion and future scope of the present work is provided.
Interpretable forecasting refers to the capacity to comprehend and transparently explain the outputs generated by predictive models. It encompasses methodologies that allow users to discern the rationale behind the model’s predictions, the contributing variables, and the associated confidence levels. This is particularly critical in domains where decisions rely on anticipated outcomes, as understanding the logic behind predictions enhances trust in the model and its outputs.
The N-BEATS (Neural Basis Expansion Analysis for Time Series) model represents a state-of-the-art neural architecture for time series forecasting. Unlike traditional time series models that leverage recurrent (e.g., LSTMs, GRUs) or convolutional layers, N-BEATS employs a fully connected feedforward architecture. The model is structured into a sequence of blocks, where each block specializes in modeling specific time series patterns, such as trends and seasonality. These blocks iteratively refine predictions using a residual stacking mechanism.
Each block in the N-BEATS architecture comprises two key components:
- Backcast: This reconstructs the input time series by modeling already captured patterns. The backcast output is subtracted from the input, allowing subsequent blocks to focus on the residual errors.
- Forecast: This generates predictions for future time series values by identifying and extrapolating patterns present in the data.
To model the inherent patterns in time series data, the blocks utilize basis expansion techniques, such as polynomials or Fourier series, which facilitate the decomposition of complex signals into interpretable components. The architecture’s residual stacking ensures that each block refines the forecast by minimizing the residuals from the previous blocks, leading to improved accuracy.
A defining characteristic of N-BEATS is its interpretable architecture. The blocks can be explicitly designed to capture distinct components of the time series, such as long-term trends or periodic fluctuations, making the outputs more transparent and explainable. This modularity and interpretability are particularly advantageous in applications where understanding the model’s reasoning is as critical as the predictions themselves.
II. Related work
To fulfil the growing demand for poultry industry, it is imperative to forecast India’s poultry output. Several researchers are working in this domain. This section presents the state of art work related to poultry disease forecasting. India’s chicken business has grown steadily because of increased demand and advancements in agricultural techniques. However, production levels are subject to change and are impacted by a number of variables, including epidemics of disease [8]. In the conventional poultry farming industry, information about poultry and the breeding environment is sensed by manual observation and empirical judgement. But as the scale of breeding keep growing, these labor-and time-intensive manual approaches are becoming more and more expensive. Furthermore, it is impossible to guarantee the regularity and accuracy of information sensing, which severely restricts the effectiveness of poultry information sensing and the advancement of precision farming methods [9]. Due to the widespread recognition of poultry as an excellent source of protein, several nations are going to be compelled to raise production, which will lead to a surge in the total number of farms housing birds at high densities. In addition, the production of eggs and meat from poultry farming contributes to food and nutrition security at the household, regional, and national levels, making it essential to the socioeconomic development of emerging nations. Moreover, it boosts a nation’s gross domestic product by more than $100 million and gives the populace cash income [10–13].
Over the past few decades, significant advancements have been made in the domains of computer vision, image processing, and recognition of patterns. These days, industries and researchers are using deep learning and its sophisticated algorithms to handle a wide range of problems, from basic item detection to intricate scene analysis [14]. When it comes to infectious disease, it is crucial to identify the exact time and location of infection on chicken farms so that prompt action can be done to stop further infection and losses. This is crucial for human safety as well as production gains, as some poultry infections, such avian influenza virus (AIV), carry a high risk of infection and potential pandemics in human populations [15–18].
Gizachew et. al. recorded the incidence rate of illness and death in smallholder poultry farms in a few Sidama Region areas. According to the study’s findings, a significant barrier to raising chicken production in the research areas seems to be infectious diseases. The design of poultry development initiatives and the establishment of systems for livestock production and health monitoring should thus take this into consideration, according to the veterinary and livestock authorities [19]. Further Jake et.al. reviewed several technologies that can help with the accurate and timely detection and diagnosis of illnesses in poultry. When sickness is discovered, quick action may be taken with the help of rapid detection and diagnostic methods, which reduces the risk of additional bird transmission and related expenses. Furthermore, data from quick disease detection systems can be used to inform decision support systems that forecast the location and timing of poultry disease outbreaks [20–23]. On the similar grounds, Arnas et. al. suggested the amalgamation of sophisticated computer vision methods and artificial intelligence as a considerable potential for non-invasive health evaluations in the poultry sector [24–26].
Several researchers have worked in the area of poultry diagnostics, and it can be safely concluded that the poultry industry needs the proliferation of AI based tools to predict the future incidences of the diseases and hence which can be used for prevention and thus leading to higher profits. Most of the researchers have worked on the single variate time series. Forecasting for disease diagnostics. The research gap identified is the application of multivariate dataset for the better forecasting of poultry disease and applying the explainable forecasting techniques to interpret the results. Therefore, the research presented in this paper focuses on the implementation of deep learning-based models for the multivariate dataset and providing the XAI framework for the interpretation of result.
III. Proposed framework
The ability to comprehend and transparently explain the forecasts produced by forecasting models is known as interpretable forecasting. It includes the methods that let users understand the reasoning behind a model’s prediction, the variables that shaped it, and the degree of confidence the model has in its forecasts. This is especially crucial in fields where decisions must be made in light of anticipated results, since knowing the logic behind the projections may foster confidence in the model and its predictions. The proposed framework employs N-BEATS, which is demonstrated in Fig 1. N-BEATS uses stacks of fully connected layers to model time series data, differing from other time series models that use recurrent or convolutional layers. The model consists of a series of blocks, each providing an ensemble of forecasts.
These blocks are structured in a way to incrementally capture different time series patterns like trends and seasonality. Each block in the architecture uses basis expansion techniques, like polynomials or Fourier series, to model the underlying patterns within the time series. Each block outputs two things: a forecast for the future values of the series and a backcast that reconstructs the input series. The backcast helps in removing already captured patterns from the data, enabling the subsequent blocks to focus on learning the residuals. The architecture uses a residual stacking approach where each block refines the forecast by focusing on the residuals left by the previous blocks. A key feature of N-BEATS is its interpretability. Each block can potentially be designed to capture specific types of patterns, making the model outputs more explainable. In the context of poultry data, which is likely multi-dimensional (involving various metrics like temperature, humidity, feed consumption, growth rates, etc.), N-BEATS can be adapted in the following ways.
- (1) Multiple Time Series
Each dimension of the poultry data can be treated as a separate time series, and N-BEATS can be used to model each one individually or jointly. The input layer is accommodated to handle multiple input variables. To extend the N-BEATS model to a multivariate setting with 3 independent variables, denoted as().The basis of N-beats architecture are the blocks, which contains fully connected layers. It consists of forecast and backcast outputs viz. eq. (1) and eq. (2)
Further, N-Beats uses Basis expansion for the linear decomposition of the back cast and forecast outputs. This is done for the better mapping of the components of the time series components viz. patterns, trend and seasonality. The basis expansion is represented by eq. (3) and eq. (4)
At the end, N-beats employ the stack architecture which is stack of various blocks where every block represents different aspect of the time series such as pattern, trend and seasonality. The output function is the aggregation of the output as demonstrated in eq. (5). The objective of the training function is to minimize the loss during the training round as demonstrated in eq. (6). The definition of the all the symbols used in the above equations are given in Table 1.
- (2) Feature Decomposition: The model can decompose the poultry data into trend, seasonality, and any poultry-specific cycles, which would help in understanding and forecasting the data more effectively.
- (3) Explanatory Variables: Since there are external factors that affect the poultry metrics, such as weather conditions or market prices, these can be incorporated into the N-BEATS model to improve the quality and explain ability of the forecasts. Custom blocks can be designed for specific patterns or cycles that are unique to poultry data, enhancing the model’s ability to forecast and explain these patterns. BEATS provides a robust and flexible framework for forecasting time series data, and with appropriate adaptation, it can be an effective tool for making explainable forecasts in the context of multi-dimensional poultry data.
The superior performance of the N-BEATS model in time series forecasting, including its application to multi-dimensional poultry data, can be attributed to the following key factors:
- 4. Architecture Design
- Fully Connected Layers Instead of RNNs or CNNs.
- Unlike recurrent networks (LSTM, GRU) or convolutional networks (CNNs), which are inherently sequential and rely on memory-based or localized filters, N-BEATS uses fully connected layers. This ensures that:
- All input information is globally accessible to the network.
- The model can better capture long-range dependencies in time series data.
- Unlike recurrent networks (LSTM, GRU) or convolutional networks (CNNs), which are inherently sequential and rely on memory-based or localized filters, N-BEATS uses fully connected layers. This ensures that:
- Residual Stacking Mechanism.
- Each block of N-BEATS focuses on modeling residuals left by previous blocks. This iterative refinement approach ensures:
- The model incrementally captures finer details of the data patterns, including trend, seasonality, and residual noise.
- Enhanced robustness to noise and complex dynamics in multivariate time series.
- Each block of N-BEATS focuses on modeling residuals left by previous blocks. This iterative refinement approach ensures:
- Fully Connected Layers Instead of RNNs or CNNs.
- 5. Interpretability Features
- Decomposition of Patterns.
- N-BEATS explicitly models different components of the time series, such as:
- Trends: Captures long-term directional changes.
- Seasonality: Identifies repeating patterns.
- Residuals: Focuses on unexplained components.
- N-BEATS explicitly models different components of the time series, such as:
- Basis Expansion.
- The use of basis expansion functions (e.g., polynomials and Fourier series) facilitates a linear decomposition of the forecast and backcast outputs:
- These basis functions effectively map time series components, enabling the model to decompose the data into interpretable sub-components.
- This approach aids in understanding how the model arrives at its predictions.
- The use of basis expansion functions (e.g., polynomials and Fourier series) facilitates a linear decomposition of the forecast and backcast outputs:
- Decomposition of Patterns.
- 6. Mathematical Foundation
- Forecast and Backcast Outputs.
- Each block outputs:
- Backcast (X^\hat{X}X^): Reconstructs the input time series by capturing already identified patterns, enabling the removal of these patterns for subsequent analysis.
- Forecast (Y^\hat{Y}Y^): Predicts future values based on refined residual patterns.
- This dual-output mechanism ensures:
- No overlap or redundancy between blocks.
- Effective handling of noise and outliers in data.
- Each block outputs:
- Aggregated Stacking.
- The outputs from multiple blocks are aggregated (YfinalY_{\text{final}}Yfinal) to form the final forecast. This aggregation improves:
- Model accuracy by averaging out errors.
- Robustness through ensemble-like behavior.
- The outputs from multiple blocks are aggregated (YfinalY_{\text{final}}Yfinal) to form the final forecast. This aggregation improves:
- Forecast and Backcast Outputs.
- 7. Loss Function Optimization
- The training objective minimizes the squared error loss (LLL), ensuring:
- High fidelity in backcast reconstruction.
- Accurate forecast generation.
- By optimizing over all dimensions (nnn) and time steps (ttt), the model achieves a balanced fit across the dataset.
- The training objective minimizes the squared error loss (LLL), ensuring:
- 8. Adaptability to Multivariate Data
- Handling Multiple Inputs.
- The input layer accommodates multiple variables, allowing the model to jointly process dimensions like temperature, humidity, and feed consumption.
- Incorporating Explanatory Variables.
- External factors (e.g., weather, market prices) can be integrated into the architecture, enhancing the explanatory power and predictive accuracy.
- Feature-Specific Blocks.
- Custom blocks can be designed for poultry-specific cycles or unique patterns, further improving the forecasting relevance.
- Handling Multiple Inputs.
- 9. Robustness Against Overfitting and Underfitting
- The residual stacking ensures that each block only learns the unexplained portion of the data, which prevents:
- Overfitting to noise or spurious patterns.
- Underfitting due to insufficient modeling capacity.
- The positive R-squared values compared to negative values from other models highlight this robustness.
- The residual stacking ensures that each block only learns the unexplained portion of the data, which prevents:
- 10. Generalization Across Time Horizons
- N-BEATS is highly flexible in handling varying forecast horizons (FFF):
- Short-term predictions benefit from fine-grained residual modeling.
- Long-term forecasts leverage aggregated trend and seasonality components.
- N-BEATS is highly flexible in handling varying forecast horizons (FFF):
In conclusion, N-BEATS outperforms traditional deep learning models due to its unique combination of residual stacking, basis expansion, interpretability, and adaptability. These factors enable it to produce highly accurate, robust, and explainable forecasts, making it particularly effective for multi-dimensional and domain-specific data like poultry farming metrics.
IV. Materials and methods
The research employed utilizes several deep learning methods to check the applicability on multivariate dataset and to make the comparative performance analysis with respect to the proposed framework. The various deep learning models employed are explained as under:
(1) LSTM (Long short-term memory)
Three gates make up the framework of LSTMs: input, forget, and output gates. The cell state preserves long-term relationships in the model by allowing data to flow through the network unaltered with only small linear interactions. Because the gates regulate the information’s entry and exit from the cell state, long-term support vector machines (LSTMs) can learn from data sequences. Through these gates, complicated nonlinear transformations are typically incorporated into LSTM units. The formula for these gates is as follows: tanh stands for the hyperbolic tangent function, and σ for the sigmoid function. Following equations demonstrate the working of various gates. Eq (7) describes the forget gate that decides the information to be discarded from the cell state. Eq (8) represents the input gate that decides which information is to be updates. Eq (9) is the output gate which decides what next hidden state should be. Eq (10) & Eq (11) determines the cell candidate and cell state respectively. Eq (12) determines the hidden state.
(2) GRU (Gated Recurrent Unit)
GRUs combine the forget and input gates into a single “update gate” and combine the hidden state and cell state, simplifying the LSTM design. This leads to a simpler model with fewer tensor operations, which may lessen training time and computational load. It consists of two gates; Update Gate to decide the amount of the past information that needs to be passed along to the future and secondly Reset Gate to determine the amount of the past information to forget.
(3) Simple RNN
The various steps followed in the execution of the model architecture are elaborated below:
The most basic type of RNN, which has a gateless fundamental architecture and may cause problems with vanishing and exploding gradients. The input received at the present state and the hidden state from earlier are used to update the hidden state:
(4) CNN (Convolutional Neural Network)
In order to capture temporal patterns in sequence data, 1D CNNs apply convolutional layers and slide filters throughout the sequence. Every convolutional layer uses a nonlinear activation function and the convolution process to alter the input data. To minimize dimensionality, pooling layers are frequently added afterward:
V. Methodology
The section demonstrates the steps followed in methodology of the framework:
1. Dataset overview
The dataset utilized in this study is multivariate, encompassing four distinct time-series variables: temperature, humidity, poultry feed, and disease outbreaks. Among these, disease outbreak serves as the dependent variable, while temperature, humidity, and poultry feed are the independent predictor variables (Fig 2).
2. Data characteristics
- Temporal Resolution: The dataset contains hourly observations spanning from January to an unspecified end date.
- Data Size: The dataset comprises a total of 35,126 records, providing a comprehensive and granular view of the temporal dynamics within poultry operations.
- Multivariate Representation: The inclusion of both environmental (temperature, humidity) and operational (feed usage, disease outbreaks) variables ensures robust modeling potential for analyzing interdependencies critical to poultry health and productivity.
3. Applicability in poultry industry
- Disease Prediction: The dataset enables predictive modeling for early detection of disease outbreaks, facilitating proactive intervention strategies to mitigate risks to poultry health.
- Environmental Impact Analysis: Understanding the influence of environmental factors (temperature and humidity) on poultry feed consumption and disease occurrence can optimize farm management practices.
- Resource Optimization: By analyzing feed consumption patterns in relation to environmental conditions, the dataset supports efficient resource utilization and cost-effective feeding strategies.
- Operational Insights: The dataset’s high temporal resolution allows for fine-grained monitoring, enabling real-time decision-making to enhance poultry farm operations.
The dataset’s rich multivariate nature and high temporal granularity make it a critical resource for deploying advanced data-driven solutions in the poultry industry, enhancing operational efficiency, health management, and predictive analytics.
4. Data preprocessing
Observing the timeseries of the data, the first major issue is the lack of a uniformly sampled temporal space. Hence, the first step is to make the data uniform by taking into account a continuous temporal space and adding the values associated with the nearest day that is available, we may solve this problem. The second issue is that Fourier analysis is limited to stationary data, whereas this time series is obviously growing over time. To identify the best-fit polynomial function that fits the data, we will specifically employ polynomial regression as demonstrated in Fig 3. After that, we’ll cut this line to get the stationary time-series as given in Fig 4.
Next step is to filter the data. All frequencies lower than a specific threshold are being eliminated. The maximum amplitude is utilized as a reference point to set this level. The result of filtering the data is given in Fig 5.
If each filtered Fourier transform is inversed, the result is depicted in Fig 6. After denoising, After the data is scaled and divided into testing and training set. The next step is to apply the various techniques and explore the applicability of various methods for time series analysis and forecasting.
VI. Result and discussion
The subsequent section compares the performance of the N-BEATS architecture with that of other deep learning models namely LSTM, GRU, RNN, CNN, Bi-LSTM on various prediction steps in this experiment. N-BEATS is the best model among tested deep learning architectures for both a prediction step of one hour and eight hours, despite the fact that GRU outperforms LSTM based on results given below. Additionally, as the prediction step increases, the errors observed in all models also increase, which is consistent with our intuition that the more uncertain about the future, the greater errors models are likely to make.
1. Evaluation of training and validation loss
This evaluation demonstrates the visual depiction of learning process of how the accuracy improves for classification issues and the loss lowers with each epoch in these plots. A model is said to be learning the training set if its loss decreases, and it is said to be well-generalizing to new data if its loss decreases on the validation set. From the Fig 7, it can be easily concluded that LSTM, GRU and CNN are underfitting as their training and validation losses remain at high levels where as Simple RNN shows overfitting as its training loss keeps on decreasing whereas its validation loss keeps on increasing. Looking at the Fig 7, it can be safely concluded that N-beats generalizes well with the data.
2. Impact of batch size and embedding layers on model efficacy
The impact of batch size and embedding layers on the efficacy of the Federated Learning (FL) model is quantified through the F1 score. The analysis was conducted using two batch sizes: 32 and 128. The numerical findings are summarized as follows:
- (a) Batch Size Impact:
- Batch size of 32: The F1 score achieved an average of 0.89 across the participating clients and the central server.
- Batch size of 128: The F1 score slightly improved to an average of 0.91, indicating a marginally better performance.
- The variance in performance across the two batch sizes remained minimal, showcasing the robustness of the model to batch size changes.
- (b) Embedding Layer Contribution:
Models utilizing pre-trained embeddings demonstrated an F1 score improvement of 2–3% compared to models without embedding layers. This enhancement highlights the efficacy of pre-trained embeddings in capturing semantic representations and improving generalization.
- (c) Overall Findings
While smaller batch sizes (32) were slightly more effective in certain scenarios, larger batch sizes (128) showed a minor edge in maintaining stability and achieving a finer gradient update. The data underscores the adaptability of the FL algorithm and suggests that careful hyperparameter tuning can optimize performance without significantly increasing computation time.
These results provide actionable insights into designing efficient and effective FL systems tailored to specific application requirements.
3. Evaluation of forecasting on training and testing data
In order to evaluate the performance of deep learning models on the multi-variate dataset, it is essential to take into account both test and training performance when assessing the models. When a model fits training data well but does not perform well on test data, it can be overfitting, meaning it has picked up noise instead of the underlying pattern. In contrast, a model that exhibits middling performance on both training and test data may be underfitting or overly basic in order to fail to capture the complexity of the data. For our poultry dataset, the disease outbreaks in the poultry dataset are probably complex, non-linear effects of the input variables. There may be threshold effects for temperature and humidity, whereby extreme values markedly raise the risk of illness. A more direct linear association between the feed variable and disease rates resulting from malnutrition may exist, with lower feed levels being associated with greater disease rates. The graph in Fig 8 implies that all the models were able to capture some patterns in the data. However, the variations in generalization are indicated by the variability in test predictions.
N-BEATs model represented by green color demonstrates good fit and generalization without overfitting indicating test and training predictions that are well matched with each other and with the real data. The result of the N-beats individually is being depicted in the Fig 9, demonstrating the well analysis of the patterns.
4. Analysis of performance metrics of various models
Various performance metrics are analyzed in order to understand the applicability of the models on the dataset. The metrics utilized are he Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), Mean Squared Logarithmic Error (MSLE), the coefficient of determination (R-squared), Root Mean Squared Logarithmic Error (RMSLE). MAE calculates the average magnitude of mistakes in a series of predictions without taking direction into account. It is the average of each forecast error’s absolute value. RMSE is the square root of the average squared discrepancies between the actual observation and the prediction. Larger errors are assigned a higher weight. The ratio between the true and anticipated values is measured by the MSLE (. More so than overestimating, MSLE will penalize underestimating R-Squared measures how well the model replicates observed outcomes based on the percentage of total outcomes variance that the model explains. RMSLE, favours underestimates over overestimates in terms of penalty. The results of various models are given in Table 2.
5. Visual Discussion of the Results of Deep Learning Models for Multivariate Dataset
A visual description of the results provided in Table 2 is provided in Fig 10. LSTM has shown moderate errors. However, the R-squared value of the LSTM model is negative, which means that the LSTM may be performing worse than a horizontal line at the mean of the actual values and hence it can be easily seen that model is not effectively capturing the variation of the dataset. Further, looking at the performance of GRU, it is seen that the GRU model outperforms the LSTM, indicating improved error and prediction ratio management. But the R-squared value is still negative, though, suggesting that there might be a problem with how well the model fits the data. The performance of CNN model is almost similar to LSTM. N-beats is the best-performing model among the employed models as it has the positive R-squared and the lowest errors across all metrics. It appears to be the best at capturing the patterns in the dataset without overfitting to noise or outliers. The performance of RNN is also not preferable as it has negative R-squared value and higher value of other performance metrics. The results of various models are given in Table 2.
VII. Conclusion and future work
The research work focuses on the applicability of various deep learning models for capturing the patterns in multi-variate poultry dataset. The dataset consists of three independent variables and the predictor variable i.e. disease incidence is modelled on the basis of independent variables. After the pre-processing step which includes removing of nose, smoothening of data, and filtration, various models are employed. Based on the results achieved, the N-Beats model is clearly the most applicable for this specific forecasting task. The lesser value of error metrics of the N-Beats model indicates that it is less susceptible to noise and possible outliers, making it more effective at capturing the trend and volatility in the data. This can be easily attributed to its architecture, which is made to efficiently handle a range of data patterns. For all other models, the negative R-squared values raise a concern as it suggests that that none of the models fits the data well indicating that the models are underfitting. Further, with the exception of N-Beats, all of the models’ RMSE values indicates that none of them performs noticeably worse or better at managing huge errors, which are highly penalised in RMSE and this metric also proves the competency of N-beats for the dataset. The future scope of the proposed work includes exploring various strategies for handling negative R-squared values which include feature engineering, hyperparameter tuning, or a detailed data pre-processing steps.
References
- 1. Wu D, Cui D, Zhou M, Ying Y. Information perception in modern poultry farming: A review. Computers and Electronics in Agriculture. 2022;199:107131.
- 2. Government of Punjab. Punjab Economic Survey. Accessed February 12, 2024. https://finance.punjab.gov.in/uploads/10Mar2023/Economic_Survey.pdf
- 3.
Gulati A. Why India Cannot Afford to Ignore the GM Crop Revolution. The Indian Express. Accessed February 12, 2024. https://indianexpress.com/article/opinion/columns/why-india-cannot-afford-to-ignore-the-gm-crop-revolution-8266595/
- 4. Rajesh Mehta RM, Nambiar RG, Joshi PK. India’s poultry sector: trade prospects. International trade and food security: the future of Indian agriculture. 2016:115–23.
- 5. Poultry India. Three Common Diseases in Poultry Farming. Accessed February 12, 2024. https://www.poultryindia.co.in/three-common-diseases-in-poultry-farming/
- 6. Puszkarski B, Hryniów K, Sarwas G. Comparison of neural basis expansion analysis for interpretable time series (N-BEATS) and recurrent neural networks for heart dysfunction classification. Physiol Meas. 2022;43(6):10.1088/1361-6579/ac6e55. pmid:35537407
- 7. Sharma N, Mangla M, Mohanty SN, Pattanaik CR. Employing stacked ensemble approach for time series forecasting. Int j inf tecnol. 2021;13(5):2075–80.
- 8. Dickstein J. Sacred Cows & Chicken Manchurian: The Everyday Politics of Eating Meat in India. HIMALAYA. 2021;40(2):164–6.
- 9. Xu X, Rothrock MJ Jr, Mohan A, Kumar GD, Mishra A. Using farm management practices to predict Campylobacter prevalence in pastured poultry farms. Poult Sci. 2021;100(6):101122. pmid:33975043
- 10. Degu MZ, Simegn GL. Smartphone based detection and classification of poultry diseases from chicken fecal images using deep learning techniques. Smart Agricultural Technology. 2023;4:100221.
- 11. Sharma N, Sharma KP, Mangla M, Rani R. Breast cancer classification using snapshot ensemble deep learning model and t-distributed stochastic neighbor embedding. Multimed Tools Appl. 2022;82(3):4011–29.
- 12.
Kelland K, Reuters . Proliferation of Bird Flu Outbreaks Raises Risk of Human Pandemic. Scientific American. Accessed February 12, 2024. https://www.scientificamerican.com/article/proliferation-of-bird-flu-outbreaks-raises-risk-of-human-pandemic1/
- 13. Hailegebreal G, Molla Tanga B, Woldegiorgis W, Sulayeman M, Sori T. Epidemiological investigation of morbidity and mortality of improved breeds of chickens in small holder poultry farms in selected districts of Sidama Region, Ethiopia. Heliyon. 2022;8(8):e10074. pmid:35992007
- 14. Astill J, Dara RA, Fraser EDG, Sharif S. Detecting and Predicting Emerging Disease in Poultry With the Implementation of New Technologies and Big Data: A Focus on Avian Influenza Virus. Front Vet Sci. 2018;5:263. pmid:30425995
- 15. Nakrosis A, Paulauskaite-Taraseviciene A, Raudonis V, Narusis I, Gruzauskas V, Gruzauskas R, et al. Towards Early Poultry Health Prediction through Non-Invasive and Computer Vision-Based Dropping Classification. Animals (Basel). 2023;13(19):3041. pmid:37835647
- 16. Oreshkin D, Carpov D, Chapados N, Bengio Y. N-BEATS: Neural Basis Expansion Analysis for Interpretable Time Series Forecasting. Proceedings of the International Conference on Learning Representations (ICLR). 2020.
- 17. Singh R, Gupta R, Kaur J. Advancements in poultry farming techniques using IoT and big data analytics. Journal of Agricultural Informatics. 2023;14(1):45–56.
- 18. Patel S, Sharma V. Role of genetically modified crops in addressing food security in India. Current Science. 2023;125(4):567–73.
- 19. Bhardwaj N, Mehta M, Joshi A. Economic Impacts of Poultry Farming on Rural Development in India. Agricultural Economics Research Review. December 2022 35;2:199–210.
- 20. Zhou X, Liu Y, Feng J. AI-Powered disease diagnosis in livestock: challenges and opportunities. Computers and Electronics in Agriculture. 2024;211(2):108417.
- 21. Mishra P, Roy T. Analysis of poultry waste management practices in India: A sustainable approach. Environmental Research and Development. 2023;17(3):78–85.
- 22. Ali S, Khan Z, Rahman S. Emerging trends in poultry nutrition and feed formulation. Poultry Science and Technology. 2023;19(5):155–68.
- 23. Carter J, Lee H. Impact of Climate Change on Poultry Farming Practices. Journal of Climate Resilience. 2023;8(2):122–31.
- 24. Kumar N, Das P. Application of deep learning models for predictive analytics in poultry health management. Neural Computing and Applications. 2023;35(10):8991–9002.
- 25. Singh A, Verma R. Economic analysis of broiler and layer farming in northern India. Agricultural Finance Review. 2023;83(3):334–45.
- 26. Park SJ, Kim DH. Recent developments in sensor technologies for smart poultry farming. Sensors. 2024;24(5):2197.