Figures
Abstract
Computational modelling of biological processes poses multiple challenges in each stage of the modelling exercise. Some significant challenges include identifiability, precisely estimating parameters from limited data, informative experiments and anisotropic sensitivity in the parameter space. One of these challenges’ crucial but inconspicuous sources is the possible presence of large regions in the parameter space over which model predictions are nearly identical. This property, known as sloppiness, has been reasonably well-addressed in the past decade, studying its possible impacts and remedies. However, certain critical unanswered questions concerning sloppiness, particularly related to its quantification and practical implications in various stages of system identification, still prevail. In this work, we systematically examine sloppiness at a fundamental level and formalise two new theoretical definitions of sloppiness. Using the proposed definitions, we establish a mathematical relationship between the parameter estimates’ precision and sloppiness in linear predictors. Further, we develop a novel computational method and a visual tool to assess the goodness of a model around a point in parameter space by identifying local structural identifiability and sloppiness and finding the most sensitive and least sensitive parameters for non-infinitesimal perturbations. We demonstrate the working of our method in benchmark systems biology models of various complexities. The pharmacokinetic HIV infection model analysis identified a new set of biologically relevant parameters that can be used to control the free virus in an active HIV infection.
Citation: Jagadeesan P, Raman K, Tangirala AK (2023) Sloppiness: Fundamental study, new formalism and its application in model assessment. PLoS ONE 18(3): e0282609. https://doi.org/10.1371/journal.pone.0282609
Editor: Attila Csikász-Nagy, Pázmány Péter Catholic University: Pazmany Peter Katolikus Egyetem, HUNGARY
Received: October 5, 2022; Accepted: February 18, 2023; Published: March 8, 2023
Copyright: © 2023 Jagadeesan et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the paper and its Supporting information files.
Funding: PJ acknowledges funding from the Ministry of Education, Government of India. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The Authors declare no Competing Financial or Non-Financial Interests.
Introduction
The dynamics of biological systems are generally considered complex owing to their non-linearity, interconnections and multi-scale nature. Complex dynamical systems are often modelled as non-linear ordinary differential equations (ODEs) with many states and parameters. Dynamical models of biological systems are central to many applications. Recent advances in whole-cell modelling [1] and quantitative systems pharmacology (QSP) [2, 3] underscore the rising importance of dynamical modelling for complex biological systems. Identifying such a complex dynamical system from data poses multiple challenges in each stage of the identification exercise [4]. Identifiability, precise parameter estimation and an-isotropic sensitivity are a few crucial challenges in the computational modelling of complex dynamical systems. One of the crucial and often overlooked sources of these challenges is the possible presence of large regions in the parameter space over which the model predictions are nearly identical. This phenomenon is known as sloppiness [5]. Sloppiness is closely related to another property known as identifiability, where multiple parameter sets result in identical predictions [6].
Identifiability is a well-established concept in the domain of system identification. Loss of identifiability can result from either the model structure (structural identifiability) or data [7, 8]. The loss of structural unidentifiability implies multiple solutions to the estimation problem. Thus, checking for structural identifiability before parameter estimation is imperative. There are several analytical methods available for assessing the structural identifiability of a model structure [6, 9–13]. A differential geometric approach using observability condition is proposed in [14]. However, most of the analytical methods are not scalable to large models. A numerical method is proposed in [15] to assess local structural unidentifiabilities. A global method has been proposed to detect identifiability, active subspaces and parameter sensitivity exploiting the underlying relationship with sensitivity Fisher Information Matrix (sFIM) is proposed in [16].
While identifiability is a ‘binary’ situation, sloppiness lies somewhere between identifiability and loss of identifiability; moreover, the closer it is to the loss of identifiability, the more problematic it is. The presence of sloppiness often results in huge uncertainties in parameter estimates [5]. To illustrate sloppiness consider the simple bi-exponential model in (1), with a true parameter vector, . The model output y(t) is computed for 5 seconds by fixing one parameter and varying the others at a time.
(1)
From Fig 1a, it is evident that for a range of θ2 values in parameter space, the model predictions are nearly identical, and hence, θ2 direction is considered sloppy. However, in Fig 1b, the change in parameter θ1 results in clearly distinctive model outputs. Here, θ1 is a stiff direction. From Fig 1c, it is seen that there are certain directions over which the model predictions are nearly identical yet significantly vary in other directions. However, in many cases, instead of individual parameters being sloppy and stiff, there will be directions in the parameter space that are sloppy and stiff. The nearly identical model outputs for a significant range of parameter sets might reflect as large standard errors in a subset of parameter estimates. These large standard errors in the parameter estimates, often referred to as practical/numerical identifiability [17] is one of the crucial challenges while modelling sloppy models [5, 15, 18].
(a) shows the model output while varying θ2 from 9 to 11 and fixing θ1 = 1. The output of the model does not significantly vary and is qualitatively indistinguishable. (b) shows the model output while varying θ1 from 0 to -2 and fixing θ2 = 10. The output significantly varies while changing θ1 in the same range. (c) shows the outputs for a range of parameters varied over the circumference of a circle with a fixed radius, r = 1 and θ1 = 1, θ2 = 10 as center. Visually evident chunks of outputs that are nearly identical. This implies certain directions in the parameter space over which model outputs are nearly identical.
Practical identifiability is formally defined as the ability to estimate parameters of a structurally identifiable model with acceptable precision given a data set [17, 19]. Practical identifiability is function of both data (inputs and measurement errors) [20] and model structure. Practical identifiability/precision of the parameter estimates and their relationship with input and cost function has been extensively studied in the domain of non-linear system identification [7, 21]. Practical identifiability is assessed by the width of the confidence interval of the parameter estimate [17]. However, there are other numerical methods proposed to assess practical identifiability [11, 22, 23]. The relationship between structural identifiability and sloppiness is well-established in [15]. A sloppy model is always structurally identifiable. Whereas in the case of practical identifiability, though sloppiness is closely related to practical identifiability, the exact relationship is still unclear [15]. This has been one of the sources of ambiguity in understanding the impact of sloppiness analysis as a part of any modelling exercise. The relationship between sloppiness, structural, and practical identifiability is depicted in the Fig 2. Sloppy models intersect with practically unidentifiable models. However, how much sloppiness affects practical identifiability is the question that needs to be answered.
Impact of sloppiness in optimization, experiment design and model validation
Along with structural and practical identifiability, impact of sloppiness on various facets of modelling has been extensively studied in the past decade [5, 15, 24–28]. The impact of sloppiness on designing optimal experiments, optimization algorithms, and uncertainty quantification has been addressed widely in a variety of studies, as we discuss hereon. It has been shown that sloppy models have nearly flat cost surfaces in the vicinity of the optimal parameter set [24]. In such cases, it is observed that non-linear optimization algorithms may get trapped in sloppy regions. A differential geometric approach has been employed to improve the convergence of the Levenberg-Marquardt algorithm. In the case of validating models, it is observed that the uncertainty estimates are unreliable in the case of sloppy models [25]. Henry et al. show that maximizing mutual information between uninformative prior and posterior distribution in Bayesian estimation effectively selects simple models [29]. They have suggested modified Markov Chain Monte Carlo (MCMC) simulations to circumvent this issue. An attempt to design optimal experiments for a precise estimate of parameters in sloppy models may result in compromising prediction accuracy [18].
Numerous attempts have been made to resolve the issues arising from sloppy models, but very few works have attempted to find the root cause of sloppiness [17, 26, 30]. Though the source of sloppiness is attributed to both model and data [5], in most cases, the source of sloppiness is attributed to the model structure [5, 15, 18, 25, 28, 31, 32]. A computational study has been carried out to study the relationship of sloppiness measure with structural identifiability, practical identifiability, and experimental design [17]. The result suggests that the relationship of sloppiness measure with practical identifiability is inconclusive. Tönsing et al [30] have shown that the root causes of sloppiness include both the model and experimental conditions. They suggest that by changing the experimental condition, it is possible to cure sloppiness. Apgar et al [31] have shown it is possible to estimate parameters precisely for sloppy models, by careful design of experiments. Sloppiness has also been reported to have connections with biological phenomena such as robustness and evolvability [11]. Ben Matcha et al. suggest in [33] that the sloppiness is essentially an effect of the multi-scale nature of the underlying process. Further, they show that specific models are not sloppy at microscopic scales, and sloppiness emerges only when fit to collective behaviour. Transtrum et al. and Katherine et al. suggest that the existence of sloppiness (dependence of model parameters to only a few macroscopic parameter directions) is the origin of simplicity in science and responsible for the emergence of comprehensible macroscopic theories from highly complex microscopic processes and [34, 35]. Evangelou et al. have developed a method to identify important and unimportant parameter combinations using manifold learning techniques (Dmaps). Further, they use a special type of auto-encoder known as a Y-shaped conformal auto-encoder to disentangle the unimportant parameter combinations. Finally, the identified effective parameters are mapped back to physical parameters [36].
The utility of sloppiness analysis in modelling has been questioned in [4, 17]. These studies argue through computational study that the presence of sloppiness does not guarantee either structural or practical unidentifiability. Based on the results, they conclude that using sloppiness analysis to assess the identifiability of parameters can be misleading and suggest identifiability analysis as a better tool. The relationship of sloppiness with Fisher information-based experiment design criteria has also been inconclusive. Though the study is convincing, the effect of sloppiness on errors in the parameter estimate is inevitable. The relationship of sloppiness with practical identifiability is widely accepted but not formally established [15, 17].
In all of the above studies, the exact role of sloppiness on various aspects of modelling, such as precision of the parameter estimates, prediction uncertainty and experiment design, is mainly inconclusive because of the lack of formal mathematical relationship between the measure of sloppiness and other properties. A formal mathematical definition of sloppiness is the missing piece in the puzzle. In order to mathematically formalize sloppiness, we first give two motivating examples to point out that with the existing measure of sloppiness, in the case of non-linear predictors, it is not possible to attribute sloppiness to model structure alone decisively. In order to circumvent the ambiguity, in this work, first, we formulate two new theoretical definitions of sloppiness: (i) sloppiness and (ii) conditional sloppiness. The conditional sloppiness is conditioned on the experiment space. The proposed definitions of sloppiness for autonomous ODE systems are defined in an augmented space of parameters and initial conditions. This implies that both model and experimental conditions together are responsible for sloppiness. Secondly, we establish a mathematical relationship between practical identifiability and sloppiness in the case of linear predictors.
Further, we propose a scalable computational method based on the proposed new definitions of sloppiness to assess the goodness of model structure in viable parameter space. The proposed method facilitates detecting conditional sloppiness, insensitive parameters, and local structural unidentifiability. The proposed method can be used to assess the goodness of the model structure in the viable space before estimation. The benefit of analyzing the model in the viable space is two-fold—once we know that the viable space of the model is in a sloppy region, the probability of the estimated model landing in the sloppy region is high. In such cases, appropriate experiments can be designed to circumvent the effects. On the other hand, if the viable space of the model (valid values of parameters and experimental conditions) is not in a sloppy region. If the estimated model is sloppy, we can fine-tune the estimation algorithm and other controllable experimental conditions. The proposed tool can also detect multi-scale sloppiness described in [15].
Motivating examples
This section provides two motivating examples that illustrate the challenges in the current measure of sloppiness in answering the questions of interest in this work. Following are the two challenges in the existing measure of sloppiness (i) Assessing the role of model structure in sloppiness with the existing sloppiness measure and (ii) Detecting practical identifiability with the existing sloppiness measure. We consider a non-sloppy model and show that it can be turned sloppy by changing experimental conditions, and a model that is regarded as sloppy can be made non-sloppy again by changing the experimental condition.
Example 1: The source of sloppiness in linear predictors is data.
Consider the linear predictor given in (2). Let z be a vector of m observations. The parameters of the model are estimated using the method of least squares. The Hessian of the cost function is given in (4).
(2)
Now, let us define
(3)
(4)
For the purpose of discussion, we use an existing measure of sloppiness given in (5), the ratio of smallest to largest eigenvalues of the Hessian of the cost function.
(5)
Linear predictors are not generally observed to be sloppy [18]. However, in the above example, the ratio of eigenvalues is only a function of data and not the model parameters—hence, we can see that it is possible for the linear predictor to show sloppiness for some experimental conditions. Moreover, sloppiness in linear predictors given in the above example is purely an artefact of data/input and not due to the nature of the model structure itself.
Example 2: Both data and model structure are responsible for sloppiness in nonlinear predictors.
Consider the state space model in (6). The Hessian of the least-squares cost function is given in (7).
(6)
(7)
Sloppiness is computed in the grid 1 ≤ θ1 ≤ 100 and 1 ≤ θ1 ≤ 100. From Fig 3a, it is evident that sloppiness is a function of parameter space. In Fig 3b, sloppiness is computed over a grid of initial conditions 1 ≤ x1(0)≤100 and 1 ≤ x2(0)≤100 around the point θ1 = 1, θ2 = 100. The model is simulated for t = 0 to t = 20 seconds with a sampling interval of 0.5 seconds. The model is structurally identifiable for all the initial condition and parameters considered.
(a) Shows sloppiness in the parameter space. There are regions in the parameter space where the system is non-sloppy. (b) Shows sloppiness for various initial conditions. For a subset of initial conditions, the model becomes non-sloppy.
From Fig 3a, we can observe that the change in experimental condition can cure sloppiness, which indicates that sloppiness is a function of information contained in a data set. In our previous work [37], we have demonstrated using simulations that parameters contributing to sloppy directions have very low information gain in the Bayesian framework. For initial condition > 50 of x2, the model becomes non-sloppy. This has been extensively studied in [11, 30, 31]. Tonsing et al. (2014) showed that the structure of the sensitivity matrix is the reason for sloppiness in ODEs and also concluded that both experiment and model structure are the root causes of sloppiness. Fig 4 shows that the relationship between sloppiness and standard errors of the parameters does not follow any definitive trend. Though it has been argued that sloppiness and practical identifiability are two distinct concepts and that they are incorrectly conflated [18], the effect of sloppiness on practical identifiability cannot be ignored [5].
It is seen that there is no specific relationship between sloppiness and standard errors.
Following are two important observations emanate from our study: (i) for a non-linear least-squares estimation problem, it is impossible to attribute sloppiness to the model structure alone. Sloppiness is a function of both parameter space and initial conditions, and hence, labelling a model to be sloppy with the current analysis method of sloppiness can be misleading [5, 18] (ii) sloppiness often results in loss of practical identifiability. However, the exact relationship is not revealed by the current measure of sloppiness.
The rest of the paper is organized as follows: Section 2 provides perspectives on sloppiness by revisiting the concept of sloppiness and rightly positioning it relative to identifiability. Section 3 presents key results, including two motivating examples to highlight the challenges in using the current measure of sloppiness. Further, it provides two new definitions of sloppiness and their relationship with practical identifiability in the case of linear predictors. Section 4 illustrates the proposed method to assess the sloppiness of non-linear predictors. The paper ends with concluding remarks in Section 5.
Problem statement
The demonstrated challenges of the current measure of sloppiness results in the need to answer the following questions: (i) what is the source of sloppiness? (ii) what is the formal relationship of sloppiness with parameter uncertainty / practical identifiability? (iii) what is the relationship of sloppiness with inputs and estimation algorithm? In this work, we propose two new mathematical definitions of sloppiness to circumvent these challenges. Further, we develop an improved computational method to detect local structural identifiability and sloppiness along with sensitive and insensitive parameters in the viable region of the parameter space.
Perspectives
Nature of sloppiness
Sloppiness, at its core, quantifies the sensitivity of output with respect to changes in parameters or directions in the parameter space. When the gradient of the predictor qualitatively does not change as the parameter is varied significantly, then the system is sloppy. Sloppiness can be observed as pockets of regions in the parameter space; in such cases, the gradient of the predictor varies significantly in the parameter directions.
In the existing literature, a method of quantifying this insensitivity of predictions to changes in parameters is captured by (8).
(8)
where H is the Hessian of the cost function, in general, for a non-linear parameter space, it is difficult to identify all such pockets of sloppy regions. Hence, sloppiness is determined locally around a parameter of interest by computing the sensitivity of the output to the change in the parameter. The sloppiness is characterized by equally spaced eigenvalues of the Hessian of the cost function in the log scale and is quantified by (8).
Even though there is no strict cut-off, a model is considered to be sloppy if S ≤ 10−6 [32]. The eigenvectors corresponding to maximum and minimum eigenvalues indicate the stiff and sloppy direction, respectively. It can be observed that the Hessian of the cost function is representative of the derivative of the output with respect to the parameters. The above measure of sloppiness is valid only if the estimation algorithm is least-squares; for a different estimation algorithm (for example, L1 norm), the Hessian of the cost function may not be representative of the sensitivity of the output with respect to the parameters.
Sloppiness and identifiability
As a first step in answering the questions raised above, the notion of sloppiness needs to be rightly positioned relative to well-established concepts such as structural identifiability and practical identifiability. Additionally, when a model is sloppy, it is important to attribute the source of sloppiness to the appropriate factors.
Table 1 shows the list of model properties, their definitions, and the corresponding method of assessment. Even though qualitative definitions for practical identifiability and sloppiness are available, a formal mathematical framework for defining sloppiness is not available. In this work, we formalize sloppiness mathematically. Mathematical definitions are given in the results section.
All three properties of interest that are under discussion arise due to either one or more of the data, model, and estimation algorithm. The data itself is characterized by (i) signal-to-noise ratio (SNR), (ii) sample size, and (iii) the input. While input and SNR determine the quality of data, the quantity of data is determined by sample size. The model’s contribution towards these properties is characterized by predictor gradient, and finally, the estimation method is generally characterized by the cost function.
The relationship of sloppiness with practical identifiability is still an open problem [15]. From Table 2 it is clear that both the input and the gradient of the output with respect to parameters affect sloppiness and practical identifiability. This is the reason why most parameters in a sloppy model become practically unidentifiable when the gradient is large. In the following section, we provide this work’s primary results: The mathematical formalism of sloppiness and its application to model assessment.
Results
In answering the crucial questions about sloppiness, we formulate two new mathematical definitions of sloppiness. The proposed mathematical definitions of sloppiness are more accurate descriptions of sloppiness as they are formulated from the fundamental nature of sloppiness.
A new mathematical definition of sloppiness
The above-perceived challenges of the present sloppiness analysis motivated us to formulate two new mathematical definitions of sloppiness. The new definitions of sloppiness are based on the following remarks.
Remark 1. Sloppiness is assessed across an augmented space (parameters, initial conditions, and inputs) rather than parameter space alone. In the case of autonomous ODE models, the augmented space is
Remark 2. A significant change in the subset of augmented space (parameter space) results in a small change in prediction space.
—All possible parameter values the model
can take.
—Identifiable region,
—Set of all experiments for which the model structure
is identifiable.
—of all parameters that belong to the sloppy region.
Definition 1. A model is (ϵ, δ) sloppy with respect to an experiment space
at
, if
(9)
(10) for every (θ1, θ*) satisfying (9) & (10). ϵ is arbitrarily small. δ ≫ ϵ.
Definition 2. A model is conditionally (ϵ, δ) sloppy with respect to an experiment space
at
, if
(11)
(12) for every (θ1, θ*) satisfying (11) & (12). ϵ is arbitrarily small. δ ≫ ϵ.
Eqs (9) & (10) say that a model is considered sloppy, if the (ϵ, δ) condition holds true for all possible experimental conditions . Similarly, (11) & (12) convey that the model is conditionally sloppy if the (ϵ, δ) condition holds true only for a subset (
) of all possible experimental conditions (Z).
No longer the proposed sloppiness and conditional sloppiness are used in a generic sense; rather, they have to be used in conjunction with ϵ and δ. By virtue of our definitions, we believe, by itself, the sloppiness is qualitative, and where ever quantitative sloppiness is to be discussed, the (ϵ, δ) from should be used. For a large δ if the ϵ is negligibly small, then the system is considered to be (δ, ϵ) sloppy. Note that the proposed measure of sloppiness is different from the multi-scale sloppiness proposed in [15]. Multi-scale sloppiness is the ratio of maximum to minimum prediction error for a non-infinitesimal perturbation from a reference parameter (θ*), whereas, in the proposed definition, sloppiness is a function of both prediction error and parameter perturbation which is a more natural way to define and understand the notion sloppiness.
Sloppiness analysis of a linear predictor.
Consider the linear predictor in (13)
(13)
(14)
The model is identifiable if X is full column rank. Consider another parameter vector θ1 and the corresponding prediction y1. The model is sloppy if X(θ* − θ1) = (y − y1) for ||θ* − θ1||2 > δ and ||y − y1||2 < ϵ.
Let θd = (θ* − θ1), ||θd||2 ≫ ϵ.
Consider the extreme case for p = 2,
(16)
For extremely small ϵ and extremely large δ, the ||X|| ≈ 0 and will become numerically unstable and that will result in loss of identifiability. This is the reason why in most cases, linear least-square problems are observed to be non-sloppy. Once the system is found practically unidentifiable, then a careful design of the experiment will constrain the parameter uncertainty making the matrix norm of data significantly large. In such a case, the possibility of sloppiness is almost eliminated.
Relationship between conditional sloppiness and practical identifiability.
For a structurally identifiable model structure and a data set z, precision of parameter estimate θi ∈ θ is a measure of practical identifiability. Practical identifiability is assessed for a data set z given a model structure
. A parameter θi is said to be practically unidentifiable for a given z, if
where
is the standard error of the parameter estimate θi and
is the prediction error. Practical identifiability can only be assessed post estimation because it is a function of the data set, whereas sloppiness can be evaluated at any given unknown space (ϕ), which includes parameter space and data set. Ideally, the infinite width of the confidence interval of a parameter estimate is considered practically unidentifiable. However, in practical scenarios, significant standard errors in the estimates are considered practically unidentifiable.
Consider the linear predictor given in (13). Choose another parameter set such that
.
(17)
(18)
If each xi is a vector of m observations, then (18) becomes,
(19)
Inverting the diagonals of (4) gives the covariance matrix for the parameter estimates. The standard error of a parameter ai is given by
(20)
Using (19), the relationship between the sloppiness and practical identifiability is derived as
(21)
This relationship holds if the data is derived from a Gaussian distribution and the least-squares cost function. From (21), it is clear that there is a relationship between sloppiness and practical identifiability. The corresponding parameter will become unidentifiable for a small ϵ and a very large δi. The above result agrees with the result obtained in (17). For a generalised non-linear predictor, it is challenging to construct a relationship analytically. Hence, we propose a numerical method to analyse non-linear predictors for sloppiness and identifiability.
Equivalence between traditional sloppiness and new definition of sloppiness
While traditional sloppiness measures the relative skewness between the largest and smallest eigenvalues for infinitesimal parameter perturbation, the proposed definition measures the skewness between the prediction and the variation in the parameter for non-infinitesimal perturbation.
Theorem 1. For a positive definite Hessian of (10) with distinct eigenvalues and δ being the length of largest semi-axis, λmin(∇2(C(θ))) → 0, if limϵ→0 and limδ→α.
Proof. Let
(22)
The Hessian of (22) equated at θ* is given by
(23)
where, n is the number of observations. At nominal value θ*, the limϵ→0, the Hessian of the cost function can be approximated to
(24)
The length of the largest semi-axis and (9) are related by
(25)
For an arbitrarily large α,
(26) α is an arbitrarily large scalar representing the length of the largest semi-axis of the hyper ellipsoid around the parameter on which the Hessian of the cost function is evaluated. α cannot be infinity because that results in a singular Hessian, which leads to the loss of identifiability. This concludes the proof.
As a consequence of Theorem 1
Corollary 1. For a positive definite Hessian of C(θ) with distinct eigenvalues and δ being the length of largest semi axis, when limϵ→0 and limδ→α. Then, sloppiness measure (S) increases.
Proof. Sloppiness measure is given by
(27)
from (17)
(28)
Smaller the S, sloppier the system is. This concludes the proof.
The (δ, ϵ) measure guarantees the existence of at least one sloppy direction. However, the proposed definition is a super-set of the existing definition. Two important distinct features of the proposed definition are (i) The proposed definition does not reflect the total aspect ratio. Rather, it defines the region of the parameter space over which model predictions are less than an arbitrarily small value (ϵ) (ii) The crucial problem in the current measure of sloppiness is that the notion of small or large eigenvalues needs to be better defined, as pointed out in [16]. The prime reason is that the parameter estimates are affected by the eigenvalue magnitude and not the ratio. For an infinitesimal perturbation, the minimum eigenvalue of the FIM/Hessian of the cost is a function of δ as shown in Theorem 1. Moreover, the boundaries of the viable space of parameters will dictate the magnitude of δ; by this, the user for a particular modelling exercise can assess the effect of eigenvalues on the precision of the parameter estimates.
Sloppiness analysis of non-linear predictors
We provide three numerical examples to demonstrate the sloppiness analysis of nonlinear predictors. We propose a novel numerical method of analysing sloppiness based on the new definitions. The detailed working of the method is illustrated in the methods section. The first illustrative example demonstrates all three different scenarios (i) a sloppy region, (ii) a non-sloppy region and (iii) an unidentifiable region. The second example is one of the hallmark models used by Gutenkunst et al. in [5] to demonstrate sloppiness. The third example is a realistic pharmacodynmaic HIV infection model. Along with detecting identifiability and sensitivity analysis, we show that the proposed method also gives new biological insight into the effect of HIV infection on T-Cells. Two more examples are provided in the supplementary to demonstrate the working of the proposed examples in detecting complete unidentifiability and in a high-dimensional model.
An illustrative example.
In this example, we use a simple two-parameter state-space model to demonstrate the working of our method. Three different scenarios are considered to cover the loss of structural identifiability, sloppiness, and an ideal scenario. They are shown in Table 3. Consider the state-space model in (29).
(29)
Parameters are sampled inside an n-ball with radius δ around θ*. The model output y(t, θ) is computed for all the sampled parameters and sum-squared error γ is computed. Further, the model-sensitive index is computed using (33).
Fig 5a and 5b show the plot for model sensitivity index and minimum deviation from reference θ*. In the case of the non-sloppy region, the curve in Fig 5b is monotonically increasing. The model’s behaviour is distinguishable from the reference point (θ*) as δ increases. Additionally, the curve in the model sensitivity plot in Fig 5a deviates away from the unity value as δ increases. In the case of a sloppy region, in Fig 5b, the curve has an increasing trend, but the curve is almost flat and not distinguishable as compared to the non-sloppy region from the reference point. In the third scenario, where the system is locally structurally unidentifiable, from Fig 5b, it is seen that γmin is very close to zero/ numerically zero for a non-zero δ. From Fig 5c, we can infer that both the parameters are unidentifiable in the region because at an absolute distance δi from the reference value, the prediction error goes to zero, which indicates that there is another parameter θ ≠ θ* for which the prediction error is zero. In addition to that, in Fig 5a, the curve in model sensitivity plot touches the unity value for a non-zero δ.
(a) shows the model sensitivity plot for all three regions; for the non-sloppy region, the curve deviates from the unity value for the given δ, indicating local structural identifiability. The curve is very close to the unity value for the slope region, indicating an-isotropic sensitivity/ multi-scale sloppiness. The curve numerically hits unity for the unidentifiable region, indicating local structural unidentifiability for the given δ. (b) shows the γmin vs δ plot for all three regions; for the non-sloppy region, the γmin is numerically significant and has a constant slope for given δ. For sloppy region, γmin is numerically insignificant for a significantly large δ. The slope of the curve is constant, but the value of the slope is numerically closer to the zero value indicating a sloppy region in the given δ. The curve numerically hits zero value respectively at δ = 0.14 indicating a local structural unidentifiability. (c) Shows γmin vs for unidentifiable region, in this case both the γmin is zero for non-zero
indicating local structural unidentifibility for both the parameters.
In order to assess local structural non-identifiability, the model sensitivity plot in conjunction with γmin plot may be used. However, a numerically zero value of γmin for a non-zero δ is sufficient to assess local structural non-identifiability. The model sensitivity index (ψ) quantifies the asymmetry between the most sensitive and least sensitive parameter directions in the parameter space. For a given prediction error ϵ = γmin, the conditional sloppiness can be assessed by the pair (ϵ, δ). For a given δ, which can be chosen as a function of acceptable parameter range among the set of parameters, if the γmin is too low, then we can say the system is conditionally sloppy with a high probability of practically unidentifiable parameters.
Mitotic oscillator.
The minimal cascade model for the mitotic oscillator is an ODE model with three states and ten parameters. This model is one of the sixteen system biology models that were shown to be sloppy [5]. The original model and the nominal parameters are obtained from [38]. Here, we analyze the behaviour of the model by constructing the visual plot.
The true parameters around which the model’s behaviour is analyzed are taken from [38]. It can be observed from Fig 6a that for δ < 0.1, there is a huge asymmetry between the minimum and maximum deviation, which implies that the system will be extremely sloppy with respect to the traditional definition of sloppiness (8) and multi-scale sloppiness [15]. From Fig 6b, it can be seen that as δ increases, γmin also increases but goes nearly flat after δ > 0.12, indicating negligible change in the γmin as δ changes, which implies that the model is (δ, ϵ) sloppy for δ > 0.12.
(a) The system is initially sloppy till 0 < δ < 0.03 and becomes non-sloppy for 0.05 < δ < 0.15 and again sloppy for 0.15 < δ < 0.3 (b) the system is locally structurally identifiable.
Figs 7 and 8 show how γmin and γmax changes for each parameter as the absolute value changes. It can be observed the system is insensitive to parameters kd, vd, K1, K2 and V4 as there is no significant change in γmin for the relative change in δi > 0.01. On the other hand, the parameters K3, Vi, V2 and V4 are highly sensitive for δi < 0.01. This gives us a good idea of how the system behaves with respect to the changes in specific parameter intervals. This can be used to fix the initial values of the parameter during an estimation exercise in order to avoid the sloppy region, which is one of the crucial challenges in a sloppy model [24]. Our method can give a good range of initial parameter values for a nonlinear optimization algorithm.
The parameters K1 and vd are highly insensitive within the given δ.
The parameters K4, V4&K2 are insensitive and contributes to the sloppy direction.
In summary, we found that (i) the model is locally structurally identifiable, (ii) the model is sloppy in the traditional sense of sloppiness also from the proposed definition of sloppiness, (iii) our method has identified insensitive parameters and most sensitive parameters (iv) the proposed method has also identified an interval for each parameter (θi) over which the model predictions are most and least sensitive.
Analysis of a pharmacodynamic HIV infection model.
In this example, we analyze a published pharmacodynamic HIV infection model. The differential equation of the model and the nominal parameters are given in [39]. The model describes the effect of HIV infection on CD4+ T-cells. The model consists of four species: Uninfected T cells, latently infected T cells, actively infected T cells and free viruses. The model has four states and eight parameters. All the parameters have biological significance. The block diagram of the model is given in Fig 9. The block diagram is adapted from [40].
T- uninfected T cells, T* latently infected T cells, T**- actively infected T-cells and V- free virus. The parameters include s—the rate of supply/production of CD4+ T-cells, μT—the death rate of uninfected and latently infected T cells. r- the rate of growth of CD4+ T cell population, k1-rate at which CD4+ T cells become infected, k2—the rate at which latently infected T cells become actively infected, μb-death rate of actively infected T cells, Nv number of free cells produced by lysing a CD4+ T-cell, μv—the death rate of the free virus, Tmax—total number of T cells.
In addition to detecting the local structural identifiability of the parameters, the analysis shall reveal the key parameters and critical parameter combinations that influence the free virus. The amount of free virus at any point in time influences depletion of CD4+ T cells [39] and hence finding parameters that influence free virus shall help significantly enhance the treatment regimen.
The parameters are transformed into log scale in order to avoid scaling issues. The initial number of samples N0 = 100000 and α = 1000. Fig 10a and 10b show the model sensitivity plot and γminvsδ plot. It can be observed that the model has high an isotropic sensitivity as ψ is uniform and close to unity for all δi. However γmin is significantly higher than zero for all δi indicating local structural identifiability inside the radius δ.
(a) The curves indicate a constant slope sensitivity ratio. (b) The curve has a significantly small slope and numerically small γmin, which indicate the sloppiness for the given δ (c). The curve does not hit the unity value indicating local structural identifiability.
Now that we know that the model has high anisotropic sensitivity inside given δ, the next step is to identify the sensitive and insensitive directions and how they are distributed. Hence, we construct a heat map to characterize the directions of the most and least sensitive directions in prediction/cost space. Fig 11a shows the cos(θ) between every vector corresponding to minimum directions to every other vector and Fig 11b show the cos(θ) between every vector corresponding to maximum directions to every other vector. It is observed that the vectors corresponding to minimum deviation are spread across large dimensions. Conversely, the vectors corresponding to maximum directions are constrained to a few dimensions. This indicates that most of the directions in the parameter space are sloppy while very few contribute to the model output (free virus V).
(a) The heat map indicates several minimum directions at each δi (b) On the other hand, the direction of maximum deviation is almost identical for each δi.
Further to finding the maximum directions, we construct global sensitivity plots (γminvsθi) and (γmaxvsθi) for every parameter. From Figs 12 and 13, the parameter μb is the most sensitive of all the parameter followed by the parameter r. This indicates that a very slight movement of the maximum deviation vector in the direction of μb(the death rate of the actively infected CD4+ T cells) creates a large change in the free virus (V). This result differs from the global sensitivity analysis carried out using the Sobol method in [40]. The biological implication of the result is that the death rate of actively infected T cells increases the free virus (V). This result is an interesting perspective as the death of T-cells can produce free viruses, which can progress the diseases further. One possible suggestion from the current analysis is to delay the death rate of CD4+ T cells. From a control theory perspective, the death rate of T-cells can be used as a manipulated variable to control the free virus and disease progression in HIV infection.
y-axis on the left is the minimum sum-square deviation, and on the right maximum sum-square deviation from y*(t). Parameter r is more sensitive when compared to other parameters in the direction of maximum deviation.
y-axis on the left is the minimum sum-square deviation, and on the right maximum sum-square deviation from y*(t). Parameter μb is sensitive compared to other parameters, and the sensitivity of the parameter Nv linearly increases in the direction of maximum deviation. Summery of results is given in Table 4.
Methodology
In this section, we discuss the visual tool introduced in the previous section to detect and quantify conditional sloppiness and detect the loss of structural identifiability for generic ODE models.
Mathematical model formulation
We consider the model of the form,
(30)
where
is a state vector.
is an nu-dimensional input vector, and
is an ny-dimensional output vector. The vector
is the vector of parameters. The system has state function (f) and the observation function (h). The observation function can be modified by an experiment scheme, whereas the state function is fixed.
Proposed method
The visual tool is based on the definition of sloppiness proposed in the previous section. The primary idea is to study the behaviour of the model structure around the point of interest in the parameter space. A Euclidean ball of radius δ is sampled around the parameter of interest θ* using a multivariate Gaussian distribution. The radius δ is subdivided into l equally spaced segments. The behaviour of the model is evaluated in each of the sub-Euclidean balls. The maximum and minimum deviation of predictions from the reference parameter vector is plotted against the radius vector (δ).
Procedure.
- Divide the radius δ into l equal segments, δ as δk, k = 1, 2…..l
- Fix the sample size N for δk = 1 and sample parameters from a Euclidean ball B of radius δk around θ* in the parameter space using uncorrelated multivariate Gaussian with standard deviation as δk.
(31)
- Simulate the model output y*(t) at the optimal/true parameter θ*
- Simulate the model output for all the parameters in the Euclidean ball and compute the sum square error γ with the optimal output y*(t).
- Compute γmin and γmax from each increment δk
- Update sample size
(32)
- Repeat steps 4 to 6 while k ≤ l.
- Compute the model sensitivity index
(33)
- Plot (δ,γmin) and (δ,ψ)
Sampling n-ball using multivariate Gaussian distribution is one of the efficient methods. However, for a sufficiently large dimension, with a high probability, the distance between all points will be the same, and the volume of the n-ball goes to zero [41]. To overcome this issue, we need to generate points from independent Gaussian distribution and normalise each vector [41]. While increasing the radius δk, the sample size N has to be increased to avoid missing the regions of unidentifiability. For each increment in δk, we derived Eq (32) to update the sample size. The parameter α can be used as a turning parameter to attain certain accuracy in the sampling, and n is the dimension of the parameter space. Fig 14 illustrates the procedure to construct the visual tool.
Discussion
Sloppiness, practical identifiability, and structural identifiability are the most frequently encountered challenges in computational modelling, particularly in complex dynamical systems. Assessing the model structure for sloppiness and parameter unidentifiability becomes imperative for successful parameter estimation. The concept of structural identifiability is well-established in the domain of system identification, and hence a great deal of work has been done on assessing the structural identifiability of a model structure. However, most analytical methods are limited to a specific type of non-linearity and miniature models. A few numerical methods also have been developed to assess structural identifiability. On the other hand, sloppiness has not been investigated with similar rigour.
Though there has been considerable discussion on sloppiness in the literature, there were a few crucial unanswered questions: (i) What is the source of sloppiness? (ii) what is the relationship of sloppiness with parameter uncertainty / practical identifiability? (iii) what is the relationship of sloppiness with inputs and estimation algorithm? This work provides definitive answers to all these questions by developing mathematical formalism of sloppiness. The new formalism defines sloppiness in an augmented space of parameters, initial conditions, and inputs.
We position the notion of sloppiness relative to well-established concepts such as structural identifiability and practical identifiability. Further, using simulations, we demonstrated the challenges in applying the current measure of sloppiness in a modelling exercise and propose to argue that the ambiguity in understanding sloppiness and related questions is due to the lack of mathematical formalism. We developed two new theoretical definitions of sloppiness, namely sloppiness and conditional sloppiness. Conditional sloppiness is conditioned on the experiment space. Using the proposed definition of sloppiness, we showed that the linear predictors cannot be sloppy but eventually become unidentifiable. A mathematical relationship between practical identifiability and conditional sloppiness has been derived for generalized linear predictors.
A numerical method is proposed to assess the conditional sloppiness for generalized non-linear predictors. The proposed method helps determine the model’s behaviour around a point of interest in the parameter space. An n-ball of radius δ is constructed with the reference parameter as the centre. Deviation in the prediction from the reference parameter is computed for all the parameters sampled from the n-ball. The minimum and maximum deviations in the prediction from the nominal parameter were plotted against the delta. The proposed tool also helps find the most sensitive and least sensitive parameters with an interval. The method can also detect structural unidentifiability.
The proposed tool was applied to three different models, including a hallmark system biology model. The proposed method detected sloppiness, structural unidentifiabilities, and sensitive and insensitive parameters. The analysis gave a holistic picture of the system’s behaviour in a subset of a region in the parameters space. The parameter interval obtained from the proposed method can be used to fix the initial parameter values for the parameter optimization in sloppy models, primarily to avoid flat regions where optimization algorithms may get stuck.
In summary, we see four crucial contributions in this study (i) rightly positioning the concept of sloppiness in relationship with identifiability (ii) elucidating the challenges in the current measure of sloppiness (iii) a new mathematical definition of sloppiness, and finally, (iv) an improved framework to assess sloppiness, structural identifiability, and parameter sensitivity. We believe this work has reconciled key aspects of sloppiness in the light of system identification and will pave the way for further synergies between the systems biology and system identification communities.
Supporting information
S1 File. Additional examples.
This file contains the analysis of two more examples (a) High dimensional biochemical pathway to demonstrate the efficacy of the method of handle large models (b) A Cholesterol absorption model to demonstrate the ability to detect local structural unidentifiability of all the parameters.
https://doi.org/10.1371/journal.pone.0282609.s001
(PDF)
S2 File. Model equations.
The ordinary differential equation models used in the examples and the link for the code for simulating the visual plot are provided in this supporting information file.
https://doi.org/10.1371/journal.pone.0282609.s002
(PDF)
Acknowledgments
PJ acknowledges the HTRA fellowship from the Ministry of Education, Government of India. The authors thank M. Jolly and K. Hari (IISc), for useful comments on the manuscript.
References
- 1. Karr JR, Sanghvi JC, Macklin DN, Gutschow MV, Jacobs JM, Bolival B Jr, et al. A whole-cell computational model predicts phenotype from genotype. Cell. 2012;150(2):389–401. pmid:22817898
- 2. Liu GS, Ballweg R, Ashbaugh A, Zhang Y, Facciolo J, Cushion M, et al. A quantitative systems pharmacology (QSP) model for Pneumocystis treatment in mice. BMC Systems Biology. 2018;12. pmid:30016951
- 3. Kumar R, Thiagarajan K, Jagannathan L, Liu L, Mayawala K, de Alwis D, et al. Beyond the single average tumor: Understanding IO combinations using a clinical QSP model that incorporates heterogeneity in patient response. CPT: pharmacometrics & systems pharmacology. 2021;10(7):684–695. pmid:33938166
- 4. Villaverde A, Banga J. Reverse engineering and identification in systems biology: Strategies, perspectives and challenges. Journal of the Royal Society, Interface / the Royal Society. 2014;11:20130505. pmid:24307566
- 5. Gutenkunst RN, Waterfall JJ, Casey FP, Brown KS, Myers CR, Sethna JP. Universally Sloppy Parameter Sensitivities in Systems Biology Models. PLoS Computational Biology. 2007;3(10). pmid:17922568
- 6. Bellman R, Åström KJ. On structural identifiability. Mathematical Biosciences. 1970;7(3):329–339.
- 7.
Tangirala A. Principles of System Identification: Theory and Practice; 2013.
- 8. Pia Saccomani M, Audoly S, D’Angiò L. Parameter identifiability of nonlinear systems: the role of initial conditions. Automatica. 2003;39(4):619–632.
- 9. Chis OT, Banga JR, Balsa-Canto E. Structural identifiability of systems biology models: A critical comparison of methods. PLoS ONE. 2011;6(11). pmid:22132135
- 10. Villaverde AF, Evans ND, Chappell MJ, Banga JR. Input-Dependent Structural Identifiability of Nonlinear Systems. IEEE Control Systems Letters. 2019;3(2):272–277.
- 11. Chachra R, Transtrum M, Sethna J. Comment on “Sloppy models, parameter uncertainty, and the role of experimental design”. Molecular BioSystems. 2011;7:2522; author reply 2523–4. pmid:21503370
- 12. Castro M, de Boer RJ. Testing structural identifiability by a simple scaling method. PLoS Computational Biology. 2020;16(11):1–15. pmid:33141821
- 13. Bates DJ, Hauenstein JD, Meshkat N. Identifiability and numerical algebraic geometry. PLOS ONE. 2019;14(12):1–23. pmid:31834904
- 14. Villaverde AF, Barreiro A, Papachristodoulou A. Structural Identifiability of Dynamic Systems Biology Models. PLOS Computational Biology. 2016;12(10):1–22. pmid:27792726
- 15. Raman DV, Anderson J, Papachristodoulou A. Delineating parameter unidentifiabilities in complex models. Physical Review E. 2017;95(3). pmid:28415348
- 16.
Brouwer AF, Eisenberg MC. The underlying connections between identifiability, active subspaces, and parameter space dimension reduction; 2018. Available from: https://arxiv.org/abs/1802.05641.
- 17. Chis Ot, Villaverde AF, Banga JR, Balsa-canto E. Mathematical Biosciences On the relationship between sloppiness and identifiability. Mathematical Biosciences. 2016;282:147–161.
- 18. White A, Tolman M, Thames HD, Withers HR, Mason A, Transtrum MK. The Limitations of Model-Based Experimental Design and Parameter Estimation in Sloppy Systems. PLoS Computational Biology. 2016; p. 1–26. pmid:27923060
- 19. Joseph DiStefano III, Dynamic systems biology modeling and simulation. Science Progress. 2019;102(4):378–378.
- 20. Nguyen VK, Klawonn F, Mikolajczyk R, Hernandez-Vargas EA. Analysis of Practical Identifiability of a Viral Infection Model. PLOS ONE. 2016;11(12):1–16. pmid:28036339
- 21. Schoukens J, Ljung L. Nonlinear System Identification: A User-Oriented Road Map. IEEE Control Systems Magazine. 2019;39(6):28–99.
- 22. Gábor A, Villaverde AF, Banga JR. Parameter identifiability analysis and visualization in large-scale kinetic models of biosystems. BMC Systems Biology. 2017;11(1):1–16. pmid:28476119
- 23. Borisov I, Metelkin E. Confidence intervals by constrained optimization—An algorithm and software package for practical identifiability analysis in systems biology. PLOS Computational Biology. 2020;16(12):1–13. pmid:33347435
- 24. Transtrum MK, Machta BB, Sethna JP. Why are nonlinear fits to data so challenging? Physical Review Letters. 2010;104(6):2–5. pmid:20366807
- 25. Gutenkunst RN, Casey FP, Waterfall JJ, Myers CR, Sethna JP. Extracting falsifiable predictions from sloppy models. Annals of the New York Academy of Sciences. 2007;1115:203–211. pmid:17925353
- 26.
Mannakee BK, Ragsdale AP, Transtrum MK, Gutenkunst RN. In: Geris L, Gomez-Cabrero D, editors. Sloppiness and the Geometry of Parameter Space. Cham: Springer International Publishing; 2016. p. 271–299.
- 27. Daniels BC, Chen YJ, Sethna JP, Gutenkunst RN, Myers CR. Sloppiness, robustness, and evolvability in systems biology. Current Opinion in Biotechnology. 2008;19(4):389–395. pmid:18620054
- 28.
Waterfall J. Universality in Multiparameter Fitting: Sloppy Models; 2006.
- 29. Mattingly HH, Transtrum MK, Abbott MC, Machta BB. Maximizing the information learned from finite data selects a simple model. Proceedings of the National Academy of Sciences. 2018;115(8):1760–1765. pmid:29434042
- 30. Tönsing C, Timmer J, Kreutz C. Cause and cure of sloppiness in ordinary differential equation models. Physical Review E—Statistical, Nonlinear, and Soft Matter Physics. 2014;90(2):1–16. pmid:25215847
- 31. Apgar JF, Witmer DK, White FM, Tidor B. Sloppy models, parameter uncertainty, and the role of experimental design. Molecular BioSystems. 2010;6(10):1890–1900. pmid:20556289
- 32. Waterfall J, Casey F, Gutenkunst R, Brown K, Myers C, Brouwer P, et al. Sloppy-Model Universality Class and the Vandermonde Matrix. Physical review letters. 2006;97:150601. pmid:17155311
- 33. Machta BB, Chachra R, Transtrum MK, Sethna JP. Parameter Space Compression Underlies Emergent Theories and Predictive Models. Science. 2013;342(6158):604–607. pmid:24179222
- 34. Transtrum MK, Machta BB, Brown KS, Daniels BC, Myers CR, Sethna JP. Perspective: Sloppiness and emergent theories in physics, biology, and beyond. The Journal of Chemical Physics. 2015;143(1):010901. pmid:26156455
- 35. Quinn KN, Abbott MC, Transtrum MK, Machta BB, Sethna JP. Information geometry for multiparameter models: New perspectives on the origin of simplicity. Reports on Progress in Physics. 2021.
- 36. Evangelou N, Wichrowski NJ, Kevrekidis GA, Dietrich F, Kooshkbaghi M, McFann S, et al. On the parameter combinations that matter and on those that do not: data-driven studies of parameter (non)identifiability. PNAS Nexus. 2022;1(4). pmid:36714862
- 37. Jagadeesan P, Raman K, Tangirala AK. A new index for information gain in the Bayesian framework. IFAC-PapersOnLine. 2020;53(1):634–639.
- 38. Goldbeter A. A minimal cascade model for the mitotic oscillator involving cyclin and cdc2 kinase. Proceedings of the National Academy of Sciences of the United States of America. 1991;88(20):9107–9111. pmid:1833774
- 39. Perelson AS, Kirschner DE, De Boer R. Dynamics of HIV infection of CD4+ T cells. Mathematical Biosciences. 1993;114(1):81–125. pmid:8096155
- 40. Zhang X, Trame M, Lesko L, Schmidt S. Sobol Sensitivity Analysis: A Tool to Guide the Development and Evaluation of Systems Pharmacology Models. CPT: Pharmacometrics & Systems Pharmacology. 2015;4. pmid:27548289
- 41.
Blum A, Hopcroft J, Kannan R. Foundations of Data Science. Cambridge University Press; 2020.