Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Large-deviations of disease spreading dynamics with vaccination

Abstract

We numerically simulated the spread of disease for a Susceptible-Infected-Recovered (SIR) model on contact networks drawn from a small-world ensemble. We investigated the impact of two types of vaccination strategies, namely random vaccination and high-degree heuristics, on the probability density function (pdf) of the cumulative number C of infected people over a large range of its support. To obtain the pdf even in the range of probabilities as small as 10−80, we applied a large-deviation approach, in particular the 1/t Wang-Landau algorithm. To study the size-dependence of the pdfs within the framework of large-deviation theory, we analyzed the empirical rate function. To find out how typical as well as extreme mild or extreme severe infection courses arise, we investigated the structures of the time series conditioned to the observed values of C.

Introduction

Due to the high relevance for the societies, the studying of the spread of diseases has long since become a very important problem in a variety of disciplines like biology, applied mathematics, statistics and statistical physics [15]. Beyond the analysis of the dynamics of existing diseases one main target is to understand how to fight them. One of the best ways to mitigate the impact of a disease is the application of vaccines, which have been successfully used to eradicate smallpox [6]. For other diseases like rabies [7] and measles [8], vaccines have also led to tremendous success, although this success is undermined by anti-vaccination movements [912]. The creation and evaluation of new vaccines is a delicate, complex and time consuming process [1315]. In particular, the safety and effectiveness of new vaccines has to be proven rigorously [16, 17].

To model the effect of vaccinations one has to first model the disease itself. A very famous and influential model, the Susceptible-Infected-Recovered (SIR) model, was introduced by Kermack and McKendrick in 1927 [18] who build upon previous research by Ross and Hudson [1923]. Initially the model was investigated for a fully-mixed population [24, 25]. This means, the disease can propagate in between all individuals with the same likelihood. In physics this corresponds to a mean-field model. Naturally those early studies ignored the effect of heterogeneity within the population and thus later studies [26, 27] have incorporated contact networks, where each node represents, e.g., an individual and the edges represent the contacts between them.

Note that the SIR process can be understood as special case of chemical reaction networks, where the fractions of susceptible, infected and recovered correspond to the concentrations of chemical species [28, 29]. In these networks the nodes typically denote the species while the edges describe the chemical reactions. Although diffusion of species is often included in the networks, this and all other reactions are usually described on mean-field level, similar to what was done for the first SIR models.

In its simplest form, vaccination can be modeled as granting perfect immunity to the disease, i.e., an individual that has been vaccinated cannot contract the disease forevermore. Here we consider the case that a vaccination is applied before the first infected individual is introduced in the network. This could be the case for a population of domestic animals which is partially vaccinated before the disease is first introduced to the population via a wild animal. This approach is equivalent to studying the disease dynamics on a new network, where all nodes that could successfully be vaccinated have been removed from the original network. Thus, the disease can only propagate within connected components [30] of the new network and one can model the vaccination as site percolation problem [31] and usually aims at reducing the network below its percolation threshold [30, 32].

This viewpoint means that one has to achieve a rather high fraction of vaccinated individuals and ignores the effects of stochastic fluctuations that can play an important role in disease extinction [33]. Therefore it makes sense to model the stochastic process that governs the disease propagation if we want to fully understand the impact of vaccines.

To gain an understanding of the impact a disease has before going extinct one can, e.g., look at the probability density function P(C) for the cumulative number C of infections, i.e., how many nodes contracted the disease over its lifespan. In order to obtain a comprehensive description of such stochastic processes, one should obtain the desired distribution over a large range of its support. For few very simple models this can be done analytically, but for most interesting cases one has to apply numerical simulations [34]. In order to access a distribution even in the large-deviation tails which exhibit probabilities as small as 10−50 or lower one has to go beyond typical-event simulation and has to use special large-deviation techniques [35], like the ones the authors of the present work have applied in a previous study [36] where we investigated the pure SIR model comprehensively. In particular we found that the large-deviation principle [3739] is fulfilled, which indicates that the model belongs to a standard class in large-deviation theory. Furthermore, our approach allowed us to obtain correlation patterns with other measurable quantities, which further deepens the understanding of the disease model and of different phases which are visible in the distribution P(C).

In the present work we study the impact of two types of widely investigated vaccination strategies, i.e., random vaccination and target vaccination where nodes with a high-degree are vaccinated first [40]. For an overview of different vaccination strategies and the state of research in general we refer to detailed review articles [41, 42]. In our work we aim at obtaining P(C) over a large range of the support, identify different phases in the distribution, relate these phases to different patterns of disease evolution and verify whether the mathematical large-deviation principle still holds.

The paper is structured as follows. First we introduce the SIR model and the network ensemble. Then we introduce the algorithms we applied, which includes the vaccination strategies. Next we briefly investigate the available parameter space to choose which parameters we use for the later analysis. We present the large-deviation analysis where we display the calculated probability density functions as well as corresponding rate functions and correlations with other quantities. We finish the paper with a summary and an outlook.

Model

The SIR model we apply is defined as follows. For a given network, each node is in either of four states Susceptible (S), Infected (I), Recovered (R) or Vaccinated (V), where the vaccinated state is equivalent to the recovered state and the distinction is just made for our convenience.

To begin a simulation we assign all nodes the state S and then first apply one of the vaccination strategies that will be later discussed in Sec Vaccination strategies to assign a subset of nodes the vaccinated state V.

Then, different from the case without vaccinations [36], we choose 5 initial infected nodes, randomly and uniformly drawn from all possible nodes without allowing duplicates. Thus, the not so relevant case that just by chance the initial infection dies out quickly is much less common. This means, the trivial peak of P(C) at C ≈ 0 is much less prominent in our results. Note that vaccinated nodes can also be chosen among the initial infected one. In this case initial infection takes prevalence for the initialization, to obtain actually 5 infected nodes. However, for the later times of the disease dynamics, vaccinated nodes can never be infected, as desired.

The actual disease dynamics is performed as follows: At each time step τ we iterate over all susceptible neighbors of all currently infected nodes. Note that this implies susceptible nodes with multiple infected neighbors will be iterated over multiple times. With the transmission probability λ we flag them to become infected. Next we iterate over all currently infected nodes again, this time we let each of them recover with the recovery probability μ. To finalize the time step we infect all nodes that were flagged to become infected.

At time step τ we denote the current fraction of infected and recovered nodes by i(τ) and r(τ), respectively. From those fractions we can obtain the cumulative fraction of infected nodes, i.e., c(τ) = i(τ) + r(τ).

This can also be used to create an indicator for the severity of a disease outbreak by defining (1) where the latter equality obviously only holds for μ > 0.

Network ensemble

We chose a small-world network [43] to represent our contact network. This is motivated by the fact that many real-world contact-networks exhibit small-world properties [44].

The creation of a network with N nodes works as follows. First a initial ring structure is created where each node i is connected to its next and second next neighbors given by i + 1 mod N and i + 2 mod N, respectively. Since all edges are undirected this means each node starts with an initial degree of 4.

Next, we iterate over all edges {i, j} and rewire them with probability p to a random node j′ ≠ i, i.e., we exchange {i, j} → {i, j′}. Throughout this paper we use a rewiring probability of p = 0.1. Note that we only draw a random node for j because each edge is “rooted” at one node and thus the minimal degree a node exhibits is 2. This rewiring is responsible for the small-world properties of the network as it introduces shortcuts and leads to a rather small diameter [30], i.e., from each node one can reach any other node via a short path of edges.

Algorithms

Large deviation algorithms

We apply special large-deviation algorithms [45]. Such algorithms exist since the 1950s [46]. In statistical physics their popularity first increased for studying the dynamics of molecules via transition path sampling [47, 48]. Since then, such algorithms were applied for many different models, e.g., the resilience of power grids [49, 50], random walks [5154], random graph properties [5558], longest increasing subsequences [59, 60], ground states of Ising spin glasses [61], and the Kardar-Parisi-Zhang equation [62].

The general idea for the SIR model is that the large-deviation simulation embraces the SIR simulation [36], i.e., it is able to manipulate the SIR dynamics in a controlled fashion to get access to the desired properties.

This manipulation works as follows: In a standard SIR simulation one would draw random numbers that are uniformly distributed in [0, 1] on demand. These numbers are then used for comparisons with the recovery probability μ or the transmission probability λ to decide whether a node becomes recovered or infected. Instead of creating the random numbers on demand one could, of course, draw them beforehand and store them in the two vectors ξμ and ξλ, such that these vectors contain a distinct random number for each node and each time step τ.

As long as the random number vectors contain enough numbers such that the disease outbreak dies out before running out of numbers, this in and of itself will not change the outcome. We refer to [36] for a detailed explanation of the estimation of the required vector length. Note that the general idea remains the same as in the referenced work, but we have slightly modified the criterion by increasing the threshold and factor compared to our previous study.

Before the SIR simulation is started the vaccination is applied, which essentially equals the removal of the vaccinated nodes. The specifics of how the nodes for the vaccination are chosen depends on the applied heuristics. All heuristics, however, require a vector ξord to give a deterministic outcome, as explained in Sec Vaccination strategies. This vector is created by shuffling the vector [0, 1, …, N − 1] of node IDs.

Lastly we also need to decide which nodes will be set to the infected state before the SIR simulation is started, i.e., which nodes should be “patient zero”. For this we create the vector ξ0 by shuffling a new vector initialized with [0, 1, …, N − 1]. In this work we always start with 5 initial infections. Thus the first five entries of the vector determine the initially infected nodes.

To be able to sample rare events we now control the values contained within the vectors (ξμ, ξλ, ξord, ξ0) with a Markov-Chain-Monte-Carlo (MCMC) approach that is sitting on top of the SIR simulation, i.e., the outcome of the SIR simulation now depends deterministically on the random numbers encountered during the MCMC. This MCMC is set up in a way to allow the estimation of a large range of the probability-density function (pdf) P(C) for a given network. To be more precise we apply the 1/t Wang-Landau (WL) algorithm [63], which is a slight modification of the original Wang-Landau algorithm [64] that prevents error saturation [63, 6567].

Since the approach we use here is very similar to the one we used for [36] we will only outline the general idea. The WL approach starts with a non-normalized estimate for the probability distribution, i.e., P(C) = 1 ∀C. Each step t = 0, 1, … of the Markov-Chain consists of the generation of a new trial configuration , which is generated from the current configuration as explained in Sec MCMC Moves. Since both configurations of random vectors deterministically determine the outbreak courses they correspond to, they also determine the resulting cumulative fractions C′ and C(t) of infections, which are calculated by simulating a complete run of the SIR simulation. The trial configuration is accepted, i.e., Ξ(t+1) = Ξ′ and thus C(t+1) = C′, with the Metropolis-Hastings probability min{1, P(C(t))/P(C′)}. If the trial configuration is not accepted we keep the old configuration, i.e., Ξ(t+1) = Ξ(t).

A multiplicative factor f > 1 is used to update the pdf estimate at C(t+1), i.e., P(C(t+1)) = fP(C(t+1)). All other values of P(C) are not changed for the current step. Due to this the probability of generating the trial configuration with value C becomes, in the long run, inversely proportional to the probability P(C) of occurrence. This allows for a rather uniform sampling of the space of C values while the pdf estimate is continuously refined. Usually a rather large factor f = e ≈ 2.71 is used initially, which forces a quick but rough estimate of P(C). Then some schedule is used to iteratively reduce f during the process to change the estimate on an increasingly finer scale. Here lies the main difference between the 1/t WL algorithm we applied [63] and the original WL algorithm [64].

This method is able to measure very rare events that are inaccessible via typical-event sampling methods, also called simple sampling, and thereby allows for the sampling of distinct features of the pdf over a very large range or even the full support.

Since P(C) is updated continuously, the WL method does not strictly fulfill detailed balance [68]. To ensure a correct statistics in the end, we subsequently employ entropic sampling [69]. This approach uses the pdf estimated by WL as a starting point and works as the WL approach, but P(C) is not updated any more. The obtained histogram for the encountered values of C is used to finally correct the pdf obtained by WL. This slightly refines the accuracy of our measurements even in the region of very low probabilities like 10−100. But more than that it enables us to more evenly sample disease infection time series over the range of C values. Overall this rigorous numerical approach ensures high confidence in the results and has proven very successful in the past.

MCMC moves

Here we will explain in detail how the trial configuration is created.

With a probability of 1% a rotation move is performed. That means that the vectors ξμ and ξλ are rotated by N elements to the right (50%) or left (otherwise), which roughly corresponds to shifting the resulting time series by one time step to the left or right. Note that this can be implemented quite efficiently by only storing a offset variable and thus removing the need to actually copy a lot of memory for the operation.

With a probability of 0.5% we perform a exchange patient move, which is done by drawing a ∈ {0, …, 4} and b ∈ {5, …, N − 1} and then swapping the values ξ0[a] ↔ ξ0[b], which effectively exchanges one of the initial infected nodes.

With a probability of 2% we perform a walk patient move that was found to improve convergence. The basic idea of this move is letting one randomly chosen patient of the initial patients perform one step of a random walk on the underlying (not vaccinated) network structure while making sure that each node has the same probability of occurring. This is ergodic with respect to the positions of the initial patients because we demand a connected network for this study in the first place.

For this move, we first draw one of the initial patients p0 ∈ {ξ0[0], …, ξ0[4]}. Then we draw a uniformly distributed random number u ∈ [0, 1]. Let A be the adjacency list of p0, i.e., a list that contains every node p0 is connected to via an edge. Let dmax be the maximal degree of the network. Then the new patient zero becomes (2) and ξ0 is changed accordingly. Note that the finite probability of nothing changing is necessary to ensure detailed balance: The naive approach of always moving the patient would bias the system in such a way that nodes with higher degree would have a higher probability of being chosen.

With a probability of 1% we do a reset start move, i.e., we redraw all random numbers from ξμ, ξλ that are associated with the time step τ = 0. This move was included since our testing suggested that it improves the convergence of the algorithm.

Lastly, if none of the above moves was selected, which happens with a probability of 95.5%, we just randomize the dynamics a bit. Thus, we repeat the following 3000 times: Draw a uniformly distributed random number u ∈ [0, 1], draw one of the vectors ξ ∈ {ξμ, ξλ} and uniformly choose a random index i of the vector to exchange the random number of the entry, i.e., ξ [i] = u. The number 3000 of times was chosen to achieve an acceptance rate of roughly 50% of this move.

Histogram interval width

Since we are vaccinating at the beginning of the SIR simulation and our vaccination assumes an effectiveness of 100% this means that we have to estimate how many nodes the disease can still reach. It is, e.g., possible that some nodes are surrounded by vaccinated nodes and are therefore unreachable by the disease, if the disease does not start on them. From an algorithmic-technical point of view, we need to determine a maximum for C because Wang-Landau will not work if a bin of the histogram is impossible to reach.

So we initialize the system and employ a greedy heuristics that only uses part of the Markov chain. For each step of the Markov chain we calculate the largest connected component that contains an initial infection by first removing all vaccinated nodes except those chosen as initial patients and then using a depth first search starting with the patients zero. Then we do a Markov move, where we only allow moves that change ξord or ξ0 and calculate again. We reject the move if the size of the largest connected component decreased.

After 50000 steps we take the maximal size of the largest connected component that we encountered as estimate. If we later notice that the large deviation algorithm is unable to reach the largest C values we remove the bins that the algorithm was not able to reach and rerun the simulation, which happened only a few times.

Note that nodes that are vaccinated can still be chosen as patients zero and in that case are not removed from the network because we consider that the vaccination was performed too late and was therefore ineffective.

Vaccination strategies

We denote the total number of vaccination doses with Nv and its fraction by nv = Nv/N, where N is the total number of nodes, as usual.

We consider a total of three vaccination strategies:

  1. A uniformly random vaccination heuristics. It is the simplest vaccination strategy [40] that does not assume any knowledge of the topology whatsoever. For this we randomly and uniformly draw Nv nodes, while not allowing duplicates. As explained in Sec Large deviation algorithms, the large-deviation scheme will supply the vector ξord as randomness input for the vaccination schemes. Conveniently we can obtain the indices of the nodes we want to vaccinate by just taking the first Nv entries of ξord.
  2. A natural choice for a targeted vaccination strategy is to mainly target highly connected nodes [32] which is the main idea of the high degree heuristics. For this we take the Nv nodes with the highest degrees possible.
    Note that the choice of the Nv nodes is rarely unique as usually there are many nodes with the same degree. Thus, the choice of the small-degree vaccinated nodes must be done by some algorithm. In this case we randomly draw from the list of nodes with the corresponding degree. This can be accomplished by selecting nodes of the same degree in an order as given by the vector ξord, starting with the nodes of highest degree. Note that for this heuristics to be efficient, the degree distribution is obviously not allowed to be homogeneous. This vaccination scheme has been proven to be especially effective for scale-free networks [70].
  3. The adaptive high degree heuristics is taking the idea one step further. This time we first take the node with the highest degree (if this is not unique we randomly and uniformly choose one of the possible nodes by using the ξord vector). Then, and this is the only difference to the previous strategy, we reduce the degree of all of the neighbors of the node that is now set to be vaccinated. This process is repeated until Nv nodes are chosen. In the literature this heuristic is also called recalculated degree removal [71].

Simple sampling

Before performing the relatively expensive large-deviation simulations we wanted to determine suitable values of the parameters to investigate. We chose to keep μ = 0.14 from our previous study [36] and chose a rather large value of λ = 0.4 for the transmission probability, which corresponds to a parameter-space location within the epidemic phase if it was without vaccinations.

For each network size we want to study later we randomly drew one network, which means that we were very likely to obtain a typical network of our ensemble, and measured within simple-sample simulations the average and its variance σ2(C) for different fractions nv of vaccinated nodes. The results for N = 3200 are shown in Fig 1 as example.

thumbnail
Fig 1. Average cumulative fraction of infected nodes for a typical network of N = 3200 nodes for different fractions nv of vaccinated nodes and the three applied vaccination strategies.

The inset shows the variance σ2(C). The data is obtained by averaging over 10000 samples per point and error bars are not shown as they are smaller than symbol size. The dashed lines indicate the positions of the peaks of the variance, respectively.

https://doi.org/10.1371/journal.pone.0287932.g001

As expected the random vaccination strategy gives the worst results and requires the largest fraction of vaccinated nodes to achieve a certain value of . The adaptive high degree heuristics is just slightly better than the not-adaptive one only in a very small window around nv ≈ 0.1. On the other hand the non-adaptive one is better in the interval nv ∈ [0.1, 0.23]. There is basically no difference between the two anywhere else. Below we discuss how the counter-intuitive behavior that the adaptive vaccination strategy in some cases performs worse can be understood.

But first we discuss how we obtained critical threshold values . We fitted Gaussians (not shown) to the peak regions of the variance curves to find the exact position of the peak, which we tread as network-specific critical vaccination values . In Fig 1 these critical values are visualized by the dashed lines. Note that the worse performance of the adaptive strategy on the interval [0.1, 0.23] results in a higher value of for this strategy.

We have performed this type of simulation and analysis for all networks that we want to study with the large-deviation algorithm. Since we always used the exact same networks when comparing different strategies, we use only one network for each simple-sampling study per system size N that is considered later on.

Note that within our preliminary simulations we also tried to average over multiple networks for each considered size N and then use finite-size scaling to find the critical vaccination doses for an infinite system, similar to what we did to obtain λc in [36]. We found out that the actual distributions such as P(C) still depend somehow on the actual network, especially for the high degree heuristics. Therefore, the infinite-size critical vaccination does not help much in selecting suitable values of nv. Thus, we opted to use network-specific values of , which fluctuate slightly: We obtained critical values for N ∈ [1414, 6400] and found for the random heuristics, for the adaptive high-degree heuristics and for the non-adaptive high degree heuristics.

To understand why the adaptive strategy is performing worse than the non-adaptive one we initialized the network as if we wanted to perform an outbreak simulation. We then removed all vaccinated nodes and calculate the size of the remaining connected components and summed up the size of the components that contained any infected nodes, as this size represents an upper limit for the number of nodes that the disease can reach. We denote this quantity with S. The results are displayed in Fig 2, where we also include an upper bound which is given by the total number of remaining non-vaccinated nodes.

thumbnail
Fig 2. Sum S of relatives sizes of the connected components that contain the initial infections, shown as function of nv.

Results are for a network of size N = 3200, for the different vaccination strategies. Panels (a), (b) and (c) contain the results for different rewiring probabilities p ∈ {0, 0.1, 1}. Each data point was averaged over 20000 samples. Error bars are smaller than the symbol size. We also include an upper bound for the cluster size, which is the fraction of the network that was not yet vaccinated.

https://doi.org/10.1371/journal.pone.0287932.g002

In panel (a) of the figure we see the results for the actual value p = 0.1 of the rewiring probability used throughout the paper. Clearly the size S decreases more rapidly for the non-adaptive high degree heuristics compared to the adaptive one.

By looking at (b) we can understand why. A rewiring probability of p = 0 means that we never rewire and thus retain the original ring structure the network is initiated with. Here every node has degree 4, which means that the non-adaptive high degree heuristics and the random vaccination heuristics are equivalent. For this reason, these two cannot be distinguished in the figure. The adaptive high degree heuristics however will, due to its adaptive nature, distribute the vaccinations more or less evenly over the ring, as nodes whose neighbors have been vaccinated have a reduced degree. This makes splits of components less likely. This leads to S(nv) following the upper bound for quite some range of values of nv before the network finally splits up into more components.

In contrast if we look at (c), where we chose a rewiring probability of p = 1, which results in random networks similar to Erdőes-Rényi [72] networks with the difference being that due to the construction each node has at least degree 2, we see a different picture. The underlying ring structure is destroyed. Therefore the adaptive strategy performs better leading to reduced values of S as compared with the other heuristics.

Note that in all three cases for any value of nv by using adaptive high-degree heuristics more edges are effectively removed than when using the other two strategies (not shown), which corresponds to the intuitive expectation. As we have seen, it is only the special small-world network structure which makes the adaptive vaccination strategy perform in some range of values for nv worse than the non-adaptive high-degree strategy. A similar effect, i.e., that the average inverse geodesic length, which can be used to quantify the severity of vertex removal in other contexts, of the adaptive high-degree heuristics can perform worse than the one of the non-adaptive heuristics, was observed before [71].

Results: Large deviation sampling

By using the large-deviation approach we can determine the probability density function of interest, i.e., P(C), over a large range of or even on its full support.

In Fig 3 we show the probability density functions for the random heuristics for different system sizes N measured at their respective critical vaccination fraction .

thumbnail
Fig 3. Probability density of the cumulative number C of infections for different system sizes N at their critical vaccination doses with the random vaccination strategy as measured by the large-deviation algorithm.

It also includes simple sampling data for the largest system size. Linear scale in the inset.

https://doi.org/10.1371/journal.pone.0287932.g003

The pdf exhibits two peaks, one near C ≈ 0 one around C ≈ 0.25. Both peaks move towards smaller values for increasing system size and become sharper. We can also see that the data gathered by large-deviation sampling agrees very well with the simple sampling results—at least in the range where simple sampling was able to reach.

From the pdfs we can calculate the empirical rate functions (3) where is a constant that shifts the resulting rate function such that they all have their respective minimum at Φmin ≡ minC Φ(C, N) = 0. The idea behind calculating the rate function is to verify whether the probability density functions obey the so called large-deviation principle as described by large-deviation theory [35, 3739]. If it obeys this principle the shape of the pdf has to follow the standard finite-size behavior (4) Note that the size N is primarily important as a linear prefactor to the rate function, while other dependencies on the size are less important, as denoted by o(N), i.e., their contributions vanish relatively in the limit N → ∞. From Eq (4) it is also clear why at the minimum Φmin = 0 should hold, because if Φmin > 0 the probability would converge to zero everywhere while if Φmin < 0, P(C) would diverge for some values of C. To check the large deviation principle is of interest, because models for which the principle holds are better accessible by analytical approaches, e.g., by using the Gärtner-Ellis theorem [35, 3739].

The rate functions we calculated from the pdfs of Fig 3 are shown in Fig 4. In the range C ∈ [0, 0.25] it looks like the functions converge to a limiting shape for N → ∞. Right to the second minimum of the rate function, which corresponds to the maximum of the pdf at around C ≈ 0.25, the case is less clear and some of the functions for different system sizes even overlap. If we shift each function horizontally such that their minimum appears at the same value of C (not shown) the functions exhibit a monotonous N-dependence also in the region of large values of C, indicating a convergence as well.

thumbnail
Fig 4. Rate functions Φ(C) for the random vaccination heuristics for different system sizes N at their respective critical vaccination.

The gray inset shows the position of the minimum of the rate function as a function of N as well as the fit fit-exp with a = 0.142(5), b = 0.00044(4) and . The other inset shows a zoom for better visibility.

https://doi.org/10.1371/journal.pone.0287932.g004

The position of the minimum is also shown in Fig 4 and seems to follow an exponential function. We therefore fitted the function (5) which worked pretty well.

Additional pdfs and rate functions for the other heuristics at their respective critical points look very similar and are thus moved to S1 Appendix.

In Fig 5 we show the comparison of the different vaccination strategies at their respective critical vaccination value. The probability density functions for both high-degree heuristics are almost identical even though the adaptive high-degree heuristics vaccinated about 3% more nodes than the non-adaptive one. The shape for the random heuristics is a bit different. The peaks of P(C) for this vaccination heuristics are shifted to the left and appear sharper as compared to the two high-degree heuristics. Also the probabilities for large values of C are significantly smaller and the decline in probability is steeper.

thumbnail
Fig 5. Probability density functions of the cumulative fraction C of infections for N = 3200 for the different vaccination heuristics at their respective critical point.

For the adaptive high-degree heuristics that corresponds to Nv = 530, for the non-adaptive one to Nv = 424, and for the random heuristics to Nv = 1133. Linear scale in the inset.

https://doi.org/10.1371/journal.pone.0287932.g005

In Fig 6 we compare all three heuristics with the same vaccination doses Nv = 530, which is the critical vaccination value for the adaptive high-degree heuristics for the system size N = 3200. Looking at the different pdfs we can see three different general shapes. The pdf of the non-adaptive high degree heuristics exhibits one maximum at C ≈ 0.05 and strongly declines afterwards. The pdf for the adaptive high degree heuristics exhibits two maxima that are comparable in height (better visible in S1 Appendix) and a sharp decline in probability after the second one. Lastly the random heuristics also exhibits two maxima, however the second maximum is at a very large C value and the first peak is at noticeably lower probabilities. The probability also declines below 10−20 in between both maxima, i.e., they are much stronger separated.

thumbnail
Fig 6. Probability density functions of the cumulative fraction C of infections for N = 3200 and Nv = 530 for the different vaccination strategies.

Linear scale in inset.

https://doi.org/10.1371/journal.pone.0287932.g006

Overall we can see that the non-adaptive high-degree heuristics proved very effective, followed by the adaptive high degree heuristics whereas the random heuristics is the least-efficient of them.

Now we want to see how the distributions look like when we choose values of Nv such that the distributions exhibit a maximum roughly at the same values of C, i.e., where the non-adaptive high degree heuristics exhibits its maximum in Fig 6. This allows us to compare the mere shape of the distributions.

The achieve this, we first performed a few simple sampling simulations for various values of Nv to align the C positions of the maxima before running the large-deviation simulations. It turned out that it is not possible to find a value of Nv to align the random heuristics well, because here the peak position is too insensitive to the value of Nv, as demonstrated by Fig 7. Thus we had to compromise for the random heuristics. The measured pdfs are displayed in Fig 8.

thumbnail
Fig 7. Probability density functions of the cumulative fraction C of infections for the random heuristics, N = 3200, and for different number Nv of vaccinated nodes.

The histograms were measured via simple sampling using 8000000 samples each. Linear scale in the inset. The dashed line indicates the position of the maximum that we aimed for.

https://doi.org/10.1371/journal.pone.0287932.g007

thumbnail
Fig 8. Probability density functions of the cumulative fraction C of infected nodes for N = 3200 for the different vaccination strategies.

The non-adaptive high-degree heuristics used Nv = 530, the adaptive one used Nv = 670 and the random heuristics used Nv = 1240. Linear scale in inset.

https://doi.org/10.1371/journal.pone.0287932.g008

Both high-degree heuristics display a very similar behavior, whereas the random heuristics displays two peaks in the range of small values of C and exhibits a much steeper decline in probability for increasing values of C resulting in much smaller probabilities for the tail. Thus, while exhibiting about the same typical number of infections, strong fluctuations are much more reduced for the random heuristics, at the price of a much larger number of vaccinated nodes.

With the same idea in mind we now wanted to align the maxima to the position of the maximum for the random heuristics from Fig 6. Again, we used simple sampling to align the maxima and did large deviations afterwards. This time no issue occurred during the alignment and the results of the large-deviation simulations can be found in Fig 9.

thumbnail
Fig 9. Probability density functions of the cumulative fraction C of infections for N = 3200 for the different vaccination strategies.

The non-adaptive high-degree heuristics used Nv = 315, the adaptive one used Nv = 305 and the random heuristics used Nv = 530. On the left we show the data in logarithmic scale and on the right the range C ∈ [0.65, 0.85] in linear scale.

https://doi.org/10.1371/journal.pone.0287932.g009

Again both high-degree heuristics give very similar results. The general shape of the pdf for the random heuristics is also similar, and all curves exhibit two maxima at roughly the same positions and with the same magnitudes. Apart from that, for the random heuristics the probability to observe intermediate values C is much lower as compared to the other two heuristics, i.e., the fluctuations away from typical values are again much more reduced.

Infection course analysis

In previous work on the SIR dynamics without protections measure [36] we have observed that the infection dynamics are remarkably different for different regions of the distributions P(C). This allowed us to pin-point characteristic properties of mild, typical and severe outbreaks. This might help to design or improve suitable protection measures. To investigate whether such characteristic properties can be identified in the presence of vaccinations, we stored, during the entropic sampling, every X’th disease time series, where we calculated X in such a way that we store 2000000 time series in total per entropic sampling simulation. These disease time series were binned according to their respective value of C.

Disparity.

To compare the time series with one another we now computed their disparity [36], which is defined as follows. Let us first focus on two arbitrary time series X1, X2, where X(τ) corresponds to c(τ) or i(τ) with the only difference that we normalize them by dividing every entry X(τ) by the maximum, i.e., maxτX(τ), with the intention of only investigating the shape and not the magnitude of the curves. Note that the time series X1 and X2 may have different lengths l1, l2 and we thus define lmax = max{l1, l2}. If necessary we continue a time series beyond its given length by repeating its last value, which makes sense because for larger time usually no further changes occur. In this way both series have lmax entries.

This lets us define the distance (6)

We now define the disparity VX(C1, C2) for two different values of C as averaged distance d. The average is performed over all pairs of time series where time series X1 exhibits the bin value C = C1 and the other C = C2, respectively.

In Fig 10 we show the disparity heat maps for the three vaccination heuristics as calculated from the normalized i(τ) time series. The pdfs P(C) that correspond to these disparities were already displayed in Fig 6. Note that the diagonals represent the comparisons within a single bin, i.e., of time series that lead to the same cumulative number of infections C. Thus, they represent the amount of variation of the time series for a given value of C.

thumbnail
Fig 10. Color-coded disparity Vi(C1, C2) of the i(τ) time series for N = 3200, Nv = 530.

On the left we show the results for the non-adaptive high-degree vaccination, in the middle the results for the adaptive high-degree vaccination and on the right the results for the random vaccination. The time series are binned with their corresponding value of C and for each bin 1500 randomly drawn time series are used for the calculation of the disparity Vi(C1, C2). For the shaded area no time-series exist as it was outside of the interval used for WL.

https://doi.org/10.1371/journal.pone.0287932.g010

Looking at the heat map for the non-adaptive high degree heuristics we make a few observations. Along the diagonal, the disparity increases with growing value of C. Thus two different disease-evolutions which exhibit the same value of C are more similar if C is small. Apart from that one can see a cone-shape of the boundaries between similar values of V.

The heat map for the adaptive strategy in Fig 10 looks rather similar. Though this time, along the diagonal, the time series are becoming more dissimilar until C ≈ 0.3 and become more similar again if C is increased further, though this effect is very small and almost not visible.

The heat map for the random heuristics, displayed on the right in Fig 10 looks different. Here we can see three regions, the first, C < 0.1, is a region of curves that are similar to one another. The second, 0.1 ≤ C < 0.35, consists of a slightly higher disparity and the last region, 0.35 ≤ C, consists of curves that are very similar to one another. Thus, also for this quantity of measurement, which tells more about the details of the disease evolution, the random heuristics leads to smaller fluctuations as compared to the degree-based ones. Interestingly, when fixing C1 at a small value like 0.05 and varying C2, one observes that first the disparity increases, but at much larger values of C2 the shape of the time evolution become more similar again. This resembles “reentrant” behavior sometimes observed for phase transitions.

In Fig 11 we show some examples of the time-series used to create the heat maps to better understand the behavior we observed when analyzing the disparities.

thumbnail
Fig 11. Infection time series i(τ) for different values of C.

These time series were created during the entropic sampling with a Network of size N = 3200 and a vaccination doses of Nv = 530. On the left we show the results for the non-adaptive high-degree heuristics, in the middle for the adaptive high-degree heuristics and on the right for the random heuristics. For each value of C we plot 30 time series. We also included one arbitrary time series for each C value that we highlighted for clarity.

https://doi.org/10.1371/journal.pone.0287932.g011

On the left we see the curves for the non-adaptive high degree heuristics. For small values of C, i(τ) peaks close to the beginning before the disease quickly goes extinct. For larger values of C the time-series display a strongly fluctuating behavior, where no clear peak exists and the disease takes way longer to go extinct, though the actual duration varies a lot. Note that strong variations make the evolution of diseases very hard to predict. On average, as the value of C increases, the duration of the outbreak and the maximum number of simultaneously infected nodes also increase. However, the maximum number of affected nodes is relatively small, with at most 6% of the network affected at once

In the center of Fig 11 we present the time series corresponding to the adaptive heuristics. Compared to the plot on the left we see a tendency for the curves to peak earlier, i.e., within the first 100 time steps, and the disease dies out quicker overall. Also some remnants of the strongly fluctuating behavior of the left plot can be seen. However, for the adaptive heuristics, the most likely trajectories are associated with higher values of C in comparison to the non-adaptive heuristics meaning that the fluctuations themselves correspond to more probable outbreak scenarios.

The time-series corresponding to the random heuristics are displayed in Fig 11 on the right. We see that the time series for large C display a clear peak and Gaussian-like shape. Still, the time series for values of C near C = 0.3 are still a bit more fluctuating and correspond to slowly-developing diseases as were observed also for the case without vaccinations [36]. These outbreaks tend to take the longest to become extinct. Overall the heat map and time series that are seen for the random case here look very similar to the ones without vaccinations [36]. Obviously, random vaccination alters the structure of the network of non-vaccinated nodes not very strongly.

Conditional densities.

For the health care system it is important to estimate the total number of people that will be infected at once or rather the maximum of this quantity, because this should not exceed the capacity of the healthcare resources. We therefore look at the maximum (7) To investigate how M correlates with the cumulative number of infections we calculate the conditional density ρ(M|C), i.e., the probability for a disease outbreak to exhibit a specific value of M given the value for C. In Fig 12 we show this conditional density for N = 3200 and Nv = 530, i.e., calculated during the same simulation we used to generate Fig 6, such that we can use that figure to know how probable a given range of C values is. Note that most parts of ρ(M|C) are for values of C where P(C) is extremely small, i.e. they are accessible only by using the large-deviation approach.

thumbnail
Fig 12. Color-coded conditional densities ρ(M|C) for a network of N = 3200 nodes and Nv = 530 vaccination doses for the three analyzed vaccination strategies.

The blue stripes indicate the C range that was outside of the Wang-Landau interval.

https://doi.org/10.1371/journal.pone.0287932.g012

For the non-adaptive high degree heuristics the system is beyond its critical vaccination value and C < 0.1 is highly likely. Since the disease dies out quickly in this range, it results in very small values for M. Only for very small values of C, the maximum M increases strongly. Even if we look at the less likely outcomes, M always stays below 200.

The density for the adaptive heuristics looks up to C ≈ 0.4 very similar to the non-adaptive one. For larger values of C, outbreaks tend to lead to higher values of M, though overall M stays below 400. When considering the pdf in Fig 6, we can conclude that M < 200 is the most likely outcome here and the large values of M are not so relevant when deciding about the capacities of health-care systems.

For the random heuristics up until C < 0.25 its conditional density is very similar to the other two heuristics. Then, however, M almost linearly increases with C and the incline is much steeper than seen for the high-degree heuristics. For the random heuristics C > 0.7 is the most probable range and thus a typical outbreak exhibits M > 600, leading to a much higher load on the health-care system.

Next, we wanted to investigate the speed with which the disease propagates through the network. One could investigate the entire duration of the disease, but it makes more sense to go about it differently. In the real-world it is unlikely that the disease will get detected immediately. We want a quantity to reflect that. We use the c(τ) time series and measure the number of time steps τ1 it takes until c(τ1) = 0.1 × c(∞) and the number of time steps τ2 it takes until c(τ2) = 0.9 × c(∞) is reached. The we define , i.e., we measure the number of time steps the disease required to reach 90% of its final cumulative number of infections once it already reached 10%, for each run.

We calculated the conditional densities for the different vaccination strategies for N = 3200 and Nv = 530 and show the results in Fig 13. Again we can use Fig 6 to see how probable each value of C is.

thumbnail
Fig 13. Conditional density that shows the probability of for any given value of C for N = 3200 and Nv = 530 for the different vaccination strategies.

The blue stripes indicate the C range that was outside of the Wang-Landau interval.

https://doi.org/10.1371/journal.pone.0287932.g013

All strategies exhibit a similar incline for C ≤ 0.05. The disease durations are rather short and do not fluctuate much. This can be seen as one specific phase within range of possible C values.

On the other hand, for the non-adaptive strategy we see a large spread of duration times for C > 0.05 and this spread increases for increasing value of C. This means it will be hard to predict the duration of this disease even if the value of C that will be reached in total was known. This occurrence of a strong-fluctuation phase fits with the observations from Fig 11 where we saw a fluctuation-rich behavior for the corresponding time series.

The adaptive strategy shows a similar behavior for C > 0.05, though the durations tend to be a bit lower and the spread does not become as large as for the non-adaptive case. For C > 0.5 the durations tend to become a tiny bit shorter again and the spread of duration times also decreases, but only slightly. This correspond to the slightly higher values of M, i.e., the diseases are more concentrated in time.

The spread of durations for the random heuristics is much smaller overall and peaks at around C ≈ 0.2, which is in a rather improbable range. For C > 0.5 the spread is almost non-existent and becomes constant, which is also reflected in Fig 11 where we can see a very regular behavior. This can be seen as the presence of a third phase for large values of C, in contrast to the two degree-based heuristics.

Thus, for the random heuristics the disease outbreaks are stronger, but can be predicted better due to smaller fluctuations.

Summary and outlook

We investigated the effect vaccinations on the dynamics of disease spreading for the standard SIR model. We considered three different vaccination strategies, namely a random vaccination heuristics and two different heuristics which target high-degree nodes first.

For this purpose we chose a rather large transmission probability of λ = 0.4 such that the model would, without vaccination, exhibit a strong epidemic outbreak.

Using large-deviation methods and especially the 1/t Wang-Landau algorithm [63] we were able to calculate the probability distribution P(C) of the cumulative fraction C of infected nodes over a large range of its support and down to probabilities as small as 10−80. While probability densities such as 10−10, which are out of range of standard approaches, might be even useful for practical considerations, obtaining P(C) over a large range of the support is in particular satisfying from a scientific point of view, because we are able to obtain a quantity which is analytically not accessible.

Note that some results exist for chemical reaction networks [28]. Since they are treated typically on mean-field level, the corresponding distribution functions usually have a simple structure, like a multidimensional simple Gaussian. Still, there are also known cases where distributions with multiple peaks have been found, like for a simple four-species reaction network [29]. Therefore the multi-modal structure of the distributions we observe might not be due to the complex underlying network, although all details of P(C) will certainly depend on the network structure.

In particular we investigated the empirical rate functions and found that the convergence properties were less clear than in the pure SIR case [36]. Nevertheless, when considering that the finite-size dependence includes horizontal shifts on the C axis as well as changes on the probability axis, it appears likely that the large-deviation principle is fulfilled. Thus it appears likely, that analytical calculations of the probability densities are possible with, e.g., the Gärtner-Ellis theorem.

By additionally applying an entropic sampling algorithm [69], we were able to investigate the actual time-series corresponding to different ranges of the value of C. Having available these sample time series of typical as well as extreme behavior might allow one to understand strong or weak outbreaks better which could lead to more sophisticated strategies.

We compared the time-series of the three vaccination strategies given the same network and same vaccination amount. We observed that for most of the cases with targeted high-degree vaccination a behavior with strong fluctuations dominates. This effect itself is less pronounced for the adaptive heuristics. For the random heuristics, in principle three phases can be observed along the C axis. In general, fluctuations are mostly suppressed, making disease propagation better predictable, although the strategy is less efficient as compared to the degree-based ones.

For practical applications, one could draw from these results the strategy to use a targeted high-degree heuristics to vaccinate enough nodes to exclude a pandemic, but to additionally vaccinate a certain fraction of randomly chosen nodes, to reduce the fluctuations such that the infection dynamics becomes more regular and predictable. This would probably also increase the acceptance of vaccination measures in the society.

The approaches used here is very general, such that in can be applied for many different disease-spreading problems, such as those exhibiting different infected states, including the implementation of other counter measures, or processes which take place on dynamic networks.

Furthermore, the large-deviation approach could be even more useful, if one considers coupled networks of multiple species. This would allow one to investigate the transfer of zoonoses between animals and humans. In this way one could study the fundamental processes or conditions that are required to let a disease transfer between different host species and possibly result in a huge outbreak in the new host species. Note that for relevant outbreaks these transfer probabilities are necessarily small, because for the cases with a high transfer probability, they will have happened during evolution already. Thus, a large-deviation approach is very suitable in this case.

References

  1. 1. Salathé M, Jones JH. Dynamics and Control of Diseases in Networks with Community Structure. PLOS Computational Biology. 2010;6(4):1–11. pmid:20386735
  2. 2. Apolloni A, Poletto C, Ramasco JJ, Jensen P, Colizza V. Metapopulation epidemic models with heterogeneous mixing and travel behaviour. Theoretical Biology and Medical Modelling. 2014;3. pmid:24418011
  3. 3. Wang W, Tang M, Stanley HE, Braunstein LA. Unification of theoretical approaches for epidemic spreading on complex networks. Reports on Progress in Physics. 2017;80(3):036603. pmid:28176679
  4. 4. Walters CE, Meslé MMI, Hall IM. Modelling the global spread of diseases: A review of current practice and capability. Epidemics. 2018;25:1–8. pmid:29853411
  5. 5. Chen Y, Liu F, Yu Q, Li T. Review of fractional epidemic models. Applied Mathematical Modelling. 2021;97:281–307. pmid:33897091
  6. 6. Bazin H. The Eradication of Smallpox: Edward Jenner and The First and Only Eradication of a Human Infectious Disease. Academic Press; 1999.
  7. 7. Slate D, Algeo TP, Nelson KM, Chipman RB, Donovan D, Blanton JD, et al. Oral Rabies Vaccination in North America: Opportunities, Complexities, and Challenges. PLOS Neglected Tropical Diseases. 2009;3(12):1–9. pmid:20027214
  8. 8. Bester JC. Measles and Measles Vaccination: A Review. JAMA Pediatrics. 2016;170(12):1209–1215. pmid:27695849
  9. 9. Larson HJ, Cooper LZ, Eskola J, Katz SL, Ratzan S. Addressing the vaccine confidence gap. The Lancet. 2011;378(9790):526–535. pmid:21664679
  10. 10. Gangarosa EJ, Galazka AM, Wolfe CR, Phillips LM, Miller E, Chen RT, et al. Impact of anti-vaccine movements on pertussis control: the untold story. The Lancet. 1998;351(9099):356–361. pmid:9652634
  11. 11. Benecke O, DeYoung SE. Anti-Vaccine Decision-Making and Measles Resurgence in the United States. Global Pediatric Health. 2019;6:2333794X19862949. pmid:31384629
  12. 12. Sallam M. COVID-19 Vaccine Hesitancy Worldwide: A Concise Systematic Review of Vaccine Acceptance Rates. Vaccines. 2021;9(2). pmid:33669441
  13. 13. Brisse M, Vrba SM, Kirk N, Liang Y, Ly H. Emerging Concepts and Technologies in Vaccine Development. Frontiers in Immunology. 2020;11. pmid:33101309
  14. 14. Diamos AG, Pardhe MD, Sun H, Hunter JGL, Kilbourne J, Chen Q, et al. A Highly Expressing, Soluble, and Stable Plant-Made IgG Fusion Vaccine Strategy Enhances Antigen Immunogenicity in Mice Without Adjuvant. Frontiers in Immunology. 2020;11. pmid:33343565
  15. 15. Rauch S, Jasny E, Schmidt KE, Petsch B. New Vaccine Technologies to Combat Outbreak Situations. Frontiers in Immunology. 2018;9. pmid:30283434
  16. 16. Kesselheim AS, Darrow JJ, Kulldorff M, Brown BL, Mitra-Majumdar M, Lee CC, et al. An Overview Of Vaccine Development, Approval, And Regulation, With Implications For COVID-19. Health Affairs. 2021;40(1):25–32. pmid:33211535
  17. 17. Sheikh S, Biundo E, Courcier S, Damm O, Launay O, Maes E, et al. A report on the status of vaccination in Europe. Vaccine. 2018;36(33):4979–4992. pmid:30037416
  18. 18. Kermack WO, McKendrick AG. A contribution to the mathematical theory of epidemics. Proc R Soc Lond A. 1927;155:700–721.
  19. 19. Ross R. An application of the theory of probabilities to the study of a priori pathometry.—Part I. Proceedings of the Royal Society of London Series A, Containing Papers of a Mathematical and Physical Character. 1916;92(638):204–230.
  20. 20. Ross R, Hudson HP. An application of the theory of probabilities to the study of a priori pathometry.—Part II. Proceedings of the Royal Society of London Series A, Containing Papers of a Mathematical and Physical Character. 1917;93(650):212–225.
  21. 21. Ross R, Hudson HP. An application of the theory of probabilities to the study of a priori pathometry.—Part III. Proceedings of the Royal Society of London Series A, Containing Papers of a Mathematical and Physical Character. 1917;93(650):225–240.
  22. 22. Fine PE. Ross’s a priori pathometry—a perspective. Proceedings of the Royal Society of Medicine. 1975;68(9):547–551. pmid:1105597
  23. 23. Dietz K, Schenzle D. Mathematical Models for Infectious Disease Statistics. In: Atkinson AC, Fienberg SE, editors. A Celebration of Statistics. New York, NY: Springer New York; 1985. p. 167–204.
  24. 24. Bailey NTJ. The mathematical theory of infectious diseases and its applications. 2nd ed. Charles Griffin & Company Ltd; 1975.
  25. 25. Brauer F, van den Driessche P, et al JW. Mathematical Epidemiology. Springer, Berlin, Heidelberg; 2008.
  26. 26. Newman MEJ. Spread of epidemic disease on networks. Phys Rev E. 2002;66:016128.
  27. 27. Moreno Y, Pastor-Satorras R, Vespignani A. Epidemic outbreaks in complex heterogeneous networks. Eur Phys J B. 2002;26:521–526.
  28. 28. Anderson DF, Kurtz TG. In: Koeppl H, Setti G, di Bernardo M, Densmore D, editors. Continuous Time Markov Chain Models for Chemical Reaction Networks. New York, NY: Springer New York; 2011. p. 3–42.
  29. 29. Bibbona E, Sirovich R. Strong approximation of density dependent Markov chains on bounded domains. arXiv. 2017.
  30. 30. Newman M. Networks: An Introduction. Oxford University Press; 2010.
  31. 31. Madar N, Kalisky T, Cohen R, ben Avraham D, Havlin S. Immunization and epidemic dynamics in complex networks. Eur Phys J B. 2004;38:269–276.
  32. 32. Albert R, Jeong H, Barabási A. Error and attack tolerance of complex networks. Nature. 2000;406:378–382. pmid:10935628
  33. 33. Hindes J, Schwartz IB. Epidemic Extinction and Control in Heterogeneous Networks. Phys Rev Lett. 2016;117:028302. pmid:27447531
  34. 34. Hartmann AK. Big Practical Guide to Computer Simulations. Singapore: World Scientific; 2015.
  35. 35. Dembo A, Zeitouni O. Large Deviations Techniques and Applications. Berlin: Springer; 2010.
  36. 36. Feld Y, Hartmann AK. Large deviations of a susceptible-infected-recovered model around the epidemic threshold. Phys Rev E. 2022;105:034313. pmid:35428162
  37. 37. den Hollander F. Large Deviations. Providence: American Mathematical Society; 2000.
  38. 38. Touchette H. The large deviation approach to statistical mechanics. Physics Reports. 2009;478(1):1–69.
  39. 39. Touchette H. A basic introduction to large deviations: Theory, applications, simulations. In: Leidl R, Hartmann AK, editors. Modern Computational Science 11: Lecture Notes from the 3rd International Oldenburg Summer School. Oldenburg: BIS-Verlag; 2011.
  40. 40. Pastor-Satorras R, Vespignani A. Immunization of complex networks. Phys Rev E. 2002;65:036104. pmid:11909162
  41. 41. Pastor-Satorras R, Castellano C, Van Mieghem P, Vespignani A. Epidemic processes in complex networks. Rev Mod Phys. 2015;87:925–979.
  42. 42. Wang Z, Bauch CT, Bhattacharyya S, d’Onofrio A, Manfredi P, Perc M, et al. Statistical physics of vaccination. Physics Reports. 2016;664:1–113.
  43. 43. Watts DJ, Strogatz SH. Collective dynamics of’small-world’ networks. Nature. 1998;393:440–442. pmid:9623998
  44. 44. Eubank S, Guclu H, Kumar VSA, Marathe MV, Srinivasan A, Toroczkai Z, et al. Modelling disease outbreaks in realistic urban social networks. Nature. 2004;429(6988):180–184. pmid:15141212
  45. 45. Bucklew JA. Introduction to rare event simulation. New York: Springer-Verlag; 2004.
  46. 46. Hammersley JM, Morton KW. A new Monte Carlo technique: antithetic variates. Math Proc Cambr Phil Soc. 1956;52:449–475.
  47. 47. Crooks GE, Chandler D. Efficient transition path sampling for nonequilibrium stochastic dynamics. Phys Rev E. 2001;64:026109. pmid:11497653
  48. 48. Dellago C, Bolhuis PG, Csajka FS, Chandler D. Transition path sampling and the calculation of rate constants. J Chem Phys. 1998;108(5):1964–1977.
  49. 49. Dewenter T, Hartmann AK. Large-deviation properties of resilience of power grids. New Journal of Physics. 2015;17(1):015005.
  50. 50. Feld Y, Hartmann AK. Large-deviations of the basin stability of power grids. Chaos: An Interdisciplinary Journal of Nonlinear Science. 2019;29(11):113103. pmid:31779340
  51. 51. Claussen G, Hartmann AK, Majumdar SN. Convex hulls of random walks: Large-deviation properties. Phys Rev E. 2015;91:052104. pmid:26066116
  52. 52. Schawe H, Hartmann AK. Large Deviations of Convex Hulls of the “True” Self-Avoiding Random Walk. Journal of Physics: Conference Series. 2019;1290(1):012029.
  53. 53. Schawe H, Hartmann AK. Large deviations of a random walk model with emerging territories. Phys Rev E. 2020;102:062141. pmid:33466107
  54. 54. Chevallier A, Cazals F. Wang-Landau algorithm: An adapted random walk to boost convergence. Journal of Computational Physics. 2020;410:109366.
  55. 55. Engel A, Monasson R, Hartmann AK. On the large deviation properties of Erdös-Rényi random graphs. J Stat Phys. 2004;117:387–426.
  56. 56. Hartmann AK. Large-deviation properties of largest component for random graphs. Eur Phys J B. 2011;84:627–634.
  57. 57. Hartmann AK. Large-deviation properties of the largest 2-core component for random graphs. Eur Phys J Spec Top. 2017;226:567–579.
  58. 58. Schawe H, Hartmann AK. Large-deviation properties of the largest biconnected component for random graphs. Eur Phys J B. 2019;92:73.
  59. 59. Börjes J, Schawe H, Hartmann AK. Large deviations of the length of the longest increasing subsequence of random permutations and random walks. Phys Rev E. 2019;99:042104. pmid:31108578
  60. 60. Krabbe P, Schawe H, Hartmann AK. Number of longest increasing subsequences. Phys Rev E. 2020;101:062109. pmid:32688539
  61. 61. Körner M, Katzgraber HG, Hartmann AK. Probing tails of energy distributions using importance-sampling in the disorder with a guiding function. Journal of Statistical Mechanics: Theory and Experiment. 2006;2006(04):P04005.
  62. 62. Hartmann AK, Doussal PL, Majumdar SN, Rosso A, Schehr G. High-precision simulation of the height distribution for the KPZ equation. Europhys Lett. 2018;121(6):67004.
  63. 63. Belardinelli RE, Pereyra VD. Fast algorithm to calculate density of states. Phys Rev E. 2007;75:046701. pmid:17501010
  64. 64. Wang F, Landau DP. Efficient, Multiple-Range Random Walk Algorithm to Calculate the Density of States. Phys Rev Lett. 2001;86:2050–2053. pmid:11289852
  65. 65. Yan Q, de Pablo JJ. Fast Calculation of the Density of States of a Fluid by Monte Carlo Simulations. Phys Rev Lett. 2003;90:035701. pmid:12570508
  66. 66. Belardinelli RE, Manzi S, Pereyra VD. Analysis of the convergence of the 1/t and Wang-Landau algorithms in the calculation of multidimensional integrals. Phys Rev E. 2008;78:067701.
  67. 67. Belardinelli RE, Pereyra VD. Nonconvergence of the Wang-Landau algorithms with multiple random walkers. Phys Rev E. 2016;93:053306. pmid:27301004
  68. 68. Newman MEJ, Barkema GT. Monte Carlo Methods in Statistical Physics. Oxford: Clarendon Press; 1999.
  69. 69. Lee J. New Monte Carlo algorithm: Entropic sampling. Phys Rev Lett. 1993;71:211–214. pmid:10054892
  70. 70. Dezső Z, Barabási AL. Halting viruses in scale-free networks. Phys Rev E. 2002;65:055103. pmid:12059627
  71. 71. Holme P, Kim BJ, Yoon CN, Han SK. Attack vulnerability of complex networks. Phys Rev E. 2002;65:056109. pmid:12059649
  72. 72. Erdős P, Rényi A. On the evolution of random graphs. Publ Math Inst Hungar Acad Sci. 1960;5:17–61.