Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Quantifying spectral information about source separation in multisource odour plumes

  • Sina Tootoonian ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Writing – original draft, Writing – review & editing

    sina.tootoonian@crick.ac.uk

    Affiliation Sensory Circuits and Neurotechnology Laboratory, The Francis Crick Institute, London, United Kingdom

  • Aaron C. True,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Writing – original draft, Writing – review & editing

    Affiliation Department of Civil, Environmental and Architectural Engineering, University of Colorado, Boulder, CO, United States of America

  • Elle Stark,

    Roles Data curation, Formal analysis, Investigation, Software, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Department of Civil, Environmental and Architectural Engineering, University of Colorado, Boulder, CO, United States of America

  • John P. Crimaldi,

    Roles Conceptualization, Funding acquisition, Investigation, Resources, Writing – original draft, Writing – review & editing

    Affiliation Department of Civil, Environmental and Architectural Engineering, University of Colorado, Boulder, CO, United States of America

  • Andreas T. Schaefer

    Roles Conceptualization, Funding acquisition, Resources, Writing – original draft, Writing – review & editing

    Affiliations Sensory Circuits and Neurotechnology Laboratory, The Francis Crick Institute, London, United Kingdom, Department of Neuroscience, Physiology and Pharmacology, University College London, United Kingdom

Abstract

Odours released by objects in natural environments can contain information about their spatial locations. In particular, the correlation of odour concentration timeseries produced by two spatially separated sources contains information about the distance between the sources. For example, mice are able to distinguish correlated and anti-correlated odour fluctuations at frequencies up to 40 Hz, while insect olfactory receptor neurons can resolve fluctuations exceeding 100 Hz. Can this high-frequency acuity support odour source localization? Here we answer this question by quantifying the spatial information about source separation contained in the spectral constituents of correlations. We used computational fluid dynamics simulations of multisource plumes in two-dimensional chaotic flow environments to generate temporally complex, covarying odour concentration fields. By relating the correlation of these fields to the spectral decompositions of the associated odour concentration timeseries, and making simplifying assumptions about the statistics of these decompositions, we derived analytic expressions for the Fisher information contained in the spectral components of the correlations about source separation. We computed the Fisher information for a broad range of frequencies and source separations for three different source arrangements and found that high frequencies were more informative than low frequencies when sources were close relative to the sizes of the large eddies in the flow. We observed a qualitatively similar effect in an independent set of simulations with different geometry, but not for surrogate data with a similar power spectrum to our simulations but in which all frequencies were a priori equally informative. Our work suggests that the high-frequency acuity of olfactory systems may support high-resolution spatial localization of odour sources. We also provide a model of the distribution of the spectral components of correlations that is accurate over a broad range of frequencies and source separations. More broadly, our work establishes an approach for the quantification of the spatial information in odour concentration timeseries.

1 Introduction

Olfactory signals enable a multitude of crucial behaviors observed in terrestrial organisms, from male moths navigating to females from hundreds of meters away [1, 2], to mice detecting and avoiding predators [3], to honeybees foraging for nectar in a flower among a field of vegetation [4]. Such feats of olfactory navigation pose complex problems, due in part to a turbulent transport process that causes odour signals to become patchy and highly fluctuating away from the source [5]. Even in a single-source context, strategies for olfactory search span a range of complexity [6], including reactive strategies in which a searcher reorients upwind upon odour detection [7, 8], information-theoretic approaches that exploit the spatiotemporal dynamics of odour signals [911], and strategies that incorporate non-olfactory cues in the search (such as visual cues as in [12], public cues from other searchers as in [13], or direction of motion of odor as in [14]). Recent work employs machine learning, including neural networks and reinforcement learning, to discover data-driven navigational strategies from large training sets [15].

Such olfactory search strategies are often studied in the context of single-source searches, but in naturalistic odour plumes, odour signals do not arrive in isolation but instead arrive in a fluctuating mixture of multiple odorants, further complicating navigation tasks [16]. In such multi-source contexts, a navigating animal must both identify the odour(s) of interest and localize the target odour(s). The relative distance between two target odours may be of particular interest to an organism; for example, the distance of a predator to a food source or the proximity of multiple food sources to each other. As noted early by Hopfield [17], the modulation of correlation by intersource distance can help animals locate odour sources in the environment, i.e. perform olfactory scene segmentation. To resolve such correlations requires an ability to parse signals at high frequency. Insects indeed demonstrate the capacity to distinguish high-bandwidth odor signals at frequencies up to 100 Hz [18], and in a multi-source context, moths have even been shown to discriminate odor sources separated by as little as 1 mm in space or 0.001 s in time [19]. Mammalian olfaction has generally been considered a ‘slow’ modality, but recent research demonstrated the ability of mice to discriminate fast odour fluctuations at frequencies of up to 40 Hz, much higher than expected [20].

Given the sensitivity of animals to high-bandwidth information in odour signals, can the correlations between two odour signals theoretically be exploited to estimate the distance between the two sources? And if so, are some frequencies more informative than others? To investigate these questions, we develop a framework for quantifying the usefulness of correlations between odour signals at various frequencies, based on computing the Fisher information that they contain about intersource distance.

Our analysis of high bandwidth information requires consideration of the statistics and driving mechanisms of the range of signal frequencies present in naturalistic odour plumes. Odours released into turbulent flow environments are advected downstream with the mean flow. Concurrent distortion of odour streams by chaotic flow structures spanning a range of sizes (a defining characteristic of fluid turbulence) produces discrete odour filaments whose structure is subsequently altered by the turbulent mixing field. The associated stretching and folding of filaments sharpens odour concentration gradients, enhancing molecular diffusion in response to these flow-sharpened gradients and acts to broaden them [21]. The net effect of these competing processes, collectively constrained by odour mass conservation, is broadly referred to as chaotic advection, turbulent diffusion, or simply mixing [22, 23], and the resultant product is a complex, spatiotemporally dynamic concentration landscape. We refer to these as odour landscapes or plumes.

Statistics describing concentration fluctuations of passive scalar plumes (composed of odours that do not modify the local flow field) and their spatial variation are set by a number of dimensionless parameters. These include a Reynolds number (Re) describing the ambient flow environment (setting the statistics of the turbulent flow field) and a Péclet number (Pe) describing the relative importance of advective and diffusive odour transport (or alternatively a Schmidt number describing the relative diffusivities of fluid momentum and odour mass). Higher Re and Pe numbers are associated with more turbulent flows producing plume statistics dominated by turbulent diffusion. Additionally, a number of ratios describe the source configuration. These include its size relative to the largest, energy-containing eddies (characterized by an integral length scale), its proximity to solid boundaries, and its injection momentum relative to the flow. Source configuration effects are well-known (e.g. geometry [24, 25], source momentum relative to the mean flow [26], and proximity to surfaces [27, 28]). The variety of experimentally-measured odour landscapes shown in Fig 1 illustrates the effects of some of these important factors.

thumbnail
Fig 1. Variety of odour landscapes.

Diverse single odour landscapes in air and water (adapted from [29]) with varying source release configurations and flow, fluid, and odourant properties. Animals may have to operate in multiple landscapes. For example, the odour landscape relevant to a mouse may be similar to the nearbed isokinetic condition (lower left) when it is searching the soil, and more like the freestream isokinetic (upper left) when it is rearing. The plumes were measured experimentally using planar laser-induced fluorescence (PLIF) and are described in detail in [30] (air) and [31] (water).

https://doi.org/10.1371/journal.pone.0297754.g001

Plume statistics exhibit characteristic variations with distance from the source—including with streamwise and cross-stream location, parallel and normal to the mean flow vector, respectively [32]—owing to transitions between phenomenological regimes defined by differing sources of concentration variance. For example, for a small relative source size, near the source where the plume width is small relative to the large eddies, concentration variance is primarily driven by meandering of the whole plume driven by those large eddies (e.g. [33]). As the plume width grows with distance from the source, the large eddies produce less meandering and instead a range of eddy sizes are effective at mixing odour-laden and ambient (clean) fluid—a turbulent diffusive regime with enhanced relative dispersion of the plume locally around its meandering centroid. A number of studies have examined statistical structure with distance from the source both experimentally [5, 34] and analytically [16]. Broadly speaking, source configuration effects influence near-field plume statistics (close to the source), whereas the far-field structure is largely self-similar, being set by the physics of turbulent flows and the linearly coupled advection-diffusion equations governing odourant transport and diffusion.

When multiple non-reactive odours are released from initially-distant locations into turbulent flows, odour-specific plumes evolve independently and uniquely under the influence of chaotic coherent flow structures [35]. The nature of the cumulative odour landscape describing all local odour concentrations and associated gradients is then simply the superposition of these odour-specific plumes, information that is potentially exploitable by a navigating sensor or organism. Of particular relevance to the work presented here is that, in addition to the factors setting the overall statistics of each individual odour plume, pointwise concentration correlations between odour species will depend strongly on the initial source separation distance [36, 37]. In particular, the correlations of concentration fluctuations can vary non-monotonically with distance from the source, and they show strong source separation effects with weak Reynolds number effects [38, 39]. This contributes to the observations of zero, negative, and positive correlation regimes (no interaction, destructive and constructive interference, respectively) [40, 41]. As an analogue to the single-source case where the local plume width relative to the large eddy sizes is important in driving local production of concentration variability, the source separation relative to the large eddy sizes in the multisource case drives these complex correlation behaviors.

Many statistical models describing concentration fluctuations for the single-source case have been proposed in the literature, and most leverage the simplicity of the exponential family to describe the one-point probability density functions [42, 43]. While the best match for a particular distance from the source and plume realization (Re, Pe, source configuration) varies, there is recent evidence that the flexibility of the Gamma distribution can robustly account for much of this variability [44], where the shape parameter is related to the concentration fluctuation intensity [24]. Intuitively, good statistical descriptors for concentration fluctuations must also capture the intermittent nature of concentration dynamics in turbulent flows manifesting as high probabilities of zero concentration events [43, 45]. Extensions of these models to describe the concentration statistics of multiple interacting sources are less frequent in the literature, but the exponential family was also shown to provide a good description of the distributions for the sum of concentrations [36, 46]. Even fewer studies have examined the spectral properties of correlations; however, recent work showed that for a single source separation, correlations increase with distance from the source and become more spectrally uniform [41].

Due to the dynamics described above, odour plumes in naturalistic environments contain fluctuations at a wide range of frequencies, modulated by distance to the source, proximity to solid boundaries, odour source characteristics, and the ambient flow environment. At the same time, recent research has demonstrated the high frequency acuity of olfactory systems, but the purpose of such sensitivity is unknown. One possibility is that information about intersource distance is better encoded in different frequency bands depending on the olfactory context, and that the sensitivity of animal olfactory systems to rapidly fluctuating odours allows them to use this information when performing odour source localization. Thus the question we aim to answer in this study is: What is the spectral distribution of the information that correlations carry about source separation?

2 Results

To provide a testing ground for our approach to quantifying information in odour plumes we performed computational fluid dynamics (CFD) simulations of two-dimensional fluid flowing past equally spaced cylindrical obstacles in a rectangular wind tunnel, as shown in Fig 2. The interaction of the flow with the obstacles generated complex flow patterns that we used as a proxy for turbulent advection. We placed multiple odour sources in the simulated wind tunnel and measured the correlation of the resulting concentration profiles—henceforth referred to as plumes—at fixed probe locations downwind. A snapshot of the plumes generated by two adjacent odour sources is shown coloured blue and red in Fig 2.

thumbnail
Fig 2. CFD simulation domain.

The wind tunnel domain used to simulate multispecies odour landscapes. The total domain consists of an inlet flow development section followed by a plume domain section, beginning at x = 0. Most of the analyses in the Main Text are for an array of source pairs located at x = 0 and mirrored about the domain centerline at y=0 (the ‘cross-stream’ source configuration, orange arrows in the inset). We also performed analyses for sources placed oblique to the flow (brown arrow in the inset) or parallel to it (‘streamwise’, green arrow). Geometric, flow, fluid, and odourant properties are summarized in Table 1.

https://doi.org/10.1371/journal.pone.0297754.g002

Most of the results we present below are for odour sources placed ‘cross-stream’ i.e. at a fixed distance from the flow inlet, symmetric along the midline, and equally spaced transverse to mean flow. The direction along which the sources are placed is indicated by the orange arrow in the inset of Fig 2. The resulting odour signals that these sources generate are measured at a single probe location situated along the midline. To determine the generalizeability of our results we also performed our analyses for sources placed oblique to the mean flow (brown arrow in the inset) and parallel to the flow (‘streamwise’, green arrow), and with odour signals measured at 8 other probe locations. Further, we repeated these analyses for a supplementary set of simulations using similar flow parameters and source configurations but slightly different domain geometry. In Supporting Information Sec S3.1 in S1 File we describe these supplementary simulations and present some of the corresponding results of the analyses presented in the Main Text. The full set of source geometries and probe locations that we used are shown in S10 Fig.

thumbnail
Table 1. Summary of model geometry, flow, fluid, and odourant properties used for multisource plume simulations.

Properties are for air and common airborne odourants at ≈ 20°C.

https://doi.org/10.1371/journal.pone.0297754.t001

Another snapshot of a flow pattern, this time for the maximally separated sources in the cross-stream configuration, is shown in Fig 3A. In Fig 3B we show some example concentration time series from sources at three different intersource distances. We express all distances in terms of the spacing between the obstacles, called the ‘pitch,’ and indicate this by suffixing all distances with ‘ϕ’. This is a natural length scale for distance normalization since it approximates the sizes of the largest eddies.

thumbnail
Fig 3. Example plumes and correlations.

(A) Snapshot of the simulated flow patterns advecting two sources (colours). The plumes are sampled downwind at the probe location marked ‘×’. (B) Concentration profiles of the two odours at the probe location, scaled by the larger of their two standard deviations computed over the window shown. Sources were placed at the same downwind (horizontal) location and separated by cross-stream (vertical) distances indicated by s. Pearson correlations computed for the time windows shown, and for the entire simulation duration, are indicated by ρw and ρ, respectively. (C) Mean ± standard deviation of the Pearson correlations computed over 1-second boxcar windows overlapping by 500 msec., for all sources at the intersource distances indicated. All distances are measured in pitch (ϕ).

https://doi.org/10.1371/journal.pone.0297754.g003

We used the Pearson correlation to measure the cofluctuations of two plumes. For the examples shown in Fig 3B correlations decrease with increasing intersource distance. We confirmed that this effect holds for the rest of our data in Fig 3C, similar to previous observations in the literature, e.g. [47].

The Pearson correlation of two plumes x(t) and y(t) in a time window of width T seconds is (1) where and σx are the mean and standard deviation of x(t) over this window, and similarly for and σy. A key property of the Pearson correlation for our purposes is that it can be decomposed as a sum of correlations computed for each harmonic of the fundamental frequency 1/T (2) where rn is the contribution from the n’th harmonic, (3) and (an, bn) and (cn, dn) are the cosine and sine coefficients of the Fourier decomposition of x(t) and y(t), respectively, scaled by the standard deviations of these two signals (see Eq 59 in Methods).

The spectral decomposition of plume correlations clearly depends on the corresponding decomposition of the plumes themselves. Observation of long duration signals such as plumes over short time windows introduces uncertainty about their spectral content called spectral leakage [48], wherein the spectral content of the observed signal at a given frequency is not just that of the original signal at that frequency, but includes weighted contributions from all other frequencies. The effects of spectral leakage can be ameliorated by judicious weighting of the observations, called ‘windowing’. Unless stated otherwise, the results below use a 1-second Hann window. In Fig 4 we have plotted the distribution of some of these correlation components as a function of distance.

thumbnail
Fig 4. Decomposition of correlations.

(A) Mean ± S.D. of Pearson correlations between the plumes emitted by two sources the specified distance apart, computed using 1-second Hann windows (compare to Fig 3C which used 1-second boxcar windows). (B–E) The components of the correlation at the frequencies stated above each panel. (F) The means of the data in panels (B–E) scaled to have the same value at intersource distance zero and overlayed for ease of comparison.

https://doi.org/10.1371/journal.pone.0297754.g004

To measure the amount of information that correlations rn provide about intersource distance s, we used the Fisher information [49], defined as (4) where p(rn|s) is the distribution of correlations rn for harmonic n for an intersource separation s. Informally, the Fisher information measures the dependence of our estimate of s on the observed correlations rn. The more sensitive this estimate is to correlations, the more informative the correlations. The Fisher information also bounds the precision of unbiased estimation of intersource distance from correlations. Thus the higher the Fisher information, the higher the maximum precision with which intersource distance can be estimated from correlations. Note also that the Fisher information is a function of the intersource distance s. Thus the information provided by correlations will in general depend on the (true) intersource distance.

2.1 Outline of our approach

Our approach to determining the spectral distribution of information in the correlations is as follows. First, we

  1. 1. Use simplifying assumptions about the statistics of the Fourier coefficients of the component waveforms to determine analytically tractable approximations to the distribution p(rn|s) needed to compute the Fisher information (Sec 2.2, Eq 21, and Fig 7);

Although we make some of these assumptions for mathematical tractability rather than because we believe they hold in actual flows, we nevertheless test their validity numerically in Sec S4 in S1 File.

Next, we

  1. 2. Use the distribution of correlations thus derived to compute analytic expressions for the Fisher information (Sec 2.3 and Eq 23);
  2. 3. Fit tractable distributions for the quantities required in the analytic expressions of Fisher information to our simulation data (Eq 24, Figs 8 and 9);
  3. 4. Compute the Fisher information using fitted parameters for each harmonic and intersource distance (Eq 26 and Fig 10).

2.2 The distribution of correlations

To determine an analytically tractable expression for p(rn|s), we began by using the rules of probability to relate the correlation rn to the Fourier coefficients an, bn, cn and dn. In our description so far we have assumed a single pair of sources at a given distance s apart. In reality, the distribution of correlations is formed by combining the contribution of all pairs of sources separated by that distance. For example, in our simulations, there will frequently be many pairs of sources a given distance apart. To quantify effects that depend only on the relative location of sources, rather than the specific locations of specific sources, we made two location-independence assumptions.

First, we assumed that sources were close enough together relative to the animal and to the geometry of the flow so that

  1. 1. The distribution of coefficients from each source is the same for all sources, and
  2. 2. The distribution of coefficients from one source given those at another source depends only on the distance between them.

We tested these assumptions numerically and found that they both broadly held (see S2 and S3 Figs).

We show in Methods Sec 5.3.3 that these two assumptions allow us to write the distribution of correlations at a given source separation as (5)

The first term in the integrand describes the dependence of correlations rn on the Fourier coefficients and the intersource distance. From Eq 3 we know that rn is determined entirely by the Fourier coefficients. We can express this fact probabilistically by concentrating all of the probability density at the value in Eq 3 using the Dirac δ function, (6)

The second term in the integrand of Eq 5 is the probability of observing the pair of coefficients an, bn from a source. To arrive at analytically tractable solutions, we made our next assumption:

  1. 3. Concentration profiles are Gaussian processes.

This assumption implies that the joint distribution of coefficients from each source, p(an, bn), is a two-dimensional Gaussian distribution. We emphasize that we made this assumption for analytic tractability; turbulent flows are well known to exhibit non-Gaussianity, for example in their velocity fields a [50] or the concentrations of advected substances [43]. Indeed, when we tested this assumption numerically we found that at most locations in the plume it only held for frequencies below 10 Hz (see S4 Fig). We consider some of the implications of this non-Gaussianity in the Discussion.

To specify the mean and covariance of our assumed Gaussian distribution, we made our next assumption,

  1. 4. Concentration profiles are temporally stationary.

This means that the statistical properties of plumes do not change with time. Temporal stationarity has the following important implications for the Fourier coefficients (see Sec S6.1 in S1 File for details). First, it implies that the marginal distribution of the sine and cosine coefficients must be the same, that is (7) Second, it implies that the coefficients at non-zero frequency have mean zero, that is (8) Third, it implies that the coefficients are uncorrelated, so (9) The expectations in Eqs 8 and 9 are over time windows. We tested all three implications numerically and found that they broadly held (see S5 Fig).

By combining our Gaussian process assumption with our assumptions of location independence and temporal stationarity we can completely specify the second term in Eq 5: (10) where is the marginal variance of the coefficients for harmonic n.

The third term in Eq 5 is p(cn, dn|an, bn, s) and expresses the probability of observing the pair of coefficients (cn, dn) from the second source given the observed coefficients (an, bn) from the first source located at an intersource distance of s. To specify it, we first represented the joint distribution of coefficients p(an, bn, cn, dn|s), as the graphical model in Fig 5A. The marginal independence of the sine and cosine coefficients an and bn in Eq 10 is reflected in the network through the lack of connections between an and bn. Importantly, the network indicates that the coefficients cn and dn are conditionally independent given the observed values of an and bn. This is because the only routes by which cn and dn can influence each other are through an and bn, which serve as common input to an and bn, and these are observed [51]. Therefore, (11)

thumbnail
Fig 5. Decomposition of phase relationships.

(A) Knowing the cosine (a) and sine (b) coefficients at one source can inform about the corresponding coefficients (c and d, respectively) at another. The lack of arrows between a and b indicates their marginal independence. (B) Cosine (a) vs. sine (b) coefficients from the same source at 5 Hz, pooled across all sources. (C) Cosine coefficients from one source (a) vs. those (c) from the closest source in the positive vertical direction, pooled across all such pairs. (D) As in panel C but plotted against the sine coefficient from the neighbouring source. The Pearson correlation of the coefficients in each panel is indicated with ρ.

https://doi.org/10.1371/journal.pone.0297754.g005

To specify a form for the conditional distributions on the right-hand side of Eq 2, we must determine the uncertainty that remains about the coefficients cn and dn of the component waveform at the second source when we know the corresponding coefficients an and bn at the first. To do so, we expressed the component waveform at the second source as a scaled and phase-shifted version of the first, plus a residual. Doing so, we made our final assumption,

  1. 5. The conditional distribution p(cn, dn|an, bn, s) of coefficients at one source (cn and dn) given those at another (an and bn), is bivariate Gaussian.

We performed a coarse test of this assumption and found that it held, at least within the limitations of this test; see S6 Fig.

With this final assumption in hand we concluded that the conditional distribution of the individual coefficients is (12a) (12b) Here βn(s) and the θn(s) are the best-fit scaling and phase-shifts, and ηn(s)2 is the variance of the residual (see Methods Sec 5.3.2).

Substituting these equations into Eq 11 and combining with Eq 10 we arrive at the form of the joint distribution of the coefficients (13a) (13b) Note that by the location independence assumption, the marginal variance of the coefficients does not depend on s.

The covariance in Eq 13 specifies two types of relationship between the coefficients from the two sources. Those that are ‘in-phase’ relate coefficients of the same type, for example the cosine coefficients an and cn. Those that are ‘out-of-phase’ or ‘quadrature’ relate coefficients of different types, for example the cosine coefficient an at the first source and the sine coefficient dn at the second. Determining the spatial information that correlations provide requires incorporating both types of relationship by using the full set of covariances specified in Eq 13.

However, for our data the out-of-phase contribution is very small, and the in-phase relationship dominates. This can be seen, for example, in the high in-phase correlation of the data in Fig 5C vs. the near zero correlation of the out-of-phase data in Fig 5D (see also S17 Fig). Therefore, it will be convenient to focus on the marginal distribution relating only the in-phase coefficients. For the pair of cosine coefficients this is (14a) (14b) where we’ve defined the in-phase correlation (15) Note that this expression is not meant to constrain the dependence of ρn on intersource distance s, and merely defines it in terms of the quantities βn and θn at that distance. The dependence on intersource distance can in principle be arbitrary, although we will find below (see Fig 8) that it is adequately described for small intersource distances by exponential decay.

The relationship relating the pair of sine coefficients bn and dn has the same form as Eq 14. Since the off-diagonal element of is the covariance of pairs of sine or cosine coefficients, we can use Eq 3 to relate the in-phase correlations to the average component correlations as (16)

The relationship of the out-of-phase coefficients can be described in a similar way. For example, (17a) (17b) where the out-of-phase correlation is defined as (18) The distribution p(an, dn|s) has the same form, but with negative the covariance between the two variables.

In Fig 6 we have plotted examples of the joint distribution of in-phase (panels A-D) and out-of-phase (panels E-H) coefficients. These plots confirm that the out-of-phase correlations, while not zero, are much smaller than the in-phase correlations.

thumbnail
Fig 6. Fourier coefficients vs. distance.

Distribution of the indicated coefficients for the 5 Hz component of the plume from one source vs. the indicated coefficient computed for the second source, computed for all pairs of sources separated by the intersource distance stated above panels A-D. Crosses indicate the mean of each dataset; ellipses indicate first three standard deviations of bivariate normal fit to the data; Pearson correlations (ρ) of the data in each plot are indicated in the top-left corner. (A-D) Cosine coefficient at source 1 vs. cosine coefficient at source 2. (E-H) Sine coefficient at source 1 vs. cosine coefficient at source 2.

https://doi.org/10.1371/journal.pone.0297754.g006

Armed with the joint distribution of Fourier coefficients given intersource distance specified in Eq 13, we can return to Eq 5 and derive the distribution of correlations to be the asymmetric Laplacian (see Methods Sec 5.3.3) (19) where the normalizing constant is defined as (20)

A simplifying assumption.

Because the out-of-phase correlations of our data are typically very close to zero, in what follows we will assume that these correlations are zero. In that case, and Eq 19 simplifies to (21)

To evaluate the agreement of the asymmetric Laplacian in Eq 21 with the observed distribution of correlations we compare their cumulative distribution functions (CDFs). In Fig 7A–7C we make the comparison for the 5 Hz correlation data at three different intersource distances. We quantified the agreement between the CDFs as 1 minus the largest absolute difference between them. A value of 1 would indicate perfect agreement, while, a value of 0, the smallest possible would indicate non-overlapping distributions. A heatmap of the match over the full range of intersource distances and frequencies (Fig 7G) reveals a match of ∼0.7 and higher over most of this range.

thumbnail
Fig 7. Modeling the distribution of observed correlations.

(A–C) Examples of observed (Fdata, thick pink) and predicted cumulative distribution function for the 5 Hz component of correlations, for three different intersource distances (indicated above each panel). Predictions were for the asymmetric Laplacian distributions (Falap, pink) in Eq 21, asymmetric Laplacian with intermittency (Fint, light blue), and the best-fitting model when testing both intermittent and non-intermittent versions of the asymmetric Laplacian, Gamma, and generalized inverse Gaussian (Fbest, navy). The parameters of Falap were derived from Gaussian fits to the distributions of Fourier coefficients, the parameters for the remaining fits were found by directly fitting the correlation data. (D–F) Absolute difference between the predicted and observed cumulative distribution functions in the corresponding panels in the first row. Data in A-F may be optimistic because fits were computed using some (though not all) of the data they are being qualitatively evaluated against in those panels. See panel H for performance on held-out data. (G) Fit quality for the asymmetric Laplacian model, measured as 1 minus the largest absolute difference between the predicted and observed CDFs, computed for all distances and frequencies listed. Higher values indicated better fits. Values may be optimistic because parameters were both trained and evaluated on the same (full) dataset. Coloured dots correspond to data points shown in the left three panels. (H) As in panel G, but for the best fitting model when testing both intermittent and non-intermittent versions of the asymmetric Laplacian, Gamma, and generalized inverse Gaussian, fitted directly to the correlations, and evaluated on unseen data. See Methods Sec 5.2.4.

https://doi.org/10.1371/journal.pone.0297754.g007

Although there is good qualitative agreement between the data distributions of correlations and the corresponding asymmetric Laplacians of Eq 21, the plots in Fig 7A–7C also suggest that there are systematic differences between the two. In Fig 7D–7F we have plotted the differences between the data distribution and the asymmetric Laplacian fits. These plots have similar shapes, consisting of a narrow positive lobe followed by a broader negative lobe, eventually decaying to zero as both CDFs approach 1. This reveals that the data had more correlation values concentrated near zero, and correspondingly fewer large correlation values, than the asymmetric Laplacians.

To achieve a better fit ot the observed distribution of correlations, we elaborated the asymmetric Laplacian model in two ways to better capture small correlations (see Methods Sec 5.3.4). First, turbulent flows produce intermittent signals [50] so some very small correlations may be noise, not actual correlations. Therefore, similar to [43] for single-source concentration models, we incorporated intermittency using a binary random variable z that indicated whether an observation yn was of a correlation rn drawn from our correlation model p(rn|s), or noise w: (22) Second, we generalized the exponential decays in our correlation model of Eq 21 to the generalized inverse Gaussian p(r) ∝ rk−1 eα(r/λ+λ/r)/2 which also includes the Gamma distribution as a special case.

We fit both intermittent and non-intermittent versions of these models to the correlations at each frequency and intersource separation separately. Examples of such fits are shown in Fig 7, showing improved agreement with the observed correlations. The performance of the best-fitting models on unseen data over the full range of frequencies an intersource distances are shown Fig 7H, demonstrating a very good fit over the entire range.

2.3 Computing the Fisher information

Having now determined expressions for the distribution of correlations given intersource distance, we can use Eq 4 to determine the Fisher information. Evaluating that expression requires not just a form for the correlation distributions themselves, but also for how they change with intersource distance. These changes in turn are determined by how the various parameters of our correlation model, Eq 19, change with intersource distance.

The three correlation models we have considered, the asymmetric Laplacian, Gamma and generalized inverse Gaussian, have 2, 4, and 6 parameters, respectively. The intermittent version of each model requires an additional two parameters to capture the intermittency. To avoid the complexity of modeling how large numbers of parameters change with intersource distance, we will use our simplest model, the non-intermittent asymmetric Laplacian of Eq 19, to estimate Fisher information.

The general expression using Eq 19 is complex but simplifies significantly when the out-of-phase correlation is zero, which is approximately the case for our data. Therefore in what follows, we will use the simplified expressions. The Fisher information is then (see Methods Sec 5.3.5) (23) As basic checks of this expression we observe that

  • It’s non-negative since 0 ≤ ρn(s)2 ≤ 1;
  • It depends on how the correlations change with distance, via the n(s)/ds term, so that when this dependence is zero, there is no information in the correlations, as expected;
  • For a fixed amount of distance dependence n(s)/ds, information is least when there is no correlation (ρn(s) = 0).

Armed with Eq 23, we still require the distance dependence of correlations ρn(s) to compute the Fisher information. To motivate a parametric form for this dependence, we require that at the very least it should peak at an intersource distance of zero and decrease with distance. A simple parametric form that meets this requirement is exponential decay with distance s. (24) The two parameters of the model are the length scale of the decay γn, and the constant offset bn that determines the value at very large intersource distances. In Fig 8 we demonstrate the fit of this parametric form to some example correlation data.

thumbnail
Fig 8. Parametric fits to ρn(s).

(A–D) Bootstrap medians (coloured points) and 5–95th percentiles (coloured bars) of ρn(s) at each intersource distance, and a few example fits using bootstrap parameters (gray traces), for a few example frequencies. (E) Length scale parameter (on a logarithmic axis) for the fits to the data in (A–D). Boxes indicate inter-quartile range (IQR), central lines are medians, whiskers extend to last data points within 1.5 IQR, circles are outliers. Fits were computed for intersource distances up to 1 pitch, indicated by the dotted vertical lines in panels A-D.

https://doi.org/10.1371/journal.pone.0297754.g008

Interestingly, when we examined the decay of correlation with distance, we observed that higher frequencies decayed faster than lower frequencies (Fig 9A). We plotted the length constants of the generalized exponential fits to the decays against frequency (Fig 9B) and observed that they decreased at about one pitch every 10 Hz (Fig 9E). According to Eq 23, Fisher information depends on the rate of change of correlations with intersource distance. Therefore, the differences in decay rates that we observed suggest that the amount of information carried by correlations at different frequencies varies. As a comparison, we generated surrogate data with a similar power spectrum to our simulated plumes, but for which all frequencies were equally informative (see Methods Sec 5.2.7). We observed very little change in length constants with frequency in this surrogate data (Fig 9C–9E).

thumbnail
Fig 9. The decay of correlations with frequency.

(A) Normalized correlations at different frequencies. (B) Length constants γ of exponential fits to the decay of correlations for different frequency components. Dots and lines are the bootstrap median and 5th-95th percentiles, respectively (C,D) As in panels A and B but for surrogate data where all frequencies are equally informative. (E) Bootstrap distribution of coefficients when regressing the length constants in (B,D) on frequency. Negative values indicate length constants shorten with frequency. Central lines are medians, boxes are interquartile range, whiskers are [5th- 95th] percentiles, circles are outliers. The simulation data (but not the surrogate data) had very large time constant at 1 Hz so only the data from 2—20 Hz are shown above and used in the analysis.

https://doi.org/10.1371/journal.pone.0297754.g009

Given the parametric form Eq 24 of the distance dependence we derived an analytic expression for the Fisher information. To simplify the expression we first normalized distance s by the length-scale parameter γn and defined (25) In Methods Sec 5.3.5 we show that the Fisher information about intersource distance provided by the n’th harmonic component of correlations can be expressed in terms of normalized distance sn as (26) In Fig 10A we’ve plotted the bootstrap median and 5th-95th percentiles of Fisher information at a range of distances for a few example harmonics. The heatmap in Fig 10B shows the Fisher information computed using all the data (i.e. not bootstrapped) at a broad range of frequencies and intersource distances.

thumbnail
Fig 10. Fisher information vs. intersource distance.

(A) Bootstrap medians (points) and 5–95th percentiles (bars) of Fisher information vs. intersource distance for five example frequencies. (B) Fisher information heatmap, computed using all the data (i.e. not bootstrapped). Note the nonlinear scaling of the x-axis.

https://doi.org/10.1371/journal.pone.0297754.g010

2.3.1 Which frequencies are more informative?

The Fisher information heatmaps for our CFD plumes suggest that high frequencies are more informative when sources are close together. To see what such heatmaps would look like when the ground-truth distribution of information over frequencies is known, we generated surrogate plume datasets with similar power spectra to our CFD plumes, but with different distributions of information across frequencies (see Methods Sec 5.2.7).

To do so, we used the fact that the Fisher information depends on the in-phase correlations ρn(s) according to Eq 23, and that these in turn are determined by the correlation (equivalently, covariance, since coefficients have zero mean) the in-phase Fourier coefficients by Eq 14b. We therefore generated surrogate data for R odour sources at a given harmonic by sampling from a zero-mean R-dimensional Gaussian and adjusted the correlation of its dimensions, representing odour sources, to decay with intersource distance to produce the desired Fisher information at that harmonic.

In the first such dataset, all frequencies were set to be equally informative. To achieve this, we set the correlation between two surrogate sources to decay at the same exponential rate with their intersource distance at each harmonic (see Table 2). In Fig 11B we have plotted the Fisher information heatmap for this dataset. The information content varies with intersource distance, but is homogeneous with frequency, as expected, and unlike that of our CFD plumes (reproduced in Fig 11A).

thumbnail
Fig 11. Fisher information for simulations and surrogate data.

(A) Fisher information heatmap for our simulated plumes (reproduced from Fig 10B). (B) Heatmap for surrogate data with power spectrum similar to our simulated data and where all frequencies are equally informative. (C) Heatmap for surrogate data with power spectrum similar to our simulated data and where frequencies >12.5 Hz have the same, higher value of Fisher information than the lower frequencies. (D) Fisher information for our simulated plumes when using a 1-second boxcar window.

https://doi.org/10.1371/journal.pone.0297754.g011

thumbnail
Table 2. Summary of surrogate datasets and the power and correlation functions used.

The kernel for each dataset was the product of the power and correlation functions: k(i, j, n) = S(n)G(|ij|, n).

https://doi.org/10.1371/journal.pone.0297754.t002

Next, we generated a second surrogate dataset in which correlations decayed with intersource distance six times faster for the higher half of the frequency range than the lower half. This should result in the high frequencies being more informative at small intersource distances. At large intersource distances, high-frequency correlations will have decayed to zero due to their fast decay rate, yielding little information. Low frequencies will not have decayed completely to zero and will remain informative. Thus we expected to see high frequencies being more informative at small intersource distances, and low frequencies more informative at large intersource distances. This is indeed what we saw when we plotted the Fisher information heatmap for this dataset in Fig 11C.

The similarity of the information heatmap for our CFD plumes (Fig 11A) to that of the surrogate dataset where high-frequencies were a priori more informative (Fig 11C), and its dissimilarity to the heatmap of the dataset where all frequencies were a priori equally informative, supports the conclusion that for our CFD plumes high frequencies are more informative when sources are close together.

To emphasize the importance of windowing to these results, in Fig 11D we plot the Fisher information heatmap for our CFD plumes, but when analyzed with a 1-second boxcar window. The boxcar window, also known as a rectangular window, weights all samples equally. The poor spectral leakage properties of this window coupled with the power-law power spectrum of our data (S14 Fig) results in the low-frequency information masking that of the higher frequencies. This then results in an heatmap where information appears to be homogeneously distributed across frequencies, similar to Fig 11B.

From Eq 23, Fisher information depends on the rate at which correlations change with distance. The fact that correlations decay faster for high frequencies than for low frequencies means that when sources are closer together, the correlations at high frequencies will be more informative. This is shown in the left-most panel of Fig 12A, where we have plotted the information available at each frequency for an intersource distance of ∼0.1 pitch. At that intersource distance information shows a positive trend with frequency. As sources move farther apart, high-frequency correlations will have decayed, eventually changing at the same rate with distance as the slower decaying low-frequency correlations. This will result in all frequencies being similarly informative, as shown in the middle panel of Fig 12A for an intersource separation of ∼1 pitch. As sources move even farther apart, the lower frequencies will become more informative, as shown in the right panel of Fig 12A for an intersource separation of ∼2 pitches. The overall change in the trend as intersource distance increases is shown in Fig 12C and shows a distinctive elbow shape, so in what follows we will frequently refer to such data as ‘elbow plots.’ For surrogate data in which all frequencies are equally informative, trends in information with frequency are minimal at all intersource separations; see Fig 12B and 12C. Note that the exponential decay in correlations with frequency means that regardless of which frequencies are more informative, the absolute amount of information available will decrease with intersource distance. This is because information is derived from the rate of change of correlation with intersource distance, and this decreases as sources move farther apart. This effect is shown by the decrease in the down-scaling applied to the data in the panels of Fig 12A and 12B.

thumbnail
Fig 12. Fisher information vs. frequency.

(A) Base 10 logarithm of the Fisher information at each frequency component (dots) and linear fits (dashed lines), for intersource separations of ∼0.1, 1, and 2 pitches (left, middle and right panel, respectively), for the simulated plumes using a 1-second Hann window. Coefficients β of the linear fits are stated in the legends. Data have been scaled by the amount shown in each legend to ease comparison. (B) As in panel A but for surrogate data where all frequencies are equally informative. (C) Bootstrap median (dots) and 5th-95th percentiles (bands) of the coefficients β of linear fits to the logarithm of Fisher information regressed on frequency for the CFD plumes (blue) and 10 surrogate datasets where all frequencies are equally informative (gray). Positive coefficients indicate information increases with frequency, negative coefficients that it decreases. Centres of orange markers indicate the integral length scale (LU, see Methods Sec 5.2.8) in the vertical direction (orthogonal to the mean flow) of the vertical velocity field, computed at the midline where the odour sources were located (‘origin’, LU = 0.85ϕ) and at the probe location (‘probe’, LU = 2.6ϕ). Linear fits were computed using robust regression, see Methods Sec 5.2.5.

https://doi.org/10.1371/journal.pone.0297754.g012

The analyses above relied on the use of windows that reduced spectral leakage when the amount of power in different frequency bands is different, as is the case in real plumes and in our simulations. By using the Hann window to reduce leakage we were able to distinguish the different rates at which correlations decay in different frequency bands, and therefore the different amounts of information that they contain. As we show in S7 and S8 Figs these effects were not limited to the 1-second Hann window used above, and we also observed them when we used other windows that similarly reduced spectral leakage.

2.3.2 Testing other probe locations and source geometries.

The results we have presented above were computed at a single probe location, located on the midline of the flow downstream of the odour sources. Because animals will find themselves oriented at a variety of locations relative to odour sources of interest, it is important to determine the extent to which our results generalize to other locations within the plume. Therefore, we selected 8 additional probe locations within the plume, chosen to be dispersed enough to test the generalizability of our results, while also staying within the plume and avoiding boundary effects (see S1 Fig). These locations are indicated with coloured ‘×’s in Fig 13A. The blue ‘×’ marks the probe location for the results we’ve presented so far, while the remaining ‘×’s mark the new probe locations within the plume.

thumbnail
Fig 13. Results for other probe locations and source geometries.

(A) Plume snapshot and probe locations for the cross-stream source geometry. Sources are equally spaced at x = 0 along a line transverse to the mean flow. The plumes (colours) are generated by the most distal pair of sources (white ‘o’s). Probe locations are marked with coloured ‘×’s. The cross-stream source geometry is the one analyzed in most of the Main Text, using the data measured at the blue probe location. (B) The decay of correlations with intersource distance for a few example frequency bands, at the nine di.erent probe locations in panel A, indicated by the background colour of each panel. The data in the blue panel is for the principal probe location analyzed in the Main Text and was previously shown in Fig 9A. (C) Elbow plots for each probe location in panel A, and coloured accordingly. The blue curve is for the principal probe location analyzed in the Main Text and is the same as in Fig 12C. Only the bootstrap medians, and not the confidence intervals, are shown, for clarity. (D,E,F) As in panels A-C, but for the oblique source configuration, where sources are equally spaced along a line at 45-degrees to the mean flow. (F,I) As in panels A-C, but for the streamwise source configuration, where sources are equally spaced along a line parallel to the mean flow.

https://doi.org/10.1371/journal.pone.0297754.g013

We performed the analysis describe in the previous section for each new probe location, culminating in ‘elbow’ plots like those of Fig 12C indicating whether high or low frequences were more informative at each intersource distance. The panels in Fig 13B show how correlations decay with intersource distance at each probe location. These panels reveal that at every probe location high-frequency correlations decay faster than lower frequency correlations when sources are close together. In Fig 13C we have overlayed the resulting elbow plots for all probe locations, showing only the medians for clarity, and coloured according to the probe locations in Fig 13A. The blue curve is the mean data previously shown in Fig 12C, while the remaining curves show the results for the new probe locations. All of the curves show the same qualitative behaviour: each starts at a positive value for β at low intersource distance, indicating that high frequencies are more informative than low frequencies when sources are close together. As intersource distance increases, all the curves trend downwards, and most eventually reach negative values of β, indicating that low frequencies become more informative when sources are sufficiently far apart.

Our results thus far are for sources located perfectly ‘cross-stream’, transverse to the mean flow. Natural odour sources will assume many other orientations relative to the mean flow. It is therefore important to determine whether our results generalize to other source orientations. To do so, we performed simulations using two new source orientations.

In the first, odour sources were equally spaced on a line at 45 degrees to the mean flow. In Fig 13D we have plotted a snapshot of the plumes generated by two most distant sources in this configuration. The panels in Fig 13E show how correlations decay at each probe location for this source configuration, again revealing that high frequencies decay faster than low frequencies when sources are close together. The resulting elbow plots of Fig 13F reflect this and show that at all probe locations in this source configuration, high frequences are more informative than low frequencies when sources are close together.

In the second new source configuration, we placed our sources parallel to the flow (streamwise). Fig 13G shows the plumes generated by the most distant sources in this configuration. The correlation decay results in Fig 13H reveal the same qualitative trends as in the corresponding panels for the previous two configurations, with high-frequency correlations decaying faster than those at low frequencies when sources are close together. This is reflected in the the elbow plots of Fig 13I which show that high frequencies are more informative than low frequencies when intersource distances are low.

In summary, our analysis of 8 new probe locations and two new source configurations all reveal the same same qualitative trends that we observed in our original probe location and source configuration, namely that high-frequency correlations decay faster than low-frequency correlations when sources are close together (compare Figs 9A, 13B, 13E and 13H), resulting in high frequencies being the most informative when sources are closer together, with the balance shifting towards low frequencies as intersource distance increases (compare Figs 12C, 13C, 13F and 13I). We saw the same qualitative trends when we performed our analyses using longer windows (2-second Hann, S21 Fig), shorter windows (0.5-second Hann, S22 Fig) and windows with different shape (1-second Kaiser-16, S23 Fig).

Finally, we repeated our entire procedure for a supplementary set of simulations, and observed the same qualitative effects (S24 Fig), though there were exceptions for some probe locations and source configurations. For example, high frequencies decay more slowly than low frequencies for small intersource distances at the red and brown probe locations in the oblique plume (S24E Fig), which may contribute to low frequencies being more informative for nearby sources than high frequencies at the red probe location (note the inverted elbow in S24F Fig). Another prominent example comes from the same source configuration but analyzed using 500 msec. window, for which the elbows from a few additional probe locations are inverted (S26F Fig). However, all of these exceptions are for probe locations that are at the edge of the plume. This is partially due to the fact that odour sources are also closer to the probes in the supplementary simulations. Both of these properties may contribute to the behaviour we observe at those locations.

When computing the Fisher information for the new source geometries above, we continued to use the in-phase component of the correlations. Initially, we had expected that the out-of-phase component of the correlations would play a larger role in these new geometries. For example, in the setting sources are arranged parallel to the flow direction, the velocity of the flow and the spacing between any pair of sources imply an initial phase difference between them, which may survive until the flow reaches the probe locations and manifest as strong out-of-phase correlations. However, we found that the magnitude of the out-of-phase correlations for the two new source configurations (S18D and S19D Figs) was small and similar to that for the original, transverse source configuration (S17D Fig), justifying our approximation of the Fisher information to use only the in-phase correlations for those source geometries as well.

3 Discussion

3.1 Quantifying information in plume correlations

The concentration timeseries from two odour sources measured at a downwind location tend to become more correlated when the sources move closer together. Therefore, correlations contain information about the spatial separations of the odour sources that generated them. Correlations can be decomposed as the sum of component correlations at different frequencies. Recently it was shown that mice can discriminate correlated vs. anti-correlated concentration timeseries at up to 40 Hz [20]. Could this high bandwidth sensitivity be caused by additional spatial information those frequencies might carry? Motivated by this question we set out to quantify the information contained in the component correlations about the spatial location of odour sources, to determine whether high frequencies contain more spatial information than lower frequencies.

Our approach is based on parametric modeling of how component correlations decay with intersource distance. We first express the component correlation at each frequency in terms of the corresponding Fourier decomposition coefficients of the odour timeseries being correlated. We then fit parametric models to these correlations and how they change with intersource distance. Using these parametric models we then derived a closed form expression for the Fisher information contained in the component correlations about the intersource distance in terms of the correlation of the Fourier coefficients and how these change with intersource distance (Eq 23).

We applied our approach to two computational fluid dynamics simulations of two-dimensional grid turbulence to determine which frequency bands were most informative about relative source locations. To verify our approach we also applied it to several surrogate datasets which we constructed to contain different patterns of spatial information in their component correlations. Analyzing our data using a 1-second Hann window to reduce spectral leakage, we observed that high-frequency correlations decayed faster than those at low frequencies (Fig 9). This meant that high frequencies were more informative when sources were less than ∼1 pitch apart, and low frequencies when sources were farther apart than this Fig 12. We saw this effect for other similar leakage-reducing windows (S7 and S8 Figs), and not in surrogate data in which all frequencies were equally informative.

We tested the generalizability of our analysis by testing two additional source configurations and eight additional probe locations. In all of these cases, we found broad agreement with our orginal simulation, showing high frequencies were more informative than low frequencies Fig 13 when sources were close together. We observed the same qualitative effects when we tried different window sizes (S21 and S22 Figs) and window shapes (S23 Fig).

We repeated this procedure for our Supplementary simulations, and again saw the same qualitative effects as in the Main simulations for most source configurations and probe locations (see S24 Fig). There were, however, exceptions. For example, the high-frequency correlations at the red and brown probe locations in the oblique plume shown S24E Fig decay more slowly than the low-frequency correlations, which may contribute to high frequencies being less informative for discriminating sources close together using the plume signal at the red location. However, these and the other exceptions we found (e.g. in S26E Fig) all occured at the edges of the plume, where odour signals are more sparse. In addition, the sources in the Supplementary simulations are closer to the probes than those in our Main simulations. These two effects may account for the unusual signal properties we measured there.

In summary, in all 9 probe locations for all three different source geometries, three different window sizes and two window shapes, we saw the same effect in our Main simulations, namely, that high frequencies were more informative than low frequencies when sources were close together. We also observed this at most corresponding points in our Supplementary simulations, with the exceptions occuring at plume edges. We therefore believe our results are robust for probe locations within the plume.

3.1.1 Implications of statistical assumptions.

To determine the Fisher information contained in correlations about source separation, we required a probabilistic model of these correlations and how they vary with source separation. In our initial approach, we made simplifying assumptions about the Gaussianity of the plumes and their interactions and derived the predicted correlation distributions to be an asymmetric Laplacian (Eq 19). While this distribution fit the data well (Fig 7), we found that it deviated from the observations systematically due to its inability to describe the high probabilities of low correlation events (i.e. intermittency). This is perhaps unsurprising given that instantaneous plume distributions are known to be non-Gaussian [43] with the exception of large distances from the source, exhibiting transitions from exponential-like behaviors near the source through a log-normal transition to the far-field Gaussian regime [24]. To address these deficiencies in the correlation models, we then modeled the correlations directly using the asymmetric Laplacian as a template. We firstly extended it to incorporate intermittency by attributing very small correlations to noise. Secondly, in addition to the asymmetrical Laplacian, we also evaluated the ability of the Gamma and generalized inverse Gaussian distributions to describe the distribution of observations flagged as true correlations. We found that the extended models were able to capture the observed correlation distributions quite well across all source separations and frequencies (Fig 7H). The improved fits also indicate that the joint distribution of coefficients is not mulitvariate Gaussian. An interesting avenue for future investigation would be to determine the joint distribution of Fourier coefficients that would predict the same parameteric form for the distribution of correlations as our model when we fitted the correlations directly.

Ideally, we would have used these more sophisticated correlation models to compute the Fisher information; however, this computation would require describing how additional parameters depend on intersource distance, greatly increasing the difficulty of our analysis. We therefore used the simpler Gaussian assumption to compute the Fisher information. Would our findings hold if we used more accurate assumptions to derive Fisher information? Our simplified analysis suggested that an important quantity that determines Fisher information is the speed with which correlations decrease with intersource distance. A robust observation in our data, independent of any statistical assumptions, was that high-frequency correlations decreased more rapidly than low-frequency correlations. Because Fisher information is a measure of how rapidly the distribution of an observed quantity (like correlations) changes with respect to variations in estimated parameter (like intersource distance), the faster changes in high-frequency correlations with intersource distance that we observe will likely mean that more sophisticated computations of Fisher information using more accurate statistical models will still yield our core observation that high frequencies are more informative than low frequencies at small source separations.

3.1.2 Other considerations.

We aimed to determine whether high frequencies contained more spatial information than low frequencies. We found that this was the case when sources were close together. At intermediate source separations, all frequencies were equally informative. Low frequencies were more informative when odour sources were far apart. However, the absolute amount of information in these latter cases was much lower, and presumably more easily obscured by noise. Thus, when correlations contain enough spatial information to rise above the noise floor, it may be contained mainly in high frequencies. Therefore the sensitivity to high-frequency correlations recently observed in mice [20] may endow animals with fine spatial resolution when locating odour sources in the environment. We note, however, that diffusion acts to smear out high-frequency information more quickly than low-frequency information, given the sharper concentration gradients associated with these fluctuations. The distance that high-frequency information can persist in a given olfactory context is therefore set by the mean wind speed, the strength of turbulent straining, and the odourant molecular diffusivity.

To test the generalizability of our results we additionally analyzed plumes generated by sources arranged oblique and streamwise to the mean flow. An interesting result of this analysis was that the roll-off of the elbows shifted to larger intersource distances as the source separations transitioned from cross-stream to streamwise configurations (compare Fig 13C, 13F and 13I). In our developing grid turbulence flows, local integral length scales grow with streamwise distance (see S1A Fig), and these length scales relative to the source size set the plume dispersion characteristics. Therefore, sources separated in the cross-stream direction are introduced into similar length scales in the flow and thus disperse similarly, while sources separated in the streamwise direction disperse under the influence of different length scales in the flow. Furthermore, for the array of fixed probe locations considered here, sources separated in the cross-stream, oblique, and streamwise directions are located at different relative streamwise and lateral distances from the probe locations. These two effects likely drive the different roll-off behaviors observed in the elbow plots for different source configurations. Future work could investigate these connections.

We observed that correlations dropped approximately exponentially with intersource distance, see e.g. Fig 3C. This exponential decrease with intersource distance has been previously observed in the fluid dynamics literature [52]. These correlations are composed of the component correlations at each frequency component. Therefore, there is considerable latitude in the way the component correlations may decay: they are bounded by the variance at each frequency and must sum at each intersource separation to the overall correlation at that separation. Nevertheless, we observed that the component correlations all had the same, nearly exponential form (Fig 9A). Explaining the reason why the component correlations decay exponentially would be an interesting topic of future work.

Our work highlights the importance of window shape when analyzing signals where information is distributed across the frequency spectrum and where spectral power spans a wide range. Our information measures were computed for 1-second windows. Longer windows would reduce leakage effects, but would lengthen the time needed for behvioural decisions. An interesting question is how the distribution of spectral changes depends on the time windows used—or in ethological terms, which frequencies the olfactory system should attend to if odour source localizations have to be made quickly. We leave the investigation of these important questions to future work.

Fisher information is a ‘local’ measure about how a given signal can distinguish a given parameter value from neigbouring values. Our expression for Fisher information Eq 23 is accordingly local and depends on the value of component correlations ρn(s) at a given intersource separation and how they change around that point. To actually evaluate the Fisher information we used parametric fits to these component correlations (Eq 24). In contrast to the local nature of Fisher information, these parameteric fits are ‘global’ in that the fitted value at each intersource separation is influenced by all the data, not just those local to the separation. For example, deviations at large intersource separations can affect the length constant γ of the fit, which will in turn affect the value of Fisher information reported for small intersource distances (and all others). It will therefore be useful to investigate whether fitting approaches that are more local, such as splines, produce better estimates of the Fisher information.

We chose to compare the two plumes by computing their Pearson correlations. We chose the Pearson correlation because it is insensitive to the mean and scale of the signals being compared and only registers their covariation, and it is this covariation that reflects the relative separation of the odour sources. By our choice of Pearson correlation we do not mean to imply that it is the optimal measure for decoding intersource distance. It is, rather, a simple measure whose information content provides a lower bound for the total information present in the interacting plumes about source separation. Other measures, for example those based on event timings (see e.g. [53]) may be more effective for odour source localization.

We used Fisher information as our information measure. An obvious alternative is mutual information between the intersource distances and correlations. One advantage of Fisher information is that it is based on the likelihood function p(rn|s) only and does not require the specification of a prior distribution on intersource distances. Mutual information requires the joint distribution p(rn, s) of intersource distances and correlations and thus does require a prior on intersource distances—or alternatively a prior p(rn) and likelihood p(s|rn) for correlations, which seem harder to specify. However, this is not a major shortcoming as a prior on intersource distances would not be hard to motivate or to determine empirically for a given environment. The type of information provided by Fisher and mutual information are also different. Fisher information is ‘local’ in the unknown parameter (intersource distance) and indicates how discriminable a given value of that parameter is from another value infinitesimally close. Hence it is a function of the unknown parameter, as seen in Fig 10. It also bounds the variance of unbiased estimates of intersource distance from correlations, and would thus be particularly relevant if the animal needs to accurately localize the relative locations of odour sources. Mutual information is a ‘global’ measure in that its computations involves averaging over the full joint distribution of intersource distances and correlations. We leave the task of determining which of these two measures of information is most relevant for quantifying the spatial information in plumes to future work.

3.2 Dispersion, coalescence, and coherent flow structures in naturalistic plumes

Our multisource plume datasets were generated in two-dimensional, spatially-decaying grid turbulence. The statistics of these plumes are broadly consistent with those observed in diverse naturalistic olfactory contexts, regardless of interesting phenomenological differences between two-dimensional and three-dimensional turbulent flows [54]. These consistencies include the observation of positive correlation regimes far from the source with clear source separation effects (Fig 3), the importance of intermittency in describing correlation distributions (Fig 7), the success of exponential-like distributions in describing correlations (Eq 19 and Fig 7), and the exponential decay of correlations with increasing source separation (Fig 8). Given these broad consistencies, the analytical toolbox we present for quantifying the Fisher information in the spectral components of correlations about source separations is likely well-suited for applications to many odor landscapes relevant to diverse olfactory contexts.

Given the evidence for spectral information contained in correlations about source separation, what are the dynamical processes that drive these correlations? Lagrangian coherent structures (LCS) provide an intuitive framework for understanding the physical processes that drive coalescence of initially-distant odour sources and thus the statistics of their correlations. LCS arise from a dynamical systems view of fluid turbulence, where stable and unstable manifolds arise in regions of exponential fluid straining [55]. These manifolds represent repelling and attracting regions of the flow, respectively. Attracting structures are regions of disproportionate importance in odour dispersion because they closely correspond to spatial regions in which single odours evolve and initially distant odours coalesce in chaotic flow environments [56]. A representative instantaneous attracting LCS computed using the backwards finite-time Lyapunov exponent (FTLE) is shown in Fig 14 alongside the underlying flow-field. Ridge lines in the FTLE field show strong spatial correlations with odour pair coalescence.

thumbnail
Fig 14. Lagrangian coherent structures.

(A) An instantaneous snapshot of the flow field (vectors) and the concentration fields of a pair of odours (c1a in red and c1b in blue). (B) The corresponding attracting LCS quantified through the backwards-time FTLE field (grey) and “reaction rate” of the odour pair (the product of their concentrations, purple).

https://doi.org/10.1371/journal.pone.0297754.g014

The exact structures of LCS and their evolution in space and time will vary among turbulent flows occurring in diverse odor landscapes. The LCS dynamics will largely determine the correlations between initially-separate odour sources, and therefore the specific flow signature inherent in the LCS field will manifest in any analysis of plume information. Thus, we suggest that LCS be used to properly contextualize our analyses and findings in order to better relate our findings to naturalistic plumes. A full quantitative linking between spectral plume information and statistical characteristics of the LCS field could be an intriguing subject for future work.

3.3 Connections to optimal navigation and control

Our framework for quantifying spectral information about source separation may have implications for optimization of multi-source search strategies, both in an olfactory context and more widely in optimal navigation and control applications. In olfactory search, an important goal is not just to find a strategy to reach the target source, but to find the optimal strategy to locate the source, namely, one that minimizes time to arrival. Model-based approaches for search optimization rely on the agent possessing an internal model of probability distributions characterizing the environment (presumably through instinct in biological contexts). Concretely, recent research situated olfactory search within optimality theory by framing it as a partially-observable Markov decision process (POMDP), using Bayesian updates to source location belief based on an internal model of odor detection likelihood across plume regions [57]. This POMDP strategy optimized for a simplified single-source scenario, and in the more complicated multi-source case, integrating a model of inter-source distance likelihood based on high bandwidth signal correlations could prove useful in optimizing the search.

Even in model-free reinforcement learning approaches, the insights from our findings can be incorporated to enhance the agent’s policy development. By structuring rewards around the acquisition of high-resolution odor signals and their correlations, agents can learn to incorporate information gain about inter-source distances in their strategies. The selection between model-based and model-free approaches involves consideration of the trade-off between strategies that are most efficient in specific environments, versus those that are more robust to changes in, or imperfect knowledge of, the search environment. Typically, model-based approaches tend to be more efficient if the search landscape matches the agent’s internal model (i.e. result in lower mean arrival times), while model-free approaches tend to be more robust to external perturbations of the expected environment (i.e. result in shorter tails in arrival time distribution across a variety of scenarios) [58]. Our study could potentially improve optimal strategy development in either context, by improving the agent’s internal model in model-based approaches, or by guiding reward structure and/or observation resolution in model-free approaches.

Optimal navigation is often situated in optimal control contexts, in which control policies guide an agent’s movement based on continuous feedback from the environment [59]. Control policies could potentially benefit from leveraging high-bandwidth odor correlations to adjust the control policies dynamically. For instance, when high-frequency information indicates that sources are close together, control systems could fine-tune the agent’s trajectory to exploit this data for rapid convergence to the source. Conversely, when sources are farther apart, control policies could be adjusted to rely more on low-frequency signals, which are more stable over longer distances. More generally, recent investigations of optimal control in the navigation of complex environments suggested that search strategies improve with the amount of information available to the agent [60]. Our findings identify an additional source of information about relative source locations available at the local level, information that may provide significant value in multi-source search control strategies.

4 Conclusion

In conclusion, we have outlined an approach for quantifying the spectral information in odour plume correlations about odour source locations. Our analysis revealed that high frequencies are more informative about intersource distance than low frequencies for sources that are close together. We tested the robustness of our findings at nine probe locations for three different source configurations, three time window lengths and two window shapes, using two different simulated two-dimensional turbulent flows, and saw the same qualitative effects in all settings except occasionally at plume edges. Owing both to the variety of odour landscape characteristics in naturalistic plume environments and to the preeminent role of CFS in driving both dispersion and coalescence processes in natural as well as simulated plumes, we expect that the method and findings outlined here will also be applicable to describing the spatial information contained in three-dimensional plumes.

5 Methods

All of our simulation data, and the code for our simulationes and analyses are provided at https://github.com/stootoon/fisher-plumes.

5.1 Computational fluid dynamics

We numerically modeled two-dimensional grid-turbulence in a wide wind tunnel and introduced an array of unique odour sources to generate chaotic multisource plumes (Fig 2). These data allowed us to i. investigate spectral correlations between odour sources as a function of source separation, ii. compute spectral Fisher information content encoding source separation and location, and iii. evaluate spatial structure in these entities. To test the robustness of the analyses to generalized plume environments, we conducted a second set of numerical simulations using different flow and odour parameters yielding an additional plume dataset with different spatiotemporal characteristics. Below we detail the CFD methods used for all analyses in the Main Text. We provide an alternative set of supplemental data derived from other numerical implementations in the Supporting Information Sec S3.1 in S1 File.

5.1.1 Model domain, simulations, and meshing.

The model domain (Fig 2) consists of a wide rectangular wind tunnel (0.62 m x 0.4225 m, x by y) with a mixing grid cylinder array distributed across the flow development inlet section (0.12 m in streamwise length). An array of 16 odour sources (8 source pairs mirrored about the domain centerline at y=0) of width Lf was distributed spatially across a transect a short distance downstream of the mixing grid at x = 0, covering a range of source separations s for analysis (Table 1). Source y-locations were constrained to the central 1/4 of the total inlet width to minimize walls effects, i.e. any impacts on cross-stream plume spread due to growth of the velocity boundary layer along the lateral walls. A uniform inlet velocity upstream of the mixing grid produces a chaotic flow field downstream through the nonlinear interactions of wake structures shed by the mixing grid. Simulations spanned a dimensional duration 70 s consisting of i. a 10 s start up period to establish fully chaotic flow conditions throughout the model domain and ii a subsequent 60 s analysis period containing dynamic multispecies plumes generated by the source array issuing into the chaotic flow environment.

Numerical simulations were performed via finite element discretization of the Navier-Stokes and continuity equations (Eqs S1 and S2 in S1 File) governing fluid flow, and the coupled, non-reactive advection-diffusion equation (Eq S4 in S1 File) governing odour transport and diffusion (see Sec S2 in S1 File). The COMSOL Multiphysics package (ver. 6.0) was used to generate the mesh in the model domain described below and to solve the system of equations resulting from the weak-form discretization of the governing equations, subjected to prescribed initial and boundary conditions described below. Geometric, flow, fluid, and odourant properties for model runs in the domain depicted in Fig 2 are summarized in Table 1.

The unstructured finite-element mesh contained 289,607 total triangular and quadrilateral elements (110 elements/cm2 on average), locally refined in spatial regions with strong fluid velocity or odour concentration gradients, notably along the solid surfaces of the mixing grid cylinders and lateral walls and near the finite-width odour sources where strong initial odour gradients persist. The level of mesh refinement (increasing model degrees of freedom) was iterated to a final spatial resolution sufficient to resolve the smallest anticipated fluid velocity and odour concentration gradients. The final mesh implemented for all model runs had typical maximum and minimum element sizes (spatial resolution) of approximately 2 mm and 0.2 mm, respectively, with approximately 35 and 10 mesh elements per pitch ϕ and per source diameter Lf, respectively.

5.1.2 Initial and boundary conditions.

Initial conditions were u* = 0, p* = 0 and c* = 0 everywhere. The mixing grid and lateral walls were no-slip conditions, u* = 0, with zero total odour flux normal to the surfaces, −n ⋅ (J + u*c*) = 0, where the diffusive flux is J = −nDc* and u*c* is the advective flux (n is the unit vector normal). The velocity inlet condition was normal, uniform flow with speed Uo, u = Uon and the outlet condition on the downstream end of the domain was zero pressure p* = 0. For solver stability we ramped the inlet velocity from zero to Uo over short interval in time. Odour sources of strength c* = 1 were introduced as constant concentration constraints from finite-width locations (Table 1), where the source profile at the origins was a smoothed top hat. The inlet and outlet odour boundary conditions were no diffusive flux normal to the boundary −nJ = −nDc* = 0.

5.1.3 Discretization, solvers, and convergence.

Lagrangian shape functions were used for weak-form discretization of the fluid velocity, pressure, and odour concentration fields. The order of the integration scheme was matched to the element order for all dependent variable (first-order here). The time-dependent solver employs an implicit backward differentiation formula (BDF) method with maximum second order schemes, balancing numerical stability and damping tendencies for our smoothly varying velocity and scalar concentration gradients. BDF methods use variable-order, variable step-size backward differentiation and are known for their stability [61, 62]. The variable step size taken by the solver was informed by a prescribed absolute tolerance for the nonlinear solver and an implicit formulation of the mesh Courant-Friedrichs-Lewy (CFL) number. The solution sequence for the resulting systems of equations was fully-coupled in all dependent variables (fluid velocity and pressure, scalar concentration) for each solver iteration using an affine invariant form of the damped Newton method [63]. The nonlinear systems of equations were solved iteratively with specified convergence criteria using the direct PARDISO solver optimized for parallelized solutions of sparse systems of equations [64, 65]. The discretization schemes and solvers detailed above on the described mesh yielded good numerical stability and solution convergence with acceptable memory and computation requirements.

5.1.4 Lagrangian coherent structures.

To investigate the dynamics of the Lagrangian flow structures responsible for odour mixing we computed backwards finite-time Lyaponov exponent (FTLE) fields per [55]. The computation followed the methods described in [56], and the resulting backwards time FTLE fields provided a proxy for the attracting LCS in the flow. The integration time TLCS was set to 0.6 seconds based on preliminary investigations of the spatially-varying integral time scales. The temporal resolution of the LCS computation matched the resolution of the underlying velocity data (20 ms), and the initial spatial resolution the Lagrangian tracer grid was 250 μm. These integration parameters yielded well-defined FTLE fields with strong ridge lines that were in good qualitative agreement with spatial regions odour pair coalescence, suggestive of convergence in the FTLE computation.

5.2 Numerical methods

5.2.1 Computing plume correlations.

Fourier decomposition and windowing of plumes. To compute the correlations between a pair of plumes, a short-time Fourier transform (STFT) was applied to each plume using the scipy.signal.stft [66]. The STFT was computed for a specified window length using rectangular (boxcar) windows, with an overlap of half the window length, and using no padding or boundary. The application of different window shapes was implemented through the detrending function supplied to the STFT. The detrender first applied the desired window before z-scoring the result through mean-subtraction followed by scaling by its standard deviation. This detrending was performed so that the coefficients of the decomposition of two signal into trigonometric coefficients could be easily combined as in Eq 59 to yield the components of the Pearson correlation contributed by each harmonic. The conversion of Fourier coefficients to their trigonometric equivalents is completely standard so we have provided the derivation in the Supporting Information. The derivation shows that a signal x[n] of length L with discrete Fourier transform coefficients X[k] = uk + jvk can be expressed in terms of sines and cosines as (27a) (27b) (27c) where ⌊⋅⌋ is the floor function.

Computing correlations from Fourier coefficients. The component of the correlation between two plumes for a given time window in a given harmonic was computed by combining the sine and cosine coefficients from each source according to Eq 3. To compute the full Pearson correlation we summed the harmonic correlations over all harmonics as in Eq 2. To compute the distribution of correlations at a given intersource distance we pooled all correlations for all time windows computed for all pairs of sources separated by the given intersource distance. Cumulative distribution functions for some of this data are plotted in the top panels of Fig 7.

We fitt the distribution of correlations.

5.2.2 Modeling the dependence of Fourier coefficients on intersource distance.

To compute the empirical distribution of the Fourier coefficients from two sources at a fixed intersource distance s and for a given harmonic n, we first listed all pairs of sources that were the desired distance apart. Our source locations only differed in the y-coordinate, and we used a signed distance, so dist(yi, yj) ≜ yiyj. Therefore, a desired intersource distance d yielded a list of ordered pairs of sources (yi, yj) such that yiyj = d. We then concatenated the sine and cosine coefficients in the given harmonic for all time windows across all sources in the first coordinate of each pairing. This yielded a vector of length (28) where the factor of 2 is because of the concatenation of both sine and cosine coefficients. The length is a function of intersource distance s because the number of pairs a given distance apart was distance-dependent. We then stacked this vector on top of the same concatenation of data applied to the sources in the second coordinate of each pairing. This yielded a 2 × N(s) matrix Datan(s), (29) where each element of the first row was one Fourier coefficient computed in one time window from one source, and the corresponding element in the second row was the corresponding coefficient for another source a distance s apart. We plotted some of this data in Fig 6.

The columns of Datan(s) represent samples from the joint distribution p(an, cn|s) of Fourier coefficients for the n’th harmonic, from sources a distance s apart. To fit a zero-mean bivariate Gaussian to Datan(s) as in Eq 14b we computed the principle variances λn(s) and μn(s) along the major (y = x) and minor (y = −x) axes, respectively. To determine these we computed the variances of the projections of the data on the unit vectors and , respectively, (30a) (30b) We computed the variance as the mean of the principle variances at s = 0, (31) This is equivalent to computing the pooled variance of all coefficients from all sources for the given harmonic.

5.2.3 Parametric fits to ρn(s) as a function of distance.

We next fit each ρn(s) to the parametric form in Eq 24 by non-linear least squares using curve_fit from the scipy.optimize package [66]. The parameters to be learned were the length scale ι and the offset b. We constrained the parameters to be non-negative, and initialized ι to 1 and b to 0. All other parameters to curve_fit were kept at their default values. Because the data sometimes exhibited changes in length scale with intersource distance (see e.g. Fig 8B) we only used the data for intersource distances up to one pitch to compute fit parameters. Some example fits and parameters learned are in Fig 8.

5.2.4 Fitting the distribution of correlations.

We fit intermittent and non-intermittent versions of three different models to the correlation data. The models differed in the probability distributions they used to fit the correlations. Each model was based on a probability distribution on non-negative values. To extend the domain to cover negative values, the same base distribution was used but applied to the absolute value of the negative values, and with new set of parameters to cover these values, and the overall distribution normalized to one. That is, for a base distribution on non-negative values with parameters θ+, (32) the extended distribution was defined as (33) For example, the asymmetric Laplacian distribution in Eq 21 corresponds to a base Exponential distribution, with parameters λ = (1 + ρn(s))/2 covering the positive correlations, and μ = (1 − ρn(s))/2 covering the negative correlations.

Note: In what follows we will frequently refer to the extended models by their base distributions, so e.g. Exponential when referring to the asymmetric Laplacian distribution.

Evaluating models by comparing CDFs.

We evaluated correlation models by comparing their predicted cumulative distribution functions (CDFs) to those we observed.

To compute the empirical CDF for a set of N observed correlations, we first sorted the values in ascending order, to yield We then computed the empirical CDF evaluated at each data point ri as (34) The predicted CDF can be derived using the probability density function of Eq 95, which is (35) where ρ and η are the correlation and noise distributions, respectively (see Eq 99 for a particular case). The CDF is then a linear combination of the CDFs for the correlation and noise distributions: (36)

When comparing CDFs the R2 value can be overly optimistic since both CDFs start at zero and increase monotonically to 1. Therefore, we compared CDFs more stringently by computing the largest absolute difference in their values, which is bounded below by 0 and above by 1. That is, we computed (37)

When numerically computing the predicted CDFs we sometimes encountered values outside of the valid [0, 1] range. This was particularly the case when using scipy.stats.geninvgauss to compute the CDF of the generalized inverse Gaussian. Therefore, when computing fit quality we only used the subset of points for which the CDF values were valid.

Fitting models using expectation-maximization.

To fit models with intermittency we had to determine which of a set of observed correlations {y1, …, yN} at a given harmonic and intersource distance were attributed to noise. From Eq 95 this requires determining the values of the binary latent variables {z1, …, zN} corresponding to each of the observations. To fit the parameters θ of a model while accounting for these latent variables, we used the Expectation-Maximization algorithm [67]. Initialized with an initial guess of the model parameters, this algorithm consists of repeatedly estimating the values of the latent variables (the ‘E’-step), then using those estimates to update the estimates of the model parameters (the ‘M’-step). Below, we summarize the parameter updates. Derivations are supplied in Sec S5 in S1 File.

The E-step update for all models at the t+ 1’st iteration has the simple form (38) Here ρ(yi|θt) and η(yi, θt) are the probabilities of the observation yi under the correlation, and noise, distributions, respectively, θt is the current estimate of the model parameters, and ιt is the current estimate of the intermittency parameter. The E-step therefore declares an observation yi to have been a correlation rather than noise if its probability according to the correlation distribution is greater than its probability according to the noise distribution by a threshold that depends on the current estimated intermittency level. For the non-intermittent models, all observations were marked as being correlations.

During the M-step model parameters were updated using the current estimate of the latent variables {z1,t, z2,t, …, zN,t}. All models had two parameters that related to intermittency: the intermittency level ι, and the standard deviation σ of the noise distribution. These parameters were updated the same way for all models. For the non-intermittent models, ι was fixed at 1, and σ at 0. Otherwise, these parameters were updated as described below.

To update the intermittency level ι, it is helpful to define the data intermittency at time t as (39) The intermittency level is updated by balancing the data intermittency against a prior on intermittency, , (40) This can be interpreted as weighting observations with an intermittency of against N observations at the data intermittency . The hyperparameter was fitted using grid search (see below). The hyperparameter α, which sets the relative strength of the prior, was fixed at 1.

We updated the noise level by setting its variance to be the mean sum of squares of the observations attributed to noise in the E-step. Intuitively, this is because under our assumption that the noise distribution has mean zero, the mean sum of squares of the noise observations is an estimate of its variance. If the E-step did not flag any observations as noise we set the estimated noise variance to a small fraction of the variance of the observed correlations. We did this because setting the noise level to exactly zero would mean that the estimated noise variance would remain at zero for all future iterations. This is because the probability of any non-zero observations would be 0 under this degenerate noise distribution, so all observations in the next E-step would also be flagged as correlations and not noise, leaving the noise variance stuck at zero.

Letting be the number of observations attributed to noise after the current E-step, (41) where is the variance of the observed correlations.

The rest of the parameters varied by model so their updates will be described for each model separately.

Exponential model. In addition to the intermittency parameters, the Exponential model had two parameters, λ and μ, describing the decay rate of the positive, and negative, correlations, respectively. The M-step updates for these parameters had the closed form (42a) (42b) Here is the number of observations that were designated as correlations in the E-step. These can be split into positive and negative correlations, and and are the averages of the absolute values of the corresponding observations. The maximum operations prevents μt from being set to zero when the E-step did not designate any observations as noise. The added factor of 10−8 in the denominators prevents division by zero when Nt = 0 i.e. all observations were estimated to be noise.

Gamma model. The Gamma model has all the parameters of the Exponential model, plus two shape parameters, k and m, to describe the positive and negative correlations. To update the parameters during the M-step we minimized the negative log likelihood of the observations that were marked as correlations in the E-step: (43a) (43b) The minimization was performed using Nelder-Mead method as implemented by the scipy.minimize [66] function. The bounds of the search were [10−6, 10] for λ and μ, and [0, 10] for k and m. The search was initialized at the parameter values from the previous iteration, clipped to lie within these bounds.

Generalized inverse Gaussian. The Generalized inverse Gaussian model has all the parameters of the Gamma model, plus two additional shape parameters, α and β, for the positive and negative correlations, respectively. During the M-step we performed a damped update of the parameters towards the minimum of the negative log likelihood of the observations that marked as correlations in the E-step: (44a) where Kv(a, b) is the modified Bessel function of the second kind of real order a evaluated at b. Damping was used to aid convergence, with the damping factor δ fixed at 0.5. The minimization was performed the same way as for the Gamma model, with the new parameters α and β bounded to [106, 10]. The search was initialized at the parameter values from the previous iteration, clipped to lie within these bounds.

Ancestral initialization.

The Gamma distribution subsumes the Exponential distribution, because the former reduces to the latter when the shape parameter is set to 1. In turn, the generalized inverse Gaussian distribution subsumes the Gamma distribution, because the former reduces to the latter when its β parameter (the coefficient of 1/x in the argument of the exponential) is set to zero. Because models with fewer parameters can be easier to fit, to aid the fitting process we fit our models ancestrally. That is, to fit a model using the Gamma distribution we first fit the data using an Exponential distribution, and then initialized the Gamma fit with the parameters of the Exponential, initializing the shape parameter of the Gamma to zero. Similarly, when fitting a generalized inverse Gaussian, we initialized its β parameter at zero, and initialized the remaining parameters from those of the best fit Gamma. The Gamma in turn was initialized with the parameters of the best fit Exponential.

Evaluating models using nested cross-validation.

We fit a range of models to the correlation data by testing all combinations of the following hyperparameters:

  • Whether the model was intermittent or not;
  • If intermittent, the mean of the prior Beta distribution on the intermittency parameter, for the values {0.1, 0.2, …, 0.9}.
  • The base probability for the correlations, whether Exponential, Gamma, or Generalized Inverse Gamma.

The remaining hyperparameters, listed below, were held constant at the stated values

  • The minimum (0) and maximum (10) shape parameters for the Gamma and Generalized Inverse Gaussian distributions.
  • The strength (1) of the Beta prior on intermittency.

We characterized these models in two ways. First, we estimated the performance of each model in fitting the correlation data. To do so, we repeatedly split the data into training (67%) and test (33%) sets. Each model was fitted to the training set and evaluated on the test set. Its performance was estimated as its average performance over three such random splits.

Secondly, we estimated the rank of each model, so that we could say e.g. which model was best overall, or which was the best among those with intermittency. One way to do this would be simply by ranking the performances we computed above. However, doing so would risk overfitting. To see why, note that by ranking the models we are in effect fitting the hyperparameters. Ranking the models using the same data used to evaluate their performance would then mean estimating performance with the same data used to fit the (hyper)parameters, risking overfitting.

To estimate the rank of each model while avoiding making this estimation on the same data that was used to estimate performance, we performed an inner loop of cross validation for each iteration of the outer loop. In each iteration of the inner loop, the training set provided by the outer loop was further split at random into a training subset (67%) and a validation set (33%). Models for every setting of the hyperparameters were fit on the training subset, and evaluated on the validation set. Models were ranked by their mean validation performance over three such random splits. The resulting model ranks were then averaged over iterations of the outer loop.

To see why this nested cross-validation procedure avoids overfitting, observe that each iteration of the outer loop yields, first, an estimate of model performance as computed on the test data, and second, model rank as determined from the training data by the inner cross-validation loop. Therefore, in each outer loop iteration, model rankings and model performance are estimated using different data. We then reduce the noise in these estimates by averaging over iterations of the outer loop.

5.2.5 Regressing information on frequency.

To produce the results in Fig 12 we regressed, for each value of intersource distance, Fisher information on frequency. The Fisher information at some intersource distances exhibited outliers that overly influenced the results of a standard linear regression, as judged by visual inspection. In addition to filtering out data points for which the computed Fisher information had NaN, infinite, or negative values, we took two measures to make the results more robust to these outliers. First, we limited the range of frequencies considered to be 1 Hz to 15 Hz. This was because some of the data, particularly in the supporting simulations, showed sudden drops in Fisher information above ∼15 Hz. Secondly, we replaced ordinary linear regression with robust regression. We used Huber regression, as implemented in scikit-learn’s linear_model.HuberRegressor. We used the default parameters, except for max_iter which we set to 10,000.

5.2.6 Bootstrapping procedure.

To estimate the variability of the various statistics used in our approached we computed their bootstrap distributions. These statistics were computed from the trigonometric form of the Fourier decompositions computed for each plume in each time window. Therefore we first computed these coefficients for all time windows, and then created each bootstrap dataset by sampling time windows with replacement until we had as many time windows as the original dataset. We generated 50 bootstrap datasets in this way. We then computed our various statistics using the data for the time windows chosen for each bootstrap dataset. This then gave us as many point estimates of each statistics as bootstrap datasets, from which we computed e.g. the 5th, 50th and 95th percentiles of Fisher information shown in Fig 10A.

5.2.7 Generating surrogate data.

In brief, we generated surrogate data consisting of artificial plumes with covariance structure similar to real plumes but for which we could adjust the amount of spatial information. To generate these plumes we randomly generated coefficients according to the bivariate normal model of Eq 14b using covariance kernel K(n, s) to relate the coefficients from plumes a distance s apart. Specifically, (45) where as usual, am and cn are the cosine coefficients for the m and n’th frequency components, generated by two plumes a distance s apart, bm and dn are the corresponding sine coefficients, and expectations are taken over time windows, and the δ function δ(m, n) ensures that coefficients for different harmonics are uncorrelated. We combined these coefficients with their corresponding sine and cosine waveforms to generate the surrogate signals.

In detail, we generated signals for M sources, each of length 2⌊L/2⌋ + 1, by first creating the kernel relating the trigonometric coefficients of their Fourier decompositions. The signals had zero-mean, therefore for each source, we needed to specify ⌊L/2⌋ sine and cosine coefficients, corresponding to the normalized frequencies , covering all positive harmonics of the fundamental up to the Shannon limit. The kernel needed to specify how each harmonic at every source was correlated with each harmonic at every other source, a total of M2L/2⌋2 values. We organized these values into a symmetric kernel matrix of size ML/2⌋ × ML/2⌋. The elements of this matrix were determined as (46) where k(i, j, n) was the desired covariance of the trigonometric coefficients for the n’th harmonic for sources i and j (see below). We assumed coefficients at different harmonics were uncorrelated.

To compute cosine coefficients with covariance K, we first computed the Cholesky decomposition K = LLT. We then generated ML/2⌋ i.i.d. samples from a standard normal, u, and determined the cosine coefficients as c = Lu. We generated a second set of samples, v, from the standard normal in the same way, and used those to produce the sine coefficients s = Lv. Indexing these coefficients by source m and harmonic n, we produced the signals at each source by combining the coefficients with their corresponding trigonometric waveforms, (47)

Kernel functions.

We used several different kernel functions k(i, j, n) to generate the surrogate data in the text. Each of these kernel functions was the product of a term S(n) determining the power at the n’th harmonic, and a function G(i, j, n) determining the correlation between sources i and j, that in some cases depended on the harmonic. In every case, this correlation function was a function only of the absolute difference s = |ij| of the sources, so it simplified to G(s, n). Thus our kernels were of the form (48)

Our kernels differed by whether they used a flat power spectrum or one similar to that of the simulations. We captured both cases by using a 1/f power spectrum: (49) where fs was the sample rate in Hz. The maximum operation provides a frequency cutoff at 1 Hz below which the power is set to 1. To achieve a ‘white’ spectrum, we set α = 0. To achieve a ‘pink’ spectrum similar to our CFD simulations, we set α = 4.

Our surrogate data also differed in the correlation functions used. For the data where all frequencies were equally informative, we set (50) For the data where the high frequencies were more informative for sources close together than the lower frequencies, we set (51) Thus the length scale of decay was 6 times shorter for high harmonics (those in the upper half of the range) than for low harmonics.

The table below summarizes the surrogate datasets and the power and correlation functions used in each.

5.2.8 Computing integral length scales.

We computed integral length scales from velocity autocorrelation functions as defined in [68]. For example, to compute the integral length scale in the y-direction at a location (x, y) of interest, we first computed the y-velocity autocorrelation function in the y direction at that location by computing the time average of the product of the y-velocity at the location of interest, uy(t;x, y) and that at a fixed positive displacement r in the y-direction, uy(t;x, y + r). That is, we computed (52) We also computed the symmetric average in the negative y-direction, (53) We then averaged these two to arrive at the autocorrelation for a single, positive value of r (54) If e.g. (x, y + r) was not in the domain, then would not be computed and Qyy took the value of , and similarly if (x, yr) was not in the domain.

We then computed the longitudinal velocity correlation function by normalizing the velocity correlation function by its value at r = 0 (55) Finally, we computed the integral length scale by integrating this function from 0 to ∞: (56) The r values we used were discrete and spaced Δr = 0.02ϕ apart, so we approximated the integral above as (57) where N was the number of r values used.

In S20 Fig we have plotted the longitudinal velocity autocorrelation functions in the x and y directions, and the corresponding integral length scales, for two locations of interest.

5.3 Analytical methods

5.3.1 Decomposing Pearson correlation.

We can express two concentration profiles x(t) and y(t) observed over a window of width T in terms of their Fourier decompositions, (58a) (58b)

Using the orthogonality of different harmonics of the fundamental frequency we can express the Pearson correlation in terms of the Fourier components as (59)

5.3.2 Relating the coefficients at one source to those at another.

To determine the uncertainty about the coefficients cn, dn of a component waveform at a second source, (60) given the coefficients an, bn of a first, (61) we express the former as the sum of a deterministic component formed of a scaled and phase-shifted version of the component waveform from the first source, plus noisy residual. That is, (62) where βn and θn are the best-fit scaling and phase-shift of the first source. These can be found to satisfy (see below) (63a) (63b) where as usual expectations are over time windows.

The residual ε(t) therefore captures the uncertainty remaining about the second component waveform, given knowledge of the first. To determine its statistical properties we express it by subtracting Eq 62 from Eq 61 as (64) This is a linear combination of sinusoidal waveforms at the same frequency ω, so we can write it as (65) By performing trigonometric expansions, the coefficients are found to be (66a) (66b) For n > 0, an, bn, cn, and dn have mean zero (Eq 8). Therefore the coefficients above, and in turn ε(t), as linear combinations of zero-mean random variables, are also zero mean.

To determine the variance of the residuals, we first show in Supporting Information Sec S6.2 in S1 File that the coefficients and are uncorrelated. Therefore, from Eq 65, the residual variance is the scaled sum of the variances of the coefficients. To compute these variances, we have (67) (68) where we’ve used the relations in Eq 66 to express coefficient correlations in terms of βn and θn, and the final equality follows by temporal stationarity. We then arrive at (69) Knowledge of the mean and variance would completely specify the probability distribution of the residual if it were a Gaussian random variable. Although it can be expressed as a linear combination of the Gaussian random variables an, bn, cn and dn (Eq 64), these latter variables are not necessarily independent. Therefore, their combination is not necessarily Gaussian. To enforce this requirement we made the assumption 2.2.

To determine the scaling βn and phase-shift θn, we equate Eq 62 an Eq 62 and match coefficients of cos(ωt) and sin(ωt), to get (70a) (70b) Correlating these equations against an and bn, respectively, and using Eq S29 in S1 File and the definition Eq 10 of the coefficient variances , we get (71a) (71b) Correlating against bn and an, respectively, and using Eq S30 in S1 File, we get (72a) (72b)

5.3.3 Deriving the distribution of correlations.

Here we derive the distribution p(rn|s) of the component correlations rn for the n’th harmonic for sources separated by a distance s, given our assumptions about the distribution of Fourier coefficients, an and bn, from the first source and cn and dn from the second source.

Accounting for location dependence.

We must first account for the fact that the distribution of correlations is formed by the contribution of all pairs of sources separated by the given distance. To do so, we first assign integer indices to the sources. A pair of sources is then (i, j), the distance between them, Dij, and the number of pairs a distance s apart, Ns. We can then use the rules of probability to express the distribution of correlations in terms of the contributions from each pair (73) The first term in the summand is the probability of the pair (i, j) being selected given the sources are a distance s apart. Because pairs at this distance will be selected uniformly and all others will not, (74) Our decomposition then simplifies to (75)

The summand is the distribution of correlations, given the pair of sources generating them and the distance between those sources. However, once the pair generating the correlations is specified the distance between the pair is redundant and we can remove it from the conditioners, yielding (76)

The new summand is the distribution of correlations generated by a given pair of sources. This in turn is related to the Fourier coefficients from each source. Let an, bn be the cosine and sine coefficients from the first source in the pair, and cn, dn the corresponding coefficients from the second source. Then (77) The distribution of correlations for a given intersource distance is then (78)

This expression depends explicitly on the identity of the sources, because the joint distribution of coefficients generated by each pair of sources can be different for each pair. To emphasize the relationship of coefficients from one source to those from the other we decompose the joint distribution as, (79) To quantify effects that depend only on the relative location of sources, rather than the specific locations of specific sources we made location our independence assumptions. The first was that the distribution of coefficients from a source was the same for all sources, so (80) The second was that the distribution of coefficients cn, dn from source j given the coefficients an, bn from source i depends only on the distance between them, (81) Substituting these two assumptions into Eq 79, and that result into Eq 79 yields Eq 5.

Deriving the distribution of correlations.

Dropping the n subscripts and the conditioning on s for clarity, (82) From Eq 12 We have that (83) By defining z = (a, b), x = (c, d) we can write Eq 83 as (84) Combining this with Eq 10 to get p(z) we have (85) where we’ve used Eq 10 to specify p(z). The integrand in Eq 85 is rotationally invariant in z since any rotation in z is matched by x through the δ function. So we can compute Eq 85 by computing it for one radial slice of z, say z = z e1 = (z, 0), and multiplying the result by 2π. Switching z to polar coordinates z = (z, θ), for which (86) we have (87) To compute the latter integral, we switch x from (c, d) to (u, t) coordinates, defined for z = (z, 0) as (88) Then the area element dcdd becomes 2z−1dudt and we have for the latter integral in Eq 87, (89) (90) (91) Now for our slice z = (z, 0), , so (92) Hence Eq 91 evaluates to (93) Substituting this in for the latter integral in Eq 87, we have Finally, to arrive at Eq 19, we simplify the length constant of the exponential by relating η2 in Eq 69 to Z in Eq 20 as η2 = −σ−2Z2σ2β2cos(ϕ)2, and use the definition of ρ in Eq 15 to get (94)

5.3.4 Improving the model of correlations.

Our correlation model captures intermittency by introducing a binary random variable z that determined whether the observation yn was of a correlation rn, or noise w. That is, (95) so that when the masking variable z = 1, we observe a ‘true’ correlation rn, and when z = 0, we observe an instance of noise instead. We assume that the masking variable z obeys a Bernoulli distribution, (96) The intermittency parameter ιn(s) ∈ [0, 1] determines the fraction of observations we deem to be true correlations, and can depend both on the harmonic n and the intersource distance s. The true correlations rn are drawn from the asymmetric Laplacian distribution of correlations in Eq 21, (97) while the noisy observations w come from a zero-mean Gaussian (98) Like the intermittency parameter, the variance of the noise can depend on both the harmonic and intersource distance. The predicted distribution of correlations for our ‘intermittent asymmetric Laplacian’ model is then a linear combination of the ‘true’ correlation distribution p(rn|s) and the noise distribution determined by the intermittency parameter ιn(s) (99) Setting ιn(s) to 1 corresponds to our initial asymmetric Laplacian model, while setting it zero represents a fully intermittent case where all correlations are in fact noise. The amount of intermittency and the variance of the noise were determined from the data (see Methods, Sec 5.2.4).

To demonstrate the effect of including intermittency in our model we have added the CDFs for the distributions in Eq 99 fitted to the data in Fig 7A–7C to the corresponding panels in those plots (light blue traces), revealing an improved qualitative agreement. In Fig 7D–7F we have added the deviations between the data CDFs and those of the new model that includes intermittency. The panels show that both the positive and negative deviations have been reduced, as expected, but not eliminated.

The fact that both positive and negative deviations remain in Fig 7D–7F indicates that the data have more small values than even augmentation with intermittency can capture. Therefore, we extended our model by replacing the Exponential distributions in Eq 21 that constitute the positive and negative halves of the asymmetric Laplacian with distributions that had higher densities near zero, but that included the Exponential as special cases. One candidate is the Gamma distribution, p(r) ∝ rk−1er. Just like the Exponential distribution, the Gamma distribution has a scale parameter, λ for the positive correlations above. In addition, it introduces a shape parameter, k > 0. When k = 1, the Gamma distribution reduces to the Exponential distribution. When k < 1, the density approaches infinity as r → 0. This singularity at 0 aids in modeling the large number of low correlation values we observe. Analogously to Eq 21, we ‘sandwich’ two Gamma distributions together so that we can also cover negative values (see also Methods Sec 5.2.4), and arrive at our correlation distribution (100) Here Z is the normalizing function of the Gamma distribution, k and λ are the shape and scale parameters for the positive correlations, and m and μ are the corresponding parameters for the negative correlations. We have omitted the dependence of these parameters on the harmonic n and the intersource distance s for clarity.

Another probability distribution, which includes the Gamma (and therefore the Exponential) as a special case, is the generalized inverse Gaussian. The parameterization of this distribution that we use gives the probability density function p(r) ∝ rk−1eα(r/λ+λ/r)/2. This distribution introduces a further shape parameter α > 0 and a term proportional to 1/r in the argument of the exponential that provide additional flexibility in modeling small observations. The scale of the distribution can be taken to be the length constant of the r term in the exponential, 2λ/α. Keeping this constant while reducing αλ towards zero, one approaches the Gamma distribution. As before, we sandwich two such distributions together to arrive at our distribution of correlations, (101) Here Z is the normalizing function of the generalized inverse Gaussian distribution, k, λ, and α are the parameters for the positive correlations, and k and μ and β are the corresponding parameters for the negative correlations. As before, we have omitted the dependence of these parameters on the harmonic n and the intersource distance s for clarity.

Our elaborated models of the correlations were thus of the same form as Eq 99, with p(rn|s) set to either asymmetric Laplacian, or sandwiched versions of the Gamma, or generalized inverse Gaussian distributions.

5.3.5 Fisher information.

To compute Fisher information for the distribution in Eq 19, we need (dropping subscripts for clarity), Then, defining λ = σ2ρ (102) Now defining λ = σ2ρ, we have so (103) Then the square is (104) We need to compute the expectation of this with respect to r under p(r|s). We can split this into the contributions from each component.

Constant term.

The expectation of the constant term is just (105)

Linear term.

Defining we have (106)

Quadratic term.

Defining we have (107) (108)

Fisher Information.

Collecting terms, we arrive at (109) This expression simplifies significantly if we assume that λ(s)≈0, since then Zσ2, Z′ ≈ 0, and (110) which is Eq 23. For ρ = (1 − b)es/γ + b as in Eq 24 we get, in terms of normalized distances , (111) which is Eq 26.

Supporting information

S1 Fig. Time-averaged statistics of simulated flow field.

(A) horizontal (top) and vertical (bottom) components of the mean velocity (left) and integral length scale (right) over the simulated plume domain. (B) mean turbulent isotropy (top) and turbulent kinetic energy (bottom) over the simulated plume domain.

https://doi.org/10.1371/journal.pone.0297754.s002

(TIF)

S2 Fig. Location independence.

p-values of the energy test measuring the difference between the distribution of coefficients at one source (‘First source’) and those at another, at each frequency. Each column corresponds to the comparison for one pair of sources.

https://doi.org/10.1371/journal.pone.0297754.s003

(TIF)

S3 Fig. Joint distribution of coefficients.

p-values of the energy test measuring the difference between the distributions of coefficients from all pairs of sources at a given intersouce distance (indicated on the x-axis), for each frequency. Each column corresponds to the comparison of one pair of sources. There were multiple pairs of odour sources at each intersource distance, indicated by the sizes of each block. Colours as in S2 Fig.

https://doi.org/10.1371/journal.pone.0297754.s004

(TIF)

S4 Fig. Gaussian coefficients.

p-values of the energy test measuring the difference between the distribution of coefficients at each source to those of a bivariate Gaussian with the same mean and covariance, at each frequency. Colours as in S2 Fig, white indicates p < 0.001.

https://doi.org/10.1371/journal.pone.0297754.s005

(TIF)

S5 Fig. Stationarity.

(A) p-values for the energy test comparing the distribution of sine coefficients to that of the cosine coefficients, at each source and frequency. (B,C) p-values of the Wilcoxon signed-rank test for whether the medians of the distributions of the sine, and cosine, coefficients is zero, for the data from each each source and each frequency. (D) p-values of the Wilcoxon signed-rank test for whether the median of the product of the sine and cosine coefficients is zero. Colours as in S2 Fig, white indicates p < 0.001.

https://doi.org/10.1371/journal.pone.0297754.s006

(TIF)

S6 Fig. Conditional Gaussianity.

p-values of the energy test measuring the difference between the conditional distribution of coefficients at one source given those at another (the ‘conditioning source’), and a bivariate Gaussian with the same mean and covariance, for each frequency. Each column corresponds to the comparison for one pair of sources. Colours as in S2 Fig, white indicates p < 0.001.

https://doi.org/10.1371/journal.pone.0297754.s007

(TIF)

S7 Fig. Frequency decomposition of correlations for the Main Text data, using a Kaiser-16 window.

Compare to Fig 9.

https://doi.org/10.1371/journal.pone.0297754.s008

(TIF)

S8 Fig. Fisher information vs. frequency for the Main Text data, using a Kaiser-16 window.

Compare to Fig 12.

https://doi.org/10.1371/journal.pone.0297754.s009

(TIF)

S9 Fig. Description of the second set of computational fluid dynamics simulations.

(A) Domain schematic and simulation parameters. Flow was from left-to-right. Vorticity was introduced into the flow by twelve cylindrical obstacles, vertically spaced evenly at a horizontal distance of 20 cm from the inlet (see also panel B). Point odour sources were placed at various vertical locations on the dashed line 40 cm horizontally from the flow inlet. Flow over these sources carried the odours to downstream probes. (B) Details of the obstacles and the simulation mesh. The cylindrical obstacles were 38 mm in diameter and evenly spaced 38 mm apart. The resulting center-to-center distances of 76 mm defined the pitch. The domain was discretized using a mesh with 2 mm resolution.

https://doi.org/10.1371/journal.pone.0297754.s010

(TIF)

S10 Fig. Probe locations and source geometries.

The full set of simulations and probe locations used in our study. Each panel shows a snapshot of the plumes from the most distal sources, with the two source locations indicated by the white ‘o’s. All simulations used 16 equally spaced sources, inclusive, between the two distal locations indicated, except for the simulations in panel F, which used 32 sources. The nine probe locations in each simulation are marked with ‘×’. (A) The main simulations used in our study, with sources transverse to the direction of the flow. The principal probe location that we discuss in the Main Text is at the blue ‘×’. (B) Simulations with sources at 45 degrees to the flow. (C) Simulations with sources parallel to the flow. (D) The principal supplementary simulations we analyze in the main text, with sources transverse to the flow. The principal probe location that we analyze is at the blue ‘×’. (E) Supplementary simulations with sources at 45 degrees to the flow. (F) Supplementary simulations with sources parallel to the flow. This set of simulations used 32 sources, rather than 16.

https://doi.org/10.1371/journal.pone.0297754.s011

(TIF)

S11 Fig. Example plumes and concentration profiles for the supplementary set of simulations.

Compare to Fig 3.

https://doi.org/10.1371/journal.pone.0297754.s012

(TIF)

S12 Fig. Frequency decomposition of correlations for the second set of simulations, using a 1s Hann window.

Compare to Fig 9.

https://doi.org/10.1371/journal.pone.0297754.s013

(TIF)

S13 Fig. Fisher information vs. frequency for the data in the second set of simulations, using a 1s Hann window.

Compare to Fig 12. The fits are at for intersource distances of 0.1, 0.4 and 0.7 ϕ. The intersource distances for the supplementary data start at lower pitch values than for the data in the Main Text because the pitch for the supplementary simulations is ∼3× larger.

https://doi.org/10.1371/journal.pone.0297754.s014

(TIF)

S14 Fig. Amplitude spectra of odour concentration profiles from all odour sources measured at the probe location, for the two simulated flows and the three surrogate datasets used in the text.

Discrete Fourier transforms were computed for consecutive 1-second windows that overlapped by 500 msec, amplitudes were averaged and scaled to have the same value at 1 Hz. Surrogate datasets are indexed by their information content (‘all=’: all frequencies equally informative; ‘high>low’: high frequencies more informative than low frequencies). The surrogate datasets (all =) and (high > low) were used in Fig 11 panels B, and C, respectively.

https://doi.org/10.1371/journal.pone.0297754.s015

(TIF)

S15 Fig. Modeling the distribution of observed correlations.

As in Fig 7 but showing the fits to the 1 Hz data, highlighting the poor fits to the data.

https://doi.org/10.1371/journal.pone.0297754.s016

(TIF)

S16 Fig. Modeling the distribution of observed correlations.

As in Fig 7 but showing the fits to the 1 Hz data, and when computing all statistics over 2-second Hann windows instead of the 1-second windows used in Fig 7.

https://doi.org/10.1371/journal.pone.0297754.s017

(TIF)

S17 Fig.

Coupling of concentration profiles from two sources at each frequency as a function of intersource separation, expressed in terms of (A) phase and (B) strength of the coupling. (C) Strength (saturation) and phase (hue) together. (D) Out-of-phase correlations, computed as β sin(θ).

https://doi.org/10.1371/journal.pone.0297754.s018

(TIF)

S18 Fig. As in S17 Fig but for sources that are arranged parallel to the flow (see e.g. Fig 13A).

https://doi.org/10.1371/journal.pone.0297754.s019

(TIF)

S19 Fig. As in S17 Fig but for sources that are arranged at 45 degrees to the flow (see e.g. Fig 13C).

https://doi.org/10.1371/journal.pone.0297754.s020

(TIF)

S20 Fig. Velocity autocorrelation functions and corresponding integral length scales (LU) for the simulations in the Main Text, evaluated along the midline (y = 0) at x = 0 (the x location of the odour sources, labeled ‘origin’) and at the probe location (labeled ‘probe’).

(A) Velocity in the x-direction (parallel to the mean flow) autocorrelated along the same direction. There are fewer data points at the ‘probe’ location since the largest x-displacement at that location is to the origin, while the largest x-displacement for the ‘origin’ extends past the probe location. (B) Velocity in the y-direction (perpendicular to the mean flow), autocorrelated along the same direction. The maximum displacement is approximately half that of panel A because the data in that panel spans the entire width of the simulation domain, while the data in this panel only extends from the midline to the upper and lower boundaries.

https://doi.org/10.1371/journal.pone.0297754.s021

(TIF)

S21 Fig. Elbow plots for other probe locations and source geometries.

As in Fig 13 but using a 2-second Hann window.

https://doi.org/10.1371/journal.pone.0297754.s022

(TIF)

S22 Fig. Elbow plots for other probe locations and source geometries.

As in Fig 13 but using a 0.5-second Hann window.

https://doi.org/10.1371/journal.pone.0297754.s023

(TIF)

S23 Fig. Elbow plots for other probe locations and source geometries.

As in Fig 13 but using a 1-second Kaiser-16 window.

https://doi.org/10.1371/journal.pone.0297754.s024

(TIF)

S24 Fig. Elbow plots for other probe locations and source geometries, for the supplementary dataset.

As in Fig 13 but for the Supplementary simulations.

https://doi.org/10.1371/journal.pone.0297754.s025

(TIF)

S25 Fig. Elbow plots for other probe locations and source geometries.

As in S24 Fig but using a 2-second Hann window.

https://doi.org/10.1371/journal.pone.0297754.s026

(TIF)

S26 Fig. Elbow plots for other probe locations and source geometries.

As in S24 Fig but using a 0.5-second Hann window.

https://doi.org/10.1371/journal.pone.0297754.s027

(TIF)

S27 Fig. Elbow plots for other probe locations and source geometries.

As in S24 Fig but using a 1-second Kaiser-16 window.

https://doi.org/10.1371/journal.pone.0297754.s028

(TIF)

Acknowledgments

We thank the members of the Crimaldi and Schaefer labs, the Neuroscience Interest Group at the Francis Crick Institute, the Latham lab at the Gatsby Computational Neuroscience Unit, and the members of Odor2Action IRG3 for their feedback. We also thank Jonathan Victor for his detailed comments on this work, in particular for pointing out the need to consider the influence of out-of-phase correlations, and for helping to clarify and simplify our model assumptions.

References

  1. 1. Wall C. and Perry J. N. Range of action of moth sex‐attractant sources. Entomologia Experimentalis et Applicata, 44(1):5–14, June 1987. ISSN 0013-8703, 1570-7458. URL https://onlinelibrary.wiley.com/doi/10.1111/j.1570-7458.1987.tb02232.x.
  2. 2. Murlis John, Elkinton Joseph S, Carde Ring T, et al. Odor plumes and how insects use them. Annual review of entomology, 37(1):505–532, 1992.
  3. 3. Ferrero David M., Lemon Jamie K., Fluegge Daniela, Pashkovski Stan L., Korzan Wayne J., Datta Sandeep Robert, et al. Detection and avoidance of a carnivore odor by prey. Proceedings of the National Academy of Sciences, 108(27):11235–11240, July 2011. ISSN 0027-8424, 1091-6490. URL https://pnas.org/doi/full/10.1073/pnas.1103317108. pmid:21690383
  4. 4. Howell Anna D. and Alarcón Ruben. Osmia bees (Hymenoptera: Megachilidae) can detect nectar-rewarding flowers using olfactory cues. Animal Behaviour, 74(2):199–205, August 2007. ISSN 00033472. URL https://linkinghub.elsevier.com/retrieve/pii/S0003347207001005.
  5. 5. Crimaldi JP and Koseff JR. High-resolution measurements of the spatial and temporal scalar structure of a turbulent plume. Experiments in Fluids, 31(1):90–102, 2001.
  6. 6. Baker Keeley L., Dickinson Michael, Findley Teresa M., Gire David H., Louis Matthieu, Suver Marie P., et al. Algorithms for Olfactory Search across Species. The Journal of Neuroscience, 38(44):9383–9389, October 2018. ISSN 0270-6474, 1529-2401. URL https://www.jneurosci.org/lookup/doi/10.1523/JNEUROSCI.1668-18.2018. pmid:30381430
  7. 7. David C. T., Kennedy J. S., and Ludlow A. R. Finding of a sex pheromone source by gypsy moths released in the field. Nature, 303(5920):804–806, June 1983. ISSN 0028-0836, 1476-4687. URL https://www.nature.com/articles/303804a0.
  8. 8. Álvarez-Salvado Efrén, Licata Angela M, Connor Erin G, McHugh Margaret K, King Benjamin MN, Stavropoulos Nicholas, et al. Elementary sensory-motor transformations underlying olfactory navigation in walking fruit-flies. Elife, 7:e37815, 2018. pmid:30129438
  9. 9. Vergassola Massimo, Villermaux Emmanuel, and Shraiman Boris I. ‘Infotaxis’ as a strategy for searching without gradients. Nature, 445(7126):406–409, January 2007. ISSN 0028-0836, 1476-4687. URL https://www.nature.com/articles/nature05464. pmid:17251974
  10. 10. Boie Sebastian D, Connor Erin G, McHugh Margaret, Nagel Katherine I, Ermentrout G Bard, Crimaldi John P, et al. Information-theoretic analysis of realistic odor plumes: What cues are useful for determining location? PLoS computational biology, 14(7):e1006275, 2018. pmid:29990365
  11. 11. Jayaram Viraaj, Kadakia Nirag, and Emonet Thierry. Sensing complementary temporal features of odor signals enhances navigation of diverse turbulent plumes. Elife, 11:e72415, 2022. pmid:35072625
  12. 12. Gilad Tomer, Bahar Ori, Hasan Malak, Bar Adi, Subach Aziz, and Scharf Inon. The combined role of visual and olfactory cues in foraging by Cataglyphis ants in laboratory mazes. Current Zoology, 69(4):401–408, August 2023. ISSN 1674-5507, 2396-9814. URL https://academic.oup.com/cz/article/69/4/401/6655976. pmid:37614920
  13. 13. Durve Mihir, Piro Lorenzo, Cencini Massimo, Biferale Luca, and Celani Antonio. Collective olfactory search in a turbulent environment. Physical Review E, 102(1):012402, July 2020. ISSN 2470-0045, 2470-0053. URL https://link.aps.org/doi/10.1103/PhysRevE.102.012402. pmid:32794953
  14. 14. Kadakia Nirag, Demir Mahmut, Michaelis Brenden T, DeAngelis Brian D, Reidenbach Matthew A, Clark Damon A, and Emonet Thierry. Odour motion sensing enhances navigation of complex plumes. Nature, 611(7937):754–761, 2022. pmid:36352224
  15. 15. Reddy Gautam, Murthy Venkatesh N., and Vergassola Massimo. Olfactory Sensing and Navigation in Turbulent Environments. Annual Review of Condensed Matter Physics, 13(1):191–213, March 2022. ISSN 1947-5454, 1947-5462. URL https://www.annualreviews.org/doi/10.1146/annurev-conmatphys-031720-032754.
  16. 16. Celani Antonio, Villermaux Emmanuel, and Vergassola Massimo. Odor landscapes in turbulent environments. Physical Review X, 4(4):041015, 2014.
  17. 17. Hopfield J. J. Olfactory computation and object perception. Proceedings of the National Academy of Sciences, 88(15):6462–6466, January 1991. ISSN 0027-8424, 1091-6490. URL http://www.pnas.org/content/88/15/6462. pmid:1862075
  18. 18. Szyszka Paul, Gerkin Richard C., Galizia C. Giovanni, and Smith Brian H. High-speed odor transduction and pulse tracking by insect olfactory receptor neurons. Proceedings of the National Academy of Sciences, 111(47):16925–16930, November 2014. ISSN 0027-8424, 1091-6490. URL https://pnas.org/doi/full/10.1073/pnas.1412051111. pmid:25385618
  19. 19. Fadamiro H. Y., Cossé A. A., and Baker T. C. Fine-scale resolution of closely spaced pheromone and antagonist filaments by flying male Helicoverpa zea. Journal of Comparative Physiology A: Sensory, Neural, and Behavioral Physiology, 185(2):131–141, August 1999. ISSN 0340-7594, 1432-1351. URL http://link.springer.com/10.1007/s003590050372.
  20. 20. Ackels Tobias, Erskine Andrew, Dasgupta Debanjan, Marin Alina Cristina, Warner Tom P. A., Tootoonian Sina, Fukunaga Izumi, Harris Julia J., and Schaefer Andreas T. Fast odour dynamics are encoded in the olfactory system and guide behaviour. Nature, 593(7860):558–563, May 2021. ISSN 1476-4687. URL https://www.nature.com/articles/s41586-021-03514-2.
  21. 21. Ottino Julio M. Mixing, chaotic advection, and turbulence. Annual Review of Fluid Mechanics, 22(1):207–254, 1990.
  22. 22. Roberts Philip JW and Webster Donald R. Turbulent diffusion, 2002.
  23. 23. Villermaux Emmanuel. Mixing versus stirring. Annual Review of Fluid Mechanics, 51(1):245–273, 2019.
  24. 24. Nironi Chiara, Salizzoni Pietro, Marro Massimo, Mejean Patrick, Grosjean Nathalie, and Soulhac Lionel. Dispersion of a passive scalar fluctuating plume in a turbulent boundary layer. part i: Velocity and concentration measurements. Boundary-layer meteorology, 156:415–446, 2015.
  25. 25. Young David L, Larsson Ann I, and Webster Donald R. Structure and mixing of a meandering turbulent chemical plume: concentration and velocity fields. Experiments in Fluids, 62(12):1–13, 2021.
  26. 26. Crimaldi JP and Koseff JR. Structure of turbulent plumes from a momentumless source in a smooth bed. Environmental Fluid Mechanics, 6(6):573–592, 2006.
  27. 27. Crimaldi John P, Wiley Megan B, and Koseff Jeffrey R. The relationship between mean and instantaneous structure in turbulent passive scalar plumes. Journal of Turbulence, 3(1):014, 2002.
  28. 28. Oskouie Shahin N, Wang Bing-Chen, and Yee Eugene. Numerical study of dual-plume interference in a turbulent boundary layer. Boundary-Layer Meteorology, 164:419–447, 2017.
  29. 29. Crimaldi John, Lei Hong, Schaefer Andreas, Schmuker Michael, Smith Brian H, True Aaron C, Verhagen Justus V, and Victor Jonathan D. Active sensing in a dynamic olfactory world. Journal of computational neuroscience, 50(1):1–6, 2022. pmid:34591220
  30. 30. Connor Erin G, McHugh Margaret K, and Crimaldi John P. Quantification of airborne odor plumes using planar laser-induced fluorescence. Experiments in Fluids, 59(9):1–11, 2018.
  31. 31. Colter C Ritsch. Experimental quantification of spatiotemporal structure of scalar plumes using planar laser-induced fluorescence. Master’s thesis, University of Colorado at Boulder, 2019.
  32. 32. Schmuker Michael, Bahr Viktor, and Huerta Ramón. Exploiting plume structure to decode gas source distance using metal-oxide gas sensors. Sensors and Actuators B: Chemical, 235:636–646, 2016.
  33. 33. Gifford Frank Jr. Statistical properties of a fluctuating plume dispersion model. In Advances in geophysics, volume 6, pages 117–137. Elsevier, 1959.
  34. 34. Yee Eugene, Kosteniuk PR, Chandler GM, Biltoft CA, and Bowers JF. Statistical characteristics of concentration fluctuations in dispersing plumes in the atmospheric surface layer. Boundary-Layer Meteorology, 65(1):69–109, 1993a.
  35. 35. Soltys MA and Crimaldi JP. Joint probabilities and mixing of isolated scalars emitted from parallel jets. Journal of Fluid Mechanics, 769:130–153, 2015.
  36. 36. Davies BM, Jones CD, Manning AJ, and Thomson DJ. Some field experiments on the interaction of plumes from two sources. Quarterly Journal of the Royal Meteorological Society, 126(565):1343–1366, 2000.
  37. 37. Shoaei F and Crimaldi JP. Effect of instantaneous stirring process on mixing between initially distant scalars in turbulent obstacle wakes. Experiments in Fluids, 58(4):1–10, 2017.
  38. 38. Warhaft Z. The interference of thermal fields from line sources in grid turbulence. Journal of Fluid Mechanics, 144:363–387, 1984.
  39. 39. Costa-Patry E and Mydlarski L. Mixing of two thermal fields emitted from line sources in turbulent channel flow. Journal of Fluid Mechanics, 609:349–375, 2008.
  40. 40. Sawford BL, Frost CC, and Allan TC. Atmospheric boundary-layer measurements of concentration statistics from isolated and multiple sources. Boundary-layer meteorology, 31(3):249–268, 1985.
  41. 41. Oskouie Shahin N, Wang Bing-Chen, and Yee Eugene. Study of the interference of plumes released from two near-ground point sources in an open channel. International Journal of Heat and Fluid Flow, 55:9–25, 2015.
  42. 42. Yee Eugene. The shape of the probability density function of short-term concentration fluctuations of plumes in the atmospheric boundary layer. Boundary-layer meteorology, 51:269–298, 1990.
  43. 43. Yee Eugene, Wilson DJ, and Zelt BW. Probability distributions of concentration fluctuations of a weakly diffusive passive plume in a turbulent boundary layer. Boundary-Layer Meteorology, 64(4):321–354, 1993b.
  44. 44. Yee Eugene and Chan R. A simple model for the probability density function of concentration fluctuations in atmospheric plumes. Atmospheric Environment, 31(7):991–1002, 1997.
  45. 45. Webster DR, Rahman S, and Dasi LP. Laser-induced fluorescence measurements of a turbulent plume. Journal of Engineering Mechanics, 129(10):1130–1137, 2003.
  46. 46. Yee Eugene, Gailis Ralph M, and Wilson David J. The interference of higher-order statistics of the concentration field produced by two point sources according to a generalized fluctuating plume model. Boundary-layer meteorology, 106:297–348, 2003.
  47. 47. Kaplan H and Dinar N. The interference of two passive scalars in a homogeneous isotropic turbulent field. Journal of Fluid Mechanics, 203:273–287, 1989.
  48. 48. Oppenheim Alan V., Schafer Ronald W., and Buck John R. Discrete-Time Signal Processing. Prentice-hall Englewood Cliffs, second edition, 1999.
  49. 49. Dayan Peter and Abbott Larry F. Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems. MIT Press, Cambridge, Mass., new ed edition edition, September 2005. ISBN 978-0-262-54185-5.
  50. 50. Firsch U. Turbulence. Cambridge University Press, 1995.
  51. 51. Koller D. and Friedman N. Probabilistic Graphical Models: Principles and Techniques. MIT Press, 2009.
  52. 52. Kree Mihkel, Duplat Jérôme, and Villermaux Emmanuel. The mixing of distant sources. Physics of Fluids, 25(9):091103, September 2013. ISSN 1070-6631. URL https://aip.scitation.org/doi/10.1063/1.4820015.
  53. 53. Nik Dennler, Damien Drix, Shavika Rastogi, André van Schaik, and Michael Schmuker. Rapid inference of geographical location with an event-based electronic nose. In Proceedings of the 2022 Annual Neuro-Inspired Computational Elements Conference, pages 112–114, 2022.
  54. 54. Boffetta Guido, Ecke Robert E, et al. Two-dimensional turbulence. Annual review of fluid mechanics, 44(1):427–451, 2012.
  55. 55. Haller George and Yuan Guocheng. Lagrangian coherent structures and mixing in two-dimensional turbulence. Physica D: Nonlinear Phenomena, 147(3-4):352–370, 2000.
  56. 56. Pratt Kenneth R, Meiss James D, and Crimaldi John P. Reaction enhancement of initially distant scalars by lagrangian coherent structures. Physics of Fluids, 27(3):035106, 2015.
  57. 57. Heinonen R. A., Biferale L., Celani A., and Vergassola M. Optimal policies for Bayesian olfactory search in turbulent flows. Physical Review E, 107(5):055105, May 2023. ISSN 2470-0045, 2470-0053. URL https://link.aps.org/doi/10.1103/PhysRevE.107.055105. pmid:37329026
  58. 58. Piro Lorenzo, Golestanian Ramin, and Mahault Benoît. Efficiency of navigation strategies for active particles in rugged landscapes. Frontiers in Physics, 10:1034267, November 2022a. ISSN 2296-424X. URL https://www.frontiersin.org/articles/10.3389/fphy.2022.1034267/full.
  59. 59. Jiang Jialin, Yang Zhengxin, Ferreira Antoine, and Zhang Li. Control and Autonomy of Microrobots: Recent Progress and Perspective. Advanced Intelligent Systems, 4(5):2100279, May 2022. ISSN 2640-4567, 2640-4567. URL https://onlinelibrary.wiley.com/doi/10.1002/aisy.202100279.
  60. 60. Piro Lorenzo, Mahault Benoît, and Golestanian Ramin. Optimal navigation of microswimmers in complex and noisy environments. New Journal of Physics, 24(9):093037, September 2022b. ISSN 1367-2630. URL https://iopscience.iop.org/article/10.1088/1367-2630/ac9079.
  61. 61. Brayton Robert K, Gustavson Fred G, and Hachtel Gary D. A new efficient algorithm for solving differential-algebraic systems using implicit backward differentiation formulas. Proceedings of the IEEE, 60(1):98–108, 1972.
  62. 62. Hindmarsh Alan C, Brown Peter N, Grant Keith E, Lee Steven L, Serban Radu, Shumaker Dan E, and Woodward Carol S. Sundials: Suite of nonlinear and differential/algebraic equation solvers. ACM Transactions on Mathematical Software (TOMS), 31(3):363–396, 2005.
  63. 63. Deuflhard Peter. A modified newton method for the solution of ill-conditioned systems of nonlinear equations with application to multiple shooting. Numerische Mathematik, 22(4):289–315, 1974.
  64. 64. Schenk Olaf, Gärtner Klaus, Fichtner Wolfgang, and Stricker Andreas. Pardiso: a high-performance serial and parallel sparse linear solver in semiconductor device simulation. Future Generation Computer Systems, 18(1):69–78, 2001.
  65. 65. Schenk Olaf and Gärtner Klaus. Solving unsymmetric sparse systems of linear equations with pardiso. Future Generation Computer Systems, 20(3):475–487, 2004.
  66. 66. SciPy 1.0: fundamental algorithms for scientific computing in Python | Nature Methods. URL https://www.nature.com/articles/s41592-019-0686-2.
  67. 67. Bishop Christopher M. Pattern Recognition and Machine Learning. Springer, 2006.
  68. 68. Davidson P.A. Turbulence. Oxford University Press, second edition, 2015.