## Figures

## Abstract

We present novel results that relate energy and information transfer with sensitivity to initial conditions in chaotic multi-dimensional Hamiltonian systems. We show the relation among Kolmogorov-Sinai entropy, Lyapunov exponents, and upper bounds for the Mutual Information Rate calculated in the Hamiltonian phase space and on bi-dimensional subspaces. Our main result is that the net amount of transfer from kinetic to potential energy per unit of time is a power-law of the upper bound for the Mutual Information Rate between kinetic and potential energies, and also a power-law of the Kolmogorov-Sinai entropy. Therefore, transfer of energy is related with both transfer and production of information. However, the power-law nature of this relation means that a small increment of energy transferred leads to a relatively much larger increase of the information exchanged. Then, we propose an “experimental” implementation of a 1-dimensional communication channel based on a Hamiltonian system, and calculate the actual rate with which information is exchanged between the first and last particle of the channel. Finally, a relation between our results and important quantities of thermodynamics is presented.

**Citation: **Antonopoulos CG, Bianco-Martinez E, Baptista MS (2014) Production and Transfer of Energy and Information in Hamiltonian Systems. PLoS ONE 9(2):
e89585.
https://doi.org/10.1371/journal.pone.0089585

**Editor: **Daniele Marinazzo, Universiteit Gent, Belgium

**Received: **October 3, 2013; **Accepted: **January 23, 2014; **Published: ** February 28, 2014

**Copyright: ** © 2014 Antonopoulos et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

**Funding: **The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. The authors were supported by the “EPSRC EP/I032606/1” grant.

**Competing interests: ** The authors have declared that no competing interests exist.

## Introduction

After the pioneering work by Shannon [1] on information, it became clear that it is a very useful and important concept as it can measure the amount of uncertainty an observer has about a random event and thus provides a measure of how unpredictable it is. The degree of disorder of a chaotic dynamical system is related to the degree of its chaotic behavior which is, in turn, characterized by the rate of exponential divergence of neighboring initial conditions, that is by the magnitude of the positive Lyapunov exponents [2]. It is the sensitive dependence on initial conditions [2]–[4] that produces information since two different but indistinguishable initial conditions at a certain precision will evolve into distinguishable states after a finite time [3]. This relation between production of information and sensitive dependence was made clear for systems that have absolutely continuous conditional measures [5], [6], by:(1)where represents the Kolmogorov-Sinai or KS entropy (Shannon's entropy per unit of time) and are the positive Lyapunov exponents of the dynamical system [2], [3], [7], [8], which measure how sensitive to initial conditions the system is. This is a property that has been found to be true for many dynamical systems [3]. In general, for bounded systems [9].

Energy and information can be produced in a system or transferred between its different “parts” or “constituents” [10]–[12]. If transferred, there are always at least two “entities” involved. In general, they can be nodes, modes, or related functions that can be defined on subspaces or projections of the phase space of the system.

Another related concept to the Shannon entropy that can characterize random complex systems is the Mutual Information (MI) [1] which is a measure of how much uncertainty one has about a state variable after observing another state variable. For deterministic systems that present correlations, a more appropriate quantity for measuring the transfer of information is the Mutual Information Rate (MIR), MI per unit of time. In Refs. [10], [13]–[15], the authors have developed alternative methods to overcome problems that stem from the definition of probabilities for these quantities and proposed the use of bounds for the MIR. In Ref. [10], the authors have derived an upper bound for the MIR between two nodes or two groups of nodes that depend on the largest Lyapunov exponents of the subspace of the network formed by the nodes. In particular, they have showed that:(2) where and are the two finite time and size Lyapunov exponents calculated in the bi-dimensional observation space which for simplicity will be referred herein as the Lyapunov exponents of the bi-dimensional subspace. In our study, when the observation space is formed by the kinetic () and potential () energy variables of the Hamiltonian, then the upper bound for the MIR in the space (i.e. and ) represents the upper bound for the information transferred per unit of time between the kinetic and potential energies. The use of the space to study the relationship between energy and information exchange is justifiable because the transfer of energy from kinetic to potential energy is easy and well understood. However, we will also study this relationship in other bi-dimensional subspaces such as those formed by any two nodes of the Hamiltonian system.

The main result of our work is Eq. (26), which states that when considering specific energy subintervals, the time rate of energy transferred from the kinetic to the potential variable during a time step is a power-law function of either the largest Lyapunov exponent of the Hamiltonian or of the upper bound for the MIR of the bi-dimensional space.

We then present the generalization of these power-law relations when considering much larger energy intervals of chaotic behavior with initial conditions set initially in different parts of the phase space of the same Hamiltonian system. We also consider different Hamiltonian systems in which we illustrate how they can be used to create communication systems.

The second main result is Eq. (1) of Information S1 which states that the upper bound for the MIR exchanged between the potential and kinetic energy is smaller than the upper bound for the MIR between two groups of oscillators formed each by half the oscillators of the Hamiltonian, and this is in turn smaller than the whole time rate of information produced by the system expressed by . We provide a proof of this result in Information S1. This result implies that, when one observes a Hamiltonian system through its kinetic and potential energy (i.e. in its space), one should not expect to have more information about the Hamiltonian system than when observing it directly (i.e. by observing half of its nodes or all of its variables).

The relation among energy, entropy, and information is a long lasting problem in physics. Nineteenth century saw the discovery of the two laws of thermodynamics, almost happening at the same time. The first law relates the rate of change of the energy of a body with the heat and work produced and the second, the rate of the change of the entropy of the body with the heating, implying the growth of its entropy during an adiabatic and irreversible process. Thermodynamics turned out to be a very important mathematical theory that can describe successfully macroscopic systems in equilibrium, based on the thermodynamic laws and provides a link between work, energy, and entropy as a universal competition, i.e. when a body approaches equilibrium, energy tends to a minimum and entropy to a maximum (see for example Ref. [16]).

In 1929, after a long lasting controversy, Leó Szilárd [17] and more recently the authors in Refs. [18], [19], showed that Maxwell's hypothetical demon does not contradict the second law of thermodynamics, implying that in principle one can convert information to free energy. By free energy we mean the portion of the energy of a system that is available to perform work mediated by thermal energy. It was only very recently in 2010 [20], that an experimental demonstration of this information to energy conversion has been achieved.

In Ref. [11], the authors study the energy transfer in terms of the classical dynamics of two particles that move in harmonic potential wells, interacting with the same external environment of noninteracting chaotic systems. They found that the oscillators can exchange energy through the environment when in almost-perfect resonance and in Ref. [12], a simple and solvable model of a device that transfer energy from a cold to a hot system by rectifying thermal fluctuations is presented. In order for this to happen, the device requires a memory register to which it can write information. The subtle issue of the connection between work and information processing is presented in Ref. [21] in a solvable model of an autonomous Maxwell's demon. The authors studied and explained a device that makes measurements about the system states, stores this information into a register, and delivers work by rectifying thermal fluctuations.

In this work however, we are interested in providing the relation between energy transfer and information production and transfer in multi-dimensional chaotic Hamiltonian systems, e.g. in isolated systems where the total energy of the system remains constant and no exchange of heat or matter with the surroundings exists. Such a relation could allow one to realize how much information a sort of Maxwell's demon would need in order to be able to transfer a certain amount of energy between oscillatory modes in Hamiltonian systems. Hamiltonian systems such as those we study herein differ from thermodynamic systems in the sense they are far from the thermodynamic limit, i.e. they have a small dimensionality. However, in the Discussion session, we provide a link between our results and important quantities of thermodynamics.

The paper is organized as follows: In Sec. Materials and Methods we present the basic material needed in our study. This includes the presentation of the two Hamiltonian systems and some of its important properties, the definition of the bi-dimensional observation space and a brief discussion about important quantities from the theory of information such as upper bound for MIR and KS entropy. In Subsec. Relation Between Largest Lyapunov Exponent of the Bi-dimensional Space and of the Hamiltonian, we present the relation between the largest Lyapunov exponent of the Hamiltonian system and that of the bi-dimensional space of the kinetic and potential energy. Then, in Subsec. Relation Between Production and Transfer of Information in the Small Energy Regime, we explain how one can arrive at Eq. (26) about the relation between production and transfer of information when considering specific energy subintervals of chaotic behavior. In Subsec. Generalization of our Study, we generalize the main results of our study by considering the case of different Hamiltonian systems for much larger energy intervals and with initial conditions set in different parts of the phase space of the systems. Then, in Sec. Hamiltonian Communication System, we illustrate how one can implement a 1-dimensional communication channel based on a Hamiltonian system, and calculate the actual rate with which information is exchanged between the first and last particle of the channel. In the Discussion section we briefly recall the main results of our study, their implications and relation with quantities of thermodynamics. Finally, in Information S1 we provide a proof of the inequality presented in Eq. (1) in there.

## Materials and Methods

### Fermi-Pasta-Ulam Hamiltonian

In this work we use two different Hamiltonian systems. We first consider the 1-dimensional lattice of particles with equal masses and nearest neighbour interactions with quartic nonlinearities (-model) given by the Fermi-Pasta-Ulam (FPU) Hamiltonian [22]:(3)adopting fixed boundary conditions:

Here, and is the position and conjugate momentum of the th particle, respectively.

For this system, we use initial conditions in the neighborhood of two particular simple periodic orbits of (3) which are called SPO1 and SPO2 [23], [24]. The reason for this choice is that they allow us to control in a systematic way the increase of the energy of the system so that chaotic motion will be sustained. Any other way of increasing the energy of the system so that chaotic behavior can exist may be equally used as well.

SPO1 is specified by the conditions:(4)and exists for all odd , keeping every even particle stationary at all times. It is not difficult to show that this is, in fact, the mode of the linear lattice (i.e. ) with frequency . The remarkable property of this solution is that it is continued in precisely the same spatial configuration in the nonlinear lattice as well, due to the form of the equations of motion associated with Hamiltonian (3):(5)which reduce, upon using (4) with the fixed boundary conditions to a single second order nonlinear differential equation for :(6)describing the oscillations of all moving particles of SPO1, with . For the stationary particles (i.e. ) we have . The solution of (6) is well known in terms of Jacobi elliptic functions [25] and can be written as:(7)where:(8)and is the modulus of the elliptic function. The energy per particle of SPO1 is then found to be:(9)by substituting simply the solution of Eq. (7) in Hamiltonian (3).

SPO2 is defined in a similar way. In particular, it exists for and corresponds to the case where every third particle is fixed, while the two in between move in opposite directions (in an out of phase fashion). Following similar arguments as for the SPO1 mode, the energy per particle of SPO2 is given by [23]:

We treat as a control parameter for the chaoticity of the FPU system (3). From now on, we drop the time-dependence notation of all involved variables for simplicity but use it wherever is needed.

### Bose-Einstein Condensate Hamiltonian

The second Hamiltonian system we use in this paper is the Bose-Einstein Condensate (BEC) model [24] which is given by:(10)where , is the position and conjugate momentum of the th particle (i.e. boson), respectively.

It possesses the second integral of motion:(11)and therefore chaotic behavior can only occur for .

We impose periodic boundary conditions in Eq. (10):(12)and use, for the same reason as in the FPU case, initial conditions set in the neighborhood of the out-of-phase mode (OPM):(13)with being even.

### Observation Subspaces and Quantities Calculated on Them

The FPU system (3) can be simply written in the form:(14)where:(15)

However, the BEC system (10) is not written in the same form and this will allow us to generalize the results of our study in the case where the space is not implied directly by the Hamiltonian form.

In our analysis, we define and study quantities like Lyapunov exponents initially in the bi-dimensional space, since is a meaningful physical quantity. Potential energy can be easily measured as well or estimated since . However, we also consider the observation space which is constructed by the position coordinates of the first and last particle of the Hamiltonian. For the FPU case, we know that:(16)where:(17)

Equation (16) is valid since the FPU Hamiltonian (3) is a global integral of the motion and thus a conserved quantity during time evolution.

Along the lines of ideas presented in Ref. [10], we compute the upper bound for the between any two groups of nodes each. The upper bound for the is defined as (see supplementary material in Ref. [10]):(18)where is half the number of positive Lyapunov exponents measured in the subspace. Naturally, . However, for the simulations we have performed we have set . So, represent the greater than or equal to zero Lyapunov exponents of the -dimensional projection constructed using scalar time series , for , which can be calculated in many ways, for example by calculating the finite size and finite time Lyapunov exponents or expansion rates [10]. represents the sum of all greater than or equal to zero Lyapunov exponents of the projection (i.e. an approximation for the KS entropy) and . Herein, we estimate them by computing the Lyapunov exponents of the Hamiltonian following [7], [8] and by keeping only those that are positive.

We also need to compute the upper bound for the MIR in the bi-dimensional space representing the maximum information exchanged between the kinetic () and potential () energies. Using the ideas from Ref. [10], is given by:(19)where and are the two positive Lyapunov exponents of the space with . In the case where , we have and thus it turns out that (see Ref. [10]).

In a series of papers [24], [26]–[31], the authors report for dynamical systems ranging from different kinds of billiards to multi-dimensional Hamiltonian systems, that the largest Lyapunov exponent of the system scales with the energy with a power-law of the form:(20)where is a real positive constant. This power-law dependence is valid for a rather large energy interval that can support chaotic behavior.

To numerically calculate we use:from which we can define the time average of the absolute value of the transfer of kinetic energy per unit of time through:(21)where denotes the time average over the integration of the trajectory up to . is the absolute value of the argument and we use it because we want to relate the quantities of Eq. (21) to positive average quantities, such as the positive Lyapunov exponents. Accordingly, is the amount of kinetic energy being transferred between and during a time step.

Since the BEC Hamiltonian (10) is not of the form as the FPU system, we reside on the calculation of a similar quantity based on the kinetic energy of any of its particles, for example of the first particle :(22)where is the kinetic energy of the first particle. Equation (22) is similar to the quantity of the left hand side of Eq. (25).

### Set of Initial Conditions

We prepare the two systems in a systematic way to reside in a chaotic regime and be able to produce information. For example, for the SPO2 we follow Ref. [23] and consider and varying the energy and initial condition appropriately as following: For each fixed energy of Hamiltonian (3), an initial condition is chosen (where and ) so that it lies in the neighborhood of SPO2. By neighborhood we mean that we perturb the equations of motion by a controllable small perturbation (i.e. ) so that the perturbed initial condition will be at the same constant energy of SPO2. Easily, we can fullfil this requirement by solving Eq. (3) for and then substitute it in the initial condition. A demonstration of the importance of this can be found in Sec. Results, where we present the relation between the largest Lyapunov exponent of the space and of the FPU Hamiltonian.

We thus end up with 14 nodes, each interacting with its nearest neighbours in a 1-dimensional lattice with fixed ends. In our example, SPO2 is destabilized at the energy and restabilized again at [23]. Thus, as increases in , SPO2 is unstable and gives rise initially to weakly and then to strongly chaotic behaviour in its neighborhood. For each we numerically integrate the corresponding initial condition and compute the Lyapunov exponents following Refs. [2], [7], [8] until they show a clear tendency to converge to a value. We subsequently record their values at the final integration time . In our case, we have checked that this convergence happens at about . We denote them as arranged in descending order. In terms of the numerical integration, we try to satisfy the condition that the relative energy error is kept between and . We follow a similar approach for the initial conditions we set in the neighborhood of SPO1 mode of FPU and OPM mode of BEC so that we can guarantee chaotic behavior with the increase of the energy of the system.

## Results

### Relation Between Largest Lyapunov Exponent of the Bi-dimensional Space and of the Hamiltonian

The dynamics on the space is driven by the dynamics of the Hamiltonian system and we have no explicitly given equations of motion for the space. As we have already pointed out, we choose initial conditions on the same energy as the SPO2, and this implies that points and belong to the line . The motion takes place on this 1-dimensional subspace and thus, there is only one Lyapunov exponent that leads to .

In Fig. 1, one can see schematically the time evolution after one time step of a deviation vector (denoted as an arrow) along the direction of the Lyapunov exponent defined for the dynamics on the line . Here and are two trajectories in the phase space of Hamiltonian (3) on the same energy as SPO2, started initially in its neighborhood and being infinitesimally close. Then, is the rate of expansion of the deviation vector defined by the points , and ,. Here, and denote the lengths of the initial and after one time step deviation vectors respectively.

and are two trajectories in the phase space of Hamiltonian (3) that drive the dynamics along this line. We denote with and the lengths of the two deviation vectors initially and after one time step, respectively.

can be defined for infinitesimally close-by points on the 1-dimensional space of of Fig. 1 by keeping track of the evolution of their distance. In particular, for such points , and ,, their distance is given by:(23)

and combining it with Eq. (23) we obtain:(24)

We denote as the largest Lyapunov exponent in the neighborhood of SPO2, and reside on numerical simulations to show in Fig. 2 that is actually . In the example of Fig. 2 we have set , resulting in the relation . However, we have checked that the above result is valid for all energies we considered in . We observe that tends to zero in the course of time and that at some point starts to saturate at about due to round off numerical errors. In other words, we have showed that the largest Lyapunov exponent of the 1-dimensional space is equal to the largest Lyapunov exponent of Hamiltonian (3), i.e. .

Here, is well inside the interval . Note that both axes are logarithmic.

To achieve this result, we integrated simultaneously two infinitesimally close trajectories and (e.g. at an initial distance of the order of ) on the same energy as SPO2 and consider thus that , and replace the limits in Eq. (24) by a finite time , computing as a time average [8], i.e. as finite size and finite time Lyapunov exponent. Since for chaotic trajectories, the distance between and quickly saturates, we periodically renormalize their separation without altering their relative orientation in phase space and then compute the new distance setting . To avoid any numerical overflows, we preferred to do this at every time step.

Note that is not an unexpected result, since the largest Lyapunov exponent should be obtained in typical low-dimensional linear projections or embedding spaces [3], [32]. By typical here we mean bi-dimensional subspaces or projections that are not oriented along Lyapunov vectors. However, the space is a highly nonlinear projection still maintaining the largest positive Lyapunov exponent of the Hamiltonian as we have demonstrated. Every initial condition creates a trajectory with only one positive Lyapunov exponent in the subspace. Therefore, .

Concluding this part, we have demonstrated that the transfer of information from to is mediated by the largest Lyapunov exponent of the Hamiltonian. We finally obtain:

The last result implies that the upper bound for the between kinetic and potential energies is equal to the largest Lyapunov exponent of the Hamiltonian and consequently, can not be bigger than this exponent.

### Relation Between Production and Transfer of Information in the Small Energy Regime

To start with, we present in a log-log plot in Fig. 3 the quantities of Eq. (18) in red dashed line with points, of Eq. (1) in green dashed line with rectangles, in black solid line with lower triangles and of Eq. (21) in blue dashed line with upper triangles for the SPO2 case of the FPU system with parameters as defined in Subsec. Fermi-Pasta-Ulam Hamiltonian. Here is the time step of the integration (i.e. ). The time derivative of the kinetic energy accounts for the rate of transfer from kinetic to potential energy. We see that all quantities follow the same morphology (i.e. share the same functional form) as the energy of the initial condition is increased in the interval . Moreover, is an upper bound of the upper bound for the between two groups formed each by 7 nodes. We will prove a related inequality in Information S1.

Note that both axes are logarithmic.

The approach we shall follow to relate with for the transfer of information between and is meaningful as long as the motion in the Hamiltonian phase space is chaotic (e.g. as long as ). If the motion is periodic or quasi-periodic there is no exchange of information between the nodes (i.e. by knowing the position of a particular node one can predict the position and momenta of another one). Our results show that is related by a power-law to the largest Lyapunov exponent of the Hamiltonian and to the upper bound for the transfer of information between kinetic and potential energies. Surprisingly, we have found that this is valid for sufficiently large enough subintervals, i.e. for .

Here, we need to make use of only one neighboring initial condition of SPO2 and denote for simplicity by . With the help of Eq. (21) and we have found numerically that:(25)for the same energy interval that Eq. (20) applies where is a real positive constant. By substituting Eq. (20) in Eq. (25), we obtain:(26)where we have used (see Subsec. Relation Between Largest Lyapunov Exponent of the Bi-dimensional Space and of the Hamiltonian). It is straightforward to show that the same power-law (26) applies to due to Eqs. (14) and (16) respectively. We emphasize that is a time-ratio that depends on time, and that and are time invariant averages.

Fig. 4A shows in a log-log scale the quantity of Eq. (18) in red dashed line with points and of Eq. (1) in green dashed line with rectangles. In panel B, we plot with red points and the power-law fitting:(27)with green line. The agreement is remarkable. In Fig. 4C we plot of Eq. (25) and fit with the power-law:(28)showed as green line. We find that and that . In panel D of the same figure we plot with red points as a function of for values that correspond to the same energy interval of panels A, B and C. The power-law fitting:(29)plotted in green dashed line gives and which is in good agreement with the value of . The above arguments directly imply that:(30)where the proportionality constant and . To arrive at Eq. (30) we have used of Subsec. Relation Between Largest Lyapunov Exponent of the Bi-dimensional Space and of the Hamiltonian and Eq. (2) presented in Information S1. Equation (30) relates the production and transfer of information in the space with and . Therefore, the larger the transfer of energy is between the kinetic and potential energy, the larger is the upper bound for the MIR between the kinetic and potential energies and the larger the KS entropy of the system will be. In other words, exchange of information between and implies exchange of energy, and vice-versa. However, a relatively small increment of energy transfer produces a larger relative increase of the information transferred since .

Plot of quantities with red points with the power-law fitting of Eq. (27) in green line (panel B). Plot of with red points with the power-law fitting of Eq. (29) in green line (panel C). Power-law dependence of to in red points, in the interval that corresponds to the energy interval of panels A, B and C and of the power-law fitting of Eq. (29) in green dashed line (panel D). Note that all axes are logarithmic.

In Information S1 we prove another important result which is the inequality:(31)and thus justify the result presented in Fig. 3.

### Generalization of our Study

Here, we extend our study and present the generalization of our predicted upper bounds for the MIR and the connection with the transfer of energy of the previous section by considering higher energy intervals with initial conditions set in different parts of the phase space of two Hamiltonian systems: the FPU (3) and BEC (10).

We will show that if one considers a much larger energy interval for these systems with initial conditions set in different parts of their phase spaces, then Eqs. (27), (28) and (29) can be generalized, as:(32)(33)

We prefer to call Eqs. (32) and (33) as generalized power-law functions. Here, stands for the bi-dimensional space of observation. In the case of the FPU system (3) we consider as a bi-dimensional space the space while for the BEC system (10) we consider the observation space constructed by observing the pair of variables and , that is by the position of the first and last particle. In Sec. Hamiltonian Communication System, where we study an “experimental” setup of a 1-dimensional communication channel based on the FPU system, we will use this particular observation space as well.

By eliminating from both Eqs. (32) and (33), one arrives at the relation between transfer of energy per unit of time (i.e. ) and upper bound of information transmitted in the bi-dimensional space (i.e. ):(34)

Parameters can be determined by performing a non-linear fitting of the numerical data by the functions (32), (33) and (34). We have used Matlab to perform these fittings.

### FPU SPO2

In the case of the SPO2 studied in Subsec. Relation Between Production and Transfer of Information in the Small Energy Regime, the fit of Fig. 4 was performed in the energy interval . Here we generalize Eqs. (27), (28) and (29) in the larger energy interval for which the dynamics around SPO2 is chaotic as indicated by the Lyapunov exponents. This allows the creation and transfer of information and energy. We have used the same parameters and setup (e.g. 14 particles) to allow for a direct comparison between Figs. 4 and 5.

Note that both axes are logarithmic. Plot of with red points with the power-law fitting of Eq. (32) in green line (panel B). Plot of with red points with the power-law fitting of Eq. (33) in green line (panel C). Power-law dependence of to in red points, in the interval that corresponds to the energy interval of panels A, B and C and of the power-law fitting of Eq. (33) in green dashed line (panel D). Note that all axes are logarithmic.

By doing a similar analysis as in Subsec. Relation Between Production and Transfer of Information in the Small Energy Regime, we present in Fig. 5 the plots of all relevant quantities for the larger energy interval. By fitting the new data with the generalized power-laws of Eqs. (32), (33) and (34) we have: , , for Eq. (32), , , , for Eq.(33) and finally: , , , , , and for Eq. (34).

### FPU SPO1

Here we extend our study to a another part of the phase space of the FPU Hamiltonian with initial conditions set in the neighborhood of the periodic orbit SPO1 (see Eq. (4) of Subsec. Fermi-Pasta-Ulam Hamiltonian). We have chosen this particular part of the phase space as SPO1 does not restabilize at some bigger energy as it happens with SPO2 and thus allows to reach as high energies as desired. We will show that the same generalized power-laws of Eqs. (32), (33) and (34) can still be used to fit the data of the upper bounds for MIR such as , and . In more details, for Eq. (32) we have: , , , for Eq.(33) we have: , , , and finally, for Eq. (34) we have: , , , , , and .

In Fig. 6 we present the corresponding plots and fits for the energy interval considering 15 particles and . Following Ref. [23], for these values we know that the dynamics around SPO1 is chaotic and thus allows the production and exchange of energy and information in the FPU chain.

Note that both axes are logarithmic. Plot of with red points with the power-law fitting of Eq. (32) in green line (panel B). Plot of with red points with the power-law fitting of Eq. (33) in green line (panel C). Power-law dependence of to in red points, in the interval that corresponds to the energy interval of panels A, B and C and of the power-law fitting of Eq. (33) in green dashed line (panel D). Note that all axes are logarithmic.

### BEC OPM

Next, we proceed and study the same problem for a different system, namely the BEC Hamiltonian given in Eq. (10). We have chosen this system as it allows us to study the relation between transfer and exchange of energy and information in a different Hamiltonian system than the FPU. Furthermore, because it is not written in the form as the FPU does (compare Eqs. (3) and (10)). It will thus permit us to demonstrate the validity of the upper bounds for the MIR and the connection between the exchange of energy and information in different observation spaces.

In particular, we consider here a small version of the system with degrees of freedom (particles) with initial conditions set in the neighborhood of the OPM periodic orbit given in Eq. (13) with periodic boundary conditions (see Eq. (12)). In Fig. 7, we show the results of a similar study as we did in the cases of SPO1, SPO2 of the FPU system, for the energy interval for which we have been able to study numerically in terms of the preservation of the accuracy of the computed energy. For this energy interval we know that the dynamics is chaotic (see Ref. [24]). Since, as we have already pointed out, BEC is not given by the sum of the kinetic and potential energy, we adopt a different strategy and reside on the calculation of the similar quantity based on the kinetic energy of the first particle (see Eq. (22)). However, the kinetic energy of any other particle can be used as well. By fitting the data with the generalized power-laws of Eqs. (32), (33) and (34) we have: , , for Eq. (32), , , , for Eq.(33) and finally: , , , , , and for Eq. (34).

Note that both axes are logarithmic. Plot of with red points with the power-law fitting of Eq. (32) in green line (panel B). Plot of with red points with the power-law fitting of Eq. (33) in green line (panel C). Power-law dependence of to in red points, in the interval that corresponds to the energy interval of panels A, B and C and of the power-law fitting of Eq. (33) in green dashed line (panel D). Note that all axes are logarithmic.

### Hamiltonian Communication System

In this section we present an “experimental” implementation of a 1-dimensional communication channel based on the FPU Hamiltonian system of Eq. (3), and show the relation between our proposed upper bounds for the MIR with the actual MIR measured for the exchange of information between the first and last particle of the channel.

In more details, we consider the FPU chain of oscillators as a 1-dimensional communication channel where information and energy flow from one end to the other, i.e. from the first particle to the last one and vice versa. To extend the applicability of our theoretical results obtained in the previous sections for different cases of chaotic dynamics, we will use the dynamics around SPO1 and SPO2 and consider as a bi-dimensional observation space the one constructed by the evolution of the pair of position variables of the first and last particle of the FPU chain of Eq. (3). The computation of the actual MIR value between the two observation nodes ad was based on the theory presented in Ref. [10]. Here, we consider 15 oscillators (degrees of freedom) for the SPO1 and 14 for the SPO2.

In panel A of Fig. 8 we show the results of our study for the SPO2 case. We have plotted in red dashed line with points the quantity of Eq. (18), as defined by Eq. (1) in green dashed line with rectangles, of Eq. (19) in black solid line with lower triangles and MIR_{1,14} in blue dashed line with upper triangles as a function of the energy . Here, MIR_{1,14} stands for the actual mutual information rate measured for the exchange of information between and . From our theoretical results derived in the previous sections we expect that MIR_{1,14} should be smaller or equal than . This is indeed what one observes as the MIR_{1,14} curve is smaller than the previously mentioned upper bound and more importantly, it follows the same morphology (functional form) as , and . We have performed the same analysis for the SPO1 case as well showed in panel B of the same figure and arrive again at the same conclusions, i.e. MIR_{1,15} lies below as expected by our study and follows the same morphology as the upper bounds , and . In this case, MIR_{1,15} denotes the actual mutual information rate measured for the exchange of information between and .

Same as in panel A for initial conditions set in the neighborhood of SPO1 (panel B). Note that all axes are logarithmic.

## Discussion

In this paper we have studied the relation among the transfer of energy from kinetic () to potential () energies, the transfer of information between these two quantities and between different particles, the production of information, and Lyapunov exponents in Hamiltonian systems.

Our first result is that the largest Lyapunov exponent of the bi-dimensional space of the kinetic and potential energy is equal to the maximal Lyapunov exponent of the Hamiltonian in the case it is given by the sum of the kinetic and potential energy. Consequently, we were able to show that the upper bound for the MIR in the subspace is given by the largest Lyapunov exponent of the Hamiltonian (). This implies that the more information the Hamiltonian system produces (), the more information can be exchanged between and .

The second important result we have found is a power-law relation between the rate of transfer from kinetic to potential energy, the largest Lyapunov exponent of the Hamiltonian, and the Kolmogorov-Sinai entropy of the Hamiltonian. The more chaotic and the more information the Hamiltonian system produces ( and ) respectively, the larger is the time average of the absolute value of energy transferred between and per unit of time (i.e. ).

The other important result is the proof of the inequality shown in Information S1. It implies that, when one observes a Hamiltonian system through its kinetic and potential energies (thus obtaining ), one measures less information about the Hamiltonian system than when observing half of its variables (thus obtaining ) or all of its variables (thus obtaining ).

Finally, we have proposed an “experimental” implementation of a 1-dimensional communication channel based on a Hamiltonian system, and have calculated the actual rate with which information is exchanged between the first and last particle of the channel and compared that with the upper bounds we have proposed. As expected from our theoretical analysis, in all cases we have studied the actual MIR values were found to be smaller than our proposed upper bounds of MIR.

It is challenging to sketch here a possible connection between our results and the free energy , entropy , temperature and Hamiltonian energy in the thermodynamic limit, i.e. when and grow indefinitely while their ratio remains constant. According to the definition attributed to Helmholtz, is equal to the internal energy of the system minus the product of the (absolute) temperature multiplied by , i.e. . is an important macroscopic quantity since its definition goes back to the early days of thermodynamics. Maxwell had realized that when the Hamiltonian has the special form:(as is the case of the FPU system we have studied in this work) the canonical ensemble average of is the temperature of the system. Thus, if one assumes ergodicity and equivalence of ensembles of initial conditions, it suffices to measure the time average of during the evolution of the system in order to compute (see for example Ref. [33]). Then, in this context is the fixed energy of the Hamiltonian (e.g. FPU) and can be calculated by the KS entropy as , where has the unit of time, since KS entropy is simply Shannon's entropy (equivalent to Gibb's entropy) per unit of time. Therefore, one can have:(35)and by solving Eq. (35) to obtain:(36)

If Eq. (30) remains still valid in the thermodynamic limit, then by substituting Eq. (36) in the right hand side of Eq. (30) one has:(37)which relates the rate of transfer from kinetic to potential energy and the largest Lyapunov exponent of the Hamiltonian with the free energy and temperature of the system. This provides a direct relation between the results of this paper and important quantities of thermodynamics and Statistical Mechanics as long as the same conditions required for the derivation of the main results of our paper hold for Eq. (37) as well. Equation (37) implies that the larger the gap between the energy of the Hamiltonian and the available energy to do work (the free energy) the smaller the transfer of energy and information from to is.

In a series of papers [20], [34]–[36], the authors discuss about technological applications of the transfer of energy and information in communication, interference and graphical networks and show how one can reuse part of the energy for successive communication tasks. These ideas are based on results from physics showing that any system that exchanges information via the transfer of given physical resources such as radio waves, particles, etc., can reuse part of the received resources. If chaotic Hamiltonian systems could be used to create a communication system such that energy of the transmitting signal could be reused to transmit more information, from Eq. (37) it is clear that must be different than zero implying that less information can be transmitted.

We believe that our work provides a viable pathway to establish similar relations between production and transfer of energy and information in other Hamiltonian systems for which the Lyapunov exponents have different dependences with the increase of the energy of the system as compared to those we have found here. Moreover, the choice of the bi-dimensional observation space is not restrictive and a plausible one can be constructed by the position coordinates of any two particles of the system. Of course, in these cases it is expected that our power-law relations will be replaced by new ones reflecting the different properties of the systems.

## Acknowledgments

Ch. A. would like to dedicate this paper to his recently born nephew, sharing both a common period of being “in preparation”.

## Author Contributions

Conceived and designed the experiments: CA MSB. Performed the experiments: CA EBM. Analyzed the data: CA EBM. Contributed reagents/materials/analysis tools: CA EBM. Wrote the paper: CA MSB.

## References

- 1. Shannon CE (1948) A mathematical theory of communication. The Bell System Technical Journal 27: 379.
- 2. Skokos C (2010) The Lyapunov characteristic exponents and their computation. Lecture Notes in Physics 790: 63–135.
- 3. Eckmann JP, Ruelle D (1985) Ergodic theory of chaos and strange attractors. Rev Mod Phys 57: 617–656.
- 4.
Bountis T, Skokos H (2012) Complex Hamiltonian Dynamics. Springer-Verlag Berlin Heidelberg.
- 5. Pesin YB (1976) Invariant manifold families which correspond to nonvanishing characteristic exponents. Izv Akad Nauk SSSR Ser Mat 40: 1332.
- 6. Pesin YB (1977) Lyapunov characteristic exponents and smooth ergodic theory. Uspekhi Matematicheskikh Nauk 32: 196.
- 7. Benettin G, Galgani L, Giorgilli A, Strelcyn JM (1980) Lyapunov characteristic exponents for smooth dynamical systems and for Hamiltonian systems: A method for computing all of them. Part 1: Theory. Meccanica 15: 9–20.
- 8. Benettin G, Galgani L, Giorgilli A, Strelcyn JM (1980) Lyapunov characteristic exponents for smooth dynamical systems and for Hamiltonian systems: A method for computing all of them. Part 2: Numerical application. Meccanica 15: 21–30.
- 9. Ruelle D (1978) An inequality for the entropy of differentiable maps. Bol Soc Bras Mat 9: 83.
- 10. Baptista MS, Rubinger RM, Viana ER, Sartorelli JC, Parlitz U, et al. (2012) Mutual information rate and bounds for it. PLoS One 7 10: e46745.
- 11. Machiori MA, Fariello R, de Aguiar MAM (2012) Energy transfer dynamics and thermalization of two oscillators interacting via chaos. Phys Rev E 85: 041119.
- 12. Mandal D, Quan HT, Jarzynski C (2013) Maxwell's refrigerator: An exactly solvable model. Phys Rev Lett 111: 030602.
- 13. Baptista MS, Kurths J (2005) Chaotic channel. Phys Rev E 72: 045202.
- 14. Baptista MS, Kurths J (2008) Transmission of information in active networks. Phys Rev E 77: 026205.
- 15. Baptista MS, de Carvalho JX, Hussein MS (2008) Finding quasi-optimal network topologies for information transmission in active networks. PLoS ONE 3: e3479.
- 16. Müller I (2008) Entropy and energy, - a universal competition. Entropy 10: 462–476.
- 17. Szilárd L (1929) On the decrease of entropy in a thermodynamic system by the intervention of intelligent beings. Z Phys 53: 840–856.
- 18. Landauer R (1961) Irreversibility and heat generation in the computing process. J Res Dev 5: 183–191.
- 19. Bennett CH (1982) The thermodynamics of computation-a review. Int J Theor Phys 102: 905–940.
- 20. Toyabe S, Sagawa T, Ueda M, Muneyuki E, Sano M (2010) Experimental demonstration of information-to-energy conversion and validation of the generalized Jarzynski equality. Nature Physics 6: 988–992.
- 21. Mandal D, Jarzynski C (2012) Work and information processing in a solvable model of Maxwell's demon. PNAS 109: 11641–11645.
- 22. Fermi E, Pasta J, Ulam S (1955) Studies of nonlinear problems. Los Alamos document LA-1940 27: 379.
- 23. Antonopoulos C, Bountis T (2006) Stability of simple periodic orbits and chaos in a Fermi-Pasta-Ulam lattice. Phys Rev E 73: 056206.
- 24.
Antonopoulos C, Bountis T, Skokos C (2006) Chaotic dynamics of
*N*-degree of freedom Hamiltonian systems. Int J Bif Chaos 16: 1777–1793. - 25.
Abramowitz M, Stegun I (1965) Handbook of mathematical functions. Dover, New York.
- 26. Rechester AB, Rosenbluth MN, White RB (1979) Calculation of the Kolmogorov entropy for motion along a stochastic magnetic field. Phys Rev Lett 42: 1247.
- 27. Pettini M, Landolfi M (1990) Relaxation properties and ergodicity breaking in nonlinear Hamiltonian dynamics. Phys Rev A 41: 768.
- 28. Pettini M, Cerruti-Sola M (1991) Strong stochasticity threshold in nonlinear large Hamiltonian systems: Effect on mixing times. Phys Rev A 44: 975.
- 29. Pettini M, Casetti L, Cerruti-Sola M, Franzosi R, Cohen EGD (2005) Weak and strong chaos in Fermi-Pasta-Ulam models and beyond. Chaos 15: 015106.
- 30. Benettin G (1984) Power-law behavior of Lyapunov exponents in some conservative dynamical systems. Physica D 13: 211–220.
- 31. Livi R, Politi A, Ruffo S (1986) Distribution of characteristic exponents in the thermodynamic limit. J Phys A: Math Gen 19: 2033–2040.
- 32. Wolf A, Swift JB, Swinney HL, Vastano JA (1985) Determining Lyapunov exponents from a time series. Physica D 16: 285.
- 33. Rugh HH (1997) Dynamical approach to temperature. Phys Rev Lett 78: 772–774.
- 34.
Popovski P, Fouladgar A, Simeone O (2013) Interactive joint transfer of energy and information. IEEE Transactions on Communications 61..
- 35.
Mohammad Fouladgar A, Simeone O (2013) Information and energy flow in graphical networks with energy transfer and reuse. IEEE Wireless Communications Letters PP.
- 36.
Park J, Clerckx B (2013) Joint wireless information and energy transfer in a two-user MIMO interference channel. CoRR abs/1303.1693.