Abstract
In a thermodynamic process with measurement and feedback, the second law of thermodynamics is no longer valid. In its place, various second-law-like inequalities have been advanced that each incorporate a distinct additional term accounting for the information gathered through measurement. We quantitatively compare a number of these information measures using an analytically tractable model for the feedback cooling of a Brownian particle. We find that the information measures form a hierarchy that reveals a web of interconnections. To untangle their relationships, we address the origins of the information, arguing that each information measure represents the minimum thermodynamic cost to acquire that information through a separate, distinct measurement procedure.
Export citation and abstract BibTeX RIS
Content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.
1. Introduction
The Kelvin–Planck statement of the second law of thermodynamics forbids the existence of a cyclically operating device whose sole effect is to convert heat from a single thermal reservoir into an equal amount of work [1]. However, we can circumvent this restriction, if our device operates via measurement and feedback: a possibility first envisioned by Szilard in his famous thought experiment [2]. Recently, there has been renewed interest in this old idea spurred by the development of a collection of distinct, second-law-like inequalities that quantify the interplay between the information gathered through measurement and the work that can be extracted in response through feedback. For continuously operating devices at temperature T, all these predictions bound the extracted work rate as
by some information acquisition rate, generically denoted here as , which differs in each second-law-like inequality, and is Boltzmannʼs constant. The first inequality of this form was derived by Sagawa and Ueda for a single feedback loop [3], but subsequently has been extended to include the repeated use of feedback, allowing for the application to continuously operating information engines [4–11]. Earlier results bounding entropy flows in feedback systems can be found in [12, 13]. In these cases, the information rate is identified as the rate of growth of the transfer entropy [14] from the system to the measurement device (or feedback controller) [8, 10, 15, 16]. An alternative inequality identifies the information rate with the flow of mutual information between the system and a continuously-interacting auxiliary measurement device. This information flow approach has been developed for small systems modeled as continuous diffusion processes [17], discrete Markov jump processes [16, 18], and for stochastic processes interacting discretly [19, 20]. Yet another version has been suggested by Kim and Qian specifically for the feedback cooling of a harmonically-trapped Brownian particle, where the extracted work is bounded by a term they call entropy pumping [21]. To date there is no clear information-theoretic interpretation of this term. Nevertheless, this result conforms to the second-law-like structure in (1). Further developments in this direction are the inclusion of measurement errors and delay [22–24]. At first glance, this plethora of seeming similar predictions is confusing and raises questions about the interpretation as well as the utility of these information bounds. To help clarify the situation, a number of studies have compared some of these measures from different points of view [16, 17, 25, 26]. Our goal in this paper is to build on these works by providing a comprehensive, pedagogical comparison of all these information measures within a single framework in order show clearly their relationships and limitations.
There are essentially two ways to view (1). The first is to treat (1) simply as a numerical bound on the extracted work without reference to the physical underpinnings of . This is the point of view we typically take when investigating feedback (or information) engines [25, 27–31], where our goal is to optimally extract the maximum amount of work; the maximum being any or all of the possible information measures. In this respect, having so many bounds is problematic, since we are unsure which is the most appropriate. Nevertheless, this is the approach we take in the first half our paper in section 3. There we investigate the quantitative relationship between the various information measures by analytically calculating them in a Brownian particle model of feedback cooling, which we introduce in section 2. We use this particular model, since it has been studied theoretically [21–23] and could be implemented experimentally in the setups of [32, 33]. The analytical tractability of this model further lets us examine these information measures from the point of view of optimal control theory [11, 54], which reveals intimate connections among them. The second way to interpret (1) is to take seriously its resemblance to the second law, and ask how far can we push this analogy? In particular, the traditional statement of the second law dictates that the entropy production of the Universe—system and surroundings—during a thermodynamic process must be positive [1]. In feedback-driven systems, the surroundings not only include the traditional thermodynamic reservoirs, such as heat baths or chemical baths, but in addition they include an auxiliary system that records the measurement and feeds back that information. In this case, does (1) still represent the entropy production of the system and its surroundings, except now the surroundings contain the feedback device? This is the question we address in the second half of our paper in section 4. There we observe that the transfer entropy rate and information flow have clear interpretations as the minimum entropy production required to acquire that information. However, each one is associated with a different physical measurement scenario, that is with a distinct surroundings in much the same way a particle reservoir differs from a thermal reservoir.
2. Feedback cooling model
Throughout, we will illustrate the different information concepts with a model for the feedback cooling of an underdamped Brownian particle [21–23]. This will allow us to discuss each measure using the same language. We therefore in this section introduce the dynamics of the model, both on the individual trajectory level and the ensemble level, as well as collect germane results regarding its energetics and thermodynamics.
2.1. Dynamics, energetics, and thermodynamics without feedback
Our quantity of interest is the time-dependent velocity vt of a trapped, underdamped Brownian particle of mass m, coupled to a thermal reservoir at temperature T with viscous damping coefficient , evolving according to the Langevin equation [34]
where ft is an externally controlled force, and is zero-mean Gaussian white noise with covariance . Starting here we set Boltzmannʼs constant to unity, .
In the absence of control, ft = 0, the velocity vt relaxes to an equilibrium Boltzmann distribution . In the following, we will vary ft using feedback in order to cool the particle, that is damp its thermal fluctuations, thereby reducing its kinetic temperature . Before we get to that, it is helpful to first review the energetics and thermodynamics of a driven, underdamped Brownian particle without feedback, so that we can appreciate the differences that arise in the presence of feedback.
To this end, we require the Fokker–Planck equation associated with (2) for the time-dependent probability density pt(v) [35],
where we have introduced the (probability) current Jvt. Anticipating our discussion of the thermodynamics, we divide the current into its irreversible half, which is anti-symmetric under time-reversal, and its reversible half, which is time-reversal symmetric, as [35–37]
Key to this splitting is treating the force ft as even under time reversal, as typically assumed for a force arising from an external potential. With this identification, the irreversible portion of the current arises solely due to the forces imparted on the particle by its surroundings: the friction, , and the fluctuating force, .
Moving on to the thermodynamics, we have from stochastic energetics an unambiguous identification of the heat flow into the system as the work done by the thermal reservoir on the particle [22, 36, 38, 39], which on average reads
It notably only depends on the irreversible current arising from the forces due to the thermal reservoir. The particleʼs (internal) energy is its average kinetic energy
By differentiating E with time and substituting in the Fokker–Planck equation (3), we are able to identify the extracted work rate via the first law of thermodynamics ,
as the average power delivered against the external force ft.
From stochastic thermodynamics, we also have the (irreversible) entropy production rate [36, 37, 39]
where we have the traditional splitting into the time variation of the systemʼs Shannon entropy ,
and the reversible entropy exchange with the environment
Notably, the entropy production only depends on the irreversible current, since it is a measure of the time-reversal symmetry breaking of the dynamics [36]. This property is what allowed us to pullout the contribution due to the heat, which is also only a function of the irreversible current.
2.2. Dynamics and energetics with feedback
Our main focus in this paper is feedback cooling, where we vary ft in response to measurements of the velocity. Following [23], we consider a feedback protocol where we measure the velocity vt obtaining outcomes yt with some error, and then feed back those measurements by applying a force that acts as an additional friction, extracting work. A simple way to incorporate measurement error is to add to our read-out of vt Gaussian white noise of zero mean and covariance , with quantifying the measurement uncertainty: for example as . However, white noise fluctuations are very violent. To make the problem more tractable, we smooth over the noise by applying a low-pass filter with time constant τ to the measurements: [40]. We are therefore led to the following modified dynamics including measurement and feedback [23]
where a is the feedback gain. It is important to note at this point that yt is merely a model of measurement outcomes. We are not making any assumption about the physical system that records the measurements, nor implements the feedback in response.
In general, the joint system relaxes to a time-independent, nonequilibrium steady state, where heat is continuously being extracted as work to maintain the particle at the cooled kinetic temperature. This is the scenario we focus on in the following.
To discuss the energetics, we need the equivalent description of the dynamics in (13) in terms of the Fokker–Planck equation for the time-dependent probability density ,
with (probability) currents
Again we can split the velocity current Jvt into irreversible and reversible pieces, as in (5),
This splitting singles out the irreversible current as solely due to the thermal reservoir as before (see (5)), which is required to correctly link the heat and entropy production in the following. Again, this division relies on choosing as time-reversal symmetric, just as in the preceding section.
Our focus is the steady state solution, which due to the linear, Gaussian dynamics is the Gaussian probability density [34],
where the steady-state covariance matrix is
and the associated steady-state currents are and . The entries of can be determined by plugging (18) into (14), as detailed for a more general model in [23]; however their precise expressions are unilluminating and therefore relegated to appendix
where the inequality is only satisfied in the regime of good cooling, . Otherwise too much measurement noise is fed back into the velocity, effectively heating it.
Again from stochastic energetics the heat current is identified as the energy lost due to the irreversible current arising from the thermal noise [22, 38, 39]
which importantly only depends on the velocity as in (5), since the measurement and feedback do not affect the interaction with the thermal environment. In a similar way as before (9) the extracted work rate is
due to the correlations between the feedback force and the particle. In the steady state, can be simplified using the defining equations for the elements of the covariance matrix in appendix
in terms of the velocityʼs relaxation rate, . When the feedback is successful, and we have reduced the kinetic temperature , we must be extracting work, , recovering the results of [23].
We finally will require the fluctuating-trajectory solutions of (13) up to time t, and . We can obtain the probability densities for these trajectories by discretizing time and then using the usual procedure for obtaining path-integral densities, which we sketch in appendix
suitably normalized, and
as
with initial probability density . It cannot be under emphasized that each is not the conditional probability of the feedback process, i.e., , since vt and yt influence each other when there is feedback [8]. Instead, we can understand by first imagining that we fix the entire velocity trajectory v0t, and then evolve yt alone according to (13). This procedure has no feedback and the probability to observe a particular measurement trajectory is exactly . A similar interpretation holds for as well. This distinction between and will become important in section 3.1 when we introduce the transfer entropy rate.
3. Information
In this section, we present the definitions of the various measures of information that can be used to bound the extracted work during a feedback process. In the next section, section 4, we will discuss the physics behind them.
3.1. Transfer entropy rate
The first information measure we discuss is the transfer entropy rate from vt to yt. The transfer entropy is a directional measure of information, which quantifies in an information-theoretic manner how much the dynamics (or more specifically the transition probabilities) of yt are influenced by vt [14]. For our continuous stochastic process, it reads
In appendix
To compare with the other information measures, we calculate its value in our model of feedback cooling. The calculation is facilitated by noting that for stationary Gaussian processes, as we have, integrals of the form (27) can be conveniently expressed in terms of the power spectra—Fourier transforms of the correlation functions. For (27), we demonstrate in appendix
where is the power spectrum of yt, and is the Fourier transform of the variance of yt given a fixed trajectory v0t. We have carried out the integral in appendix
New information is acquired at the relaxation rate of vt, ; that is we learn new information about vt only as fast as vt changes enough to detect. In addition, the transfer entropy rate does not depend on the feedback parameters a and , but only on the measurement accuracy through the dimensionless signal-to-noise ratio , which quantifies the relative size of the measurement accuracy to the thermal diffusion of the velocity. As a result, for perfect measurements without error, , the diverges and with it the transfer entropy rate. Thus, error-free measurement corresponds to infinite information, consistent with the notion that infinite information is required to localize a continuous variable with perfect precision.
3.2. Information flow
We next consider the information flow, whose origin is in the exchange of information between the velocity and the auxiliary measurement device implementing the control. It was first considered in the context of interacting diffusion processes [17], but subsequently has been introduced in the analysis of the thermodynamics of continuously-coupled, discrete stochastic systems [16, 18, 43]. When the coupling is not continuous, but each system takes turns evolving, the information flow simplifies to the mutual information [18–20]. In order to facilitate connections to the other information measures, we sketch in this section the basic arguments leading to the information flow, following the program outlined in [18], and then calculate its value in our feedback cooling model.
First, we must note that strictly speaking this approach requires that yt be the degree of freedom of a physical system, not simply an abstract measurement outcome. Still, in this section we would like not to comment on the precise thermodynamics of yt, taking it only as a generic thermodynamic system. We will come back to its precise interpretation in section 4 when we compare the physics underlying the different information measures.
The key insight in this approach is that the (irreversible) entropy production of the joint system of vt and yt can be divided as
with positive contributions arising due to the irreversible current in the v-direction (16),
and separately from yt, . The next step is to perform the traditional splitting of into the variation of the Shannon entropy due to vt (see (11)),
and the heat (21) as
The additional contribution due to the influence of yt is an information-theoretic piece
which is (minus) the variation of the mutual information3
between vt and yt, due to the fluctuations of vt [44]. The mutual information is a measure of correlations, quantifying how knowledge of the measurement outcomes reduces uncertainty in the velocity. While may be positive or negative, in the regime of good cooling where we are extracting work, we will always have . In the steady state, and , so that (33) reduces to [16, 18]
in the form of (1).
Employing the steady-state solution in (18), we have for the steady-state information flow
where denotes the determinant of . Unfortunately, we have been unable to formulate a more transparent expression in general. Even still, the information rate again only grows as fast as the relaxation rate of the velocity .
3.3. Entropy pumping
For the feedback cooling of a Brownian particle without errors an entropy pumping bound has been introduced by Kim and Qian [21]. This approach has subsequently been developed by Ge [45] and extended to the setup in (13) by Munakata and Rosinberg [22–24], which we discuss in this section.
The entropy pumping approach is based on a coarse graining of the Fokker–Planck equation (14). Following [23], we formally integrate out yt from (14) to obtain the reduced Fokker–Planck equation
where we have identified an effective feedback force
Furthermore, we treat as time-reversal symmetric, as we would expect for an external force [23]. In which case, we single out from the coarse-grained current the irreversible current exactly as for the no-feedback case (5),
This will allow us to connect the entropy production in the environment with the heat.
Equation (38) is not a closed equation for pt (v); the measurement dynamics are required to solve it. Nevertheless, the entropy pumping approach is to treat (38) as a thermodynamically consistent equation for pt (v) with an effective external force . In this case, the entropy balance is developed in analogy to the no-feedback setup, as in (10),
where the second equality follows by substituting in the definition of the coarse-grained current in (40). Here, is equivalent to the expression for the rate of change of the systemʼs Shannon entropy including feedback in (32), and the additional entropy pumping term arises due to the coarse-grained feedback force
As pointed out in [23], the feedback force is proportional to the minimum mean square error estimate of yt given vt. Other than that though, there does not appear to be a crisp interpretation of the entropy pumping as a form of information, like for the transfer entropy rate and information flow.
Using the steady-state distribution in (18), we have for the steady-state entropy pumping [23]
with positivity guaranteed when there is cooling .
3.4. Trajectory mutual information
Another information measure that has aroused some attention is the mutual information rate between the entire v0t and y0t trajectories [15, 46]. For continuous stochastic processes, the trajectory mutual information rate is [44]
It quantifies how much the uncertainty about the entire velocity trajectory v0t is reduced given knowledge of the entire measurement trajectory y0t, and vice versa, as it is symmetric.
The bound on the extracted work follows readily once we observe a close connection between the trajectory mutual information and the transfer entropy pointed out in [16]; by substituting with (26) in , it follows that
after identifying the transfer entropy rate from yt to vt, , defined analogously to (27). The positivity of the transfer entropy implies that
giving (1) for the trajectory information, which is always weaker than the transfer entropy bound.
The trajectory information rate has been studied in numerous contexts and has a well-known expression in terms of power spectra [47–50] that we recall in appendix
in appendix
Comparing with (45), we have as a byproduct the transfer entropy rate from yt to vt,
3.5. Maximum work
A final bound on the extracted work is simply to maximize in (23) with respect to the measurement parameters a and . While the result is not general, remarkably for linear Guassian processes it has a close connection with the transfer entropy rate, as first noticed by Sandberg et al [11]. Using standard calculus methods, the extracted work is bounded above by its maximal value
akin to (1), for parameter values
The optimal measurement has no low-pass filtering: it is immediately fed back into the particle to control it. Remarkably, the extracted work is again bounded by the transfer entropy rate, except multiplied by the cooled kinetic temperature of the particle, instead of T.
3.6. Discussion
3.6.1. Quantitative comparison of information measures
To better understand the relationship between all of these information measures, we plot them all together with in figure 1 as a function of the feedback gain a and measurement error in the range where cooling occurs (). As expected, each information measure bounds the extracted work.
The most striking feature of figure 1 is the hierarchy of information measures
apart from , which does not actually have a generic information interpretation. In fact, this ranking holds quite generally. We have already seen that in section 3.4 when discussing the second-law-like inequality for the trajectory information. The middle inequality, , has been demonstrated by Hartich et al [16] for continuous-time, discrete Markov jump processes. For diffusion processes, a similar conclusion was reached by Allahverdyan et al [17] except for a slightly different transfer entropy rate that uses only the most recent measurement, which upper bounds the transfer entropy rate considered here, as pointed out in [16]. Nevertheless, the proof for jump processes in [16] can be carried over to diffusion processes, once their evolution is discretized. The last inequality between the information flow and the entropy pumping also is generic. This follows by bounding the steady-state entropy production of vt in the information-flow description (31) using a coarse-graining inequality [23] to connect it to the coarse-grained, entropy-pumping approach
where we have employed the entropy balance of entropy pumping in (41). Clearly
As a lower bound on all other information measures, the entropy pumping can be given an information-theoretic interpretation, which till now has been lacking, as a minimal information requirement for successful feedback cooling.
An alternative perspective on this hierarchy of information measures comes from considering the efficiency of work extraction
By utilizing the smaller information measures, we will estimate higher efficiencies, even without changing the measurement or feedback procedure. This conclusion is somewhat surprising, since it makes the notion of efficiency somewhat arbitrary. We will come back to this observation later, after discussing the physical origins of the different information measures.
We also see in figure 1 that the transfer entropy rate and the trajectory mutual information diverge as the measurement error tends to zero, ; whereas the other measures remain finite. Munakata and Rosinberg have also observed that the entropy pumping displays a nontrivial structure, attaining a maximum at the maximum cooling rate [23]. Figure 1 demonstrates that displays a similar structure, but its maximum does not quite correspond to the maximum cooling. Most likely, this discrepancy arises due to the effect of coarse-graining.
3.6.2. Optimal control and the Kalman–Bucy filter
Interestingly, closer connections exist between the information flow, transfer entropy rate, and maximum extractable work that are revealed by re-examining our feedback problem from the perspective of optimal control theory.
The feedback cooling we have been addressing is a special case of a classic problem in optimal control theory: the characterization of feedback controllers that minimize quadratic performance objectives of the form
where is a constant parameter used to tune the trade-off between keeping small fluctuations in vt and applying a strong control force ft, for example [51]. For the special case of cooling, we have been focused on minimizing alone, which corresponds to .
Assuming linear dynamics and Gaussian noise, the optimal feedback controller with access to noisy measurements can be written in the form
where is the abstract dynamical state of the controller, and G and K are carefully chosen constants. According to the separation principle [51, 52], these parameters G and K can be determined as the solutions to two independent optimization problems: the optimal gain G is obtained by minimizing , temporarily assuming there is no measurement noise, ; whereas the optimal K is obtained by minimizing the estimation error, see below, and is independent of the tuning parameter . While the exact expression for the optimal gain G is of little interest to us here, we do note that it tends monotonically to infinity as . This makes intuitive sense, since means we only care about minimizing the variance and assess no cost for large control forces . On the other hand, optimal filtering theory selects an optimal K by minimizing the estimation error
given all the past measurements . The steady-state optimum, achieved for
is
Thus, represents the best estimate of vt given all past measurements. In fact, no other filter, nonlinear or otherwise, can produce a better estimate than the one described here, which is known as the Kalman–Bucy filter [51, 53].
Remarkably, the optimal controller (57) with Kalman–Bucy filter can always be realized using the feedback cooling dynamics in (13) by a simple rescaling
and choosing the parameters a and τ as
This mapping allows us to investigate our information measures from a new point of view by replacing yt with the optimal .
For starters, maximal cooling, which coincides with the maximum extracted work (50), is obtained when , in which case the optimal controller (62) becomes
recovering a* and in (51) as expected.
The optimal controller also extracts the maximum amount of information4 . To see this, first note that optimality of the estimate implies that the estimation error is stochastically orthogonal to the estimate for all t [51]. This property greatly simplifies the steady-state covariance matrix
where the variance of the estimate is simply
Note optimal cooling is achieved by , forcing , so that fluctuations in the velocity are only caused by estimation error. Furthermore, by exploiting the structure of in (64), the expression for the steady-state information flow (37) greatly simplifies,
for all G. This is a very interesting observation, supporting the claimed optimality of the Kalman–Bucy filter. We already know that . What we see here is that the class of controllers given by (57), i.e., with K fixed (59) and G free, saturates the bound, maximizing the information flow. Hence, a controller with a small gain G (zero even) only uses information to create an optimal estimate of the process, whereas a high gain cools as well. To gain further insight into equality (66), we have to look at the transfer entropy rate and information flow from a different perspective. Namely, the transfer entropy rate can also be defined as the rate of growth of the mutual information between vt and the entire trajectory of measurement outcomes y0t, that is the change in . On the other hand, the information flow is the rate of growth of the mutual information between vt and just the most recent measurement yt, that is the change in . The inequality is then related to the simple idea that the entire trajectory of measurements contains more information than just the last. Now, it is known that the Kalman–Bucy filter is a sufficient statistic for the conditional distribution of vt given the measurements [54]. In other words, everything useful in a collection of measurements for predicting vt is contained in just , or in terms of the mutual information . This equality translated into rates implies (66).
In figure 2, we illustrate how the extracted work depends on G, and how the maximum is asymptotically achieved. In addition, we see that holds for all G. We can also conclude that with certain choices of a and τ (namely and in (62)) our original setup (13) can always saturate , which is indeed observed in figure 1 for .
Download figure:
Standard image High-resolution image4. Energetics of information and measurement
We have seen that there are various, distinct measures of information that each offer a nontrivial bound for the extracted work. However, there does not seem to be an obvious reason to prefer any one of these measures. To this end, we investigate their origins in this section. We will find that the transfer entropy rate and the information flow both correspond to the information that is recorded in an auxiliary system, or memory, and therefore is subject to the limits of thermodynamics, as originally suggested by Landauer [2]. In particular, we show that these two information measures both bound the minimum energy required to gather that information through distinct thermodynamic processes, implying that the energy that we are able to extract as work originates in the (free) energy supplied by the memory.
4.1. Information flow
Let us start with the simpler measurement scenario corresponding to the information flow . Actually, we have already touched on its physical interpretation when we introduced it in section 3.2. Recall, there we considered the measurement outcomes yt to correspond to a physical degree of freedom of an auxiliary system. We now clarify that interpretation by taking yt to be the position of a secondary, harmonically-trapped, overdamped Brownian particle. To be thermodynamically consistent, the origin of the measurement noise must be a thermal reservoir, which requires imposing the Fluctuation-Dissipation theorem [39]
We have chosen the temperature of the measurement device to be the same as the controlled system, which is the customary choice. From this point of view, (13) is the equation of motion for an overdamped Brownian particle with viscous damping coefficient τ trapped in a harmonic potential of unit spring constant, centered about the velocity, as illustrated in figure 3. Alternatively, such a coupling can be implemented in an electric circuit as was presented in [11]. The result is that the position of the measurement oscillator yt feels a fluctuating force making it track the velocity vt, thereby establishing and maintaining correlations. Roughly speaking, the measurement oscillator is constantly learning new information at a rate , which keeps getting rewritten in the value of its position.
Download figure:
Standard image High-resolution imageWhen introducing the information flow, we divided the entropy production into two positive contributions (30), one due to the velocity , and another due to the measurements . When studying the extracted work , we focused on . However, a similar analysis also holds for , which verifies that the y-system must consume at least free energy to sustain the correlations that promote feedback. Observing that as a position yt is even under time-reversal (consistent with our previous analysis in section 2.1), we develop its thermodynamics by splitting its current J yt (15) into irreversible and reversible portions as
Notice that here the irreversible current is the time-symmetric contribution, since yt is even under time-reversal [36]. Then, repeating the analysis in section 3.2, we have that in the steady state [16–18]
where is the heat flow rate into yʼs reservoir. Thus, in the steady state
In order to track the velocity, yʼs environment continually absorbs heat at a rate . We verify this bound in figure 4, where is plotted with . The minimum is reached when the measurement device operates reversibly in the limit , so that yt rapidly relaxes to its instantaneous equilibrium centered about vt: .
Download figure:
Standard image High-resolution imageIn addition, we have already argued that the entropy pumping lower bounds the information flow, (54). As a result, offers a weaker lower bound on the energy required for an auxiliary system to provide the entropy-pumping feedback, , which is verified in figure 4 as well.
4.2. Transfer entropy rate
The transfer entropy rate can also be understood as the minimum free energy required to measure, but with an alternative measurement scenario. In the previous section, the information flow was shown to bound the energy required to constantly rewrite a single memory with each new measurement yt. By contrast, the setup for the transfer entropy rate is much closer to that envisioned by Landauer and Bennett in their thermodynamics of computation [2, 55]: each measurement is recorded separately in one of a collection of memories through a specific driven thermodynamic process; one example of which was recently described in [25].
The central idea is that each measurement outcome is recorded in a distinct memory. Therefore, to track the system over any finite time interval, say from time s = 0 to t, we require an infinite number of memories in which to record the infinity of measurements. However, this is difficult to analyze. So to proceed, we discretize time as , with and , where the measurement outcome at time sk is denoted simply as , and similarly . To store these measurement outcomes, we imagine a collection of N auxiliary memories with phase space positions mk, prepared initially in positions distributed according to . The measurement is a thermodynamic process during a time interval of length θ in which the kth memory is manipulated, with the velocity fixed, in such a way to reproduce the correlations with embodied in the measurement statistics of yk. In other words, we demand that the statistics of the kth memory after the measurement are (equality in distribution).
To see how these ideas play out in our model system, consider the discretized version of the Langevin equation (13)
where the are independent Gaussian random variables of zero mean and covariance . Equation (71) is a rule that tells us how the measurement outcome yk at time sk depends on the velocity as well as the past measurement outcome stored in a previous memory. Such measurements that depend on past outcomes are sometimes called non-Markovian measurements [42]. Specifically, yk is characterized by a Gaussian probability density
roughly centered about the velocity with a variance depending on the measurement error. Now, in view of our previous discussion, we desire a physical system to act as a memory and a measurement process that prepares that system in a statistical state with the probability density in (72). A natural choice is an overdamped harmonic oscillator coupled to a thermal reservoir at temperate T. Initially each memory oscillator is prepared in equilibrium with an arbitrary initial spring constant k0 centered about zero, as illustrated in figure 5. Since each measurement is performed in sequence, it is attractive to visualize the phase spaces of the N measurement oscillators aligned in a row, or tape. Then one by one we couple each measurement oscillator to the system as well as past memories, so as to establish correlations. The density in (72) suggests that the measurement protocol for the kth oscillator should be the quasistatic turn-on of an interaction that shifts the center of the harmonic oscillator to —which includes interactions with the past memories—while simultaneously expanding the spring constant to , which results in the interaction potential
as depicted in figure 5. As a result, upon completion of the kth measurement the memoryʼs position has settled into an equilibrium distribution equivalent to (72). To complete the measurement, we must freeze the state of the memory to lock in the correlations, and remove the interactions by turning off V. One possible, though admittedly idealized, option is to instantaneously set V = 0, and then immediately turn off the dynamics of the measurement oscillator—perhaps by quenching the temperature to zero—so that the oscillator can no longer move. By repeating this sequence of actions on each new memory, we store a collection of measurement outcomes, each in a different physical memory. Now to be precise each measurement has to be performed instantaneously from the point of view of the velocity. This merely means that the time-scale of the evolution of the individual memories must be much faster than the velocity, , so that the measurement is completed before the velocity changes appreciably [25]. However, this assumption is not unreasonable, since measurements are usually assumed to read out the instantaneous state of the system.
Download figure:
Standard image High-resolution imageHaving described how we can mimic the measurement statistics in a physical situation, we now address the thermodynamics from a general point of view, applying the methods of [19, 20, 25]. Our analysis is based on the following second-law-like inequality that relates the work performed in an isothermal process to the increment in the nonequilibrium free energy [56, 57]: for a thermodynamic system with microscopic states z, the work W performed along an isothermal process during which the systemʼs probability density transitions from p(z) to is bounded as
where is the change in the nonequilibrium free energy defined in terms of the average internal energy U(z) and entropy . The nonequilbirium free energy is a natural extension of the equilibrium free energy to systems characterized by an arbitrary probability density, since it reduces to the equilibrium free energy for systems in equilibrium.
We begin by focusing on the work done during the kth measurement, Wk, during which the kth memory becomes correlated with not only the velocity but all the past memories through the velocity which depends on the entire past. Applying (74), we have
where is the change in the nonequilibrium free energy of the kth memory, corresponding to the change in the conditional density from —due to the initial independence of each memory—to . We single out the new correlations by introducing the mutual information between and conditioned on all the past measurements as [44]. Substituting in this definition, (75) becomes
where is the change in free energy conditioned on just the past memories: . Summing over all measurements, we find
where is the work to perform all N measurements, is the change in entire tapeʼs free energy, and we have identified the discrete version of the transfer entropy [10]
which is reviewed in appendix
Taking the limit as the number of measurements go to infinity while the time between them goes to zero, we obtain
Thus, the transfer entropy rate is the minimum rate at which free energy is consumed to write to the memories. The slow protocol that we described previously saturates this bound, since it is quasistatic and therefore thermodynamically reversible.
At this point, it is worthwhile to make a connection to a class of Maxwell-demon models that exploit a tape of low entropy, auxiliary systems or cells, similar to what we have just described [26, 58–62]. Apart from the study in [62], these models use an ideal tape that has no internal energy, and therefore cannot exchange energy with the system, but only entropy; a setup sometimes referred to as an information reservoir [26, 60]. Under these conditions, a second-law-like inequality has been predicted that shows that the extracted work is bounded by the increase in entropy of each individual auxiliary system, ignoring the correlations between the different cells. Our memories, on the other hand, have internal energy and therefore the natural thermodynamic quantity to consider is the free energy instead of the entropy. Therefore to fit our measurement model into this tape-model framework, we must relate our information bound on the work to measure to a bound that ignores the correlations. To this end, we start with the bound for the energy to measure in (77), which includes through all the correlations between different memories. By noting that ignoring correlations and conditioning increases the entropy, [44], we can conclude that ignoring the correlations decreases the free energy . As a result, we have from (77) and the initial independence of each memory the series of inequalities
For the ideal tape with no internal energy this reduces to recovering the ideal-tape bound [26, 58–62] in our setup. Equation (81) may lead us to conclude that the bound on the extracted work from the tape-model framework, , is weaker than that provided by the transfer entropy. However, this would be too hasty, because these tape models allow a more general interaction between the tape cells and the system. Whereas, in our setup the memory evolution is assumed to occur separately with the velocity fixed, the tape models consider a dynamics where the memory (or cell) would be allowed to evolve simultaneously with the velocity. From this point of view, the measurement model we have presented is a special case of these more general tape models, and it is exactly our assumption that the velocity is frozen during measurement that allows us to tighten the tape-model bound using the transfer entropy. Further comparisons of such tape models with other information measures and more traditional statements of the second law can be found in [25, 26].
Finally, it should be noted that the preceding second law analysis can be viewed as a specific implementation of the information flow framework (outlined in sections 3.2 and 4.1) applied to a nonautonomously driven auxiliary memory composed of a sequence of many subsystems, see [18].
4.3. Discussion
The transfer entropy rate and information flow both bound the energy consumed during measurement. However, each measurement scenario is distinct, and in general each of these information measures will not bound the energy consumption for the otherʼs measurement scenario. An example where is possible is presented in [16]; thus, the transfer entropy rate does not generally lower bound the amount of heat dissipated by a single memory being constantly rewritten. Our model corroborates this observation, as verified in figure 4 by the crossing of and . The one exception is if the the controller implements the Kalman–Bucy filter (57). In which case, the equality of the information measures, , implies that there is a unique lower bound to the energy required for measurement.
To conclude this section, we take a broader perspective. Our observation that the transfer entropy rate and information flow both represent the minimum (free) energy consumed (or alternatively the entropy produced) in the auxiliary memory to create that information, suggests that it is reasonable to interpret some second-law-like inequalities as actually telling us something about the thermodynamics of the system and its surroundings, where the surroundings include the measurement device. This allows us to incorporate information into the standard statement of the second law of thermodynamics through a kind of information reservoir on equal footing with other traditional thermodynamic reservoirs, similar to what was suggested for tapes in [26, 60]
which is equivalent to (1) in the steady state. Here, represents the minimum entropy produced in the environment that allows for feedback, with the minimum attained for reversible measurement. The appropriate choice of —transfer entropy rate or information flow—depends on which type of information reservoir we wish to use. From this point of view, the efficiency introduced in (55) is a true measure of energetic efficiency that quantifies how faithfully the energy supplied by a reversible memory is extracted back out as work.
5. Summary
We have explored a collection of information measures that appear in second-law-like inequalities for measurement and feedback, using the tools of stochastic thermodynamics and optimal control theory. We have seen that these measures form a hierarchy of bounds on the extracted work and that the Kalman–Bucy filter optimally will extract information and energy. Even though each measure offers a different numerical bound on the extracted work, they also each correspond to different ways of gathering information. With this distinction in mind, these second-law-like inequalities can be seen as manifestations of the second law of thermodynamics, since they include the entropy production of the system and surroundings, including the controller.
Acknowledgments
We would like to thank Martin Rosinberg for a carefully reading of this manuscript. JMH is supported financially by the ARO MURI grant W911NF-11-1-0268 and HS is supported financially by the Swedish Research Council under grant 2013-5523.
Appendix A.: Steady state probability density
The Gaussian steady state probability density in (18) is completely characterized by its means, which are zero, and the covariance matrix . The elements of can be determined by exploiting the Fokker–Planck equation (14) to develop a collection of equations for the variances , , and , as described in [63] for example. The time-independent steady state solutions can then be shown to satisfy the algebraic equations
whose solutions can be obtained after some lengthy algebra
Appendix B.: Path probabilities and the transfer entropy rate
In this appendix, we demonstrate how we arrive at (26) for the trajectory probability density , and how this structure allows the compact expression for the transfer entropy rate in (27).
The analysis precedes by discretizing the evolution over the time interval s = 0 to t into steps of width as for with and . We are interested in determining the probability density to observe the pair of discrete trajectories and . To this end, we discretize the Langevin equation (13) as
where () are independent, zero-mean, Gaussian random variable with covariance (). From this we deduce that to lowest order in the transition probability splits into separate v and y evolutions as [17]
Thus, the joint trajectory probability takes the form
with arbitrary initial density . Since the evolution naturally divides, it suggests introducing the trajectory conditional probabilities
in terms of which the joint trajectory probability becomes
Equations (24)–(26) are the continuous time versions of the preceding equations obtained in the limit .
In this discretized setup, we can directly apply the theory of discrete feedback [3, 5, 7, 16, 27, 41]. Here, the transfer entropy after N measurements is given as
We see that the transfer entropy is the relative entropy between the transition probability of y given v, , and the unconditioned transition probability, , averaged over . Recall that the relative entropy between two probability densities f(x) and g(x) is [44]. In this way, the transfer entropy measures the affect the velocity has on the measurement dynamics, that is, how distinguishable the measurement dynamics given the velocity are from the measurement dynamics without the velocity. Expanding the sum we can rewrite (B.7) as
The continuous time version appears in (27).
Appendix C.: Power spectra formulae for information rates
In this appendix, we sketch how entropy rates for stationary Gaussian processes can be expressed in terms of the processes' correlation functions, following the developments in [49, 50].
Let us consider a discretization with spacing of a Gaussian stochastic process . It is completely characterized by its mean and covariance matrix with elements , which we assume to be time-independent, , an example being a stationary process:
The power spectra formulae for the information rates follow from the observation that the entropy of such a Gaussian distribution is completely characterized by the covariance matrix
Since the process is time-invariant, the covariance matrix has a Toeplitz structure, , which allows us to diagonalize it in the limit using its Fourier transform , with . In which case, the entropy rate can be expressed as [50]
The transfer entropy is the difference in entropy rate between the trajectory of measurement outcomes and the entropy rate for :
Taking the continuous time limit , we recover the expression in (27). Similarly, the trajectory mutual information is
where is the Fourier transform of the covariance matrix of the joint measurement and velocity process. One can show, as in [49], that
which when substituted into (C.5) recovers (47) after the taking .
Appendix D.: Calculation of information rates
In this appendix we calculate in (28) and in (47). As a first step, we must determine the Fourier transforms of various correlation functions. To this end, we begin by Fourier transforming the equations of motion for vt and yt in (13)
with and .
Let us start by determining , which requires two correlation functions obtained from the solutions of (D.1) as
and
where is known as the sensitivity function of the feedback system [40]. Thus, the transfer entropy rate is
These integrals can be performed by exploiting the formula [64]
In particular,
and
which recovers (29). That logarithmic integrals of the sensitivity function, such as B, equals zero holds with great generality. In fact, it represents a well-known conservation principle in control theory known as Bodeʼs integral formula [40].
To determine , we first note that . Since we already know , it remains to determine
The power spectra are obtained from (D.1) as
Therefore, recognizing that the contribution from the sensitivity function is zero, we have
by virtue of (D.5).
Footnotes
- 3
- 4
A different perspective on the information-theoretic optimality of the Kalman–Bucy filter can be found in [54].