This site uses cookies. By continuing to use this site you agree to our use of cookies. To find out more, see our Privacy and Cookies policy.
Paper The following article is Open access

Measurement-feedback formalism meets information reservoirs

, and

Published 21 January 2016 © 2016 IOP Publishing Ltd and Deutsche Physikalische Gesellschaft
, , Citation Naoto Shiraishi et al 2016 New J. Phys. 18 013044 DOI 10.1088/1367-2630/18/1/013044

1367-2630/18/1/013044

Abstract

There have been two distinct formalisms of thermodynamics of information: one is the measurement-feedback formalism, which concerns bipartite systems with measurement and feedback processes, and the other is the information reservoir formalism, which considers bit sequences as a thermodynamic fuel. In this paper, we derive a second-law-like inequality by applying the measurement-feedback formalism to information reservoirs, which provides a stronger bound of extractable work than any other known inequality in the same setup. In addition, we demonstrate that the Mandal-Jarzynski model, which is a prominent model of the information reservoir formalism, is equivalent to a model obtained by the contraction of a bipartite system with autonomous measurement and feedback. Our results provide a unified view on the measurement-feedback and the information-reservoir formalisms.

Export citation and abstract BibTeX RIS

Original content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

1. Introduction

Evaluating the amount of work extraction from reservoirs is an important issue in thermodynamics and nonequilibrium statistical mechanics. Here, we focus on the work extraction by using information, where two distinct formalisms exist. One formalism concerns the measurement-feedback processes between two systems: a celebrated example is Maxwell's demon. In this case, the generalizations of the second law and the fluctuation theorem have been obtained by including the mutual information[1] between the system and the controller (demon) [213], and it has been applied to biochemical sensing [1416]. We call this avenue of research the measurement-feedback (MF) formalism. The other formalism concerns the work extraction from bit sequences called information reservoirs, where the Shannon entropy of a long bit sequence is consumed to extract work (see also figure 1(a)). A second-law-like inequality stronger than the conventional second law has been obtained in this formalism [1725]. We call this as the information-reservoir (IR) formalism. Although both of these two formalisms address the fundamental link between information and thermodynamics, the relation between the MF and the IR formalisms has not been fully understood. In previous researches, an important observation is that the IR formalism gives a stronger inequality than the MF formalism in some setups [2325].

Figure 1.

Figure 1. (a) Schematic of the simplified MJ model. The horizontal lines represent the energy levels corresponding to the potential energy of the weight, and the green ball represents the height of the weight, which plays the role of a work storage. A long bit sequence with 0 and 1 moves from left to right, which can flip its own states only at the central site (i.e., yellow cage in the figure). The flip of a bit is coupled to the up-down of the weight. (b) The labeling of the states of the weight corresponding to the incoming bit. If the incoming bit is 0 (1), the state of the weight is labeled as d (u), and the higher (lower) energy level is labeled as u (d). (c) The state space and transition rates of the simplified MJ model. The two red arrows constitute the subset Ω Th.

Standard image High-resolution image

In contrast, in this paper we show that the MF formalism is also applicable to IR models, and leads to a stronger second-law-like inequality than the IR formalism. The central idea here is the additive decomposition of the total entropy production, which was first investigated in the context of a single measurement and feedback [7], and then extended to general Markov jump processes [12]. This decomposition assigns the partial entropy production to each transition path, for which a generalized second law is obtained.

In addition, we show that the Mandal-Jarzynski (MJ) model, a prominent IR model, can be regarded as a contracted model of an autonomous MF model even at the level of stochastic trajectories. In this special case, the newly obtained inequality is shown to be equivalent to the inequality with the mutual-information flow in terms of the MF formalism [4, 9, 10]. Our result would lead to the comprehensive understanding of the link between information and thermodynamics.

2. Partial entropy production

We briefly review the notion of the partial entropy production introduced in [12], which is an extension of the MF formalism. Throughout this paper, we consider Markov jump processes with discrete states and continuous time, and focus on the stationary states in the presence of a single heat bath. We normalize the inverse temperature β as 1 for simplicity. The transition from a state w' to another state w is written as w' →w, whose transition rate is denoted by P(w'→w). The stationary distribution P(w) satisfies the stationary condition

Equation (1)

We define the total entropy production σtot with time interval $0\leqslant t\leqslant T$ as [26]

Equation (2)

where we denoted the number of jumps by N, and the i-th jump occurs from wi − 1 to wi at time ti ($1\leqslant i\leqslant N$). The state at time t is written as w(t). $Q({w}^{\prime }\to w)=-\mathrm{ln}(P({{\rm{w}}}^{\prime }\to {\rm{w}})/P({\rm{w}}\to {{\rm{w}}}^{\prime }))$ is the heat absorption by the system from the heat bath with the transition w' → w, and s(w):= − lnP(w) represents the stochastic Shannon entropy at the state w. We note that the local detailed balance condition [27] is assumed.

We now define the partial entropy production σΩ for a set of transitions, written as Ω, which is a subset of all possible transitions. We first define the partial probability flux at the state w as ${J}_{{\rm{\Omega }}}(w):= {\sum }_{\{{{w}^{\prime }}| ({w}^{\prime }\to w)\in {\rm{\Omega }}\}}J({{w}^{}}^{\prime} \to w)$. We then define the partial entropy production σΩ as

Equation (3)

Here, we defined ${Q}_{{\rm{\Omega }}}:= {\sum }_{i=1}^{N}Q({w}_{i-1}\to {w}_{i}){\delta }_{{\rm{\Omega }}}({w}_{i-1}\to {w}_{i})$, where δΩ (w'→w) takes 1 if ${{w}^{}}^{\prime} \to w\in {\rm{\Omega }}$ and takes 0 otherwise. We also defined ${\rm{\Delta }}{s}_{{\rm{\Omega }}}:= {s}_{{\rm{\Omega }},{\rm{jump}}}-{\displaystyle \int }_{0}^{T}{J}_{{\rm{\Omega }}}(w(t))/P(w(t)){dt}$ with ${s}_{{\rm{\Omega }},{\rm{jump}}}\;:= {\sum }_{i=1}^{N}$ $(s({w}_{i})-s({w}_{i-1}){\delta }_{{\rm{\Omega }}}({w}_{i-1}\to {w}_{i}))$. It is crucial that the partial entropy production σΩ is an additive decomposition of the total entropy production

Equation (4)

where Ωc is the complement of Ω. The partial entropy production satisfies the fluctuation theorem:

Equation (5)

which leads to the generalized second law for the partial entropy production:

Equation (6)

It has been shown that the fluctuation theorem for autonomous measurement and feedback processes is obtained as a special case of equation (6) [12]. We further discuss the significance of this inequality in section 4.

3. Mandal–Jarzynski model

We here explain the MJ model which extracts work from a bit sequence [17, 24]. In particular, we consider a simplified version of the MJ model introduced in [24]. The simplified MJ model is a composite system of a bit sequence and a weight with the coupling reaction $(0,d)\leftrightarrow (1,u)$ (see figure 1(a)). Here, 0 and 1 represent states of a bit, u and d represent (relative) positions of the weight on ladder-like energy levels with energy interval mgΔh. The weight plays the role of a work storage. The bit sequence moves in one direction, from left to right, with the Poisson process with the rate γ. The probability that the incoming bit is 1 is epsilon. When a new bit comes into the central site, the state of the weight is instantly labeled as d (u) corresponding to the state of the incoming bit 0 (1), and the higher (lower) energy level is labeled as u (d) (see figure 1(b)). Under this labeling, the possible states of the composite system are only (0, d) and (1, u).

There are two possible transitions between u and d (see figure 1(c)). One is the stochastic jump induced by the thermal bath with rates ${P}_{{\rm{heat}}}^{{\rm{MJ}}}$(du) = k+ and ${P}_{{\rm{heat}}}^{{\rm{MJ}}}$(ud) = k, which satisfy k+/k = emgΔh. This is the coupling reaction of the up-down of the weight and the flip of a bit such that $(0,d)\leftrightarrow (1,u)$. We denote the set of these transitions by Ω Th. The other is induced by the bit stream with rates ${P}_{{\rm{bit}}}^{{\rm{MJ}}}(d\to u)=\gamma \epsilon $ and ${P}_{{\rm{bit}}}^{{\rm{MJ}}}(u\to d)=\gamma (1-\epsilon );$ from the above-mentioned rule of the state labeling, the stream of the bit sequence also changes the states u and d. If epsilon is small (i.e., the entropy of the bit sequence is low), the weight can be raised on average by increasing the entropy of the bit sequence. The stationary probability of u is given by ${p}_{\tau }:= \tau p+(1-\tau )\epsilon $, where $\tau := ({k}_{+}+{k}_{-})/({k}_{+}+{k}_{-}+\gamma )$ and $p:= {k}_{+}/({k}_{+}+{k}_{-})$. Since pτ is the probability that the outgoing bit is 1, the average energy gain per single incoming bit is given by ${W}_{\tau }={mg}{\rm{\Delta }}h({p}_{\tau }-\epsilon )$, which we regard as work extraction. We note that the entropy change in the work storage is irrelevant to our argument, because the fluctuation of the state of the work storage does not affect the procedure of autonomous work extraction. This type of work storages for autonomous work extraction have been discussed in [19, 31, 32].

4. Stronger second law from (6)

We now apply the fluctuation theorem with the partial entropy production, discussed in section 2, to the simplified MJ model. We set Ω in section 2 to Ω Th in section 3 (the red arrows in figure 1(c)). Equality (5) and inequality (6) are then written as

Equation (7)

and

Equation (8)

By using the stationary condition $(1-{p}_{\tau })\cdot {k}_{+}-{p}_{\tau }\cdot {k}_{-}={p}_{\tau }\cdot \gamma (1-\epsilon )-(1-{p}_{\tau })\cdot \gamma \epsilon =\gamma ({p}_{\tau }-\epsilon )$, inequality (8) reduces to a simple form:

Equation (9)

The left-hand side of (9) represents the entropy change induced by the heat bath.

We next extend our foregoing argument to general Markov jump processes with states $\{{w}^{1},\cdots ,{w}^{M}\}$. In line with [25], we divide pairs of transitions into two groups, which we refer to as thermal transitions and informational transitions. We denote these sets of transitions by Ω Th and Ω I, respectively (see figure 2). The work is extracted during the thermal transitions. By setting Ω Th to Ω, equation (6) reduces to

Equation (10)

where we defined ${Q}_{{\rm{Th}}}:= {\sum }_{{w}^{i}\leftrightarrow {w}^{j}\in {{\rm{\Omega }}}_{{\rm{Th}}}}J({w}^{i}\to {w}^{j})Q({w}^{i}\to {w}^{j})$.

Figure 2.

Figure 2. State space of a Markov jump process. We divide pairs of transitions into thermal transitions (red arrows) and informational transitions (black arrows). We consider the partial entropy production associated with the thermal transitions.

Standard image High-resolution image

5. Hierarchy of second-law-like inequalities

We now show the hierarchical structure of three upper bounds on Wτ with the simplified MJ model: the conventional second law, an inequality derived from the IR formalism [17, 24], and the newly derived inequality (9). First, we compare the conventional second law and the inequality obtained from the IR formalism. To clarify the thermodynamic energy cost of the one-way stream of the bit sequence, we model the bit sequence by non-interacting particles with two components 0 and 1 (figure 3), where the transition rates are the same as ${P}_{{\rm{bit}}}^{{\rm{MJ}}}$. Particles in the bath stochastically come and go to the reaction site, which corresponds to the yellow cage of the MJ model. The ratio of the particles 0 and 1 in the bath is $1-\epsilon \quad :\epsilon $, and therefore the gain (loss) of the chemical potential with the transition ud (du) is ${\rm{\Delta }}\mu =\mathrm{ln}((1-\epsilon )/\epsilon )$. Although the kinetics of the particle-bath model is equivalent to that of the bit sequence, the entropy production in the particle-bath and that in the bit sequence are different as shown below. The conventional second law for the total system with the particle-bath is given by

Equation (11)

On the other hand, the second law for the total system with the bit sequence has been derived from the IR formalism, which is given by [17, 24]

Equation (12)

We refer to this inequality as the MJ inequality. Subtracting the left-hand side (lhs) of (12) from the lhs of (11), we obtain

Equation (13)

where $D(\cdot \parallel \cdot )$ represents the Kullback-Leibler divergence that is non-negative [1]. Inequality (13) confirms that the MJ inequality gives a stronger bound than the conventional second law.

Figure 3.

Figure 3. Schematic of the particle-bath model of a bit sequence. A particle labeled as 0 or 1 stochastically enters the reaction site (the yellow cage) with the probability ratio $1-\epsilon :\epsilon $, and leaves the space with 1−pτ : pτ.

Standard image High-resolution image

Their difference $D({p}_{\tau }\parallel \epsilon )$ characterizes an extra cost to move the bit sequence in one direction. In the particle-bath case, the outgoing particle from the reaction site is in state 1 with probability pτ, while a particle in the particle bath is in state 1 with probability epsilon. Thus, when a particle leaves the reaction site to the particle bath, the entropy production in the particle bath is calculated as $-{p}_{\tau }\mathrm{ln}\epsilon -(1-{{p}}_{\tau })\mathrm{ln}(1-\epsilon )$. In contrast, in the bit sequence case, the ratio of bit 1 in the bit sequences on the left side of the cage (i.e., the particle bath releasing the incoming bit) and that on the right side of the cage (i.e., the particle bath receiving the outgoing bit) are different. Therefore, a bit on the right side of the cage is in state 1 with probability pτ, and when a bit leaves the reaction site to the bath, the entropy production in the bath is calculated as $-{p}_{\tau }{\rm{ln}}\quad {{p}}_{\tau }-(1-{{p}}_{\tau })\mathrm{ln}(1-{{p}}_{\tau })$. The difference between the foregoing two entropy productions is equal to $D({p}_{\tau }\parallel \epsilon )$. This implies that the strength of the MJ inequality comes not from the use of the bit sequence carrying information but from the one-directional operation, which inevitably needs the additional energy cost.

We now compare two inequalities (12) and (9) (i.e., the MJ inequality and the inequality derived from the partial entropy production method). Subtracting the lhs of (9) from the lhs of (12), we get

Equation (14)

Here, $D(\epsilon \parallel {p}_{\tau })$ characterizes the dissipation corresponding to the fact that the time evolution is not quasi-static [28]. The left-hand side of the MJ inequality (12) is just characterized by the entropy change between the initial state (i.e., the state of the incoming bit) and the final state (i.e., the state of the outgoing bit). Therefore, the MJ inequality does not take into account the dissipation due to the free relaxation of a bit, while inequality (9) can take it into account. Such dissipation is unavoidable, as autonomous information engines inevitably involve relaxation processes, because of the very fact that they are autonomous. This is the physical reason why the partial entropy production method gives a stronger bound of the work extraction.

More generally, we can show that inequality (10) is a stronger inequality than the generalized version of the MJ inequality for the general IR formalism [25]. The setup for the generalized MJ inequality is the same as that for inequality (10). The general version of the MJ inequality (12) is written as [25]

Equation (15)

where we defined ${p}_{{ij}}:= P({w}^{i})/(P({w}^{i})+P({w}^{j}))$ and ${\gamma }_{{ij}}:= P({w}^{i}\to {w}^{j})+P({w}^{j}\to {w}^{i})$. We then compare two inequalities (15) and (10). Subtracting the lhs of (10) from lhs of (15), we obtain

Equation (16)

where we defined a positive constant ${\epsilon }_{{ij}}$ as ${\epsilon }_{{ij}}:= P({w}^{j}\to {w}^{i})/(P({w}^{i}\to {w}^{j})+P({w}^{j}\to {w}^{i}))$ for each informational transition ${w}^{i}\leftrightarrow {w}^{j}$. Therefore, inequality (10) is stronger than (15).

6. Contraction from a MF model to the MJ model

We next consider an autonomous measurement-feedback (AMF) model [8], and show the equivalence of the simplified MJ model and the AMF model. This equivalence allows us to directly compare the MJ inequality and the second-law-like inequality with the mutual information flow. The AMF model consists of a fluctuating particle on the tilted one-dimensional discrete lattice and the delta-function-like wall (see figure 4(a)). The position on the lattice is labeled as $x\in \{0,1\}$ in mod 2, and the delta-function-like walls are at the right side of 0 or 1, which we label as y = e or o, respectively. The energy difference of the particle between a site and its right neighbor is mgΔh. The transition rate of the position of the particle due to the thermal diffusion is given by

Equation (17)

with k+/k = emgΔh. The transition rate of the positions of the walls is given by

Equation (18)

which are equal to ${P}_{{\rm{bit}}}^{{\rm{MJ}}}$(ud) and ${P}_{{\rm{bit}}}^{{\rm{MJ}}}(d\to u)$, respectively. We note that Δμ is regarded as the chemical potential difference that drives the switching of the walls. Owing to the chemical driving, the particle climes up against the potential bias when mgΔh < Δμ.

Figure 4.

Figure 4. (a) State space of the AMF model, which consists of the tilted potential and the delta-function-like walls. The position of the particle is labeled as 0 or 1, and the position of the walls are labeled as e or o. The switching of the position of the walls is driven by the chemical potential, and the position of the particle changes due to the thermal diffusion. (b) Contraction of the AMF model. The states (0, o) and (1, e) are contracted to d, and (0, e) and (1, o) to u. The contracted state space and transition rates are exactly equal to those of the MJ model illustrated in figure 1(c).

Standard image High-resolution image

We stress that the particle and the walls do not directly change the energy, but exchange only information. In this sense, the AMF model is a typical example of an autonomous information-driven engine [12, 13, 29, 30]. Analogous to Maxwell's demon, the walls measure the position of the particle, and change their own position depending on the measurement outcome, which is the feedback control. The AMF model is a bipartite system with two variables (x,y), and thus we can define the entropy production associated only with the particle as ${\sigma }_{x}:= -{Q}_{x}+{\rm{\Delta }}{s}_{x}$, where Qx is the heat absorbed by the subsystem x, and ${\rm{\Delta }}{s}_{x}:= -\mathrm{ln}P({\rm{x}}({\rm{T}}))+\mathrm{ln}P({\rm{x}}(0))$ is the change in the stochastic Shannon entropy of x. Due to the presence of the autonomous measurement and feedback, σx can be negative on average and does not satisfy the fluctuation theorem in the conventional form. However, by taking into account the flow of stochastic mutual information $I(x;y)=\mathrm{ln}P({\rm{x}},{\rm{y}})/P({\rm{x}})P({\rm{y}})$ [7], we obtain the modified fluctuation theorem in the stationary state [12]:

Equation (19)

Here we defined the mutual information flow by ${\rm{\Delta }}{I}_{x}:= {I}_{x,{\rm{jump}}}+{\int }_{0}^{T}{J}_{x}(x,y)/P(x,y){dt}$, where ${I}_{x,{\rm{jump}}}:= {\sum }_{i=1}^{N}$ $(I({x}_{i};{y}_{i})-I({x}_{i-1};{y}_{i-1})$ ${\delta }_{{y}_{i},{y}_{i-1}})$ with ${w}_{i}=: ({x}_{i},{y}_{i})$ and Kronecker delta δ, and ${J}_{x}(x,y):= {\sum }_{{x}^{\prime }}J(({{x}^{}}^{\prime} ,y)\to (x,y))$. We stress that equation (19) is a special case of equation (5): if we apply equation (5) to the AMF model by setting all transitions of x to Ω, we straightforwardly obtain equation (19). By applying Jensen's inequality to equation (19), the generalized second law with the mutual information flow is obtained as [4, 9, 10, 12]

Equation (20)

where $\langle {\dot{\sigma }}_{x}\rangle $ represents the stationary entropy production rate of x. This inequality is stronger than the conventional second law of the composite system.

We now consider the contraction of the AMF model by contracting two states (0, o) and (1, e) to a single state d, and (0, e) and (1, o) to u (see figure 4(b)). This contraction means that the absolute positions of the particle and walls are replaced by the relative positions. Since the AMF model has the two types of transition paths corresponding to the thermal diffusion of the particle and the chemical switching of the walls, its contracted model also has two types of transition paths between u and d, which we refer to as the thermal transition and the informational transition. Then, by denoting the transition rates of the thermal and informational transitions by ${P}_{{\rm{heat}}}^{{\rm{cAMF}}}$ and ${P}_{{\rm{bit}}}^{{\rm{cAMF}}}$ respectively, the contracted AMF model is regarded equivalent to the simplified MJ model introduced in section 3. Correspondingly, each stochastic trajectory in the AMF model is contracted to that in the simplified MJ model, where the transition rates are kept unchanged:

Equation (21)

Equation (22)

Equation (23)

Equation (24)

We denote as PcAMF(Γ ) and PMJ(Γ ) the probabilities that trajectory Γ from time 0 to T and its contraction are realized in the AMF model and the simplified MJ model, respectively. We then obtain

Equation (25)

where the factor 1/2 comes from the contraction of the initial state: If the initial state after the contraction is d (u), the initial state before the contraction is either (0, o) or (1, e) ((0, e) or (1, o)) with probability 1/2.

We next show that equation (19) directly reduces to equation (7), and correspondingly, inequality (20) reduces to inequality (9). Therefore, for the case of the simplified MJ model we can see the direct application of the MF formalism to an IR model without invoking the partial entropy production method. We first note that

Equation (26)

holds for $x={x}^{\prime },y\ne {y}^{\prime }$ or $x\ne {x}^{\prime },y={y}^{\prime }$, where every line is just zero if x = x', $y\ne {y}^{\prime }$. We then obtain

Equation (27)

for any trajectory Γ of the AMF model. Equality (27) implies that the two fluctuation theorems, equations (19) and (7), are equivalent, and thus inequalities (20) and (9) are also equivalent. This clarifies the physical meaning of inequality (9) from the viewpoint of the MF formalism: The bit sequence plays the equivalent role to the chemical fuel that drives the autonomous measurement and feedback.

7. Concluding remarks

We have established the fundamental relation between the MF and the IR formalisms. Inequality (9) derived from the partial entropy production method provides a stronger inequality than that derived from the IR formalism (12). Moreover, we have demonstrated that the contraction of the AMF model, to which the fluctuation theorem with the mutual information flow is applicable, is equivalent to the simplified MJ model. This allows us the direct comparison of the inequality with the mutual information flow and the MJ inequality.

We have also shown the hierarchy of the three second-law-like inequalities for the simplified MJ model: the conventional second law (11), the MJ inequality (12), and the inequality (9) with the partial entropy production. An important insight is that the more source of dissipation is taken into account, the stronger inequality we obtain. From this perspective, the reason why inequality (9) gives the strongest bound is the following: It specifies the entropy production only associated with the thermal transitions, and it takes into account the dissipation induced by the free relaxation of the bit. In general, the partial entropy production method would be useful to investigate stronger second-law-like inequalities for a variety of nonequilibrium dynamics.

In addition, we discuss a possible extension of our result to quantum cases. Recently, the refinement of quantum thermodynamics has been intensively studied from several perspectives [31, 3336]. These studies have revealed novel and fundamental aspects of the second law of thermodynamics. In addition, both of the MF formalism [37, 38] and the IR formalism [39] for quantum systems has already been discussed. Unifying these two formalisms in light of quantum thermodynamics would be an interesting future problem.

Acknowledgments

The authors thank Sosuke Ito and Kyogo Kawaguchi for fruitful discussions and helpful comments. NS is supported by Grant-in-Aid for JSPS Fellows Number 26–7602. TS is supported by JSPS KAKENHI Grant No. 25800217 and No. 22340114, by KAKENHI No. 25103003 'Fluctuation & Structure', and by Platform for Dynamic Approaches to Living System from MEXT, Japan.

Please wait… references are loading.
10.1088/1367-2630/18/1/013044