This site uses cookies. By continuing to use this site you agree to our use of cookies. To find out more, see our Privacy and Cookies policy.
Paper The following article is Open access

Device-independent test of causal order and relations to fixed-points

and

Published 1 April 2016 © 2016 IOP Publishing Ltd and Deutsche Physikalische Gesellschaft
, , Focus on Device Independent Quantum Information Citation Ämin Baumeler and Stefan Wolf 2016 New J. Phys. 18 035014 DOI 10.1088/1367-2630/18/3/035014

1367-2630/18/3/035014

Abstract

Bell non-local correlations cannot be naturally explained in a fixed causal structure. This serves as a motivation for considering models where no global assumption is made beyond logical consistency. The assumption of a fixed causal order between a set of parties, together with free randomness, implies device-independent inequalities—just as the assumption of locality does. It is known that local validity of quantum theory is consistent with violating such inequalities. Moreover, for three parties or more, even the (stronger) assumption of local classical probability theory plus logical consistency allows for violating causal inequalities. Here, we show that a classical environment (with which the parties interact), possibly containing loops, is logically consistent if and only if whatever the involved parties do, there is exactly one fixed-point, the latter being representable as a mixture of deterministic fixed-points. We further show that the non-causal view allows for a model of computation strictly more powerful than computation in a world of fixed causal orders.

Export citation and abstract BibTeX RIS

Original content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

1. Introduction

Device-independent tests of assumptions depend only on some input-output behaviour, whereas the 'internals' of the concrete physical systems, as well as the used devices are ignored. Examples of such assumptions are free randomness, the impossibility of superluminal signaling, or predefined causal structures [14]. A combination of basic assumptions leads to inequalities composed of probabilities to observe certain outcomes. Whenever a theory or an experiment violates a device-independent inequality, we can apply the contrapositive, and thus are forced to drop at least one of the assumptions made. One of the most prominent findings of this line of reasoning is Bell-nonlocality [5], which is not only of fundamental interest but also leads to a series of device-independent protocols in cryptography [612].

1.1. Historical background on space, time, and causality

The debate about as how fundamental space and time are to be seen has a long history in within natural philosophy. In pre-Socratic time, the opposite standpoints on the question have arisen in the views of Parmenides as opposed to Heraclitus: for the latter, the stage set by a fundamental space–time structure is where the play of permanent change—for him synonymous to existence—happens. Parmenides' world view, on the other hand, is static and such that space and, in particular, time emerge only subsequently and only subjectively. The described opposition can be seen as a predecessor of the famous debate between Newton and Leibniz [13], centuries later. Whereas Newton starts from an initially given and static space–time, Leibniz was criticising that view: space, for instance, is for him merely relational and not absolute. The course of occidental science decided to go for Newton's (overly successful) picture, until Leibniz' relational view was finally adopted by Mach. Indeed, Mach's principle states that inertial forces are purely relational, and it was the crystallisation point of Einstein's general relativity although the latter did, in the end, not satisfy the principle; however, it does propose a dynamic space–time structure (still absolute, though) in which, additionally, space and time become closely intertwined where they have been seen independently in Newton's picture.

The absolute space–time resulting in general relativity has been challenged by the following observations. First, Gödel [14] showed the possibility of solutions to the general-relativistic field equations corresponding to closed space–time curves. Second, Bell's non-local correlations [5] do not seem to be well explainable according to Reichenbach's principle [15], stating that any correlation between two space–time events in a causal structure must be due to a common cause or a direct influence from one event to the other. These facts can serve as a motivation to drop the causal structure as being fundamental in the first place. That idea, as sketched here, is not as novel as it may seem to be. In 1913, Russell [16] wrote that 'the law of causality $[\ldots ]$ is a relic of a bygone age, surviving, like the monarchy, only because it is erroneously supposed to do no harm'. In the view adopted here, the causal space–time structure arises only together with instead of prior to the pieces of classical information coming to existence, e.g., in the context of a quantum-measurement process3 . A consequence of taking that standpoint is that the definition by Colbeck and Renner [17] of freeness of randomness in terms of causal structure can be turned around: may the free random bit be the fundamental concept from which the usual space–time causal structure emerges?

1.2. Results

This article studies causal inequalities that are derived from the two basic assumptions of free randomness and predefined causal structures, and their violations. The inequalities are obeyed by input–output behaviours between multiple parties that are consistent with a predefined causal structure. Any violation of such inequalities forces us to drop at least one of the two assumptions. Indeed, in section 2 we define causal order based on free randomness; in that perspective at least, a fundamental causal structure seems unnecessary. In theory, if one drops the assumption of a global time and adheres to logical consistency only, such inequalities can be violated—quantumly (see section 3) as well as classically (see section 4). Thus, we are lead to the statement: the assumption of a predefined causal structure is not a logical necessity.

In section 5 we show a relation between logical consistency and the uniqueness of fixed-points in functions. By pushing this line of research further, we obtain a new class of circuits that are logically consistent, yet where the gates can be connected arbitrarily (see section 6). Such circuits describe a new model of computational that is strictly more powerful than the standard circuit model. We conclude the work with a list of open questions.

1.3. Related work

Hardy [18, 19] challenged the notion of a global time in quantum theory. His main motivation is to reconcile quantum theory with general relativity. While indeterminism, a feature of the former theory, is absent in the latter, the latter theory is more general compared to the former as its space–time is dynamic. Thus, a theory that is probabilistic and has a dynamic space–time is a reasonable candidate for quantum gravity. Chiribella et al [20] and Chiribella et al [21] introduced the notion of 'quantum combs', which are higher-order transformations, e.g. transformations from operations to operations. An interesting feature of these 'quantum combs' is that they allow for superpositions of causal orders [22]. Such superpositions have lead to a computational advantage in certain tasks [2225]. In general, 'quantum combs' can also describe resources beyond superpositions of causal orders. Another framework to study such resources was introduced by Oreshkov et al [26] (see also [27]).

2. Definitions of causal relations and orders

We describe causal relations between random variables, and in a next step, between parties. Since we are interested in theories where causal structures are not fundamental, we distinguish between input and output random variables such that a structure emerges from these notions.

(Input and output).

Definition 1 An input random variable carries a hat, e.g., $\hat{A}$. The distribution over input random variables does not need to be specified. An output random variable is denoted by a single letter, e.g., X. Output random variables come with a distribution that is conditioned on the inputs, e.g., ${P}_{X| \hat{A}}$.

This difference allows us to define causal future and causal past for random variables.

(Causal future and causal past).

Definition 2 Let $\hat{A}$ be an input and X be an output random variable. If and only if $\hat{A}$ is correlated with X, i.e., $\exists {P}_{\hat{A}}\;:\;{P}_{\hat{A}}{P}_{X}\;\ne \;{P}_{\hat{A}}{P}_{X| \hat{A}}$, then $\hat{A}$ is said to be in the causal past of X and, equivalently X is said to be in the causal future of $\hat{A}$. Furthermore, $\hat{A}$ is called the cause and X is called the effect. This relation is denoted by $\hat{A}\preceq X$.

Such a definition follows the interventionists' approach to causality, e.g., as defined by Woodward. The intuition behind this definition is that we are allowed to manipulate only certain physical systems. If such a manipulation influences another physical system, then the former manipulation causes the latter (see figure 1). This allows us to derive the causal structure from observed correlations—the causal structure emerges from the correlations. Note that this definition does not allow an input to be an effect, and does not allow an output to be a cause. For further studies, we define parties and causal relations on parties.

Figure 1.

Figure 1. If the input random variable $\hat{A}$ is manipulated and the output random variable X is correlated to $\hat{A}$, then $\hat{A}$ is the cause of X.

Standard image High-resolution image
(Party).

Definition 3 A party $S=(\hat{A},X,{\boldsymbol{ \mathcal E }})$ with ${\boldsymbol{ \mathcal E }}\;:\hat{A}\times {I}_{{\rm{S}}}\to X\times {O}_{{\rm{S}}}$, consists of an input random variable $\hat{A}$, an output random variable X, and a map ${\boldsymbol{ \mathcal E }}$ that maps $\hat{A}$ together with a physical system that S receives from the environment to X and a physical system that is returned to the environment (see figure 2). A party interacts at most once with the environment, where we consider one reception and one transmission of a system as a single interaction.

Now, we can define causal relations between parties.

Figure 2.

Figure 2. Party $S=(\hat{A},X,{\boldsymbol{ \mathcal E }})$ with ${\boldsymbol{ \mathcal E }}\;:\hat{A}\times {I}_{{\rm{S}}}\to X\times {O}_{{\rm{S}}}$, where IS is a physical system obtained from the environment and OS is a physical system returned to the environment.

Standard image High-resolution image
(Causal future and causal past for parties).

Definition 4 Let $R=(\hat{A},X,{\boldsymbol{ \mathcal E }})$ and let $S=(\hat{B},Y,{\boldsymbol{ \mathcal F }})$ be two parties. We say R is in the causal past of S or S is in the causal future of R if and only if $\hat{A}$ is correlated with Y and $\hat{B}$ is uncorrelated with X. This relation is denoted by $R\preceq S$.

2.1. Predefined versus indefinite causal order

In this work we distinguish between predefined and indefinite causal orders. In particular, we will show that if one defines causality based on the inputs and outputs (see definition 2), then correlations that are not compatible with a predefined causal order could arise.

(Compatibility with two-party predefined causal order).

Definition 5 Let $R=(\hat{A},X,{\boldsymbol{ \mathcal E }})$ and $S=(\hat{B},Y,{\boldsymbol{ \mathcal F }})$ be two parties. A conditional probability distribution ${P}_{X,Y| \hat{A},\hat{B}}$ is called consistent with two-party predefined causal order if and only if the conditional probability distribution can be written as a convex combination of the orderings $R\preceq S$ and $S\preceq R$ i.e.

for some $0\leqslant p\leqslant 1$. The first term in the sum represents $R\preceq S$ and the second term represents $S\preceq R$.

This definition determines a polytope of probability distributions ${P}_{X,Y| \hat{A},\hat{B}}$ that are compatible with two-party predefined causal order. All facets for binary random variables were recently enumerated [28].

For three parties or more, a distribution that is compatible with a predefined causal order can be more general than a distribution that can be written as a convex combination of all orderings. The reason for this is that a party in the causal past of some other parties could in principle influence everything which lies in its causal future—therefore, it could also influence the causal order of the parties in its causal future.

(Compatibility with three-party predefined causal order).

Definition 6 Consider the three parties $R=(\hat{A},X,{\boldsymbol{ \mathcal E }})$$S=(\hat{B},Y,{\boldsymbol{ \mathcal F }})$, and $T=(\hat{C},Z,{\boldsymbol{ \mathcal G }})$. A conditional probability distribution ${P}_{X,Y,Z| \hat{A},\hat{B},\hat{C}}$ is called consistent with predefined causal order if and only if the probability distribution can be written as a convex combination of all orderings where one party Q is in the causal past of the other two, and where the causal order of these two parties is determined by Q.

A generalised version of definition 6 to any number of parties can be found in [29].

Lemma 1. [30] (Necessary condition for predefined causal order) A necessary condition for predefined causal order is that at least one party is not in the causal future of any party.

If a conditional distribution is incompatible with predefined causal order, then we call it indefinite.

(One-way signaling).

Example 1 Let $R=(\hat{A},X,{\boldsymbol{ \mathcal E }})$ and $S=(\hat{B},Y,{\boldsymbol{ \mathcal F }})$ be two parties. The probability distribution over binary random variables

is compatible with predefined causal order, because it can be written as

with

and

(Two-way signaling).

Example 2 Let $R=(\hat{A},X,{\boldsymbol{ \mathcal E }})$ and $S=(\hat{B},Y,{\boldsymbol{ \mathcal F }})$ be two parties. The probability distribution over binary random variables

is incompatible with predefined causal order (has an indefinite causal order), because it cannot be written as described in definition 5.

3. Assuming quantum theory locally

In 2012, Oreshkov et al [26] introduced the process-matrix framework for quantum correlations without predefined causal order. In that framework, a party $R=(\hat{A},X,{\boldsymbol{ \mathcal E }}\;:\hat{A}\times {I}_{{\rm{R}}}\to X\times {O}_{{\rm{R}}})$ receives a quantum state on the Hilbert space IR from the environment and returns a quantum state on the Hilbert space OR to the environment. The map ${\boldsymbol{ \mathcal E }}$ is completely positive, because we assume the validity of quantum theory within the laboratories of every party. Let4  ${{\boldsymbol{ \mathcal E }}}_{x,a}$ be the corresponding Choi–Jamiołkowski [31, 32] map which is an element in ${I}_{{\rm{R}}}\otimes {O}_{{\rm{R}}}$, where a is the input value and x is the output value. Let $S=(\hat{B},Y,{\boldsymbol{ \mathcal F }}\;:\hat{B}\times {I}_{{\rm{S}}}\to Y\times {O}_{{\rm{S}}})$ be another party with Choi–Jamiołkowski map ${{\boldsymbol{ \mathcal F }}}_{y,b}$. The most general probability distribution ${P}_{X,Y| \hat{A},\hat{B}}$ that is linear in the local operations is

where W is a matrix living in the Hilbert space ${I}_{{\rm{R}}}\otimes {O}_{{\rm{R}}}\otimes {I}_{{\rm{S}}}\otimes {O}_{{\rm{S}}}$. The matrix W is called process matrix.

Since ${P}_{X,Y| \hat{A},\hat{B}}$ is designed to be a probability distribution, and since both parties can arbitrarily choose their local operations ${\boldsymbol{ \mathcal E }}$ and ${\boldsymbol{ \mathcal F }}$, the following two conditions must be satisfied:

Equation (1)

Equation (2)

(Logically consistent process matrix).

Definition 7 We call a process matrix W logically consistent if and only if W satisfies the conditions (1) and (2).

Condition (1) implies that W must be a completely positive trace-preserving map from ${O}_{{\rm{R}}}\otimes {O}_{{\rm{S}}}$ to ${I}_{{\rm{R}}}\otimes {I}_{{\rm{S}}}$. Therefore, we can interpret W as a quantum channel (see figure 3). If we assumed a global causal structure (what we do not), then W would be a back-in-time channel. Rather, it can be interpreted as the environment which lies outside space–time—the causal structure is designed by W and emerges from the correlations, as will become clear later. First, we can understand W as a generalised notion of a state and a channel: whereas it describes a quantum state in example 3, it models a quantum channel in example 4.

Figure 3.

Figure 3. Two-party process matrix W as a channel from the Hilbert space ${O}_{{\rm{R}}}\otimes {O}_{{\rm{S}}}$ to ${I}_{{\rm{R}}}\otimes {I}_{{\rm{S}}}$.

Standard image High-resolution image

(Representation of a quantum state).

Example 3 This logically consistent process matrix

describes the quantum state ρ that is sent to both parties.

(Representation of a quantum channel).

Example 4 This logically consistent process matrix

with $| {\rm{\Psi }}\rangle =(| 0,0\rangle +| 1,1\rangle )/\sqrt{2}$, describes a qubit channel from party R to party S.

Besides quantum states and quantum channels, a logically consistent process matrix can also describe superpositions of quantum channels. For instance, we could have a channel from R to S to T superposed with a channel from S to R to T, where $T=(\hat{C},Z,{\boldsymbol{ \mathcal G }}\;:\hat{C}\times {I}_{{\rm{T}}}\times {I}_{{{\rm{T}}}^{\prime }}\to Z)$ is a party which does not return a system to the environment, and receives two systems from the environment (the target on IT and the control on ${I}_{{{\rm{T}}}^{\prime }}$) (see example 5).

(Superposition of channels).

Example 5 This logically consistent process matrix

with

describes a superposition of a channel from R to S to T and a channel from S to R to T.

The process matrix from example 5 can be used to solve certain tasks more efficiently. Suppose you are given two black boxes ${\rm{B}}$ and ${\rm{C}}$ that act on qubits, and you are guaranteed that ${\rm{B}}$ and ${\rm{C}}$ either commute or anti-commute. In the standard circuit model, you would need to query each box twice in order to determine the (anti)commutativity. Whereas by using the process matrix Wsuperposedchannel, a single query suffices [2224, 33, 34].

3.1. Non-causal process matrices

While the above examples of process matrices are compatible with predefined causal order, some logically consistent process matrices lead to correlations that cannot be obtained in a world with a predefined causal ordering of the parties—such process matrices are called non-causal.

(Causal and non-causal process matrices).

Definition 8 A process matrix W is called causal if and only if for any choice of operations of the parties the resulting probability distribution is compatible with predefined causal order (according to definition 6). Otherwise, it is called non-causal.

To show that a process matrix is non-causal, we define a game, give an upper bound on the winning probability for this game under the assumption of a predefined causal order, and show that this bound can be violated in the process-matrix framework.

Game 1. [26] (Two-party non-causal game) Let $R=(\hat{A},X,{\boldsymbol{ \mathcal E }})$ and $S=((\hat{B},{\hat{B}}^{\prime }),Y,{\boldsymbol{ \mathcal F }})$ be two parties that aim at maximising

where all random variables are binary and where all inputs are uniformly distributed. Informally, if ${b}^{\prime }=0$, then party S is asked to send her input to R, otherwise, party R is asked to send her input to S.

We give an upper bound on the success probability of game 1 under the assumption of a predefined causal order.

Theorem 1. [26] (Upper bound on success probability of game 1) Under the assumption of a predefined causal order, game 1 can at best be won with probability 3/4.

Proof. By lemma 1, at least one party is not in the causal future of any party. Without loss of generality, let R be this party. Then, R can only make a random guess, which means $\mathrm{Pr}(X=\hat{B}\;| \;{b}^{\prime }=0)=1/2$. Therefore, we obtain the upper bound

Then again, this upper bound given by theorem 1 can be violated in the process-matrix framework.

Theorem 2. [26] game 1 can be won with probability $(2+\sqrt{2})/4$ if we drop the assumption of a global causal order.

The logically consistent process matrix

where we omit the ⨂ symbols for better presentation, and the local operations

for an arbitrary ρ, can be used to violate the bound given by theorem 1 up to $(2+\sqrt{2})/4$ [26]. Brukner [35] proved under certain assumptions that within the process-matrix framework, this violation cannot be exceeded.

This raised the question whether non-causal logically consistent process matrices in the classical realm also exist—which was answered negatively in the two-party case [26, 36]. However, for three parties or more, this is not true anymore; this is shown in section 4.

Logically consistent process matrices can be understood as a new resource for quantum operations. Oddly enough, these resources cannot be composed: if we take two logically consistent process matrices W1 and W2, where the former is a quantum channel from R to S and the latter is a quantum channel from S to R, then ${W}_{1}\otimes {W}_{2}$ is not logically consistent; R could in principle alter the state on IR she receives from the environment, leading to a causal paradox. The impossibility of composing process matrices within the model reflects the fact that a process matrix is supposed to describe the environment as a whole.

4. Assuming classical probability theory locally

The classical analogue of the process-matrix framework was recently developed [30], and has been found to give rise to non-causal correlations for three parties or more. It is the classical analogue in the sense that, instead of assuming the validity of quantum theory locally, classical probability theory is assumed to hold locally. Let $R=(\hat{A},X,{\boldsymbol{ \mathcal E }}\;:\hat{A}\times {I}_{{\rm{R}}}\to X\times {O}_{{\rm{R}}})$ be a party. Here, the spaces IR and OR are not Hilbert spaces (as in the process-matrix framework), but describe random variables. Therefore, the local operation ${\boldsymbol{ \mathcal E }}$ of a party is a conditional probability distribution ${P}_{X,{O}_{{\rm{R}}}| \hat{A},{I}_{{\rm{R}}}};$ locally we assume the validity of probability theory as opposed to quantum theory. Let $S=(\hat{B},Y,{\boldsymbol{ \mathcal F }})$ with ${\boldsymbol{ \mathcal F }}\;:\hat{B}\times {I}_{{\rm{S}}}\to Y\times {O}_{{\rm{S}}}$ and $T=(\hat{C},Z,{\boldsymbol{ \mathcal G }})$ with ${\boldsymbol{ \mathcal G }}\;:\hat{C}\times {I}_{{\rm{T}}}\to Z\times {O}_{{\rm{T}}}$ be two further parties. The most general probability distribution ${P}_{X,Y,Z,{I}_{{\rm{R}}},{I}_{{\rm{S}}},{I}_{{\rm{T}}},{O}_{{\rm{R}}},{O}_{{\rm{S}}},{O}_{{\rm{T}}}| \hat{A},\hat{B},\hat{C}}$ that is linear in the local operations is

Equation (3)

where E is called classical process. Since all three parties can arbitrarily choose their local operations ${\boldsymbol{ \mathcal E }}$${\boldsymbol{ \mathcal F }}$, and ${\boldsymbol{ \mathcal G }}$, the following conditions must be satisfied (we use ${\boldsymbol{i}}$ as shorthand expression for $({i}_{{\rm{R}}},{i}_{{\rm{S}}},{i}_{{\rm{T}}})$, likewise for ${\boldsymbol{o}}$${\boldsymbol{I}}$, and ${\boldsymbol{O}}$):

Equation (4)

Equation (5)

(Logically consistent clasical process).

Definition 9 We call a classical process E logically consistent if and only if E satisfies the conditions (4) and (5).

Condition (4) implies that the classical process E is a conditional probability distribution ${P}_{{I}_{{\rm{R}}},{I}_{{\rm{S}}},{I}_{{\rm{T}}}| {O}_{{\rm{R}}},{O}_{{\rm{S}}},{O}_{{\rm{T}}}}$. Therefore, E can be interpreted in the same way as process matrices: it maps the systems OROS, and OT to IRIS, and IT (see figure 4).

Figure 4.

Figure 4. Three-party classical process E as a channel from $({O}_{{\rm{R}}},{O}_{{\rm{S}}},{O}_{{\rm{T}}})$ to $({I}_{{\rm{R}}},{I}_{{\rm{S}}},{I}_{{\rm{T}}})$.

Standard image High-resolution image

We rewrite the conditions (4) and (5) by using stochastic matrices. This helps to check whether a classical process is logically consistent or not. We write ${\boldsymbol{E}}$ to denote the stochastic matrix that models the classical process E. Then, the non-negativity condition (4) becomes

Equation (6)

where ${{\boldsymbol{E}}}_{m,n}$ are the matrix elements. For simplicity, we fix all inputs $(\hat{A},\hat{B},\hat{C})$ to $(a,b,c)$ and consider the operations ${{\boldsymbol{ \mathcal E }}}^{\prime }\;:{I}_{{\rm{R}}}\to {O}_{{\rm{R}}}$ only, and likewise for the other parties. Let ${{\boldsymbol{ \mathcal E }}}^{\prime }$ be the corresponding stochastic matrix. The probability that the operation ${{\boldsymbol{ \mathcal E }}}^{\prime }$ produces oR conditioned on iR is

where ${{\boldsymbol{o}}}_{{\boldsymbol{R}}}$ is the stochastic vector that models oR, e.g., for a binary random variable OR the value oR = 0 is modelled by

Having this, we rewrite condition (5) as

Indeed, since any local operation ${{ \mathcal E }}^{\prime }$ can be written as a convex combination of deterministic operations, the total-probability condition

Equation (7)

where ${\boldsymbol{ \mathcal D }}$ is the set of all deterministic local operations, is sufficient.

4.1. Non-causal classical processes

In analogy to the process-matrix framework, we can define non-causal classical processes.

(Non-causal classical processes).

Definition 10 A classical process E is called non-causal if and only if there exists a choice of operations of the parties such that the resulting probability distribution does not satisfy the necessary condition for predefined causal order (see lemma 1).

It is known that for two parties, non-causal logically consistent classical processes do not exist [26, 36]. For three parties or more, however, such processes do exist [30, 37]. We describe two non-causal logically consistent classical processes.

Game 2.  [37] (Three-party non-causal) Let $R=((\hat{A},\hat{M}),X,{\boldsymbol{ \mathcal E }})$$S=((\hat{B},\hat{M}),Y,{\boldsymbol{ \mathcal F }})$, and $T=((\hat{C},\hat{M}),Z,{\boldsymbol{ \mathcal G }})$ be three parties that aim at maximising

where all random variables but $\hat{M}$ are binary, and where $\hat{M}$ is a shared ternary random variable. All input random variables are uniformly distributed. Informally, if the shared random variable $\hat{M}$ takes value 0, then party R has to guess the parity of the inputs of S and T, and likewise for the alternative values the random variable $\hat{M}$ can take.

We give an upper bound on the success probability of game 2 under the assumption of a predefined causal order.

Theorem 3. [37] In a predefined causal order, game 2 can at best be won with probability 5/6.

Proof. A necessary condition for predefined causal order is that at least one party is not in the causal future of any party (see lemma 1). Without loss of generality, let R be this party. Whenever the shared trit $\hat{M}$ takes the value 0, party R can give a random guess only. Therefore, $\mathrm{Pr}(X=\hat{B}\oplus \hat{C}\;| \;m=0)=1/2$. This gives the upper bound

Theorem 4. [37] game 2 can be won perfectly if we drop the assumption of a global causal order.

Consider the logically consistent classical process E

This process is the uniform mixture of the identity channel from R to S to T to R and the bit-flip channel from R to S to T to R (see figure 5). By combining it with the local operations

Game 2 is won perfectly. The operations can be understood as follows. The party who has to make the guess simply uses the system that she receives from the environment as guess. If the party who has to produce a guess is the next in the chain (R to S to T to R) from the point of view of the actual party, then the actual party returns the parity of the system she obtained from the environment and her input. In the third case, if the party who has to guess is two steps ahead in the chain, then the actual party returns her input to the environment.

Figure 5.

Figure 5. Logically consistent classical process used to perfectly win game 2.

Standard image High-resolution image

Game 3.  [30] (Three-party non-causal game) Let $R=(\hat{A},X,{\boldsymbol{ \mathcal E }})$$S=(\hat{B},Y,{\boldsymbol{ \mathcal F }})$, and $T=(\hat{C},Z,{\boldsymbol{ \mathcal G }})$ be three parties that aim at maximising

where all random variables are binary, and where $\mathrm{maj}(a,b,c)$ is the majority of $(a,b,c)$. All inputs are uniformly distributed. In words, if the majority of the inputs is 0, then the parties have to guess their neighbour's input in one direction. Otherwise, if the majority of the inputs is 1, then the parties have to guess their neighbour's inverted input in the opposite direction.

We give an upper bound on the success probability of game 3 under the assumption of a predefined causal order.

Theorem 5.   [30] Under the assumption of a predefined causal order, game 3 can at best be won with probability 3/4.

Proof. The truth table of game 3 is given in table 1. A necessary condition for predefined causal order is that at least one party is not in the causal future of any other party (see lemma 1). Without loss of generality, let R be this party. Therefore, party R has to output X without learning any other random variable than $\hat{A}$. By inspecting table 1, we see that X = 0 is R's best guess. Thus, in at least two out of the eight cases, the game is lost, resulting in ${p}_{\mathrm{succ}3}\leqslant 3/4$.

Table 1.  Conditions for winning game 3.

$\hat{A}$ $\hat{B}$ $\hat{C}$ $\mathrm{maj}(\hat{A},\hat{B},\hat{C})$ X Y Z
0 0 0 0 0 0 0
0 0 1 0 1 0 0
0 1 0 0 0 0 1
0 1 1 1 0 0 1
1 0 0 0 0 1 0
1 0 1 1 1 0 0
1 1 0 1 0 1 0
1 1 1 1 0 0 0

Theorem 6.   [30] By using the logically consistent classical process framework, game 3 can be won perfectly.

The logically consistent classical process E to perfectly win game 3 is

The local operations are

A graphical representation of the classical process is given in figure 6.

Figure 6.

Figure 6. Logically consistent classical process used to win game 3.

Standard image High-resolution image

4.2. Geometric representation

The set of logically consistent classical processes forms a polytope [30], which is defined by the linear conditions (6) and (7). Both classical processes used to perfectly win games 2 and 3 are extremal points of the mentioned polytope. In contrast to the process of game 3, the classical process of game 2 is a mixture of logically inconsistent processes. Such a classical process must be fine-tuned: tiny variations of the probabilities make the classical process logically inconsistent. This motivates the definition of a smaller polytope, where all extremal points represent deterministic classical processes. This smaller polytope is called deterministic-extrema polytope. Qualitative representations of both polytopes are given in figure 7.

Figure 7.

Figure 7. Qualitative representation of the polytopes discussed in the article. The outer polytope consists of all logically consistent classical processes. The middle polytope is the same as the outer except that all probabilistic extremal points are omitted. The innermost polytope is the polytope of all causal classical processes.

Standard image High-resolution image

5. Characterising logical consistency with fixed-points

The above considerations on classical processes allow us to characterise logically consistent classical process via functions with unique fixed-points. For that purpose, we redefine party $R=({\boldsymbol{ \mathcal A }},{\boldsymbol{ \mathcal X }},{\boldsymbol{f}})$ with ${\boldsymbol{f}}\;:{\boldsymbol{ \mathcal A }}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}\to {\boldsymbol{ \mathcal X }}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}$ where ${\boldsymbol{ \mathcal A }}$${\boldsymbol{ \mathcal X }}$${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$, and ${{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}$ are sets. Similarly, we redefine $S=({\boldsymbol{ \mathcal B }},{\boldsymbol{ \mathcal X }},{\boldsymbol{g}})$ with ${\boldsymbol{g}}\;:{\boldsymbol{ \mathcal B }}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}\to { \mathcal Y }\times {{\boldsymbol{ \mathcal O }}}_{{\rm{S}}}$ and $T=({\boldsymbol{ \mathcal C }},{\boldsymbol{ \mathcal Z }},{\boldsymbol{h}})$ with ${\boldsymbol{h}}\;:{\boldsymbol{ \mathcal C }}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}\to {\boldsymbol{ \mathcal Z }}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{T}}}$. Further, let $f\;:{{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}\to {{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}$$g\;:{{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}\to {{\boldsymbol{ \mathcal O }}}_{{\rm{S}}}$, and $h\;:{{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}\to {{\boldsymbol{ \mathcal O }}}_{{\rm{T}}}$ be functions.

Theorem 7.  (Deterministic extremal points modelled by functions with unique fixed-point) A function ${\boldsymbol{e}}\;:{{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{T}}}\to {{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$ represents an extremal point of the deterministic-extrema polytope if and only if

where $\exists !$ is the uniqueness quantifier.

Proof. Recall the total-probability condition (7) that we express using stochastic matrices

where ${\boldsymbol{ \mathcal D }}$ is the set of all deterministic local operations. By definition, a classical process E that is an extremal point of the deterministic-extrema polytope is deterministic. This implies that the matrix

for some ${{\boldsymbol{ \mathcal E }}}^{\prime },{{\boldsymbol{ \mathcal F }}}^{\prime },{{\boldsymbol{ \mathcal G }}}^{\prime }\in {\boldsymbol{ \mathcal D }}$, is a stochastic matrix with trivial probabilities. The diagonal ${\boldsymbol{q}}$ of ${\boldsymbol{M}}$ consists of 0's and exactly one 1. In particular, by applying the stochastic matrix ${\boldsymbol{M}}$ to ${\boldsymbol{q}}$ results in ${\boldsymbol{q}}$, i.e., ${\boldsymbol{M}}{\boldsymbol{q}}={\boldsymbol{q}}$, which means that ${\boldsymbol{q}}$ is a unique fixed-point of ${\boldsymbol{M}}$.

Theorem 7 implies that deterministic causal and non-causal classical processes can be expressed as functions with one and only one fixed-point if the former is composed with arbitrary deterministic local operations. Such a function for a causal classical process is given in example 6 and for a non-causal classical process in example 7.

(Causal classical process as a function).

Example 6 Let the sets ${\boldsymbol{ \mathcal A }}$${\boldsymbol{ \mathcal X }}$${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$, and ${{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}$, and likewise for the other two parties, be $\{0,1\}$. Consider the causal classical process E with an identity channel from party R to S and an identity channel from party S to T, and where R receives the element 0 from the environment. We express E as a function ${\boldsymbol{e}}\;:{{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{T}}}\to {{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$:

Since the sets ${{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}$ and ${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$, and likewise for the other parties, are binary, the local operation of a party is one out of four functions:

For any choice $f,g,h$ of the local operations, the function

has a unique fixed-point. We discuss a few cases. First, let $f=g=h={d}_{\mathrm{id}}$. Then, the fixed-point is $(0,0,0)$, as can be verified by table 2, which describes the composed function ${\boldsymbol{e}}\circ ({d}_{\mathrm{id}},{d}_{\mathrm{id}},{d}_{\mathrm{id}})$. In the case $f=g={d}_{\mathrm{id}}$ and $h={d}_{\mathrm{not}}$, the fixed-point is $(0,0,0)$ as well. In the case $f=g=h={d}_{\mathrm{not}}$, the fixed-point is $(0,1,0)$. A final case we express explicitly is $f=g=h={d}_{1};$ the fixed-point is $(0,1,1)$.

(Non-causal classical process as a function).

Example 7 We use the same parties as they are described in example 6. Take the logically consistent classical process that can be used to win game 3. A graphical representation of this classical process is given in figure 6. We rewrite the process as a function ${\boldsymbol{e}}\;:{{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{T}}}\to {{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$:

If the parties use the identity as local operations, then ${\boldsymbol{e}}\circ ({d}_{\mathrm{id}},{d}_{\mathrm{id}},{d}_{\mathrm{id}})$ has a unique fixed-point $(0,0,0)$, as can be verified by inspecting table 3. For the local operations $f=g=h={d}_{\mathrm{not}}$, the unique fixed-point is $(0,0,0)$ as well. In another case, if the local operations are $f=g={d}_{\mathrm{id}}$ and $h={d}_{\mathrm{not}}$, then the unique fixed-point is $(1,0,0)$.

We present a statement similar to theorem 7, yet for any logically consistent classical process (probabilistic or deterministic).

Table 2.  Classical process from example 6 as a function composed with the identity as local operations of all three parties.

${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$
0 0 0 0 0 0
0 0 1 0 0 0
0 1 0 0 0 1
0 1 1 0 0 1
1 0 0 0 1 0
1 0 1 0 1 0
1 1 0 0 1 1
1 1 1 0 1 1

Table 3.  Classical process from example 7 as a function composed with the identity as local operations of all three parties.

${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$
0 0 0 0 0 0
0 0 1 1 0 0
0 1 0 0 0 1
0 1 1 0 0 1
1 0 0 0 1 0
1 0 1 1 0 0
1 1 0 0 1 0
1 1 1 0 0 0
(Logically consistent classical process and fixed-points).

Theorem 8 A conditional probability distribution closeup-->$E={P}_{{I}_{{\rm{R}}},{I}_{{\rm{S}}},{I}_{{\rm{T}}}| {O}_{{\rm{R}}},{O}_{{\rm{S}}},{O}_{{\rm{T}}}}$openup-->is a logically consistent classical process if and only if one of the decomposition closeup-->$E={\sum }_{i}{p}_{i}{D}_{i}$closeup-->, where closeup-->$\forall i\;:{D}_{i}$openup--> is a deterministic distribution, has the property

where ${{\boldsymbol{d}}}_{i}\;:{{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{T}}}\to {{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$ is a function modelling Di.

Proof. The classical process ${\boldsymbol{E}}$ can be decomposed as

where $\forall i\;:\;{D}_{i}$ is a deterministic distribution. The total-probability condition (7), that we express using stochastic matrices, is

where ${\boldsymbol{ \mathcal D }}$ is the set of all deterministic local operations, and where ${\boldsymbol{j}}$ is a stochastic vector with all 0's and exactly one 1. The expression

is thus the number of fixed-points of the stochastic matrix ${{\boldsymbol{D}}}_{i}({{\boldsymbol{ \mathcal E }}}^{\prime }\otimes {{\boldsymbol{ \mathcal F }}}^{\prime }\otimes {{\boldsymbol{ \mathcal G }}}^{\prime })$. From this we get ${\sum }_{i}{p}_{i}{n}_{i}=1$.

The informal statement of theorem 8 is: the average number of fixed-points of a logically consistent classical process is one. Clearly, the classical processes of examples 6 and 7 can be rewritten as a convex combination of deterministic processes fulfilling this property. However, there are classical processes which lie outside the deterministic-extrema polytope—theorem 7 does not apply to those—to which we can apply theorem 8. Theorems 7 and 8 can naturally be extend to more than three parties.

(Non-causal classical process as a mixture of functions).

Example 8 The classical process used to win game 2 is

A graphical representation of this classical process is given in figure 5. The classical process can be written as a convex combination $E=({E}_{0}+{E}_{1})/2$ with

Now, we switch to the parties as defined in example 6 and rewrite E0 and E1 as functions ${{\boldsymbol{e}}}_{0}\;:{{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{T}}}\to {{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$ and ${{\boldsymbol{e}}}_{1}\;:{{\boldsymbol{ \mathcal O }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal O }}}_{{\rm{T}}}\to {{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}\times {{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$:

For the choice $f=g=h={d}_{\mathrm{id}}$ of local operations, the function ${{\boldsymbol{e}}}_{0}\circ ({d}_{\mathrm{id}},{d}_{\mathrm{id}},{d}_{\mathrm{id}})$ has two fixed-points, and the function ${{\boldsymbol{e}}}_{1}\circ ({d}_{\mathrm{id}},{d}_{\mathrm{id}},{d}_{\mathrm{id}})$ has no fixed-points (see table 4). Thus, the average number of fixed-points is $(2+0)/2=1$. An alternative choice of local operations is $f=g={d}_{\mathrm{id}}$ and $h={d}_{\mathrm{not}}$. Then ${{\boldsymbol{e}}}_{0}\circ ({d}_{\mathrm{id}},{d}_{\mathrm{id}},{d}_{\mathrm{not}})$ has 0, and ${{\boldsymbol{e}}}_{1}\circ ({d}_{\mathrm{id}},{d}_{\mathrm{id}},{d}_{\mathrm{not}})$ has 2 fixed-points: $(0,1,0)$ and $(1,0,1)$. As a last choice, consider $f=g=h={d}_{0}$, in which both composed functions have 1 fixed-point: $(0,0,0)$ for the former function, and $(1,1,1)$ for the latter function.

6. A non-causal model for computation

Above we took the operational approach, where we think in terms of parties that can choose to perform arbitrary experiments within their laboratories. A model for computation can be designed if we depart from this point of view and rather think in terms of circuits. We briefly sketch such a model [38]. In the logical-consistency conditions of the process-matrix framework and its classical analogue, we consider any local operation of the parties. In a circuit, however, there are no parties. Thus, the logical-consistency condition has to be adopted. Let ${\boldsymbol{ \mathcal C }}$ be a classical circuit that consists of wires and gates.

Table 4.  The pair of functions that yield the classical process from example 8 when mixed, as functions composed with the identity as local operations. The left table describes the function ${{\boldsymbol{e}}}_{0}\circ ({d}_{\mathrm{id}},{d}_{\mathrm{id}},{d}_{\mathrm{id}})$, and the right table describes the function ${{\boldsymbol{e}}}_{1}\circ ({d}_{\mathrm{id}},{d}_{\mathrm{id}},{d}_{\mathrm{id}})$.

${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{R}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{S}}}$ ${{\boldsymbol{ \mathcal I }}}_{{\rm{T}}}$
0 0 0 0 0 0
0 0 1 1 0 0
0 1 0 0 0 1
0 1 1 1 0 1
1 0 0 0 1 0
1 0 1 1 1 0
1 1 0 0 1 1
1 1 1 1 1 1
0 0 0 1 1 1
0 0 1 0 1 1
0 1 0 1 1 0
0 1 1 0 1 0
1 0 0 1 0 1
1 0 1 0 0 1
1 1 0 1 0 0
1 1 1 0 0 0
(Logically consistent circuit).

Definition 11 A classical circuit ${\boldsymbol{ \mathcal C }}$ is called logically consistent if and only if there exists a probability distribution over all values on the wires where the probabilities are linear in the choice of the inputs.

This definition allows gates to be connected in a circular way—which is not allowed in the standard circuit model. The wires which are connected to a gate on one side only are either the inputs or the outputs of the circuit. Intuitively, if a closed circuit (when ignoring the open wires) admits a unique fixed-point, then it is logically consistent. The wires, in that case, carry the value of the fixed-point. This non-causal model for computation is more powerful than the standard circuit model. As a demonstration, consider the following task.

Task 1. (Fixed-point search) Suppose you are given a black-box ${\rm{B}}$ with the promise that ${\rm{B}}$ has exactly one fixed-point. The task is to find the fixed-point with as few queries to ${\rm{B}}$ as possible.

In the worst case, the minimal number of queries to ${\rm{B}}$ in the standard circuit model is $n-1$, where n is the input size.

(task 1 can be solved with one query to B ).

Theorem 9 By using this new model of computation, a single query to black box ${\rm{B}}$ suffices to solve task 1.

Proof. To solve the task, we use the logically consistent non-causal circuit given by figure 8 with a = 0. The stochastic matrix of the $\mathrm{CNOT}$ gate C is

and the stochastic matrix of the black-box ${\rm{B}}$ is

with

Since ${\rm{B}}$ is a black-box, the values ei are unknown to us. We write the probability of getting $(b,x,y)$ conditioned on the input a with stochastic matrices and stochastic vectors:

where ${\boldsymbol{x}}$ is the stochastic vector of the value x, as described on page , and where ${{\boldsymbol{b}}}^{{\rm{T}}}{\boldsymbol{B}}{\boldsymbol{y}}$ is the probability of obtaining b when y is given as input to ${\rm{B}}$. Now, if we use a = 0, we find the fixed-point x:

Figure 8.

Figure 8. Circuit to find the fixed-point x of ${\rm{B}}$.

Standard image High-resolution image

So far we ignore how to deal with black boxes that have an unknown number of fixed-points, and whether, in that case, a computational advantage is achievable at all.

7. Open questions

The main open question in this field of research is whether nature violates causal inequalities. Contrary to some superpositions of causal orders [34], no experiment has been proposed or carried out to violate such inequalities. Even though some results [35, 39] indicate a strong connection between non-causal correlations and non-local correlations—the maximally achievable violation of a two-party causal inequality in the quantum setting is $(2+\sqrt{2})/4$, which is the same value as Cirel'son's bound [40], and for three parties, the algebraic maximum is achievable [41, 42]—the connection remains unknown. A violation of a two-party causal inequality up to the algebraic maximum by the use of higher-dimensional systems is also unknown so far; such a result would strengthen the connection between both types of correlations and could give insight on how these correlations are connected. To our knowledge, it is also unknown which process matrices can be purified. Since purification is of great importance in quantum information [43], purification of process matrices is also desired. It has been suspected that the process matrices that can be purified cannot violate causal inequalities. However, a counterexample has been provided recently [44, 45]. The same was conjectured for classical processes. This has been refuted as well: for the classical case, it is known that classical processes from the deterministic-extrema polytope can be made reversible [44, 45]—this fact is the classical analogue of purification. Of great interest are also information-theoretic considerations of non-causal correlations, similar to [46, 47]. For instance, it might be possible to quantify the amount of 'superposition' versus the amount of 'loops' a process matrix carries. Finally, little is known on the computational power of process matrices, classical processes, and of our model of computation. While the D-CTC model [48] can efficiently solve problems that are in PSPACE [49], it is unknown what the complexity class is of the efficiently solvable problems in our model of computation.

Acknowledgments

We thank Mateus Araújo, Veronika Baumann, Časlav Brukner, Fabio Costa, Adrien Feix, Christina Giarmatzi, and Thomas Strahm for helpful discussions. Furthermore, we thank the anonymous referees for their detailed comments. The present work was supported by the Swiss National Science Foundation (SNF) and the National Centre of Competence in Research 'Quantum Science and Technology' (QSIT).

Footnotes

  • The key to understanding how exactly this happens is perhaps hidden in thermodynamics and a suitable interpretation of quantum theory.

  • Throughout this work we use bold letters for vectors and matrices.

Please wait… references are loading.
10.1088/1367-2630/18/3/035014