Brought to you by:
Paper The following article is Open access

The space of logically consistent classical processes without causal order

and

Published 14 January 2016 © 2016 IOP Publishing Ltd and Deutsche Physikalische Gesellschaft
, , Citation Ämin Baumeler and Stefan Wolf 2016 New J. Phys. 18 013036 DOI 10.1088/1367-2630/18/1/013036

1367-2630/18/1/013036

Abstract

Classical correlations without predefined causal order arise from processes where parties manipulate random variables, and where the order of these interactions is not predefined. No assumption on the causal order of the parties is made, but the processes are restricted to be logically consistent under any choice of the parties' operations. It is known that for three parties or more, this set of processes is larger than the set of processes achievable in a predefined ordering of the parties. Here, we model all classical processes without predefined causal order geometrically and find that the set of such processes forms a polytope. Additionally, we model a smaller polytope—the deterministic-extrema polytope—where all extremal points represent deterministic processes. This polytope excludes probabilistic processes that must be—quite unnaturally—fine-tuned, because any variation of the weights in a decomposition into deterministic processes leads to a logical inconsistency.

Export citation and abstract BibTeX RIS

Original content from this work may be used under the terms of the Creative Commons Attribution 3.0 licence. Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.

1. Motivation and main result

An assumption often made in physical theories, sometimes implicitly, is the existence of a global time. In particular, quantum theory is formulated with time as an intrinsic parameter. If one relaxes this assumption by requiring local validity of some theory and logical consistency only, then a larger set of correlations can be obtained, called correlations without predefined causal order. The processes that lead to such correlations are called processes without predefined causal order. Two motivations to study such correlations are quantum gravity and quantum non-locality. Quantum gravity motivates this research in the sense that on the one hand, relativity is a deterministic theory equipped with a dynamic spacetime; on the other hand, quantum theory is a probabilistic theory embedded in a fixed spacetime. This suggests that quantum gravity is relaxed in both aspects, i.e., it is a probabilistic theory equipped with a dynamic spacetime [1]. Quantum non-local correlations [24] motivate this study since the possibility of a satisfactory causal explanation [5] for such correlations is questionable [3, 613]. Dropping the notion of a global time or of an a priori spacetime—as has been suggested from different fields of research [1423]—dissolves this paradox. This can be achieved by defining causal relations based on free randomness (see figure 1) as opposed to defining free randomness based on causal relations [24, 25]. Such an approach gives a dynamic character to causality; causal connections are not predefined but are derived from the observed correlations.

Figure 1.

Figure 1. If the random variable A is an input (here, visualized by a knob), the random variable X is an output, and A is correlated to X, then A can signal to X which implies that X is in the causal future of A ($X\succeq A$).

Standard image High-resolution image

Relaxations of quantum theory where the assumption of a global time is dropped have recently been studied widely [1, 2645] (see [46] for a review). Our work follows the spirit of an operational quantum framework for such correlations developed by Oreshkov et al [31]. Some correlations appearing in their quantum framework—for two parties or more—cannot be simulated by assuming a predefined causal order of the parties. Such correlations are termed non-causal. Analogously to non-locality, non-causal correlations could be witnessed by violating so-called causal inequalities [31, 32, 35, 43]. All causal inequalities in the two-party scenario and for binary inputs and outputs are presented in [43]. In a previous work [35], we showed that in the classical limit of the quantum framework, i.e., if it is restricted to probability theory, classical non-causal correlations can arise as well. This result holds for three parties or more. In the present work we follow this path and give a representation of all classical—as opposed to quantum—processes without predefined causal order as polytopes. Such a representation helps in optimizing winning strategies for causal games [31, 43]—the optimization problem can be stated as a linear program—, and for finding new causal games.

First, we present the framework of classical correlations without predefined causal order. Then, we describe the polytope of processes that lead to such correlations implicitly and explicitly for scenarios with up to three parties and binary inputs and outputs. In the general case, we give an implicit description of the polytope. In addition, we construct the smaller polytope of classical processes without predefined causal order where all extremal points describe deterministic processes. We call this polytope the deterministic-extrema polytope. The processes from this polytope can be thought of as being 'more physical' in the sense that its extremal points are not proper mixtures of logically inconsistent processes [47], i.e., this set contains processes that can be written as a convex combination of deterministic ones from within the polytope only. Our motivation for this is that some proper mixtures need to be fine-tuned [13], i.e., tiny variations of the mixtures renders the processes logically inconsistent. The fine-tuned proper mixtures are the probabilistic extremal points of the larger polytope. A qualitative representation of these polytopes is given in figure 2.

Figure 2.

Figure 2. A qualitative representation of processes without predefined causal order studied in this work is given. The dashed region describes all processes that are achievable in a predefined causal order—it also forms a polytope [42, 43]. The polytope with the dashed–dotted lines is the polytope of processes without predefined causal order. The region in-between marked with the solid lines is the polytope of processes without predefined causal order restricted to deterministic extremal points.

Standard image High-resolution image

2. Modelling classical correlations without predefined causal order

2.1. Causality, predefined causal order, and a framework of classical correlations without predefined causal order

We describe an operational framework without global assumptions (other than logical consistency). Causal relations are defined as in the interventionists' approach to causality [48, 49]: Outputs can be correlated to inputs and inputs are manipulated freely (see figure 1). Defining causality based on free randomness is the converse approach to the one used in recent literature [24, 25]; there, free randomness is defined based on causal relations.

Definition 1. (Causality [35]). For two correlated random variables X and A, where X is an output and A is an input, i.e., A is chosen freely, we say that X is in the causal future of A, or equivalently, that A is in the causal past of X, denoted by $X\succeq A$ or $A\preceq X$. The negations of these relations are denoted by $\not\hspace{-2pt}{\succeq }$ and $\not\hspace{-2pt}{\preceq }$.

Consider N parties ${\{{S}_{j}\}}_{0\leqslant j\lt N}$, where party Sj has access to an input random variable Aj and generates an output random variable Xj. This allows us to causally order parties: if Aj is correlated to Xk, then Sj is in the causal past of Sk (${S}_{j}\preceq {S}_{k}$). To simplify the presentation, we write $\vec{X}=({X}_{0},...,{X}_{N-1})$ and likewise for $\vec{A}$$\vec{O}$, and $\vec{I}$.

Definition 2. (Two-party predefined causal order). A two-party predefined causal order is a causal ordering of party S with input A, output X, and party T with input B, output Y, such that the distribution ${P}_{X,Y| A,B}$ can be written as a convex combination of one-way signaling distributions

for some $0\leqslant p\leqslant 1$.

A definition for multi-party predefined causal order is given in [42]. Such a definition turns out to be more subtle since a party Sj in the causal past of some other parties ${\{{S}_{{\ell }}\}}_{L}$ can in principle influence everything in her causal future; in particular, Sj can influence the causal order of the parties ${\{{S}_{{\ell }}\}}_{L}$. We just state a lemma that follows from such a definition and that is sufficient to prove our claims.

Lemma 1. (Necesarry condition for predefined causal order). A necessary condition for a predefined causal order is that the probability distribution ${P}_{\vec{X}| \vec{I}}$ can be written as a convex combination

with ${\sum }_{k}{p}_{k}=1$ and $\forall k\;:{p}_{k}\geqslant 0$, such that in every distribution Pk at least one party is not in the causal future of any other party, i.e.

where ${\not\hspace{-2pt}{\succeq }}^{{P}_{k}}$ stands for the causal relation that is deduced from the distribution Pk.

In the framework without predefined causal order, each party Sj receives a random variable Ij from the environment E on which Sj can act. After the interaction with Ij, party Sj outputs a random variable Oj to the environment. Both random variables Ij and Oj are output random variables. The only input random variable a party has is Aj. The operation of Sj is a stochastic process mapping ${A}_{j},{I}_{j}$ to ${X}_{j},{O}_{j}$ (see figure 3). A stochastic process is a probability distribution over the range conditioned on the domain; in this case, the stochastic process of party Sj (which in the following will also be called the local operation of party Sj) is ${P}_{{X}_{j},{O}_{j}| {A}_{j},{I}_{j}}$. All parties are allowed to apply any possible operation described by probability theory. Furthermore, they are isolated from each other, which means that they can interact only through the environment. Because we do not make global assumptions (beyond logical consistency), the most general picture is that the random variables that are sent from the environment E to the parties are the result of a map on the random variables fed back by all parties to the same environment E (see figure 4). Such a composition of parties with the environment combines states and communication channels in one framework.

Figure 3.

Figure 3. A single party Sj describes a stochastic process ${P}_{{X}_{j},{O}_{j}| {A}_{j},{I}_{j}}$. The variables Aj, and Xj model the input and the output. The variable Ij is obtained from the environment E; the party Sj feeds the variable Oj into the same environment.

Standard image High-resolution image
Figure 4.

Figure 4. The box E describes the environment. Because no predefined causal order is assumed between the parties, the random variable obtained by the parties is the result of E applied to the outgoing random variable of all parties. This picture combines states and channels, i.e., signaling and no-signaling correlations. For example, assume that S0 is in the causal past of all other parties. In that case, the random variable I0 is constant, whereas the random variable ${I}_{j(\ne 0)}$ could depend on A0. For three parties or more, this framework gives rise to a new quality: E can describe a map where no Ij is a constant, yet where no contradiction arises. Such correlations are called non-causal. Similarly to the parties, the box E is a stochastic process ${P}_{{I}_{0},...,{I}_{N-1}| {O}_{0},...,{O}_{N-1}}$.

Standard image High-resolution image

A party Sj has access to the four random variables XjOjIj, and Aj, where Aj is chosen freely. If we consider all parties together, we should get a probability distribution ${P}_{\vec{X},\vec{I},\vec{O}| \vec{A}}$. Furthermore, we ask the environment E to be a multi-linear functional of all local operations. The motivation for this is that linear combinations of local operations should carry through to the probabilities ${P}_{\vec{X},\vec{O},\vec{I}| \vec{A}}$. This brings us to a definition of logical consistency.

Definition 3. (Logical consistency). An environment E is called logically consistent if and only if it is a multi-linear positive map on any choice of local operations ${\{{P}_{{X}_{j},{O}_{j}| {A}_{j},{I}_{j}}\}}_{0\leqslant j\lt N}$ of all parties such that the composition of E with the local operations results in a probability distribution ${P}_{\vec{X},\vec{I},\vec{O}| \vec{A}}$.

The linearity and positivity conditions from definition 3 imply theorem 1, which states that the environment must be a stochastic process (conditional probability distribution).

Theorem 1. (Logical consistent environment as stochastic process). The environment E is a stochastic process ${P}_{\vec{I}| \vec{O}}$ that maps $\vec{O}$ to $\vec{I}$.

Proof. The environment is a multi-linear positive map ${ \mathcal E }$ on the probabilities (we omit the arguments for the sake of presentation)

that party Sj outputs oj to the environment and generates xj conditioned on the setting aj and on ${I}_{j}={i}_{j}$. Therefore, we write

Since ${ \mathcal E }$ is a multi-linear positive map and since it depends on $\vec{O}$ and $\vec{I}$ only, the above probability can be written as

Equation (1)

where $E(\vec{o},\vec{i})$ is a number. This number must be non-negative, as otherwise the above expression (1) is not a probability. By fixing $\vec{A}=\vec{a}$ and by summing over $\vec{x}$, we get

where

Let us fix the local operations ${p}_{j}^{\prime }$ of all parties to be

From the total-probability condition we obtain

By repeating this calculation for different choices of local operations where the parties deterministically output a value, we get

Therefore, E is a stochastic process ${P}_{\vec{I}| \vec{O}}$. □

The following corollary follows from theorem 1.

Corollary 1. A logical consistent environment ${P}_{\vec{I}| \vec{O}}$ fulfills the property that under any choice of the local operations ${\{{P}_{{X}_{j},{O}_{j}| {A}_{j},{I}_{j}}\}}_{0\leqslant j\lt N}$ of all parties, the expression ${P}_{\vec{I}| \vec{O}}{\prod }_{j=0}^{N-1}{P}_{{X}_{j},{O}_{j}| {A}_{j},{I}_{j}}$ form a conditional probability distribution ${P}_{\vec{X},\vec{I},\vec{O}| \vec{A}}$.

Note that not every conditional distribution ${P}_{\vec{I}| \vec{O}}$ is logically consistent. Some stochastic processes lead to grandfather-paradox-type [50] inconsistencies. Consider the following two extreme examples of such inconsistencies. We describe the examples in the single-party scenario as depicted in figure 5 and where OIX, and A are binary random variables.

Figure 5.

Figure 5. Party S is described by ${P}_{X,O| A,I}$ and the environment E is ${P}_{I| O}$.

Standard image High-resolution image

Example 1. Let the environment as well as the party S forward the random variable, i.e., the operation of the environment is

and the operation of the party S is

Since the environment E and the party S forward the random variable, we have $\mathrm{Pr}(O=I)=1$. However, it is unclear what value the probability ${P}_{O}(0)$ should take. This is also known as the causal-loop paradox.

Example 2. We alter the local operation of party S to negate the binary random variable

Now, we are faced with the grandfather paradox: if party S receives i = 1 from the environment, then she sends the value o = 0 to the environment. But in that case, she should receive i = 0 and not i = 1.

2.2. Mathematical model of states, operations, evolution, and composition

Let $\{{q}_{0},{q}_{1},...\}$ be the sample space of a random variable Q with the probability measure PQ.

Definition 4. (States, operations, evolution, and composition). We represent a state corresponding to a random variable PQ as the probability vector

A stochastic process ${P}_{R| Q}$ from Q to a random variable R with sample space $\{{r}_{0},{r}_{1},...\}$ describes an operation and is modeled by the stochastic matrix

The result PR of evolving the random variable PQ through the operation ${P}_{R| Q}$ is given by the matrix multiplication

Finally, vectors and matrices are composed in parallel using the Kronecker product $\otimes $.

For example, by this definition, the output of a stochastic process ${P}_{R| {Q}_{0},{Q}_{1}}$ taking two inputs and producing one output is expressed by

2.3. Set of logically consistent processes without predefined causal order

We derive the conditions on the environment E (stochastic process) such that it is logically consistent. For simplicity, we start with the single-party scenario as depicted in figure 5; the party is denoted by S and the environment by E. We can further simplify our picture by fixing the value of A to a and by summing over X:

The stochastic process of the environment E is ${P}_{I| O}$. For now, let us assume that S performs a deterministic operation ${D}_{O| I}$. This assumption is dropped later. The operation applied by S can be written as a function

where i is a deterministic input value. By embedding f into the process of E, we get

This can be interpreted as a probability measure of party S receiving the value i from the environment E:

For $Q(i)$ to represent a probability measure, the values of Q for every deterministic value i must be non-negative and have to sum up to 1:

We express both conditions in the matrix picture. Non-negativity is achieved whenever all entries of the matrix ${\hat{P}}_{I| O}$ are non-negative. The total-probability conditions are formulated in the following way. The value $f(i)$ that is fed into the environment E is

The matrix ${\hat{P}}_{I| O}$ fixed to providing the state $\vec{i}$ to the party S is

Therefore, the probability of party S observing i is

and the law of total probability requires

This condition remains the same if we relax the input to a stochastic input and the operation of S to a stochastic process ${P}_{O| I}$. The reason for this is that any stochastic input can be written as a convex combination of deterministic inputs, and any stochastic process can be written as a convex combination of deterministic operations. Therefore, the logical-consistency requirement asks the environment E to be restricted to those processes $\hat{E}$ where, under any choice of the local operation ${P}_{O| I}$ of party S, the law of total probability

Equation (2)

and the non-negativity condition

Equation (3)

hold. Because a stochastic process can be written as a convex mixture of deterministic operations, it is sufficient to ask for

for every operation $\hat{D}$ from the set ${ \mathcal D }$ of all deterministic operations. Thanks to linearity, we can straightforwardly extend these requirements to multiple parties, and arrive at theorems 2 and 3.

Theorem 2. (Total probability). The law that the sum of the probabilities over the exclusive states the parties receive is 1 is satisfied if and only if

where ${\hat{D}}_{j}$ represents a deterministic operation of party Sj.

Theorem 3. (Non-negative probabilities). The law that the probability of the parties observing a state is non-negative is satisfied if and only if

2.4. Equivalence to the quantum correlations framework in the classical limit

The ingredients of the framework by Oreshkov et al [31] are process matrices and local operations—described by matrices as well. All the matrices are completely positive trace-preserving quantum maps in the Choi–Jamiołkowski [51, 52] picture. In the classical limit, the matrices become diagonal in the computational basis [31, 35]. In the single-party scenario, the process matrix W is a map from the Hilbert space ${{ \mathcal H }}_{O}$ to the Hilbert space ${{ \mathcal H }}_{I}$. The party's local operation A then again is a map from the Hilbert space ${{ \mathcal H }}_{I}$ to the Hilbert space ${{ \mathcal H }}_{O}$. The conditions a process matrix W in a single-party scenario has to fulfill [31] are

Equation (4)

Equation (5)

where ${ \mathcal M }$ is the set of all completely positive trace-preserving maps from the space ${{ \mathcal H }}_{I}$ to the space ${{ \mathcal H }}_{O}$. Intuitively, the condition given by equation (4) 'short-circuits' both maps and enforces the probabilities of the outcomes to sum up to 1.

Theorem 4. (Equivalence). The quantum framework given by equations (4) and (5) in the classical limit is equivalent to the description of classical correlations without predefined causal order given by equations (2) and (3).

Proof. The process matrix W in the quantum framework corresponds to the stochastic process of the environment E in our framework, and the local operations correspond to the stochastic process of the parties. We show a bijection between process matrices and stochastic processes of the environment, and between local operations and stochastic processes of the parties.

A stochastic matrix $\hat{E}$, representing the environment E in our framework, can be translated into the quantum framework by

where $| k\rangle $ and $| {\ell }\rangle $ are computational-basis states of the same dimension as $\hat{E}$, and where the subscripts denote the respective Hilbert spaces. This completely positive trace-preserving map (expressed in the Choi–Jamiołkowski picture) acts in the same way as the stochastic matrix $\hat{E}$: the state $| k\rangle $ is mapped to $\hat{E}| k\rangle $. The function $d(\rho )$ takes the matrix ρ and cancels all off-diagonal terms, i.e.

We can rewrite ${W}_{\hat{E}}$ as

Analogously, the stochastic matrix ${\hat{P}}_{O| I}$ of the party can be translated into the quantum framework and becomes

The reverse direction of the bijection follows from the description above.

Now, we show that the conditions (4) and (5) in a single-party scenario on a process matrix W coïncide with the conditions (2) and (3) in our framework. The non-negativity condition (5) forces the probabilities of the outputs of W to be non-negative; the same holds for the condition (3) in our framework. That the condition (4) coincides with the condition (2) is shown below. Forcing W and A to be diagonal in the computational basis gives

Substituting W with ${W}_{\hat{E}}$ and A with ${A}_{{\hat{P}}_{O| I}}$ yields

which proves the claim. The multi-party case follows through linearity.□

3. Polytope of classical processes without predefined causal order

3.1. Polytopes

Convex polytopes can be represented in two different ways: the H-representation is a list of half-spaces where the intersection is the polytope, and the V-representation is a list of the extremal points of the polytope. Algorithms like the double-description method [53, 54] enumerate all extremal points of the polytope given the H-representation. We used cdd+ [55] for vertex enumeration. The inverse problem is solved by its dual: a convex-hull algorithm.

Here, we derive the polytope of classical processes without predefined causal order. This polytope is represented by the dashed–dotted lines in figure 2. A projection of the polytope for three parties and binary inputs/outputs onto a plane is given in figure 6.

Figure 6.

Figure 6. Here, we see a projection of the polytope of classical processes without predefined causal order among three parties and with binary inputs/outputs. The circular identity channel C and the circular bit-flip channel $\bar{C}$ are logically inconsistent; they can be used to reproduce the grandfather's paradox. The solid lines mark the deterministic-extrema polytope and the dashed–dotted lines mark the additional space of logically consistent processes. Point ${\hat{E}}_{\mathrm{ex}1}$ is an extremal point of the polytope and is a uniform mixture of the deterministic processes C and $\bar{C}$. The behaviour of this point is shown in figure 12. Point ${\hat{E}}_{\mathrm{det}1}$ is an extremal point of the deterministic-extrema polytope, and is described in figure 11.

Standard image High-resolution image

3.2. Single party, binary input, and binary output

We start with the polytope for one party (see figure 5) with a binary input and a binary output. In this case, a process is described by a square matrix of dimension 2. The most general process of the environment E is

consisting of four variables. The deterministic operations party S can apply are

where ${\hat{D}}_{0}$${\hat{D}}_{1}$ produce a constant 0, 1, respectively, and where the matrix ${\hat{D}}_{2}$ is the identity and ${\hat{D}}_{3}$ the negation. The equalities

Equation (6)

Equation (7)

Equation (8)

enforce

This is shown as follows:

By eliminating three variables using the total-probability conditions (6)–(8) from above, we get

with the non-negativity conditions

This solution set is a one-dimensional polytope with the extremal points 0 and 1. All solutions describe a state. This implies that all correlations that can be obtained in this framework with a single party and binary input and output, can also be obtained in a framework without feedback, i.e., these correlations can be obtained causally (see figure 7).

Figure 7.

Figure 7. All logically consistent single-party correlations that can be obtained with a feedback channel (see figure 5) can be simulated without a feedback channel.

Standard image High-resolution image

3.3. Two parties, binary inputs, and binary outputs

In the two-party case with a binary input and a binary output for each party, the process $\hat{E}={\hat{P}}_{{I}_{0},{I}_{1}| {O}_{0},{O}_{1}}$ of the environment is described by a square matrix of dimension 22. The conditions are

Equation (9)

With a similar argument as above, one can show that the operation ${\hat{D}}_{3}$ does not need to be considered for either party. The matrix $\hat{E}$ consists of 42 unknowns, out of which 32 are eliminated by the total-probability conditions given by equation (9). Thus, we are left with seven unknowns, forming a seven-dimensional polytope with 16 inequalities.

The resulting V-representation of the polytope consists of 12 extremal points, all of which represent deterministic processes:

In the following, we use $A={S}_{0}$ and $B={S}_{1}$. The first four processes ${\hat{E}}_{0},{\hat{E}}_{1},{\hat{E}}_{2},{\hat{E}}_{3}$ represent the four constants $(0,0),(0,1),(1,0),(1,1)$ as inputs to the parties A and B. The next four processes represent a constant input to party A (processes ${\hat{E}}_{4}$ and ${\hat{E}}_{5}$ produce the constant 0, and the other two processes produce the constant 1 ) and a channel from party A to party B; the processes ${\hat{E}}_{4}$ and ${\hat{E}}_{6}$ describe the identity channel, and ${\hat{E}}_{5}$ and ${\hat{E}}_{7}$ describe the bit-flip channel. The last four processes are analogous, with a channel from B to A and where party B receives a constant. All these 12 processes act deterministically on bits for two parties where at least one party receives a constant (see figure 8). Therefore, every such channel can be simulated in a causal fashion. This result generalized to higher dimensions was already shown by taking the classical limit ofthe framework for quantum correlations without predefined causal order [31].

Figure 8.

Figure 8. (a) Both parties A and B receive a constant each. (b) Party A receives a constant and sends a bit through the identity $(c=0)$ or the bit-flip $(c=1)$ channel to B. (c) Same as (b), where the parties are interchanged.

Standard image High-resolution image

3.4. Three parties, binary inputs, and binary outputs

The process of the environment E in a three party setup with binary inputs and outputs is described by a square matrix $\hat{E}={\hat{P}}_{{I}_{0}{I}_{1}{I}_{2}| {O}_{0}{O}_{1}{O}_{2}}$ of dimension 23. The matrix $\hat{E}$ consists of 43 variables, out of which 33 can be eliminated with the total-probability conditions

Equation (10)

resulting in a 37 -dimensional polytope with 43 linear constraints (non-negative probabilities):

Equation (11)

Solving this polytope yields ${710}^{\prime }760$ extremal points. Only 744 extremal points out of these ${710}^{\prime }760$ are deterministic, i.e., consist of 0–1 values; the remaining extremal points are so-called proper mixtures of logically inconsistent processes. Such proper mixtures are not convex combinations of deterministic extremal points inside the polytope, but are convex combinations of deterministic points where some lie outside of the polytope—any process from outside of the polytope leads to logical inconsistencies. Interestingly, this smaller polytope (hence, also the polytope described by the equations (10) and (11)) consists of processes that cannot be simulated using a predefined causal order, i.e., processes where no party receives a constant, implying that every party causally succeeds some other party. The 744 deterministic extremal points are discussed in section 4 along with the general polytope restricted to the deterministic extremal points.

3.5. General case

We describe the polytope for logically consistent classical processes without predefined causal order in the general case. Let n be the number of parties and let d be the dimension of the states entering and leaving every laboratory. This leaves us with a ${d}^{n}\times {d}^{n}$ stochastic matrix $\hat{E}$ describing the environment. Every party can perform an operation that is a convex mixture of all dd deterministic operations. The set of all deterministic operations is denoted by ${ \mathcal D }$. For every party, under any choice of deterministic operation $D\in { \mathcal D }$, the trace of the environment $\hat{E}$ multiplied with the local operations is constrained to give 1 (see theorem 2). However—as in the binary-input/output case above—, some of these constraints are redundant.

Theorem 5. (Sufficient set for total-probability conditions). The total-probability conditions to this family of operations

where j is output for input i and 0 otherwise, imply the total-probability conditions for all remaining deterministic operations of the same dimension, i.e.

Proof. We restrict ourselves to the single-party scenario—the multi-party case follows through linearity. Let ${\vec{v}}_{i}$ be the d-dimensional vector with a 1-entry at position i and 0 's everywhere else. We can write a d-dimensional matrix ${\hat{D}}_{i,j}$ as

A general deterministic matrix $D\in { \mathcal D }$ of the same dimension, where k is mapped to ak, is expressed as

On the one hand, using the antecedent above, we have

Equation (12)

with ${\ell }=| \{k| {a}_{k}\;\ne \;0\}| $. On the other hand, we can rewrite ${\sum }_{k:{a}_{k}\ne 0}{\hat{D}}_{k,{a}_{k}}$ as

Therefore

which, with the identity (12), implies

The family $\{{\hat{D}}_{i,j}| i,j\in I\}$ of deterministic operations with the set $I=\{0,...,n-1\}$ has size $d(d-1)+1$.

Theorem 6. (Polytope). The H-representation of the polytope of logically consistent classical processes without predefined causal order is

The polytope has ${d}^{2n}$ facets and dimension

which is exponential in the number of parties.

4. The deterministic-extrema polytope

Definition 5. (Deterministic-extrema polytope). The deterministic-extrema polytope is defined as the polytope of logically consistent processes without predefined causal order where all extremal points are deterministic processes (see polytope with the solid lines in figure 2).

The deterministic-extrema polytope excludes proper mixtures of logically inconsistent processes. Such mixtures (consistent mixture of inconsistent points) are convex combinations of deterministic points where at least one deterministic point is outside of the polytope. To find this polytope, one can first solve the extremal points of the general polytope and thereafter select the boolean solutions. These boolean solutions form the V-representation of the polytope in discussion.

4.1. Three parties, binary inputs, and binary outputs

We discuss the deterministic-extrema polytope in the setting of three parties and binary inputs and outputs. To simplify the presentation, we use $A={S}_{0}$$B={S}_{1}$, $C={S}_{2}$, ${O}_{A}={O}_{0}$${O}_{B}={O}_{1}$${O}_{C}={O}_{2}$${I}_{A}={I}_{0}$${I}_{B}={I}_{1}$, and ${I}_{C}={I}_{2}$. As described in section 3.4, this polytope has 744 extremal points. They can be characterized as follows.

Assume that $\hat{E}$, when the parties locally apply the identity operation, maps $(0,0,0)$ to $(0,0,0)$, i.e., $(0,0,0)$ is a fixed-point. Then, any other extremal point ${\hat{E}}^{\prime }$ is obtained by the local operations identity ${\hat{D}}_{3}$ and bit-flip ${\hat{D}}_{4}$, where we embed these local operations into the environment. Let ${\hat{L}}_{i,j,k}$ be the local operation of the three parties

i.e., party A performs the identity if i = 0 and the bit-flip operation if i = 1—the other parties' local operations are defined in the same way. The extremal point ${\hat{E}}^{\prime }$ can be described as

where, as described above, the operations are embedded into the environment (see figure 9). Logical consistency of the environment ${\hat{E}}^{\prime }$ follows because we started with a logically consistent $\hat{E}$ and the operations act on single parties. The process ${\hat{E}}^{\prime }$ maps ($i,j,k$) to $(i,j,k)$. Thus, starting with $\hat{E}$, for any choice of $i,j,k$, we obtain a different extremal point. There are ${2}^{3}-1$ alternative extremal points that can be constructed in this fashion. From this we conclude that $744/8=93$ extremal points are such, that $(0,0,0)$ is a fixed-point under locally applying the identity. We restrict our analysis to these 93 extremal points; all others can be obtained by the above construction. The following analysis is structured depending on the number of parties that receive a constant from the environment.

Figure 9.

Figure 9. By starting from a logically consistent environment E and for any choice of $i,j,k\in \{0,1\}$, one can construct another logically consistent environment ${E}^{\prime }$.

Standard image High-resolution image

There exists only one extremal point where $(0,0,0)$ is mapped to $(0,0,0)$ under applying identity locally, and where every party receives a constant: the constant $(0,0,0)$.

Assume exactly two parties receive a constant, which leaves us with three possibilities of choosing them. Fix these parties to be A and B. The third party C receives a value that depends on the operation of A or of B or of both. Thus, we are in the case $A\preceq C$ or $B\preceq C$. Inevitably, the constant must be $(0,0);$ otherwise the fixed-point $(0,0,0)$ is not recovered. Party C receives a value that depends on the value fed back by A and B; there exist ${2}^{3}-1=7$ such functions where we have excluded the constant and all operations where C receives a value different from 0 on inputs $(0,0)$ to the environment from A and B. Therefore, under all permutations of the parties, 21 extremal points give a constant to two parties and have the fixed-point $(0,0,0)$ when the identity is applied locally.

In a next step, assume that exactly one party receives a constant. This assumption, again, allows for three different setups, as we can choose which party receives a constant. Without loss of generality, let A be this party, i.e., $A\preceq B$ and $A\preceq C;$ the constant must be 0 again in order to comply with the requirement of the fixed-point. Now, we are left with several possibilities on how B and C depend on A and on each other. As a first case, we assume that B and C do not depend on each other, but depend on A only ($A\preceq B$ and $A\preceq C$). This dependency cannot be different from the identity channels from A to B and from A to C; the alternative would be the bit-flip channel that would not reproduce the desired fixed-point $(0,0,0)$. This gives us  three different extremal points under all permutations of the parties. Another possibility on the dependencies is that B depends on A, and C depends on B, i.e., $A\preceq B\preceq C$, and the interchange of parties B and C. The channels—by following the same reasoning above—again must be the identity channels: this gives us six extremal points. Now, we look at the case where B depends on A and where C depends on both, A and B, i.e., $A\preceq B$$A\preceq C$, and $B\preceq C$, and any permutation of the parties. There are six permutations. The constant that A receives must be 0, party B must depend trivially on A (the identity channel) and party C can depend in  five different ways on A and B: these are all 2 -to- 1 -bit functions where $(0,0)$ is mapped to 0 (23) minus the constant and minus the dependencies on A only and on B only. In total, there are $6\times 5=30$ such extremal points. We are left with the last scenario: B depends on A and on C, and C depends on A and on B, i.e., $A\preceq B$$C\preceq B$$A\preceq C$, and $B\preceq C$. The constant, as above, is 0. Given the random variable OA fed to the environment by A, the environment can either describe a channel from B to C ($B\preceq C$) or describe a channel from C to B ($C\preceq B$); any other channel would lead to a causal loop. The direction of the channel must differ under different values oA fed-back by A, as otherwise B and C would not mutually depend on each other. We have two possibilities on the direction given the value fed-back by A is ${O}_{A}=0$. Assume the direction to be $B\preceq C$. For the case ${O}_{A}=0$, the channels from A to B and from B to C are the identity channels in order to comply with the fixed point $(0,0,0)$. In the other case ${O}_{A}=1$, the direction of the channel between B and C is in the reverse direction compared to ${O}_{A}=0$, i.e., $C\preceq B$. Then, because of ${O}_{A}\;\ne \;0$, the random variables IB and IC are not forced to be $(0,0);$ there exist two channels from A to C and another two channels from B to C. Therefore, we are left with 3 × 2 × 4 = 24 possibilities. An overview over these setups is given in figure 10.

Figure 10.

Figure 10. (a) Every party receives a constant. For an environment with fixed point $(0,0,0)$ when the parties locally apply the identity map, this constant must be $(0,0,0)$. (b) Two parties receive a constant $(0,0)$, the third party receives a value depending on the other parties' state fed to the environment. For each of the three cases, seven different functions exist. (c) For each of the three cases, the identity function only is consistent with the setup. (d) Here as well, only the identity channel is consistent with the fixed-point. (e) Five different functions are possible per setup. (f) Here, eight functions per setup are consistent with the fixed-point. (g) No party receives a constant, yet no contradiction arises under any choice of local operations. For a fixed-point $(0,0,0)$ where the parties locally apply the identity map, eight different functions that fulfill these requirements exist. In total, the number of deterministic extremal points where $(0,0,0)$ is mapped to $(0,0,0)$ when the parties apply the identity operation is $1({\rm{a}})+3\times 7({\rm{b}})+3({\rm{c}})+6({\rm{d}})+6\times 5({\rm{e}})+3\times 8({\rm{f}})+8({\rm{g}})=93$.

Standard image High-resolution image

The last setup (see figure 10(g)) where no party receives a constant builds a family of  eight extremal points. All extremal points are equivalent up to relabelling of the inputs to and outputs from the environment. One such extremal point is

The behaviour of this extremal point is

Equation (13)

where $\bar{x}$ is the negation of x. This solution is depicted in figure 11. Its location in the polytope is shown in figure 6.

Figure 11.

Figure 11. The left channel is chosen if the majority of the values fed into the environment is 0, otherwise, the right channel is chosen (see equation (13)).

Standard image High-resolution image

5. Causal games

Given a deterministic extremal point where every party is in the causal past of all other parties, a causal game can be constructed that can be won perfectly in the framework presented here, but that is lost if one assumes a global time.

Definition 6. (Causal game). Let a deterministic process map $O=({O}_{0},{O}_{1},...)$ to ${I}^{O}=({I}_{0}^{O},{I}_{1}^{O},...)$, where the ith entry belongs to party Si, and where for every i${I}_{i}^{O}$ depends on all other parties' inputs to the environment. We define a causal game where party Si gets a random Ai and has to produce ${X}_{i}={I}_{i}^{A}$. The parties are allowed to communicate in a predefined causal order [32, 35, 42]. Let the guesses of all parties be $X=({X}_{0},{X}_{1},...)$, and let the random inputs to all parties be $A=({A}_{0},{A}_{1},...)$. In a setup with n parties and where every party obtains and sends a d-dimensional state, the game's winning probability is

Equation (14)

Let ${p}_{\mathrm{succ}}^{{\rm{C}}}$${p}_{\mathrm{succ}}^{\mathrm{NC}}$ be the success probability of the game (14) with, without the assumption of a predefined causal order, respectively.

Theorem 7. (No winning strategy with predefined order). Using a predefined causal order, the success probability (14) is strictly less than 1, i.e., ${p}_{\mathrm{succ}}^{{\rm{C}}}\lt 1$.

Proof. For every party Si, the random variable ${I}_{i}^{A}$ party Si has to guess depends on the other parties' inputs ${A}_{j(\ne i)}$. In a predefined causal order, however, at least one party is not in the causal future of any other party. Let Si be that party, i.e., $\forall j\;\ne \;i:{S}_{i}\;\not\hspace{-2pt}{\succeq }\;{S}_{j}$ (see lemma 1). Then, at least for one input ${A}_{i}={a}^{\prime }$ to Si, the party Si cannot guess perfectly. The success probability ${p}_{\mathrm{succ}}^{{\rm{C}}}$ is upper bounded by

The guessing probability for the non-perfect guess is upper bounded by

because for at least one input A with ${A}_{i}={a}^{\prime }$, party Si guesses wrongly. Therefore, we obtain

Theorem 8. (Winning strategy without predefined causal order). If we drop the assumption of a predefined causal order, then the causal game (14) can be won perfectly, i.e., ${p}_{\mathrm{succ}}^{\mathrm{NC}}=1$.

Proof. To win the game perfectly, the parties use the process that maps the random variable $O=({O}_{0},{O}_{1},...)$ to the random variable ${I}^{O}=({I}_{0}^{O},{I}_{1}^{O},...)$ deterministically, forward their random input to the environment $({A}_{i}={O}_{i})$, and use the value obtained from the environment as their guess $({X}_{i}={I}_{i}^{O})$.□

For other games, a larger gap between the success probability with a predefined causal order and the success probability without a predefined causal order can be achieved—as is shown in the examples below.

6. Examples

We briefly discuss two examples in the three-party scenario. Let ABC be random input bits to the three parties ABC, respectively, and let XYZ be the corresponding output bits.

Example 3. An extremal point of the first class of polytopes for three parties and binary inputs/outputs is

its behaviour is shown in figure 12. This extremal point is a proper mixture of logically inconsistent processes, as it cannot be written as a convex combination of deterministic points from within the polytope; the left and the right channels from figure 12 individually describe a causal loop and, hence, are logically inconsistent. Initially, this process was used to show that in the classical scenario with three parties or more, correlations without predefined causal order can arise [35]. A causal game that can be formulated for this extremal point is

where, depending on the shared random trit m, the party selected by m has to guess the parity of the other two parties' inputs. If we assume a predefined causal order, then this causal game can be won with probability at most 5/6 [32]. The reason for this is that at least one party is not in the causal future of the others. This party, hence, can guess the parity with a probability of 1/2 only. However, by using the environment from figure 12, one can win the game perfectly. To achieve this, if m = 1, then party B forwards the random input $({O}_{B}=B)$, party C forwards the parity of the random input and the random variable obtained from the environment $({O}_{C}=C\oplus {I}_{C})$, and party A uses the random variable obtained from the environment as its guess $(X={I}_{A})$. For the cases m = 2 and m = 3, the same strategy is used, but where the parties are permuted.

Example 4. Another example [56] is depicted in figure 11 and is a deterministic extremal point of the polytope with three parties and binary inputs/outputs (see also equation (13)). Consider the causal game

where $\mathrm{maj}(A,B,C)$ is the majority of the three bits AB, and C. Whenever the majority of the inputs is 0, i.e., $\mathrm{maj}(A,B,C)=0$, then the parties play the 'guess-your-neighbours-input' game [57, 58]: party A guesses the input of party B, party B guesses the input of party C, and finally party C guesses the input of party A; the game is won if all guesses are correct simultaneously. If the majority of the inputs is 1, then they play the same game in reverse direction and flip the output bits. The success probability of winning this game in a world with a predefined causal order is upper bounded by 3/4. This can be seen by the following reasoning. In a predefined causal order, at least a single party has to make a guess without learning anything from the other parties. For example, if party A causally precedes B and C, i.e., ($A\preceq B$ and $A\preceq C$), then party A at best always outputs 0 (see table 1). By making such a guess, however, in 2 out of 8 cases, the parties will loose the game, yielding an upper bound of 3/4 to the success probability. The same upper bound is achieved by choosing party B or party C as causally preceding the others.

By using the environment shown in figure 11, the game can be won perfectly. The parties simply forward their inputs to the environment and use the bits obtained from the environment as the guesses.

7. Conclusion and open questions

We describe the polytope for classical multi-party processes without predefined causal order but where the arising correlations are logically consistent. We also describe the polytope formed by deterministic extremal points; it excludes processes that are proper mixtures of logically inconsistent processes, i.e., processes that cannot be written as a convex combination of deterministic processes from within the polytope. For three parties or more, these polytopes contain processes that cannot be simulated by using a predefined causal order among the parties—this is shown by violations of so-called causal inequalities.

Figure 12.

Figure 12. Channel where the circular identity channel is uniformly mixed with the circular bit-flip channel.

Standard image High-resolution image

Table 1.  Conditions for winning the game of example 4.

A B C $\mathrm{maj}(A,B,C)$ X Y Z
0 0 0 0 0 0 0
0 0 1 0 1 0 0
0 1 0 0 0 0 1
0 1 1 1 0 0 1
1 0 0 0 0 1 0
1 0 1 1 1 0 0
1 1 0 1 0 1 0
1 1 1 1 0 0 0

A representation with polytopes helps for finding new causal games as well as for optimizing the processes for winning causal games; the optimization problem can be stated as a linear program.

In comparison, it has been shown that the set of causal correlations, i.e., correlations with predefined causal order, also forms a polytope [42, 43]. A complete characterization for the two-party case is given [43], however, in the multi-party case, a characterization is missing. Such a characterization is interesting as then one could subtract it from the polytope studied in this work; this yields an exact characterization of the non-causal processes. Another open question is to decide for which causal games the quantum correlations without causal order outperform their classical counterpart.

Acknowledgments

We thank Mateus Araújo, Veronika Baumann, Cyril Branciard, Časlav Brukner, Fabio Costa, Adrien Feix, Arne Hansen, Alberto Montina, and Benno Salwey for helpful discussions. Furthermore we thank the anonymous referees for the detailed comments. The present work was supported by the Swiss National Science Foundation (SNF) and the National Centre of Competence in Research 'Quantum Science and Technology' (QSIT).

Please wait… references are loading.