ZapperZ pointed out this paper[I'll refer to this as SA] (which also led me to this paper and to this earlier paper). All of these are open access. The notation and the language that has become commonplace in quantum information/computation has come to obscure the issues involved, IMO, so I wanted to run what these papers do, reformulated in terms of weak von Neumann measurements, through PhysicsOverflow.
A weak von Neumann measurement associated with a self-adjoint operator ˆA that has spectral decomposition ˆA=∑iaiˆPAi, where ∑iˆPAi=1, transforms a state ρ:ˆO↦Tr[ˆρˆO] to the state ρA:ˆO↦∑iTr[ˆPAiˆρˆPAiˆO]=∑iTr[ˆρˆPAiˆOˆPAi],
which discards all terms
Tr[ˆρˆPAiˆOˆPAj] for which
i≠j. SA refers to such operations as "measure-and-reprepare operations" [SA, last sentence in the first complete paragraph on page 4].
SA takes Alice to be able to carry out a weak von Neumann operation while Bob can only carry out a unitary operation, ρU:ˆO↦Tr[ˆρˆU†ˆOˆU] [SA, second complete paragraph on page 4], but they would prefer both to be able to carry out a weak von Neumann operation, so I'll work with that (just remove the sum over j and replace ˆPBj...ˆPBj by ˆU†B...ˆUB everywhere, if preferred).
If we perform two weak von Neumann measurements in sequence, we obtain ρAB:ˆO↦∑i,jTr[ˆρˆPAiˆPBjˆOˆPBjˆPAi] orρBA:ˆO↦∑i,jTr[ˆρˆPBjˆPAiˆOˆPAiˆPBj],
which in general will be different states, whenever
[ˆA,ˆB]≠0, so any convex linear combination
λρAB+(1−λ)ρBA is also a state [cf. SA, Eq. (6)]. The papers above then introduce a third measurement
ˆC, which is actually measured. Supposing we again introduce a spectral decomposition, we have the expectation values
∑i,jTr[ˆρ(λˆPAiˆPBjˆPCkˆPBjˆPAi+(1−λ)ˆPBjˆPAiˆPCkˆPAiˆPBj)]
to work with. The task at this point is to construct what they call a "causal witness", which is essentially to construct a set of density matrices
ˆρi that verify that, given that
ˆA and
ˆB are being applied as weak von Neumann measurements before the
ˆC measurement,
0<λ<1.
It seems noteworthy that this construction does not need the apparatus of qubits. If we do use qubits, one qubit is enough. I'm not touching on the experimental implementation at all here, but am I missing something of the theoretical construction? I take the measure-and-reprepare operation to be understandable as a logical matrix operation on the density matrix that is essentially atemporal beam-line ordered. If this approach fails, is there another approach that does not make such a big deal of causal/temporal order (my ulterior motive, cf also ZapperZ)?
Update: Quantum Optics is a "beam-line formalism". That is, the order of operations is determined by position along each beam-line (each of which is ordered by where the source of the beam is). Quantum Optics removes time from the formalism by taking beam-line ordering as a proxy for temporal ordering. If we are to talk about causality, however minimally, it seems that time should be added back into the formalism. If measurements are made in different orders, then we might, for example, write λρA1B2+(1−λ)ρB1A2 for the compound measurement, acknowledging that in a more complete description of the experiment (something that goes beyond a small-dimensional beam-line Hilbert space formalism in which ˆA1=ˆA2, ˆB1=ˆB2) a measurement ˆA1 at a time t1 that is before a measurement ˆB2 at time t2 would be different from a measurement ˆA2 at a time t2 that is after a measurement ˆB1 at time t1. That is, in a more temporally complete description the statistical mixture would be of different pairs of operations rather than of a different causal ordering. In any case, I see no reason to reason here for doubting microcausality.