The arrow of time in RIS

January 18, 2017*

*Last modified 11-Nov-19

Tags: math, physics, RIS, Landauer, talk

I gave an informal talk today on the arrow of time in repeated interaction systems and I thought I’d write about it here.

Repeated interaction systems (RIS)

For a recent review of RIS, see arxiv/1305.2472. We also introduce them in my and my coauthor’s work Landauer’s Principle in Repeated Interaction Systems. However, I’ll include what we need here below.

For our purposes, a repeated interaction system consists of a finite dimensional system S\mathcal{S} interacting with a sequence (or “chain”) of finite dimensional systems E1,E2,\mathcal{E}_1,\mathcal{E}_2,\ldots called “probes”.

Associated to S\mathcal{S} is an Hilbert space HS\mathcal{H}_\mathcal{S}, self-adjoint operator hSB(HS)h_\mathcal{S}\in \mathcal{B}(\mathcal{H}_\mathcal{S}) called the Hamiltonian, and an inital state ρiB(HS)\rho^\text{i}\in \mathcal{B}(\mathcal{H}_\mathcal{S}) with ρi0\rho^\text{i}\geq 0, trρi=1\operatorname{tr}\rho^\text{i}=1.

Similarly, each probe has a Hilbert space HE,k\mathcal{H}_{\mathcal{E},k}, self-adjoint Hamiltonian hE,kB(HE)h_{\mathcal{E},k} \in \mathcal{B}(\mathcal{H}_\mathcal{E}), and initial state ξkB(HE)\xi_k\in \mathcal{B}(\mathcal{H}_\mathcal{E}) (with ξk0\xi_k\geq 0, trξk=1\operatorname{tr}\xi_k =1). We will assume the probes’ Hilbert spaces are isomorphic, HE,kHE\mathcal{H}_{\mathcal{E},k}\equiv \mathcal{H}_\mathcal{E}, and that the initial state ξk\xi_k is a thermal state also called a Gibbs state at inverse temperate β>0\beta>0. That is, ξk=exp(βhE,k)tr(exp(βhE,k)). \xi_k = \frac{\exp(-\beta h_{\mathcal{E},k})}{\operatorname{tr}(\exp(-\beta h_{\mathcal{E},k}))}.

The system S\mathcal{S} interacts with each probe, one at a time, for some duration τ>0\tau>0. This can be described inductively as follows. If the system is in the state ρk\rho_k after iteracting with the kkth probe, then S\mathcal{S} couples with Ek\mathcal{E}_k to form the joint (uncorrelated) state ρkξk+1 \rho_k \otimes \xi_{k+1} The interaction of the system and kkth probe is described by some interaction Hamiltonian vkB(HSHE)v_k\in \mathcal{B}(\mathcal{H}_\mathcal{S}\otimes \mathcal{H}_\mathcal{E}) so that after time τ\tau, the joint state of S\mathcal{S} and Ek\mathcal{E}_k is given by Ukρkξk+1Uk U_k \rho_k \otimes \xi_{k+1} U_k^* where Uk:=exp(τ(hSid+idhE,k+vk)). U_k := \exp(-\tau(h_\mathcal{S}\otimes \operatorname{id}+ \operatorname{id}\otimes h_{\mathcal{E},k} + v_k)).

The state on S\mathcal{S} alone then, after interacting with the kkth probe, is given by ρk+1=trHE(Ukρkξk+1Uk). \rho_{k+1} = \operatorname{tr}_{H_\mathcal{E}} (U_k \rho_k \otimes \xi_{k+1} U_k^*). This is the initial state of S\mathcal{S} for the interaction with the next probe.

Note that we may define the reduced dynamics on S\mathcal{S} for step kk. This is the map Lk:B(HS)B(HS)ηtrHE[UkηξkUk]. \begin{aligned} \mathcal{L}_k : \qquad \mathcal{B}(\mathcal{H}_\mathcal{S}) &\to \mathcal{B}(\mathcal{H}_\mathcal{S}) \\ \eta &\mapsto \operatorname{tr}_{\mathcal{H}_\mathcal{E}} [ U_k \eta\otimes \xi_k U_k^*]. \end{aligned} In this language, ρk+1=Lk(ρk)\rho_{k+1} = \mathcal{L}_k (\rho_k), and thus ρk+1=LkLk1L1(ρi). \rho_{k+1} = \mathcal{L}_k \circ \mathcal{L}_{k-1}\circ\dotsm\circ \mathcal{L}_1(\rho^\text{i}).

Introduction to arrow of time

Now that we understand the setup of an RIS, we can consider what it means to test the arrow of time in this context. We will imagine two black boxes F\blacksquare_F and B\blacksquare_B.

For such a box, we input an RIS; that is, we give it all the temperatures, Hamiltonians, initial states, etc., of an RIS with some large number of probes; let’s call the number of probes TT. Afterwards, we may push a button and recieve a printout.

To produce its printout, F\blacksquare_F does the following when we press the button:

  1. Make a fresh RIS identical to the one we gave it
  2. Measure the initial state of the system (labelling this outcome by aa) and the energy of each probe (labelling the energy level of the kkth probe by iki_k)
  3. Time evolve the system and probes by having them interact as described above
  4. Measure the final state of the system (yielding outcome bb) and the energies of the probes again (where jkj_k is the energy level of the kkth probe), and
  5. Print out the measurement outcomes (a,b,i1,,iT,j1,,jT)(a,b,i_1,\dotsc,i_T,j_1,\dotsc,j_T).

To produce its printout, B\blacksquare_B does the following:

  1. Make a fresh RIS by putting S\mathcal{S} in the state

ρf:=LTLT1L1(ρi). \rho^\text{f}:= \mathcal{L}_T \circ \mathcal{L}_{T-1}\circ\dotsm\circ \mathcal{L}_1(\rho^\text{i}).

and for k=1,,Tk=1,\dotsc,T, putting Ek\mathcal{E}_k in the state ξk\xi_k. 2. Measure S\mathcal{S} (labelling the outcome by bb) and the energy of each probe (outcomes jkj_k). 2. Time evolve by UkU_k^*, i.e. UkU_k with τ\tau replaced by τ-\tau. 3. Measure the state of S\mathcal{S} (calling the outcome aa) and the energy of each probe (labelling the outcomes by iki_k). 4. Printout (a,b,i1,,iT,j1,,jT)(a,b,i_1,\dotsc,i_T,j_1,\dotsc,j_T).

Now, we are a given a box \blacksquare, and we input an RIS. Our task now is to determine if we were given B\blacksquare_B, or if we were given F\blacksquare_F.

This is the task of determine whether we can experimentally detect whether or not time has evolved in the forward direction, corresponding to F\blacksquare_F, or in the backwards direction, corresponding to B\blacksquare_B, for this particular RIS.

A more careful description of the forward and backward processes

To proceed, we’ll need to calculate the probability of getting a particular set of measurement outcomes for each process. Let us analyse the forward process.

The system starts in the state ρi\rho^\text{i}, and the chain of probes in the initial state Ξ:=k=1Tξk\Xi := \bigotimes_{k=1}^T \xi_k.

We write ρi=aaπai\rho^\text{i}= \sum_a a \pi_a^\text{i} the spectral decomposition of ρi\rho^\text{i}, where {a}\{a\} are the eigenvalues of ρi\rho^\text{i}, and πai\pi_a^\text{i} is the projection onto the eigenspace associated to aa, and hE,k=iEi(k)Πi(k)h_{\mathcal{E},k} = \sum_i E_i^{(k)} \Pi_i^{(k)} the spectral decomposition of the kkth probes Hamiltonian.

We measure S\mathcal{S} in the eigenbasis of ρi\rho^\text{i} and each probe in their Hamiltonian’s eigenbasis, obtaining outcomes aa and i=(ik)k=1T\vec i = (i_k)_{k=1}^T with probability tr[(ρiΞ)(πaiΠi)] \operatorname{tr}[(\rho^\text{i}\otimes \Xi) (\pi_a^\text{i}\otimes \Pi_{\vec i})] where Πi=k=1TΠik(k)\Pi_{\vec i} = \bigotimes_{k=1}^T \Pi_{i_k}^{(k)}. By performing these measurements, we have projected e.g. the kkth probe into it’s iki_kth energy level. We have also projected the system into the state πai/dimπai\pi_a^\text{i}/\dim \pi_a^\text{i}. If aa is a non-degenerate eigenvalue, this means that S\mathcal{S} is now in a pure state.

Next, S\mathcal{S} interacts with each probe, one at a time, starting at k=1k=1 until k=Tk=T, via the time evolution Uk:=exp(τ(hSid+idhE,k+vk)). U_k := \exp(-\tau(h_\mathcal{S}\otimes \operatorname{id}+ \operatorname{id}\otimes h_{\mathcal{E},k} + v_k)).

Now, we measure S\mathcal{S} in the eigenbasis of ρT=LTLT1L1(ρi)\rho_T = \mathcal{L}_T \circ \mathcal{L}_{T-1}\circ\dotsm\circ \mathcal{L}_1(\rho^\text{i}) and all of the probes in their energy eigenbasis; that is, we measure Ek\mathcal{E}_k in the eigenbasis of hE,kh_{\mathcal{E},k}. We will write the spectral decomposition ρT=bbπb(T)\rho_T = \sum_b b \pi_b^{(T)}. The probability of obtaining outcome bb from measuring S\mathcal{S} and outcomes j=(jk)k=1T\vec j = (j_k)_{k=1}^T from measuring the probes is tr[UTU1(πaiΠi)(ρiΞ)(πaiΠi)U1UT(πb(T)Πj)]. \operatorname{tr}[ U_T \dotsm U_1 (\pi_a^\text{i}\otimes \Pi_{\vec i}) (\rho^\text{i}\otimes \Xi) (\pi_a^\text{i}\otimes \Pi_{\vec i}) U_1^* \dotsm U_T^* (\pi_b^{(T)} \otimes \Pi_{\vec{j}})]. Note we only need to write the measurement operators for the second measurement once, by cyclicity of the trace.

We call this quantity PF(T)(a,b,i,j)\mathbb{P}_F^{(T)}(a,b,\vec i, \vec j); this is the probability that when we press the button on F\blacksquare_F we obtain the outcomes (a,b,i,j)(a,b,\vec i, \vec j). It is in fact a discrete probability distribution.

Similarly, the probability of getting the measurement outcomes (a,b,i,j)(a,b,\vec i, \vec j) when we press the button on B\blacksquare_B is given by tr[UTU1(πb(T)Πj)(ρTΞ)(πb(T)Πj)U1UT(πaiΠi)] \operatorname{tr}[ U_T^* \dotsm U_1^* (\pi_b^{(T)} \otimes \Pi_{\vec j}) (\rho_T \otimes \Xi) (\pi_b^{(T)} \otimes \Pi_{\vec j}) U_1 \dotsm U_T (\pi_a^\text{i}\otimes \Pi_{\vec{i}})] which we will call PB(T)(a,b,i,j)\mathbb{P}_B^{(T)}(a,b,\vec i, \vec j).

Hypothesis testing on the arrow of time

Let us assume we press the button of our box \blacksquare nn times. We collect a printout each time. Based on the observed frequencies of each set of measurement outcomes, we want to know: are these outcomes likely being drawn from the distribution PF(T)\mathbb{P}_F^{(T)}, or from PB(T)\mathbb{P}_B^{(T)}?

This is a classic question in hypothesis testing, and has been well-studied; e.g. Cover & Thomas Ch. 11 is a clear reference.

We can consider our nn uses of \blacksquare as drawing one sample (x1,,xn)(x_1,\dotsc,x_n) from nn independent and identically distributed random variables X1,,XnX_1,\dotsc,X_n.

We wish to choose a set AnA_n of sequences of outcomes such that if (x1,,xn)An(x_1,\dotsc,x_n) \in A_n, we output FF, and we output BB otherwise.

There are two errors we can make: we can say that our box \blacksquare is F\blacksquare_F when really it is B\blacksquare_B, or vice-versa. Let us define the error probabilities αn(An)=Pr((X1,,Xn)̸An=F)βn(An)=Pr((X1,,Xn)An=B) \begin{aligned} \alpha_n(A_n) &= \text{Pr} ( (X_1,\dotsc,X_n) \not \in A_n |\, \blacksquare = \blacksquare_F)\\ \beta_n(A_n) &= \text{Pr} ( (X_1,\dotsc,X_n) \in A_n |\, \blacksquare = \blacksquare_B) \end{aligned} which depend on the set AnA_n we choose.

We will minimize a particular type of error which leads to a particular nice formula for how the error rate decays as we increase nn.

Let us define βn(ϵ)=minAn:αn(An)<ϵβn(An). \beta_n(\epsilon) = \min_{A_n: \alpha_n(A_n) < \epsilon} \beta_n(A_n). That is, the smallest probability of “guessing time evolved forwards when it really evolved backwards”, subject to the constraint that the error that we “guess that time evolved backwards when it really evolved forwards” is kept small.

That is, we are sure as we can be that if we say “forwards”, it was indeed forwards, while keeping the other error tolerably small.

For 0<ϵ<1/20< \epsilon < 1/2, the Chernoff-Stein lemma tells us that

limn1nlogβn(ϵ)=D(PF(T)PB(T))=a,b,i,jPF(T)(a,b,i,j)logPF(T)(a,b,i,j)PB(T)(a,b,i,j) \lim_{n\to\infty} -\frac{1}{n}\log \beta_n(\epsilon) = D(\mathbb{P}_F^{(T)}|\mathbb{P}_B^{(T)}) = \sum_{a,b,\vec i, \vec j}\mathbb{P}_F^{(T)}(a,b,\vec i, \vec j) \log \frac{\mathbb{P}_F^{(T)}(a,b,\vec i, \vec j)}{\mathbb{P}_B^{(T)}(a,b,\vec i, \vec j)} is the relative entropy between the two probability distributions. That is, βn(ϵ)exp(nD(PF(T)PB(T))) \beta_n(\epsilon) \sim \exp(-n D(\mathbb{P}_F^{(T)}|\mathbb{P}_B^{(T)}) ) the error βn(ϵ)\beta_n(\epsilon) asymptotically decays exponentially with a rate given by the relative entropy D(PF(T)PB(T))D(\mathbb{P}_F^{(T)}|\mathbb{P}_B^{(T)}).

Connection to Landauer’s Principle

Let me refer to my post on Landauer’s principle. At each step of the RIS, we are in the setup of Landauer’s principle described there: a system interacts unitarily with a thermal reservoir, in this case Ek\mathcal{E}_k. Thus, at each step of the RIS run without measurement, we have an entropy production σk0\sigma_k \geq 0. It turns out that D(PF(T)PB(T))=k=1Tσk. D(\mathbb{P}_F^{(T)}|\mathbb{P}_B^{(T)}) = \sum_{k=1}^T \sigma_k. This is very nice! In fact, k=1Tσk\sum_{k=1}^T \sigma_k was the object of study in the recent work of myself and my coauthors in Landauer’s Principle in Repeated Interaction Systems (which I’m happy say has now been published).

I won’t prove the equality here; however, if all goes well, it’ll been on the arxiv soon™ [Edit: it’s up, arxiv:1705.08281].