The arrow of time in RIS

January 18, 2017*

*Last modified 26-Aug-18

Tags: math physics RIS Landauer talk

I gave an informal talk today on the arrow of time in repeated interaction systems and I thought I’d write about it here.

Repeated interaction systems (RIS)

For a recent review of RIS, see arxiv/1305.2472. We also introduce them in my and my coauthor’s work Landauer’s Principle in Repeated Interaction Systems. However, I’ll include what we need here below.

For our purposes, a repeated interaction system consists of a finite dimensional system \(\mathcal{S}\) interacting with a sequence (or “chain”) of finite dimensional systems \(\mathcal{E}_1,\mathcal{E}_2,\ldots\) called “probes”.

Associated to \(\mathcal{S}\) is an Hilbert space \(\mathcal{H}_\mathcal{S}\), self-adjoint operator \(h_\mathcal{S}\in \mathcal{B}(\mathcal{H}_\mathcal{S})\) called the Hamiltonian, and an inital state \(\rho^\text{i}\in \mathcal{B}(\mathcal{H}_\mathcal{S})\) with \(\rho^\text{i}\geq 0\), \(\operatorname{tr}\rho^\text{i}=1\).

Similarly, each probe has a Hilbert space \(\mathcal{H}_{\mathcal{E},k}\), self-adjoint Hamiltonian \(h_{\mathcal{E},k} \in \mathcal{B}(\mathcal{H}_\mathcal{E})\), and initial state \(\xi_k\in \mathcal{B}(\mathcal{H}_\mathcal{E})\) (with \(\xi_k\geq 0\), \(\operatorname{tr}\xi_k =1\)). We will assume the probes’ Hilbert spaces are isomorphic, \(\mathcal{H}_{\mathcal{E},k}\equiv \mathcal{H}_\mathcal{E}\), and that the initial state \(\xi_k\) is a thermal state also called a Gibbs state at inverse temperate \(\beta>0\). That is, \[ \xi_k = \frac{\exp(-\beta h_{\mathcal{E},k})}{\operatorname{tr}(\exp(-\beta h_{\mathcal{E},k}))}. \]

The system \(\mathcal{S}\) interacts with each probe, one at a time, for some duration \(\tau>0\). This can be described inductively as follows. If the system is in the state \(\rho_k\) after iteracting with the \(k\)th probe, then \(\mathcal{S}\) couples with \(\mathcal{E}_k\) to form the joint (uncorrelated) state \[ \rho_k \otimes \xi_{k+1} \] The interaction of the system and \(k\)th probe is described by some interaction Hamiltonian \(v_k\in \mathcal{B}(\mathcal{H}_\mathcal{S}\otimes \mathcal{H}_\mathcal{E})\) so that after time \(\tau\), the joint state of \(\mathcal{S}\) and \(\mathcal{E}_k\) is given by \[ U_k \rho_k \otimes \xi_{k+1} U_k^* \] where \[ U_k := \exp(-\tau(h_\mathcal{S}\otimes \operatorname{id}+ \operatorname{id}\otimes h_{\mathcal{E},k} + v_k)). \]

The state on \(\mathcal{S}\) alone then, after interacting with the \(k\)th probe, is given by \[ \rho_{k+1} = \operatorname{tr}_{H_\mathcal{E}} (U_k \rho_k \otimes \xi_{k+1} U_k^*). \] This is the initial state of \(\mathcal{S}\) for the interaction with the next probe.

Note that we may define the reduced dynamics on \(\mathcal{S}\) for step \(k\). This is the map \[ \begin{aligned} \mathcal{L}_k : \qquad \mathcal{B}(\mathcal{H}_\mathcal{S}) &\to \mathcal{B}(\mathcal{H}_\mathcal{S}) \\ \eta &\mapsto \operatorname{tr}_{\mathcal{H}_\mathcal{E}} [ U_k \eta\otimes \xi_k U_k^*]. \end{aligned} \] In this language, \(\rho_{k+1} = \mathcal{L}_k (\rho_k)\), and thus \[ \rho_{k+1} = \mathcal{L}_k \circ \mathcal{L}_{k-1}\circ\dotsm\circ \mathcal{L}_1(\rho^\text{i}). \]

Introduction to arrow of time

Now that we understand the setup of an RIS, we can consider what it means to test the arrow of time in this context. We will imagine two black boxes \(\blacksquare_F\) and \(\blacksquare_B\).

For such a box, we input an RIS; that is, we give it all the temperatures, Hamiltonians, initial states, etc., of an RIS with some large number of probes; let’s call the number of probes \(T\). Afterwards, we may push a button and recieve a printout.

To produce its printout, \(\blacksquare_F\) does the following when we press the button:

  1. Make a fresh RIS identical to the one we gave it
  2. Measure the initial state of the system (labelling this outcome by \(a\)) and the energy of each probe (labelling the energy level of the \(k\)th probe by \(i_k\))
  3. Time evolve the system and probes by having them interact as described above
  4. Measure the final state of the system (yielding outcome \(b\)) and the energies of the probes again (where \(j_k\) is the energy level of the \(k\)th probe), and
  5. Print out the measurement outcomes \((a,b,i_1,\dotsc,i_T,j_1,\dotsc,j_T)\).

To produce its printout, \(\blacksquare_B\) does the following:

  1. Make a fresh RIS by putting \(\mathcal{S}\) in the state

\[ \rho^\text{f}:= \mathcal{L}_T \circ \mathcal{L}_{T-1}\circ\dotsm\circ \mathcal{L}_1(\rho^\text{i}). \]

and for \(k=1,\dotsc,T\), putting \(\mathcal{E}_k\) in the state \(\xi_k\). 2. Measure \(\mathcal{S}\) (labelling the outcome by \(b\)) and the energy of each probe (outcomes \(j_k\)). 2. Time evolve by \(U_k^*\), i.e. \(U_k\) with \(\tau\) replaced by \(-\tau\). 3. Measure the state of \(\mathcal{S}\) (calling the outcome \(a\)) and the energy of each probe (labelling the outcomes by \(i_k\)). 4. Printout \((a,b,i_1,\dotsc,i_T,j_1,\dotsc,j_T)\).

Now, we are a given a box \(\blacksquare\), and we input an RIS. Our task now is to determine if we were given \(\blacksquare_B\), or if we were given \(\blacksquare_F\).

This is the task of determine whether we can experimentally detect whether or not time has evolved in the forward direction, corresponding to \(\blacksquare_F\), or in the backwards direction, corresponding to \(\blacksquare_B\), for this particular RIS.

A more careful description of the forward and backward processes

To proceed, we’ll need to calculate the probability of getting a particular set of measurement outcomes for each process. Let us analyse the forward process.

The system starts in the state \(\rho^\text{i}\), and the chain of probes in the initial state \(\Xi := \bigotimes_{k=1}^T \xi_k\).

We write \(\rho^\text{i}= \sum_a a \pi_a^\text{i}\) the spectral decomposition of \(\rho^\text{i}\), where \(\{a\}\) are the eigenvalues of \(\rho^\text{i}\), and \(\pi_a^\text{i}\) is the projection onto the eigenspace associated to \(a\), and \(h_{\mathcal{E},k} = \sum_i E_i^{(k)} \Pi_i^{(k)}\) the spectral decomposition of the \(k\)th probes Hamiltonian.

We measure \(\mathcal{S}\) in the eigenbasis of \(\rho^\text{i}\) and each probe in their Hamiltonian’s eigenbasis, obtaining outcomes \(a\) and \(\vec i = (i_k)_{k=1}^T\) with probability \[ \operatorname{tr}[(\rho^\text{i}\otimes \Xi) (\pi_a^\text{i}\otimes \Pi_{\vec i})] \] where \(\Pi_{\vec i} = \bigotimes_{k=1}^T \Pi_{i_k}^{(k)}\). By performing these measurements, we have projected e.g. the \(k\)th probe into it’s \(i_k\)th energy level. We have also projected the system into the state \(\pi_a^\text{i}/\dim \pi_a^\text{i}\). If \(a\) is a non-degenerate eigenvalue, this means that \(\mathcal{S}\) is now in a pure state.

Next, \(\mathcal{S}\) interacts with each probe, one at a time, starting at \(k=1\) until \(k=T\), via the time evolution \[ U_k := \exp(-\tau(h_\mathcal{S}\otimes \operatorname{id}+ \operatorname{id}\otimes h_{\mathcal{E},k} + v_k)). \]

Now, we measure \(\mathcal{S}\) in the eigenbasis of \(\rho_T = \mathcal{L}_T \circ \mathcal{L}_{T-1}\circ\dotsm\circ \mathcal{L}_1(\rho^\text{i})\) and all of the probes in their energy eigenbasis; that is, we measure \(\mathcal{E}_k\) in the eigenbasis of \(h_{\mathcal{E},k}\). We will write the spectral decomposition \(\rho_T = \sum_b b \pi_b^{(T)}\). The probability of obtaining outcome \(b\) from measuring \(\mathcal{S}\) and outcomes \(\vec j = (j_k)_{k=1}^T\) from measuring the probes is \[ \operatorname{tr}[ U_T \dotsm U_1 (\pi_a^\text{i}\otimes \Pi_{\vec i}) (\rho^\text{i}\otimes \Xi) (\pi_a^\text{i}\otimes \Pi_{\vec i}) U_1^* \dotsm U_T^* (\pi_b^{(T)} \otimes \Pi_{\vec{j}})]. \] Note we only need to write the measurement operators for the second measurement once, by cyclicity of the trace.

We call this quantity \(\mathbb{P}_F^{(T)}(a,b,\vec i, \vec j)\); this is the probability that when we press the button on \(\blacksquare_F\) we obtain the outcomes \((a,b,\vec i, \vec j)\). It is in fact a discrete probability distribution.

Similarly, the probability of getting the measurement outcomes \((a,b,\vec i, \vec j)\) when we press the button on \(\blacksquare_B\) is given by \[ \operatorname{tr}[ U_T^* \dotsm U_1^* (\pi_b^{(T)} \otimes \Pi_{\vec j}) (\rho_T \otimes \Xi) (\pi_b^{(T)} \otimes \Pi_{\vec j}) U_1 \dotsm U_T (\pi_a^\text{i}\otimes \Pi_{\vec{i}})] \] which we will call \(\mathbb{P}_B^{(T)}(a,b,\vec i, \vec j)\).

Hypothesis testing on the arrow of time

Let us assume we press the button of our box \(\blacksquare\) \(n\) times. We collect a printout each time. Based on the observed frequencies of each set of measurement outcomes, we want to know: are these outcomes likely being drawn from the distribution \(\mathbb{P}_F^{(T)}\), or from \(\mathbb{P}_B^{(T)}\)?

This is a classic question in hypothesis testing, and has been well-studied; e.g. Cover & Thomas Ch. 11 is a clear reference.

We can consider our \(n\) uses of \(\blacksquare\) as drawing one sample \((x_1,\dotsc,x_n)\) from \(n\) independent and identically distributed random variables \(X_1,\dotsc,X_n\).

We wish to choose a set \(A_n\) of sequences of outcomes such that if \((x_1,\dotsc,x_n) \in A_n\), we output \(F\), and we output \(B\) otherwise.

There are two errors we can make: we can say that our box \(\blacksquare\) is \(\blacksquare_F\) when really it is \(\blacksquare_B\), or vice-versa. Let us define the error probabilities \[ \begin{aligned} \alpha_n(A_n) &= \text{Pr} ( (X_1,\dotsc,X_n) \not \in A_n |\, \blacksquare = \blacksquare_F)\\ \beta_n(A_n) &= \text{Pr} ( (X_1,\dotsc,X_n) \in A_n |\, \blacksquare = \blacksquare_B) \end{aligned} \] which depend on the set \(A_n\) we choose.

We will minimize a particular type of error which leads to a particular nice formula for how the error rate decays as we increase \(n\).

Let us define \[ \beta_n(\epsilon) = \min_{A_n: \alpha_n(A_n) < \epsilon} \beta_n(A_n). \] That is, the smallest probability of “guessing time evolved forwards when it really evolved backwards”, subject to the constraint that the error that we “guess that time evolved backwards when it really evolved forwards” is kept small.

That is, we are sure as we can be that if we say “forwards”, it was indeed forwards, while keeping the other error tolerably small.

For \(0< \epsilon < 1/2\), the Chernoff-Stein lemma tells us that

\[ \lim_{n\to\infty} -\frac{1}{n}\log \beta_n(\epsilon) = D(\mathbb{P}_F^{(T)}|\mathbb{P}_B^{(T)}) = \sum_{a,b,\vec i, \vec j}\mathbb{P}_F^{(T)}(a,b,\vec i, \vec j) \log \frac{\mathbb{P}_F^{(T)}(a,b,\vec i, \vec j)}{\mathbb{P}_B^{(T)}(a,b,\vec i, \vec j)} \] is the relative entropy between the two probability distributions. That is, \[ \beta_n(\epsilon) \sim \exp(-n D(\mathbb{P}_F^{(T)}|\mathbb{P}_B^{(T)}) ) \] the error \(\beta_n(\epsilon)\) asymptotically decays exponentially with a rate given by the relative entropy \(D(\mathbb{P}_F^{(T)}|\mathbb{P}_B^{(T)})\).

Connection to Landauer’s Principle

Let me refer to my post on Landauer’s principle. At each step of the RIS, we are in the setup of Landauer’s principle described there: a system interacts unitarily with a thermal reservoir, in this case \(\mathcal{E}_k\). Thus, at each step of the RIS run without measurement, we have an entropy production \(\sigma_k \geq 0\). It turns out that \[ D(\mathbb{P}_F^{(T)}|\mathbb{P}_B^{(T)}) = \sum_{k=1}^T \sigma_k. \] This is very nice! In fact, \(\sum_{k=1}^T \sigma_k\) was the object of study in the recent work of myself and my coauthors in Landauer’s Principle in Repeated Interaction Systems (which I’m happy say has now been published).

I won’t prove the equality here; however, if all goes well, it’ll been on the arxiv soon™ [Edit: it’s up, arxiv:1705.08281].