Optimization, Game Theory, and Quantum Mechanics, Part I
This post is the first in what I hope is a two-post or three-post sequence about framing things in quantum mechanics through the window of optimization. These pieces were inspired by things I heard while sitting with my friend in Charman's OH for Physics 112, Statistical Mechanics, and some things I learned in CS 270 with Rao. I would highly recommend both classes; I haven't actually taken Physics 112 yet, but from what my friend says, it seems amazing. From these discussions, I was inspired to write about this topic for my Physics 137B term paper, but I decided against it to be safe and switched to writing about the path-integral formulation last minute. This blog post comes from notes I wrote while working on the term paper.
Introduction
In physics, we create mathematical models for the universe. These models take the form of laws and equations, which are often some sort of differential or linear algebra equation: \[ \begin{align}\frac{d\textbf{p}}{dt} &= - \nabla_r \, V(r) && \text{Newton's Second Law} \\ i\hbar \, \partial_t \Psi &= \left ( -\frac{\hbar^2}{2m} \nabla^2 + V \right) \Psi && \text{Time-dependent Schrödinger Equation} \end{align}\] Obviously, not all theories are built with this machinery, but of the physics that I know, most are. These are often easier to work with, since they follow strict equations that determine the evolution of the system. But many theories are also built with some sort of optimization built-in. Light takes the path of least time, and particles follow the path of stationary action. Statistical mechanics can also be viewed as an optimization problem. How about quantum mechanics? Let's first start with the basics of the theory.The Basic Premises of Quantum Mechanics
In quantum mechanics, systems obey the following postulates:- The state of any quantum system is completely described by the wavefunction \(\ket{\Psi}\); it contains all possible information about the system. Here \(\ket{\Psi}\) resides in the Hilbert space (an infinite-dimensional vector space).
- All measurable, physical quantities, such as position, momentum, and energy, correspond to specific linear and Hermitian operators that act on the wavefunction.
- The only outcomes of a measurement are the measurement's corresponding operator's eigenvalues, with the probability of an outcome being an eigenvalue equal to the projection of the wavefunction onto the eigenstate corresponding to that eigenvalue.
- Upon measurement, the wavefunction instantly collapses into the eigenstate corresponding to the measured eigenvalue.
- If we define the Hamiltonian operator \(\hat{H}\) to be \[ \hat{H} = -\frac{\hbar^2}{2m} \nabla^2 + \hat{V},\] where \(\hat{V}\) is the operator associated with a potential \(V\), the time-evolution of the wavefunction, and hence the system, satisfies \[ i\hbar \, \partial_t \ket{\Psi} = \hat{H} \ket{\Psi},\] the Time-dependent Schrödinger Equation (TDSE). If we separate the equation into a temporal one and a spatial one, we obtain the Time-independent Schrödinger Equation (TISE): \[ \hat{H} \ket{\psi} = E \ket{\psi},\] where \(E\) is the energy of the eigenstate, and \(\ket{\Psi } = e^{-iEt/\hbar} \ket{\psi}\).
The Static Picture: Energy as a Cost
We claim that $\ket{\psi} \in \mathcal{C}(\phi)$, where $\phi$ is the solution to the following optimization problem \[ \phi = \text{argmax}_{\phi} \, \, \, \text{tr}(\phi^\dagger H \phi ) \, \, \text{subject to } \, \phi^\dagger \phi = I. \] To prove this, we simply setup the optimization Lagrangian. We have \[ \mathcal{L} = tr(\phi^\dagger H \phi) - tr(\Lambda^T (\phi^\dagger \phi - I)).\] The associated KKT conditions are \[ \nabla_{\phi} \mathcal{L} = 2 H \phi - 2 \phi \Lambda = 0,\] or $H\phi = \phi \Lambda$, as desired. Hence, this optimization problem is equivalent to the TISE. From this, we can notice that $\Lambda$, the dual variable that corresponds to the normalization constraint of the wavefunction, is energy. Energy is the cost the universe pays for the uinitarity of time-evolution. We can also rewrite our Lagrangian by noting the cyclic property of the trace; we have \[\begin{align} \mathcal{L} &= tr(\phi^\dagger H \phi) - tr(\Lambda^T (\phi^\dagger \phi - I)) \\ &= tr(H \phi \phi^\dagger) - tr(\Lambda^T (\phi^\dagger \phi - I)) \end{align}.\] The first term may feel familiar to physics students, as \[ \phi \phi^\dagger = \sum_n \ket{\psi_n} \bra{\psi_n} = \sum_n \rho_n,\] where $\rho_n$ is called the density matrix. Hence, \[\mathcal{L} = \sum_n tr(H \rho_n ) - tr(\Lambda^T (\phi^\dagger \phi - I)).\]The Dynamic Picture: Encoding Evolution
More such insights open up if we can encode the time-evolution of the wavefunction straight into the optimization problem. This leads us to the following: we claim that $\Psi \in \mathcal{C}(\Phi)$, where $\Phi$ evolves according to the following optimization problem: \[ \begin{align} \max_{\gamma} \left ( \min_{\Phi} tr(\Phi^\dagger H \Phi ) - tr(F(\Phi^\dagger R \Phi - I)) \right) & \text{ subject to constraint } F - \hbar I = 0 \end{align}, \] where $R$ is the operator associated with $i \partial_t$. The associated optimization Lagrangian is \[ \mathcal{L} = \min_{\Phi} \left [ tr(\Phi^\dagger H \Phi) - tr(F(\Phi^\dagger R \Phi - I ))) \right] + tr(\Omega (F - \hbar I)).\] To show that this is equivalent to the TDSE, we invoke the KKT conditions. If we define \[ g(F) = \min_{\Phi} \left[ tr(\Phi^\dagger H \Phi) - tr(F(\Phi^\dagger R \Phi - I))\right],\] our conditions are \[\begin{align} \nabla_{\Phi} \mathcal{L} &= \nabla_{\Phi} g = 2 H \Phi - 2 FR \Phi = 0. \\ \nabla_F \mathcal{L} &= \nabla_F g + \Omega^\dagger = \Omega - (\Phi^\dagger R \Phi - I)^\dagger = 0 \\ \nabla_\Omega \mathcal{L} &= (F - \hbar I)^{\dagger} = 0\end{align}\] Plugging the third condition into the first condition yields \[H \Phi = FR \Phi = \hbar I R \Phi = i\hbar \partial_t \Phi,\] which is the TDSE.From the TD Lagrangian to the TI Lagrangian
M: Well, I'm not hiding in there if that's your brilliant plan.
Bond: We're changing vehicles. Trouble with company cars is they have trackers.
M: Oh, and I suppose that's [Bond's Aston Martin DB5] completely inconspicuous.
Skyfall, 2012
- \( H = H_s \otimes I_t\) ($H$ is time-independent)
- $R = I_s \otimes R_t$ ($R$ is spatial-independent)
- We can write $\Phi$ as the sum of tensor products of rank-1 tensors: \[ \Phi = \sum_n \zeta_n \otimes \tau_n. \]
Optimization and the Principle of Stationary Action
Gordon: What about the Riddler? He’s gonna kill again.
Batman: It’s all connected. Like it or not, it’s his game now. We wanna find Riddler, we gotta find that rat.
The Batman, 2022
Killing Particles: Decreasing Probability
Maverick: Put that in your Pentagon budget!
Hondo: Alright, Mav, that’s it. Back it off. We got what we came for.
Maverick: Just a little more. Just a little. Come on... come on...
Top Gun: Maverick, 2022
Lemma. For a constraint $\mathcal{C}$ with an associated constant $c$ for an optimization problem with reward $Q$, at optimality $x^*$, \[ \frac{d Q}{d c} = \lambda^*,\] where $\lambda$ is the dual variable asociated with constraint $c$.
Proof. By strong duality, at optimality, $Q(x^*) = \mathcal{L}(x^*)$. Hence, \[ \begin{align} \frac{d Q}{d c} = \frac{d\mathcal{L}}{dc} &= \frac{\partial \mathcal{L}}{\partial x^*} \frac{dx^*}{dc} + \frac{\partial \mathcal{L}}{\partial y^*} \frac{dy^*}{dc} + \frac{\partial \mathcal{L}}{\partial c} \\ &= \frac{\partial \mathcal{L}}{\partial c} \\ &= \lambda^* \end{align} \] where we note that that optimality, we have stationarity, and we recall the structure of the Lagrangian for the last step.
In our current TIL setup, with the perturbation, to first-order, \[ \begin{align} Q &= Q^* - \frac{dQ}{dN} \Delta(t) \\ &= Q^* - tr(\Lambda^* \Delta(t) ) \end{align} \] If we take the derivative of this with respect to time, recalling $Q$ for the Lagrangian written in terms of the density matrices, and we assume that $H$ is time-independent, we get \[ \begin{align}\frac{dQ}{dt} &= - tr\left(\Lambda^* \frac{d\Delta}{dt}\right) \\ &= \sum_n tr\left(H \frac{d\rho_n}{dt}\right) \end{align} \] Hence, \[ \sum_n tr(H \dot{\rho}_n) + tr(\Lambda^* \dot{\Delta}) = 0. \] We can also note that $tr(\rho_n) = 1 - \Delta^{(n)}(t)$, where $\Delta^{(n)}(t)$ is the $n$th diagonal entry of $\Delta$, so that \[ \frac{d(tr(\rho_n))}{dt} = tr(\dot{\rho}_n) =- \frac{d\Delta^{(n)}}{dt}.\] Summing over $n$, we have \[ \sum_n tr(\dot{\rho}_n) + tr(\dot{\Delta}) = 0. \label{asdf}\] There is a remarkable similarity between the equations; both are of the form \[ \sum_n tr(O \dot{\rho}_n) + tr(\lambda_O \dot{\Delta}) = 0,\] where $O$ is some operator, and $\lambda_O$ are the equilibrium eigenvalues of $O$. This actually holds for all operators that commute with $H$, and hence all conserved quantities: \[\begin{align} O\ket{\Psi_n} = \lambda_O \ket{\Psi_n} &\Rightarrow O \rho_n = \lambda_O \rho_n \\ &\Rightarrow O\dot{\rho}_n = \lambda_O \dot{\rho}_n \\ &\Rightarrow tr(O \dot{\rho}_n) = \lambda_O \, tr(\dot{\rho}_n) \\ &\Rightarrow tr(O \dot{\rho}_n) = -\lambda_O (\dot{\Delta}^{(n)}); \end{align}\] summing over $n$ yields our identity. We can also see this geometrically from our optimization: the main lemma that we use is invariant (up to a constant scaling factor) if we swap the operator from $H$ to $O$ on the set of eigenstates. Note that these equations are very similar to the first law of thermodynamics. Hence, for commuting operators, and hence conserved quantities, similar "first law" equations exist.
Closing Remarks
In this post, I ranted about some thoughts about encoding the TDSE as an optimization I had for my term paper. There are likely tons of mistakes and other problems in what I've written here. I thought the coolest part was the suprising connection between the optimization Lagrangian and the Principle of Stationary Action. If I do write another blog post on this topic (hopefully), I think it will be more about that connection, with hopefully other geometric insights.Sid