Notes on quantum control

Introduction

Why is quantum control theory and quantum dynamics can help us, in developping interesting and powerful quantum software tools, as user-side quantum engineers ?

What is a parametric quantum circuit ? Within the quantum circuit framework, it is rigorously some quantum interactions build on entanglement layers (CNOT) and some parametric (XYZ) set of qubit rotations. Quantum control ask the (really reasonnable question) of wether or not it is physically and technologically possible to apply those rotations using pulses control.

For any of our quantum devices, especially dominating ones, pulses (and associated pulse shapes) are the control parameters of micro or meso quantum systems. In the case of superconducting transmon, variable configurations of correlators, Josephson junctions and layout for electrical quantum control is layed out in a quantum circuit, design and topologies are engineered around the idea of quantum control using microwave pulses, whose shapes, sequences and timing have also to be described through (optimal) quantum control. It is moreover the case also for any dominating quantum technologies, Rydberg atoms are being moved around within laser tweezers and raised or lowered from one quantum state to the other also using engineered control pulses, this time LASER based. Finally, it is the same for ion traps, with EM and MW pulses.

This list is of course non-exhaustive, however it goes almost without saying that quantum control is an essential part of the control electronics of a given quantum system. Now one can raise the question, why would I need quantum control theory as an user-side quantum engineer ? I trust the hardware side to have well defined qubits and well defined program to pulses compiler and low latencies? This would be a completely valid question.

It turns out however that in the NISQ era, effort and research is going more and more towards parametric quantum algorithns. Those are hybrid quantum routines with as its core philosophy the offloading of some complicated task to some trainable quantum sytem. For exampl for quantum chemistry tasks, it is beleived that a well designed, physics-inspired quantum antzats could be more efficient than DFT based classical methods at simulating complex quantum molecules. In addition we have other quantum algorithms, such as HHL, QAOA, VQE, and some emergent QML methods which are entirely based on this hybrid quantum classical model, for example the latest demonstration that quantum denoiser models can be achieved.

Parametric quantum circuits are believed to be the emerging candidates for somewhat early quantum advantage, however the only physics-motivated way of expressing their behavior is through Lie Algebras and their quantum dynamical structure. This structure is analyzed through rank analysis, and by extention expressivity analysis, a conceptual category of quantum control theory.

Table of contents

  1. Intuitive approach: writeup of the David J. Tannor, commented
  2. Some paragraph
    1. Sub paragraph
  3. Another paragraph

Introduction and context

The Tannor book, introduction to Quantum Machanics, a time-dependent perspective is mainly organized around a relatively new field of chemical reaction control using light pulses. Chapter 16, Design of Femtosecond pulse Sequences to Control Chemical Reactions is a classic in the genre and quite a niche but interesting activity. However it is interesting for us as it introduces a lot of quantum control ideas that translate immediately to our interest in quantum information science.

Controllability Theory

In the field of system engineering two concepts are central to signal and system control: controllability and observability. Those concepts are relatively easy to think about, for any dynamical "black box" system, defined by it's time evolution S(Li,t)i[0,n]S(L_i, t) \forall i \in [0, n], any state of the system in its parameter space (or all configurations of the dynamical system) can be reached from one state to the other, as well as for any time t[0,T]t\in [0, T], the set of parameters can be measured and observed. This is absolutely foundamental for control, as it allows to create objective and feedback control procedures to automatically bring a system to a desirable value.

In the case of quantum systems it is quite different, for once quantum internal states are not necesserally accessible. A sequence of measurement can discover the state at a certain time but famously in quantum mechanics, measurement interacts with the state, hence constant observation of the system is not possible. It is quite clear that quantum control will have its own perks and particularities.

Quantum control answers the question:

For any state ψi\psi_i at t=0t=0, is it possible to reach any final state ψf\psi_f at t=Tt=T?

(We will see later when going into quantum informational context, this definition might be insufficient)

It turns out that the classical, general theory of controllability can be applied to finate-dimensional dynamical quantum systems without any radical changes. It makes sense as even though quantum mechanics can seem like abstract and alien theory of Nature, the structure of it and predictions is governed by a well understood and defined differential equation (the Schrodinger's equation), and then the dynamical time evolution of the quantum states (defined by either density operators or quantum state vectors) follow some kind of trajectory. The only difference is the direct accessibility of this trajectory, which will create some interesting perks within quantum control. At this time however, infinite-dimensional quantum control is however not completely understood.

The mathematical theory of linear quantum control

Let us suppose a simple multi-dimensional dynamical system described by the following equation:

x˙(t)=Ax(t) \dot{x}(t) = Ax(t)

It is a simple first order equation with AA some n×nn \times n matrix representing the inner dynamics of the system. We know from system theory and calculus 1 that a forcing parameter can be added to transform this free Cauchy problem first order ODE into a dynamically controlled equation:

x˙(t)=Ax(t)+Bu(t) \dot{x}(t) = Ax(t) + Bu(t)

uu is the control (forcing) parameter of MM componants where xx is the state vector of NN componant. We note that for this simple clase:

The equation is said to be controllable if for every x0x_0 and x1x_1 and every T>0T > 0 there exist such a u(t)u(t) such as if x(0)=x0x(0) = x_0 then x(T)=x1x(T)=x_1.

A necessery and sufficient condition for controllability, is

Rank(B,AB,...,AN1B)=N Rank(B, AB, ..., A^{N-1}B) = N

Let's go into more details of this condition: the Rank is the amount of linearly independant columns. I.e. the amount of linearly independant columns of the control and vector matrices (and their products) from the left of the vector matrix, should all from 0 to NN be equal to N. Meaning that if at some point, one of those Rank isn't NN (most probably lower), the system is not controllable, as it exist configurations from one initial state to the final state, not every state can be reached. In terms of pure algebraic concepts, it is clear that if the rank of some configuration cannot be met, it means that the target vector is of lower dimension that the full representation, and then the target space is not being fully spanned and is thus a subspace of all xfx_f.

Let's show this result as the link to Rank is foundamental for quantum control once we translate into the Lie Algebra language.

The formal solution of the first order differential equation write:

x(t)=t0teA(tt)Bu(t)dt+eA(tt0)x(t0) x(t) = \int_{t_0}^{t}e^{A(t - t')}Bu(t')dt' + e^{A(t - t_0)}x(t_0)

If you do not remember how to reach this, it is the usual method to solve ODEs, solving for the reduced form, then finding one solution of the full and summing both, then using the full initial conditions of the Cauchy problem. For simplicity and because we just want some theoretical understanding, let's assume that x(t0)=0x_(t_0) = 0. We have:

x(t)=t0teA(tt)Bu(t)dt x(t) = \int_{t_0}^{t}e^{A(t - t')}Bu(t')dt'

Let us introduce another real analysis tool that is foundamental to this derivation: the Cayley-Hamilton theorem. It states:

Every square matrix over a commutative ring satisfies its own characteristic equation

We note that a ring is a set equipped with two binary operations, the addition and multiplication with the usual properties of fields (with an obsorbant and neutral for both addition and multiplication, internality properties, etc). However a ring doesn't have to be commutative for multiplication nor needs to have a multiplicative inverse. There's plenty of proofs of this, but let's see how we should use it.

Those results are particularely important to understand because the mathematical structure in the quantum context. The matrix exponantial is a matrix function on square matrices analogeous to the ordinary exponantial function. In the theory of Lie groups, the matrix exponantial gives the exponantial map between a matrix Lie Algebra and the corresponding Lie group. We will come back to this.

Let XX be an n×nn\times n real or complex matrix, the exponantial map is given by:

eX=k=01k!Xk e^{X} = \sum_{k=0}^{\infty}\frac{1}{k!}X^k

By interpreting the Cayley-Hamilton as is, combined with a Taylor expention we can find some useful result to determine the minimum representation of the characteristic polynomial. Let's show this.

We assume a scalar function f(s)f(s) that is analytic in a region of the complex plane (hint: that every Cauchy series converges in a singular point in the viscinity of ss). In that region of ss, it is safe to assume that the function can be expressed as a power serie of the form:

f(s)=k=0βksk f(s) = \sum_{k=0}^{\infty}\beta_ks^k

Introduction to the proof

By definition of matrix determinant, if we have a n×nn\times n matrix AA with characteristic polynomial Δ(s)\Delta(s) and eigenvalues λi\lambda_i. As a reminder, the characteristic polynomial is given:

Δ(s)=detsIA=sn+cn1sn1+...+c0 \Delta(s) = det|sI - A| = s^n +c_{n-1}s^{n-1} + ... + c_0

In this equation it is quite clear that if we set s=As = A, we get the characteristic equation:

Δ(A)=An+cn1An1+...+c0I \Delta(A) = A^n + c_{n-1}A^{n-1} + ... + c_0 I The Cayley-Hamilton theorem states that every matrix satisfies it's own characteristic equation, seeing the definition of the characteristic equation it -of course- makes sense as this equation then should always bring to the zero matrix.

Δ(A)[0] \Delta(A) \equiv [0]

Polynomial manipulation allows us to re-write the equation representing the characteristic polynomial, let us assume still AA and a polynomial in ss, let's say P(s)P(s), we can write right away P(s)P(s) as:

P(s)=Q(s)Δ(s)+R(s) P(s) = Q(s)\Delta(s) + R(s)

Where Q(s)Q(s) is found by long division and R(s)R(s) is simply the reminder of the long division like we did in high school. We note two things about this reminder, first it is clear that as AA is of dimension n×nn\times n, R(s)R(s) can only be of rank (n1)(n-1) or less (or Q(s)Q(s) would be a number, i.e. a trivial polynomial). Moreover by definition of eigenvalues, if s=λi,i=1,...,ns=\lambda_i, i=1,...,n,

P(λi)=R(λi) P(\lambda_i) = R(\lambda_i)

The same way, for the matrix itself satisfying it's own characteristic equation (CH theorem)

P(A)=R(A) P(A) = R(A)

Limited series of the matrix exponantial

We saw that locally any analytical fonction will have its Cauchy series converge on one unique point, if this particular analytical function can be represented in the same quotient incorporating the characteristic equation of any matrix, we have:

f(s)=Δ(s)Q(s)+R(s) f(s) = \Delta(s)Q(s) + R(s)

And in particular if we pick the surroundings of s=λis=\lambda_i. We remind ourselves that when we write the extention f(s)=k=0βkskf(s) = \sum_{k=0}^{\infty}\beta_k s^k, we are building the function as a set of orthogonal polynomials spanning the functional space in which ff is defined. If we restrict the space of ss we know that the exponantial expantion will look exactly the same but limited at rank (n1)(n-1) as we showed in the quotient form, when the characteristic equation is being satisfied. We get:

f(λi)=R(λi)=k=0n1αkλik \begin{align} f(\lambda_i) &= R(\lambda_i)\\ &= \sum_{k=0}^{n-1}\alpha_k\lambda_i^k \end{align}

Because the set of eigenvalues is known, the previous equation defines a set of simultaneous linear equations that will generate the (α0,...αn1)(\alpha_0,...\alpha_{n-1}) coefficients. By construction and because the matrix satisfies it's own characteristic equation, the exact same argument can be build for f(A)f(A). We have:

f(A)=k=0βksk=Δ(A)k=0βksk+R(A)=R(A) \begin{align} f(A) &= \sum_{k=0}^{\infty}\beta_ks^k\\ &= \Delta(A)\sum_{k=0}^{\infty}\beta_k s^k + R(A)\\ &= R(A) \end{align}

Since Δ(A)[0]\Delta(A) \equiv [0],

f(A)=k=0n1αkAk f(A) = \sum_{k=0}^{n-1}\alpha_kA^k

Proof

This result will unlock the previous situation, we had the explicit solution for some ODE with control parameter, with matrix exponantial maps. We can write using the previous result:

eA(tt=i=0N1αiAi(tt)i e^{A(t-t'} = \sum_{i=0}^{N-1}\alpha_iA^i(t-t')^i

Where the set {αi}\{\alpha_i\} are some scalar coifficients. By substitution we can place this form in the dynamical solution, we get:

x(t)=t0ti=0N1αiAi(tt)iBu(t)dt x(t) = \int_{t_0}^t\sum_{i=0}^{N-1}\alpha_iA^i(t-t')^iBu(t')dt'

The vector in the form AiBu,i=0,...,N1A^iBu, i=0, ..., N-1 corresponds to directions in the Hilbert space. This last formulation requires some additional interpretation: let's imagine the time interval [t0,t][t_0, t] is set, then the integral is being done at each interval on NN subdivisions of the sum of those vectors weighted by the values of uu at each infinitecimal time step. Let's expand this for more clarity:

x(t)=t0t(α0A0(tt)0Bu(t)dt+α1A1(tt)1Bu(t)dt+...+αN1AN1(tt)N1Bu(t)dt) x(t) = \int_{t_0}^t\big(\alpha_0A^0(t-t')^0Bu(t')dt' + \alpha_1A^1(t-t')^1Bu(t')dt' + ... + \alpha_{N - 1}A^{N-1}(t-t')^{N-1}Bu(t')dt'\big)

We can thus interpret the sum as a set of independant controls, whose strength is being dictated by the values of uu. i.e. NN independant controls during during the (t0t)(t_0 - t) interval.

In principle, once can create any space spanned by the set {AiB}\{A^iB\}, thus if Rank(B,AB,...,AN1B)=NRank(B, AB, ..., A^{N-1}B) = N, then the reachable space is the full Hilbert space, i.e. we have enough independant controls to reach any state.

It turns out that the claim of linear control theory is stronger than reachability, that all the states in this span can be reached within an arbitrary amount of time, so arbitrary short intervals. Within this frame of reference, regardless on how short the interval tt0t - t_0 is, it can be subdivized into NN subintervals giving NN independant controls. Infinitely short times would mean infinitely strong controls, in which regime linear equation of motion are for sure not valid anymore.

Links can be made with strong measurement theory and Hamiltonian driving. If we have some quantum dynamical system described by density operator ρ\rho and total hamiltonian H(t)H(t), the driving from of the Lindbladian can be written

tϱ^(t)=i[H^,ϱ^(t)]+σ2(P)D(ϱ^(t)) \partial_t\hat{\varrho}(t) = -i[\hat{H}, \hat{\varrho}(t)] + \langle \sigma^2(P)\rangle D(\hat{\varrho}(t))

Where we chose a metric system such as =1\hbar = 1, in the driving Von Neumann model, we chose to drive the system by measuring the momentum operator PP, whose diffusor DD is weighted by the measurement probe's standart deviation σ2(P)\langle \sigma^2(P)\rangle. I.e. infinitely fast drive of a quantum system impose infinitely squeezed measurement aparatus or driving parameters. This generalizes really easily in the complex plane.

The mathematical theory of bilinear quantum control

[to be continued]