We begin with an interesting observation.
Consider some n x n unitary matrix, and look at its first two columns. They will be orthogonal complex vectors.
import qutip as qt
import numpy as np
n = 4
U = qt.rand_unitary(n)
col0, col1 = U[:,0], U[:, 1]
print(np.vdot(col0, col1))
Then form n spinors, by taking the first component of each from the first column, and the second component of each from the second column.
$ \begin{pmatrix} a & b & c & d \\ e & f & g & h \\ i & j & k & l \\ m & n & o & p \end{pmatrix} \rightarrow \begin{pmatrix} a & b \\ e & f \\ i & j \\ m & n \end{pmatrix} \rightarrow \Big{\{} \begin{pmatrix} a \\ b \end{pmatrix}, \begin{pmatrix} e \\ f \end{pmatrix}, \begin{pmatrix} i \\ j \end{pmatrix}, \begin{pmatrix} m \\ n \end{pmatrix} \Big{\}}$
These spinors in general won't be normalized.
spinors = [qt.Qobj(np.array([col0[i], col1[i]])) for i in range(n)]
print([s.norm() for s in spinors])
Here's the interesting observation. If you convert all those spinors into $(x, y, z)$ vectors, and add them up: their sum will always be 0.
def spinor_xyz(spinor):
return np.array([qt.expect(qt.sigmax(), spinor), qt.expect(qt.sigmay(), spinor), qt.expect(qt.sigmaz(), spinor)])
xyzs = [spinor_xyz(s) for s in spinors]
print(sum(xyzs))
There is a famous theorem due to Minkowski about polyhedra. He tells us that you can describe a polyhedron in terms of its vertices, sure, but you can also uniquely specify a polyhedron by giving: the unit normals to its faces, $\vec{u_{i}}$, and the areas of those faces $a_{i}$.
Given $n$ faces, you can package up this information into $n$ vectors: $a_{i}\vec{u_{i}}$: each such vector points in the direction normal to a face, and has a length equal to the area of the face. The condition that the polyhedron is legit, in other words, that it's closed is just:
$\sum_{i} a_{i}\vec{u_{i}} = 0$
We can therefore see that we can interpret the first two columns of our unitary matrix as specifying a polyhedron: the corresponding spinors the unit normals scaled by the areas. In fact, any two orthogonal complex vectors specify a polyhedron in terms of the components paired off, interpreted as spinors.
import qutip as qt
import numpy as np
import polyhedrec
from vhelper import *
scene = vp.canvas(background=vp.color.white)
n = 6
U = qt.rand_unitary(n)
def unitary_spinors(U):
col0, col1 = U[:,0], U[:, 1]
return [qt.Qobj(np.array([col0[i], col1[i]])) for i in range(n)]
def spinor_xyz(spinor):
return np.array([qt.expect(qt.sigmax(), spinor), qt.expect(qt.sigmay(), spinor), qt.expect(qt.sigmaz(), spinor)])
def make_poly(spinors):
xyzs = [spinor_xyz(s) for s in spinors]
areas = [np.linalg.norm(xyz) for xyz in xyzs]
unormals = [xyz/areas[i] for i, xyz in enumerate(xyzs)]
return polyhedrec.reconstruct(unormals, areas)
vp.sphere(radius=0)
VisualPolyhedron(make_poly(unitary_spinors(U)))
In the above, we use the "polyhedrec" library (available here) to reconstruct the polyhedron from its areas and unit normals: the calculation is a little involved! Also, note that the polyhedron is defined up to 3D rotations. (Also, for $n = 3$, we get a polygon in the plane, which will give the library some trouble, etc.)
Now one thing that this implies is that there is an action of the unitary group on polyhedra.
import qutip as qt
import numpy as np
import polyhedrec
from vhelper import *
scene = vp.canvas(background=vp.color.white)
n = 6
dt = 0.001
U = qt.rand_unitary(n)
V = (-1j*qt.rand_herm(n)*dt).expm()
def unitary_spinors(U):
col0, col1 = U[:,0], U[:, 1]
return [qt.Qobj(np.array([col0[i], col1[i]])) for i in range(n)]
def spinor_xyz(spinor):
return np.array([qt.expect(qt.sigmax(), spinor), qt.expect(qt.sigmay(), spinor), qt.expect(qt.sigmaz(), spinor)])
def make_poly(spinors):
xyzs = [spinor_xyz(s) for s in spinors]
areas = [np.linalg.norm(xyz) for xyz in xyzs]
unormals = [xyz/areas[i] for i, xyz in enumerate(xyzs)]
return polyhedrec.reconstruct(unormals, areas)
vpoly = VisualPolyhedron(make_poly(unitary_spinors(U)))
print("total area: %f" % sum([np.linalg.norm(spinor_xyz(s)) for s in unitary_spinors(U)]))
for t in range(100):
U = V*U
vpoly.update(make_poly(unitary_spinors(U)))
vp.rate(500)
print("total area: %f" % sum([np.linalg.norm(spinor_xyz(s)) for s in unitary_spinors(U)]))
In fact, we can see that not only does the action of a unitary on the polyhedron preserve the closure constraint, but also: it preserves the total surface area of the polyhedron!
Now suppose we had any old random set of vectors in 3D, which doesn't necessarily satisfy the closure constraint. It turns out we can easily calculate a Lorentz boost, which acts on the spinors, and closes the polyhedron.
We form the projector for each spinor, and sum them all: $S = \sum_{i} \mid \phi_{i} \rangle \langle \phi_{i} \mid$. We then use singular value decomposition to split $S$ into $S = UDV$, where $U$ and $V$ are unitaries, and $D$ is a diagonal matrix. We then form a boost $B = det(D)^{-\frac{1}{4}}U\sqrt{D}$, and actually the boost we want is just the inverse: $B^{-1}$. If we apply this $SL(2, \mathbb{C}$ matrix to each spinor, we'll end up with a closed polyhedron.
One implication is that if we imagine our spinors refering to the directions and energies of incoming photons on the celestial sphere, there's always a velocity we could be traveling so that the sum of the energy/direction vectors is 0: and they form a closed polyhedron.
import qutip as qt
import numpy as np
def spinor_xyz(spinor):
return np.array([qt.expect(qt.sigmax(), spinor), qt.expect(qt.sigmay(), spinor), qt.expect(qt.sigmaz(), spinor)])
def close_spinors(spinors):
S = sum([spinor*spinor.dag() for spinor in spinors])
U, d, V = np.linalg.svd(S.full())
D = np.diag(d)
rho = np.linalg.det(D)
B = np.dot(U, np.sqrt(D)*rho**(-1./4.))
Bi = qt.Qobj(np.linalg.inv(B))
return Bi, [Bi*spinor for spinor in spinors]
def check_closure(spinors):
return sum([spinor_xyz(s) for s in spinors])
n = 5
spinors = [np.random.randn(1)[0]*qt.rand_ket(2) for i in range(n)]
print("closure: %s" % check_closure(spinors))
boost, closed_spinors = close_spinors(spinors)
print("closure: %s" % check_closure(closed_spinors))
Furthermore, it's worth noting we can act with our unitary directly on the set of spinors. For each spinor $\phi_{i}$:
$\phi_{i} \rightarrow \sum_{i,j} U_{i,j}\phi_{j}$
import qutip as qt
import numpy as np
import polyhedrec
from vhelper import *
scene = vp.canvas(background=vp.color.white)
n = 6
dt = 0.001
U = qt.rand_unitary(n)
V = (-1j*qt.rand_herm(n)*dt).expm()
def unitary_spinors(U):
col0, col1 = U[:,0], U[:, 1]
return [qt.Qobj(np.array([col0[i], col1[i]])) for i in range(n)]
def spinor_xyz(spinor):
return np.array([qt.expect(qt.sigmax(), spinor), qt.expect(qt.sigmay(), spinor), qt.expect(qt.sigmaz(), spinor)])
def make_poly(spinors):
xyzs = [spinor_xyz(s) for s in spinors]
areas = [np.linalg.norm(xyz) for xyz in xyzs]
unormals = [xyz/areas[i] for i, xyz in enumerate(xyzs)]
return polyhedrec.reconstruct(unormals, areas)
spinors = unitary_spinors(U)
vpoly1 = VisualPolyhedron(make_poly(spinors), [-1, 0, 0])
vpoly2 = VisualPolyhedron(make_poly(spinors), [1, 0, 0])
for t in range(50):
U = V*U
spinors = [sum([V.full()[i][j]*spinors[j]\
for j in range(n)])\
for i in range(n)]
vpoly1.update(make_poly(unitary_spinors(U)))
vpoly2.update(make_poly(spinors))
vp.rate(500)
s0, s1 = spinors, unitary_spinors(U)
print(np.array([np.isclose((s0[i]-s1[i]).full().T[0], np.array([0,0])).all() for i in range(n)]).all())
So it's interesting that we can fully specify a classical polyhedron in terms of its unit normals and areas, expressed for example as a set of spinors. This provides a clue for how to quantize polyhedra and obtain the quantum polyhedron, which we can regard as an "atom of space". In fact, we've already done it, we just didn't realize what we were doing!
The basic idea is this. We quantize each spinor, representing a polyhedral face, as a spin-$j$ particle. The $J^{2}$ operator $J^{2} = X^{2} + Y^{2} + Z^{2}$ becomes the area operator (squared) of the face. We then consider the tensor product of the $n$ spin-$j$ particles representing the $n$ faces. We form the total spin operator $J_{total}^{2} = (\sum_{i} X_{i})^{2} + (\sum_{i} Y_{i})^{2} + (\sum_{i} Z_{i})^{2}$, and demand that our states live in the angular momentum $0$ eigenspace. This constraint is the quantum equivalent of the closure constraint, as we will see.
We've actually already considered this.
For example, suppose we have four spin-$\frac{1}{2}$ states. The angular momentum 0 subspace of their tensor product is two dimensional.
$H_{\frac{1}{2}} \otimes H_{\frac{1}{2}} \otimes H_{\frac{1}{2}} \otimes H_{\frac{1}{2}} = (H_{0} \oplus H_{0}) \oplus (H_{1} \oplus H_{1} \oplus H_{1}) \oplus (H_{2})$.
This means that we can use a single qubit to parameterize the angular momentum 0 subspace of four spin-$\frac{1}{2}$ states. Each of the qubit basis vectors is sent to one of the two eigenvectors of $J_{total}^{2}$ with eigenvalue 0. In general, this subspace will be larger (or smaller), but we could always use a nominal spin-$j$ particle to parameterize the angular momentum 0 subspace of the tensor product. Depending on the $j$ values of the tensored spins, there may not even be an angular momentum 0 subspace at all: and so these spins can't form the faces of a quantum polyhedron.
For each face spin, we have a set of $X$, $Y$, $Z$ operators, each tensored appropriately with identities so they act on only their own spin. Now you might think that the vector corresponding to each face would just be given by the expectation values of X, Y, Z on the face spin, but that's not how it works.
Instead: define the following "inner product operators" between two spins $\Phi_{i}$ and $\Phi_{j}$:
$ \Phi_{i} \cdot \Phi_{j} = X_{i}X_{j} + Y_{i}Y_{j} + Z_{i}Z_{j}$
For concreteness, suppose we have four faces. Consider the following matrix of operators:
$\begin{pmatrix} \Phi_{0} \cdot \Phi_{0} & \Phi_{0} \cdot \Phi_{1} & \Phi_{0} \cdot \Phi_{2} & \Phi_{0} \cdot \Phi_{3} \\ \Phi_{1} \cdot \Phi_{0} & \Phi_{1} \cdot \Phi_{1} & \Phi_{1} \cdot \Phi_{2} & \Phi_{1} \cdot \Phi_{3} \\ \Phi_{2} \cdot \Phi_{0} & \Phi_{2} \cdot \Phi_{1} & \Phi_{2} \cdot \Phi_{2} & \Phi_{2} \cdot \Phi_{3} \\ \Phi_{3} \cdot \Phi_{0} & \Phi_{3} \cdot \Phi_{1} & \Phi_{3} \cdot \Phi_{2} & \Phi_{3} \cdot \Phi_{3} \end{pmatrix}$
Along the diagonal, we have the area operators. And the off-diagonal terms correspond to all the possible inner products between the face vectors, which relate to the cosines of the angles between the faces: we can think of these like "angle" operators.
Given some polyhedron state $\mid \pi \rangle$, we can take its expectation values on each of the operators above, and form a matrix out of them:
$G_{i,j} = \langle \pi \mid \Phi_{i} \cdot \Phi_{j} \mid \pi \rangle$
We then decompose $G$ using the singular value decomposition: $G = UDV$. We consider the first three rows of $V$, and turn them into columns and make a matrix $V^{\prime}$. We also consider the first three columns/rows of $\sqrt{D}$, and form $D^{\prime}$. We then form the matrix:
$ M = V^{\prime}D^{\prime}$.
The rows of this matrix will be the (x, y, z) coordinates of the face vectors. In other words, their norms will be the areas of the faces, and their directions will be the unit normals to each face.
The quantum closure constraint guarantees that these (x, y, z) vectors all sum to 0, and so represent a closed polyhedron: although recall that now these vectors are obtained via the expectation values of quantum operators!
Before we demonstrate the quantum version, let's just make sure that this procedure works in the case of a classical polyhedron!
import qutip as qt
import numpy as np
import polyhedrec
def unitary_spinors(U):
col0, col1 = U[:,0], U[:, 1]
return [qt.Qobj(np.array([col0[i], col1[i]])) for i in range(n)]
def spinor_xyz(spinor):
return np.array([qt.expect(qt.sigmax(), spinor), qt.expect(qt.sigmay(), spinor), qt.expect(qt.sigmaz(), spinor)])
def vectors(spinors):
return [spinor_xyz(s) for s in spinors]
def inner_products(vectors):
return np.array([[np.dot(vectors[i], vectors[j]) for j in range(len(vectors))] for i in range(len(vectors))])
def recover_vectors(G):
U, D, V = np.linalg.svd(G)
M = np.dot(V[:3].T, np.sqrt(np.diag(D[:3])))
return [m for m in M]
n = 6
U = qt.rand_unitary(n)
vecs = vectors(unitary_spinors(U))
G = inner_products(vecs)
vecs2 = recover_vectors(G)
G2 = inner_products(vecs2)
print(G)
print()
print(G2)
[[ 0.3120559 -0.07817091 -0.01945402 -0.09603503 -0.11907589 0.00067995] [-0.07817091 0.07792043 -0.00179758 -0.04054626 0.08233279 -0.03973846] [-0.01945402 -0.00179758 0.01470451 0.03393009 -0.0417283 0.01434531] [-0.09603503 -0.04054626 0.03393009 0.13429133 -0.09117618 0.05953605] [-0.11907589 0.08233279 -0.0417283 -0.09117618 0.23895166 -0.06930408] [ 0.00067995 -0.03973846 0.01434531 0.05953605 -0.06930408 0.03448124]] [[ 0.3120559 -0.07817091 -0.01945402 -0.09603503 -0.11907589 0.00067995] [-0.07817091 0.07792043 -0.00179758 -0.04054626 0.08233279 -0.03973846] [-0.01945402 -0.00179758 0.01470451 0.03393009 -0.0417283 0.01434531] [-0.09603503 -0.04054626 0.03393009 0.13429133 -0.09117618 0.05953605] [-0.11907589 0.08233279 -0.0417283 -0.09117618 0.23895166 -0.06930408] [ 0.00067995 -0.03973846 0.01434531 0.05953605 -0.06930408 0.03448124]]
We can see that if we have a bunch of vectors satisfying the closure constraint, we can form $G$, the matrix of the inner products between them. We can then recover the vectors via the above method up to a 3D rotation. Indeed, we don't get exactly the vectors back, but we can see that they are the same up to a 3D rotation, since the $G$ matrix formed from the recovered vectors is identical.
Putting it all together, we can construct a quantum polyhedron in the following way. We pick $j$ values for each of the $n$ faces, and form the tensor product of $n$ spins, one for each face. We constrain ourselves to work only in the spin-$0$ subspace of the tensor product, and so depending on the dimensionality $d$ we can represent this as a "qudit" of $d$ dimensions. We construct the inner product operators out of all the $X_{i}$'s, $Y_{i}$'s, and $Z_{i}$'s, and given some polyhedron state $\mid \pi \rangle$, we form the matrix of the expectation values on all these operators. We then recover the face vectors using the singular value decomposition. At last, using Minkowski's Theorem, we can construct a polyhedron out of these statistical averages. Because we're stuck in the spin-$0$ subspace, the face vectors will always form a (possibly degenerate) closed polyhedron.
In fact, for efficiency's sake, we can downgrade all the inner product operators to act on the qudit that spans the spin-$0$ subspace instead of the much larger tensor product. We'll already have to construct an operator $S$ that takes the qudit state to a state in the original tensor product, so that $\mid \pi \rangle = S\mid q \rangle$. We can use this same operator to upgrade operators that act on $\mid q \rangle$ to operators that act on $\mid \pi \rangle$: $O_{big} = SO_{tiny}S^{\dagger}$, and downgrade operators that act on $\mid \pi \rangle$ to operators that act on $\mid q \rangle$: $O_{tiny} = S^{\dagger}O_{big}S$. And so, we can downgrade all our inner product operators to act on the qudit.
And actually, this is a good idea anyway, since the eigenstates of inner product operators acting on the tensor product don't all live in the spin-$0$ subspace. So downgrading the operators makes sure their eigenstates all live in the spin-$0$ subspace. Lest this worry you, this doesn't effect the expectation values in any way, and if you downgrade one of these operators and upgrade it back, the resulting operator commutes with the original operator.
We can then evolve our quantum polyhedron under some unitary matrix just as we could evolve our classical polyhedron, by choosing some unitary that acts on the underlying qudit.
Check out quantum_polyhedron.py!
Choose a set of $j$ values for your faces. You can then evolve the polyhedron in time using a Hamiltonian that acts on the underlying qudit: p.evolve(H=qt.rand_herm(p.d), dt=0.01, t=1000)
. You can also measure one of the operators, for example: p.measure(p.INNER_PRODUCTS[0][1], tiny=False)
. Operators are assumed to act on the tensor product, and are automatically downgraded to act on the qudit. If you already have a "tiny" operator, set tiny=True
. If the polyhedron is degenerate, or if something otherwise goes wrong, the underlying vectors will be displayed instead of the polyhedron. Finally, you'll notice that the polyhedron glitches out from time to time: this is because the polyhedron is only defined up to 3D rotation, and so when we recover the polyhedron at each time step, we might end up with a arbitrarily rotated version. It would be nice to easily calculate the 3D rotation that will make sure to align the orientation of the polyhedron over time, but it's a little tricky!
Note as well that just as before: the total area of the polyhedron is preserved under unitary evolution, and under measurements too, for that matter. And actually, it's a little more than before: the areas of each of the individual faces is preserved as well.
Although we can visualize a quantum polyhedron as a classical polyhedron built of expectation values, don't be mislead: there is something distinctively quantum going on here.
The crucial fact is that, in general, the inner product operators don't commute. This means that we can't "simultaneously know" all the angles between the faces of the quantum polyhedron at the same time.
For example, if we measure the angle between the first two faces, we'll end up in an eigenstate of that operator. But this will lead to uncertainty in the angles between the other faces: we won't be in an eigenstate of those operators.
from quantum_polyhedron import *
p = QuantumPolyhedron([1/2, 1/2, 1/2, 1/2])
IP01 = p.INNER_PRODUCTS_[0][1]
IP12 = p.INNER_PRODUCTS_[1][2]
print("Eigenvalues for angle between faces 0 and 1: %s" % IP01.eigenstates()[0])
print("Eigenvalues for angle between faces 1 and 2: %s" % IP12.eigenstates()[0])
print()
print(p.inner_products())
print()
p.measure(p.INNER_PRODUCTS[0][1])
print()
print(p.inner_products())
Eigenvalues for angle between faces 0 and 1: [-0.75 0.25] Eigenvalues for angle between faces 1 and 2: [-0.75 0.25] [[ 0.75 -0.71718557 -0.17059889 0.13778447] [-0.71718557 0.75 0.13778447 -0.17059889] [-0.17059889 0.13778447 0.75 -0.71718557] [ 0.13778447 -0.17059889 -0.71718557 0.75 ]] -0.75 [[ 0.75 -0.75 0. 0. ] [-0.75 0.75 0. 0. ] [ 0. 0. 0.75 -0.75] [ 0. 0. -0.75 0.75]]
We can see above that if we measure the inner product between faces 0 and 1 or between faces 1 and 2, we can either $-\frac{3}{4}$ or $\frac{1}{2}$ as eigenvalues. If we measure the first operator, we collapse to one of its eigenstate. So the $G_{0,1}$ entry of the matrix goes to $-\frac{3}{4}$ or $\frac{1}{2}$: but we also see that if we do so, then the $G_{1,2}$ entry of the matrix will never be one of the eigenvalues! And hence, there is a quantum uncertainty about the inner product between them: in fact, in this case, the matrix entry is either $-\frac{1}{2}$ or $0$.
In fact, we see here that when the angle between 0 and 1 is definite, then so is the angle between 2 and 3. So we can know two of the angles at once, but not the others!
In other words, the polyhedron can't simulatenously be in an eigenstate of all its inner product operators, and so there is a complementarity between knowing the different angles between pairs of faces. And yet, from the expectation values of the quantum polyhedron, we can form a classical polyhedron which satisfies the closure condition.
It's analogous to the case of spin, say, a qubit. We can form a representation of its rotation axis as an (x, y, z) vector formed from the expectation values of the X, Y, Z operators. This is just the state of a classical spinning object, but it is built out of statistical averages. The really quantum part comes in when we go to measure the spin: the X, Y, Z operators don't commute, and so if we collapse to an X eigenstate, this will lead to uncertainty in Y and Z: we'll end up in a superposition of Y or Z eigenstates. So that choosing to measure along X, Y, or Z precludes us from knowing the spin in the other directions. Nevertheless, being in an X eigenstate, with uncertainty in Y and Z, corresponds to: yet another classical state of the spinning object, this time with its rotation axis pointing along the X direction. Moreover, we can recover the original state by performing the experiment many times on identically prepared states, measuring X, Y, Z in turn, over and over again.
Similarly, in the case of a quantum polyhedron, we can form a representation of it as a classical polyhedron, built out of expectation values. The inner product operators don't all commute, and so if we collapse to an eigenstate of one of them, this leads to uncertainty in the other operators, which precludes us from knowing those angles. By measuring the one angle, we've renounced knowledge of the others. Nevertheless, the resulting state corresponds to: yet another classical polyhedron in terms of the expectation values. Moreover, we can recover the original state by performing repeated experiments on identically prepared polyhedra, measuring the different angles in turn, over and over again.
One way of interpreting a tensor product of spins in the spin-$0$ subspace is as an interaction vertex that preserves angular momentum. In other words, it represents a possible interaction such that the total angular momentum coming in is the same as the total angular momentum going out.
To see this, let us have recourse to our symmeterized spinor formalism for a moment.
Suppose we have two spin-$\frac{1}{2}$ states and we want to contract them in an angular momentum invariant way.
Back in the string diagram days, we had a way of contracting states: the cup.
When we work with spins, it's useful to use a different set of cups and caps:
Instead of using the $\mid \uparrow \uparrow \rangle + \mid \downarrow \downarrow \rangle$ state, we use the antisymmetric state. This is often called $\epsilon$.
Suppose we want to contract two spin-$j$ states in an angular momentum way. What's the right cap? For concreteness, take two spin-$\frac{3}{2}$ states. Construct the permutation symmetric tensor product representation of each. Each state has three stars, and so can be represented as a symmeterized tensor product of three spin-$\frac{1}{2}$ states. Use the antisymmetric cup to contract each of the spin-$\frac{1}{2}$'s with their pair on the other side.
We can also see that we can use the $\epsilon$ to raise and lower indices: to go from bra's to ket's in an angular momentum invariant way. The point is that if you yank a spin, its spin axis should flip: incoming and outgoing spins should have opposite senses.
If we wanted to work with the spin states in the $\mid j, m \rangle$ representations directly, we could imagine constructing the $\epsilon_{j}$ by tensoring $2j$ copies of the $\epsilon_{\frac{1}{2}}$ together, permuting them so that all $\epsilon$ inputs are on the left and all the $\epsilon$ outputs are on the right. Call this $\epsilon_{sym-j}$. We have some $S$, for which $S\mid\psi_{sym}\rangle = \mid\psi_{spin}\rangle$: it acts on a permutation symmetric state of $2^{2j}$ dimensions and downgrades us to the $2j+1$ dimensional $\mid j, m\rangle$ state. So we apply $(S \otimes S)$ to $\epsilon_{sym-j}$ to get $\epsilon_{j}$, which let's us contract two spin-$j$ particles directly.
import qutip as qt
import numpy as np
from magic import *
from quantum_polyhedron import *
def repair(q):
q.dims[1] = [1]*len(q.dims[0])
return q
def epsilonj(j):
n = int(2*j)
epsilons = qt.tensor(*[np.sqrt(2)*qt.bell_state("11")]*(n))
epsilons = epsilons.permute(list(range(0, 2*n, 2))+list(range(1, 2*n, 2)))
S = sym_spin(n)
return repair(qt.tensor(S, S)*epsilons)
def contract(qpoly, input_spin, index, verbose=True, restore_norm=True):
input_j = qpoly.js[index]
epsilon = (np.sqrt(2*input_j +1) if restore_norm else 1)*epsilonj(input_j)
state = qt.tensor(epsilon, input_spin, qpoly.zero_map*qpoly.spin)
output_spins = qt.tensor_contract(state, (1, index+3))
output_spins = qt.tensor_contract(output_spins, (0, 1))
if verbose:
X0, Y0, Z0, J20 = make_total_spin([input_j])
removed_indices = qpoly.js[:]
del removed_indices[index]
X1, Y1, Z1, J21 = make_total_spin(removed_indices)
input_xyz = np.array([qt.expect(sum(X0), input_spin),\
qt.expect(sum(Y0), input_spin),\
qt.expect(sum(Z0), input_spin)])
output_xyz = np.array([qt.expect(sum(X1), output_spins),\
qt.expect(sum(Y1), output_spins),\
qt.expect(sum(Z1), output_spins)])
print("input ang mo: %s " % input_xyz)
print("output ang mo: %s " % output_xyz)
print("difference: %s" % (input_xyz-output_xyz))
return repair(output_spins)
p = QuantumPolyhedron([1/2,1/2,1/2,1/2])
contract(p, qt.rand_ket(2), 0)
input ang mo: [-0.22010705 -0.258395 -0.36713065] output ang mo: [-0.22010705 -0.258395 -0.36713065] difference: [1.11022302e-16 5.55111512e-17 6.10622664e-16]
Above we have a quantum tetrahedron with four spin-$\frac{1}{2}$'s as faces. We consider an incoming spin-$\frac{1}{2}$ at random. Using the $\epsilon$, we contract the incoming spin with the first face. This steers us to a state of three outgoing spin-$\frac{1}{2}$'s, and comparing the angular momentum of the incoming spin with the total angular momentum of the outgoing spins, we find that they are exactly equal. This would not be true if the four spins corresponding to the faces weren't in the spin-$0$ subspace. In other words, the closure of our polyhedron corresponds to the conservation of angular momentum.
Let's dig a little deeper.
Below we have code that lets you contract arbitrary numbers of spins onto the faces of a polyhedron. In general, however, the norm of the resulting state may not be 1: this happens when it's impossible to satisfy the angular momentum constraint given the input up to a scalar. The more you contract it may even be impossible to satisfy the angular momentum constraint at all. You may be asking too much if you just choose any old input spins. If you contract all the spins, you get a scalar: the probability that those incoming spins would emerge as the output spins identical to the spins of the polyhedron. In other words, that those spins, coming together, would form the polyhedron: or that: those spins could pass through the polyhedron unaffected, merely flipping from incoming to outgoing.
We also have code that can turn a bra spin-$j$ into a ket spin-$j$ using the $\epsilon$. We also check the consistency of the scheme by a) finding the (antisymmetric) inner product between polyhedron and itself to be 1 b) Showing that we can vary the order of contraction and still get the same final amplitude c) Contracting ever more indices, and checking whether the angular momentum is conserved. (You make note some trickiness with the normalization.)
import qutip as qt
import numpy as np
from magic import *
from quantum_polyhedron import *
def repair(q):
q.dims[1] = [1]*len(q.dims[0])
return q
def epsilonj(j):
n = int(2*j)
epsilons = qt.tensor(*[np.sqrt(2)*qt.bell_state("11")]*(n))
epsilons = epsilons.permute(list(range(0, 2*n, 2))+list(range(1, 2*n, 2)))
S = sym_spin(n)
return repair(qt.tensor(S, S)*epsilons)
def raise_indices(spins):
input_js = [(d-1)/2 for d in spins.dims[0]]
n = len(input_js)
epsilons = qt.tensor(*[epsilonj(input_j)\
for input_j in input_js])
state = repair(qt.tensor(epsilons, spins))
pairs = [(2*i+1, 2*n+i) for i in range(n)]
return repair(qt.tensor_contract(state, *pairs)).conj()
def contract(poly_spins, input_spins, indices, verbose=False, restore_norm=False):
js = [(d-1)/2 for d in poly_spins.dims[0]]
ns = poly_spins.dims[0][:]
n = len(js)
input_js = [js[index] for index in indices]
input_ns = [ns[index] for index in indices]
if len(indices) == n:
epsilons = qt.tensor(*[epsilonj(input_j) for input_j in input_js])
epsilons = epsilons.permute(list(range(0, 2*n, 2))+list(range(1, 2*n, 2)))
output = epsilons.dag()*qt.tensor(input_spins, poly_spins)
output.dims = [[1],[1]]
return output
else:
epsilons = qt.tensor(*[epsilonj(input_j)*(np.sqrt(2*input_j+1) if restore_norm else 1) for input_j in input_js])
offset = 3*len(indices)
state = repair(qt.tensor(epsilons, input_spins, poly_spins))
pairs = [(2*i, 2*len(indices)+i) for i in range(len(indices))]+\
[(2*i+1, index+offset) for i, index in enumerate(indices)]
output_spins = repair(qt.tensor_contract(state, *pairs))
if verbose:
X0, Y0, Z0, J20 = make_total_spin(input_js)
removed_indices = np.delete(js, indices)
X1, Y1, Z1, J21 = make_total_spin(removed_indices)
input_xyz = np.array([qt.expect(sum(X0), input_spins),\
qt.expect(sum(Y0), input_spins),\
qt.expect(sum(Z0), input_spins)])
output_xyz = np.array([qt.expect(sum(X1), output_spins),\
qt.expect(sum(Y1), output_spins),\
qt.expect(sum(Z1), output_spins)])
uinput_xyz = input_xyz/np.linalg.norm(input_xyz)
uoutput_xyz = output_xyz/np.linalg.norm(output_xyz)
print("input ang mo direction: %s " % uinput_xyz)
print("output ang mo direction: %s " % uoutput_xyz)
print("difference: %s" % (uinput_xyz-uoutput_xyz))
print("norm of output: %s" % output_spins.norm())
print("actual difference: %s" % (input_xyz-output_xyz))
return output_spins
####################################################################################
p = QuantumPolyhedron([1,1/2,1/2,2])
print("epsilon norm: %s" % contract(p.big(), raise_indices(p.big()), list(range(p.nfaces)))[0][0][0].real)
print()
def T1(p, cut):
test1 = qt.rand_ket(np.prod(p.ns[:cut]))
test1.dims = [p.ns[:cut], [1]*len(p.ns[:cut])]
test2 = qt.rand_ket(np.prod(p.ns[cut:]))
test2.dims = [p.ns[cut:], [1]*len(p.ns[cut:])]
test = qt.tensor(test1, test2)
output = contract(p.big(), test1, list(range(cut)))
output2 = contract(output, test2, list(range(p.nfaces-cut)))
output_ = contract(p.big(), test, list(range(p.nfaces)))
print(output2[0][0][0], output_[0][0][0])
for i in range(1, p.nfaces):
T1(p, i)
print()
def T2(p, cut):
test = qt.rand_ket(np.prod(p.ns[:cut]))
test.dims = [p.ns[:cut], [1]*len(p.ns[:cut])]
return contract(p.big(), test, list(range(cut)), verbose=True, restore_norm=True)
for i in range(1, p.nfaces+1):
print(T2(p, i))
print()
epsilon norm: 1.0 (0.09017490519585734+0.1550025552826686j) (0.09017490519585734+0.15500255528266868j) (0.025834970065418843+0.11292303722730597j) (0.02583497006541885+0.11292303722730598j) (0.012665934530754444+0.10244795227287351j) (0.012665934530754461+0.1024479522728735j) input ang mo direction: [-0.43990165 -0.07490741 0.89491643] output ang mo direction: [-0.43990165 -0.07490741 0.89491643] difference: [-1.16573418e-15 7.35522754e-16 -4.44089210e-16] norm of output: 1.0000000000000009 actual difference: [ 4.99600361e-16 6.31439345e-16 -2.66453526e-15] Quantum object: dims = [[2, 2, 5], [1, 1, 1]], shape = (20, 1), type = ket Qobj data = [[ 0. +0.j ] [ 0. +0.j ] [-0.00463925-0.26619788j] [-0.10543949+0.15077893j] [-0.31335062-0.09386608j] [ 0. +0.j ] [ 0.00568189+0.32602449j] [ 0.08609098-0.12311048j] [ 0.15667531+0.04693304j] [ 0. +0.j ] [ 0. +0.j ] [ 0.00568189+0.32602449j] [ 0.08609098-0.12311048j] [ 0.15667531+0.04693304j] [ 0. +0.j ] [-0.01136378-0.65204898j] [-0.10543949+0.15077893j] [-0.12792485-0.03832066j] [ 0. +0.j ] [ 0. +0.j ]] input ang mo direction: [-0.76286876 0.64096815 -0.08480026] output ang mo direction: [-0.6665285 0.61026386 -0.42815626] difference: [-0.09634026 0.03070429 0.343356 ] norm of output: 1.0341542911199557 actual difference: [ 0.10837021 -0.1216866 0.22792596] Quantum object: dims = [[2, 5], [1, 1]], shape = (10, 1), type = ket Qobj data = [[ 0. +0.j ] [-0.09703006+0.12170861j] [-0.05472108+0.3337494j ] [-0.45954519-0.18306765j] [ 0.36519082+0.26509928j] [ 0.19406012-0.24341722j] [ 0.06701936-0.40875787j] [ 0.37521708+0.14947411j] [-0.18259541-0.13254964j] [ 0. +0.j ]] input ang mo direction: [ 0.3223151 -0.89603035 0.3053565 ] output ang mo direction: [ 0.82539529 -0.54845303 -0.13387262] difference: [-0.5030802 -0.34757732 0.43922912] norm of output: 1.0258516556700248 actual difference: [-0.21848746 -0.10499153 0.16665054] Quantum object: dims = [[5], [1]], shape = (5, 1), type = ket Qobj data = [[ 0.34144121+0.105266j ] [-0.51998612+0.03919313j] [-0.28119882+0.23753023j] [-0.56829395+0.36022808j] [ 0.25372703-0.01392283j]] Quantum object: dims = [[1], [1]], shape = (1, 1), type = bra Qobj data = [[0.05014693+0.01986517j]]
Also check out interaction_vertex.py! Choose a random polyhedral vertex, and some spins for inputs. You'll see the density matrices of the inputs displayed below, and the density matrices of the outputs above. You can evolve the polyhedron with evolve("polyhedron", qt.rand_herm(p.d))
, keeping the inputs the same. Or you can evolve the inputs, keeping the polyhedron the same, e.g. evolve("input", qt.jmat(1/2, 'x'))
. The opacity of the density matrix's sphere is its norm. And the opacities of the stars are their probabilities. Stars of the color belong to the same eigenstate of the DM. Try filling all the input slots to form an amplitude!
Another thing we could do is glue polyhedra together. We simply contract two faces together.
import qutip as qt
import numpy as np
from quantum_polyhedron import *
scene = vp.canvas(background=vp.color.white, width=1000, height=500)
def glue(p1, p2, pair, verbose=True, pos=[0,0,0]):
p1_state = p1.big()
p2_state = p2.big()
j = (p1_state.dims[0][pair[0]]-1)/2
epsilon = np.sqrt(2*j+1)*epsilonj(j)
state = qt.tensor(epsilon, p1.big(), p2.big())
state = repair(qt.tensor_contract(state, (0, 2+p1.nfaces+pair[1]), (1, 2+pair[0])))
new_js = [(d-1)/2 for d in state.dims[0]]
if verbose:
X, Y, Z, J2 = make_total_spin(new_js)
print("glued polyhedron closed?: total ang mo = %s" % qt.expect(J2, state))
new_poly = QuantumPolyhedron(new_js, show_poly=True, pos=pos, initial=state)
return new_poly
p1 = QuantumPolyhedron([1/2,1/2,1/2,1/2], show_poly=True, pos=[-1,0,0])
p2 = QuantumPolyhedron([1/2,1/2,1/2,1/2], show_poly=True, pos=[1,0,0])
p3 = glue(p1, p2, (0,0), pos=[0,1,0])
glued polyhedron closed?: total ang mo = -1.945965821749531e-17
Combining polyhedra. Spin networks. Entangled polyhedra.
Tensor product: a site of interaction: an atom of space The loops picture.
Variable spin-j: quantize arbitrary polyhedra: coherent polyhedra. In the limit of infinite faces: a quantum sphere.
one spin goes thru