# A Galactic Simulation of Planck-Scale Physics

Let’s imagine simulating a lattice field theory on a quantum computer. To
summarize how this works, for the uninitiated: *physical* space (the thing
we’re simulating) is discretized by a lattice. To each point on this lattice we
associate some number of qubits, which will represent the state of the fields
at and near that point. Then we apply some sequence of quantum gates mimicking
physical time evolution. If this seems excessively simple, it’s because it
really is (conceptually) very simple.

An unambitious fellow might wish to simulate the standard model; that is, the
fields represented would be the fields of the standard model. In a sufficiently
large such simulation, of course, you’re not *just* simulating the standard
model. The standard model gives rise to protons and neutrons, and you’re
simulating them too. The standard model yields amino acids, and you’re
simulating them too. From the standard model emerges sentient life, and, well,
in a sufficiently large simulation…

But that’s what an unambitious fellow wants. Today I’m feeling ambitious. We’re going to put near-Planck-scale physics on the quantum computer. Take whatever effective field theory holds right below the Planck scale (“right below” means go far enough down that we can discretize space without thinking too much about quantum gravity), and put that on the quantum computer. And now, run as large of a simulation as possible, to see as much emergent behavior at longer distance scales as possible.

The question is, if we do this, how low energy can we probe? Obviously, with infinite space and time, there’s no limit: if you want lower energies, just build a larger quantum computer and simulate a larger lattice on it. But that’s not the world we live in. In the real world, there’s a limit to the computational power of any computer we can build.

The Margolus-Levitin theorem is not the only theorem that gives physical bounds on computation, but it’s the most relevant one I know of. It states that a single computational “operation”—defined as the process of transitioning from one quantum state to an orthogonal state—can happen no faster than \(\hbar / 4 E\), where \(E\) is the average energy of the system in question (relative to the ground state). By the way, the interpretation of this theorem can be tricky; if you want to employ it, it’s important to be confident that transitioning between two orthogonal states really is the right notion of an “operation”.

Okay, back to our quantum simulations. First, what happens if we use the whole universe? The mass of the observable universe is something like \(10^{54}\) kg. By the ML theorem, that means we can accomplish one operation every \(3 \times 10^{-97}\) seconds. If we allow the computation to take as long as the current age of the universe, that’s a total of only \(10^{125}\) operations. “Only”? Well, assume that:

- We run the simulation just long enough for a simulated photon to traverse the entire lattice
- Each site requires exactly one operation per time step (the true value is probably much more)
- The size of a time step is equal to the time it take for light to move from one site to an adjacent site.

Now our lattice is restricted to a side length of a bit more than \(10^{31}\) sites. The Planck length is \(10^{-35}\) meters, so our all-consuming simulation will only manage to simulate a cubic millimeter, and only for \(10^{-12}\) seconds. That’s not bad though: it’s certainly enough to see a bit of chemistry, or maybe even biology, emerge all the way from the Planck scale!

(If you’re in the mood for some numerology, feel free to find significance in the fact that the all-encompassing simulation manages to simulate almost exactly half of the orders of magnitude that we believe exist in our universe.)

Using the entire visible universe and letting the simulation run for the age of the universe is a bit unrealistic. Let’s limit it to a galaxy-sized simulation, and let it run for just 1000 years. The mass of the Milky Way is about \(10^{12}\,M_\odot\), giving us \(10^{104}\) operations to work with. Now we get \(10^{26}\) sites per side, for a simulated length scale of one nanometer. So, we can see an atom. Not terrible.

If we use only the full mass of the Earth, and only for 100 years, the news is still good. We have around \(10^{86}\) operations to work with, giving a simulated size of \(10^{-13}\) meters. This is certainly enough to see the standard model emerge; in fact, it’s roughly comparable to modern lattice QCD simulations!

At some point, though, we need to come to terms with the fact that we’re not going to use the entire mass-energy of the Earth, for 100 years, just to perform a single computation. With a one-ton computer, we get \(10^{64}\) operations. This is sufficient for a length scale of \(10^{-19}\) meters. This is no longer quite enough for us to see the standard model emerge.

Why do this ridiculous calculation? Mainly because I’m putting off doing any actual work. But also, because it serves to illustrate an important truth. No matter what quantum hardware advances come our way, we will never be able to perform a “naive” simulation that demonstrates the emergence of the standard model from Planck scale physics.

This might be fine. One can, for instance, imagine presenting a sequence of effective field theories, starting near the Planck scale and working down. All that needs to be done is convincingly argue that the \(n\)th theory indeed emerges from the \((n-1)\)th, and if the energy scales don’t differ by more than a few orders of magnitude, that argument can even be made nonperturbative with lattice calculations.

Or, we might not be fine. For example, what if there just isn’t any good EFT description between a near-Planck EFT and the standard model? The near-Planck EFT could be, in some sense, as “nice” as possible—a traditional, local quantum field theory with no funny stringy stuff—but we would still be powerless to verify that the standard model indeed emerges.

Less dramatically, even given a set of regularly spaced EFTs spanning the gap
between standard model physics and Planck-scale physics, it’s not immediately
obvious that errors don’t build up enough over “time” to make the whole
calculation untrustworthy. It could be so, but as far as I know *nobody’s asked*.

In general, any plausible calculation to confront this problem is going to need to handle multiple scales with a computational complexity that is strongly sublinear in the ratio of largest to shortest distance. As I write this, to my knowledge, whether such algorithms exist in general is unexplored.