## Friday, May 23, 2014

### Conference Fees for Speakers

Discuss!

## Wednesday, January 29, 2014

### Questions to the inter webs: classical 't-Hooft-limit and path integral entanglement

I am coming back to you with a new format: Questions. Let me start with two questions I have been thinking about recently but that I don't know a good answer to.

#### 't Hooft limit of classical field equations

#### Path Integral Entanglement

**the**distinguishing feature of quantum theory as compared to classical physics. It is closely tied to the non-comutativity of the observable algebra and is responsible for things like the violation of Bell's inequality.

## Tuesday, November 06, 2012

### A Few Comments On Firewalls

Now, of course I had to read (some of) the papers and I have to say that I am confused. I admit, I did not get the point. Even more, I cannot understand a large part of the discussion. There is a lot of prose and very little formulas and I have failed to translate the prose to formulas or hard facts for myself. Many of the statements taken at face value do not make sense to me but on the other hand, I know the authors to be extremely clever people and thus the problem is most likely on my end.

In this post, I would like to share some of my thoughts in my endeavor to decode these papers but probably they are to you even more confusing than the original papers to me. But maybe you can spot my mistakes and correct me in the comment section.

I had a long discussion with Cristiano Germani on these matters for which I am extremely grateful. If this post contains any insight it is his while all errors are for course mine.

### What is the problem?

### The classical situation

A decaying black hole, image stolen from Sabine Hossenfelder. |

Two regions M and N on a Cauchy surface C with their causal shadows |

### States

First of all (and quite trivially), a state is always at one instant of time, that is it lives on a Cauchy surface (or at least a space-like hyper surface, as our space-time might not be globally hyperbolic), not in a region of space-time. Hilbert space, as the space of (pure) states thus also lives on a Cauchy surface (and not for example in the region behind the horizon). If one event is after another (i.e. in its forward light-cone) it does not make sense to say they belong to different tensor factors of the Hilbert (or different Hilbert spaces for that matter).

Furthermore, a state is always a global concept, it is everywhere (in space, but not in time!). There is nothing like "the space of this observer". What you can do of course is restrict a state to a subset of observables (possibly those that are accessible to one observer) by tracing out a tensor factor of the Hilbert space. But in general, the total state cannot be obtained by merging all these restricted states as those lack information about correlations and possible entanglement.

This brings me to the next confusion: There is nothing wrong with states containing correlations of space-like separated observables. This is not even a distinguishing property of quantum physics, as this happens all the time even in classical situations: In the morning, I pick a pair of socks from my drawer without turning on the light and put it on my feet. Thus I do not know which socks I am wearing, in particular, I don't know their color. But as I combined matching socks when they came from the washing machine (as far as this is possible given the tendency of socks going missing) I know by looking at the sock on my right foot what the color of the sock on my left foot is, even when my two feet are spatially separated. Before looking, the state of the color of the socks was a statistical mixture but with non-local correlations. And of course there is nothing quantum about my socks (even if in German "Quanten" is not only "quantum" but also a pejorative word for feet). This would even be true (and still completely trivial) if I had put one of my feet through an event horizon while the other one is still outside. This example shows that locality is not a property that I should demand of states in order to be sure my theory is free of time travel. The important locality property is not in the states, it is in the observables: The measurement of an observable here must not depend of whether or not I apply an operator at a space-like distance. Otherwise that would imply I could send signals faster than the speed of light. But it is the operators, not the states that have to be local (i.e. commute for spatial separation).

If two operators, however, are time-like separated (i.e. one is after the other in its forward light cone), I can of course influence one's measurement by applying the other. But this is not about correlations, this is about influence. In particular, if I write something in my notebook and then throw it across the horizon of a black hole, there is no point in saying that there is a correlation (or even entanglement) between the notebook's state now and after crossing the horizon. It's just the former influencing the later.

Which brings us to entanglement. This must not be confused with correlation, the former being a strict quantum property whereas the other can be both quantum or classical. Unfortunately, you can often see this in popular talks about quantum information where many speakers claim to explain entanglement but in fact only explain correlations. As a hint: For entanglement, one must discuss non-commuting observables (like different components of a the same spin) as otherwise (by the GNS reconstruction theorem) one deals with a commutative operator algebra which always has a classical interpretation (functions on a classical space). And of course, it is entanglement which violates Bell's inequality or shows up in the GHZ experiment. But you need something of this complexity (i.e. involving non-commuting observables) to make use of the quantumness of the situation. And it is only this entanglement (and not correlation) that is "monogamous": You cannot have three systems that are fully entangled for all pairs. You can have three spins that are entangled, but once you only look at two they are no longer entangles (which makes quantum cryptography work as the eavesdropper cannot clone the entanglement that is used for coding).

And once more, entanglement is a property of a state when it is split according to a tensor product decomposition of the Hilbert space. And thus lives on a Cauchy surface. You can say that a state contains entanglement of two regions on a Cauchy surface but it makes no sense to say to regions that are time-like to each other to be entangled (like the notebook before and after crossing the horizon). And therefore monogamy cannot be invoked with respect to also taking the outgoing radiation in as the third player.

## Monday, September 24, 2012

### The future of blogging (for me) and in particular twitter

This is not only true for me writing blog posts but also about reading: Until about a year ago, I was using google reader not to miss a single blog post of a list of about 50 blogs. I have completely stopped this and systematically read blogs only very occasionally (that is other than being directed to a specific post by a link from somewhere else).

What I still do (and more than ever) is use facebook (mainly to stay in contact with not so computer affine friends) and of course twitter (you will know that I am @atdotde there). Twitter seems to be the ideal way to stay current on a lot of matters you are interested in (internet politics for example) while not wasting too much time given the 140 character limit.

Twitter's only problem is that they don't make (a lot of) money. This is no problem for the original inventors of the site (they have sold their shares to investors) but the current owners now seem desperate to change this. From what they say they want to move twitter more to a many to one (marketing) communication platform and force users to see ads they mix among the genuine tweets.

One of the key aspects of the success of twitter was its open API (application programmers interface): Everybody could write programs (and for example I did) that interacted with twitter so for example everybody can choose their favourite client program on any OS to read and write tweets. Since the recent twitter API policy changes this is no longer the case: A client can now have only 100,000 users (or if they already have more can double the number of users), a small number given the allegedly about 4,000,000 million twitter accounts. And there are severe restrictions how you may display tweets to your users (e.g. you are not allowed to use them in any kind of cloud service or mix them with other social media sites, i.e. blend them with Facebook updates). The message that this sends is clearly: "developers go away" (the idea seems to be to force users to use the twitter website and their own clients) and anybody who still invests in twitter developing is betting on a dead horse. But it is not hard to guess that in the long run this will also make the while twitter unattractive to a lot of (if not eventually all) their users.

People (often addicted to twitter feeds) are currently evaluating alternatives (like app.net) but this morning I realized that maybe the twitter managers are not so stupid as they seem to be (or maybe they just want to cash in what they have and don't care if this ruins the service), there is still an alternative that would make twitter profitable and would secure the service in the long run: They could offer to developers to allow them to use the old API guidelines but for a fee (say a few $/Euros per user per month): This would bring in the cash they are apparently looking for while still keeping the healthy ecosystem of many clients and other programs. twitter.com would only be dealing with developers while those would forward the costs to their users and recollect the money by selling their apps (so twitter would not have to collect money from millions of users).

But maybe that's too optimistic and they just want to earn advertising money NOW.

## Tuesday, February 07, 2012

### AdS/cond-mat

My attitude towards this attempt has somewhat changed from "this will never work" to "it's probably as good as anything else" and in this post I will explain why I think this. I should mention as well that Sean Hartnoll has been essential in this phase transition of my mind.

Let me start by sketching (actually: caricaturing) what I am talking about. You want to understand some material, typically the electrons in a horribly complicated lattice like bismuth strontium calcium copper oxide, or BSCCO. To this end, you come up with a five dimensional theory of gravity coupled to your favorite list of other fields (gauge fields, scalars with potentials, you name it) and place that in an anti-de-Sitter background (or better, for finite temperature, in an asymptotically anti-de-Sitter black hole). Now, you compute solutions with prescribed behavior at infinity and interpret these via Witten's prescription as correlators in your condensed matter theory. For example you can read off Green functions and (frequency dependent) conductivities, densities of state.

How can this ever work, how are you supposed to guess the correct field content (there is no D-brane/string description anywhere near that could help you out) and how can you ever be sure you got it right?

The answer is you cannot but it does not matter. It does not matter as it does not matter elsewhere in condensed matter physics. To clarify this, we have to be clear about what it means for a condensed matter theorist to "understand" a system. Expressed in our high energy lingo, most of the time, the "microscopic theory" is obvious: It is given by the SchrÃ¶dinger equation for $10^23$ electrons plus as similar number of noclei feeling the Coulomb potential of the nuclei and interacting themselves with Coulomb repulsion. There is nothing more to be known about this. Except that this is obviously not what we want. These are far too many particles to worry about and, what is more important, we are interested in the behavior at much much lower energy scales and longer wave lengths, at which all the details of the lattice structure are smoothed out and we see only the effect of a few electrons close to the Fermi surface. As an estimate, one should compare the typical energy scale of the Coulomb interactions, the binding energies of the electrons to the nucleus (Z times 13.6 eV) or in terms of temperature (where putting in the constants equates 1eV to about 10,000K) to the milli-eV binding energy of Cooper pairs or the typical temperature where superconductivity plays a role.

In the language of the renormalization group, the Coulomb interactions are the UV theory but we want to understand the effective theory that this flows to in the IR. The convenient thing about such effective theories is that they do not have to be unique: All we want is a simple to understand theory (in which we can compute many quantities that we would like to know) that is in the same universality class as the system we started from. Differences in relevant operators do not matter (at least to leading order).

Surprisingly often, one can find free theories or weakly (and thus almost free) theories that can act as the effective theory we are looking for. BCS is a famous example, but Landau's Fermi Liquid Theory is another: There the idea is that you can almost pretend that your fermions are free (and thus you can just add up energies taking into account the Pauli exclusion principle giving you Fermi-surfaces etc) even though your electrons are interacting (remember, there is always the Coulomb interaction around). The only effect the interactions have, is to renormalize the mass, to deform the Fermi surface away from a ball and to change the hight of the jump in the T=0 occupation number. Experience shows that this is an excellent description in more than one dimension (that has the exception of the Luttinger liquid) and can probably traced back to the fact that a four-Fermi-interaction is non-renormalizable and thus invisible in the IR.

Only, it is important to remember that the fields/particles in that effective theories are not really the electrons you started with but just quasi-particles that are build in complicated ways out of the microscopic particles carrying around clouds of other particles and deforming the lattice they move in. But these details don't matter and that is the point.

It is only important to guess the effective theory in the same universality class. You never derive this (or: hardly ever). Following an exact renormalization group flow is just way beyond what is possible. You make a hopefully educated guess (based on symmetries etc) and then check that you get good descriptions. But only the fact, that there are not too many universality classes makes this process of guessing worthwhile.

Free or weakly coupled theories are not the only possible guesses for effective field theories in which one can calculate. 2d conformal field theories are others. And now, AdS-technology gives us another way of writing down correlation functions just as Feynman-rules give us correlation functions for weakly coupled theories. And that is all one needs: Correlation functions of effective field theory candidates. Once you have those you can check if you are lucky and get evidence that you are in the correct universality class. You don't have to derive the IR theory from the UV. You never do this. You always just guess. And often enough this is good enough to work. And strictly speaking, you never know if your next measurement shows deviations from what you thought would be an effective theory for your system.

In a sense, it is like the mystery that chemistry works: The periodic table somehow pretends that the electrons in atoms are arranged in states that group together like for the hydrogen atom, you get the same n,l,m,s quantum numbers and the shells are roughly the same (although with some overlap encoded in the Aufbau principle) as for hydrogen. This pretends that the only effect of the electron-electron Coulomb potential is to shield the charge of the nucleus and every electron sees effectively a hydrogen like atom (although not necessarily with integer charge Z) and Pauli's exclusion principle regulates that no state is filled more than once. One could have thought that the effect of n-1 electrons on the last is much bigger, after all, they have a total charge that is almost the same of the nucleous, but it seems, the last electron only sees the nucleus with a 1/r potential although with reduced charge.

If you like, the only thing one should might worry about is that the Witten prescription to obtain boundary correlators from bulk configurations really gives you valid n-point functions of

*quantum theory (if you feel sufficient mathematical masochism for example in the sense of Wightman) but you don't want to show that it is*

**a***the*quantum field theory corresponding to the material you started with.

## Friday, February 03, 2012

### Write-ups

First, there are the notes of a block course that I have in the summer on how to fix some mathematicla lose ends in QFT (notes written by our students Mario Flory and Constantin Sluka):

# How I Learned to Stop Worrying and Love QFT

Lecture notes of a block course explaining why quantum field theory might be in a better mathematical state than one gets the impression from the typical introduction to the topic. It is explained how to make sense of a perturbative expansion that fails to converge and how to express Feynman loop integrals and their renormalization using the language of distribtions rather than divergent, ill-defined integrals.

Then there are the contributions to a seminar on "Foundations of Quantum Mechanics" (including an introduction by your's truly) that I taught a year ago. From the contents:

- C*-algebras, GNS-construction, states, (Sebastian)
- Stone-von-Neumann Theorem (Dennis)
- Pure Operations, POVMs (Mario)
- Measurement Problem (Anupam, David)
- EPR and Entanglement, Bell's Theorem, Kochen–Specker theorem (Isabel, Matthias)
- Decoherence (Kostas, Cosmas)
- Pointer Basis (Greeks again)
- Consistent Histories (Hao)
- Many Worlds (Max)
- Bohmian Interpretation (Henry, Franz)

## Wednesday, November 30, 2011

### More than one nature for natural units

The first thing is that strictly speaking, there is not *the* natural unit system, it depends on the problem you are interested in. For example, if you are interested in atoms, the typical mass is that of the electron, so you will likely be interested in masses as multiples of $m_e$. Then, interactions are Coulomb and you will want to express charges as multiples of the electron charge $e$. Finally, quantum mechanics is your relevant framework, so it is natural to express actions in multiples of $\hbar$. Then a quick calculation shows that this unit system of setting $m_e=e=\hbar=1$ implies that distances are dimensionless and the distance $r=1$ happens to be the Bohr radius that sets the natural scale for the size of atoms. Naturalness here lets you guess the size of an atom from just identifying the electron mass, the electric charge and quantum mechanics to be the relevant ingredients.

When you are doing high energy particle physics quantum physics and special relativity are relevant and thus it is convenient to use units in which $\hbar=c=1$ which is Bee's example. In this unit system, masses and energy have inverse units of length.

If you are a classical relativist contemplating solutions of Einstein's equations, then quantum mechanics (and thus $\hbar$) does not concern you but Newton's constant $G$ does. These people thus use units with $c=G=1$. Confusingly, in this unit system, masses have units of length (and not inverse length as above). In particular, the length scale of a black hole with mass M, the Schwarzschild radius is $R=2M$ (the 2 being there to spice up life a bit). So you have to be a bit careful when you convert energies to lengths, you have to identify if you are in a quantum field theory or in a classical gravity situation.

My other remark is that it is conventional how many independent units you have. Many people think, that in mechanics you need three (e.g. length, mass and time, meters, kilograms and seconds in the SI system) and a fourth if you include thermodynamics (like temperature measured in Kelvins) and a fifth if there is electromagnetism (like charge or alternatively current, Amperes in SI). But these numbers are just what we are used to. This number can change when we change our understanding of a relation from "physical law" to "conversion factor". The price is a dimensionful constant: In the SI system, it is a law that in equipartition of energy $E=\frac 12k_bT$ and Coulombs law equates a mechanical force to an electrostatic expression via $F=\frac{qQ} 1{4\pi\epsilon_0r}$ and it is a law that light moves at a speed $c=s/t$.

But alternatively, we could use these laws to define what we actually *mean* by Temperature (then measured in units of energy), charge (effectively setting $4\pi\epsilon_0$ to unity and thereby expressing charge in mechanical units) and length (expressing a distance by the time light need to traverse it). This eliminates a law and a unit. What remains of the law is only the fact that one can do that without reference to circumstances, that a distance from here to Paris does not depend for example on the time of the year (and thus on the direction of the velocity of the earth on its orbit around the sun and thus potentially relative to the ether). If the speed of light would not be constant and we would try to measure distances by the time it takes light to traverse them then distances would suddenly vary when we would say that the speed of light varies.

There is even an example that you can increase the number of units to more than what we are used to (although a bit artificial): It is not god given what kinds of things we consider 'of the same type' and thus possible to be measured in the same units. We are used to measuring all distances in the same unit (like for example meters) or derived units like kilometers or feet (with a fixed numerical conversion factor). But in nautical situations it is common to treat horizontal distance to be entirely different from vertical distances. Horizontal distances like the way to the next island you would measure in nautical miles while vertical distances (like the depth of water) you measure in fathoms. It is then a natural law that the ratio between a given depth and a given horizontal distance is constant over time and there is dimensionful constant (fathoms per mile) of nature that allows to compute a horizontal distance from a depth.