From Gender to Gleason

… or, The Case of Adam Becker’s What Is Real? (2018).

It is easy to argue that the founders of quantum mechanics made statements which are opaque and confusing. It is fair to say that their philosophical takes on the subject are not infrequently unsatisfying. We can all use reminders that human flaws and passions are a part of physics. So, it would be nice to have a popular book on these themes, one that makes no vital omissions, represents its sources accurately and lives up to its own ideals.

In Re “CopenHagen” and “COLLAPSE”

I was having an e-mail conversation the other day with a friend from olden days — another MIT student who made it out with a physics degree the same year I did — and that led me to set down some thoughts about history and terminology that may be useful to share here.

My primary claim is the following:

We should really expunge the term “the Copenhagen interpretation” from our vocabularies.

What Bohr thought was not what Heisenberg thought, nor was it what Pauli thought; there was no single unified “Copenhagen interpretation” worthy of the name. Indeed, the term does not enter the written literature until the 1950s, and that was mostly due to Heisenberg acting like he and Bohr were more in agreement back in the 1920s than they actually had been.

For Bohr, the “collapse of the wavefunction” (or the “reduction of the wave packet”, or whatever you wish to call it) was not a singular concept tacked on to the dynamics, but an essential part of what the quantum theory meant. He considered any description of an experiment as necessarily beginning and ending in “classical language”. So, for him, there was no problem with ending up with a measurement outcome that is just a classical fact: You introduce “classical information” when you specify the problem, so you end up with “classical information” as a result. “Collapse” is not a matter of the Hamiltonian changing stochastically or anything like that, as caricatures of Bohr would have it, but instead, it’s a question of what writing a Hamiltonian means. For example, suppose you are writing the Schrödinger equation for an electron in a potential well. The potential function $V(x)$ that you choose depends upon your experimental arrangement — the voltages you put on your capacitor plates, etc. In the Bohrian view, the description of how you arrange your laboratory apparatus is in “classical language”, or perhaps he’d say “ordinary language, suitably amended by the concepts of classical physics”. Getting a classical fact at your detector is just the necessary flipside of starting with a classical account of your source.

(Yes, Bohr was the kind of guy who would choose the yin-yang symbol as his coat of arms.)

To me, the clearest expression of all this from the man himself is a lecture titled “The causality problem in atomic physics”, given in Warsaw in 1938 and published in the proceedings, New Theories in Physics, the following year. This conference is notable for several reasons, among them the fact that Hans Kramers, speaking both for himself and on behalf of Heisenberg, suggested that quantum mechanics could break down at high energies. More than a decade after what we today consider the establishment of the quantum theory, the pioneers of it did not all trust it in their bones; we tend to forget that nowadays.

As to how Heisenberg disagreed with Bohr, and what all this has to do with decoherence, I refer to Camilleri and Schlosshauer.

Do I find the Bohrian position that I outlined above satisfactory? No, I do not. Perhaps the most important reason why, the reason that emotionally cuts the most deeply, is rather like the concern which Rudolf Haag raised while debating Bohr in the early 1950s:

I tried to argue that we did not understand the status of the superposition principle. Why are pure states described as [rays] in a complex linear space? Approximation or deep principle? Niels Bohr did not understand why I should worry about this. Aage Bohr tried to explain to his father that I hoped to get inspiration about the direction for the development of the theory by analyzing the existing formal structure. Niels Bohr retorted: “But this is very foolish. There is no inspiration besides the results of the experiments.” I guess he did not mean that so absolutely but he was just annoyed. […] Five years later I met Niels Bohr in Princeton at a dinner in the house of Eugene Wigner. When I drove him afterwards to his hotel I apologized for my precocious behaviour in Copenhagen. He just waved it away saying: “We all have our opinions.”

Why rays? Why complex linear space? I want to know too.

The Mills of Institutional Review Boards

Back in 2017, nominal intellectuals Peter Boghossian and James Lindsay proved that they were willing to invest a lot of energy in lying. Not content to rest on those laurels, they joined with Helen Pluckrose and soon redoubled their efforts. I didn’t have much to say about their “Grievance Studies” brouhaha when that story broke last fall, apart from the observation that the Bogdanov brothers had a better success rate — six out of six bullshit papers accepted! — and I’m still waiting for Steven Pinker to call the whole of physics a heap of postmodern rubbish.

(He won’t; he’s too busy appropriating its respectability.)

Today, I learned that Boghossian’s institutional review board found that his actions in the “Grievance Studies” hoax constitute research misconduct.

Sporadic SICs and exceptional Lie algebras

A while back, I had a bit of a sprawling conversation about certain geometrical oddities over multiple threads at the n-Category Café. I finally got organized enough to gather these notes together, incorporating edits for clarity and recording one construction I haven’t found written in the literature anywhere.

Sometimes, mathematical oddities crowd in upon one another, and the exceptions to one classification scheme reveal themselves as fellow-travelers with the exceptions to a quite different taxonomy.

UPDATE (30 March 2019): Thanks to a kind offer by John Baez, we’re going through this material step-by-step over at a blog with a community, the n-Category Café:

• Part 1: Definitions and preliminaries
• Part 2: Qutrits and E6
• Part 3: The Hoggar lines, E7 and E8

Triply Positive Matrices

One more paper to round out the year!

J. B. DeBrota, C. A. Fuchs and B. C. Stacey, “Triply Positive Matrices and Quantum Measurements Motivated by QBism” [arXiv:1812.08762].

We study a class of quantum measurements that furnish probabilistic representations of finite-dimensional quantum theory. The Gram matrices associated with these Minimal Informationally Complete quantum measurements (MICs) exhibit a rich structure. They are “positive” matrices in three different senses, and conditions expressed in terms of them have shown that the Symmetric Informationally Complete measurements (SICs) are in some ways optimal among MICs. Here, we explore MICs more widely than before, comparing and contrasting SICs with other classes of MICs, and using Gram matrices to begin the process of mapping the territory of all MICs. Moreover, the Gram matrices of MICs turn out to be key tools for relating the probabilistic representations of quantum theory furnished by MICs to quasi-probabilistic representations, like Wigner functions, which have proven relevant for quantum computation. Finally, we pose a number of conjectures, leaving them open for future work.

This is a sequel to our paper from May, and it contains one minor erratum for an article from 2013.

QBism and the Ithaca Desiderata

Time again for the New Paper Dance!

B. C. Stacey, “QBism and the Ithaca Desiderata” [arXiv:1812.05549].

In 1996, N. David Mermin proposed a set of desiderata for an understanding of quantum mechanics, the “Ithaca Interpretation”. In 2012, Mermin became a public advocate of QBism, an interpretation due to Christopher Fuchs and Ruediger Schack. Here, we evaluate QBism with respect to the Ithaca Interpretation’s six desiderata, in the process also evaluating those desiderata themselves. This analysis reveals a genuine distinction between QBism and the IIQM, but also a natural progression from one to the other.

The State Space of Quantum Mechanics is Redundant

There was some water-cooler talk around the office this past week about a paper by Masanes, Galley and Müller that hit the arXiv, and I decided to write up my thoughts about it for ease of future reference. In short, I have no reason yet to think that the math is wrong, but what they present as a condition on states seems more naturally to me like a condition on measurement outcomes. Upon making this substitution, the Masanes, Galley and Müller result comes much closer to resembling Gleason’s theorem than they say it does.

So, if you’re wanting for some commentary on quantum mechanics, here goes:
Continue reading The State Space of Quantum Mechanics is Redundant

Proceedings of the Royal Society of Smegheads

So, the news from a little while back was that a new Journal of Controversial Ideas is in the pipeline, with a big part of the motivation being to protect “academic freedom” from the (nonexistent) Campus Free Speech Crisis. If this sounds to you like a way for the hateful to spout toxic ravings about marginalized peoples from behind a screen of anonymity, then I’d say you have a low opinion of human nature, a low opinion that is entirely merited by the data. If it also sounds to you like a good way to part a mark from his dollar with “peer review” that amounts to a vanity pay-to-publish scheme, then I’d say your sense of cynicism is nicely calibrated.

When I heard about J. Con. Id., I couldn’t help thinking that I have myself supported some unpopular scientific opinions. A few times, that’s where my best professional judgment led me. When my colleagues and I have found ourselves in that position, we set forth our views by publishing … in Nature.

(I have to admit that the 2010 comment is not as strong as it could have been. It was a bit of a written-by-committee job, with all that that implies. I recommend that every young scientist go through that process … once. Better papers in the genre came later. And for my own part, I think I did a better job distinguishing all the confusing variants of terminology when I had more room to stretch, in Chapter 9 of arXiv:1509.02958.)

The Rise of Ironic Physics and/or Machine Physicists?

Over on his blog, Peter Woit quotes a scene from the imagination of John Horgan, whose The End of Science (1996) visualized physics falling into a twilight:

A few diehards dedicated to truth rather than practicality will practice physics in a nonempirical, ironic mode, plumbing the magical realm of superstrings and other esoterica and fret­ting about the meaning of quantum mechanics. The conferences of these ironic physicists, whose disputes cannot be experimentally resolved, will become more and more like those of that bastion of literary criticism, the Modern Language Association.

OK (*cracks knuckles*), a few points. First, “fretting about the meaning of quantum mechanics” has, historically, been damn important. A lot of quantum information theory came out of people doing exactly that, just with equations. The productive way of “fretting” involves plumbing the meaning of quantum mechanics by finding what new capabilities quantum mechanics can give you. Let’s take one of the least blue-sky applications of quantum information science: securing communications with quantum key distribution. Why trust the security of quantum key distribution? There’s a whole theory behind the idea, one which depends upon the quantum de Finetti theorem. Why is there a quantum de Finetti theorem in a form that physicists could understand and care about? Because Caves, Fuchs and Schack wanted to prove that the phrase “unknown quantum state” has a well-defined meaning for personalist Bayesians.

This example could be augmented with many others. (I selfishly picked one where I could cite my own collaborator.)

It’s illuminating to quote the passage from Horgan’s book just before the one that Woit did:

This is the fate of physics. The vast majority of physicists, those employed in industry and even academia, will continue to apply the knowledge they already have in hand—inventing more versatile lasers and superconductors and computing devices—without worrying about any underlying philosophical issues.

But there just isn’t a clean dividing line between “underlying philosophical issues” and “more versatile computing devices”! In fact, the foundational question of what the nature of “quantum states” really are overlaps with the question of which quantum computations can be emulated on a classical computer, and how some preparations are better resources for quantum computers than others. Flagrantly disregarding attempts to draw a boundary line between “foundations” and “applications” is my day job now, but quantum information was already getting going in earnest during the mid-1990s, so this isn’t a matter of hindsight. (Feynman wasn’t the first to talk about quantum computing, but he was certainly influential, and the motivations he spelled out were pretty explicitly foundational. Benioff, who preceded Feynman, was also interested in foundational matters, and even said as much while building quantum Hamiltonians for Turing machines.) And since Woit’s post was about judging whether a prediction held up or not, I feel pretty OK applying a present-day standard anyway.

In short: Meaning matters.

But then, Horgan’s book gets the Einstein–Podolsky—Rosen thought-experiment completely wrong, and I should know better than to engage with what any book like that on the subject of what quantum mechanics might mean.

To be honest, Horgan is unfair to the Modern Language Association. Their convention program for January 2019 indicates a community that is actively engaged in the world, with sessions about the changing role of journalism, how the Internet has enabled a new kind of “public intellectuals”, how to bring African-American literature into summer reading, the dynamics of organized fandoms, etc. In addition, they plainly advertise sessions as open to the public, which I can only barely imagine a physics conference doing more than a nominal jab at. Their public sessions include a film screening of a documentary about the South African writer and activist Peter Abrahams, as well as workshops on practical skills like how to cite sources. That’s not just valuable training, but also a topic that is actively evolving: How do you cite a tweet, or an archived version of a Wikipedia page, or a post on a decentralized social network like Mastodon?

Dragging the sciences for supposedly resembling the humanities has not grown more endearing since 1996.
Continue reading The Rise of Ironic Physics and/or Machine Physicists?

What I Do

At the moment, I’m taking a quick break from reading some rather dense mathematical prose, and I spent yesterday plugging away at a draft of my research group’s next technical publication. This led me to reflect on a lesson that I think a lot of science education leaves out: Even in a technical article, you have to have a story to carry the progression through. “These are all the boffo weird roadside attractions we found while proving the theorems in our last paper” is honest, but not adequate.

Our research project is the reconstruction of the mathematical formalism of quantum theory from physical principles. We tease apart the theory, identify what is robustly strange about it — for many more quantum phenomena can be emulated with classical stochasticity than are often appreciated — and try to build a new representation that brings the most remarkable features of the physics to the forefront. In special relativity, we have Einstein’s postulates, and the dramatic tension between them: Inertial observers can come to agree upon the laws of physics, but they cannot agree upon a standard of rest. In thermodynamics, we have the Four Laws, which come with their own dramatic tension, in that energy is conserved while entropy is nondecreasing. Both of these theories are expressed in terms of what agents can and cannot do, yet they are more than “mere” engineering, because they apply to all agents. Or, to say it another way, it is to the benefit of any agent to pick up the theory and use it as a guide.

What, then, is the analogue for quantum theory? If the textbook presentation of quantum physics is like the formulae for the Lorentz transform, with all those square roots and whatnot, or the Maxwell relations in thermo, with all those intermingling partial derivatives that we invent hacks about determinants to remember, what is quantum theory’s version of Einstein’s postulates or the Four Laws?

That’s the grandiose version, anyway. The reason I got invited to speak at an American Mathematical Society meeting is that the geometric structures that arise in this work are vexingly fascinating. You want about Galois fields and Hilbert’s 12th problem? We’ve got ’em! How about sphere packing and unexpected octonions? We’ve got those, too! And the structure that leads down the latter path turns out, on top of that, to yield a new way of thinking about Mermin’s 3-qubit Bell inequality. It is all lovely, and it is all strange.

The SIC problem gives us the opportunity to travel all throughout mathematics, because, while the definition looks pretty small, the question is bigger on the inside.

J. B. DeBrota and B. C. Stacey, “FAQBism” [arXiv:1810.13401].

We answer several questions that have been Frequently Asked about QBism. These remarks (many of them lighthearted) should be considered supplements to more systematic treatments by the authors and others.

New Paper Dance (-ing With Myself)

B. C. Stacey, “Misreading EPR: Variations on an Incorrect Theme” [arXiv:1809.01751].

Notwithstanding its great influence in modern physics, the EPR thought-experiment has been explained incorrectly a surprising number of times.

22 pages; an unknown number of bridges burned.

New Paper Dance, Encore

This time, it’s another solo-author outing.

B. C. Stacey, “Is the SIC Outcome There When Nobody Looks?” [arXiv:1807.07194].

Informationally complete measurements are a dramatic discovery of quantum information science, and the symmetric IC measurements, known as SICs, are in many ways optimal among them. Close study of three of the “sporadic SICs” reveals an illuminating relation between different ways of quantifying the extent to which quantum theory deviates from classical expectations.

New Papers Dance

In spite of the “everything, etc.” that is life these days, I’ve managed to do a bit of science here and there, which has manifested as two papers. First, there’s the one about quantum physics, written with the QBism group at UMass Boston:

J. B. DeBrota, C. A. Fuchs and B. C. Stacey, “Symmetric Informationally Complete Measurements Identify the Essential Difference between Classical and Quantum” [arXiv:1805.08721].

We describe a general procedure for associating a minimal informationally-complete quantum measurement (or MIC) and a set of linearly independent post-measurement quantum states with a purely probabilistic representation of the Born Rule. Such representations are motivated by QBism, where the Born Rule is understood as a consistency condition between probabilities assigned to the outcomes of one experiment in terms of the probabilities assigned to the outcomes of other experiments. In this setting, the difference between quantum and classical physics is the way their physical assumptions augment bare probability theory: Classical physics corresponds to a trivial augmentation — one just applies the Law of Total Probability (LTP) between the scenarios — while quantum theory makes use of the Born Rule expressed in one or another of the forms of our general procedure. To mark the essential difference between quantum and classical, one should seek the representations that minimize the disparity between the expressions. We prove that the representation of the Born Rule obtained from a symmetric informationally-complete measurement (or SIC) minimizes this distinction in at least two senses—the first to do with unitarily invariant distance measures between the rules, and the second to do with available volume in a reference probability simplex (roughly speaking a new kind of uncertainty principle). Both of these arise from a significant majorization result. This work complements recent studies in quantum computation where the deviation of the Born Rule from the LTP is measured in terms of negativity of Wigner functions.

To get an overall picture of our results without diving into the theorem-proving, you can watch John DeBrota give a lecture about our work.

Second, there’s the more classical (in the physicist’s sense, if not the economist’s):

B. C. Stacey and Y. Bar-Yam, “The Stock Market Has Grown Unstable Since February 2018” [arXiv:1806.00529].

On the fifth of February, 2018, the Dow Jones Industrial Average dropped 1,175.21 points, the largest single-day fall in history in raw point terms. This followed a 666-point loss on the second, and another drop of over a thousand points occurred three days later. It is natural to ask whether these events indicate a transition to a new regime of market behavior, particularly given the dramatic fluctuations — both gains and losses — in the weeks since. To illuminate this matter, we can apply a model grounded in the science of complex systems, a model that demonstrated considerable success at unraveling the stock-market dynamics from the 1980s through the 2000s. By using large-scale comovement of stock prices as an early indicator of unhealthy market dynamics, this work found that abrupt drops in a certain parameter U provide an early warning of single-day panics and economic crises. Decreases in U indicate regimes of “high co-movement”, a market behavior that is not the same as volatility, though market volatility can be a component of co-movement. Applying the same analysis to stock-price data from the beginning of 2016 until now, we find that the U value for the period since 5 February is significantly lower than for the period before. This decrease entered the “danger zone” in the last week of May, 2018.