Scooped … by Newton and Leibniz

I just learned via the FQXi twitter feed about this paper that was actually published in a medical journal in 1994:

A Mathematical Model for the Determination of Total Area Under Glucose Tolerance and Other Metabolic Curves

Abstract

OBJECTIVE To develop a mathematical model for the determination of total areas under curves from various metabolic studies.

RESEARCH DESIGN AND METHODS In Tai’s Model, the total area under a curve is computed by dividing the area under the curve between two designated values on the X-axis (abscissas) into small segments (rectangles and triangles) whose areas can be accurately calculated from their respective geometrical formulas. The total sum of these individual areas thus represents the total area under the curve. Validity of the model is established by comparing total areas obtained from this model to these same areas obtained from graphic method Gess than ±0.4%). Other formulas widely applied by researchers under- or overestimated total area under a metabolic curve by a great margin.

RESULTS Tai’s model proves to be able to 1) determine total area under a curve with precision; 2) calculate area with varied shapes that may or may not intercept on one or both X/Y axes; 3) estimate total area under a curve plotted against varied time intervals (abscissas), whereas other formulas only allow the same time interval; and 4) compare total areas of metabolic curves produced by different studies.

CONCLUSIONS The Tai model allows flexibility in experimental conditions, which means, in the case of the glucose-response curve, samples can be taken with differing time intervals and total area under the curve can still be determined with precision.

That’s right — the author discovered the trapezoidal rule from integral calculus and named it after herself. A subsequent issue of the journal contained several letters pointing this out. (Both links may be paywalled for all I know.)

It’s pretty embarrassing that such a paper (a) got written in the first place and (b) passed peer review, but  the really shocking part is that the paper has garnered 161 citations.

Why astrophysicists should measure immigration delays at Heathrow

As I’ve mentioned before, I like the BBC’s podcast More or Less, which examines the use and misuse of statistics in the news. It generally conveys some reasonably sophisticated ideas engagingly and correctly. Here’s a nice example I learned from them recently.

The British Government has set a goal for the time people spend in immigration queues at Heathrow. According to an article by Tim Harford, host of More or Less,

The Border Force is supposed to ensure that passengers from outside the EU get through immigration checks within 45 minutes 19 times out of 20, while EU-based passengers should get through within 25 minutes, again 19 times out of 20.

They then measure whether this goal has been met:

At regular intervals [once per hour, to be precise] they pick somebody joining the back of the queue and then time how long it takes for that person to clear immigration.

At first glance, this might sound like a reasonable method, but in fact it’s not. The reason should certainly be familiar to astrophysicists (and probably to lots of other sorts of scientists). I’ll put a page break here just in case you want to think about it for a minute.

Continue reading Why astrophysicists should measure immigration delays at Heathrow

ESA’s going to Jupiter

The European Space Agency’s next large mission will be JUICE, a probe to study Jupiter’s icy moons. People who study other areas of astrophysics (like me) are disappointed that ESA didn’t choose a mission to the stuff we’re terribly excited about. In particular, some people are very disappointed that the X-ray observatory ATHENA lost out.

I don’t really know what should have happened, but this blog post (which I learned about from Peter Coles, by the way) does a pretty good job of explaining why JUICE is interesting.

The most striking thing to me is the extremely long time scale. JUICE isn’t scheduled to get to Jupiter until the 2030s. I know it’s sometimes necessary to plan way in advance, but it does seem like a big gamble to devote a bunch of resources to something that far off. How certain are we that the questions that seem interesting now will still seem interesting then?

We still don’t know if the wavefunction is physically real

Nature has a piece claiming that a newly-published paper shows that the quantum mechanical wavefunction is physically real, as opposed to merely encoding information about our knowledge of the state of a system. As I mentioned back in the fall, I can’t see how the published result shows anything like this.

The paper shows that a certain class of hidden-variable theories would lead to predictions that differ from standard quantum mechanics, and hence that experiments can in principle tell the difference between these theories and quantum mechanics. But that doesn’t show that the wavefunction is real, unless you believe that this particular class of hidden-variable theories is the only thing that the wavefunction-isn’t-real camp could possibly believe. There’s certainly no evidence that this is the case.

Personally, I’m not sure the question of whether the wavefunction is “physically real” is meaningful. I am pretty sure that, even if it is, this paper doesn’t resolve it.

 

As long as you’re boycotting Elsevier anyway …

why not add to your list all the journals on Allen Downey’s Hall of Shame? Elsevier may charge exorbitant prices, but at least they don’t (as far as I know) require authors to write in the passive voice.

I’m not a big passive-basher. I’m with Geoffrey Pullum: teaching students never to use the passive voice is largely passing on a supersition. But the reverse rule (always use the passive voice), which some scientists seem to have been taught, is far worse. The never-use-the-passive superstition is harmless, maybe even mildly helpful. The always-use-the-passive superstition, on the other hand, is wholly pernicious.

If you’re a scientist, use the active voice whenever it sounds better (which is most of the time). If an editor won’t let you, fight back.

For what it’s worth, I’ve never had an editor or referee complain about my use of the active voice in any of my papers. Fortunately, I don’t publish in the ICES Journal of Marine Science or the other hall of shame inductees.

Haven’t posted anything for a while. Busy. But I thought I’d at least throw up a link to my old friend Allen Downey’s post in which he offers a bounty for anyone who can find a scientific journal whose style guide explicitly requires or recommends the passive voice.

As I mentioned once, I think that the blanket advice to avoid the passive voice is often overstated. But the idea that you’re always supposed to use the passive in scientific writing is incredibly silly. If the active voice is better at clearly and concisely indicating who did what to whom (as it often is), use the active voice. If a journal editor tells you you can’t, well, at least you can collect $100 from Allen.

Rational conspiracy theorists

Here’s an interesting study on belief in conspiracy theories:

Conspiracy theories can form a monological belief system: A self-sustaining worldview comprised of a network of mutually supportive beliefs. The present research shows that even mutually incompatible conspiracy theories are positively correlated in endorsement. In Study 1 (n = 137), the more participants believed that Princess Diana faked her own death, the more they believed that she was murdered. In Study 2 (n = 102), the more participants believed that Osama Bin Laden was already dead when U.S. special forces raided his compound in Pakistan, the more they believed he is still alive. Hierarchical regression models showed that mutually incompatible conspiracy theories are positively associated because both are associated with the view that the authorities are engaged in a cover-up (Study 2). The monological nature of conspiracy belief appears to be driven not by conspiracy theories directly supporting one another but by broader beliefs supporting conspiracy theories in general.

Based on this, one might be tempted to think that conspiracy theorists are just crazy! How can they simultaneously believe contradictory things?

Maybe they are crazy, of course, but these data don’t actually provide strong evidence for it. The problem is that the abstract uses a shorthand that seems quite reasonable at first but is actually a bit misleading. Instead of saying

The more participants believed that Princess Diana faked her own death, the more they believed that she was murdered.

They should say

The more participants believed that Princess Diana might have faked her own death, the more they believed that she might have been murdered.

That’s what the surveys probably revealed. They asked people to rate their belief in the various statements on a 7-point Likert scale from “Strongly Disagree” to “Strongly Agree”. Anyone who gave strongly positive responses (say 6’s or 7’s) to two contradictory options would indeed be irrational, but as far as I can tell the results are consistent with the (more likely, in my opinion) scenario that lots of people gave 3’s and 4’s to the various contradictory options. And there’s nothing irrational at all about saying that multiple mutually contradictory options are all “somewhat likely.”

In fact, although the positive correlation between contradictory possibilities is amusing, it’s actually not surprising. The last couple of sentences of the abstract lay out a sensible explanation: if you’re generally conspiracy-minded, then you believe that shady people are trying to conceal the truth from you. Given that premise, it’s actually rational to bump up your assessment of the probabilities of a wide variety of conspiracies, even those that contradict each other.

I’m not saying, of course, that the original premise is rational, just that the conclusions apparently being drawn are rational consequences of it.

And I should add that, as far as I can tell, a careful reading of the paper indicates that the authors understand all this. A quick reading of the abstract might lead one to the wrong conclusion (that conspiracy theorists simultaneously believe contradictory things), but on a more careful reading the paper doesn’t say that.

Neutrinos (probably) don’t go faster than light

Apparently the experimenters who found that neutrinos go faster than light have identified some sources of error in their measurements that may explain the result. The word originally went out in an anonymously-sourced blog post over at Science:

According to sources familiar with the experiment, the 60 nanoseconds discrepancy appears to come from a bad connection between a fiber optic cable that connects to the GPS receiver used to correct the timing of the neutrinos’ flight and an electronic card in a computer. After tightening the connection and then measuring the time it takes data to travel the length of the fiber, researchers found that the data arrive 60 nanoseconds earlier than assumed. Since this time is subtracted from the overall time of flight, it appears to explain the early arrival of the neutrinos. New data, however, will be needed to confirm this hypothesis.

Science is of course a very reputable source, but people are rightly skeptical of anonymously sourced information. But apparently it’s legitimate: a statement from the experimental group later confirmed the main idea, albeit in more equivocal language. They say they’ve found two possible sources of error that may eventually prove to explain the result.

Nearly all physicists (probably including the original experimenters) expected something like this to happen: it always seemed far more likely that there was an experimental error of some sort than that revolutionary new physics was occurring. If you asked them why, they’d probably trot out the old saying, “Extraordinary claims require extraordinary evidence,” or maybe the version I prefer, “Never believe an experiment until it has been confirmed by a theory.”

The way to make this intuition precise is with the language of probabilities and Bayesian inference.

Suppose that you hadn’t heard anything about the experimental results, and someone asked you to estimate the probability that neutrinos go faster than light. Your answer would probably be something like 0.000000000000000000000000000000000000000000000000001, but possibly with a lot more zeroes in it. The reason is that a century of incredibly well-tested physics says that neutrinos can’t go faster than light.

We call this number your prior probability. I’ll denote it ps (s for “superluminal,” which means “faster than light”).

Now suppose that someone asked you for your opinion about the probability that this particular group of experimenters would make a mistake in an experiment like this. (Let’s say that you still don’t know the result of the experiment yet.) I’ll call your answer pm (m for “mistake”, of course). Your answer will depend on what you know about the experimenters, among other things. They’re a very successful and reputable bunch of physicists, so pm will probably be a pretty low number, but these experiments are hard, so even if you have a lot of respect for them it won’t be incredibly low the way ps is.

Now someone tells you the result: “These guys found that neutrinos go faster than light!” The theory of probability (specifically the part often called “Bayesian inference”) gives a precise prescription for how you will update your subjective probabilities to take this information into account. I’m sure you’re dying to know the formula, so here it is:

 

Here ps means the final (posterior) probability that neutrinos go faster than light.

For any given values of the two input probabilities, you can work out your final probability. But we can see qualitatively how things are going to go without a tedious calculation. Suppose that you have a lot of respect for the experimenters, so that pm is small, and that you’re not a crazy person, so that ps is extremely small (much less than pm). Then to a good approximation the numerator in that fraction is ps and the denominator is pm + ps, which is pretty much pm. We end up with

ps‘ = ps/pm .

If, for instance, you thought there was only a one in a thousand chance that the experimenters would make a mistake, then ps‘ would be 1000 ps. That is, the experimental result makes it 1000 times more likely that neutrinos go faster than light. But you almost certainly thought that ps was much smaller than 1/1000 to begin with — it was more like 0.0000000000000000000000000000000000001 or something. So even after you bump it up by a factor 1000, it’s still small.

The situation here is exactly the same as a classic example people love to use in explaining probability theory. Suppose you take a test for some rare disease, and the test comes back positive. You know that the test only fails 1 in 1000 times. Does that mean there’s only a 0.1% chance of your being disease-free? No. If the disease is rare (your prior probability of having it was low), then it’s still low even after you get the test result. You would only conclude that you probably had the disease if the failure rate for the test was at least as low as the prior probability that you had the disease.

One sociological note: people who talk about probabilities and statistics a lot tend to sort themselves into Bayesian and non-Bayesian camps. But even those scientists who are fervently anti-Bayesian still believed that the superluminal neutrino experiment was wrong, and even those people were (by and large) not surprised by the recent news of a possible experimental error. I claim that those people were in fact unconsciously using Bayesian inference in assessing the experimental results, and that they do so all the time. There’s simply no other way to reason in the face of uncertain, probabilistic knowledge (i.e., all the time). Whether or not you think of yourself as a Bayesian, you are one.

There’s an old joke about a person who put a horseshoe above his door for good luck. His friend said, “C’mon, you don’t really believe that horseshoes bring good luck, do you?” The man replied, “No, but I hear it works even if you don’t believe in it.” Fortunately, Bayesian inference is the same way.

 

Do we need more scientists in Congress?

Yesterday I posted a criticism of John Allen Paulos’s blog post asking “Why Don’t Americans Elect Scientists?” I focused on the numbers, arguing that scientists are if anything overrepresented in Congress. But it’s worth stepping back and looking at the bigger question: Why might we want more scientists in Congress?

The usual answer, as far as I can tell, is that technology-related issues are important, so we should have representatives who understand them. As Paulos puts it, “given the complexities of an ever more technologically sophisticated world, the United States could benefit from the participation and example of more scientists in government.”

I guess that’s true, but there are lots of areas in which one might wish for expertise in Congress, and I’m not sure technology’s all that near the top. My wish list might include people with expertise in economics, diplomacy, demography, ethics, sociology, and psychology above technology.

I’m not sure that “scientist” is all that great a proxy for “expert in technology” anyway. Some scientists certainly have such expertise, but many don’t, and many non-scientists do. How good a proxy it is depends in part on what you mean by the word “scientist.” Paulos seems to mean “person with some sort of technical training,” but in that case you should certainly include engineers and doctors, in which case the level of representation in Congress is quite high.

When a scientist says we need more scientists in Congress, I suspect that the real reason is not expertise in technology but some combination of the following:

  • Scientists are smart, and we need more smart people in Congress.
  • Scientists will be more likely to base policy decisions on analytic, data-based arguments.
  • Scientists will be more likely to support increased funding for science.

I’m actually sympathetic to all of these arguments, but let’s remember that not all scientists meet these criteria and plenty of non-scientists do.