National Numeracy Network > NNN Blog

NNN Blog

QR for Dying is Dying for QR


Posted: Apr 24 2014 by Nathan Grawe

I was recently pointed to a Bloomberg slideshow on how Americans die. The piece has high production value and includes a good bit of interesting data on the causes of deaths in America from the 1968 to 2010. So far, so good.

However, the curious thing is that the text often contradicts the data. For instance, the first graph includes a caption that notes that mortality rates have dropped 17% over the period studied, but that "it looks like the progress stops in the mid-1990s." The data, on the other hand, report that the mortality rate fell almost 8% from 1995 through 2010. In other words, over 45% of the drop in mortality was experienced in the 36% of time following 1995. Far from gains "stopping" in the mid-1990s, we've apparently done better in that time.

That we've made progress in mortality since 1995 is all the more amazing given two facts. 1) Death is ultimately inevitable: As we eliminate deaths, those that remain will be awfully difficult to avoid. 2) As the second slide shows, the population aged considerably in recent years.

In another instance (slide 10), the authors note they are surprised by the fact that mortality among those ages 45-54 has been steady since 2000 because "cancer and heart disease...have become much less deadly over the years." This caption is immediately over a graph showing that deaths from these causes in this age group have actually increased over this time period.

Slide 13 is a more common mistake. The authors assert that cars generally kill younger people, but fail to note the the younger age groups include 20 birth cohorts while the older groups include only 10. Little surprise there are about twice as many deaths in the younger groups in recent years.

I think we can understand this kind of problem as a form of confirmation bias. We all tend to see evidence for what we are looking for. Now, in this case, it's ultimately evident that the authors saw evidence for their priors that really wasn't there. But this is just a reflection of our flawed humanity–we are prone to seeing evidence for what we believe even when the evidence contradicts our views. What I appreciate about QR is that it at least provides a chance that others might more readily correct our error. Better yet, maybe we can learn to see our own errors and re-evaluate our priors.

Precise Ignorance


Posted: Apr 23 2014 by Nathan Grawe

Last week the Census Bureau made a very sad announcement. Starting in the fall of 2014 they will be using revised questions in the Current Population Survey concerning insurance coverage. The old questions, used for three decades, had been widely regarded by health experts as flawed because the wording created potential ambiguity that likely led to over-stating the number of uninsured. The new questions were tested alongside the old last year and it looks like the hypothesis was right–estimates of the uninsured were a couple percentage points lower with the new questions than with the old (10.6% vs. 12.5%).

The problem is that the wording of the new questions is so different than the old that the Census Bureau does not believe the two to be comparable. And so their data on insurance coverage will include an irreconcilable break between 2013 and 2014.

This is a real blow to evidence-based policy. The Affordable Care Act is arguably the biggest federal policy reform of the last generation. But because we have chosen to change methodologies right at the point of implementation, it will be difficult to assess its effectiveness. The fact that the old questions were biased is, by comparison, a second-order concern. So long as the bias remains more or less the same over time, the resulting data would still have been useful in estimating the ACA's effect. Instead, a highly contentious empirical debate is likely to continue needlessly for want of empirical data.

The underlying issue here is of broader QR application. Often we are offered fancy methods which promise to eliminate a bias (at least in "large" samples) at the cost of larger standard errors. One example is the use of instrumental variables as a consistent alternative to ordinary least squares. I'm often left thinking I'd rather have my more-precise-but-biased estimate. If I understand the source of the bias I can often back it out of the answer. By contrast, the imprecise-yet-accurate estimator often tells me very little. I'd rather be precisely wrong (in a predictable way) than vaguely right.

And this is exactly where we are with the new Census questions. I'll trust the survey experts that these are better questions, but the sad result will be that we can say precisely nothing.

The QR of Water


Posted: Apr 15 2014 by Nathan Grawe

The New York Times reports that China is planning a large desalination plant to provide 1 million tons of fresh water to Beijing. The paper cites planners' beliefs that the plant "could account for one-third of the water consumption of Beijing, a city of more than 22 million people." Interestingly, the cost of desalinated water isn't as great as you might expect. The Times estimates a cost of $1.29 per ton or about twice the cost of tap water. (Nice job by the journalist in providing that useful context!)

This is potentially important because we consume a lot of water. The USGS estimates that we in the US use about 400 billion gallons per day. Given that a ton is about 240 gallons, that's over 1.7 billion tons of water. At an added cost of 65 cents per ton that means we save about $1 billion every day we are able to avoid desalination to meet our water needs. But with over 96% of the earth's water accounted for by sea water, it is comforting to know that for a bit less than half the cost of Social Security we could find alternative water sources. Here's hoping it doesn't come to that!

QR in Bracketology


Posted: Mar 21 2014 by Nathan Grawe

It's March Madness time. This year, a few more people may beplaying brackets. With financial backing from Warren Buffet, Quicken Loans is offering billion to anyone who filled out a perfect NCAA bracket. As the USAToday article I've linked to notes, that's not as big a risk for Buffet as it sounds like because there are 9,223,372,036,854,775,808 (9 quintillion) different possible brackets. Of course, not all of those are equally likely. Nate Silver and the folks at fivethirtyeight.com have put a lot of effort into picking the games. Silver figures that all of his number crunching dramatically improved his odds of a perfect bracket...to 1 in "7.4 billion". (Actually it is 1 in 7,419,071,319–don't forget those last 19+ million outcomes!)

So, how's it all working out? Not surprisingly, Quicken reports at least 50 people have perfect brackets after the first 16 games. Unfortunately for Nate Silver, he didn't see Dayton beating Ohio State.

Growth of the Internet


Posted: Mar 15 2014 by Nathan Grawe

This news story announces US plans to give up its central role in administering the internet. That got me wondering who in the world (literally!) uses the internet. UNData provides the answer, with a graph designed by Drastic Data. [Note: The Drastic Data site includes a link to the UN data and also allows you to re-cast the data in percentage terms. In case you are wondering, the UN internet data are reported by national statistical offices, so these statistics are likely of widely varying quality.]

The image makes two things clear:

1) Growth has been dramatic–though not literally exponential. We are apparently around 1 billion users, increasing at a pace of about 100 million users per year. At this rate we won't have the entire world online until the end of the 21st century (assuming the world population grows to approximately 9 billion and then stabilizes). Of course, because the last adopters will likely be harder to draw online due to poverty, we probably won't get anywhere near that many users.

2) Until 1997, the US made up more than half of internet users. Since then we have fallen to only 20% of the internet market.

3) Despite the US's decreasing role, we remain the dominant internet user. It looks like European users have only just recently risen equal US users. So, while "the rest of the world" has long overtaken the number of US users, no other homogeneous political group has matched the US use until just recently. Perhaps this explains why the pressure to shift internet control out of the US has taken so long to build considerable pressure.

RSS


« Previous Page      Next Page »