We are in the information business and it seems obvious that more data is better, but how much better? Would you be surprised to learn that most of the information is in a single measurement?

The Cover-Hart inequality was introduced to the world in 1967 but few people seem to know of it or its implications. If you are a statistically-oriented programmer doing data mining or AI, then it seems to have something really important to say to you.

A recent paper by Tilmann Gneiting of the Universität Heidelberg puts the question very clearly:

"Bob predicts a future observation based on a sample of size one. Alice can draw a sample of any size before issuing her prediction. How much better can she do than Bob?"

it goes on to provide the surprising answer:

"Perhaps surprisingly, under a large class of loss functions, which we refer to as the Cover-Hart family, the best Alice can do is to halve Bob’s risk.

In this sense, half the information in an inﬁnite sample is contained in a sample of size one."

At this point you are probably thinking that there must be something very odd about the class of loss functions in the Cover-Hart family, but this doesn't provide an escape.

As the paper proves, any metric and any negative definite kernel belong to the Cover-Hart family. This includes the squared error loss function - so estimation by least squares is included in the theorem.

The paper gives a list of a range of loss functions that are contained within the Cover-Hart family:

You can see that while the question remains of what constitutes the full Cover-Hart family, most of the measures of loss that we use in practice are included.

The theorem not only applies to point estimators but to probabilistic predictions where the sample is used to construct a probability distribution for the estimate. In this case, defining the expected loss is more complicated but the results are very similar - a multi-sample estimate does no better than twice the expected loss of a single sample estimate.

This has some interesting consequences. As the paper points out:

Given that under many of the most prevalent loss functions used in practice, Alice, despite having an inﬁnite sample at her disposal, can at most halve Bob’s risk, who has access to a sample of size one only, it is not surprising that empirically observed differentials in the predictive performance of competing forecasters tend to be small.

For example, this was observed in the Netﬂix contest, where predictive performance was measured in terms of the (root mean) squared error.

"Taking a much broader perspective, the Cover-Hart inequality may contribute to our understanding of the empirical success not only of nearest neighbor techniques and their ramifications, but reasoning and learning by analogy in general."

In other words, you can't get that much of an advantage no matter how clever your estimation procedure is and in many cases one look is sufficient to draw a conclusion.

To celebrate reaching 1,000 Chrome Experiments that showcase what you can do in the browser with HTML5 and JavaScript, the Google Chrome has created an experiment that visualizes every other [ ... ]

Until a few days ago the recently discovered flaw in modern memory chips was just a worrying curiosity - changing some bits at one location in a particular way could result in another distinct bit fli [ ... ]