Chance News 15
- 1 Quotation
- 2 Forsooth
- 3 More generous than they realize
- 4 More on medical studies that conflict with previous studies
- 5 Economists analyze the tv show "Deal or No Deal?"
- 6 Sexual prediction
- 7 Some children left behind
- 8 The future divined by the crowd
Take statistics. Sorry, but you'll find later in life that it's handy to know what a standard deviation is.
New York Times, March 2, 2006
This appears on a list of core knowledge that Brooks says will be sufficient to give you a great education even if you don't make Harvard.
From the February 2006 RSS news we have:
One primary school in East London
has a catchment area of 110 metres.Sunday Telegraph
23 October 2005
And from the March 2006 RSS news we have:
Fewer names appear in the top 100
than ten years ago.
(from report on most popular names
given to babies born in scotland in 2005)The Scotsman
24 December 2005
More generous than they realize
From the novel Summer Harbor, by Susan Wilson (NY: ATRIA Books, 2003):
The waitress coyly asked if they wanted change of the two twenties Kiley laid down on a thirty-two-dollar tab.
"No, we’re fine."
"That’s a twenty percent tip, Mom."
"I’m feeling generous."
Submitted by Margaret Cibes
More on medical studies that conflict with previous studies
Humans, being what they are, it is only natural that when a study's consequences seem plausible there is no need to look too closely. On the other hand, when the outcomes go against what was expected, a great deal of inspection is called for. This was discussed here.
The Wall Street Journal of February 28, 2006 details possible reasons for why the Women's Health Initiative might have had design flaws leading to "murky results." In summary, the WSJ reported:
- Calcium/Vitamin D study
- Message: Supplements don't protect bones or cut risk of colorectal cancer.
- Problem: Those in placebo group also took supplements in many cases.
- Low-fat diet study:
- Message: Doesn't cut risk of breast cancer.
- Problem: Few met the fat goal.
- A 22% drop in risk for women who cut fat the most got little emphasis.
- Hormone study:
- Message: No benefit, possible increased cancer and heart risk.
- Problem: Most in study were too old for this to apply to menopausal women.
More generally, according to the WSJ, "Design problems in all of the trials mean the results don't really answer the questions they were supposed to address. And a flawed communications effort led to widespread misinterpretation of results by the news media and public."
In particular, in order to reduce the number of participants for the studies, "more than half [of the women] took part in at least two of them, and more that 5,000 were in all three trials." As might be imagined, "Among problems this posed was simple burnout" which "contributed to compliance problems that plagued all three and hurt the reliability of their results."
Another problem was the difficulty of double blinding for the hormone study since any hot flashes would indicate to the patient (and to her physician) that she was in the placebo arm; to get around this impediment, the vast majority of the women recruited were well past menopause, thus biasing the results against the benefits of hormone replacement.
So where are we after 68,132 female participants, "fifteen years and $725 million later"? More than likely, the Women's Health Initiative study will be in and out of the news for some time to come because of its ambiguity.
Submitted by Paul Alper
Economists analyze the tv show "Deal or No Deal?"
Why game shows have economists glued to their TVs
Wall Street Journal, Jan. 12, 2006
Economists Learn from Game Show 'Deal or No Deal'
NPR, March 3, 2006, All things considered
Deal or No Deal? Decision making under risk in a large-payoff game show
Thierry Post, Martijn Van Den Assem, Guido Baltussen, Richard H. Thaler
The authors of the research paper write:The popular television game show "Deal or No Deal" offers a unique opportunity for analyzing decision making under risk: it involves very large and wide-ranging stakes, simple stop-go decisions that require minimal skill, knowledge or strategy and near-certainty about the probability distribution.
Here is a nice description of the game from the MS Math in the Media magazine:
- Twenty-six known amounts of money, ranging from one cent to one million dollars, are (symbolically) randomly placed in 26 numbered, sealed briefcases. The contestant chooses a briefcase. The unknown sum in the briefcase is the contestant's.
- In the first round of play, the contestant chooses 6 of the remaining 25 briefcases to open. Then the "banker" offers to buy the contestant's briefcase for a sum based on its expected value, given the information now at hand, but tweaked sometimes to make the game more interesting. The contestant can accept ("Deal") or opt to continue play ("No Deal").
- If the game continues, 5 more briefcases are opened in the second round, another offer is made, and accepted or refused. If the contestant continues to refuse the banker's offers, subsequent rounds open 4, 3, 2, 1, 1, 1, 1 briefcases until only two are left.
- The banker makes one last offer; the contestant accepts that offer or takes whatever money is in the initially chosen briefcase.
So the banker is always trying to buy out the player. If he fails the player will end up with the amount in his suitcase.
The best way to understand the game is to play it here on the NBC website.
We did this playing the only the first round. We chose number 13 for our briefcase. We were then asked to choose six briefcases. We chose suitcases 3,10,12,13,19, 25. Here is the result of this round:
Our six choices eliminated the amounts 1, 50, 75, 300, 400,000, 500,000. The expected value of our suitcase is now the average of the amounts in the remaining suitcases. Making this calculation We find that this is $125,900 (rounded to the nearest dollar). The banker has offered us $29,854 to quit. So we would have to be pretty risk adverse to accept the banker's offer at this point.
In order to see if the banker's offers get better, we played a second game in which we refused all the banker's offers. With this strategy we start with an expected winning given by the average of the initial amounts which we found to be $131,478.
We then played all 9 rounds. In each round we recalculated the expected amount in our briefcase and recorded the banker's offer but refused it. Here are the results:
Round Expected Value Offer Percent of Expected Value 1 130,857 30565 23 2 142,475 37361 26 3 125,186 54507 44 4 153,319 57554 38 5 170,967 67160 39 6 205,160 108950 53 7 256,425 176781 69 8 341,800 317700 93 9 512,500 367500 72
We note that the banker's offer at each round is less than the expected amount in our suitcase. While it seems clear at the beginning we are right to refuse the offer, as we get nearer the end this is not so obvious. For example, after the 8th round the only amounts available were $400, $25,000, and $1,000,000. We are offered $317,700, which is a pretty nice amount of money. If we reject it we would have a 2/3 chance of ending up with a relatively small amount and a 1/3 chance of getting the million dollars. We might well be risk averse at this point and accept the offer. However, we didn't and the suitcase the chose had the $400. Now the only amounts left were $25,000 and $1,000,000. We are now offered $367,500. Again we might think twice about refusing this. But we did refuse it and we opened the last suitcase, which had the million dollars, and so our suitcase had $25,000. If you watch the TV program you will find that most contestants refuse the banker's offers in the early rounds but accept it in one of the later rounds.
Deal or No Deal was introduced in Australia in 2003 and currently airs in 38 countries including the U.K. Those who feel that Excel is slow to fix its statistical errors will be amused by the Wikipedia account (See Predictable Sequence Controversy) of what happend to the UK Deal no Deal game when the Excel random number generater was not properly seeded. As you will see the Excel error allowed one to estimate the contents of the remaining boxes--obviously very useful information.
This error was was first reported by the U.K. Bother's Bar website. You can see the response of the producers of their home page (see Glenn Hugill writes). You can find an interesting discussion of the U.K. version of Deal or no Deal by the Brother's Bar here. In particular in their FAQ they give the following answer to question everyone asks: How does The Banker come up with the offers?Banker's software throws up a lower and higher cash sum based on the values left on the board. The Banker then puts his own spin on the proceedings by picking a figure between the two boundaries. Factors such as the attitude of the player and who is having the better run of luck are all considerations. Producer Glenn Hugill says "I don't pretend to know how he operates but I believe there is a range which is considered to mathematically cover the financial risk. However if the banker wishes to gamble or indeed play safe he can accordingly pitch his offer anywhere below or above it."
Bother's Bar also collected data which made it possible to compare the Box(Suitcase) value with the Banker's offer. You will find here the Box(suitcase) value, Peak bank offer, Prize won for all programs from October 31 2005 to the present time.
Post and his co-authors obtained videos of 53 episodes from Australia and the Netherlands. From these they determined, for each contestant and each round, the situation the contestant faced in terms of the amounts of money in the remaining briefcases, the banker's offer and their decision to accept or refuse the offer. They then analyze this data using a measure of relative risk aversion developed by Arrow and Pratt (RRA). The authors write:In every game round, a unique RRA coefficient can be determined at which the contestant would be indifferent between accepting and rejecting the bank offer. If the contestant accepts the offer, his RRA must be higher than this value; if the offer is rejected, his RRA must be lower.
They also investigate how this RRA changes according to other variables such as estimated income and education.
The authors find that, for a wealth level of €25,000, an average over all the contestants gives an RRA 1.61. This estimate decreases when the wealth level decreases and increases when it increases. The degree of risk aversion differs strongly across the contestants, some exhibiting strong risk averse behavior (RRA > 5) and others risk seeking behavior (RRA < 0). They say:The degree of risk aversion differs strongly across the contestants, some exhibiting strong risk averse behavior (RRA > 5) and others risk seeking behavior (RRA< 0). The differences can be explained in large part by the earlier outcomes experienced by the contestants in previous rounds of the game. Most notably, RRA generally decreases following losses. Contestants facing a large reduction in the expected prize during the game even exhibit risk seeking behavior.
(1) Suppose that at each round the banker offered you the expected value of the amount in your briefcase and you were only interested in maximizing your expected winning. Would it matter when you stopped?
(2) Professor Post tells us:The banker behaved in a predictable manner, with offers around 5-10% of expected value in round one and 100% in the later rounds. Also, for losers, the offers are relatively generous, sometimes exceeding 100%.
This leaves many questions about how the banker made his offers: Is there a real banker or is it all done by an algorithm? The Bother's Bar comments suggest this is not the case in the U.K. Does the offer depend only on the expected amount in a player's suitcase. If you are teaching a probability or statistics course you could ask the students to look into one of these questions or to make their own hypothesis about how the offers are made.
(3) Suppose you knew how the offer is determined. How would this effect your strategy for playing the game.
(4) Suppose you are told that the banker knows what is in your suitcase. How would this affect the way your played the game?
(5) Using the Bother's Bar data, compare the sum of the prizes given out with the sum off the Box values when the prizes are awarded.
Submited by Laurie Snell
Predicting the sex of fetus has been around for as long as children have been born. According to Roger Dobson of The Independent (March 7, 2006) there are many "theories from the wilder side."
- Coffee: If a man has coffee before sex, the Y-sperm is more active and likely to result in a boy.
- Dreams: Whatever sex of child a mother dreams of having, she will have the opposite.
- Age: As a mother gets older, her chance of conceiving a boy increases.
- Positions: Missionary position is best for a girl, says Italian folklore.
- Underwear: Men who wear loose underwear are more likely to produce boys.
And several others which we would also view as less than scientific. Now, however, modern methods of predictions with probabilities attached exist. Unfortunately, it is difficult to tell from the article whether it is P(X|Y) or P(Y|X). For example, "The Whelan method, named after Dr. Elizabeth Whelen," and depends on the timing of ovulation, "is claimed to be some 68 per cent effective for boys and 56 per cent for girls." Assuming that boys and girls are equally likely and "effective" means P(Test Boy|Boy) = .68, and P(Test Girl|Girl) = .56, translates via Bayesian inversion to a not very impressive P(Boy|Test Boy) = .61 and P(Girl|Test Girl) =.64.
"Another technique, the MicroSort method" which depends on DNA relating to the difference in size of the X and Y chromosomes, yields "success" rates of 88 per cent for girls and 73 per cent for boys. Assuming again that boys and girls are equally likely and "success" is synonymous with "effective," this translates via Bayesian inversion to a slightly more striking P(Boy|Test Boy) = .86 and P(Girl|Test Girl) =.77.
None of these methods can compare with that of the Acu-Gen Biolab which claims 99.9 per cent "accuracy," a term which sounds like "effective" or "success" but may not mean anything because some mothers who paid $25 for the kit and the $250 for the analysis wound up with a child of the other kind than what they were told. According to the Boston Globe (March 1, 2006), "Alleging that their pregnancies had been marred by fraud, 16 women have filed a class-action lawsuit against a Lowell laboratory that promises to determine the gender of an embryo by testing the mother's blood just five weeks after conception. In the suit filed in US District Court in Boston, the women charged that despite its claim of 99.9 percent accuracy, Acu-Gen Biolab of Lowell got the genders of their babies wrong, causing confusion and distress, and then refused to make good on its double-your-money-back guarantee."
Whether believer or buyer, always beware.
1. What are the ethical implications of an inexpensive, non-invasive sex prediction procedure which is available very early on in a pregnancy? What are the implications for the anti-abortion movement and for the pro-choice adherents?
2. Comment upon the linguistic dangers of using ordinary English, as in "effective," "success" and "accuracy" for well-defined statistical concepts.
3. Are you surprised that Acu-Gen Biolab "refused to make good on its double-your-money-back guarantee"? What excuses or reasons do you think the firm would give for claiming the fault lies with the woman and not with the technique?
Submitted by Paul Alper
Some children left behind
Joseph Stalin's famous quote regarding elections, "The people who count the votes decide everything" has an analogy to the bizarre world of the No Child Left Behind Act. Regardless of the merits of the act, it is in the interests of those who will be judged by this law to look good. As a consequence, it appears that there are several ways, other than outright cheating or the students actually testing higher, to make the numbers come out better.
In Minnesota, according to James Walsh of the Minneapolis Star Tribune (March 6, 2006), the trick is to have the low-scoring students take a different test. "To some, it may look like we're gaming the system" says Tim Vansickle, the Minnesota director of assessments. Minnesota does have many students whose native language is not English and Vansickle denies that the alternative test is not meant to help schools keep off "the list" of schools deemed failing.
Another technique according to Jay Mathews of the Washington Post (February 28, 2006), is to engage in strange arithmetic. He quotes Monte Dawson, the testing and assessment director for Alexandria, Virginia: "Remediation Recovery, which has been around since 2001, means that fourth grade students who failed the third grade test in 2004, got to retake the third grade test in 2005. Up until this year (2005), if they passed the third grade test, then they were included in the numerator only of the calculation to determine the third grade passing score. As illustration, if 4 out of 5 third grade students passed and 1 out of 5 fourth grade Remediation Recovery students passed, the passing percentage would be 100 percent."
Mathews, the Washington Post's education columnist, is puzzled by this explanation so he leaves it to the spokesman for the Virginia Department of Education, Charles Pyle. Mathews translates Pyle's remarks thusly: "in 2000, the state school board changed the counting procedure to encourage more schools to do what Maury [the school featured in the article] did -- give the students who failed some extra help and let them try again. Often the second-test passing rates of students who flunk a test initially are lower than their class's overall passing rate, since they are the class's weakest students. So if those second-test results were combined with the first test results in the usual way, it would likely lower the overall percentage and make the school look worse than otherwise. School districts in Virginia figured this out and resisted the urge to work with their lowest-performing students and test them again."
Therefore, to avoid this inevitable drop in performance should a school actually try to help its students rather than game the system, Mathews uses the following illustration: "To give schools an incentive to make that effort, the school board ordered an unorthodox change in the way the school percentage would be calculated after the retesting. If a school had 100 students, with 30 failing the test the first time and 10 of those passing the test the second time, they could add 10 to the 70 who passed the first time, divide those 80 passing students by 100, and get a nice boost from 70 to 80 percent in their passing rate. Done the conventional way, they would have had to add 30 to the denominator as they added 10 to the numerator, and gotten a passing rate of only 62 percent, lower than the 70 percent rate they had before."
Stalin was undoubtedly an evil person but his remark about defining and controlling the counting and evaluation procedure has many followers who probably don't realize the inspiration for their creativity. Perhaps the moral of the story for statisticians is the well-known dictum, "the data never speaks for itself." People have great ingenuity particularly when it comes to guarding their vested interests via number manipulation.
1. Is the purpose of the No Child Left Behind Act punishment of teachers and administrators or is the purpose to improve performance of the students?
2. Why is the Act controversial?
3. If the rules of the game allow the numerator to increase while not changing the denominator, what is there to complain about?
4. If you were a teacher or administrator would you be tempted to suggest to your poorer performing students to stay home on the day of the test?
Submitted by Paul Alper
Norton Starr sugested the next article.
The future divined by the crowd
The Future Divined by the Crowd
New York Times, March 11, 2006
Michael Mauboussin is a well known Wall Street strategist who also is an adjunct professor at Columbia Business School. Every year since 1993 he has asked his students to vote for the winners in 12 categories including major and relatively obscure contests. We read:This year, the pick that got the most votes — the consensus pick, he calls it — turned out to be right in 9 of the 12 categories, including, amazingly enough, film editing and art direction. And yet, of the 47 students who participated, only one matched the accuracy of the consensus. None did better, and most did much worse; according to Mr. Mauboussin, the average number of correct answers per ballot this year was only 4.1. "It has never failed," he said. "The consensus invariably does much better than the average student."
The article also discusses other evidence that consensus predictions are often better than individual predictions such as the Iowa Electronic Market ((Chance News 13.06) , Holywood Stock Exchange (Chance News 12.02), and a new one HedgeStreet whose motto is: "It's your economy. Trade it."
(1) Choose one of the 12 categories and assume that there are 5 nominees for this category labeled 1,2,3,4,5. Assume that each of the 47 students rolls a five-sided die to determine their prediction. What is the probability that that their consensus vote is the Oscar winner for this category?(2) Assume that there are 5 contestants for each of the 12 categories and the 47 students guess the winner in each category. What is the probability that their consensus vote is the Oscar winner in 9 or more of the 12 categories?