Donald Berry
Feb 16, 2018
I'm very interested in communicating statistical results to courts, and communicating statistical ideas and conclusions to people more generally. I've testified in many cases in a variety of areas. I'll give two examples, one that I wrote about and the other in which I testified, although in an unusual way.
The first is O.J. Simpson's murder trial. It's the most important courtroom example in the U.S. where statistics and communicating statistics played critical roles, although in less than a positive way. Internationally, in my view, the most important cases regarding the importance of communicating statistics are the Sally Clark case in the U.K. and Lucia de Berk in the Netherlands. (Google them.) As regards the former, Peter Donnelly's rendition is arguably the most elegant TED talk that exists: <https://www.ted.com/talks/peter_donnelly_shows_how_stats_fool_juries/transcript>
I wrote an article (Berry DA. DNA, Statistics and the Simpson Case. Chance 7(1994)(4):9-12) in advance of the Simpson trial. The article did not predict the result, of course, but it did try to illuminate the ability for courts to understand numbers and what they mean. And that turned out to be a major, possibly defining issue in the actual trial. Here's one of the paragraphs from my article:
"A common defense tactic is to muddy the waters. A way to do this is to find experts who will testify to something different from the testimony of prosecution experts--it doesn't really matter what it is as long as it's different. Testimony involving numbers is especially vulnerable to smoke screens. Match proportions depend on assumed bin size, measurement standard deviation, database, etc. In one California case, match proportions presented to the court varied from 1 in 70,000 to 1 in 700 million, depending on assumptions made. Unsophisticated jurors can become confused--numerical calculations designed to make evidence more informative can have the opposite effect. In this particular case, the court was confused as well--it ruled that the discrepancies warranted excluding all match proportion estimates!"
I'll say here a bit more about these two very different match proportions. At issue was the molecular weights of DNA fragments. A particular sample had shown a single fragment. But there had to be two fragments, one maternal and the other paternal. The defense assumed that the other fragment had washed off the end of the gel and therefore its weight was censored at the size that could be weighed by the process. Result: 1/70,000. The prosecution expert assumed homozygosity, that the both fragments were measured and were on top of each other. Result: 10,000 times smaller. Disallowing the evidence to be presented in the case was amazingly uninformed because the evidence was strongly incriminating regardless of which expert was right.
Bruce Weir was the statistician for the prosecution in the actual Simpson trial. In his testimony on a Friday he presented some calculations of prevalence of DNA bands in the general population. Pushed by the defense, over the weekend he realized he had made a mistake. He had forgotten a factor of 2. It's the Hardy-Weinberg 2. For example, the probability that a couple's two children are a boy and a girl is approximately ½ x ½ = ¼, right? Well, only if the boy-girl question is something like, the older is a boy and the younger is a girl. If it's just boy-girl in any order then the right answer is twice ¼ or ½. Nothing is statistics is simpler than this. We all make mistakes similar to Weir's, but we usually catch them before we announce them. On Monday Weir testified to his mistake. For example, when he had calculated 1/1600 for the blood on the famous glove being Simpson's, it should have been 1/800.
The defense attorneys had a field day with Weir's mistake. They were unrelenting. They focused on his reliability and not so much on the fact that the jury had been presented with two numbers for the same quantity. See the report in the next day's LA Times: <http://articles.latimes.com/1995-06-27/local/me-17506_1_dna-analysis>.
As in the earlier example, both match probabilities, 1/800 and 1/1600, are incriminating. But the defense attorneys were successful in turning the evidence on its head and worked to convince the jury that they couldn't believe any of it. Apparently, they succeeded.
* * *
The other example is athlete doping. In 2011 I represented an Estonian Olympic gold medalist in cross-country skiing at the Court of Arbitration for Sport (CAS) in Lausanne, Switzerland. (Pro bono, by the way.) He had been banned from sport for three years for using human growth hormone in out-of-competition training. Tests had been carried out by the World Anti-Doping Agency (WADA). He appealed the ban to the CAS. For the first time, the ban was overturned. And, as clearly stated in the CAS ruling, the exclusive reason was statistics. (There has been one other ban overturned by CAS since that time. That was a Russian runner whose sample was tested beyond the statute of limitations.)
The CAS hearings are different from most hearings. The CAS agreed with the athlete's request that I be allowed to sit with the CAS hearing panel and ask questions of witnesses from both sides. Krista Fischer and I chronicled our experiences in this article: (Fischer K, Berry DA. Statisticians introduce science to international doping agency: The Andrus Veerpalu case. Chance 27(2014)(3):10-16.) Krista is one of a handful of statisticians in Estonia, the smallest Baltic country at 1.3 million people.
At the hearing the WADA statistician presented the way he determined the decision limit (DL), the measurement above which the athlete is determined to be a doper. He had fewer than 200 cases that had been collected from standard competitions. So some of these values might have been from dopers. So he deleted the highest values! Then he fit a log-normal distribution and found the DL to be the 99.99 percentile of the distribution. So he (and WADA) claimed a false-positive rate for the test of 1 in 10,000.
There were the two critical problems with this method that I pointed out at the hearing ,,, two critical "communications." First I asked the WADA statistician how he could conclude that only 1 in 10,000 non-dopers would have a higher value of the DL when his sample contained fewer than 200 people.
I also posed the other critical problem with his analysis-his rejection of "outliers"-as a question:
At the hearing, Berry asked the statistician representing WADA the following rhetorical question: "So you're saying that you cut off the tail of the distribution and then claimed that it had a small tail, is that correct?"
I could tell by follow-up questions of this witness that one particular member of the CAS really got it … he understood what I was saying. So the fact that the athlete's ban was lifted may have had more to do with this person being on the CAS panel and not my asking these questions.
If you want further reading on the statistical issues of doping: Berry DA. Commentary: The science of doping. Nature 454(2008):692-693. Up until the day the article went to press its title was "The science of doping … or lack thereof." But Nature's editors dropped the last three words from the title under threat of a law suit by WADA. Despite my raking WADA over the coals in this article and in the Chance article with Krista Fischer, I've never been contacted by them regarding how to improve their science. After the Veerpalu debacle WADA did hire some local (mediocre) statisticians in Montreal to endorse what they're doing and, surprise!, they did endorse what WADA had been doing.
Also, here's a news article regarding Veerpalu: <https://www.outsideonline.com/1925761/whats-wrong-world-anti-doping-agency>
* * *
If you're in the mood for a bit of humor, a lot of funny things happen in court … but nobody laughs! Here's the funniest anecdote from my courtroom experience. It was the 1980s. I was testifying (also pro bono) for the Attorney General of Minnesota. The State was prosecuting the defendant for operating a marketing business that was in effect a pyramid scheme with a product being sold only to cover up the scam to avoid the laws against pyramids. There was only one other witness for the State, someone from the Attorney General's office who testified that the scam was in fact operating in Minnesota. But the room was packed, mostly with lawyers from the "big city." The judge was Esther Tomljanovich. I was on the stand sitting next to her. The State's attorney began asking questions of me. One question elicited "I object" from one of the big city lawyers. He then explained his objection, talking to me! He said the question didn't deal with mathematics or statistics, which were my areas of competence. He launched into a detailed explanation and not incidentally was doing mathematics and statistics along the way … which ironically happened not to be among his competences. And he finished up addressing me saying, "And therefore it's not mathematics or statistics, is it sir?" Now, I knew place. I turned to the judge. She looked at me and said, "Do you want to rule on this or should I?"
That's the story. Remember I said this was the 1980s. But the rest of the story is kinda cute. She started to say things, looking alternatively at the suitably red-faced lawyer and me. It became clear that she didn't know what to rule. So I spoke up, asking her if it would help for me to say what I understood the State's attorney to be asking. She said "Yes." So I did. And she replied, "Objection overruled!"
I suppose the real "rest of the story" is the result of the case: the State won.
------------------------------
Donald Berry
Professor
Univ of Texas MD Anderson Cancer Center-Department of Biostatistics
------------------------------
Original Message:
Sent: 02-13-2018 11:27
From: Nick Thieme
Subject: Looking for experts on a story about statistics and the law
Hi all!
I'm science/stats journalist currently working on a freelance project with Significance Magazine at the intersection of statistics and the law. Without going into too much detail, an engineer collected data on yellow light timings in his city of residence and, after comparing them against the city's required yellow light duration, he sued the city for violating their own regulations. He ended up losing that case.
I'm writing a story about him, about his use of statistics (which was limited), and about how statistics can be used to persuade in the courtroom. I was hoping to talk to some experts in the use of statistics as a tool for legal persuasion. This seemed like a good place to turn. By any chance, would anyone be interested in a brief discussion on the topic?
I'd be happy to send you more information about the story or about my previous writing.
Thanks!
Nick
------------------------------
Nick Thieme
------------------------------