3 Unit C: Special Topics in Forensics 3 Unit C: Special Topics in Forensics
3.1 Class 9: The Persistent Past - Bitemarks 3.1 Class 9: The Persistent Past - Bitemarks
Pages 67-68 and 83-87 in Chapter 5 of the 2016 PCAST Report on “Bitemark Analysis.”
The selected pages cover the introduction to the chapter and PCAST’s assessment of bitemark evidence.
How the flawed ‘science’ of bite mark analysis has sent innocent people to prison How the flawed ‘science’ of bite mark analysis has sent innocent people to prison
Randy Balko, The Washington Post, a four-part series (Feb. 2015)
This in-depth series is an excellent assessment of the fallibility of bite mark evidence.
This is an excerpt from an excellent four part series on bitemark evidence. You can access the full story at this link.
Excerpt from Part 3 of 4:
In 2007 Mary and Peter Bush, a married couple who head up a team of researchers at the State University of New York at Buffalo, began a project to do what no one had done in the three decades — conduct tests to see whether there’s any scientific validity to the bite mark evidence presented in courts across the United States.
The Bushes sought to test the two main underlying premises of bite mark matching — that human dentition is unique and that human skin can record and preserve bite marks in a way that allows for analysts to reliably match the marks to a suspect’s teeth. The Bush team was the first to apply sophisticated statistical modeling to both questions. It was also the first to perform such tests using dental molds with human cadavers. Previous tests had used animal skins.
When they first set out on the project, the Bushes received preliminary support from some people in the bite mark analyst community. “Franklin Wright was the ABFO president at the time,” says Mary Bush. “He visited our lab, and then put up a message praising our work on the ABFO website.” They also received a small grant from the ASFO, the discipline’s non-accrediting advocacy and research organization.
“There was a lot enthusiasm at the outset,” says Fabricant. “I think some analysts were excited about the possibility of getting some scientific validation for their field.”
But when the Bushes began to come back with results that called the entire discipline into question, that support quickly dried up.
The Bushes’ research found no scientific basis for the premise that human dentition is unique. They also found no support for the premise that human skin is capable of recording and preserving bite marks in a useful way. The evidence all pointed to what critics such as Bowers had always suspected: Bite mark matching is entirely subjective. The Bushes’ first article appeared in the January 2009 issue of the Journal of Forensic Sciences. The couple have since published a dozen more, all in peer-reviewed journals.
Outside of ABFO and their supporters, the Bushes’ research has been lauded. “I think there’s a chance that because of the Bushes’ research, five years from now we aren’t going to be talking about bite mark evidence anymore,” says Risinger. “It’s that good. Their data is solid. Their methodology is solid. And it’s conclusive.”
Other legal scholars and experts on law and scientific evidence interviewed for this article shared Risinger’s praise for the Bushes’ research but were less optimistic about its implications, in part because the criminal justice system so far hasn’t recognized the significance of their work.
But from a scientific standpoint, the Bushes’ research was a direct and severe blow to the credibility of bite mark analysis. At least initially, it threatened to send the entire field the way of voice print matching and bullet lead analysis, both of which have now been discredited. And so when defense attorneys began asking the couple to testify in court, the bite mark analysts fought back with a nasty campaign to undermine the Bushes’ credibility. In a letter to the editor of the Journal of Forensic Sciences, seven bite mark specialists joined up to attack the Bushes in unusually harsh terms for a professional journal. When that letter was rejected for publication, five of the same analysts wrote another. That, too, was rejected. A toned-down but still cutting third letter was finally published.
In the unpublished letter dated November 2012, the authors — all bite mark analysts who hold or have held positions within ABFO — declared it “outrageous that any of these authors would go into courts of law and give sworn testimony citing this research as the basis for conclusions or opinions related to actual bite mark casework, especially considering that no independent research has validated or confirmed their methods or findings.”
Of course, critics would say this was a bit of rhetorical jujitsu — that the last clause could describe exactly what bite mark analysts have been doing for 35 years. For emphasis they added, “This violates important principles of both science and justice.” In the other letter, the authors referred to the Bushes’ testimony in an Ohio case, which was based upon their research, as “influenced by bias” and “reprehensible and inexcusable.”
The primary criticism of the Bushes’ research is that they used vice clamps to make direct bites into cadavers that were stationary through the entire process. This is quite a different scenario than the way a bite would be administered during an attack. During an assault, the victim would probably be pulling away, causing the teeth to drag across the skin. For the Bush tests, the clamp they used to make the bites moved only up and down. A human jaw also moves side to side. A biter might also twist his head or grind his teeth. A live body will also fight the bite at the source to prevent infection, causing bruising, clotting and various other defenses that would alter the appearance of the bite.
“We acknowledge that our lab tests are different from how bites are made in the real world,” says Mary Bush. “But to the extent that our tests differed, they should have made for better preserved samples.”
In other words, the tests that the Bushes conducted made for cleaner, clearer bites that could be easily analyzed. If they were in error, they were in error to the benefit of the claims of bite mark analysts. And they still found no evidence to support the field’s two basic principles.
“That’s exactly right,” says Risinger. “If there was any validity to bite mark analysis at all, these tests would have found it. They gave the field the benefit of the doubt. The evidence just wasn’t there. Their data is very, very strong.”
To argue that the Bushes’ experiments should be disregarded because they weren’t able to replicate real-world bites is also an implicit acknowledgment that real-world bites aren’t replicable in a lab, and therefore aren’t testable. You won’t find many people volunteering to allow someone else to violently bite them for the purposes of lab research. Even if you could, a volunteer won’t react the same way to a bite that an unwitting recipient might.
The Bushes’ research not only failed to find any scientific support for bite mark matching, but it also exposed the fact that for four decades the bite mark community neglected to conduct or pursue any testing of its own. It put the ABFO and its members on the defensive. The bite mark analysts responded by intensifying their attacks on the couple and making the attacks more personal.
At the February 2014 AAFS conference in Seattle, the ABFO hosted a dinner for its members. The keynote speaker was Melissa Mourges, an assistant district attorney in Manhattan, one of the most outspoken defenders of bite mark matching in law enforcement.
Mourges already had a high profile. The combative, media-savvy prosecutor was part of the prosecution team featured in the HBO documentary “Sex Crimes Unit,” which followed the similarly named section of the Manhattan DA’s office, the oldest of its kind in the country. Mourges herself founded a cold-case team within that unit. At the 2012 AAFS conference she spoke on a panel called “How to Write Bestselling Novels and Screenplays in Your Spare Time: Tips From the Pros.” At this year’s conference, she’ll be on a panel that’s titled “Bitemarks From the Emergency Room to the Courtroom: The Importance of the Expert in Forensic Odontology.” She’ll be co-presenting with Franklin Wright, the former ABFO president who initially supported the Bushes’ research.
Mourges was also the lead prosecutor in State v. Dean, a New York City murder case in which the defense challenged the validity of the state’s bite mark testimony. In 2013, Manhattan state Supreme Court Judge Maxwell Wiley held a hearing on the scientific validity of bite mark evidence. Mary Bush testified about the couple’s research for the defense. It was the first (and so far the only) such hearing since the NAS report was released, and both sides of the bite mark debate watched with anticipation. In September 2014, Wiley ruled for the prosecution, once again allowing bite mark evidence to be used at trial. (I’ll have more on the Dean case in part four of the series.) Mourges’s talk at the ABFO dinner was basically a victory lap.
There’s no transcript of Mourges’s speech, but those in attendance say it was basically a no-holds-barred attack on Mary Bush. Cynthia Brzozowski has been practicing dentistry in Long Island for 28 years and sits on the ABFO Board of Directors. She practices the widely accepted form of forensic dentistry that uses dental records to identify human remains, but she doesn’t do bite mark matching, and she won’t testify in bite mark cases. Brzozowski was at the dinner in Seattle and says she still can’t believe what she heard from Mourges.
“Her tone was demeaning,” Brzozowski says. “It would be one thing if she had just come out and presented the facts of the case, but this was personal vitriol against the Bushes because of their research.”
According to Brzozowski, Mourges even went after Mary Bush’s physical appearance. “At one point, she put up an unflattering photo of Mary Bush on the overhead. I don’t know where she got it, or if it had been altered. Mary Bush is not an unattractive person. But it was unnecessary. You could hear gasps in the audience. It was clear that she had chosen the least flattering image she could find. Then she said, ‘And she looks better here than she does in person.’ It was mean. I had to turn my back. I was mortified.”
Other ABFO members — including two other members of the board of directors — also complained, to both the ABFO and the AAFS. The complainants described Mourges’s attack on Bush as “malicious,” “bullying” and “degrading.” According to accounts of those in attendance, other members were also upset by Mourges’s remarks but didn’t file formal complaints for fear of professional retaliation.
A few weeks later, Loomis sent an e-mail to the ABFO Board of Directors to address the complaints. Loomis defended Mourges and her presentation. He described the dinner as a “convivial affair” where members can socialize, have a libation and “be entertained” by the invited speaker. He argued that “anyone who understands litigation” should not have been unsettled by the talk and described the presentation as “sarcastic, serious, and even light-hearted.” He stood by the decision of his predecessor, Greg Golden, to invite Mourges, calling it “a good decision,” adding, “I apologize to those who were offended. However, I do not apologize for the message.”
“‘Bullying’ is exactly what it is,” says Peter Bush. “We’re scientists. We’re used to collegial disagreement. But we had no idea our research would inspire this kind of anger.”
. . .
Addendum: After this post was published, the office of Manhattan District Attorney Cyrus Vance sent the following statement:
Melissa Mourges is a veteran prosecutor and a nationally recognized leader in her field. As Chief of the Manhattan District Attorney’s Forensic Science/Cold Case Unit, she has solved dozens of cold case homicides, including two recently attributed to “Dating Game” serial killer Rodney Alcala. In addition to being a Fellow at the American Academy of Forensic Sciences, ADA Mourges has also served as co-chief of the DNA Cold Case Project, which uses DNA technology to investigate and prosecute unsolved sexual assaults. As part of that work, she pioneered the use of John Doe indictments to stop the clock on statutes of limitation and bring decades-old sexual assaults to trial. Her work and reputation are impeccable, and her record speaks for itself.
Excerpt from Part 4 of 4:
The most significant challenge to bite mark evidence since the NAS report was released came in State v. Dean, the New York case mentioned in part three of this series. In 2013, attorneys for defendant Clarence Dean challenged the prosecution’s plan to use bite mark evidence against their client. Manhattan state Supreme Court Judge Maxwell Wiley granted a hearing to assess the validity of bite mark matching. It was the first such hearing since the NAS report was published, and both sides of the bite mark debate watched closely. Mary Bush testified for the defense, as did Karen Kafadar, chair of the statistics department at the University of Virginia and a member of the National Institute of Standards and Technology’s Forensic Science Standards Board.
The prosecutor in that case was Manhattan assistant district attorney Melissa Mourges, an aggressive 30-year prosecutor with a high profile. Mourges was featured in a 2011 HBO documentary and holds the title of chief of the District Attorney’s Forensic Science/Cold Case Unit in what is arguably the most influential DA’s office in the country. So her advocacy for bite mark matching is significant.
As reported in part three, Mourges has not only defended bite mark evidence but also seems to be on a campaign to denigrate its critics, going so far as to heckle scientific researchers Mary and Peter Bush at a panel, and then to personally attack Mary Bush during a dinner talk at a forensics conference. Her bite mark brief in the Dean case compared bite mark evidence critic Michael Bowers to the notorious bite mark charlatan Michael West. It was a particularly egregious comparison because Bowers had helped expose West back when he was still embraced by the ABFO.
In her brief, Mourges first encouraged Wiley to embrace the “soft science” approach to bite mark analysis used by the Texas court in Coronado. Conveniently, doing so would allow bite mark specialists to testify to jurors as experts with almost no scrutiny of their claims at all.
Mourges next argued that if the court must do an analysis of the validity of bite mark testimony, it do so on the narrowest grounds possible. When it comes to assessing the validity of scientific evidence, New York still goes by the older Frye standard, which states that evidence must be “generally accepted” by the relevant scientific community. The question then becomes: What is the relevant scientific community?
In her brief, Mourges urged Wiley to limit that community to analysts who “have actually done real-world cases.” In other words, when assessing whether bite mark matching is generally accepted within the scientific community, Mourges says the only relevant “community” is other bite mark analysts.
Saks offers a metaphor to illustrate what Mourges is asking. “Imagine if the court were trying to assess the scientific validity of astrology. She’s saying that in doing so, the court should only consult with other astrologers,” he says. ”She’s saying the court shouldn’t consult with astronomers or cosmologists or astrophysicists. Only astrologers. It’s preposterous.”
Saks, who submitted a brief in the case on behalf of Dean, also offers a real-world example: the now-discredited forensic field of voiceprint identification. The FBI had used voiceprinting in criminal cases in the 1970s but discontinued the practice after an NAS report found no scientific support for the idea that an expert could definitively match a recording of a human voice to the person who said it.
“If you look at the Frye hearings on voiceprint identification, when judges limited the relevant scientific community to other voiceprint analysts, they upheld the testimony every time,” Saks said. “When they defined the relevant scientific community more broadly, they rejected it every time. It really is all about how you define it.”
In urging Wiley to only consider other bite mark analysts, Mourges also casts aspersions on the scientists, academics and legal advocates urging forensics reform. She writes:
The make-up of the relevant scientific community is and should be those who have the knowledge, training and experience in bitemark analysis and who have actually done real world cases. We enter a looking-glass world when the defense urges that the Court ignore the opinions of working men and women who make up the ranks of board-certified forensic odontologists, who respond to emergency rooms and morgues, who retrieve, preserve, analyze and compare evidence, who make the reports and who stand by their reasoned opinions under oath. The defense would instead have this Court rely on the opinions of statisticians, law professors and other academics who do not and could not do the work in question.
Of course, one needn’t practice astrology or palm reading to know that they aren’t grounded in science. And if police and prosecutors were to consult with either in a case, we wouldn’t dismiss critics of either practice by pointing out that the critics themselves have never read a palm or charted a horoscope.
Mourges also attempts to both discredit the NAS report and claim that it isn’t actually all that critical of bite mark analysis. For example, she laments that the report was written by scientists and academics, not bite mark analysts themselves. This, again, was entirely the point. The purpose of the NAS report was to research the scientific validity of entire fields. If it were written by active practitioners within those fields, every field of forensics would have been deemed valid, authoritative and scientifically sound.
Mourges also misstates and mischaracterizes what the report actually says. She writes in one part of her brief that “the NAS report does not state that forensic odontology as a field should be discredited.” That’s true. But bite mark matching is only one part of forensic odontology. The other part, the use of dental records to identify human remains, is widely accepted. What the report makes abundantly clear is that there is zero scientific research to support bite mark analysis in the manner it is widely practiced and used in courtrooms.
In another portion of the brief, Mourges selectively quotes part of the the report, cutting out some critical language. She writes:
When Dr. Kafadar and her NAS committee created the NAS report, they wrote a summary assessment of forensic odontology. In it they said that “the majority of forensic odontologists are satisfied that bite marks can demonstrate sufficient detail or positive identification …
That ellipsis is important, as is the word that comes before the quote. Here’s the passage quoted in full:
Although the majority of forensic odontologists are satisfied that bite marks can demonstrate sufficient detail for positive identification, no scientific studies support this assessment, and no large population studies have been conducted. In numerous instances, experts diverge widely in their evaluations of the same bite mark evidence, which has led to questioning of the value and scientific objectivity of such evidence.
Bite mark testimony has been criticized basically on the same grounds as testimony by questioned document examiners and microscopic hair examiners. The committee received no evidence of an existing scientific basis for identifying an individual to the exclusion of all others.
The report only acknowledges the near consensus within the community of bite mark analysts for the purpose of criticizing them. Mourges’s selective quotation implies that the report says the relevant scientific community accepts bite mark matching. The full passage reveals that the report is essentially pointing out just the opposite: The insular community of bite mark analysts may believe in what they do, but the larger scientific community is far more skeptical.
One common tactic that shows up in Mourges’s brief and has also shown up in defenses of bite mark analysis across multiple forums — court opinions, forensic odontology journals and public debates — is a sort of meticulous recounting of the care and precision into which bite mark analysts collect and preserve evidence as well as the scientific-sounding nomenclature used by the field’s practitioners. Mourges devotes more than 10 pages to laying out the procedures, methods and jargon of bite mark matching.
In any field of forensics it’s of course important that evidence be carefully handled, properly preserved and guarded against contamination. But to go back to the astrology metaphor, even the most careful, conscientious, detail-oriented astrologer . . . is still practicing astrology. If the field of bite mark analysis cannot guarantee reliable and predictable conclusions from multiple practitioners looking at the same piece of evidence, if it cannot produce a margin for error, if its central premises cannot be proved with testing, then it doesn’t matter how pristine the bite mark specimens are when they’re analyzed or what the mean number of syllables may be in each word of a bite mark analyst’s report.
But ultimately, Mourges was effective. In September 2013, Wiley rejected the defense challenge to bite mark evidence in the Dean case. He never provided a written explanation for his ruling. In an e-mail, Joan Vollero, director of communications for the Manhattan District Attorney’s Office, wrote of the ruling: “Following the months-long Frye hearing, Judge Wiley denied the defendant’s motion to preclude the bite mark evidence, finding that the field of bite mark analysis and comparison comports with New York State law.”
Texas science commission is first in the U.S. to recommend moratorium on bite mark evidence Texas science commission is first in the U.S. to recommend moratorium on bite mark evidence
By Brandi Grissom, Trail Blazers Blog, Feb. 12. 2016
Texas science commission is first in the U.S. to recommend moratorium on bite mark evidence
By Brandi Grissom, Trail Blazers Blog, Feb. 12. 2016
Texas on Friday became the first state in the nation to recommend a ban on the use of bite mark analysis in criminal cases, a decision that could prompt change in courtrooms nationwide.
The Texas Forensic Science Commission recommended courts institute a moratorium on the use of bite marks until additional scientific research is done to confirm its validity. The decision comes after Steven Mark Chaney, a Dallas man, was freed from prison last fall when the court agreed that bite mark evidence used to convict him of murder was bogus. Courts nationwide have used such evidence for decades to identify suspects in murders, sexual assaults, child abuse and other violent crimes. In recent years, though, critics have posed serious questions about whether bite marks can be used accurately pinpoint the perpetrator of a crime. Texas, which has become a national leader in forensic science developments, becomes the first jurisdiction to advocate abandoning it as evidence in court until scientific criteria for its use is created.
“It shouldn’t be brought into court until those standards are done,” said Richard Alpert, an assistant criminal district attorney for the Tarrant County District Attorney’s Office and a member of the Texas Forensic Science Commission.
The New York-based Innocence Project asked the commission last fall to investigate the use of bite mark analysis as evidence. The Innocence Project and other critics of the practice argue that bite mark analysis has no scientific basis and shouldn’t be used in courtrooms.
Chris Fabricant, director of strategic litigation for the Innocence Project called the decision “incredibly significant.”
“The Texas Forensic Science Commission has taken a giant step in purging unscientific and unreliable bite mark evidence from court rooms nationwide,” he said in a prepared statement.
Members of the American Board of Forensic Odontology, which oversees the practice of bite mark analysis, have fiercely defended the validity of bite mark analysis. The dentists acknowledge some mistakes with analysis in the past but contend that when used appropriately, such evidence can be a useful investigative tool.
Bite marks, they say, can be particularly useful as evidence in child abuse cases, where the patterns left on a child’s body can help get a young one out of a dangerous household.
“What we hope for is to avoid unintended consequences,” said dentist David Senn, a clinical assistant professor at the University of Texas Health Science Center at San Antonio Dental School and member of the ABFO.
Chaney, whose case prompted the commission’s decision Friday, was freed from prison in October after Dallas County prosecutors agreed that the bite mark evidence that led to his conviction for a 1987 murder was invalid. Dr. Jim Hales told a Dallas County jury that there was a “1-to-a million” chance that someone other than Chaney left a bite mark on the arm of John Sweek, who had been stabbed to death.
Chaney’s case was hardly the first to involve bite mark evidence, though. The practice seems to date back to a 1954 robbery case in West Texas. A grocery store owner arrived one December morning to find someone had ransacked his shop and stolen two bottles of whiskey and 13 silver dollars. The culprit had left behind a mess of cold cuts and cheese on the meat counter.
Police found James Doyles’ teeth marks in a piece of cheese. They forced Doyle — who was already in jail for public intoxication — to bite into another piece of cheese.
A dentist concluded the teeth marks matched, and Doyle was sentenced to two years for burglary.
In recent years, a number of scientific studies, including a 2009 report from the National Academy of Sciences, have found that there is little to no scientific basis for conclusions that match a person to a set of apparent bite marks.
At least one study found that dentists were largely unable to agree on which injuries were the result of a bite. Other experts have raised questions about whether skin is an accurate template on which to assess bite marks and about how the passage of time affects the reliability of the patterns that teeth leave.
The Associated Press reported in 2013 that at least 24 people in the U.S., including two in Texas, had been exonerated in cases in which bite mark evidence played a central role in the conviction.
Even the ABFO, which certifies dentists who analyze bite marks, has decided the evidence can’t be used to draw the kind of strong conclusions that Hale made during Chaney’s trial.
“There’s no question that was improper,” Senn told commission members Thursday.
The commissioners left open the possibility that bite mark analysis could receive approval in the future if additional research and testing was performed that improved the reliability of it as evidence.
Senn told a panel of commission members on Thursday that certified ABFO dentists in Texas – there are nine of them – would work with state officials to keep them informed of additional research. Those dentists, he said, would also provide the commission with a list of the Texas cases in which they provided bite mark testimony.
That will help the commission and prosecutors who are undertaking the daunting task of identifying cases in which bite mark evidence may have led to a wrongful conviction. There is no central database of cases in which such testimony was crucial to a conviction, which has left the commission and prosecutors to search through legal cases on the web and comb through reams of trial transcripts.
It is unclear just how many cases involving bite mark evidence are pending in Texas courts. It is most commonly used in cases related to child abuse. A recommendation from the commission does not require judges to prohibit such evidence, but it is likely to influence their decisions about whether bite marks should be allowed as evidence.
Motion to vacate the conviction of Eddie Lee Howard, Jr. Motion to vacate the conviction of Eddie Lee Howard, Jr.
The motion seeks to vacate the death sentence for a man who was convicted in Mississippi on the basis of bite mark testimony. (The motion is a PDF so is posted on Moodle under "Class 8.")
Feel free to skim the facts (and note there are facts about sexual assault). Be sure to read (a) the Table of Contents at page 2 to understand the structure of the argument; (b) the portion of the legal argument on pages 42-63; and (c) the conclusion on page 75.
Memorandum from Robert Ferarri, Manhattan District Attorney’s Office, to Justice Maxwell Wiley re. People v. Dean (Jan. 8, 2016) Memorandum from Robert Ferarri, Manhattan District Attorney’s Office, to Justice Maxwell Wiley re. People v. Dean (Jan. 8, 2016)
One of the cases mentioned in The Washington Post’s series on bite mark evidence is the New York case of People v. Dean. This memo is from the District Attorney’s office to the judge in that case. (Because this is a PDF document it is posted on Moodle under "Class 8."
Writing Reflection #9 Writing Reflection #9
Please go to our Moodle Page and under "Class 9" you will find the prompt and submission folder for Writing Reflection #9.
3.2 Class 10: The Alarming Present - Facial Recognition 3.2 Class 10: The Alarming Present - Facial Recognition
Garbage In, Garbage Out | Face Recognition on Flawed Data
Clare Garvie, Georgetown Law Center on Privacy & Technology (May 16, 2019)
This article looks at the use of facial recognition data by the NYPD
The Racist History Behind Facial Recognition The Racist History Behind Facial Recognition
Sahil Chinoy, The New York Times (July 10, 2019)
Use the link here to access to the online article with photos - the visuals add a lot to the article. The text of the article is below.
This article describes how facial recognition technology "echo the same biological essentialism behind physiognomy" - that is, the "science" we examined on the first day of class when we cosnidered the "evidence" that the shape of a person's head and features could provide information about their character.
Researchers recently learned that Immigration and Customs Enforcement used facial recognition on millions of driver’s license photographs without the license-holders’ knowledge, the latest revelation about governments employing the technology in ways that threaten civil liberties.
But the surveillance potential of facial recognition — its ability to create a “perpetual lineup” — isn’t the only cause for concern. The technological frontiers being explored by questionable researchers and unscrupulous start-ups recall the discredited pseudosciences of physiognomy and phrenology, which purport to use facial structure and head shape to assess character and mental capacity.
Artificial intelligence and modern computing are giving new life and a veneer of objectivity to these debunked theories, which were once used to legitimize slavery and perpetuate Nazi race “science.” Those who wish to spread essentialist theories of racial hierarchy are paying attention. In one blog, for example, a contemporary white nationalist claimed that “physiognomy is real” and “needs to come back as a legitimate field of scientific inquiry.”
More broadly, new applications of facial recognition — not just in academic research, but also in commercial products that try to guess emotions from facial expressions — echo the same biological essentialism behind physiognomy. Apparently, we still haven’t learned that faces do not contain some deeper truth about the people they belong to.
Composite photographs, new and old
One of the pioneers of 19th-century facial analysis, Francis Galton, was a prominent British eugenicist. He superimposed images of men convicted of crimes, attempting to find through “pictorial statistics” the essence of the criminal face.
Galton was disappointed with the results: He was unable to discern a criminal “type” from his composite photographs. This is because physiognomy is junk science — criminality is written neither in one’s genes nor on one’s face. He also tried to use composite portraits to determine the ideal “type” of each race, and his research was cited by Hans F.K. Günther, a Nazi eugenicist who wrote a book that was required reading in German schools during the Third Reich.
Galton’s tools and ideas have proved surprisingly durable, and modern researchers are again contemplating whether criminality can be read from one’s face. In a much-contested 2016 paper, researchers at a Chinese university claimed they had trained an algorithm to distinguish criminal from noncriminal portraits, and that “lip curvature, eye inner corner distance, and the so-called nose-mouth angle” could help tell them apart. The paper includes “average faces” of criminals and noncriminals reminiscent of Galton’s composite portraits.
The paper echoes many of the fallacies in Galton’s research: that people convicted of crimes are representative of those who commit them (the justice system exhibits profound bias), that the concept of inborn “criminality” is sound (life circumstances drastically shape one’s likelihood of committing a crime) and that facial appearance is a reliable predictor of character.
It’s true that humans tend to agree on what a threatening face looks like. But Alexander Todorov, a psychologist at Princeton, writes in his book “Face Value” that the relationship between a face and our sense that it is threatening (or friendly) is “between appearance and impressions, not between appearance and character.” The temptation to think we can read something deeper from these visual stereotypes is misguided — but persistent.
In 2017, the Stanford professor Michal Kosinski was an author of a study claiming to have invented an A.I. “gaydar” that could, when presented with pictures of gay and straight men, determine which ones were gay with 81 percent accuracy. (He told The Guardian that facial recognition might be used in the future to predict I.Q. as well.)
The paper speculates about whether differences in facial structure between gay and straight men might result from underexposure to male hormones, but neglects a simpler explanation, wrote Blaise Agüera y Arcas and Margaret Mitchell, A.I. researchers at Google, and Dr. Todorov in a Medium article. The research relied on images from dating websites. It’s likely that gay and straight people present themselves differently on these sites, from hairstyle to the degree they are tanned to the angle they take their selfies, the critics said. But the paper focuses on ideas reminiscent of the discredited theory of sexual inversion, which maintains that homosexuality is an inborn “reversal” of gender characteristics — gay men with female qualities, for example.
“Using scientific language and measurement doesn’t prevent a researcher from conducting flawed experiments and drawing wrong conclusions — especially when they confirm preconceptions,” the critics wrote in another post.
Echoes of the past
Parallels between the modern technology and historical applications abound. A 1902 phrenology book showed how to distinguish a “genuine husband” from an “unreliable” one based on the shape of his head; today, an Israeli start-up called Faception uses machine learning to score facial images using personality types like “academic researcher,” “brand promoter,” “terrorist” and “pedophile.”
Faception’s marketing materials are almost comical in their reduction of personalities to eight stereotypes, but the company appears to have customers, indicating an interest in “legitimizing this type of A.I. system,” said Clare Garvie, a facial recognition researcher at Georgetown Law.
“In some ways, they’re laughable,” she said. “In other ways, the very part that makes them laughable is what makes them so concerning.”
In the early 20th century, Katherine M.H. Blackford advocated using physical appearance to select among job applicants. She favored analyzing photographs over interviews to reveal character, Dr. Todorov writes. Today, the company HireVue sells technology that uses A.I. to analyze videos of job applicants; the platform scores them on measures like “personal stability” and “conscientiousness and responsibility.”
Cesare Lombroso, a prominent 19th-century Italian physiognomist, proposed separating children that he judged to be intellectually inferior, based on face and body measurements, from their “better-endowed companions.” Today, facial recognition programs are being piloted at American universities and Chinese schools to monitor students’ emotions and engagement. This is problematic for myriad reasons: Studies have shown no correlation between student engagement and actual learning, and teachers are more likely to see black students’ faces as angry, bias that might creep into an automated system.
Classification and surveillance
The similarities between modern, A.I.-driven facial analysis and its earlier, analog iteration are eerie. Both, for example, originated as attempts to track criminals and security targets.
Alphonse Bertillon, a French policeman and facial analysis pioneer, wanted to identify repeat offenders. He invented the mug shot and noted specific body measurements like head length on his “Bertillon cards.” With records of more than 100,000 prisoners collected between 1883 and 1893, he identified 4,564 recidivists.
Bertillon’s classification scheme was superseded by a more efficient fingerprinting system, but the basic idea — using bodily measurements to identify people in the service of an intelligence apparatus — was reborn with modern facial recognition. Progress in computer-driven facial recognition has been spurred by military investment and government competitions. (One C.I.A. director’s interest in the technology grew from a James Bond movie — he asked his staff to investigate facial recognition after seeing it used in the 1985 film “A View to Kill.”)
Early facial recognition software developed in the 1960s was like a computer-assisted version of Bertillon’s system, requiring researchers to manually identify points like the center of a subject’s eye (at a rate of about 40 images per hour). By the late 1990s, algorithms could automatically map facial features — and supercharged by computers, they could scan videos in real time.
Many of these algorithms are trained on people who did not or could not consent to their faces being used. I.B.M. took public photos from Flickr to feed facial recognition programs. The National Institute of Standards and Technology, a government agency, hosts a database of mug shots and images of people who have died. “Haunted data persists today,” said Joy Buolamwini, an M.I.T. researcher, in an email.
Emotional “intelligence”
Facial analysis services are commercially available from providers like Amazon and Microsoft. Anyone can use them at a nominal price — Amazon charges one-tenth of a cent to process a picture — to guess a person’s identity, gender, age and emotional state. Other platforms like Face++ guess race, too.
But these algorithms have documented problems with nonwhite, nonmale faces. And the idea that A.I. can detect the presence of emotions — most commonly happiness, sadness, anger, disgust and surprise — is especially fraught. Customers have used “affect recognition” for everything from measuring how people react to ads to helping children with autism develop social and emotional skills, but a report from the A.I. Now Institute argues that the technology is being “applied in unethical and irresponsible ways.”
Affect recognition draws from the work of Paul Ekman, a modern psychologist who argued that facial expressions are an objective way to determine someone’s inner emotional state, and that there exists a limited set of basic emotional categories that are fixed across cultures. His work suggests that we can’t help revealing these emotions. That theory inspired the television show “Lie to Me,” about a scientist who helps law enforcement by interpreting unforthcoming suspects’ expressions.
Dr. Ekman’s work has been criticized by scholars who say emotions cannot be reduced to such easily interpretable — and computationally convenient — categories. Algorithms that use these simplistic categories are “likely to reproduce the errors of an outdated scientific paradigm,” according to the A.I. Now report.
[If you’re online — and, well, you are — chances are someone is using your information. We’ll tell you what you can do about it. Sign up for our limited-run newsletter.]
Moreover, it is not hard to stretch from interpreting the results of facial analysis as “how happy this face appears” to the simpler but inaccurate “how happy this person feels” or even “how happy this person really is, despite his efforts to mask his emotions.” As the A.I. Now report says, affect recognition “raises troubling ethical questions about locating the arbiter of someone’s ‘real’ character and emotions outside of the individual.”
We’ve been here before. Much like the 19th-century technologies of photography and composite portraits lent “objectivity” to pseudoscientific physiognomy, today, computers and artificial intelligence supposedly distance facial analysis from human judgment and prejudice. In reality, algorithms that rely on a flawed understanding of expressions and emotions can just make prejudice more difficult to spot.
In his book, Dr. Todorov discusses the German physicist Georg Christoph Lichtenberg, an 18th-century skeptic of physiognomy who thought that the practice “simply licensed our natural impulses to form impressions from appearance.”
If physiognomy gained traction, “one will hang children before they have done the deeds that merit the gallows,” Lichtenberg wrote, warning of a “physiognomic auto-da-fé.”
As facial recognition technology develops, we would be wise to heed his words.
NIST Study Evaluates Effects of Race, Age, Sex on Face Recognition Software | NIST
December 19, 2019
This article summarizes a study (linked in the article) evaluating face recognition software.
The Hidden Role of Facial Recognition Tech in Many Arrests
Khari Johnson, Wired (March 7, 2022)
This article describes (as indicated by the title) the role of facial recognition in arrests - and the NYPD's policy on using facial recognition to make arrests.
Challenging Facial Recognition Software in Criminal Court Challenging Facial Recognition Software in Criminal Court
Kaitlin Jackson, The Champion: Magazine of the National Association of Criminal Defense Lawyers (July 2019)
This article explains why it is difficult to challenge facial recognition technology in court and offers some strategies for fighting it. Beacuse it is a pdf document it is posted on Moodle under "Class 9."
Writing Reflection #10 Writing Reflection #10
Please go to our Moodle Page and under "Class 10" you will find the prompt and submission folder for Writing Reflection #10.
3.2.1 Additional Facial Recognition and Digital Evidence Resources 3.2.1 Additional Facial Recognition and Digital Evidence Resources
Digital Evidence at CSAFE
Amazon’s Face Recognition Falsely Matched 28 Members of Congress With Mugshots
Jacob Snow, American Civil Liberties Union (July 26, 2018)
Racial Discrimination in Face Recognition Technology
Alex Najibi, Harvard University Science Policy Blog (Oct. 24, 2020)
How is Face Recognition Surveillance Technology Racist?
Kade Crockford, News & Commentary | American Civil Liberties Union
Facial Recognition Laws in the United States (List of bans)
#ProjectPanoptic
Proposed federal law: S.2052 - Facial Recognition and Biometric Technology Moratorium Act of 2021
Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification
Joy Buolamwini & Timnit Gebru, Proceedings of Machine Learning Research 81:1–15, 2018
3.3 Class 11: The Clear Illustration - Arson 3.3 Class 11: The Clear Illustration - Arson
"Trial by Fire: Did Texas Execute an Innocent Man?” "Trial by Fire: Did Texas Execute an Innocent Man?”
David Grann, The New Yorker (Sept. 7, 2009)
This is an investigation of the case of Cameron Todd Willingham, who was executed in 2004 for setting a fire that killed his three children.
Trial by Fire
Did Texas execute an innocent man?
By David Grann
August 31, 2009
Cameron Todd Willingham in his cell on death row, in 1994. He insisted upon his innocence in the deaths of his children and refused an offer to plead guilty in return for a life sentence. Photograph By Ken Light
The fire moved quickly through the house, a one-story wood-frame structure in a working-class neighborhood of Corsicana, in northeast Texas. Flames spread along the walls, bursting through doorways, blistering paint and tiles and furniture. Smoke pressed against the ceiling, then banked downward, seeping into each room and through crevices in the windows, staining the morning sky.
Buffie Barbee, who was eleven years old and lived two houses down, was playing in her back yard when she smelled the smoke. She ran inside and told her mother, Diane, and they hurried up the street; that’s when they saw the smoldering house and Cameron Todd Willingham standing on the front porch, wearing only a pair of jeans, his chest blackened with soot, his hair and eyelids singed. He was screaming, “My babies are burning up!” His children—Karmon and Kameron, who were one-year-old twin girls, and two-year-old Amber—were trapped inside.
Willingham told the Barbees to call the Fire Department, and while Diane raced down the street to get help he found a stick and broke the children’s bedroom window. Fire lashed through the hole. He broke another window; flames burst through it, too, and he retreated into the yard, kneeling in front of the house. A neighbor later told police that Willingham intermittently cried, “My babies!” then fell silent, as if he had “blocked the fire out of his mind.”
Diane Barbee, returning to the scene, could feel intense heat radiating off the house. Moments later, the five windows of the children’s room exploded and flames “blew out,” as Barbee put it. Within minutes, the first firemen had arrived, and Willingham approached them, shouting that his children were in their bedroom, where the flames were thickest. A fireman sent word over his radio for rescue teams to “step on it.”
More men showed up, uncoiling hoses and aiming water at the blaze. One fireman, who had an air tank strapped to his back and a mask covering his face, slipped through a window but was hit by water from a hose and had to retreat. He then charged through the front door, into a swirl of smoke and fire. Heading down the main corridor, he reached the kitchen, where he saw a refrigerator blocking the back door.
Todd Willingham, looking on, appeared to grow more hysterical, and a police chaplain named George Monaghan led him to the back of a fire truck and tried to calm him down. Willingham explained that his wife, Stacy, had gone out earlier that morning, and that he had been jolted from sleep by Amber screaming, “Daddy! Daddy!”
“My little girl was trying to wake me up and tell me about the fire,” he said, adding, “I couldn’t get my babies out.”
While he was talking, a fireman emerged from the house, cradling Amber. As she was given C.P.R., Willingham, who was twenty-three years old and powerfully built, ran to see her, then suddenly headed toward the babies’ room. Monaghan and another man restrained him. “We had to wrestle with him and then handcuff him, for his and our protection,” Monaghan later told police. “I received a black eye.” One of the first firemen at the scene told investigators that, at an earlier point, he had also held Willingham back. “Based on what I saw on how the fire was burning, it would have been crazy for anyone to try and go into the house,” he said.
Willingham was taken to a hospital, where he was told that Amber—who had actually been found in the master bedroom—had died of smoke inhalation. Kameron and Karmon had been lying on the floor of the children’s bedroom, their bodies severely burned. According to the medical examiner, they, too, died from smoke inhalation.
News of the tragedy, which took place on December 23, 1991, spread through Corsicana. A small city fifty-five miles northeast of Waco, it had once been the center of Texas’s first oil boom, but many of the wells had since dried up, and more than a quarter of the city’s twenty thousand inhabitants had fallen into poverty. Several stores along the main street were shuttered, giving the place the feel of an abandoned outpost.
Willingham and his wife, who was twenty-two years old, had virtually no money. Stacy worked in her brother’s bar, called Some Other Place, and Willingham, an unemployed auto mechanic, had been caring for the kids. The community took up a collection to help the Willinghams pay for funeral arrangements.
Fire investigators, meanwhile, tried to determine the cause of the blaze. (Willingham gave authorities permission to search the house: “I know we might not ever know all the answers, but I’d just like to know why my babies were taken from me.”) Douglas Fogg, who was then the assistant fire chief in Corsicana, conducted the initial inspection. He was tall, with a crew cut, and his voice was raspy from years of inhaling smoke from fires and cigarettes. He had grown up in Corsicana and, after graduating from high school, in 1963, he had joined the Navy, serving as a medic in Vietnam, where he was wounded on four occasions. He was awarded a Purple Heart each time. After he returned from Vietnam, he became a firefighter, and by the time of the Willingham blaze he had been battling fire—or what he calls “the beast”—for more than twenty years, and had become a certified arson investigator. “You learn that fire talks to you,” he told me.
He was soon joined on the case by one of the state’s leading arson sleuths, a deputy fire marshal named Manuel Vasquez, who has since died. Short, with a paunch, Vasquez had investigated more than twelve hundred fires. Arson investigators have always been considered a special breed of detective. In the 1991 movie “Backdraft,” a heroic arson investigator says of fire, “It breathes, it eats, and it hates. The only way to beat it is to think like it. To know that this flame will spread this way across the door and up across the ceiling.” Vasquez, who had previously worked in Army intelligence, had several maxims of his own. One was “Fire does not destroy evidence—it creates it.” Another was “The fire tells the story. I am just the interpreter.” He cultivated a Sherlock Holmes-like aura of invincibility. Once, he was asked under oath whether he had ever been mistaken in a case. “If I have, sir, I don’t know,” he responded. “It’s never been pointed out.”
Vasquez and Fogg visited the Willinghams’ house four days after the blaze. Following protocol, they moved from the least burned areas toward the most damaged ones. “It is a systematic method,” Vasquez later testified, adding, “I’m just collecting information. . . . I have not made any determination. I don’t have any preconceived idea.”
The men slowly toured the perimeter of the house, taking notes and photographs, like archeologists mapping out a ruin. Upon opening the back door, Vasquez observed that there was just enough space to squeeze past the refrigerator blocking the exit. The air smelled of burned rubber and melted wires; a damp ash covered the ground, sticking to their boots. In the kitchen, Vasquez and Fogg discerned only smoke and heat damage—a sign that the fire had not originated there—and so they pushed deeper into the nine-hundred-and-seventy-five-square-foot building. A central corridor led past a utility room and the master bedroom, then past a small living room, on the left, and the children’s bedroom, on the right, ending at the front door, which opened onto the porch. Vasquez tried to take in everything, a process that he compared to entering one’s mother-in-law’s house for the first time: “I have the same curiosity.”
In the utility room, he noticed on the wall pictures of skulls and what he later described as an image of “the Grim Reaper.” Then he turned into the master bedroom, where Amber’s body had been found. Most of the damage there was also from smoke and heat, suggesting that the fire had started farther down the hallway, and he headed that way, stepping over debris and ducking under insulation and wiring that hung down from the exposed ceiling.
As he and Fogg removed some of the clutter, they noticed deep charring along the base of the walls. Because gases become buoyant when heated, flames ordinarily burn upward. But Vasquez and Fogg observed that the fire had burned extremely low down, and that there were peculiar char patterns on the floor, shaped like puddles.
Vasquez’s mood darkened. He followed the “burn trailer”—the path etched by the fire—which led from the hallway into the children’s bedroom. Sunlight filtering through the broken windows illuminated more of the irregularly shaped char patterns. A flammable or combustible liquid doused on a floor will cause a fire to concentrate in these kinds of pockets, which is why investigators refer to them as “pour patterns” or “puddle configurations.”
The fire had burned through layers of carpeting and tile and plywood flooring. Moreover, the metal springs under the children’s beds had turned white—a sign that intense heat had radiated beneath them. Seeing that the floor had some of the deepest burns, Vasquez deduced that it had been hotter than the ceiling, which, given that heat rises, was, in his words, “not normal.”
Fogg examined a piece of glass from one of the broken windows. It contained a spiderweb-like pattern—what fire investigators call “crazed glass.” Forensic textbooks had long described the effect as a key indicator that a fire had burned “fast and hot,” meaning that it had been fuelled by a liquid accelerant, causing the glass to fracture.
The men looked again at what appeared to be a distinct burn trailer through the house: it went from the children’s bedroom into the corridor, then turned sharply to the right and proceeded out the front door. To the investigators’ surprise, even the wood under the door’s aluminum threshold was charred. On the concrete floor of the porch, just outside the front door, Vasquez and Fogg noticed another unusual thing: brown stains, which, they reported, were consistent with the presence of an accelerant.
The men scanned the walls for soot marks that resembled a “V.” When an object catches on fire, it creates such a pattern, as heat and smoke radiate outward; the bottom of the “V” can therefore point to where a fire began. In the Willingham house, there was a distinct “V” in the main corridor. Examining it and other burn patterns, Vasquez identified three places where fire had originated: in the hallway, in the children’s bedroom, and at the front door. Vasquez later testified that multiple origins pointed to one conclusion: the fire was “intentionally set by human hands.”
By now, both investigators had a clear vision of what had happened. Someone had poured liquid accelerant throughout the children’s room, even under their beds, then poured some more along the adjoining hallway and out the front door, creating a “fire barrier” that prevented anyone from escaping; similarly, a prosecutor later suggested, the refrigerator in the kitchen had been moved to block the back-door exit. The house, in short, had been deliberately transformed into a death trap.
The investigators collected samples of burned materials from the house and sent them to a laboratory that could detect the presence of a liquid accelerant. The lab’s chemist reported that one of the samples contained evidence of “mineral spirits,” a substance that is often found in charcoal-lighter fluid. The sample had been taken by the threshold of the front door.
The fire was now considered a triple homicide, and Todd Willingham—the only person, besides the victims, known to have been in the house at the time of the blaze—became the prime suspect.
Police and fire investigators canvassed the neighborhood, interviewing witnesses. Several, like Father Monaghan, initially portrayed Willingham as devastated by the fire. Yet, over time, an increasing number of witnesses offered damning statements. Diane Barbee said that she had not seen Willingham try to enter the house until after the authorities arrived, as if he were putting on a show. And when the children’s room exploded with flames, she added, he seemed more preoccupied with his car, which he moved down the driveway. Another neighbor reported that when Willingham cried out for his babies he “did not appear to be excited or concerned.” Even Father Monaghan wrote in a statement that, upon further reflection, “things were not as they seemed. I had the feeling that [Willingham] was in complete control.”
The police began to piece together a disturbing profile of Willingham. Born in Ardmore, Oklahoma, in 1968, he had been abandoned by his mother when he was a baby. His father, Gene, who had divorced his mother, eventually raised him with his stepmother, Eugenia. Gene, a former U.S. marine, worked in a salvage yard, and the family lived in a cramped house; at night, they could hear freight trains rattling past on a nearby track. Willingham, who had what the family called the “classic Willingham look”—a handsome face, thick black hair, and dark eyes—struggled in school, and as a teen-ager began to sniff paint. When he was seventeen, Oklahoma’s Department of Human Services evaluated him, and reported, “He likes ‘girls,’ music, fast cars, sharp trucks, swimming, and hunting, in that order.” Willingham dropped out of high school, and over time was arrested for, among other things, driving under the influence, stealing a bicycle, and shoplifting.
In 1988, he met Stacy, a senior in high school, who also came from a troubled background: when she was four years old, her stepfather had strangled her mother to death during a fight. Stacy and Willingham had a turbulent relationship. Willingham, who was unfaithful, drank too much Jack Daniel’s, and sometimes hit Stacy—even when she was pregnant. A neighbor said that he once heard Willingham yell at her, “Get up, bitch, and I’ll hit you again.”
On December 31st, the authorities brought Willingham in for questioning. Fogg and Vasquez were present for the interrogation, along with Jimmie Hensley, a police officer who was working his first arson case. Willingham said that Stacy had left the house around 9 a.m. to pick up a Christmas present for the kids, at the Salvation Army. “After she got out of the driveway, I heard the twins cry, so I got up and gave them a bottle,” he said. The children’s room had a safety gate across the doorway, which Amber could climb over but not the twins, and he and Stacy often let the twins nap on the floor after they drank their bottles. Amber was still in bed, Willingham said, so he went back into his room to sleep. “The next thing I remember is hearing ‘Daddy, Daddy,’ ” he recalled. “The house was already full of smoke.” He said that he got up, felt around the floor for a pair of pants, and put them on. He could no longer hear his daughter’s voice (“I heard that last ‘Daddy, Daddy’ and never heard her again”), and he hollered, “Oh God— Amber, get out of the house! Get out of the house!’ ”
He never sensed that Amber was in his room, he said. Perhaps she had already passed out by the time he stood up, or perhaps she came in after he left, through a second doorway, from the living room. He said that he went down the corridor and tried to reach the children’s bedroom. In the hallway, he said, “you couldn’t see nothing but black.” The air smelled the way it had when their microwave had blown up, three weeks earlier—like “wire and stuff like that.” He could hear sockets and light switches popping, and he crouched down, almost crawling. When he made it to the children’s bedroom, he said, he stood and his hair caught on fire. “Oh God, I never felt anything that hot before,” he said of the heat radiating out of the room.
After he patted out the fire on his hair, he said, he got down on the ground and groped in the dark. “I thought I found one of them once,” he said, “but it was a doll.” He couldn’t bear the heat any longer. “I felt myself passing out,” he said. Finally, he stumbled down the corridor and out the front door, trying to catch his breath. He saw Diane Barbee and yelled for her to call the Fire Department. After she left, he insisted, he tried without success to get back inside.
The investigators asked him if he had any idea how the fire had started. He said that he wasn’t sure, though it must have originated in the children’s room, since that was where he first saw flames; they were glowing like “bright lights.” He and Stacy used three space heaters to keep the house warm, and one of them was in the children’s room. “I taught Amber not to play with it,” he said, adding that she got “whuppings every once in a while for messing with it.” He said that he didn’t know if the heater, which had an internal flame, was turned on. (Vasquez later testified that when he had checked the heater, four days after the fire, it was in the “Off” position.) Willingham speculated that the fire might have been started by something electrical: he had heard all that popping and crackling.
When pressed whether someone might have a motive to hurt his family, he said that he couldn’t think of anyone that “cold-blooded.” He said of his children, “I just don’t understand why anybody would take them, you know? We had three of the most pretty babies anybody could have ever asked for.” He went on, “Me and Stacy’s been together for four years, but off and on we get into a fight and split up for a while and I think those babies is what brought us so close together . . . neither one of us . . . could live without them kids.” Thinking of Amber, he said, “To tell you the honest-to-God’s truth, I wish she hadn’t woke me up.”
During the interrogation, Vasquez let Fogg take the lead. Finally, Vasquez turned to Willingham and asked a seemingly random question: had he put on shoes before he fled the house?
“No, sir,” Willingham replied.
A map of the house was on a table between the men, and Vasquez pointed to it. “You walked out this way?” he said.
Willingham said yes.
Vasquez was now convinced that Willingham had killed his children. If the floor had been soaked with a liquid accelerant and the fire had burned low, as the evidence suggested, Willingham could not have run out of the house the way he had described without badly burning his feet. A medical report indicated that his feet had been unscathed.
Willingham insisted that, when he left the house, the fire was still around the top of the walls and not on the floor. “I didn’t have to jump through any flames,” he said. Vasquez believed that this was impossible, and that Willingham had lit the fire as he was retreating—first, torching the children’s room, then the hallway, and then, from the porch, the front door. Vasquez later said of Willingham, “He told me a story of pure fabrication. . . . He just talked and he talked and all he did was lie.”
Still, there was no clear motive. The children had life-insurance policies, but they amounted to only fifteen thousand dollars, and Stacy’s grandfather, who had paid for them, was listed as the primary beneficiary. Stacy told investigators that even though Willingham hit her he had never abused the children—“Our kids were spoiled rotten,” she said—and she did not believe that Willingham could have killed them.
Ultimately, the authorities concluded that Willingham was a man without a conscience whose serial crimes had climaxed, almost inexorably, in murder. John Jackson, who was then the assistant district attorney in Corsicana, was assigned to prosecute Willingham’s case. He later told the Dallas Morning News that he considered Willingham to be “an utterly sociopathic individual” who deemed his children “an impediment to his lifestyle.” Or, as the local district attorney, Pat Batchelor, put it, “The children were interfering with his beer drinking and dart throwing.”
On the night of January 8, 1992, two weeks after the fire, Willingham was riding in a car with Stacy when swat teams surrounded them, forcing them to the side of the road. “They pulled guns out like we had just robbed ten banks,” Stacy later recalled. “All we heard was ‘click, click.’ . . . Then they arrested him.”
Willingham was charged with murder. Because there were multiple victims, he was eligible for the death penalty, under Texas law. Unlike many other prosecutors in the state, Jackson, who had ambitions of becoming a judge, was personally opposed to capital punishment. “I don’t think it’s effective in deterring criminals,” he told me. “I just don’t think it works.” He also considered it wasteful: because of the expense of litigation and the appeals process, it costs, on average, $2.3 million to execute a prisoner in Texas—about three times the cost of incarcerating someone for forty years. Plus, Jackson said, “What’s the recourse if you make a mistake?” Yet his boss, Batchelor, believed that, as he once put it, “certain people who commit bad enough crimes give up the right to live,” and Jackson came to agree that the heinous nature of the crime in the Willingham case—“one of the worst in terms of body count” that he had ever tried—mandated death.
Willingham couldn’t afford to hire lawyers, and was assigned two by the state: David Martin, a former state trooper, and Robert Dunn, a local defense attorney who represented everyone from alleged murderers to spouses in divorce cases—a “Jack-of-all-trades,” as he calls himself. (“In a small town, you can’t say ‘I’m a so-and-so lawyer,’ because you’ll starve to death,” he told me.)
The Willingham family in the days before Christmas, 1991.
Not long after Willingham’s arrest, authorities received a message from a prison inmate named Johnny Webb, who was in the same jail as Willingham. Webb alleged that Willingham had confessed to him that he took “some kind of lighter fluid, squirting [it] around the walls and the floor, and set a fire.” The case against Willingham was considered airtight.
Even so, several of Stacy’s relatives—who, unlike her, believed that Willingham was guilty—told Jackson that they preferred to avoid the anguish of a trial. And so, shortly before jury selection, Jackson approached Willingham’s attorneys with an extraordinary offer: if their client pleaded guilty, the state would give him a life sentence. “I was really happy when I thought we might have a deal to avoid the death penalty,” Jackson recalls.
Willingham’s lawyers were equally pleased. They had little doubt that he had committed the murders and that, if the case went before a jury, he would be found guilty, and, subsequently, executed. “Everyone thinks defense lawyers must believe their clients are innocent, but that’s seldom true,” Martin told me. “Most of the time, they’re guilty as sin.” He added of Willingham, “All the evidence showed that he was one hundred per cent guilty. He poured accelerant all over the house and put lighter fluid under the kids’ beds.” It was, he said, “a classic arson case”: there were “puddle patterns all over the place—no disputing those.”
Martin and Dunn advised Willingham that he should accept the offer, but he refused. The lawyers asked his father and stepmother to speak to him. According to Eugenia, Martin showed them photographs of the burned children and said, “Look what your son did. You got to talk him into pleading, or he’s going to be executed.”
His parents went to see their son in jail. Though his father did not believe that he should plead guilty if he were innocent, his stepmother beseeched him to take the deal. “I just wanted to keep my boy alive,” she told me.
Willingham was implacable. “I ain’t gonna plead to something I didn’t do, especially killing my own kids,” he said. It was his final decision. Martin says, “I thought it was nuts at the time—and I think it’s nuts now.”
Willingham’s refusal to accept the deal confirmed the view of the prosecution, and even that of his defense lawyers, that he was an unrepentant killer.
In August, 1992, the trial commenced in the old stone courthouse in downtown Corsicana. Jackson and a team of prosecutors summoned a procession of witnesses, including Johnny Webb and the Barbees. The crux of the state’s case, though, remained the scientific evidence gathered by Vasquez and Fogg. On the stand, Vasquez detailed what he called more than “twenty indicators” of arson.
“Do you have an opinion as to who started the fire?” one of the prosecutors asked.
“Yes, sir,” Vasquez said. “Mr. Willingham.”
The prosecutor asked Vasquez what he thought Willingham’s intent was in lighting the fire. “To kill the little girls,” he said.
The defense had tried to find a fire expert to counter Vasquez and Fogg’s testimony, but the one they contacted concurred with the prosecution. Ultimately, the defense presented only one witness to the jury: the Willinghams’ babysitter, who said she could not believe that Willingham could have killed his children. (Dunn told me that Willingham had wanted to testify, but Martin and Dunn thought that he would make a bad witness.) The trial ended after two days.
During his closing arguments, Jackson said that the puddle configurations and pour patterns were Willingham’s inadvertent “confession,” burned into the floor. Showing a Bible that had been salvaged from the fire, Jackson paraphrased the words of Jesus from the Gospel of Matthew: “Whomsoever shall harm one of my children, it’s better for a millstone to be hung around his neck and for him to be cast in the sea.”
The jury was out for barely an hour before returning with a unanimous guilty verdict. As Vasquez put it, “The fire does not lie.”
II
When Elizabeth Gilbert approached the prison guard, on a spring day in 1999, and said Cameron Todd Willingham’s name, she was uncertain about what she was doing. A forty-seven-year-old French teacher and playwright from Houston, Gilbert was divorced with two children. She had never visited a prison before. Several weeks earlier, a friend, who worked at an organization that opposed the death penalty, had encouraged her to volunteer as a pen pal for an inmate on death row, and Gilbert had offered her name and address. Not long after, a short letter, written with unsteady penmanship, arrived from Willingham. “If you wish to write back, I would be honored to correspond with you,” he said. He also asked if she might visit him. Perhaps out of a writer’s curiosity, or perhaps because she didn’t feel quite herself (she had just been upset by news that her ex-husband was dying of cancer), she agreed. Now she was standing in front of the decrepit penitentiary in Huntsville, Texas—a place that inmates referred to as “the death pit.”
She filed past a razor-wire fence, a series of floodlights, and a checkpoint, where she was patted down, until she entered a small chamber. Only a few feet in front of her was a man convicted of multiple infanticide. He was wearing a white jumpsuit with “DR”—for death row—printed on the back, in large black letters. He had a tattoo of a serpent and a skull on his left biceps. He stood nearly six feet tall and was muscular, though his legs had atrophied after years of confinement.
A Plexiglas window separated Willingham from her; still, Gilbert, who had short brown hair and a bookish manner, stared at him uneasily. Willingham had once fought another prisoner who called him a “baby killer,” and since he had been incarcerated, seven years earlier, he had committed a series of disciplinary infractions that had periodically landed him in the segregation unit, which was known as “the dungeon.”
Willingham greeted her politely. He seemed grateful that she had come. After his conviction, Stacy had campaigned for his release. She wrote to Ann Richards, then the governor of Texas, saying, “I know him in ways that no one else does when it comes to our children. Therefore, I believe that there is no way he could have possibly committed this crime.” But within a year Stacy had filed for divorce, and Willingham had few visitors except for his parents, who drove from Oklahoma to see him once a month. “I really have no one outside my parents to remind me that I am a human being, not the animal the state professes I am,” he told Gilbert at one point.
He didn’t want to talk about death row. “Hell, I live here,” he later wrote her. “When I have a visit, I want to escape from here.” He asked her questions about her teaching and art. He expressed fear that, as a playwright, she might find him a “one-dimensional character,” and apologized for lacking social graces; he now had trouble separating the mores in prison from those of the outside world.
The aftermath of the fire on December 23, 1991. Photograph from Texas State Fire Marshal’s Office
Photograph from Texas State Fire Marshal’s Office
When Gilbert asked him if he wanted something to eat or drink from the vending machines, he declined. “I hope I did not offend you by not accepting any snacks,” he later wrote her. “I didn’t want you to feel I was there just for something like that.”
She had been warned that prisoners often tried to con visitors. He appeared to realize this, subsequently telling her, “I am just a simple man. Nothing else. And to most other people a convicted killer looking for someone to manipulate.”
Their visit lasted for two hours, and afterward they continued to correspond. She was struck by his letters, which seemed introspective, and were not at all what she had expected. “I am a very honest person with my feelings,” he wrote her. “I will not bullshit you on how I feel or what I think.” He said that he used to be stoic, like his father. But, he added, “losing my three daughters . . . my home, wife and my life, you tend to wake up a little. I have learned to open myself.”
She agreed to visit him again, and when she returned, several weeks later, he was visibly moved. “Here I am this person who nobody on the outside is ever going to know as a human, who has lost so much, but still trying to hold on,” he wrote her afterward. “But you came back! I don’t think you will ever know of what importance that visit was in my existence.”
They kept exchanging letters, and she began asking him about the fire. He insisted that he was innocent and that, if someone had poured accelerant through the house and lit it, then the killer remained free. Gilbert wasn’t naïve—she assumed that he was guilty. She did not mind giving him solace, but she was not there to absolve him.
Still, she had become curious about the case, and one day that fall she drove down to the courthouse in Corsicana to review the trial records. Many people in the community remembered the tragedy, and a clerk expressed bewilderment that anyone would be interested in a man who had burned his children alive.
Gilbert took the files and sat down at a small table. As she examined the eyewitness accounts, she noticed several contradictions. Diane Barbee had reported that, before the authorities arrived at the fire, Willingham never tried to get back into the house—yet she had been absent for some time while calling the Fire Department. Meanwhile, her daughter Buffie had reported witnessing Willingham on the porch breaking a window, in an apparent effort to reach his children. And the firemen and police on the scene had described Willingham frantically trying to get into the house.
The witnesses’ testimony also grew more damning after authorities had concluded, in the beginning of January, 1992, that Willingham was likely guilty of murder. In Diane Barbee’s initial statement to authorities, she had portrayed Willingham as “hysterical,” and described the front of the house exploding. But on January 4th, after arson investigators began suspecting Willingham of murder, Barbee suggested that he could have gone back inside to rescue his children, for at the outset she had seen only “smoke coming from out of the front of the house”—smoke that was not “real thick.”
An even starker shift occurred with Father Monaghan’s testimony. In his first statement, he had depicted Willingham as a devastated father who had to be repeatedly restrained from risking his life. Yet, as investigators were preparing to arrest Willingham, he concluded that Willingham had been too emotional (“He seemed to have the type of distress that a woman who had given birth would have upon seeing her children die”); and he expressed a “gut feeling” that Willingham had “something to do with the setting of the fire.”
Dozens of studies have shown that witnesses’ memories of events often change when they are supplied with new contextual information. Itiel Dror, a cognitive psychologist who has done extensive research on eyewitness and expert testimony in criminal investigations, told me, “The mind is not a passive machine. Once you believe in something—once you expect something—it changes the way you perceive information and the way your memory recalls it.”
After Gilbert’s visit to the courthouse, she kept wondering about Willingham’s motive, and she pressed him on the matter. In response, he wrote, of the death of his children, “I do not talk about it much anymore and it is still a very powerfully emotional pain inside my being.” He admitted that he had been a “sorry-ass husband” who had hit Stacy—something he deeply regretted. But he said that he had loved his children and would never have hurt them. Fatherhood, he said, had changed him; he stopped being a hoodlum and “settled down” and “became a man.” Nearly three months before the fire, he and Stacy, who had never married, wed at a small ceremony in his home town of Ardmore. He said that the prosecution had seized upon incidents from his past and from the day of the fire to create a portrait of a “demon,” as Jackson, the prosecutor, referred to him. For instance, Willingham said, he had moved the car during the fire simply because he didn’t want it to explode by the house, further threatening the children.
Gilbert was unsure what to make of his story, and she began to approach people who were involved in the case, asking them questions. “My friends thought I was crazy,” Gilbert recalls. “I’d never done anything like this in my life.”
One morning, when Willingham’s parents came to visit him, Gilbert arranged to see them first, at a coffee shop near the prison. Gene, who was in his seventies, had the Willingham look, though his black hair had gray streaks and his dark eyes were magnified by glasses. Eugenia, who was in her fifties, with silvery hair, was as sweet and talkative as her husband was stern and reserved. The drive from Oklahoma to Texas took six hours, and they had woken at three in the morning; because they could not afford a motel, they would have to return home later that day. “I feel like a real burden to them,” Willingham had written Gilbert.
As Gene and Eugenia sipped coffee, they told Gilbert how grateful they were that someone had finally taken an interest in Todd’s case. Gene said that his son, though he had flaws, was no killer.
The evening before the fire, Eugenia said, she had spoken on the phone with Todd. She and Gene were planning on visiting two days later, on Christmas Eve, and Todd told her that he and Stacy and the kids had just picked up family photographs. “He said, ‘We got your pictures for Christmas,’ ” she recalled. “He put Amber on the phone, and she was tattling on one of the twins. Todd didn’t seem upset. If something was bothering him, I would have known.”
The aftermath of the fire on December 23, 1991. Photograph from Texas State Fire Marshal’s Office
Photograph from Texas State Fire Marshal’s Office
Gene and Eugenia got up to go: they didn’t want to miss any of the four hours that were allotted for the visit with their son. Before they left, Gene said, “You’ll let us know if you find anything, won’t you?”
Over the next few weeks, Gilbert continued to track down sources. Many of them, including the Barbees, remained convinced that Willingham was guilty, but several of his friends and relatives had doubts. So did some people in law enforcement. Willingham’s former probation officer in Oklahoma, Polly Goodin, recently told me that Willingham had never demonstrated bizarre or sociopathic behavior. “He was probably one of my favorite kids,” she said. Even a former judge named Bebe Bridges—who had often stood, as she put it, on the “opposite side” of Willingham in the legal system, and who had sent him to jail for stealing—told me that she could not imagine him killing his children. “He was polite, and he seemed to care,” she said. “His convictions had been for dumb-kid stuff. Even the things stolen weren’t significant.” Several months before the fire, Willingham tracked Goodin down at her office, and proudly showed her photographs of Stacy and the kids. “He wanted Bebe and me to know he’d been doing good,” Goodin recalled.
Eventually, Gilbert returned to Corsicana to interview Stacy, who had agreed to meet at the bed-and-breakfast where Gilbert was staying. Stacy was slightly plump, with pale, round cheeks and feathered dark-blond hair; her bangs were held in place by gel, and her face was heavily made up. According to a tape recording of the conversation, Stacy said that nothing unusual had happened in the days before the fire. She and Willingham had not fought, and were preparing for the holiday. Though Vasquez, the arson expert, had recalled finding the space heater off, Stacy was sure that, at least on the day of the incident—a cool winter morning—it had been on. “I remember turning it down,” she recalled. “I always thought, Gosh, could Amber have put something in there?” Stacy added that, more than once, she had caught Amber “putting things too close to it.”
Willingham had often not treated her well, she recalled, and after his incarceration she had left him for a man who did. But she didn’t think that her former husband should be on death row. “I don’t think he did it,” she said, crying.
Though only the babysitter had appeared as a witness for the defense during the main trial, several family members, including Stacy, testified during the penalty phase, asking the jury to spare Willingham’s life. When Stacy was on the stand, Jackson grilled her about the “significance” of Willingham’s “very large tattoo of a skull, encircled by some kind of a serpent.”
“It’s just a tattoo,” Stacy responded.
“He just likes skulls and snakes. Is that what you’re saying?”
“No. He just had—he got a tattoo on him.”
The prosecution cited such evidence in asserting that Willingham fit the profile of a sociopath, and brought forth two medical experts to confirm the theory. Neither had met Willingham. One of them was Tim Gregory, a psychologist with a master’s degree in marriage and family issues, who had previously gone goose hunting with Jackson, and had not published any research in the field of sociopathic behavior. His practice was devoted to family counselling.
At one point, Jackson showed Gregory Exhibit No. 60—a photograph of an Iron Maiden poster that had hung in Willingham’s house—and asked the psychologist to interpret it. “This one is a picture of a skull, with a fist being punched through the skull,” Gregory said; the image displayed “violence” and “death.” Gregory looked at photographs of other music posters owned by Willingham. “There’s a hooded skull, with wings and a hatchet,” Gregory continued. “And all of these are in fire, depicting—it reminds me of something like Hell. And there’s a picture—a Led Zeppelin picture of a falling angel. . . . I see there’s an association many times with cultive-type of activities. A focus on death, dying. Many times individuals that have a lot of this type of art have interest in satanic-type activities.”
The other medical expert was James P. Grigson, a forensic psychiatrist. He testified so often for the prosecution in capital-punishment cases that he had become known as Dr. Death. (A Texas appellate judge once wrote that when Grigson appeared on the stand the defendant might as well “commence writing out his last will and testament.”) Grigson suggested that Willingham was an “extremely severe sociopath,” and that “no pill” or treatment could help him. Grigson had previously used nearly the same words in helping to secure a death sentence against Randall Dale Adams, who had been convicted of murdering a police officer, in 1977. After Adams, who had no prior criminal record, spent a dozen years on death row—and once came within seventy-two hours of being executed—new evidence emerged that absolved him, and he was released. In 1995, three years after Willingham’s trial, Grigson was expelled from the American Psychiatric Association for violating ethics. The association stated that Grigson had repeatedly arrived at a “psychiatric diagnosis without first having examined the individuals in question, and for indicating, while testifying in court as an expert witness, that he could predict with 100-per-cent certainty that the individuals would engage in future violent acts.”
After speaking to Stacy, Gilbert had one more person she wanted to interview: the jailhouse informant Johnny Webb, who was incarcerated in Iowa Park, Texas. She wrote to Webb, who said that she could see him, and they met in the prison visiting room. A man in his late twenties, he had pallid skin and a closely shaved head; his eyes were jumpy, and his entire body seemed to tremble. A reporter who once met him described him to me as “nervous as a cat around rocking chairs.” Webb had begun taking drugs when he was nine years old, and had been convicted of, among other things, car theft, selling marijuana, forgery, and robbery.
As Gilbert chatted with him, she thought that he seemed paranoid. During Willingham’s trial, Webb disclosed that he had been given a diagnosis of “post-traumatic stress disorder” after he was sexually assaulted in prison, in 1988, and that he often suffered from “mental impairment.” Under cross-examination, Webb testified that he had no recollection of a robbery that he had pleaded guilty to only months earlier.
Johnny Webb claimed that Willingham confessed to him in prison. Photograph by Alex Garcia / Chicago Tribune
Photograph by Alex Garcia / Chicago Tribune
Webb repeated for her what he had said in court: he had passed by Willingham’s cell, and as they spoke through a food slot Willingham broke down and told him that he intentionally set the house on fire. Gilbert was dubious. It was hard to believe that Willingham, who had otherwise insisted on his innocence, had suddenly confessed to an inmate he barely knew. The conversation had purportedly taken place by a speaker system that allowed any of the guards to listen—an unlikely spot for an inmate to reveal a secret. What’s more, Webb alleged that Willingham had told him that Stacy had hurt one of the kids, and that the fire was set to cover up the crime. The autopsies, however, had revealed no bruises or signs of trauma on the children’s bodies.
Jailhouse informants, many of whom are seeking reduced time or special privileges, are notoriously unreliable. According to a 2004 study by the Center on Wrongful Convictions, at Northwestern University Law School, lying police and jailhouse informants are the leading cause of wrongful convictions in capital cases in the United States. At the time that Webb came forward against Willingham, he was facing charges of robbery and forgery. During Willingham’s trial, another inmate planned to testify that he had overheard Webb saying to another prisoner that he was hoping to “get time cut,” but the testimony was ruled inadmissible, because it was hearsay. Webb, who pleaded guilty to the robbery and forgery charges, received a sentence of fifteen years. Jackson, the prosecutor, told me that he generally considered Webb “an unreliable kind of guy,” but added, “I saw no real motive for him to make a statement like this if it wasn’t true. We didn’t cut him any slack.” In 1997, five years after Willingham’s trial, Jackson urged the Texas Board of Pardons and Paroles to grant Webb parole. “I asked them to cut him loose early,” Jackson told me. The reason, Jackson said, was that Webb had been targeted by the Aryan Brotherhood. The board granted Webb parole, but within months of his release he was caught with cocaine and returned to prison.
In March, 2000, several months after Gilbert’s visit, Webb unexpectedly sent Jackson a Motion to Recant Testimony, declaring, “Mr. Willingham is innocent of all charges.” But Willingham’s lawyer was not informed of this development, and soon afterward Webb, without explanation, recanted his recantation. When I recently asked Webb, who was released from prison two years ago, about the turnabout and why Willingham would have confessed to a virtual stranger, he said that he knew only what “the dude told me.” After I pressed him, he said, “It’s very possible I misunderstood what he said.” Since the trial, Webb has been given an additional diagnosis, bipolar disorder. “Being locked up in that little cell makes you kind of crazy,” he said. “My memory is in bits and pieces. I was on a lot of medication at the time. Everyone knew that.” He paused, then said, “The statute of limitations has run out on perjury, hasn’t it?”
Aside from the scientific evidence of arson, the case against Willingham did not stand up to scrutiny. Jackson, the prosecutor, said of Webb’s testimony, “You can take it or leave it.” Even the refrigerator’s placement by the back door of the house turned out to be innocuous; there were two refrigerators in the cramped kitchen, and one of them was by the back door. Jimmie Hensley, the police detective, and Douglas Fogg, the assistant fire chief, both of whom investigated the fire, told me recently that they had never believed that the fridge was part of the arson plot. “It didn’t have nothing to do with the fire,” Fogg said.
After months of investigating the case, Gilbert found that her faith in the prosecution was shaken. As she told me, “What if Todd really was innocent?”
III
In the summer of 1660, an Englishman named William Harrison vanished on a walk, near the village of Charingworth, in Gloucestershire. His bloodstained hat was soon discovered on the side of a local road. Police interrogated Harrison’s servant, John Perry, and eventually Perry gave a statement that his mother and his brother had killed Harrison for money. Perry, his mother, and his brother were hanged.
Two years later, Harrison reappeared. He insisted, fancifully, that he had been abducted by a band of criminals and sold into slavery. Whatever happened, one thing was indisputable: he had not been murdered by the Perrys.
The fear that an innocent person might be executed has long haunted jurors and lawyers and judges. During America’s Colonial period, dozens of crimes were punishable by death, including horse thievery, blasphemy, “man-stealing,” and highway robbery. After independence, the number of crimes eligible for the death penalty was gradually reduced, but doubts persisted over whether legal procedures were sufficient to prevent an innocent person from being executed. In 1868, John Stuart Mill made one of the most eloquent defenses of capital punishment, arguing that executing a murderer did not display a wanton disregard for life but, rather, proof of its value. “We show, on the contrary, most emphatically our regard for it by the adoption of a rule that he who violates that right in another forfeits it for himself,” he said. For Mill, there was one counterargument that carried weight—“that if by an error of justice an innocent person is put to death, the mistake can never be corrected.”
The modern legal system, with its lengthy appeals process and clemency boards, was widely assumed to protect the kind of “error of justice” that Mill feared. In 2000, while George W. Bush was governor of Texas, he said, “I know there are some in the country who don’t care for the death penalty, but . . . we’ve adequately answered innocence or guilt.” His top policy adviser on issues of criminal justice emphasized that there is “super due process to make sure that no innocent defendants are executed.”
In recent years, though, questions have mounted over whether the system is fail-safe. Since 1976, more than a hundred and thirty people on death row have been exonerated. DNA testing, which was developed in the eighties, saved seventeen of them, but the technique can be used only in rare instances. Barry Scheck, a co-founder of the Innocence Project, which has used DNA testing to exonerate prisoners, estimates that about eighty per cent of felonies do not involve biological evidence.
In 2000, after thirteen people on death row in Illinois were exonerated, George Ryan, who was then governor of the state, suspended the death penalty. Though he had been a longtime advocate of capital punishment, he declared that he could no longer support a system that has “come so close to the ultimate nightmare—the state’s taking of innocent life.” Former Supreme Court Justice Sandra Day O’Connor has said that the “execution of a legally and factually innocent person would be a constitutionally intolerable event.”
Such a case has become a kind of grisly Holy Grail among opponents of capital punishment. In his 2002 book “The Death Penalty,” Stuart Banner observes, “The prospect of killing an innocent person seemed to be the one thing that could cause people to rethink their support for capital punishment. Some who were not troubled by statistical arguments against the death penalty—claims about deterrence or racial disparities—were deeply troubled that such an extreme injustice might occur in an individual case.” Opponents of the death penalty have pointed to several questionable cases. In 1993, Ruben Cantu was executed in Texas for fatally shooting a man during a robbery. Years later, a second victim, who survived the shooting, told the Houston Chronicle that he had been pressured by police to identify Cantu as the gunman, even though he believed Cantu to be innocent. Sam Millsap, the district attorney in the case, who had once supported capital punishment (“I’m no wild-eyed, pointy-headed liberal”), said that he was disturbed by the thought that he had made a mistake.
In 1995, Larry Griffin was put to death in Missouri, for a drive-by shooting of a drug dealer. The case rested largely on the eyewitness testimony of a career criminal named Robert Fitzgerald, who had been an informant for prosecutors before and was in the witness-protection program. Fitzgerald maintained that he happened to be at the scene because his car had broken down. After Griffin’s execution, a probe sponsored by the N.A.A.C.P.’s Legal Defense and Educational Fund revealed that a man who had been wounded during the incident insisted that Griffin was not the shooter. Moreover, the first police officer at the scene disputed that Fitzgerald had witnessed the crime.
These cases, however, stopped short of offering irrefutable proof that a “legally and factually innocent person” was executed. In 2005, a St. Louis prosecutor, Jennifer Joyce, launched an investigation of the Griffin case, upon being presented with what she called “compelling” evidence of Griffin’s potential innocence. After two years of reviewing the evidence, and interviewing a new eyewitness, Joyce said that she and her team were convinced that the “right person was convicted.”
Supreme Court Justice Antonin Scalia, in 2006, voted with a majority to uphold the death penalty in a Kansas case. In his opinion, Scalia declared that, in the modern judicial system, there has not been “a single case—not one—in which it is clear that a person was executed for a crime he did not commit. If such an event had occurred in recent years, we would not have to hunt for it; the innocent’s name would be shouted from the rooftops.”
“My problems are simple,” Willingham wrote Gilbert in September, 1999. “Try to keep them from killing me at all costs. End of story.”
During his first years on death row, Willingham had pleaded with his lawyer, David Martin, to rescue him. “You can’t imagine what it’s like to be here, with people I have no business even being around,” he wrote.
For a while, Willingham shared a cell with Ricky Lee Green, a serial killer, who castrated and fatally stabbed his victims, including a sixteen-year-old boy. (Green was executed in 1997.) Another of Willingham’s cellmates, who had an I.Q. below seventy and the emotional development of an eight-year-old, was raped by an inmate. “You remember me telling you I had a new celly?” Willingham wrote in a letter to his parents. “The little retarded boy. . . . There was this guy here on the wing who is a shit sorry coward (who is the same one I got into it with a little over a month ago). Well, he raped [my cellmate] in the 3 row shower week before last.” Willingham said that he couldn’t believe that someone would “rape a boy who cannot even defend himself. Pretty damn low.”
Because Willingham was known as a “baby killer,” he was a target of attacks. “Prison is a rough place, and with a case like mine they never give you the benefit of a doubt,” he wrote his parents. After he tried to fight one prisoner who threatened him, Willingham told a friend that if he hadn’t stood up for himself several inmates would have “beaten me up or raped or”—his thought trailed off.
Over the years, Willingham’s letters home became increasingly despairing. “This is a hard place, and it makes a person hard inside,” he wrote. “I told myself that was one thing I did not want and that was for this place to make me bitter, but it is hard.” He went on, “They have [executed] at least one person every month I have been here. It is senseless and brutal. . . . You see, we are not living in here, we are only existing.” In 1996, he wrote, “I just been trying to figure out why after having a wife and 3 beautiful children that I loved my life has to end like this. And sometimes it just seems like it is not worth it all. . . . In the 3 1/2 years I been here I have never felt that my life was as worthless and desolate as it is now.” Since the fire, he wrote, he had the sense that his life was slowly being erased. He obsessively looked at photographs of his children and Stacy, which he stored in his cell. “So long ago, so far away,” he wrote in a poem. “Was everything truly there?”
Inmates on death row are housed in a prison within a prison, where there are no attempts at rehabilitation, and no educational or training programs. In 1999, after seven prisoners tried to escape from Huntsville, Willingham and four hundred and fifty-nine other inmates on death row were moved to a more secure facility, in Livingston, Texas. Willingham was held in isolation in a sixty-square-foot cell, twenty-three hours a day. He tried to distract himself by drawing—“amateur stuff,” as he put it—and writing poems. In a poem about his children, he wrote, “There is nothing more beautiful than you on this earth.” When Gilbert once suggested some possible revisions to his poems, he explained that he wrote them simply as expressions, however crude, of his feelings. “So to me to cut them up and try to improve on them just for creative-writing purposes would be to destroy what I was doing to start with,” he said.
Despite his efforts to occupy his thoughts, he wrote in his diary that his mind “deteriorates each passing day.” He stopped working out and gained weight. He questioned his faith: “No God who cared about his creation would abandon the innocent.” He seemed not to care if another inmate attacked him. “A person who is already dead inside does not fear” death, he wrote.
One by one, the people he knew in prison were escorted into the execution chamber. There was Clifton Russell, Jr., who, at the age of eighteen, stabbed and beat a man to death, and who said, in his last statement, “I thank my Father, God in Heaven, for the grace he has granted me—I am ready.” There was Jeffery Dean Motley, who kidnapped and fatally shot a woman, and who declared, in his final words, “I love you, Mom. Goodbye.” And there was John Fearance, who murdered his neighbor, and who turned to God in his last moments and said, “I hope He will forgive me for what I done.”
Willingham had grown close to some of his prison mates, even though he knew that they were guilty of brutal crimes. In March, 2000, Willingham’s friend Ponchai Wilkerson—a twenty-eight-year-old who had shot and killed a clerk during a jewelry heist—was executed. Afterward, Willingham wrote in his diary that he felt “an emptiness that has not been touched since my children were taken from me.” A year later, another friend who was about to be executed—“one of the few real people I have met here not caught up in the bravado of prison”—asked Willingham to make him a final drawing. “Man, I never thought drawing a simple Rose could be so emotionally hard,” Willingham wrote. “The hard part is knowing that this will be the last thing I can do for him.”
Another inmate, Ernest Ray Willis, had a case that was freakishly similar to Willingham’s. In 1987, Willis had been convicted of setting a fire, in West Texas, that killed two women. Willis told investigators that he had been sleeping on a friend’s living-room couch and woke up to a house full of smoke. He said that he tried to rouse one of the women, who was sleeping in another room, but the flames and smoke drove him back, and he ran out the front door before the house exploded with flames. Witnesses maintained that Willis had acted suspiciously; he moved his car out of the yard, and didn’t show “any emotion,” as one volunteer firefighter put it. Authorities also wondered how Willis could have escaped the house without burning his bare feet. Fire investigators found pour patterns, puddle configurations, and other signs of arson. The authorities could discern no motive for the crime, but concluded that Willis, who had no previous record of violence, was a sociopath—a “demon,” as the prosecutor put it. Willis was charged with capital murder and sentenced to death.
Willis had eventually obtained what Willingham called, enviously, a “bad-ass lawyer.” James Blank, a noted patent attorney in New York, was assigned Willis’s case as part of his firm’s pro-bono work. Convinced that Willis was innocent, Blank devoted more than a dozen years to the case, and his firm spent millions, on fire consultants, private investigators, forensic experts, and the like. Willingham, meanwhile, relied on David Martin, his court-appointed lawyer, and one of Martin’s colleagues to handle his appeals. Willingham often told his parents, “You don’t know what it’s like to have lawyers who won’t even believe you’re innocent.” Like many inmates on death row, Willingham eventually filed a claim of inadequate legal representation. (When I recently asked Martin about his representation of Willingham, he said, “There were no grounds for reversal, and the verdict was absolutely the right one.” He said of the case, “Shit, it’s incredible that anyone’s even thinking about it.”)
Willingham tried to study the law himself, reading books such as “Tact in Court, or How Lawyers Win: Containing Sketches of Cases Won by Skill, Wit, Art, Tact, Courage and Eloquence.” Still, he confessed to a friend, “The law is so complicated it is hard for me to understand.” In 1996, he obtained a new court-appointed lawyer, Walter Reaves, who told me that he was appalled by the quality of Willingham’s defense at trial and on appeal. Reaves prepared for him a state writ of habeas corpus, known as a Great Writ. In the byzantine appeals process of death-penalty cases, which frequently takes more than ten years, the writ is the most critical stage: a prisoner can introduce new evidence detailing such things as perjured testimony, unreliable medical experts, and bogus scientific findings. Yet most indigent inmates, like Willingham, who constitute the bulk of those on death row, lack the resources to track down new witnesses or dig up fresh evidence. They must depend on court-appointed lawyers, many of whom are “unqualified, irresponsible, or overburdened,” as a study by the Texas Defender Service, a nonprofit organization, put it. In 2000, a Dallas Morning News investigation revealed that roughly a quarter of the inmates condemned to death in Texas were represented by court-appointed attorneys who had, at some point in their careers, been “reprimanded, placed on probation, suspended or banned from practicing law by the State Bar.” Although Reaves was more competent, he had few resources to reinvestigate the case, and his writ introduced no new exculpatory evidence: nothing further about Webb, or the reliability of the eyewitness testimony, or the credibility of the medical experts. It focussed primarily on procedural questions, such as whether the trial court erred in its instructions to the jury.
The Texas Court of Criminal Appeals was known for upholding convictions even when overwhelming exculpatory evidence came to light. In 1997, DNA testing proved that sperm collected from a rape victim did not match Roy Criner, who had been sentenced to ninety-nine years for the crime. Two lower courts recommended that the verdict be overturned, but the Court of Criminal Appeals upheld it, arguing that Criner might have worn a condom or might not have ejaculated. Sharon Keller, who is now the presiding judge on the court, stated in a majority opinion, “The new evidence does not establish innocence.” In 2000, George W. Bush pardoned Criner. (Keller was recently charged with judicial misconduct, for refusing to keep open past five o’clock a clerk’s office in order to allow a last-minute petition from a man who was executed later that night.)
On October 31, 1997, the Court of Criminal Appeals denied Willingham’s writ. After Willingham filed another writ of habeas corpus, this time in federal court, he was granted a temporary stay. In a poem, Willingham wrote, “One more chance, one more strike / Another bullet dodged, another date escaped.”
Willingham was entering his final stage of appeals. As his anxieties mounted, he increasingly relied upon Gilbert to investigate his case and for emotional support. “She may never know what a change she brought into my life,” he wrote in his diary. “For the first time in many years she gave me a purpose, something to look forward to.”
As their friendship deepened, he asked her to promise him that she would never disappear without explanation. “I already have that in my life,” he told her.
Together, they pored over clues and testimony. Gilbert says that she would send Reaves leads to follow up, but although he was sympathetic, nothing seemed to come of them. In 2002, a federal district court of appeals denied Willingham’s writ without even a hearing. “Now I start the last leg of my journey,” Willingham wrote to Gilbert. “Got to get things in order.”
He appealed to the U.S. Supreme Court, but in December, 2003, he was notified that it had declined to hear his case. He soon received a court order announcing that “the Director of the Department of Criminal Justice at Huntsville, Texas, acting by and through the executioner designated by said Director . . . is hereby directed and commanded, at some hour after 6:00 p.m. on the 17th day of February, 2004, at the Department of Criminal Justice in Huntsville, Texas, to carry out this sentence of death by intravenous injection of a substance or substances in a lethal quantity sufficient to cause the death of said Cameron Todd Willingham.”
Willingham wrote a letter to his parents. “Are you sitting down?” he asked, before breaking the news. “I love you both so much,” he said.
His only remaining recourse was to appeal to the governor of Texas, Rick Perry, a Republican, for clemency. The process, considered the last gatekeeper to the executioner, has been called by the U.S. Supreme Court “the ‘fail safe’ in our criminal justice system.”
IV
One day in January, 2004, Dr. Gerald Hurst, an acclaimed scientist and fire investigator, received a file describing all the evidence of arson gathered in Willingham’s case. Gilbert had come across Hurst’s name and, along with one of Willingham’s relatives, had contacted him, seeking his help. After their pleas, Hurst had agreed to look at the case pro bono, and Reaves, Willingham’s lawyer, had sent him the relevant documents, in the hope that there were grounds for clemency.
Hurst opened the file in the basement of his house in Austin, which served as a laboratory and an office, and was cluttered with microscopes and diagrams of half-finished experiments. Hurst was nearly six and half feet tall, though his stooped shoulders made him seem considerably shorter, and he had a gaunt face that was partly shrouded by long gray hair. He was wearing his customary outfit: black shoes, black socks, a black T-shirt, and loose-fitting black pants supported by black suspenders. In his mouth was a wad of chewing tobacco.
A child prodigy who was raised by a sharecropper during the Great Depression, Hurst used to prowl junk yards, collecting magnets and copper wires in order to build radios and other contraptions. In the early sixties, he received a Ph.D. in chemistry from Cambridge University, where he started to experiment with fluorine and other explosive chemicals, and once detonated his lab. Later, he worked as the chief scientist on secret weapons programs for several American companies, designing rockets and deadly fire bombs—or what he calls “god-awful things.” He helped patent what has been described, with only slight exaggeration, as “the world’s most powerful nonnuclear explosive”: an Astrolite bomb. He experimented with toxins so lethal that a fraction of a drop would rot human flesh, and in his laboratory he often had to wear a pressurized moon suit; despite such precautions, exposure to chemicals likely caused his liver to fail, and in 1994 he required a transplant. Working on what he calls “the dark side of arson,” he retrofitted napalm bombs with Astrolite, and developed ways for covert operatives in Vietnam to create bombs from local materials, such as chicken manure and sugar. He also perfected a method for making an exploding T-shirt by nitrating its fibres.
His conscience eventually began pricking him. “One day, you wonder, What the hell am I doing?” he recalls. He left the defense industry, and went on to invent the Mylar balloon, an improved version of Liquid Paper, and Kinepak, a kind of explosive that reduces the risk of accidental detonation. Because of his extraordinary knowledge of fire and explosives, companies in civil litigation frequently sought his help in determining the cause of a blaze. By the nineties, Hurst had begun devoting significant time to criminal-arson cases, and, as he was exposed to the methods of local and state fire investigators, he was shocked by what he saw.
Many arson investigators, it turned out, had only a high-school education. In most states, in order to be certified, investigators had to take a forty-hour course on fire investigation, and pass a written exam. Often, the bulk of an investigator’s training came on the job, learning from “old-timers” in the field, who passed down a body of wisdom about the telltale signs of arson, even though a study in 1977 warned that there was nothing in “the scientific literature to substantiate their validity.”
In 1992, the National Fire Protection Association, which promotes fire prevention and safety, published its first scientifically based guidelines to arson investigation. Still, many arson investigators believed that what they did was more an art than a science—a blend of experience and intuition. In 1997, the International Association of Arson Investigators filed a legal brief arguing that arson sleuths should not be bound by a 1993 Supreme Court decision requiring experts who testified at trials to adhere to the scientific method. What arson sleuths did, the brief claimed, was “less scientific.” By 2000, after the courts had rejected such claims, arson investigators increasingly recognized the scientific method, but there remained great variance in the field, with many practitioners still relying on the unverified techniques that had been used for generations. “People investigated fire largely with a flat-earth approach,” Hurst told me. “It looks like arson—therefore, it’s arson.” He went on, “My view is you have to have a scientific basis. Otherwise, it’s no different than witch-hunting.”
In 1998, Hurst investigated the case of a woman from North Carolina named Terri Hinson, who was charged with setting a fire that killed her seventeen-month-old son, and faced the death penalty. Hurst ran a series of experiments re-creating the conditions of the fire, which suggested that it had not been arson, as the investigators had claimed; rather, it had started accidentally, from a faulty electrical wire in the attic. Because of this research, Hinson was freed. John Lentini, a fire expert and the author of a leading scientific textbook on arson, describes Hurst as “brilliant.” A Texas prosecutor once told the Chicago Tribune, of Hurst, “If he says it was an arson fire, then it was. If he says it wasn’t, then it wasn’t.”
Hurst’s patents yielded considerable royalties, and he could afford to work pro bono on an arson case for months, even years. But he received the files on Willingham’s case only a few weeks before Willingham was scheduled to be executed. As Hurst looked through the case records, a statement by Manuel Vasquez, the state deputy fire marshal, jumped out at him. Vasquez had testified that, of the roughly twelve hundred to fifteen hundred fires he had investigated, “most all of them” were arson. This was an oddly high estimate; the Texas State Fire Marshals Office typically found arson in only fifty per cent of its cases.
Hurst was also struck by Vasquez’s claim that the Willingham blaze had “burned fast and hot” because of a liquid accelerant. The notion that a flammable or combustible liquid caused flames to reach higher temperatures had been repeated in court by arson sleuths for decades. Yet the theory was nonsense: experiments have proved that wood and gasoline-fuelled fires burn at essentially the same temperature.
Vasquez and Fogg had cited as proof of arson the fact that the front door’s aluminum threshold had melted. “The only thing that can cause that to react is an accelerant,” Vasquez said. Hurst was incredulous. A natural-wood fire can reach temperatures as high as two thousand degrees Fahrenheit—far hotter than the melting point for aluminum alloys, which ranges from a thousand to twelve hundred degrees. And, like many other investigators, Vasquez and Fogg mistakenly assumed that wood charring beneath the aluminum threshold was evidence that, as Vasquez put it, “a liquid accelerant flowed underneath and burned.” Hurst had conducted myriad experiments showing that such charring was caused simply by the aluminum conducting so much heat. In fact, when liquid accelerant is poured under a threshold a fire will extinguish, because of a lack of oxygen. (Other scientists had reached the same conclusion.) “Liquid accelerants can no more burn under an aluminum threshold than can grease burn in a skillet even with a loose-fitting lid,” Hurst declared in his report on the Willingham case.
Hurst then examined Fogg and Vasquez’s claim that the “brown stains” on Willingham’s front porch were evidence of “liquid accelerant,” which had not had time to soak into the concrete. Hurst had previously performed a test in his garage, in which he poured charcoal-lighter fluid on the concrete floor, and lit it. When the fire went out, there were no brown stains, only smudges of soot. Hurst had run the same experiment many times, with different kinds of liquid accelerants, and the result was always the same. Brown stains were common in fires; they were usually composed of rust or gunk from charred debris that had mixed with water from fire hoses.
Another crucial piece of evidence implicating Willingham was the “crazed glass” that Vasquez had attributed to the rapid heating from a fire fuelled with liquid accelerant. Yet, in November of 1991, a team of fire investigators had inspected fifty houses in the hills of Oakland, California, which had been ravaged by brush fires. In a dozen houses, the investigators discovered crazed glass, even though a liquid accelerant had not been used. Most of these houses were on the outskirts of the blaze, where firefighters had shot streams of water; as the investigators later wrote in a published study, they theorized that the fracturing had been induced by rapid cooling, rather than by sudden heating—thermal shock had caused the glass to contract so quickly that it settled disjointedly. The investigators then tested this hypothesis in a laboratory. When they heated glass, nothing happened. But each time they applied water to the heated glass the intricate patterns appeared. Hurst had seen the same phenomenon when he had blowtorched and cooled glass during his research at Cambridge. In his report, Hurst wrote that Vasquez and Fogg’s notion of crazed glass was no more than an “old wives’ tale.”
Hurst then confronted some of the most devastating arson evidence against Willingham: the burn trailer, the pour patterns and puddle configurations, the V-shape and other burn marks indicating that the fire had multiple points of origin, the burning underneath the children’s beds. There was also the positive test for mineral spirits by the front door, and Willingham’s seemingly implausible story that he had run out of the house without burning his bare feet.
As Hurst read through more of the files, he noticed that Willingham and his neighbors had described the windows in the front of the house suddenly exploding and flames roaring forth. It was then that Hurst thought of the legendary Lime Street Fire, one of the most pivotal in the history of arson investigation.
On the evening of October 15, 1990, a thirty-five-year-old man named Gerald Wayne Lewis was found standing in front of his house on Lime Street, in Jacksonville, Florida, holding his three-year-old son. His two-story wood-frame home was engulfed in flames. By the time the fire had been extinguished, six people were dead, including Lewis’s wife. Lewis said that he had rescued his son but was unable to get to the others, who were upstairs.
When fire investigators examined the scene, they found the classic signs of arson: low burns along the walls and floors, pour patterns and puddle configurations, and a burn trailer running from the living room into the hallway. Lewis claimed that the fire had started accidentally, on a couch in the living room—his son had been playing with matches. But a V-shaped pattern by one of the doors suggested that the fire had originated elsewhere. Some witnesses told authorities that Lewis seemed too calm during the fire and had never tried to get help. According to the Los Angeles Times, Lewis had previously been arrested for abusing his wife, who had taken out a restraining order against him. After a chemist said that he had detected the presence of gasoline on Lewis’s clothing and shoes, a report by the sheriff’s office concluded, “The fire was started as a result of a petroleum product being poured on the front porch, foyer, living room, stairwell and second floor bedroom.” Lewis was arrested and charged with six counts of murder. He faced the death penalty.
Subsequent tests, however, revealed that the laboratory identification of gasoline was wrong. Moreover, a local news television camera had captured Lewis in a clearly agitated state at the scene of the fire, and investigators discovered that at one point he had jumped in front of a moving car, asking the driver to call the Fire Department.
Seeking to bolster their theory of the crime, prosecutors turned to John Lentini, the fire expert, and John DeHaan, another leading investigator and textbook author. Despite some of the weaknesses of the case, Lentini told me that, given the classic burn patterns and puddle configurations in the house, he was sure that Lewis had set the fire: “I was prepared to testify and send this guy to Old Sparky”—the electric chair.
To discover the truth, the investigators, with the backing of the prosecution, decided to conduct an elaborate experiment and re-create the fire scene. Local officials gave the investigators permission to use a condemned house next to Lewis’s home, which was about to be torn down. The two houses were virtually identical, and the investigators refurbished the condemned one with the same kind of carpeting, curtains, and furniture that had been in Lewis’s home. The scientists also wired the building with heat and gas sensors that could withstand fire. The cost of the experiment came to twenty thousand dollars. Without using liquid accelerant, Lentini and DeHaan set the couch in the living room on fire, expecting that the experiment would demonstrate that Lewis’s version of events was implausible.
The investigators watched as the fire quickly consumed the couch, sending upward a plume of smoke that hit the ceiling and spread outward, creating a thick layer of hot gases overhead—an efficient radiator of heat. Within three minutes, this cloud, absorbing more gases from the fire below, was banking down the walls and filling the living room. As the cloud approached the floor, its temperature rose, in some areas, to more than eleven hundred degrees Fahrenheit. Suddenly, the entire room exploded in flames, as the radiant heat ignited every piece of furniture, every curtain, every possible fuel source, even the carpeting. The windows shattered.
The fire had reached what is called “flashover”—the point at which radiant heat causes a fire in a room to become a room on fire. Arson investigators knew about the concept of flashover, but it was widely believed to take much longer to occur, especially without a liquid accelerant. From a single fuel source—a couch—the room had reached flashover in four and a half minutes.
Because all the furniture in the living room had ignited, the blaze went from a fuel-controlled fire to a ventilation-controlled fire—or what scientists call “post-flashover.” During post-flashover, the path of the fire depends on new sources of oxygen, from an open door or window. One of the fire investigators, who had been standing by an open door in the living room, escaped moments before the oxygen-starved fire roared out of the room into the hallway—a fireball that caused the corridor to go quickly into flashover as well, propelling the fire out the front door and onto the porch.
After the fire was extinguished, the investigators inspected the hallway and living room. On the floor were irregularly shaped burn patterns that perfectly resembled pour patterns and puddle configurations. It turned out that these classic signs of arson can also appear on their own, after flashover. With the naked eye, it is impossible to distinguish between the pour patterns and puddle configurations caused by an accelerant and those caused naturally by post-flashover. The only reliable way to tell the difference is to take samples from the burn patterns and test them in a laboratory for the presence of flammable or combustible liquids.
During the Lime Street experiment, other things happened that were supposed to occur only in a fire fuelled by liquid accelerant: charring along the base of the walls and doorways, and burning under furniture. There was also a V-shaped pattern by the living-room doorway, far from where the fire had started on the couch. In a small fire, a V-shaped burn mark may pinpoint where a fire began, but during post-flashover these patterns can occur repeatedly, when various objects ignite.
One of the investigators muttered that they had just helped prove the defense’s case. Given the reasonable doubt raised by the experiment, the charges against Lewis were soon dropped. The Lime Street experiment had demolished prevailing notions about fire behavior. Subsequent tests by scientists showed that, during post-flashover, burning under beds and furniture was common, entire doors were consumed, and aluminum thresholds melted.
John Lentini says of the Lime Street Fire, “This was my epiphany. I almost sent a man to die based on theories that were a load of crap.”
Hurst next examined a floor plan of Willingham’s house that Vasquez had drawn, which delineated all the purported pour patterns and puddle configurations. Because the windows had blown out of the children’s room, Hurst knew that the fire had reached flashover. With his finger, Hurst traced along Vasquez’s diagram the burn trailer that had gone from the children’s room, turned right in the hallway, and headed out the front door. John Jackson, the prosecutor, had told me that the path was so “bizarre” that it had to have been caused by a liquid accelerant. But Hurst concluded that it was a natural product of the dynamics of fire during post-flashover. Willingham had fled out the front door, and the fire simply followed the ventilation path, toward the opening. Similarly, when Willingham had broken the windows in the children’s room, flames had shot outward.
Hurst recalled that Vasquez and Fogg had considered it impossible for Willingham to have run down the burning hallway without scorching his bare feet. But if the pour patterns and puddle configurations were a result of a flashover, Hurst reasoned, then they were consonant with Willingham’s explanation of events. When Willingham exited his bedroom, the hallway was not yet on fire; the flames were contained within the children’s bedroom, where, along the ceiling, he saw the “bright lights.” Just as the investigator safely stood by the door in the Lime Street experiment seconds before flashover, Willingham could have stood close to the children’s room without being harmed. (Prior to the Lime Street case, fire investigators had generally assumed that carbon monoxide diffuses quickly through a house during a fire. In fact, up until flashover, levels of carbon monoxide can be remarkably low beneath and outside the thermal cloud.) By the time the Corsicana fire achieved flashover, Willingham had already fled outside and was in the front yard.
Vasquez had made a videotape of the fire scene, and Hurst looked at the footage of the burn trailer. Even after repeated viewings, he could not detect three points of origin, as Vasquez had. (Fogg recently told me that he also saw a continuous trailer and disagreed with Vasquez, but added that nobody from the prosecution or the defense ever asked him on the stand about his opinion on the subject.)
After Hurst had reviewed Fogg and Vasquez’s list of more than twenty arson indicators, he believed that only one had any potential validity: the positive test for mineral spirits by the threshold of the front door. But why had the fire investigators obtained a positive reading only in that location? According to Fogg and Vasquez’s theory of the crime, Willingham had poured accelerant throughout the children’s bedroom and down the hallway. Officials had tested extensively in these areas—including where all the pour patterns and puddle configurations were—and turned up nothing. Jackson told me that he “never did understand why they weren’t able to recover” positive tests in these parts.
Hurst found it hard to imagine Willingham pouring accelerant on the front porch, where neighbors could have seen him. Scanning the files for clues, Hurst noticed a photograph of the porch taken before the fire, which had been entered into evidence. Sitting on the tiny porch was a charcoal grill. The porch was where the family barbecued. Court testimony from witnesses confirmed that there had been a grill, along with a container of lighter fluid, and that both had burned when the fire roared onto the porch during post-flashover. By the time Vasquez inspected the house, the grill had been removed from the porch, during cleanup. Though he cited the container of lighter fluid in his report, he made no mention of the grill. At the trial, he insisted that he had never been told of the grill’s earlier placement. Other authorities were aware of the grill but did not see its relevance. Hurst, however, was convinced that he had solved the mystery: when firefighters had blasted the porch with water, they had likely spread charcoal-lighter fluid from the melted container.
Without having visited the fire scene, Hurst says, it was impossible to pinpoint the cause of the blaze. But, based on the evidence, he had little doubt that it was an accidental fire—one caused most likely by the space heater or faulty electrical wiring. It explained why there had never been a motive for the crime. Hurst concluded that there was no evidence of arson, and that a man who had already lost his three children and spent twelve years in jail was about to be executed based on “junk science.” Hurst wrote his report in such a rush that he didn’t pause to fix the typos.
V
“I am a realist and I will not live a fantasy,” Willingham once told Gilbert about the prospect of proving his innocence. But in February, 2004, he began to have hope. Hurst’s findings had helped to exonerate more than ten people. Hurst even reviewed the scientific evidence against Willingham’s friend Ernest Willis, who had been on death row for the strikingly similar arson charge. Hurst says, “It was like I was looking at the same case. Just change the names.” In his report on the Willis case, Hurst concluded that not “a single item of physical evidence . . . supports a finding of arson.” A second fire expert hired by Ori White, the new district attorney in Willis’s district, concurred. After seventeen years on death row, Willis was set free. “I don’t turn killers loose,” White said at the time. “If Willis was guilty, I’d be retrying him right now. And I’d use Hurst as my witness. He’s a brilliant scientist.” White noted how close the system had come to murdering an innocent man. “He did not get executed, and I thank God for that,” he said.
On February 13th, four days before Willingham was scheduled to be executed, he got a call from Reaves, his attorney. Reaves told him that the fifteen members of the Board of Pardons and Paroles, which reviews an application for clemency and had been sent Hurst’s report, had made their decision.
“What is it?” Willingham asked.
“I’m sorry,” Reaves said. “They denied your petition.”
The vote was unanimous. Reaves could not offer an explanation: the board deliberates in secret, and its members are not bound by any specific criteria. The board members did not even have to review Willingham’s materials, and usually don’t debate a case in person; rather, they cast their votes by fax—a process that has become known as “death by fax.” Between 1976 and 2004, when Willingham filed his petition, the State of Texas had approved only one application for clemency from a prisoner on death row. A Texas appellate judge has called the clemency system “a legal fiction.” Reaves said of the board members, “They never asked me to attend a hearing or answer any questions.”
The Innocence Project obtained, through the Freedom of Information Act, all the records from the governor’s office and the board pertaining to Hurst’s report. “The documents show that they received the report, but neither office has any record of anyone acknowledging it, taking note of its significance, responding to it, or calling any attention to it within the government,” Barry Scheck said. “The only reasonable conclusion is that the governor’s office and the Board of Pardons and Paroles ignored scientific evidence.”
LaFayette Collins, who was a member of the board at the time, told me of the process, “You don’t vote guilt or innocence. You don’t retry the trial. You just make sure everything is in order and there are no glaring errors.” He noted that although the rules allowed for a hearing to consider important new evidence, “in my time there had never been one called.” When I asked him why Hurst’s report didn’t constitute evidence of “glaring errors,” he said, “We get all kinds of reports, but we don’t have the mechanisms to vet them.” Alvin Shaw, another board member at the time, said that the case didn’t “ring a bell,” adding, angrily, “Why would I want to talk about it?” Hurst calls the board’s actions “unconscionable.”
Though Reaves told Willingham that there was still a chance that Governor Perry might grant a thirty-day stay, Willingham began to prepare his last will and testament. He had earlier written Stacy a letter apologizing for not being a better husband and thanking her for everything she had given him, especially their three daughters. “I still know Amber’s voice, her smile, her cool Dude saying and how she said: I wanna hold you! Still feel the touch of Karmon and Kameron’s hands on my face.” He said that he hoped that “some day, somehow the truth will be known and my name cleared.”
He asked Stacy if his tombstone could be erected next to their children’s graves. Stacy, who had for so long expressed belief in Willingham’s innocence, had recently taken her first look at the original court records and arson findings. Unaware of Hurst’s report, she had determined that Willingham was guilty. She denied him his wish, later telling a reporter, “He took my kids away from me.”
Gilbert felt as if she had failed Willingham. Even before his pleas for clemency were denied, she told him that all she could give him was her friendship. He told her that it was enough “to be a part of your life in some small way so that in my passing I can know I was at last able to have felt the heart of another who might remember me when I’m gone.” He added, “There is nothing to forgive you for.” He told her that he would need her to be present at his execution, to help him cope with “my fears, thoughts, and feelings.”
On February 17th, the day he was set to die, Willingham’s parents and several relatives gathered in the prison visiting room. Plexiglas still separated Willingham from them. “I wish I could touch and hold both of you,” Willingham had written to them earlier. “I always hugged Mom but I never hugged Pop much.”
As Willingham looked at the group, he kept asking where Gilbert was. Gilbert had recently been driving home from a store when another car ran a red light and smashed into her. Willingham used to tell her to stay in her kitchen for a day, without leaving, to comprehend what it was like to be confined in prison, but she had always found an excuse not to do it. Now she was paralyzed from the neck down.
While she was in an intensive-care unit, she had tried to get a message to Willingham, but apparently failed. Gilbert’s daughter later read her a letter that Willingham had sent her, telling her how much he had grown to love her. He had written a poem: “Do you want to see beauty—like you have never seen? / Then close your eyes, and open your mind, and come along with me.”
Gilbert, who spent years in physical rehabilitation, gradually regaining motion in her arms and upper body, says, “All that time, I thought I was saving Willingham, and I realized then that he was saving me, giving me the strength to get through this. I know I will one day walk again, and I know it is because Willingham showed me the kind of courage it takes to survive.”
Willingham had requested a final meal, and at 4 p.m. on the seventeenth he was served it: three barbecued pork ribs, two orders of onion rings, fried okra, three beef enchiladas with cheese, and two slices of lemon cream pie. He received word that Governor Perry had refused to grant him a stay. (A spokesperson for Perry says, “The Governor made his decision based on the facts of the case.”) Willingham’s mother and father began to cry. “Don’t be sad, Momma,” Willingham said. “In fifty-five minutes, I’m a free man. I’m going home to see my kids.” Earlier, he had confessed to his parents that there was one thing about the day of the fire he had lied about. He said that he had never actually crawled into the children’s room. “I just didn’t want people to think I was a coward,” he said. Hurst told me, “People who have never been in a fire don’t understand why those who survive often can’t rescue the victims. They have no concept of what a fire is like.”
The warden told Willingham that it was time. Willingham, refusing to assist the process, lay down; he was carried into a chamber eight feet wide and ten feet long. The walls were painted green, and in the center of the room, where an electric chair used to be, was a sheeted gurney. Several guards strapped Willingham down with leather belts, snapping buckles across his arms and legs and chest. A medical team then inserted intravenous tubes into his arms. Each official had a separate role in the process, so that no one person felt responsible for taking a life.
Willingham had asked that his parents and family not be present in the gallery during this process, but as he looked out he could see Stacy watching. The warden pushed a remote control, and sodium thiopental, a barbiturate, was pumped into Willingham’s body. Then came a second drug, pancuronium bromide, which paralyzes the diaphragm, making it impossible to breathe. Finally, a third drug, potassium chloride, filled his veins, until his heart stopped, at 6:20 p.m. On his death certificate, the cause was listed as “Homicide.”
After his death, his parents were allowed to touch his face for the first time in more than a decade. Later, at Willingham’s request, they cremated his body and secretly spread some of his ashes over his children’s graves. He had told his parents, “Please don’t ever stop fighting to vindicate me.”
In December, 2004, questions about the scientific evidence in the Willingham case began to surface. Maurice Possley and Steve Mills, of the Chicago Tribune, had published an investigative series on flaws in forensic science; upon learning of Hurst’s report, Possley and Mills asked three fire experts, including John Lentini, to examine the original investigation. The experts concurred with Hurst’s report. Nearly two years later, the Innocence Project commissioned Lentini and three other top fire investigators to conduct an independent review of the arson evidence in the Willingham case. The panel concluded that “each and every one” of the indicators of arson had been “scientifically proven to be invalid.”
In 2005, Texas established a government commission to investigate allegations of error and misconduct by forensic scientists. The first cases that are being reviewed by the commission are those of Willingham and Willis. In mid-August, the noted fire scientist Craig Beyler, who was hired by the commission, completed his investigation. In a scathing report, he concluded that investigators in the Willingham case had no scientific basis for claiming that the fire was arson, ignored evidence that contradicted their theory, had no comprehension of flashover and fire dynamics, relied on discredited folklore, and failed to eliminate potential accidental or alternative causes of the fire. He said that Vasquez’s approach seemed to deny “rational reasoning” and was more “characteristic of mystics or psychics.” What’s more, Beyler determined that the investigation violated, as he put it to me, “not only the standards of today but even of the time period.” The commission is reviewing his findings, and plans to release its own report next year. Some legal scholars believe that the commission may narrowly assess the reliability of the scientific evidence. There is a chance, however, that Texas could become the first state to acknowledge officially that, since the advent of the modern judicial system, it had carried out the “execution of a legally and factually innocent person.”
Just before Willingham received the lethal injection, he was asked if he had any last words. He said, “The only statement I want to make is that I am an innocent man convicted of a crime I did not commit. I have been persecuted for twelve years for something I did not do. From God’s dust I came and to dust I will return, so the Earth shall become my throne.” ♦
Published in the print edition of the September 7, 2009, issue.
David Grann, a staff writer at The New Yorker since 2003, is the author of, most recently, “The White Darkness,” and of “Killers of the Flower Moon: The Osage Murders and the Birth of the F.B.I.,” which won an Edgar Award and was a finalist for the National Book Award.
3 Men Imprisoned in 1980 Brooklyn Arson Case Are Exonerated
Stephanie Clifford, “The New York Times (Dec. 16, 2015)
This is a brief article about a local exoneration based on new information about flawed arson science. I have pasted in the article below in case you cannot access the link.
3 Men Imprisoned in 1980 Brooklyn Arson Case Are Exonerated
The exonerations of Amaury Villalobos, left, and William Vasquez in a 1980 Brooklyn arson case was celebrated in State Supreme Court. From left, Adele Bernhard, a lawyer, with Mr. Villalobos; Rita Dave, a lawyer, with Mr. Vasquez; and Mr. Mora’s widow, Janet Mora, and their daughter, Eileen Mora.
Three men were cleared on Wednesday of setting a 1980 fire in Park Slope, Brooklyn, that caused the death of a mother and her five children, after the sole eyewitness was deemed unreliable and new understanding of the evidence fires leave behind was applied.
Two of the men had each spent almost 33 years in prison on arson and murder charges, the longest time served by any defendant whose conviction has been vacated under Ken Thompson, the Brooklyn district attorney. The third defendant died in prison.
Now 70 and 66, the surviving defendants, William Vasquez and Amaury Villalobos, stood in State Supreme Court in Brooklyn. Next to them stood the widow and daughter of the third defendant, Raymond Mora. His daughter, who was 7 when he died in prison in 1989, held up a photo of her father.
“I don’t know how this case managed to proceed,” Assistant District Attorney Mark J. Hale, who oversees the Conviction Review Unit in the office, told Justice Matthew J. D’Emic as he asked him to vacate the men’s convictions.
In February 1980, a townhouse at 695 Sackett Street burned to the ground. The third-floor tenants, a mother and her five young children, were killed.
The townhouse’s owner, Hannah Quick, immediately told the police that it had been arson and that she had heard the three defendants inside the townhouse just before the fire, then saw them walk out.
Ms. Quick, a drug dealer, said she had been feuding with one or two of the men over drugs.
Mr. Hale said her account was “implausible, almost unbelievable,” because it did not comport with the evidence and had inconsistencies.
As she was dying, Ms. Quick told her daughter that she had lied about the men’s involvement in the fire. Mr. Hale said in an interview on Tuesday that Ms. Quick’s motives to lie may have included liability and an insurance payment that she received (though she testified at the men’s trial that she did not receive such a payment).
A fire marshal found what he testified was evidence of arson, but Mr. Hale said evolving fire science meant that the 1980 analysis did not hold up today. Experts’ reports that Mr. Villalobos’s lawyer and the district attorney’s office commissioned as they re-examined the case showed no evidence of arson.
The office’s investigation found that addicts who bought heroin from the witness had prepped the drug over open flames on the townhouse’s first floor, that the floor was lit only by candles and that Ms. Quick had her own connection to the electrical grid, suggesting the fire might well have been an accident.
“These men were wrongfully convicted,” Mr. Hale said.
In a news conference after the hearing, Mr. Thompson said, “This is a case that shouldn’t have been brought, and we have to own up to that.”
After Justice D’Emic granted the motion, the courtroom spectators began clapping, and several stood to applaud.
Afterward, Mr. Vasquez, who had been wiping tears from behind his dark glasses while in court — he developed glaucoma in prison, which a lawyer said was untreated, and he is now blind — used a cane to navigate down the courthouse hallway.
“Those are years that nothing in the world can give me back, no money, no nothing,” Mr. Vasquez said after the proceeding. “It’s just like I lost myself in prison, I lost 33 and a half years of my life. I went in at 30, I come out at 65, so...” Mr. Vasquez shook his head and his sentence trailed off.
Mr. Villalobos said he had been denied parole several times because he refused to show remorse for a crime he had not committed. Still, he said: “I’m not angry. Why would you be angry? I know they did something bad to me, but God is there.”
His wife, Ernesta, who testified at his trial that she had been with him the night of the fire and who stayed with him throughout his time in prison, said she felt “happy, happy, happy, happy.”
Eileen Mora, Mr. Mora’s daughter, said she had “some sweet memories” of her father, who died of a heart attack in 1989.
“My father was a good man, and I’m glad we got to prove it today in court,” she said, her voice shaking. “There’s no better Christmas gift that I can get than to say that my dad was innocent, and now everybody knows.”
Mr. Villalobos, who like Mr. Vasquez has been out on parole since 2012, said he thought “all the time” that he would get his conviction vacated. In 2012, he contacted New York Law School, where Adele Bernhard, a law professor and director of the Post-Conviction Innocence Clinic, and her students worked on the case.
“Measuring the Impact of Cognitive Bias in Fire Investigation”
Paul Bieber, Arson Research Project (July 2012)
The treatment of bias is one of the areas in which forensic science diverges most dramatically from traditional science. This article describes different types of cognitive bias (expectation bias, confirmation bias, selective re-examination, role bias, and the effect of certainty on accuracy) and provides examples of cognitive biases affecting arson investigations.
Forensic Science Assessments: A Quality and Gap Analysis (July 11, 2017)
Excerpt from Fire Investigation: A Plain Language Summary
This is a “plain language” summary of an up-to-date assessment of the scientific validity of arson investigation. (The full report is available at this link; the assignment is to read pages 11-13.)
Project Findings and Outcomes
The recommendations emerging from this report promise to be influential in the advance of fire and fire debris analysis. The report offers an agenda encouraging the injection of rigorous scientific methods and standards into efforts to better understand the strengths and weaknesses of the current scientific bases of fire investigation. The report can be used by funders, public and private, to develop “calls for proposals” that will generate investigator initiated research into questions touching, directly or indirectly, on fire investigation. This research ideally will come from a broader range of disciplines and institutions than have traditionally been engaged in the study of fires. For example, this and other forensic studies can benefit from contributions by physicists, psychologists, statisticians, and chemists. We also hope that this and future reports, and the increase in funding it may encourage, will lead to more collaborative research by teams of forensic and academic scientists and engineers.
This report should also help other key actors in the criminal justice system—law enforcement, attorneys, and judges—better understand the limits of the current scientific underpinnings of fire analysis. For instance, as judges gain a better understanding of the current state of knowledge regarding fires they will be better able to fulfill their gatekeeping responsibilities, thus bringing a “culture of science” into legal decision-making.
Fire Science
Fire investigation has been the target of well-founded skepticism as relatively newer investigative assumptions and practices are replacing those that historically were accepted “wisdom” in the field. However, not everyone is ready to abandon the standards they have relied on for their entire careers. “Police and fire investigators across the country were often slow to accept advances in fire science that called into question what had long been considered unmistakable evidence of arson” and in “many jurisdictions, those [new] rules were slow to take hold, as veteran investigators clung to what are now considered disproven theories.” In fact, some investigators have been “openly hostile to the updated science” (Mills, 2015).
As advances in fire science are being made, investigators and attorneys are revisiting old cases. What they are finding is chilling. John Lentini (one of the authors of this report) estimates that “a couple hundred” people are wrongly imprisoned because of methods used in fire investigation that have not been empirically validated. According to the National Registry of Exonerations, 63 individuals convicted of arson have already been exonerated since 1991 (http://www.exonerationregistry.org/). Michael Keller, an electrical engineer at a research facility supported by the Bureau of Alcohol, Tobacco, Firearms and Explosives, has investigated three potential capital cases in which the fires were eventually found to be accidental (Bui, 2015). We will never know for sure how many people have been wrongfully convicted based on the aforementioned methods and practices. Likewise, we will never know how many criminal arsonists have not been indicted because of inadequate forensic tools. This report sets forth an evaluation of research on fire scene investigation and fire debris analysis, and we hope, will be 6 read by judges, lawyers for the prosecution and defense, law enforcement officers, laboratory scientists, policymakers, funding agencies, and fire investigation practitioners. It also points to future directions that much-needed research should take.
CONCLUSIONS & RECOMMENDATIONS
FIRE SCENE INVESTIGATION
The conclusions presented in this section of the report are drawn from the two technical sections that follow (fire scene investigation and fire debris analysis). Recommendations are tied to the conclusions. As with the technical sections that follow, the Conclusions and Recommendations are divided between fire scene investigation and fire debris analysis. Each conclusion is accompanied by page numbers to guide the reader to the location in the technical section that supports it. The 25 recommendations can be the basis for further research by fire scientists, as well as scientists from other disciplines in academia.
Origin and Cause Determination
Conclusions
- Fire has been extensively studied, but the complex chemical and physical processes involved are still not fully understood. While much is known about the behavior of fires in building enclosures, using that knowledge to determine where a particular fire started and what caused it remains very challenging. An incorrect determination of a fire’s origin generally leads to an incorrect determination of its cause (pp. 13, 18).
- Determining the origin of a pre-flashover fire is generally a straightforward exercise based on burn pattern analysis. However, determining the origin of post-flashover fires is more difficult because post-flashover fires may create new ventilation-generated burn patterns while obscuring pre-existing burn patterns. Moreover, the longer a fire burns in a fully involved condition, the more difficult is the determination of the correct area of origin (pp. 13, 17, 19).
- The speed with which a fire spreads in residences with contemporary furnishings cannot generally be used alone to classify a fire as accidental or incendiary because fires involving modern upholstered furnishings tend to burn faster than those used in older furnishings, regardless of the cause of the fire (p. 19).
- Computer-based deterministic fire models can be used as one means of testing different hypotheses regarding the origin and development of a fire, but such models cannot generally be used alone to determine the cause of a fire. Uncertainties exist concerning these models when they are applied to fire cause determination (p. 23).
Recommendations
- To improve the analysis of a fire’s origin and cause, tests should be run in both reduced and full scale, using multiple compartments and multiple openings, fully documenting the aftermath; with the burning of different materials under a range of realistic fire conditions; and by lighting fires in identically constructed compartments. These tests should be scientifically instrumented so that information, such as temperature at various layers of the room and radiant heat fluxes are measured. A major consideration in deciding to conduct this research is the high cost of burning compartment test rooms. However, by using reduced scale testing, the cost can be lowered and may provide the same critical values related to heat as a full-scale test.
- When a physical fire test is conducted, the fire scenario being tested should also be simulated with a deterministic fire model to evaluate the accuracy of the model and to better understand uncertainties associated with the model. Based on the data obtained through such comparisons, the computer-based deterministic fire models can be continually refined to produce more accurate results, and over time may find an expanded role as a useful tool in actual investigations.
Locating Ignitable Liquid Residues in Fire Debris
Conclusions
- A well-trained canine detection team is the current “gold standard” for locating samples at the fire scene that may test positive for ignitable liquid residues (ILRs) in the fire debris analysis laboratory. Canines are advantageous because they provide immediate feedback and are mobile, allowing them to search a large space in a very short time. However, canine alerts should not be relied upon unless confirmed by laboratory analysis (pp. 18, 30).
- Many substances are produced during a fire, but investigators are currently in a position to look only for evidence of ignitable liquids, for which there is currently validated instrumentation. Other chemical markers are of little value, because there is no way to determine when they may have been created during a fire (p. 23).
Recommendations
- Testimony that relies on canine alerts only, without supporting laboratory results, should not be used in court proceedings.
- New technologies, as well as additional training aids and research on new methods need to be developed for measuring canine performance that could enhance their effectiveness.
- Comparative research assessing the effectiveness of technologically more innovative field tools against the effectiveness of canine use should be a research priority.
Reliability and Validity
Conclusion
- Little is known about the consistency and accuracy of conclusions among experienced investigators when presented with the same data (p. 25).
Recommendations
- The reliability of conclusions when fire investigators are presented with similar data of fire origin and cause should be studied. This will allow the calculation of both error rates and the reliability of investigators’ conclusions. This exercise should be repeated over several fire scenarios to help determine what types of fire scenes elicit few disagreements and what types elicit many. These tests of reliability will provide feedback on decision points that cause divergent findings among investigators.
- Not only should the reliability of investigators’ conclusions or diagnoses be established, but research should also be done on the validity of those conclusions or diagnoses. The use of “test fires” (described in recommendation #1) will help to establish a “ground truth” against which the validity of investigators’ conclusions can be assessed. Such tests will also create crucial knowledge about which cues or diagnostics are genuinely associated with various fire characteristics. 8. The data generated by the research on reliability and validity should be incorporated into a database that could be used to develop standards for identifying the origin and cause of fires and serve as a resource for the education and training of fire investigators.
Cognitive Bias
Conclusions
- Interpretation of the evidence regarding origin and cause is often subjective and depends to a significant degree on human cognitive factors (p. 13).
- Evidence from other domains, as well as within forensic science, suggests that there are practical ways to mitigate and minimize bias (such as, Linear Sequential Unmasking). The aim of such procedures is to maximize the independence of mind of forensic examiners to be as bias free as possible in both the identification of relevant evidence and the conclusion about what the evidence shows about a fire’s origin and cause. In this context, the distinction between bias and relevance should be noted. For example, eyewitness testimony although potentially biasing, can also be relevant in some instances. One of the recommendations from the OSAC (Organization of Scientific Area Committees) Fire and Explosion Investigation Subcommittee suggests research in this area (p. 26).
Recommendations
- Given what is known about the role of cognitive bias in interpretation and decision making, the work by fire scene investigators should be separated from other components of the fire investigation. Those who gather and prepare evidence should focus on scientific analysis and be as neutral as possible in deciding what evidence to collect and how to interpret it. This would help to minimize bias that might affect fire scene investigation. In some jurisdictions, this will be cost prohibitive, but in those jurisdictions where this is not the case, this recommendation should be followed.
- Case management interventions should be adopted that shield fire scene investigators from information irrelevant, but potentially biasing, for assessing the scientific evidence critical to determining fire origin and cause (see the National Commission on Forensic Science document “Ensuring That Forensic Science Is Based Upon Task-Relevant Information”).
- There should be policies and procedures that clarify what is or is not relevant for fire investigators to know at each stage of an investigation, in order to reduce the possibility of bias.
- Forensic laboratories and fire scene investigators’ professional societies should adopt policies and procedures to help implement the recommended changes in case management. Policies and procedures should reflect what is best for helping fire investigators reach accurate scientific conclusions without regard for the convenience of the labs or others associated with the investigation, such as law enforcement.
- In general scientific practices, bias is often handled by “blinded procedures,” whereby researchers are unaware of information irrelevant to their task. This should be the gold standard for fire scene investigation as well. The literature on this topic discussed in the 10 technical report includes specific suggestions for accomplishing this challenging task in the context of forensic science.
- The implementation of the recommendations for minimizing the influence of cognitive bias should be accompanied by monitoring and evaluation to assess their impact and, where appropriate, lead to modifications.
Education, Certification, and Experience
Conclusions
- There are insufficient educational and proficiency testing requirements for fire scene investigators (p. 22).
- There is currently no scientific basis for concluding that the accuracy of certified fire investigators, in particular, is better than the accuracy of non-certified fire investigators (p. 22).
Recommendations
- Education and training for fire investigators should cover the issues of human cognition and cognitive bias, as well as what has been discovered in the reliability and validity testing discussed in recommendations #6 and #7. The training could be done through live training simulations as well as online with photographs and videos or other depictions of test fires in which the origin and cause are known with certainty. Testing materials could employ the same technique with new fire depictions not used in the training materials.
- The effects of education, training, and certification on fire investigators’ ability to determine fire origin and cause should be further studied.
Writing Reflection #11 Writing Reflection #11
Please go to our Moodle Page and under "Class 11" you will find the prompt and submission folder for Writing Reflection #11.
3.4 [omitted for Spring 2023: DNA] 3.4 [omitted for Spring 2023: DNA]
PCAST Report
Pages 69-83
This section of the PCAST Report describes traditional DNA analysis (single-source and simple-mixtures samples) as well as the less reliable method of DNA analysis of complex-mixtures samples.
DNA Mixtures: A Forensic Science Explainer
Rich Press, National Institute of Standards and Technology (2019)
This plain language article identifies some of the problems with mixtures and “touch DNA.”
People v. Collins People v. Collins
15 N.Y.S.3d 564 (N.Y. Sup Ct. 2015)
In this case (litigated by New York Legal Aid Society attorneys) you can see: (a) a judge transparently trying to understand the complex topic of high-sensitivity DNA analysis; (b) assessing the who’s-who of the relevant scientific community; and (c) deciding whether there’s something wrong with the defense attorneys objecting to this evidence when proffered by the prosecution and yet relying on it if it exculpates their clients. This opinion also illustrates the extent to which a judge can engage in an assessment of the method even in a Frye opinion.
NOTE: This is a long case, but as you read you'll see I have marked numerous places where you can skim the judge's deccription of the science.
[15 NYS3d 564]
The People of the State of New York, Plaintiff, v Jaquan Collins, Defendant. The People of the State of New York, Plaintiff, v Andrew Peaks, Defendant.
Supreme Court, Kings County,
July 2, 2015
*596APPEARANCES OF COUNSEL
Seymour W. James, Jr., The Legal Aid Society, Brooklyn (Karen Faraguna of counsel), and Seymour W. James, Jr., The Legal Aid Society, New York City (Jessica Goldthwaite, Clinton Hughes, Susan Friedman and Daniel Ades of counsel), for defendant in the first above-entitled action.
Kenneth P. Thompson, District Attorney, Brooklyn (Alfred Delngeniis and Kyle C. Reeves of counsel), for plaintiff in the first above-entitled action.
Seymour W. James, Jr., The Legal Aid Society, Brooklyn (Susan Morris of counsel), and Seymour W. James, Jr., The Legal Aid Society, New York City (Jessica Goldthwaite, Clinton Hughes, Susan Friedman and Daniel Ades of counsel), for defendant in the second above-entitled action.
Kenneth P. Thompson, District Attorney, Brooklyn (Elizabeth Doerfler and Kyle C. Reeves of counsel), for plaintiff in the second above-entitled action.
*597OPINION OF THE COURT
In these unrelated cases, defendants Collins and Peaks face violent felony charges. What the cases have in common is that the People applied to introduce DNA evidence at each defendant’s trial.
In defendant Collins’ case, the People obtained two DNA “mixtures,” each apparently from three contributors, from the handlebars of a bicycle ridden by the perpetrator of a shooting. The DNA samples were very small, and so they were tested with “high sensitivity” analysis in the laboratory of New York City’s Office of the Chief Medical Examiner (OCME). After that, a new OCME software program called the Forensic Statistical Tool (FST) indicated that one DNA mixture was 972,000 times more probable if the sample originated from defendant Collins and two unknown, unrelated people than if it instead originated from three unknown, unrelated individuals. The other mixture was 19.4 times more probable if the sample originated from Collins and two unknown, unrelated people than if it instead originated from three unknown, unrelated individuals. One might reasonably expect that this evidence would be conclusive on any identity issue at trial.
In defendant Peaks’ case, the People obtained a DNA “mixture” from the bra of the victim of a sexual assault. At least one female and two males contributed DNA to the sample. Using standard DNA analysis, not “high sensitivity” analysis, and using the FST software, an analyst determined that the sample was 19.6 times more probable if the sample originated from defendant Peaks, the victim, and an unknown, unrelated person than if it instead originated from the victim and two unknown, unrelated persons. One might reasonably expect that this evidence would be highly persuasive on any identity issue at trial.
Defendant Collins has moved to preclude the DNA evidence in his case on the theory that neither “high sensitivity” DNA analysis nor the FST is generally accepted in the relevant scientific community. Defendant Peaks has moved to preclude DNA evidence on the theory that the FST is not generally accepted in the relevant scientific community. This court ordered that a Frye hearing be held (see Frye v United States, 293 F *5981013 [DC Cir 1923]), and the cases were consolidated for the hearing.1
I. The Underlying Facts and Issues
A, Standard DNA Analysis
This court recognizes that judges are, far and away, not the people best qualified to explain science. That observation is doubly applicable when novel scientific techniques are at issue — and that of course is precisely what Frye analysis involves. But courts are bound to do their best.
DNA defines who we are. Who we are depends on the genetic contributions of our natural parents. Each parent contributes half of our genetic coding, in that each provides half of the DNA at each point on our genetic map. That entire map is included in the nuclei of most cells in our bodies. But nature has it that parental contributions vary, even as to most siblings. Apart from identical twins, no one has a DNA profile that is the same as anyone else’s.
Law enforcement puts that to use. Under the right circumstances, standard DNA analysis tells whether cells left at a crime scene contain a particular person’s DNA. The entirety of the DNA in the cells need not be examined. It is enough to look on the genome at 15 places chosen for this purpose — each of those places being called a “locus,” plural “loci” — to determine whether there is a match between a crime scene DNA sample and the DNA of a suspect.2
At each locus among the select 15, DNA patterns repeat themselves in numbers that differ from person to person. The number of repeats lets us place a distinguishing label on each locus — a number that is the “allele.” If the pattern repeats 11 times, the allele at the locus is an 11. At every locus, we possess an allele from our mother and another from our father. If, at a particular one of the 15 chosen loci, one parent gave you 11 repeats and the other gave you 15, you will be considered an “11,15” at that locus, and a “heterozygote.” If, by chance, each *599parent gave you 11 repeats, you are at that locus simply an “11” and at that locus you are a “homozygote.”
Assume, counter to fact and just to simplify, that there are eight possible repeat numbers at each of the select loci, and thus 64 possible combinations of maternal and paternal numbers at each locus for unrelated people. Likewise assume, counter to fact and just to simplify, that all possibilities, maternally and paternally, are equally likely at every locus. One in 64 unrelated people would share the same numbers at a particular locus.
But there are 15 relevant loci. Assume (as the experts agree) that the results at each of the 15 select loci are independent. Then the chances that unrelated people would share the same numbers at two loci would be one in 4,096 (64 x 64). Unrelated people would share the same numbers at three loci only one time in 262,144 (64 x 4,096). By the time that you find that people have the same numbers at all 15 loci, the odds against the match are well over one trillion to one.
All of that is non-controversial here. As noted above, contested here are two new DNA tools — “high sensitivity” analysis and the FST software.
B. The Facts and Issues as to Defendant Collins
On August 15, 2010, Joshua Hamilton was walking near the Kingsborough Houses in Brooklyn. A man on a bicycle began shooting at Hamilton, jumped off the bicycle, and continued shooting. The assailant then fled, leaving the bicycle behind. Police officers who responded to the scene swabbed the handlebars and sent the two swabs for DNA testing at the OCME laboratory.
Too little DNA was present on either swab to permit standard DNA testing, in which the recovered DNA is copied 28 times before it is analyzed. The OCME lab therefore employed a relatively new procedure, pioneered in Britain, called “high sensitivity” analysis. Under this procedure, recovered DNA is copied in 31 “cycles” instead of 28. The extra three duplications make available to the analyst about eight times as much material as does the standard procedure.3
*600With the three extra cycles come not just more material to examine, but also more “stochastic” effects. The software kits that create DNA profiles vary, but generally were designed to create DNA profiles of the 15 select loci with quantities of DNA that can be analyzed after the standard 28 copying cycles. But even with standard analysis, the software reveals “stochastic” imperfections — random additions to or subtractions from the material examined.4 Interpretive protocols are adopted by laboratories to filter out these extraneous results. The use of these protocols is not controversial and does not cast doubt on the scientific validity of standard DNA testing. However, the extra three cycles employed in high sensitivity testing magnify the stochastic effects. New protocols have been created by OCME to compensate. The parties dispute whether the high sensitivity protocols adequately do so. The defense position is that the scientific community has not agreed that high sensitivity analysis yields reliable DNA profiles.
In defendant Collins’ case, not only the high sensitivity protocols are at issue. The high sensitivity analysis revealed, not a single profile, but a mixture of three peoples’ DNA. The DNA mixtures on the handlebars did not, by themselves, yield highly inculpatory results. When a mixture of DNA from two or more people is analyzed, the quantities of DNA contributed by particular individuals may differ sufficiently to permit the creation of separate DNA profiles of the contributors. In Collins’ case, that was not true; the quantities of DNA from the contributors were too nearly equal to permit such differentiation. The high sensitivity results initially indicated only that Collins “could be a contributor” to the DNA sample on one handlebar of the bicycle, because all his DNA alleles were found *601in the mixture — with many other alleles. As to the other handlebar, not all of defendant’s alleles were present, but the absence of the others could be explained. As to that sample, defendant therefore “could not be excluded” as a contributor.
Subsequently, the OCME lab analyzed the mixture results with the new FST software tool. The FST was developed in the hope that more informative statistical information could be reported as to the likelihood that a particular individual contributed DNA to a mixture. The FST was created by OCME through analysis of samples of DNA mixtures made by known contributors. As part of the process, OCME counted the stochastic effects at each of the select 15 loci by comparing the results with the contributors’ actual profiles. OCME then used the results to calculate the probabilities of additional or missing alleles at each locus. Through mathematical analysis, very importantly factoring in this probability of stochastic effects, the FST software permits an analyst to state a “likelihood ratio”: how much more likely, or less likely, a two-person or three-person DNA mixture is if a suspect is a contributor, than it would be if instead the suspect was not a contributor. In defendant Collins’ case, the FST yielded a report that the mixture on one handlebar was 972,000 times more likely if it originated from Collins and two unknown, unrelated people than if it came from three unknown, unrelated individuals. The other mixture was 19.4 times more likely if Collins was a contributor.
The validity of the math in the probability analysis underlying the FST software is not at issue. That mathematical analysis is “Bayesian” analysis. Bayes was a mathematician who worked in the eighteenth century. His methods for calculating probabilities are employed throughout all fields in which probabilities are calculated, including medicine and molecular genetics. Nor does the defense dispute the People’s position that the FST software does calculations of Bayesian probabilities like a simple calculator, performing complex mathematical functions with more precision, and certainly with far more speed, than a human could.
Instead, the defense attack on the FST centers on whether the probabilities of stochastic effects, determined in the “known contributor” tests, were produced through methods generally accepted in the relevant scientific community, and whether the methods for testing crime scene samples otherwise allow accurate input to the FST. The defense also complains that the *602FST wrongly limits analysis to single hypothetical and unscientifically prevents alternative analyses.
C. The Facts and Issues as to Defendant Peaks
At about 5:00 a.m. on July 13, 2010, an individual who will be referred to here as victim A was attacked on the sixth floor of a residential building on Stanley Avenue in Brooklyn. Victim A was grabbed around the neck and dragged into a stairwell, where her attacker touched her breasts and took her purse. When victim A’s brother came to her aid the perpetrator fled, dropping the purse and losing his Yankees cap as he did so.
At about 9:00 a.m. on August 26, 2010, an individual who will be referred to here as victim B entered the elevator of a residential building on Loring Avenue in Brooklyn. A stranger entered with her and got off with her on the seventh floor. He then put victim B in a choke hold and displayed a box cutter. After taking money from victim B’s purse, the attacker ordered her to lift her shirt and bra. When she complied, the attacker put his mouth on her breast. He soon ran from the scene.
In victim A’s case, the purse and the Yankees cap were sent to OCME for DNA analysis. The cap yielded a DNA mixture with a recognizable major contributor and a minor contributor. The major contributor’s DNA was subjected to standard analysis, and would be found in only one in over 6.8 trillion people. That profile matches the profile of defendant Peaks. No issue is presented here as to the testing in victim A’s case.
In victim B’s case, the victim’s shirt and bra were sent to OCME for DNA testing. A DNA mixture from at least three contributors was found on the bra. That mixture included both male and female DNA, but — unlike in the case of victim A— the amounts of DNA left by the contributors were too similar for individual profiles to be identified. A non-standard type of DNA analysis, YSTR testing, was then employed. The YSTR analysis, which examines only alleles contributed by males, indicated that DNA from a major male contributor and DNA from a minor male contributor were in the mixture. Defendant Peaks’ YSTR profile matched the profile of the major male contributor, indicating that defendant or a paternal male relative could be the source of the major male contribution to the mixture. This profile would be found in one in a thousand black males. This finding also is not in issue in this motion.
Finally — and very much at issue here — the results of standard DNA testing of the mixture on the bra were analyzed *603with the new FST software. Defendant’s alleles at two loci were not found during the initial testing. Nonetheless, the FST software, taking stochastic effects into account, concluded that the mixture was 19.6 times more likely if defendant, victim B, and an unknown, unrelated individual were the contributors than if the contributors were victim B and two unknown unrelated individuals.
Defendant Peaks challenges the use of the FST software. Like defendant Collins, he does not dispute Bayesian probability theory, or the use of FST software as a calculator for determinations too complex for ready performance by humans. He instead objects that the scientific community does not concur with the way that the FST assesses the probabilities of stochastic effects, and he objects as well that it does not permit assessment of various alternative hypotheses about the contributors of DNA.
II. The Frye Rule
This opinion has cited Frye, yet has not explained it. But in New York legal circles, little explication is required. Novel methods of scientific analysis can produce admissible evidence only if the relevant scientific community generally (though not necessarily unanimously) considers those methods to be reliable. (Frye v United States, 293 F 1013 [DC Cir 1923].)
Put another way, a court assessing the admissibility of evidence under Frye is not charged with deciding the validity of novel scientific procedures. It would hardly be sensible to assign that task to the judiciary, most of which is as patently unqualified to perform the task as is this court. Judges should be “counting scientists’ votes,” and not “verifying the soundness of a scientific conclusion.” (Parker v Mobil Oil Corp., 7 NY3d 434, 446-447 [2006], quoting People v Wesley, 83 NY2d 417, 439 [1994, Kaye, Ch. J., concurring].)
One New York trial court has, after a Frye hearing, upheld the use of high sensitivity DNA analysis. (People v Megnath, 27 Misc 3d 405 [Sup Ct, Queens County 2010, Hanophy, J.].) Another has, after a Frye hearing, held that the FST approach to mixture analysis is generally accepted in the DNA community. (People v William Rodriguez, Sup Ct, NY County, Oct. 24, 2013, Carruthers, J. [unreported].) With all respect to the authors of those decisions, this court has counted the scientists’ votes differently, and disagrees with their conclusions.
*604III. High Sensitivity Analysis, Often Called Low Copy Number Analysis
At OCME, standard testing of a DNA sample utilizes what is called PCR-STR analysis to identify the alleles at 15 points on the genome of the individual who supplied the DNA — and it determines the person’s sex as well. The laboratory work proceeds in six stages. First, an evidence sample thought to contain DNA is examined with “presumptive” tests to confirm that theory. Second, if the result is positive, chemicals extract DNA from the sample and purify the DNA. Third, the amount of DNA is quantified, and that quantification resolves whether the testing may continue with standard DNA analysis.
Fourth, standard analysis continues with the “amplification” of the DNA in 28 cycles. In each cycle, the DNA, which resembles a twisted ladder, is sliced in two on its vertical axis. Each half ladder then bonds with chemicals to produce a ladder identical to the original, so that there are twice the number of complete ladders as existed before. After 28 cycles, the result is a sample that is 2 to the 28th power (about 268 million) times bigger than the original.
Fifth, the technician employs a capillary electrophoresis machine. That device tracks the passage of DNA through a gel. Alleles with smaller numbers have fewer repeats of their DNA pattern than alleles with larger numbers and, being smaller, can move through the gel more quickly. For each of the 15 target loci, the machine records the speed of the alleles.
Sixth, computer software analyzes the electrophoresis data about the movement of the alleles. That permits a graph of “peaks” for the loci which identify which alleles are present at each. An analyst then takes over, disregarding stray peaks below a set threshold and resolving whether the remaining alleles match another DNA profile from a crime scene sample, a suspect, a victim, or the national DNA database.
Again as noted above, standard DNA analysis is well accepted, and is not at issue. And, with respect to defendant Peaks, it has already been mentioned that standard analysis of the major contributor of DNA on a Yankees cap worn by victim A’s attacker was someone with Peaks’ DNA profile. One in 6.8 trillion people would share that profile.
But here we deal not with standard analysis, but with high sensitivity analysis of small DNA deposits — or if you prefer, “low copy number” analysis. In the view of OCME, inclusion of *605100 picograms of DNA in a sample permits standard analysis. About 16 human cells would yield that much DNA. High sensitivity analysis comes into play if the sample is below 100 picograms. One cell will produce about six picograms of DNA, and high sensitivity analysis hopes to provide information where a recovered DNA sample contains between about 6 and about 100 picograms. In that range, amplification with 28 cycles does not produce enough DNA to allow electrophoresis analysis.
There are very many cases in which DNA samples smaller than 100 picograms are all that can be obtained. Each case is different, but very often small “touch” samples are obtained — a few cells’ worth of DNA were left by someone who touched a gun, a steering wheel, the handle of a drawer, a doorknob, or the handlebar of a bicycle. In other cases, a small DNA sample will be left by someone on a hat or a glove recovered at a crime scene, or by touching the victim of an assault with his mouth. OCME scientists — and many others — quite understandably wish to develop reliable methods to create DNA profiles from these smaller samples, and thereby increase the odds that the justice system will correctly resolve criminal cases. Of course, reliable DNA evidence can lead either to just convictions, or to proper exonerations.
Defendant Collins asserts, and the People acknowledge, that high sensitivity analysis increases stochastic effects which can impede proper DNA analysis. The People argue that high sensitivity analysis is reliable nonetheless, because OCME’s protocols create conservative interpretations of the test data and ensure trustworthy and sound scientific conclusions. This court, after conducting a Frye hearing, will not seek to state whether the People are correct; this court will instead consider the evidence about whether the relevant scientific community generally accepts the OCME protocols. But some further discussion of those protocols must precede that consideration.
This opinion has already identified the four kinds of stochastic effects that are of general concern in DNA analysis. Dropout — the failure of any DNA to register at one or more loci in a contributor’s sample — is not of great concern in high sensitivity analysis of a crime scene sample left by a guilty person. If no result appears at a locus, the odds behind any inculpatory result are simply reduced to the guilty party’s benefit. However, the innocent defendant loses one or, if there is more drop-out, more chances for an exculpatory non-matching allele to appear. *606Notably, in real-world conditions very small DNA samples are particularly subject to degradation and are thus more prone than larger samples to having alleles simply disappear.
More often problematic is drop-in. Drop-in is, in effect, contamination with extraneous alleles, either at the point of the recovery of a sample or, hopefully less frequently, during the laboratory analysis of the sample. Drop-in is troublesome in any situation, but obviously is of most concern if the contamination should supply an innocent person’s alleles to a sample and increase the perceived chance that he was the contributor.
Stutter is in some ways like drop-in, in that it adds an allele that does not belong in the profile. The cause, however, is not contamination. Stutter is a common “echo” effect appearing on the DNA profile — usually one allele before its true allele, or less frequently, just after. Stutter is of little concern in a single contributor’s standard DNA analysis, as it is suspect not only for its location, but also because it shows as a peak on the graph that is much shorter than the neighbor allele it “echoes.” It is more of a problem with high sensitivity analysis, as such analysis can involve far less informative peak heights. Stutter may show a peak as high or higher than the true allele, and thus is harder to identify.
And that brings us to peak heights, the fourth stochastic effect relevant here. With standard DNA analysis, the true peaks on a graph of a single contributor’s profile are all relatively close in height. Analysts rely on that fact to exclude peaks of extraneous alleles which will (as with stutter) usually have only a fraction of the height of those of the true alleles. But that is not the case with high sensitivity analysis. The small amounts of DNA available for analysis may not yield consistent peak heights for the true alleles and may result in extraneous alleles at some loci that produce higher peaks than do the true alleles.
OCME was of course aware of all this. Accordingly, if the quantification stage of DNA analysis shows that less than 100 picograms of DNA is present in a sample, different OCME protocols apply. First and foremost, the already small DNA sample is divided into three samples — “aliquots.” While it may seem counterintuitive to divide a small sample, each aliquot is then amplified not 28 times but 31. The analysis proceeds with tweaks intended to enhance its sensitivity. Ultimately, three profiles, one from each aliquot, are created for the analyst. The analyst combines the three profiles into one, but in doing so *607counts an allele only if it appears in two of the three aliquots. In this way OCME hopes to avoid counting drop-in alleles. And OCME expects the three extra cycles to minimize drop-out. Ultimately the combined profile is treated like a profile produced from standard analysis, and is compared to profiles from crime scenes, suspects, and victims. And at a trial, an OCME analyst will offer the same types of conclusions about the probabilities of a match from a random member of the population that he or she might have offered in any standard analysis case. The defense may of course suggest to the jury that there are flaws in high sensitivity methods, but the expert testimony will be before the jury.
At the Frye hearing, the People produced very impressive witnesses to attest to the reliability of high sensitivity analysis performed under the OCME protocols. The architect of the program, Dr. Theresa Caragine, has her Ph.D. from the Sackler Institute of Biomedical Sciences at the New York University (NYU) School of Medicine. She also has graduate and postgraduate training in statistics. The doctor began work at OCME in 1991 as a criminalist performing DNA tests and was eventually named a Deputy Director of the Forensic Biology Laboratory, in charge of, inter alia, supervising the validation and performance of the high sensitivity work there.
Dr. Mitchell Holland received his Ph.D. in biochemistry from the University of Maryland. After extensive DNA work with the Bode Technology Group he moved to the Armed Forces DNA Identification Laboratory, where he became the director and employed high sensitivity analysis. After serving as director of the Bode laboratory, Dr. Holland moved to Pennsylvania State University. There he teaches molecular biology and is director of the Forensic Science Program. At Penn State the doctor in fact teaches classes about high sensitivity DNA analysis. He has also written extensively, and has conducted workshops internationally, on the subject of high sensitivity analysis.
This court finds that both witnesses are extremely skilled professionals who offered honest opinions.5 They believe that high sensitivity analysis is reliable, and indeed that it is not *608“novel.” Dr. Caragine testified at length about the development of OCME’s high sensitivity testing, and in particular about the validation studies performed before such testing was approved for case work.
Those studies produced results that led the DNA subcommittee of the New York State Forensic Science Commission to give unanimous approval to the use of high sensitivity analysis in case work. The subcommittee was composed of seven highly regarded experts in various aspects of DNA analysis, and the recommendations of this subcommittee are binding on the Commission. The People note as well that over a dozen scientific articles about high sensitivity DNA testing have been published, and that OCME scientists have both published articles on OCME’s techniques, and given over 85 presentations about those techniques at scientific gatherings.
But obviously, our trek does not end with the People’s presentation. The defense presented expert testimony as to the high sensitivity protocols as well. First and foremost, the court heard from Dr. Bruce Budowle, currently the executive director of the Institute of Applied Genetics at the University of North Texas Health Science Center. In 1979 Dr. Budowle received his Ph.D. in genetics from Virginia Polytechnic Institute and State University. After three post-doctorate years studying genetic risk factors for certain diseases, he moved to the Federal Bureau of Investigation (FBI). In 16 years at the FBI Laboratory, Dr. Budowle performed work that arguably entitles him, as much as anyone, to be considered the father of American DNA analysis. During his long career Dr. Budowle has served as the chair of SWGDAM6 and the chair of the DNA Commission of the International Society of Forensic Genetics. He is also one of the original architects of CODIS, the national DNA database.
Dr. Eli Shapiro’s Ph.D. in biology is from Yale. After working in a biology laboratory he moved to OCME in 2000. He eventually became the Director of Training at the laboratory, retiring in 2011. As the People note, he does not have the research experience of other hearing witnesses, and he has not published scholarly articles about DNA analysis.
*609Dr. Angela Van Daal, a molecular geneticist, received her Ph.D. from Macquarie University in New South Wales. She did post-graduate study in the United States and then, for about 14 years, helped introduce forensic DNA analysis to Australia. In 2005, Dr. Van Daal took a position at Bond University, and she has since been affiliated as well with the University of North Texas.
Dr. Ranajit Chakraborty, a population geneticist, received his Ph.D. from the Indian Statistical Institute in Calcutta in 1971. He moved to the University of North Texas in 1993, and is now a full professor there. He has extensive experience with issues in DNA analysis. Dr. Chakraborty was a member of the DNA subcommittee of the New York State Forensic Science Commission when that group approved OCME’s high sensitivity analysis, and its FST software as well. He has since changed his mind.
Finally, Dr. Heather Coyle received her Ph.D. in plant biology from the University of New Hampshire in 1994. She is an associate professor in the Forensic Science Department at the University of New Haven and founded a consulting company in Connecticut that provides services to the defense bar concerning the analysis of DNA and other biological evidence. Dr. Coyle’s academic publications do not include articles about high sensitivity DNA analysis.
The People’s witnesses addressed OCME’s validation studies of its high sensitivity procedures and otherwise explained why they think OCME’s high sensitivity analysis is robust, and not novel. The defense witnesses were not of the same point of view, and some of their objections to OCME’s procedures will now be discussed.
We begin with an issue that impacts in particular on high sensitivity analysis of samples containing mixtures of DNA from two or more contributors, which are very common in high sensitivity work. When a graph of DNA alleles in a sample is prepared, alleles are represented on the graph by peaks placed along a base line. The peaks are taller or shorter depending on how much DNA at that locus, and that allele, is present. In standard analysis, each contributor’s alleles are likely to produce peaks of roughly the same height; a minor contributor’s peaks will be shorter than those of a major contributor. That can permit an understanding of which alleles are to be assigned to various individual profiles. That advantage is lost with OCME’s method of analyzing small DNA samples. The *610three extra amplifications, produced by DNA kits not designed for them, create stochastic effects that make peak heights unreliable. Dr. Budowle so concluded.
As to single-contributor samples, Dr. Coyle focused on dropout. It is not disputed that drop-out is more common with high sensitivity analysis, and that drop-out can make a heterozygous locus look like a homozygous one. Or it can make any locus “disappear” with no allele appearing. A person who did contribute to a sample is unlikely to complain. But someone who did not contribute, perhaps an innocent man accused of a crime, may lose the chance to show that someone else’s alleles, and not his, were present at relevant loci. The doctor noted that OCME protocols substitute a Z for a number at a locus where an allele may be missing, and that there are many Zs in profiles from high sensitivity analyses of low picogram DNA samples. But the Z might as well be a question mark, and use of the Z hardly substitutes for exculpatory numbers.
Dr. Coyle also addressed “drop-in” contamination of a sample, the appearance of extraneous DNA material that creates reports of false alleles. Drop-in is not a problem with standard analysis. “Negative controls” are run along with case samples. If the negative controls reveal any peaks due to contamination, the run is disallowed. But the extra three cycles used in high sensitivity analysis make contamination a particular problem. The extra amplifications make drop-in far more likely. Apparently with that in mind, the protocols for high sensitivity analysis permit the analysts to ignore contamination results in negative control tests until 10 drop-in peaks appear in at least two of the three aliquots. Dr. Coyle considered that unacceptable. She added that contaminants can appear in high sensitivity samples even if there is no contamination shown in the negative control samples. The possible appearance of peaks due to contamination makes analysis unreliable.
The defense witnesses also addressed stutter. In standard analysis, in the graph depiction a stutter peak is generally extremely small compared to a “true” peak. Dr. Budowle and others have noted that, with high sensitivity analysis, stutter peaks can be magnified dramatically. They thus can appear to be true peaks when they are not. And even if they can be recognized as stutter, their possibly inordinate peak height can mask the appearance of a true allele at the same position— particularly in mixture cases.
The defense witnesses addressed peak heights more generally as well. As noted, alleles are represented at the correct *611spot on the graph by a sharp peak that, in standard analysis, will be taller or shorter depending on how much of the contributor’s DNA is in the sample. Protocols filter peaks out of the profile if they do not reach a certain percentage of the height of the other peaks. Again as noted, in standard analysis that helps weed out stutter. More broadly, it guards against very low-level contamination. A dozen or three dozen picograms of contaminant DNA will never spoil the graph where the DNA sample contains 300 picograms; the low height of the contaminant’s peak will give it away. That rather critical advantage disappears in high sensitivity analysis, in which the false report may be based on a higher peak than the report of a “true” allele. Relative peak heights may unreliably indicate which are the true alleles at one or many loci, and result in incorrect profiles.
As a result of issues like these, no public laboratory in the United States, other than the OCME lab, employs high sensitivity analysis to develop profiles for use in criminal cases. Among the labs refusing to use high sensitivity analysis is the FBI laboratory. Moreover, CODIS, which is run by the FBI and contains the national DNA database, will not upload profiles created with high sensitivity analysis. Some laboratories, including Dr. Budowle’s lab at the University of North Texas, will use high sensitivity analysis for limited purposes. For example, after a disastrous accident like an airplane crash, high sensitivity analysis of bodily remains can be used to identify the victims of the “closed” population of possible contributors. But that is because the population is limited, and because the remains, for example bones, can be cleaned before the analysis is done. And, unfortunately, a mistaken analysis can be of no consequence to the contributor.
Except for OCME, then, no American laboratory produces high sensitivity conclusions for use as evidence in a criminal case. As Dr. Budowle notes, that does not mean that high sensitivity analysis must be considered totally irrevelant in criminal cases. Such analysis can produce “investigative leads.” Critics of high sensitivity analysis agree that if a DNA profile created through high sensitivity analysis suggests that a particular individual is the perpetrator of a crime, that profile can legitimately point investigators at the suspect. In that regard, the results of some other techniques — polygraphs and facial recognition software, for example — likewise can aid an investigation, but are not considered sufficiently reliable to be admissible at a trial.
*612This court initially wondered why the criticisms of high sensitivity analysis were not matters of weight to be considered by the jury — particularly since even defense witnesses like Dr. Budowle acknowledge that a profile produced by such analysis can be of value. Ultimately, however, that thought is trumped by Frye. The products of polygraph technology and of facial recognition technology similarly can sometimes have value, but evidence produced by those technologies is not generally accepted as reliable by the relevant scientific communities and so cannot be admitted in trials. The same should be true, at least at this time, for high sensitivity analysis. After all, if the experts in the DNA field cannot agree on the weight to be given to evidence produced by high sensitivity analysis, it would make no sense to throw such evidence before a lay jury and ask the jurors to give the evidence appropriate weight.
The People insist, however, that the relevant scientific community does accept high sensitivity analysis. It is true, as the People note, that OCME’s procedures have been described in peer-reviewed articles and in discussions at gatherings of scientists. But this court cannot accept the thesis that publication and discussions equate to general acceptance. Not only the impressive defense witnesses indicate otherwise, so too do the many peer-reviewed articles submitted as defense exhibits which question OCME’s procedures. And, as the defense notes, after all this discussion of high sensitivity analysis, no other laboratory has employed it for use in criminal cases. This court simply cannot conclude that there is a general consensus in favor of high sensitivity analysis, in the face of this contrary evidence.
The People have a more specific argument that decidedly deserves attention. High sensitivity analysis was approved by the DNA subcommittee of the New York State Forensic Commission. The conclusions of that subcommittee are binding on the Commission, and so the subcommittee numbers are the true decision-makers. And the members of the subcommittee are world-class scientists in various disciplines relevant to DNA analysis. The subcommittee approved OCME’s high sensitivity procedures, and the People suggest that this is very strong evidence of general acceptance in the relevant scientific community.
This court does not agree. It is not just that Dr. Chakraborty, one of the members of the subcommittee, has “defected,” and now has testified for the defense. The more important point is *613that no state subcommittee can be equated with the general membership of the relevant scientific community. Will we next consider the matter closed, because the members of a committee in Idaho or Florida approve of a procedure? This court knows that the members of the DNA subcommittee are indeed experts in their particular fields and that their opinions are valuable. They simply are not determinative.
Nor does the court agree with the People that the OCME validation studies and the audits at the OCME laboratory by outside reviewers are conclusive. Every laboratory validates techniques and procedures before implementing them. But a laboratory’s satisfaction with its validation results does not show general acceptance of techniques and procedures, if the validation studies fail to create such general acceptance. And OCME’s validation studies have failed to create general acceptance of high sensitivity analysis. As to audits, they appear to test whether procedures are being implemented in accordance with protocols, not whether the principles underlying the procedures are valid.
IV. The Forensic Statistical Tool
A.
In standard DNA analysis, and emphatically in high sensitivity DNA analysis, analysts encounter samples that are a mixture of the DNA of two or more contributors. These mixtures frequently present problems. Especially in standard analysis, the peak heights of one contributor may stand out, and thus readily distinguish his alleles from those of the one or more other contributors. But it is often the case, especially with relatively small contributors seen in high sensitivity analysis, that the sample contains a soup from which each individual’s alleles cannot be separated out and placed in a profile.
As a result, in the past analysts often could draw only general conclusions from a mixture. For example, a mixture containing three or four alleles at each of the select loci could be called as a two-person mixture. If all of a suspect’s alleles were present in that soup, the analyst could say that the suspect could be a contributor to the mixture. If many of the suspect’s alleles were missing, he could be pronounced a non-contributor. But if one or a few alleles were not detected, perhaps as a result of degradation or simple drop-out, all that the analyst could say was that the suspect could not be *614excluded as a possible contributor, or that no conclusion could be drawn. No statistic for the probability of inclusion could be generated.
And so OCME created the FST. The FST is a computer program that calculates a “likelihood ratio” derived from a fraction. The numerator of the fraction represents the chance that the prosecution hypothesis is true — that a particular individual was one of the contributors to a mixture. The denominator represents the chance that the defense hypothesis is true— that other random individuals, and not the one of interest to the prosecution, were the contributors. Division of the numerator by the denominator produces the likelihood ratio. That ratio could indicate, for example, that a three-person mixture is 100,000 times more likely as the result of contributions by the targeted individual and two random, unrelated individuals, than as the result of contributions by three random, unrelated individuals. Or the prosecution hypothesis might be undercut —it might be, for example, that the mixture is only one third as likely as the result of contributions by the targeted individual and two random, unrelated individuals than as the result of contributions by three random, unrelated individuals. The enormous value of such statistical results, compared to simple statements like “the individual cannot be excluded as a contributor” is obvious — if the statistics are accurate.
To arrive at the likelihood ratio, the FST employs Bayesian mathematics. As noted above, Bayesian probability calculations have been made in science for centuries, and no one disputes the mathematical principles involved. The FST simply performs the analysis more swiftly (by far) than a human could. It takes the probabilities of the prosecution hypothesis at each of the select loci and combines them into an overall number. It then does the same for the probabilities of the defense hypothesis at each select locus, and divides the first number by the second to create the result.
The process, to that point, is so non-controversial that many courts have stopped there, finding that there is nothing novel about the FST and thus that there is no basis for a Frye attack upon it. But that does no justice to the actual positions pressed by the defense. The key advance in programs like the FST is that they factor into the Bayesian calculations the likelihood that alleles have appeared or failed to appear as a result of stochastic effects. The defense contends that the manner in which the drop-in and drop-out rates are assessed at each locus *615is not generally accepted in the DNA community. The defense further argues that the FST wrongly limits analysis to a numerator and a denominator each reflecting only a single hypothesis, and thus unscientifically prevents alternative analyses.
As to the first complaint, for each locus OCME has specified the probability of drop-in and drop-out in mixtures of different DNA amounts (quants). The FST uses the figure at each locus and the particular quant in determining the probability of the prosecution and defense hypotheses. For example, if one of the targeted individual’s alleles is not present at a particular locus, the FST program considers the chance, given the quant, that this is the result of drop-out; if the targeted individual’s alleles are among those present at a locus, the FST program considers the chance that this is the result of drop-in contamination.
OCME calculated the probability of stochastic effects at most loci the old-fashioned way: they counted. The laboratory analyzed DNA mixtures from known contributors at several quants, and counted how often drop-in and drop-out occurred. But the numbers were modified to a certain extent to resolve divergence from expected patterns in the results. The numbers were also reduced by one standard deviation in an effort to be “conservative” — to err on the side of producing lower likelihood ratios than the actual counting data would have produced. At two loci, for reasons not relevant here, the probabilities of stochastic effects were calculated, not by counting after actual DNA analysis, but through computer simulations. These procedures create issues for some scientists.
As noted, the FST utilizes different figures for the probability of stochastic effects based on the amount of DNA in a mixture sample. A related complaint about the FST’s stochastic calculations concerns that fact. The volume is determined at the quantification step of DNA analysis. The quant calculation is not particularly precise; the numbers developed can be as much as 30% greater or lower than the true quant. Such discrepancies are unimportant with samples large enough to be evaluated with standard analysis. They can be far more significant in samples that are less than 100 picograms in weight.
The defense also complains that the FST is a “black box”— that is, that OCME has not published the FST program and is the sole entity able to employ it. As a result, a likelihood ratio can be deduced only for the prosecution hypothesis and the defense hypothesis propounded by the OCME analysts. A *616defense expert cannot, for example, obtain a likelihood ratio based on a hypothesis that there were a larger or smaller number of contributors to the mixture than OCME supposes, even though the number of contributors is often subject to reasonable dispute. A defense expert cannot determine a likelihood ratio based on a hypothesis that contributors are related. A defense expert cannot determine a likelihood ratio based on a different quant estimate. A defense expert cannot obtain likelihood ratios based on how much an individual of a mixed race might differ from a person who is not.
The People produced impressive expert testimony in support of the design and the testing methods underlying the FST. The principle architects were Dr. Caragine and Dr. Adele Mitchell. Dr. Mitchell owns a Master’s degree in statistical genetics and a Ph.D. in human genetics and molecular biology from Johns Hopkins School of Medicine; her doctoral thesis focused on the effects of drop-in and drop-out on the statistical analysis of alleles. She has taught in her field both at Mount Sinai School of Medicine and the NYU School of Medicine. She and Dr. Caragine received the Frederick O’Reilly Hayes Prize, an annual award for outstanding New York City employees, for their work on the FST.
Also testifying for the People was Dr. Hilda Haned of the Netherlands Forensic Institute in The Hague. Her Ph.D. studies at the University of Lyon were in statistical methods for analyzing DNA mixtures. Dr. Haned worked with Dr. Peter Gill, the DNA pioneer in Britain, to develop a likelihood ratio program modeling drop-out, and she developed a likelihood ratio program for her own laboratory as well. Dr. Haned had studied the FST and considered it a reliable method for determining a likelihood ratio. She was particularly impressed with OCME’s use of quant to calculate which drop-in and dropout probability statistics to employ. Dr. Haned hoped to implement the quant method in her own laboratory.
B.
But the defense presented experts as well. Dr. Shapiro disagreed with an FST protocol, intended to be “conservative,” that might underestimate the number of contributors to a mixture and with another that could lead to a false understanding of the drop-out rate. He criticized the fact that OCME’s validation studies included quant estimates with a “plus or minus” range of even more than 30%. The doctor also contested *617other aspects of the FST programming such as its failure to consider relatedness and its methods for determining the likelihood of alleles in members of various races.
Dr. Rori Rohlfs, who received her Ph.D. from the University of Washington, is a post-doctoral fellow at the University of California at Berkeley, and is a population geneticist. Dr. Rohlfs’ testimony focused on the FST’s false positive tests.
Dr. Noah Rosenberg’s Ph.D. in biology is from Stanford University, and he is now a population geneticist and statistician there. His post-doctoral studies at the University of Southern California were in molecular and computational biology. Dr. Rosenberg’s testimony also was in large part centered on the false positive testing of the FST.
In addition, Drs. Coyle, Van Daal, Budowle, and Chakraborty offered criticisms of the FST. These defense witnesses had many issues with the program. Dr. Budowle agreed that the FST is “novel,” and indeed unique, in how it determines which drop-in and drop-out rates to use. He also believed that this was a problem. OCME never formally tested the theory that quant could reliably determine drop-in and drop-out rates and Dr. Mitchell’s “exploratory” tests on that front were not documented. Perhaps more importantly, in Dr. Budowle’s view, the validation studies based their drop-in and drop-out percentages on the stochastic effects appearing in studies of “pristine” DNA samples created in the laboratory. That was no indicator of what the results would be in the real world, in which DNA samples, especially small samples, degrade over time and to a degree that is based on the circumstances of the case. Notably, different alleles in a sample may well degrade at different rates and as a result a uniform overall quant estimate may, even apart from the problem that it is just an estimate, mask the fact that some alleles may be far more subject to stochastic effects than others.
Next, the defense challenged the assumption by the FST architects that drop-out rates at various quant levels would increase in a linear fashion as the quant rates decreased. When the counting of drop-out was done, the assumption proved not necessarily to be the case. Nonetheless, the OCME numbers were changed from the “counted” results, to reflect the expected “linear” results. This court does not suggest for a moment that these changes were anything but the result of objective scientific judgment. But Dr. Budowle and other scientists noted by the defense do not agree that the OCME assumptions are necessarily valid.
*618Beyond that, Dr. Budowle and others differ with OCME as to the use of the same drop-in and drop-out rates in the numerator and denominator to create the fraction which becomes the likelihood ratio. For example, Dr. Gill and prosecution witness Dr. Haned have published an article opining that the denominator, the “defense hypothesis,” should reflect the views of the defense. This is particularly important given Dr. Mitchell’s testimony that, with very small DNA samples, drop-out can be extremely high.
Further, the fact that the FST software is not open to the public, or to defense counsel, is the basis of a more general objection. This court understands the City’s desire to control access to computer programming that was developed at great cost. But the FST is, as a result, truly a “black box” — a program that cannot be used by defense experts with theories of the case different from the prosecution’s. The prosecution will present a likelihood ratio based on assumptions — for example, that other possible contributors to a mixture are unrelated to a suspect. In the defense hypothesis, there is no information for the jury about whether the suspect is more or less likely than his brother to have been a contributor — even in a case in which the identity of the rider of a bicycle is in dispute.
Similarly, the FST gives numbers that are based on a conclusion about how many people contributed to a DNA mixture. But a mixture, and especially a mixture that will be classified as one requiring high sensitivity analysis, will present a challenge to one trying to determine how many people contributed to it. The fact that, for example, there are at most four alleles at each locus does not mean, necessarily, that a third person’s alleles are not in the mixture. Likewise, that there are anywhere from three to six alleles at each locus does not mean that four individuals’ DNA is not present. But the “black box” nature of the FST prevents any defense attorney from informing the jury of the likelihood ratio, should the prosecution estimate of the number of contributors be incorrect. The jury will hear only one number: the one that is produced by “the program” as it assesses the prosecution hypothesis, and a dictated so-called defense hypothesis.
Some experts would add to this “black box” criticism. OCME’s laboratory, like others, sets threshold levels for peak heights below which an allele will not be recognized. Acknowledged experts like Dr. David Balding and Dr. Mark Perrin were *619not witnesses at the hearing, but have opined that possible alleles should be considered even if below a set threshold, and that conclusion is especially weighty where high sensitivity analysis, and questionable peak heights, are involved.
Next, as noted, Drs. Rohlfs and Rosenberg spoke to the validation study testing for “false positive” hits. Dr. Rohlfs noted that the FST recognized four “races” — Asian, European, African, and Latino — and used different numbers for allele frequency for each. In the absence of data from OCME on which “false positive” test subjects identified themselves (accurately or not) as being in those categories, Dr. Rohlfs assigned particular test subjects to those racial categories based on how their individual likelihood ratios accorded with the “race” classifications. Dr. Rosenberg agreed that this method was acceptable and the best that could be done in the absence of OCME data.
What Dr. Rohlfs then discovered seemed to her to be problematic. First, only three Asian subjects contributed to the various DNA mixtures. Two were used in mixtures a few times, and one 60 times. Dr. Rohlfs did not believe that this allowed fair conclusions about error rates for the general Asian population. Whoever the three subjects were, there is no reason to expect they fairly represent genetic diversity from Hokkaido to Ceylon and from Java to Mongolia.
Second, Dr. Rohlfs thought there were simply too few mixtures (480) from too few contributors (61) for adequate statistics to emerge. Had she designed the “false positives” test she would have included hundreds of contributors in hundreds of thousands of simulated mixtures. Critically, only 11 of the 480 mixtures involved contributors who were all of the same race. That is very much a problem, as it is precisely when contributors are of the same race that alleles overlap, and false positive hits, are most likely.
Finally, and perhaps most fundamentally, Dr. Rohlfs thought the four-race classification scheme was poorly conceived. There was no thought for individuals of mixed race. There was no thought of other populations that are possibly genetically distinct from the four groups considered — the court might posit Native Americans, Semites or natives of the Aran Islands. No thought was given to misleading impacts on likelihood ratios from the existence of a distant ancestor, even one from generations in the past, who was common to a criminal and an innocent suspect. Such an ancestor’s genes could well skew a *620likelihood ratio dramatically. The jury could simply hear that a mixture was ever-so-many times more likely if defendant was a contributor and hear nothing about these and other possible variables.
Dr. Rosenberg stated that proper peer review of a software package like the FST requires the submission of that package to the independent experts — something not done by OCME with the FST. Dr. Rosenberg further opined that publication in peer-reviewed journals is not the equivalent of general acceptance in the relevant community. That must be judged from the results of publication.
As to the FST’s protocols, Dr. Rosenberg disagreed that analysts should simply make an estimate of the number of contributors to a mixture, and limit analysis to that estimate. Mixtures often cannot be said definitively to have been produced by two, or three, or four individuals. And a mistaken estimate comes with serious consequences. Mixtures created by three or four contributors result in four times as many false positive assessments as do mixtures created by two contributors.
Further, other variables that lead to false positives were not identified and assessed by OCME. The FST factors in suppositions about race and the number of contributors to mixtures. It has not factored in other, unidentified variables that tend to result in false positives. And of course, there is no way for a defense expert to perform calculations of likelihood ratios that take into account either different views about OCME’s assessment of race and the number of contributors or other possible variables that might be relevant to a particular case.
Put another way, in Dr. Rosenberg’s view, the FST does not recognize that individuals may constitute the ultimate variables. People’s alleles drop out at different rates, for reasons having no apparent connection to race. By assigning a likelihood ratio based on which of four “races” a person supposedly belongs to, the FST sweeps up all purported members of that race with a particular assumption that is not generally valid.
Finally, on the race front, Dr. Rosenberg agreed with Dr. Rohlfs’ view that the FST’s “Asian, European, African, and Latino” categorization was inadequate. And he agreed in particular that the possible use of only three Asians, one far more than the other two, was not a valid way to determine false positive rates for people with Asian ancestry.
*621c.
The prosecution answer to criticisms seems to be that the FST formula is “conservative” and automatically reduces the likelihood ratio sufficiently to compensate. But, there are no studies which show that the FST consistently underestimates the proper ratio, especially when an innocent suspect is thought to be a contributor. It may even be assumed that a true contributor’s likelihood ratio is dropped by, for example, lowering the results by one standard deviation. But it would not follow, in the absence of pertinent testing, that the same would apply to a non-contributor, who might well benefit by not having the results reduced. There is no data, either way. Dr. Budowle had more general doubts on this “conservative” issue as well. He testified that “there is a certain sweet point . . . .” Depending on the circumstances of the case, if the drop-out rate is high it will give a conservative result, but at a certain point “it will go in the opposite direction” (testimony of Dec. 9, 2013 at 829-830).
A possible response to the “black box” criticisms is that the defense can call its own experts, and can cross-examine the People’s witnesses about the fact that alternative analyses are not considered. This court does not agree with the response. An OCME expert may come to court and say that, based on his analysis, a mixture is 1,000,000 times more probable if a defendant is a contributor than if he is not. It is little comfort to the defense that the defense attorney can ask, “well, what if there were three contributors, not two,” and have the expert respond that there were only two, and that the results if there were three are unknown. Again, it is inconsistent with Frye to present scientific evidence that is not generally accepted and expect a lay jury to give it proper weight.
There is much more that can be discussed. Heavy objections were made to OCME’s statistics on false positive tests. But this court need not go beyond the issues already addressed. The evidence on the other matters is in the record, and can be reviewed in future litigation. This court concludes, based on that record, that the FST is not generally accepted in the DNA scientific community.
V. Reargument
The parties rested in December 2013, and prepared briefs for the court. This court announced its decision orally on November 7, 2014, before its written opinion was completed. The People *622almost immediately moved to reopen the hearing and to reargue. The People relied on what they considered to be new information that was not available when the parties rested. The defense has responded and the People have replied to the response. This section of the opinion will address the parties’ positions as to the reargument application.
As a preliminary matter, the defendants argue that the court should, for various reasons, simply deny reargument. Certain of defendants’ arguments may have technical merit, but this court agrees with the People that it would be unwise for this court to ignore their new submissions. These Frye proceedings have lasted for over two years. It would make no sense for this court to say that the People’s new evidence and arguments cannot be considered until another court expends the resources to duplicate what has been done here. However, defendants can take some consolation from the conclusion which the court reaches on reargument: defendants’ motions will still be granted.
A. The DNA Subcommittee
As noted above, in 2005 the DNA subcommittee of the New York State Forensic Science Commission approved the use of high sensitivity DNA analysis under the protocols promulgated and validated by OCME. This court did not find this dispositive of the issues. The People now report that, in 2014, the current members of the DNA subcommittee addressed questions more recently posed by the Forensic Science Commission. The subcommittee advised the Commission in September 2014, that there have been no material procedural changes in OCME’s high sensitivity procedures since 2005.
The People consider this to be a new endorsement of high sensitivity analysis. The court disagrees, concluding from the 2014 report only that the current members have found OCME’s practices to be materially unchanged since 2005. There was no new consideration of the reliability of high sensitivity analysis. This court’s initial decision was that the endorsement of high sensitivity analysis by the subcommittee could not be conclusive on what the scientific community as a whole believes. That conclusion remains unchanged.
B. Other Programs Assess Mixtures
The People assert that there now exist at least eight software packages which, like the FST,. state likelihood ratios as to the *623probability that a defendant is a contributor to a DNA mixture and that the defendant is not a contributor.
The People misunderstand the nature of the court’s ruling as to the FST. By no means did this court suggest that there is anything wrong with using likelihood ratios or that no method for calculating likelihood ratios can be created that will gain general approval in the scientific community. What is at issue is whether OCME’s FST program in particular, developed as it was, has general approval. That other programs are on the market has nothing much to say about that.
In that regard, it is important to remember the ways in which the FST is different from most, if not all, of the other programs. For example, it is very significant that the FST gives one ratio — OCME’s ratio. The defense cannot “tweak” the program with alternative hypotheses that may be reasonable, to see if different results emerge. Other programs now identified by the People are more flexible.
Moreover, as noted above, OCME utilized a number of procedures to produce the FST program that differ from those used to create other programs. For example, at two loci, OCME departed from its own general “counting” techniques. In addition, the FST uses quant, rather than peak heights, to determine which drop-in and drop-out rates to utilize. That practice is controversial, and it seems to remain correct that no other program for assessing mixtures utilizes “quant.” And, as noted above, OCME’s particular and unique methods for assessing drop-in and drop-out rates are in dispute.
The People assert that a prominent DNA expert, Dr. John Buckleton, has opined that the FST is as sound as other mixture programs. Dr. Buckleton’s view is plainly relevant, but does not by itself change the calculus concerning “general” acceptance. And, for the reasons noted by the defense, the doctor’s views do not appear to be based on a solid familiarity with the FST.
Considerations such as these cause respected scientists to withhold approval of the FST. That other, different programs are on the market does not change that. Nor is the court impressed with the argument that the FST “quant” approach to choosing drop-in and drop-out rates, and the alternative “peak heights” approach, both involve estimates. That one method of making an estimate is generally considered reliable does not show that an alternative method is acceptable as well. And the court’s difficulty with the “fixed parameters” of the *624FST drop-in and drop-out rates is fundamentally not only that they are “fixed” but with the method through which they were set.
C. The Practice of Defense Counsel
The People note that defense advocates are happy to rely on high sensitivity and FST testing results when those results do or might exculpate their clients. The People assert, for example, that the Innocence Project “routinely” asks OCME to perform high sensitivity DNA testing for convicted defendants, and that four exonerations had resulted by the time of the People’s re-argument motion. The court finds the argument of small moment. First, the advocate for a convicted defendant has little to lose by requesting DNA analysis even under a procedure that is not generally accepted as reliable, if there is a chance that testing will produce exculpatory results.
Second, the court has by no means concluded that high sensitivity analysis cannot show that a suspect is excluded as the contributor of a DNA sample. Nor is the court even saying that high sensitivity analysis will never correctly identify a contributor; defendant Peaks, for example, is inculpated by high sensitivity results that are consistent with a standard DNA sample as to another victim. And a lie detector may sometimes detect a lie. The court’s conclusion is simply that high sensitivity results are not now generally accepted as reliable in the relevant community.
Nor does the court find it troublesome that Legal Aid Society attorneys seek to introduce high sensitivity and FST evidence in other cases in which the testing favors the defense. That defense advocates would make inconsistent legal arguments in an unsettled area, on behalf of clients whose legal interests differ, comes as no shock to the court. It would hardly surprise the court to learn, for example, that in one case an attorney would seek to admit exculpatory polygraph or bite mark evidence, while opposing the admission of such evidence in a case in which it is inculpatory. This court’s views would require the exclusion of the types of DNA evidence contested here in any case in which either side objected to it.
D. The FBI and SWGDAM
CODIS and the FBI do not permit high sensitivity DNA profiles to be compared to profiles in the CODIS data bank. *625Since the parties rested, SWGDAM has issued guidelines for the potential validation of, and for quality assurances of, high sensitivity DNA analysis. The People in effect argue that it is only a matter of time before high sensitivity profiles can be uploaded for CODIS searches, and that this is further proof that OCME’s high sensitivity analysis is now generally accepted.
The court first makes what it supposes is the most obvious response: the time has not yet come, and there is no way to know whether OCME’s procedures will ultimately be accepted. Indeed, the SWGDAM position expressly is that it has not offered an opinion on “the viability of” high sensitivity testing. The possibility of a future SWGDAM/CODIS endorsement does no harm to the People’s position. At the same time, a simple possibility adds nothing to it.
The People add that the FBI occasionally contracts with OCME to have high sensitivity analysis performed. At least one local federal judge has found that OCME high sensitivity results are admissible (see United States v Morgan, 53 F Supp 3d 732 [SD NY 2014]) — albeit under the Daubert admissibility test, which is less rigorous than the Frye test. But the FBI laboratory still does not perform high sensitivity testing. That FBI personnel devoted to making cases are content to accept new DNA procedures tells this court little, when those procedures are not accepted by the scientists in the FBI laboratory.
The People note as well that CODIS search protocols are difficult to apply with OCME’s high sensitivity results and that this practical difficulty, not scientific issues, is responsible for the CODIS refusal to look for matches of high sensitivity samples. When OCME high sensitivity analysis identifies only one allele at a locus, OCME assigns a “wild card” — a “Z” — as the second allele, because of the real possibility that the second allele has dropped out. It could be that, at this locus, the contributor is instead simply a homozygote, with the second allele being the same as the first. Or it could be that any other possible allele was present, but dropped out. CODIS searches therefore return “hits” for all contributors in the database with the first allele and any other possible allele. The result is often a very large number of possible profile matches at the allele in question. Because high sensitivity samples produce drop-out at so many alleles, the court is told, this aspect of CODIS searches is problematic.
*626If high sensitivity test results, with their relatively high drop-out possibilities, are too problematic for the CODIS computers, that hardly speaks in favor of high sensitivity analysis. But this court cannot conclude that the Z factor would prevent the CODIS experts from trying to match high sensitivity samples if in fact OCME’s methods were considered reliable. A search could be done only on those loci in a sample where two alleles were found. Such “partial” searches are common where two alleles are found at 10 or more loci, and can yield strong evidence for or against a suspect.
VI. The Prior Case Law
This opinion will close with a discussion of five prior court opinions. None is controlling. Our slate is not entirely clean, but no New York appellate decision on the relevant issues has yet been written. Three cases deal with high sensitivity analysis, and two with the FST.
A.
The leading New York decision on high sensitivity analysis, and apparently the lead decision in the country given OCME’s unique position in that field, is a decision by Justice Hanophy. (See People v Megnath, 27 Misc 3d 405 [Sup Ct, Queens County 2010].) The holding is that OCME’s high sensitivity procedures are generally accepted in the relevant scientific community. This court, as noted, disagrees.
Megnath pointedly refuses even to acknowledge that high sensitivity analysis involves “novel” scientific procedures that would require Frye review. In light of the evidence produced for this court by very diligent attorneys on both sides, that conclusion seems impossible to defend. None of the witnesses at the hearing in this case, and none of the articles introduced as exhibits, suggest that the issue in this case is a “gimme.” Even the People’s experts were quick to acknowledge that moving from standard DNA analysis to high sensitivity analysis means crossing a border to a different world.
In particular, the Megnath opinion is dismissive as to the increase in stochastic effects that is the natural result of employing 31 cycles with software designed for 28 cycles. That, as the opinion notes, there can be stochastic effects in any DNA analysis hardly suggests that the substantial increase in such effects in high sensitivity analysis is irrevelant. And of course it does not suggest that the expert opinions explaining *627why these effects are not adequately dealt with by OCME’s protocols can be ignored.
To similar effect is United States v McCluskey (954 F Supp 2d 1224 [D NM 2013]). The McCluskey court excluded a primitive high sensitivity analysis — though, concededly, on facts enormously different from those of this case. In McCluskey, a laboratory did analysis of DNA samples with fewer DNA programs than usual, with no changes in their usual procedure. When the McCluskey court analyzed the OCME procedures described in Megnath, it noted that in several ways they were more sensible than those used in New Mexico. But the opinion ended by criticizing Megnath for suggesting that high sensitivity analysis does not involve new science. And the court also disagreed with the apparent and illogical Megnath view that the increased stochastic effects in high sensitivity analysis are, basically, no matter of concern, because there are stochastic effects even with standard analysis.
This court rejects both of those thoughts from Megnath. So do, it seems, all the experts who gave relevant evidence for either side.
B.
The leading New York decision on the FST is Justice Carruthers’ opinion, not officially reported, in People v William Rodriguez (Sup Ct, NY County, indictment No. 5471/2009, Oct. 24, 2013). Rodriguez concludes that the FST is generally accepted in the relevant scientific community. But the opinion is focused on matters not in controversy, such as the general acceptance of PCR-STR analysis; of the use of likelihood ratios in the evaluation of DNA mixtures; and of taking account of drop-in and drop-out rates in calculating likelihood ratios. Moreover, the court deprecated the “counting” of scientists’ “votes” on disputed points, giving extreme deference to the view of New York’s DNA subcommittee and of its chair, Dr. Ballantyne. This court has already noted its view that the subcommittee’s conclusions are by no means binding in Frye analysis.
Further, the Rodriguez opinion pays no attention to what, at least at this court’s hearing, was a major focus of scientific contention: the manner in which the likelihood of stochastic effects at the relevant loci was computed for the FST. Similarly touched on only lightly was the pronounced controversy over the FST’s use of “quant,” rather than peak heights, to *628determine which figures to employ in assessing the likelihood of drop-in and drop-out. In its brief remarks on the subject, the court in fact simply and inappropriately decided the scientific issue in the People’s favor. Finally, the opinion concluded that it is a positive asset to the FST that it remains a “black box,” and that it can examine only the prosecution’s hypothesis; this court has already mentioned its view to the contrary.
This court ends with a word about another opinion upholding use of the FST — People v Belle (47 Mice 3d 1218[A], 2015 NY Slip Op 50663[U] [Sup Ct, Bronx County, Apr. 29, 2015]). The Belle court first notes, as so many other courts discussing Frye and the FST have done, that Bayesian analysis is universally accepted and that the FST simply performs Bayesian calculations faster and more reliably than a human can. But the Belle court, to its credit, does not stop there, recognizing that other aspects of the FST are also in issue.
Still, the court then jumps over those aspects without discussion because the defense in that case performed its own FST-like calculations and arrived at a likelihood ratio very different from OCME’s — though still highly inculpatory. The court’s conclusion is that the FST methodology is no longer at issue in the case, and that a jury should hear the experts and decide which highly inculpatory likelihood ratio is correct. But the conclusion does not follow. Of course, the defense can maintain its legal challenge to the FST, and upon losing still take the fail-back position that the less damaging ratio should be considered.7
The preceding analysis was complete, but not yet in full opinion form, on November 7, 2014, when the court orally announced its views. The People almost immediately asked to present additional evidence of developments since the People’s case concluded. The defense objects to this court considering the People’s new submissions, but the court believes that under the circumstances of this case the People’s reargument application should be granted. This court accepts the additional information, and will simply add that the People’s additional considerations do not change the result.
*629This court concluded that evidence derived both from high sensitivity analysis and from the FST is not yet proved to be admissible under the Frye test. The court is not happy with that result. This court has heard for years about the high sensitivity initiative, with all of the incumbent expense. And this court understands as well the sincere effort that Dr. Caragine, Dr. Mitchell, and many others have put into the development of the FST. They must continue, if they are to persuade.
People v. Rodriquez People v. Rodriquez
(N.Y. Sup. Ct. 2013)
This case is unpublished so is posted in pdf form on Moodle under this week's class number.
In this case - litigated by Bronx Defenders and cited by the court in Collins - you can see a judge misapplying Frye so as deny the defense challenge to FST, the computer program used by New York's Office of the Chief Medical Examiner for analyzing DNA mixtures.
End guilt by genetic association in N.Y. End guilt by genetic association in N.Y.
Allison Lewis, New York Daily News (May 24, 2022)
This Op-Ed by our guest speaker highlights the problems – and the disparate racial impact – of familial DNA searches.
[This article is available online here and is pasted in below.]
They say you are the company you keep. Should that apply tenfold for the family you’re born into? Law enforcement certainly thinks so. But if guilt by association still strikes a chord, we should all be concerned with familial DNA searching, because it is guilt by genetic association.
Imagine learning you are the target of a police investigation for a violent crime, not based on suspicion, but because a close relative was convicted of a crime and you share DNA. Familial searching scans the convicted offender database not for full matches, but for close non-matches to crime scene DNA, hoping someone in the database is related to the person of interest. From there, anyone in their family tree is a possible suspect, and many innocent people may wind up targeted by police.
Once you become a viable target (meaning, not an infant or in jail at the time of the crime), the cops may nab your DNA profile from a discarded soda can while you’re not looking. Hopefully, the DNA will establish your innocence (though consider Lukis Anderson, whose DNA was recovered from the fingernails of a murder victim but luckily had an airtight alibi), and hopefully they won’t store it forevermore (though consider the unregulated New York City database, which stores thousands with little oversight).
Then again, the investigation could take a more perilous turn. The police question your neighbor or boss. What if you come to find out your baby brother was convicted of a crime? Or that you have a baby brother? Or your dad isn’t your biological father?
Proponents claim familial searching is a “no harm, no foul” situation and only generates potential leads. If there’s no more evidence connecting a person to a crime, they say, that’s that. Only the guilty should be worried, they say. But “generating leads” is another way of saying they get to poke around with zero probable cause or accountability. And the harms are real.
Not all New Yorkers experience the harms of the criminal justice system the same way. Communities of color, in particular, have known for years the damage done by invasive surveillance and over-policing. Because familial searching utilizes the convicted offender database (including broken windows and quality-of-life violators), the method promises to extend that disgraceful legacy.
This month, an appellate court struck down the use of familial DNA searching in New York, finding the unelected members of the Commission on Forensic Sciences overstepped their authority when they took it upon themselves to approve it in 2017. The court ruled that the decision to use this controversial technique is a matter of policy only legislators can make.
The court also acknowledged the disparate impact this method will inevitably have on communities of color: “The likelihood of being targeted because someone is a person of color (likely a male) is increased, a consideration that can only be evaluated as a matter of social policy, not science.”
Indeed. By trolling the database for potential connections to law-abiding relatives, this technique endorses the rationale that those related to criminals are more likely themselves criminals. What’s worse, the NYPD makes no secret that it agrees. Emanuel Katranakis, the NYPD’s chief of forensics, repeatedly cites a statistic that almost half of all prison inmates have a close relative who had also been in prison, arguing that “this supports the belief in and bolsters the efficacy of familial searching to produce valuable leads.” It turns out this thinking is not just dangerous but scarcely effective. As the appellate court notes, the Department of Justice reports that successful prosecutions based on this technique are rare.
All New Yorkers want to prioritize safety, which requires striking a balance with privacy. But we must remain vigilant to strike that balance the same way for all New Yorkers. Albany lawmakers have considered familial searching over the years, but repeatedly left it on the legislative cutting room floor. It seems like the perfect place for this inequitable policy.
Lewis is staff attorney with the DNA Unit at The Legal Aid Society.
The NYPD's new DNA dragnet: The department is collecting and storing genetic information, with virtually no rules to curb their use The NYPD's new DNA dragnet: The department is collecting and storing genetic information, with virtually no rules to curb their use
This Op-Ed (also by our guest speaker) critiques the NYPD’s DNA database, asserting that it’s a continuation of egregious stop-and-frisk policies
[The article is available online here and is pasted in below.]
If you think clandestine DNA dragnets and secret databases are things of science fiction, I have some upsetting news.
From the same law enforcement playbook that brought you stop-and-frisk comes the latest form of racial profiling and policing: the knock-and-spit. We’ve learned that NYPD officers are willing to knock on doors to take New Yorkers’ DNA, whether by “consent” or at the precinct by offering people a cigarette or drink to collect their spit. And once again, as the city’s primary public defenders, we see this latest law enforcement sweep is happening mainly to New Yorkers of color.
Take as an extreme example the police investigation of the Howard Beach jogger case. Before they identified a suspect, the NYPD collected well over 500 DNA profiles from men in the East New York area. Imagine police knocking on doors, in uniform, with a cheek swab in hand, asking residents to prove they didn’t kill the jogger in the nearby park.
They were willing to do it in East New York. Do you think this would happen on Park Ave.? In Park Slope? Not likely.
But things get worse from there. For those people excluded from the jogger case, the Office of the Chief Medical Examiner, the city’s crime lab, permanently keeps those profiles in their databank and routinely compares profiles to all city crimes.
As of 2017, the OCME had more than 64,000 profiles, stored regardless of whether they were identified as a suspect or cleared. Based on the lab’s claim that state laws are ambiguous and city laws are silent, the lab feels justified to invent their own rules about how this information is used.
And so, those 500 people who “chose” to cooperate with police may never know that they, and possibly their close relatives via familial searching, will be routinely compared to DNA samples from every rape, murder and violent crime in town. In this respect, they will be treated just like someone convicted of a crime.
Maybe none of this bothers you. A common refrain from law enforcement goes, if you don’t commit crimes, you needn’t worry. But innocent people have legitimate cause for concern because DNA has unparalleled power to wrongly accuse. Genetic testing can now reveal a profile based on a few cells, and we shed hundreds of thousands of cells a day. Your DNA can show up in places you’ve never been.
Consider Lukis Anderson, whose DNA was discovered under a murdered billionaire’s fingernails. Luckily, Anderson’s attorney discovered he was dead drunk in the hospital at the time and couldn’t possibly be the killer. How did his DNA get there? The authorities figure the same paramedics treated each man that night and inadvertently transferred Anderson’s DNA to the dead man. Anderson was facing a death sentence for seven months before he was cleared in 2013.
You might also take a cue from the police themselves. Under their labor contract with the city, rank-and-file officers don’t give the lab their DNA, which means the lab can’t easily rule out possible crime-scene contamination. This means that the officers knocking on doors ask people to volunteer to do what they won’t.
Basic privacy is another genuine worry. We see every day how our personal information, once set loose upon the internet, can never be recaptured. The past few years have heralded incredible expansions of the uses of DNA. Now imagine what will be possible in 30 years.
The mayor proudly considers the demise of stop-and-frisk a shining example of the city’s equality in law enforcement. It has been replaced by something arguably worse. We must stop the profiling, illegal storing and rogue searching that target people of color all over again.
Lewis is staff attorney with the DNA Unit at The Legal Aid Society.
Writing Reflection #10 Writing Reflection #10
Please go to our Moodle Page and under "Class 10" you will find the prompt and submission folder for Writing Reflection #10.
3.4.1 Optional DNA reading 3.4.1 Optional DNA reading
OPTIONAL: Can cats at a crime scene help find key DNA evidence? OPTIONAL: Can cats at a crime scene help find key DNA evidence?
Article available here and below.
Pets at a crime scene may be helpful in gathering key evidence but rarely are they considered for their role in human DNA transfer.
This research considers cats both as receptors and vectors for DNA of a person of interest—key evidence in criminal investigations.
In collaboration with the Victoria Police Forensic Services Department, forensic science researchers Heidi Monkman and Dr. Mariya Goray, from the College of Science and Engineering at Flinders, collected human DNA from 20 pet cats from multiple households.
Detectable levels of DNA were found in 80% of the samples and interpretable profiles that could be linked to a person of interest were generated in 70% of the cats tested.
"Collection of human DNA needs to become very important in crime scene investigations, but there is a lack of data on companion animals such as cats and dogs in their relationship to human DNA transfer," says Monkman.
"These companion animals can be highly relevant in assessing the presence and activities of the inhabitants of the household, or any recent visitors to the scene."
An experienced crime scene investigator Dr. Goray, an expert in DNA transfer, says this data can be very relevant when interpreting forensic DNA results obtained from a crime scene that includes pets.
"This type of data can help us understand the meaning of the DNA results obtained, especially if there is a match to a person of interest.
"Are these DNA finding a result of a criminal activity or could they have been transferred and deposited at the scene via a pet?"
"Further research is required on the transfer, persistence and prevalence of human DNA to and from cats and other pet animals and the influences animal behavioral habits, the DNA shedder status of the owners and many other relevant factors," the researchers say.
To this point, further collaborative work on cats and dogs is currently underway at the Flinders University forensic laboratory.
The article, "Is There Human DNA on Cats?," has been published in Forensic Science International: Genetic Supplement Series.
More information: Heidi Monkman et al, Is there human DNA on cats, Forensic Science International: Genetics Supplement Series (2022). DOI: 10.1016/j.fsigss.2022.10.014
3.5 Class 12: The Flip Side - False Confessions 3.5 Class 12: The Flip Side - False Confessions
This psychologist explains why people confess to crimes they didn't commit
Science | AAAS (June 2019)
This article provides an overview of the research on false confessions, highlights the work of CUNY Professor Saul Kassin, explains how false confessions can happen, and demonstrates how confessions can "trump" and "corrupt" other evidence.
“Why Do Innocent People Confess? Understanding & Challenging False Confessions” “Why Do Innocent People Confess? Understanding & Challenging False Confessions”
Deja Vishny
This article was prepared by a public defender as a tool to help other defenders challenge false confessions. It describes the Reid Technique, summarizes the science of false confessions, and spells out a complete defense strategy, including introducing expert testimony. Please read pages 1-15 (up to “theory of the case”) and feel free to skim or skip pages 15-19.
This document is a PDF so it is posted on Moodle under "Class 12."
Defense Brief on Appeal of Trial Court's Exclusion of Expert Testimony on False Confessions Defense Brief on Appeal of Trial Court's Exclusion of Expert Testimony on False Confessions
ARGUMENT
POINT I
TRIAL COURT'S REFUSAL TO ALLOW EXPERT TESTIMONY ABOUT CHARACTERISTICS OF DAN EVANS DEFENDANT AND SITUATIONAL FACTORS DURING THE INTERROGATION AND THEIR RELATION TO POTENTIAL FALSE CONFESSIONS VIOLATED EVAN'S RIGHT TO A FAIR TRIAL, AND REQUIRES REVERSAL
Dan Evans, who having had almost no contact with the criminal justice system for the entirety of his life, suddenly - in one afternoon -- confessed to two separate crimes that were committed three years apart. Given the interrogation techniques involved and Dan Evans' individual psychological make-up, there is strong basis to believe that the confessions were unreliable or false.
In their interrogation of Evans, the police employed a variety of techniques that scientific research has shown to be highly correlated with eliciting false confessions. Moreover, the police were not interrogating an individual of normal intelligence and a well-adjusted personality. Rather, Evans has severe intellectual and cognitive deficiencies that place him on the borderline of intellectual functioning. Further, extensive psychological testing shows that his personality is highly suggestible and compliant. Well-established scientific research indicates that when a person with Evans' psychological make-up is exposed to the *32 interrogation techniques, which the police freely admit to having used in this case, then any resulting confessions cannot considered reliable.
Unfortunately, the jury was not able to evaluate Evans statements in light of these findings. The trial judge decided that expert testimony on false confessions would not be admitted. As a result, the jury had no way of knowing that the techniques the police used, in light of Evans' intellectual and psychological capacities, were of vital significance. Instead, the jury was continuously exposed to the prosecutor's scientifically unsupportable idea that a false confession was not possible in this case and that the confessions were reliable. Had expert testimony been available to the jury - as required by law - Evans may very well have been found not guilty of the charged crimes, especially since there was very little evidence against him but for his confessions.
A. False confessions experts are crucial in trials in which certain specific indicia of false confessions are present
In People v. Bedessie, the Court of Appeals ruled that expert testimony may be used in trials to educate jurors about false confessions. The court ruled: “[T]here is no doubt that experts in such disciplines as psychiatry and psychology or the social sciences may offer valuable testimony to educate a jury about those factors of personality and situation that the relevant scientific community considers to be associated with false confessions.” 19 NY3d 147, 161 (2012). Specifically, expert testimony would be useful to illuminate the “conditions or characteristics of an *33 interrogation which might induce someone to confess falsely to a crime” and to explain that, “certain types of defendants are more likely to be coerced into giving a false confession - e.g., individuals who are highly compliant or intellectually impaired or suffer from a diagnosable psychiatric disorder, or who are for some other reason psychologically or mentally fragile.” 19 NY3d at 159. The Court further explained that, “just because jurors know about [false confessions] in their experience, doesn't mean they know about psychological studies about reliability of the evidence.” 19 NY3d at 156.
In this case, defense counsel sought to introduce testimony from Dr. Maria Hartwig, an expert on false confessions, to educate the jury on the phenomenon of false confessions and the factors that scientific research has found to be associated with false confessions. Scientific research demonstrates that when certain police practices are used, particularly on uniquely vulnerable suspects, what can result is a “coerced compliant false confession” in which a suspect adopts a false narrative of the crime consistent with what the suspect believes the interrogating police officer wants to hear so that the suspect can end the stressful situation they find themselves in. Christine S. Scott-Hayward, Explaining Juvenile False Confessions: Adolescent Development and Police Interrogation, 31 Law & Psychol Rev 53, 55 [2007]. Some of the specific police practices associated with “coerced compliant false confessions” are: an investigating officer that believes in *34 the suspects guilt prior to an interrogation; isolation of the suspect for a lengthy interrogation; using deception during the interrogation; suggesting to the suspect that they committed the crime accidentally; and implying that some leniency or lack of culpability will result from cooperating with the investigator. Id.; Brandon L. Garrett, The Substance of False Confessions, 62 Stan L Rev 1051, 1063-64 [2010]; Richard A. Leo, Why Interrogation Contamination Occurs, 11 Ohio St J Crim L 193, 211 [2013]. These practices are particularly likely to cause false confessions from suspects who have low intelligence, issues with cognitive functioning, suggestive and compliant personality, and who suffer from certain personality disorders. Christine S. Scott-Hayward, Explaining Juvenile False Confessions: Adolescent Development and Police Interrogation, 31 Law & Psychol Rev 53, 56 [2007]; Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 22 [2010], Saul Kassin, a leading researcher on false confessions, explains: “There is a strong consensus among psychologists, legal scholars, and practitioners that juveniles and individuals with cognitive impairments or psychological disorders are particularly susceptible to false confession under pressure.” Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 22, 34 [2010].
*35 The need for expert testimony in cases in which these indicia for false confessions is present is particularly acute because the procedures police employ that elicit false confessions are often standard techniques that are, “entirely permissible under the U.S. Constitution and recommended by police training on modern psychological interrogation techniques.” Brandon L. Garrett, The Substance of False Confessions, 62 Stan L Rev 1051, 1063 [2010]; Frances E. Chapman, Coerced Internalized False Confessions and Police Interrogations: The Power of Coercion, 37 Law & Psychol Rev 159, 164-65 [2013]. Therefore, rather than resulting from some kind of police abuse, poor training, or bad-faith, false confessions in these situations arise out of the proper application of the various techniques that are standard practice in American policing. Richard A. Leo, Why Interrogation Contamination Occurs, 11 Ohio St J Crim L 193, 197-98 [2013]; Frances E. Chapman, Coerced Internalized False Confessions and Police Interrogations: The Power of Coercion, 37 Law & Psychol Rev 159, 164-65 [2013].
It is important to note that this research does not suggest that anytime a police officer uses standard police techniques a false confession will occur. Rather, this research reveals that in situations in which many of these techniques converge, particularly in relation to a suspect with a uniquely vulnerable *36 psychological make-up, resulting statements are difficult to deem reliable.5 In the present case, then, expert testimony was essential to helping a jury connect the unique factors present in Evans' interrogation with scientific research that links those factors with false and unreliable confessions.6
B. Denying the false confessions expert was abuse of discretion because specific factors present in this case made such testimony essential
The specific indicia of false confessions that are present in this case are: (1) a defendant with severe intellectual and cognitive deficiencies and a personality that is uniquely suggestible; (2) a lead detective who believes the defendant to be *37 guilty prior to initiating the interrogation and with knowledge of all the key facts needed to contaminate the confession; (3) the employment of the standard police interrogation techniques of confrontation, minimization, deception, and possibility for leniency with cooperation; and (4) an abnormally long, isolating interrogation.
1. Evans intelligence, cognitive abilities, and personality are in line with those that make individuals uniquely vulnerable to false confessions
As explained above, courts find that vulnerability of a defendant is a crucial factor that makes false confessions more likely. In this case, psychologist Dr. Sanford Drob performed an extensive psychological evaluation of Evans and found that, “Mr. Evans exhibits traits that would render him vulnerable to producing a false confession. These traits include borderline intellectual functioning, cognitive, social and emotional immaturity, severe deficits in reality testing and deficits in the capacity to understand the actions and intentions of others, deficits in his capacity to cope with interpersonal stress, anxiety, depression, dependency, passivity and a desire to please others, and a concomitant tendency to rely on others for direction and support.” (Drob Report, 13).
With respect to intelligence, Evans was found to have an IQ of 72, which places him in the low-end of borderline intellectual functioning. (Drob Report, 8, 12). Additional testing of Evans' basic cognitive/neuropsychological processes yielded a performance in the lowest 0.4% of the population with notable deficits in *38 basic language functions, constructional capacity, learning, and memory. (Drob Report, 8). Standard tests of suggestibility and compliance showed Evans to be highly suggestive, willing to provide answers to questions for which he has no information, and a tendency to comply with requests and obey instructions even when they are against his wishes. (Drob Report, 12).
Tests of Evans' personality revealed depressive, dependent, avoidant, schizoid and negative personality traits. These tests indicated strong dependency strivings, desire to lean on others, disillusionment with those he depends on, low self-esteem, tendency to retreat into depression and fantasy, distress over personal failures, sense of humiliation, feelings of unworthiness, uselessness and guilt, and evidence that Evans is timid, shy, anxious and sensitive to rejection. (Drob Report, 10). Moreover, these tests showed that Evans suffers from several personality disorders and syndromes, including dependent personality disorder, general anxiety disorder, and posttraumatic stress disorder. (Drob Report, 10). Rorschach testing of Evans showed significant impairment in reality testing and information processing that was so severe that it potentially predisposes him to psychosis. This can lead to impaired judgment and a failure to anticipate the consequences of his behavior. This testing also suggests that Evans' thinking can become odd and peculiar, especially when he is trying to understand people and their actions. (Drop Report, 11).
*39 Evans vulnerable, dependent and compliant personality was apparent early in his interaction with the police. While hospitalized for the gunshot wound that arose from the 2009 shooting, Evans was visited by Det. Melino who interviewed him about the shooting. During the interview he cried twice, explaining that he was having problems with his girlfriend who was scared and wanted to move out of town, with their child, because of the shooting. (PTH, 14). Specifically, Evans wanted Det, Melino to speak to his girlfriend and “to convince her not to move, you know, to stay supporting him.” (PTH, 14-15). Further, he wanted Det. Melino to speak to his doctor, “because he was upset that he's been there for, he said, three days.” (PTH, 17). It seems that Evans immediately and strangely believed that Det. Melino was a confidante of his who would help him with his personal problems - something that would be easier to understand if one imagined Evans as a child.
Importantly, the intellectual, cognitive, and personality deficiencies that Evans suffers from have all been linked to false confessions. Christine S. Scott-Hayward, Explaining Juvenile False Confessions: Adolescent Development and Police Interrogation, 31 Law & Psychol Rev 53, 56 [2007]; Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 22 [2010]. However, without expert testimony stating this, the jury had no way of knowing that these risk factors were present or relevant.
*40 2. Detective Mooney's single-minded approach to Evans' involvement with the 2006 shooting is an independent risk factor in producing a false confession
Standard police training instructs detectives and officers to, “interrogate only those suspects whose culpability they ‘establish’ on the basis of their initial investigation.” Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 6 [2010]. As a result, once the interrogation begins its purpose is not to, “discern the truth, determine if the suspect committed the crime, or evaluate his or her denials... the single-minded purpose of interrogation is to elicit incriminating statements, admissions, and perhaps a full confession in an effort to secure the conviction of offenders.” Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 6 [2010].
However, researchers have discovered that when investigators begin the interrogation process presuming the guilt of the suspect, “this presumption creates expectancies that guide the investigator's behavior” Brian Cutler et. al., Expert Testimony on Interrogation and False Confession, 82 UMKC L Rev 589, 610-11 [2014]. Specifically, it introduces confirmation biases and expectancy effects that predisposes them, “to ask confirmatory questions, use persuasive tactics, and seek confessions.” Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 6 [2010], citing Hill, Memon, & *41 McGeorge, 2008; Kassin, Goldstein, & Savitsky, 2003). With these biases in place, the investigator, “discounts and manages the suspect's denials, increases the suspect's anxiety, and, in the case of an innocent suspect, fulfills the prophecy by securing a false confession from the suspect.” Brian Cutler et. al., Expert Testimony on Interrogation and False Confession, 82 UMKC L Rev 589, 610-11 [2014]. This phenomenon has been demonstrated in research as well as through case studies of known false confessions. Id. As a result, “Investigator bias,' whereby officers focus on one suspect because they are convinced he or she is guilty,” is an independent risk factor increasing the likelihood of false confessions. Frances E. Chapman, Coerced Internalized False Confessions and Police Interrogations: The Power of Coercion, 37 Law & Psychol Rev 159, 164-65 [2013].
In this case, there is clear evidence that investigators approached Evans' interrogation with a belief in his guilt and a single-minded focus to illicit a confession. . . . All this information suggests that Det. Mooney had developed a belief that Evans was guilty of the 2006 homicide and that after sizing Evans up in the hospital following the 2009 shooting, he developed a plan to arrest Evans for the 2009 shooting so that he could then interrogate him about the 2006 homicide.
3. The police deployment of standard techniques of maximization and minimization, deception, and implied offers of leniency in the context of an isolating and lengthy interrogation are unique indicia of false confessions
The standard police technique utilized by the two officers who interrogated Evans - Det. Ragolia and Det. Mooney - include the following: long and isolating interrogation; use of deception; use of maximization and minimization, and use of implied offers of leniency. These are all techniques that scientific research has revealed to be linked with coerced and unreliable confessions. Specifically, scientific research demonstrates that the cumulative stress that these techniques create leads to a situation where the immediate gain of complying with an officer's *44 wish for a confession outweighs the uncertain long-term consequences of false confession. Christine S. Scott-Hayward, Explaining Juvenile False Confessions: Adolescent Development and Police Interrogation, 31 Law & Psychol Rev 53, 55 [2007], Brandon L. Garrett, The Substance of False Confessions, 62 Stan L Rev 1051, 1063-64 [2010].
a. Isolation and Interrogation Length
Two primary risk factors associated with false confessions are the length of the confession and the isolating manner in which they occur. Empirical studies have shown that most interrogations last around one hour, and that most proven false confessions arose out of interrogations lasting six hours or longer. Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 16 [2010]; Steven A. Drizin & Richard A. Leo, The Problem of False Confessions in the Post-DNA World, 82 NCL Rev 891, 948 [2004].
Further, the leading training manual on law enforcement questioning suggests that custodial interrogations should not last for more than four hours, acknowledging that lengthier interrogations are potentially coercive and increase the likelihood of eliciting involuntary compliance from a suspect who is seeking to escape the pressures of extended confinement and accusatory questioning. Fred Inbau, John Reid, Joseph Buckley and Brian Jayne, Criminal Interrogation and Confessions, 4th Ed., Aspen Publishers, p. 206 (2001)). Separately, the technique *45 of isolating suspects in small, private rooms during their interrogations enhances the coercive effects of standard police interrogation techniques and are an independent risk factor in producing false confessions. Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 6-7 [2010].
In this case, Evans' interrogation began at 11:00 a.m. and he was questioned through the day until almost midnight (PHT. 186-7). According to Det. Ragolia, Evans likely made his first incriminating statement sometime between 1:00 pm and 2:00 pm, roughly two to three hours after Evans was placed in the interrogation room. (TT 821, 828, 922). The time is uncertain because according to Det. Ragolia, Evans initial statement indicated that he had nothing to do with the shooting. (TT 814). Then, according to Det. Ragolia, he left the room - saw Det. Mooney and then reentered the room and requested another statement from Evans, which this time contained incriminating content. (TT 820-21). However, the time at which this occurred is not clear. According to Det. Ragolia, he Mirandized Evans and then asked for this second statement at 12:45 pm. However, according to Det. Mooney, he didn't arrive to the station until 1:00 pm and was told around 1:30 pm or 1:45 pm that Evans was now making an incriminating statement. (TT 922-24). Nevertheless, the statement that Evans gave did not implicate him as the responsible party in the crime. Therefore, at around 3:45 pm *46 (TT 925, 937), Det. Mooney initiated another interrogation that lasted until 5:00 or 5:30 pm (TT 831), which yielded a more incriminating statement. So Evans did not provide a statement that was used as his confession until approximately six hours after he was placed in the interrogation room. It's important to note, however, that Evans did not experience this statement as a confession. Based on conversations he had with Det. Ragolia and Det. Mooney, he was under the impression that the statement he made did not incriminate him and that he was free to go home. It wasn't until 7:45 pm, over eight hours after Evans was brought to the interrogation room, that Det. Mooney began questioning Evans about the 2006 homicide (PHT. 169). Further, it is important to note that in the entire time that he was interrogated, Evans did not eat any food except for one bite of a brownie. (PTH 74-75).
That Evans was interrogated for such an extraordinary length of time, and, further, that he was isolated in an interrogation room with barely anything to eat for that entire time, is strongly associated with false confessions. However, without expert testimony on these risk factors, the jury had no way of knowing their significance.
b. Maximization and Minimization
Maximization (also known as confrontation) and minimization is a push-pull process whereby investigators will alternate between confronting a suspect with *47 their absolute certainty of the suspect's guilt and suggesting to the suspect that perhaps there are circumstances that would mitigate or minimize their guilt. The confrontation thrusts “the suspect into despair” while the minimization “makes a confession seem like an expedient means of escape. Interrogators are thus trained to suggest to suspects that their actions were spontaneous, accidental, provoked, peer-pressured, drug-induced, or otherwise justifiable by external factors.” Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law 8 Hum Behav 3, 18 [2010]. Brian Cutler explains: “While maximization techniques provide the pressure to confess, minimization techniques ease the way toward confession.” Brian Cutler et. al., Expert Testimony on Interrogation and False Confession, 82 UMKC L Rev 589, 608-09 [2014].
In this case, Det. Mooney freely admits that he employed the maximization and minimization techniques prior to any statement by Evans regarding the 2006 homicide:
I explained to him that you are going to get charged with murder as a result of this investigation. I don't know exactly what happened because I wasn't there, but I don't know whether you did it on purpose or by accident. I don't know. I wasn't there. Only way I'm gonna find out is if you tell me the truth about what occurred. I explained to him what mitigating meant. There were - perhaps there were circumstances beyond his control that caused this to happen. (TT, 960).
Mooney further explains: “We had a little bit more of an extended conversation where I again explained to him, I used the word ‘mitigating’ when I *48 was talking to him, and I explained to him what I meant about that. Once I thought that he understood what I was talking about, I said, ‘Just tell me what happened,’ and then he began to give a narrative about what happened.” (PHT, 178-79). Mooney is clear that before he asked Evans for his narrative, he told Evans, “I don't know whether you shot her on purpose or it was accidental. The only way that I'd find out is if you were to tell the truth about it.” (PHT, 231-32). Thus, Mooney suggested to Evans that if it was accidental, the version Evans ultimately adopted, it would not be so bad for him.
Although the interrogating officer in this case freely admits to using the interrogation technique of maximization and minimization, the jury was not allowed to know that scientific research has demonstrated that this technique significantly increases the possibility of a false confession.
c. Deception
The use of deception and presentation of false evidence in interrogations is a particular police technique that research has shown leads innocent people to confess to prohibited acts. Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 17 [2010]. False confessions expert Saul Kassin explains: “Basic psychological research reveals that misinformation renders people vulnerable to manipulation... Scientific evidence for human malleability in the face of misinformation is broad and pervasive. The forensic literature on confessions reinforces and extends this classic point, indicating that *49 presentations of false evidence can lead people to confess to crimes they did not commit.” Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 17 [2010].
Courts have recognized that such techniques, “particularly when applied to those of immature minds, may cause, with a substantial degree of probability, false confessions.” Singletary v Fischer, 365 F Supp 2d 328, 336-37 (EDNY 2005). See also People v Thomas, 22 NY3d 629, 639 (2014)(detailing how combination of police tactics can result in psychological coercion and render incriminating statements wholly unreliable).
There can be no doubt that that Det. Mooney used this technique as part of his interrogation with Evans. Mooney explained: “Commonly in the course of an interrogation and it's an accepted practice that detectives are allowed to use what's refer[red] to as a ruse in order to have the subject of the interrogation think that I know more than I actually do about what happened.” The purpose of this ruse, Mooney explained, is “to gain cooperation from the subject of the interrogation.” (TT 1064), In this case, Det. Mooney told Evans that several witnesses, when shown a photo array, had identified Evans as the shooter in the 2006 homicide although this was entirely untrue. Further, Mooney told Evans that these people would be coming to view him in a physical lineup the following morning. (TT 956; *50 PHT. 234). Det. Mooney reinforced this ruse by continuously telling Evans that he was certain of Evans' guilt. (TT 960).
Det. Ragolia suggested that he used a similar technique when questioning Evans about the 2009 shooting when he testified, that “I explained to him that we had witnesses that say that he was one of the shooters,” even though all the police had was a single child witness who identified Evans under classic suggestive circumstances. (PTH, 63).
Although both detectives used ruses in their interrogations, the jury was not allowed to find out that such ruses - particularly when utilized against an individual like Evans - are highly correlated with false confessions.
d. Implied Offers of Leniency
Social science researchers understand the technique of minimization to be particularly influential on inducing false confessions because it implies, without any explicit statement, that if a suspect provides a confession they will receive some kind of leniency. Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 18 [2010] Empirical research suggests that techniques that impliedly suggest leniency in exchange for a confession are a primary cause of false confessions. Richard A. Leo et. al., Bringing Reliability Back in: False Confessions and Legal Safeguards in the Twenty-First Century, 2006 Wis L Rev 479, 517 [2006].
*51 In this case, in addition to Det. Mooney's minimizing suggestion and associated implication of leniency - that Evans' involvement was accidental - Det. Mooney also created the impression that he was willing to provide Evans with leniency in exchange for his services as an informant. According to Det. Mooney, early in the interrogation Evans had expressed to Det. Ragolia that he wanted to be an informant. (PHT. 221-22). Mooney further explains: “One of his lines of response to me was that he was interested, he had seen on television where people sometimes would have a problem that get arrested become informants for the police... He thought that he could perhaps provide us with a lot of information on various subjects.” (PHT. 171-72). In response to Evans' offer to be an informant, Mooney told Evans: “Without you telling me the truth about things, there's no point even discussing being an informant at this point.” (PHT. 172). The statement suggests that if Evans provides a narrative in which he is the cause of the 2006 homicide, then Mooney would consider making him an informant with its implied benefits. Given that Mooney told Evans that he firmly believed Evans was guilty and fortified this stated belief with lies about the strength of the case, Evans was trapped into thinking that admitting the crime was the only way to avail himself of the opportunity to be an informant.
Det. Mooney took further action to make it seem as if he was taking Evans' offer to be an informant seriously by showing him - in direct response to his *52 request - photographs of two men who were involved in an unrelated homicide. (PHT. 224-25). Mooney explained that these were not the only photographs he showed Evans during the interrogation, and in fact, that he brought additional photos of unrelated cases for Evans to look at. (PHT. 174-75). Mooney explains: “He had already stated that he was interested in being an informant. These are a lot of open cases. Philip Diaz through all of the other non-fatal shootings. If he could have provided closure on some of those, I certainly would have tried to get that information from him.” (PHT. 239).
It is also important to note that it was at this point in the interrogation that Det. Mooney, for the first time, explained to Evans that he was under arrest and that he was not free to go home. (PHT. 170). Very significantly, it is also at this point, that Det. Mooney admits to having employed the minimization and maximization techniques described above - and it is at this point that Evans asked Det. Mooney if he could be an informant like the ones from TV. Given that Evans has borderline intellectual functioning, his personality is uniquely suggestible, and he is prone to misinterpret the intentions of others, it is understandable how Evans would be so compliant and influenced by these minimization and maximization techniques and why he might actually believe that Det. Mooney was now *53 employing him as an informant.8 For example, while Det. Mooney wrote out a statement for Evans admitting his culpability, he noticed that Mr. Evans, “was looking at those papers that were on the table. The photographs as well as the summaries that were attached to some of them.” (PHT. 186). Given the combination of techniques used, it is very likely that Evans believed that he was being used as an informant and that by stating that the shootings were accidental he would be helping the police while absolving himself of responsibility.9
4. The need for a false confessions expert in this trial was particularly acute as Evans behavior suggests that he was, in fact, producing a false confession that was being actively contaminated by the interrogating officers.
a. Evans' did not understand the severity of his situation and believed that he would be allowed to go home after providing statements that included the details interrogating officers' asked for.
. . . . Det. Mooney explained that even after Evans had given a recorded statement in which he claimed to be responsible for the 2009 shooting, Evans believed that he would be allowed to go home, “to put his baby to bed because his grandmother was baby-sitting.” (PHT. 169). . . . Even the trial judge noted that Evans' did not seem to understand the situation he was in following his initial confession, noting, “I guess it's still in his mind. He didn't think he was in custody, I guess because he's not experienced in the criminal justice system. I'm not exactly sure.” (PTH, 284-85).
Evans' reaction - and his lack of understanding of the severity of his situation - is consistent with innocent individuals with his psychological make-up experiencing the kind of interrogation that he did. Saul Kassin explains: “People with low IQ and high [suggestibility] scores are most vulnerable. They are substantially more likely to think they'll be able to go home after a confession and that they do not need legal advice if they are innocent.” Saul M. Kassin et. al., Police-Induced Confessions: Risk Factors and Recommendations, 34 Law & Hum Behav 3, 21 [2010]. Evans functions intellectually and emotionally much closer to *55 the level of a child. As a result, it would be easy for him to believe that if he simply told the police what they wanted to hear he would not suffer any consequences for it - particularly when the police reinforce this idea by telling him repeatedly that if he says that the shooting was accidental then they would understand, implying that he would not be in trouble.
The fact that even the trial court was perplexed by Evans' reaction to his admission of guilt is a clear indication that expert testimony on false confessions was essential for a jury. Evans' reaction is consistent with false confessions, but one would not think so if relying exclusively on common sense. Expert assistance, then, was necessary to explain how the specific factors present in Evans' interrogation - including his mental make-up - cause false confessions.
b. Based on police testimony Evans' behavior is entirely consistent with a false confession contaminated by police.
False confessions often appear authentic because they contain details consistent with commission of the crime. Richard A. Leo, Why Interrogation Contamination Occurs, 11 Ohio St J Crim L 193, 195-96 [2013]. However, these details can enter a suspect's narrative through a process of contamination where, “Police may, intentionally or not, prompt the suspect on he v the crime happened so that the suspect can then parrot back an accurate-sounding narrative.” Brandon L. Garrett, The Substance of False Confessions, 62 Stan L Rev 1051, 1053 [2010]. Leading false confessions expert Richard Leo explains:
*56 [T]he psychological design of American interrogation methods virtually dictate the contamination once detectives have selected an innocent target for questioning. American interrogation is a laboratory of confirmation bias. Contaminated confessions are the product of what we might call the confirmation bias effect of the assumptions, goals and techniques of guilt-presumptive, accusatory police interrogation. Richard A. Leo, Why Interrogation Contamination Occurs, 11 Ohio St J Crim L 193,211 [2013].
Importantly, however, police contamination of a given suspect's narrative is typically inadvertent and, “investigators do not appear to typically realize when they are contaminating confessions.” Id. Research indicates that even in cases in which it is known that confessions were contaminated, as the accused were subsequently exonerated by DNA evidence, detectives subjectively believed themselves to be assiduously avoiding contaminating the suspects' narratives. Brandon L. Garrett, The Substance of False Confessions, 62 Stan L Rev 1051, 1057 [2010]; Deborah Davis & Richard A. Leo, To Walk in Their Shoes: The Problem of Missing, Misunderstood, and Misrepresented Context in Judging Criminal Confessions, 46 New Eng L Rev 737, 767 [2012]. Detective Jim Trainum recounts a personal experience that demonstrates how unknowing contamination of a suspect's narrative can occur: “We had unintentionally fed her almost the entire case over a several hour period. And Kimberly - the woman - she would guess. She would guess a lot. And sometimes the guesses were right. And we wouldn't see the ones that weren't because ‘She was being evasive.’ ‘She was protecting someone.’ So that's how we wrote that off... But fortunately, we had accidentally let the tape continue to run, and we captured the whole thing on *57 video. But if we hadn't had that video, we never would have been able to go back years later and catch our mistakes.” As told to the Mid-Atlantic Innocence Project's Rachel Cicurel. Available at: http://www.exonerate.org/in-their-words/james-trainum (Last visited August 15, 2014).
In this case the entire interrogation consists of the interrogating detectives specifically asking Evans to change details in his statements so that they fit with the detectives beliefs of what occurred. Det. Mooney was clear that after reading Evans first statement - written by Det. Ragolia - that he wanted to further interrogate Evans because, “there were discrepancies between what he said and what we knew based on witness accounts.” (TT 925, 937, PHT. 145-46). Det. Mooney confronted Evans, telling him that he knew that he was lying about certain details in the written statement. (TT 830). According to Det. Mooney: “When I tried to explain to him what some of the things were that - thought were not true, he seemed to me very easily acquiesced to the idea.” (TT 927-28). The interrogation continued with Det. Mooney supplying Evans with the details he thought were true, and Evans agreeing to include them in his statement, [examples omitted]
These examples demonstrate that the narratives that are attributed to Evans came in direct response to police suggestion. The interrogating detectives are clear that they asked Evans to change details that did not conform to the evidence they had gathered and, in response, Evans complied. In addition, this interrogation featured detectives continuously lying about the case against Evans, implying that if he admits guilt he will be allowed to go home because the events were not his fault, and further insinuating that no harm would come his way because they would be using his services as an informant. Use of these techniques combined with the fact that Evans' mental functioning is closer to that of a compliant child makes it clear that his statements are not reliable. Many if not all details he supplied came from the officers. The lack of a videotape of the interrogation makes it impossible to know the true extent of contamination that occurred - but even based on the *60 detectives' own testimony, it is clear that his narratives were coached and suggested.
c. Based on the presence of so many indicia of false confessions, and evidence that the interrogation actually did induce a suggestible, compliant, confused state in the defendant, the defendant's statements cannot be considered reliable.
It is not in question that Evans is mentally handicapped and has a highly suggestible and compliant personality. It is not in question that Evans was held in an isolated interrogation room for over twelve hours as part of the interrogation that yielded his confessions. It is also not in question that detectives used techniques such as maximization and minimization, ruses, and implied offers of leniency in their interrogation. Further, it is not in question the detectives supplied Evans most of the details that he included in his statements. That all these factors appear together is hugely significant and results in statements that cannot be considered reliable. “Employing standard interrogation methods on mentally handicapped suspects and conducting lengthy interrogations... appear to play a major part in precipitating untrustworthy confessions. When either of these practices are employed, there is a very substantial likelihood that suspects will provide the answers sought by their interrogators, regardless of their own beliefs in the truth of those answers.” Welsh S. White, What Is an Involuntary Confession Now?, 50 Rutgers L Rev 2001, 2042-43 [1998].
. . . . It is only after Evans is exposed to a panoply of police techniques that are known to induce false confessions that Evans makes an admission connected to the 2006 homicide. Given that Evans is functioning at the mental level of a child, with no clear understanding of the severity of his situation, with a uniquely compliant and suggestible personality it is very likely, then, that he was simply providing Det. Mooney with the statement that Det. Mooney wanted to hear so that Evans could go home to his grandmother and baby. At minimum, however, there is no way to understand this statement as reliable given the number of indicia of false confessions that are present.
5. Lack of false confessions expert altered the entire trial by presenting the jury with a skewed world in which social science research about false confessions does not exist.
a. A lack of testimony from a false confessions expert was used by the prosecution to introduce various theories and ideas about false confessions that have no basis in fact.
Scientific research is clear that certain factors strongly increase the likelihood of false confessions. Although many of these factors are present in this case, Evans was not allowed to introduce any of this research as evidence. However, the prosecutor, particularly in their closing argument, introduced various theories about false confessions that have no basis in any scientific research, but rely exclusively on conventional wisdom that is not accurate in this case. These *64 theories stood as uncontested thus creating the impression that the prosecutor's arguments were somehow credible.
i. Length of interrogation
It is well established that the length of an interrogation is a crucial risk factor in false confessions. Nevertheless, because no expert was allowed to testify on this subject, the prosecutor was able to state in her closing summation: “In all, the whole thing took from the minute he gets to the precinct until the end of the second video is twelve hours. Certainly not any unnecessary delay and nothing that would overcome somebody's will.” (TT 1497). The idea that being held in an interrogation room for twelve hours is not a factor that increases the likelihood of false confessions is simply false. As explained above, average interrogations last one hour, and interrogations that last over six hours are highly correlated with false confessions.
ii. Individual characteristics
The prosecutor also challenged scientifically established knowledge about the role that intellectual and cognitive functioning and personality have on the likelihood of a given suspect providing a false confession. While research cited above - including by the Court of Appeals in Bedessie - establishes that individuals with low intellectual and cognitive functioning and suggestible personalities are uniquely vulnerable to providing false confessions, the prosecutor *65 states the following in her summation: “None of these traits would ever suggest that he would falsely confess, if anything, they help to explain why he went to kill a man who he felt had humiliated him, fire indiscriminately at him down a crowded street and then truthfully confessing to it.” (TT 1507-08). The prosecutor explored this theme again suggesting that rather than making defendant more vulnerable to false confession, his low intelligence, “explains why he committed these two crimes in broad daylight with lots of people looking and why he thought he could talk his way out of them and didn't refuse to talk to the police.” (TT 1500-01).
When defense counsel attempted to introduce evidence that such traits do suggest that a person might falsely confess, as an overwhelming amount of scientific research demonstrates, the trial judge refused. (TT 1349). As a result, the jury was presented with a world in which low intelligence actually decreases the likelihood of a false confession - a world exactly opposite to the one we live in.
iii. Minimization and Maximization Techniques
The prosecutor also distorted the significance of the techniques of minimization and maximization, which Det. Mooney freely admitted to having used in his interrogation of Evans, and which extensive research links to the production of false confessions. This research explains that, for example, when an interrogator insists that they know a suspect is guilty, this event induces incredible *66 stress in the suspect, makes them feel “trapped” and “hopeless” irrespective of whether or not they are guilty. Brian Cutler et. al., Expert Testimony on Interrogation and False Confession, 82 UMKC L Rev 589, 608-09 [2014]. Nevertheless, the prosecutor takes the scientifically unsupportable position that, “[c]onfronting [Evans] with the fact that witnesses have identified him is not psychological coercion. No[where] near what it would take an innocent person no matter what their IQ to confess to shooting a gun in a crowded basketball court.” (TT, 1472-73). The prosecutor goes on to say, contrary to expert findings, that only a guilty person would experience such maximization techniques as coercive:
An innocent person being told we're going to put you in a lineup, witnesses from that homicide will see you, is going to say, Great, put me in a lineup, I'm not the right guy, you'll see there is no way I'm going to get picked out. An innocent person wouldn't hear that he was about to be put in a lineup and then confess to a murder.
(TT 1499).
. . . . Of course, these characterizations by the prosecutor demonstrate why expert testimony was so essential in this case. The prosecution is presenting theories *68 about how innocent and guilty people act in this kind of situation. Scientific research demonstrates that these theories are wrong. However, because this scientific research was excluded, the jury was only exposed to those inaccurate accounts, without any corrective information.
b. The prosecutor was allowed to mischaracterize the significance of the videotaped statements while actual expertise on the videotaping was excluded
The prosecutor repeatedly suggests that the videotape of Evans statement shows that he is relaxed, which means there was no coercion in the interrogation. (TT, 1495-96). In cross-examination of Dr. Drob, the prosecutor was allowed to ask: “Isn't it true according to the literature that the videotape is the best evidence of an individual at the time that we're talking about, if we're talking about his competency at that time the video is the best evidence of that, is it not?” (TT 1302). Dr. Drob responded: “It's certainly very good evidence because it's a film of the individual at the time you're considering, yes.” (TT 1302). The prosecutor then proceeded to use Dr. Drob's testimony to suggest that Evans was not anxious in the video and, by implication, was not coerced.
The reality is, unfortunately, that these kinds of ‘final-hour’ tapings, which are commonly made at the end of lengthy interrogations, are not useful measures of what actually occurred during an interrogation. In fact, the highest courts in at least seven states have limited admissibility of recorded statements in such *69 situations and have condemned such final-hour taping of interrogations. . . . .
Nevertheless, during the redirect of Dr. Drob, defense counsel was not allowed to ask Dr. Drob about the scientific research related to such final hour tapings: . . . . the videotape is not a valuable measure of Evans likelihood of having confessed falsely. But, the jury was not allowed to be exposed to this reality.
c. Cumulative effect of judge's rulings was to create a fantasy world where scientific evidence about false confessions doesn't exist and that the only source of information about factors that contribute to false confessions was the prosecutor's inaccurate theories
Extensive scientific research suggests that certain factors substantially increase the likelihood of false confessions. Many of these factors were present in this trial. However, the scientific evidence that links these factors with false confessions was systematically excluded. Consequently, although the prosecution's case was based almost entirely on Evans' confessions, Evans was not allowed to illuminate to the jury that various indicia of false confessions were present in his interrogation.
As a result, the prosecutor could credibly state the following in summation:
*72 There is absolutely no shred of evidence presented at this trial, by either side, that the defendant falsely confessed. No one ever said it... There is absolutely no evidence whatsoever that the defendant was coerced to make this statement or any evidence that he had gotten his information from anywhere other than his own memory of the events... You can only base your verdict on evidence and here there was nothing presented to you that in any way suggest a false confession.
(TT 1500)
In other words: the prosecutor was able to capitalize on the judge's erroneous ruling's to create the false impression that nothing in Evans' case suggested that a false confession may have occurred. Yet, extensive evidence does suggest, given the unique factors present in this case, that a false confession was likely.
Defense counsel objected to the prosecutor's summation, arguing: “No one presented evidence that the confession was false. I was prevented from doing that so this reopens the issue about a false confession expert and her exploiting the fact that I wasn't able to put on a false confession expert, so I object to that. I'm just amplifying my grounds and that is also a basis for a mistrial.” (TT 1514). The trial court's only response was “Okeydokey.” (1514).
There is no way that, in light of this situation, Dan Evans received a fair trial. This case featured numerous indicia of false confessions that courts have found require explanation by a false confessions expert to a jury. The jury was denied access to this vitally important scientific research and, thus, was presented *73 with a world in which no such research existed. As a result, the prosecutor was able to blatantly lie about the causes of false confessions and the likely behavior of those subject to the kind of techniques employed by the police in this case. With no opportunity to effectively rebut these erroneous claims, Evans opportunity for a meaningful defense and his right to a fair trial was destroyed.
Constructing Rich False Memories of Committing Crime Constructing Rich False Memories of Committing Crime
Julia Shaw and Stephen Porter, Psychological Science OnlineFirst (January 14, 2015)
This is “the first study to provide evidence suggesting that full episodic false memories of committing crime can be generated in a controlled experimental setting.” When reading, consider: would you confess?
Note: You only need to read the highlighted portions (there’s a lot of highlighting at the beginning and less at the end). Beacuse this is a PDF it is posted on Moodle under "Class 12."
Writing Reflection #12 Writing Reflection #12
Please go to our Moodle Page and under "Class 12" you will find the prompt and submission folder for Writing Reflection #12.
3.5.1 OPTIONAL 3.5.1 OPTIONAL
OPTIONAL: The Accused: An interactive experience
Engage with the interactive experience at this link to get experience (in some ways) with a police interrogation.
OPTIONAL: Police Body Cameras: What Do You See?
Interactive experience from The New York Times
Evaluate the effectiveness of videotapes confessions by engaging with this interactive assessment of the effectiveness of police body cameras. (Content warning: this is dramatized and not real footage, but it does show police conduct that could be triggering.)
3.6 Class 13: Hope & Advocacy 3.6 Class 13: Hope & Advocacy
West Virginia House Judiciary Committee Unanimously Approves Legislation to Address Wrongful Convictions
Jackson Newspapers (March 16, 2021)
Hope for reopening cases that involve discredited forensics: A 2021 statutory change in West Virginia.
NY Bill No. A05891B (New York State Assembly)
Amends procedures required for the custodial interrogation of children
This is the proposed New York legislation that you just read about. The key amendment is pasted in below, but I've included the link in case you want to read or skim the full bill.
A child shall not be questioned pursuant to this section unless or until:
(a) the child and a person required to be notified pursuant to subdivision three if present, have been advised: (i) of the child's right to remain silent; (ii) that the statements made by the child may be used in a court of law; (iii) of the child's right to have an attorney present at such questioning; and (iv) of the child's right to have an attorney provided for him or her without charge if he or she is indigent; and
(b) the child has consulted with legal counsel in person, by telephone, or by video conference. This consultation may not be waived.
Oregon Bill SB 418 (Effective date, January 1, 2022)
Hope for reducing false confessions: A 2021 statutory change in Oregon.
Georgia v. Denton, Order Granting Extraordinary Motion for New Trial (Feb. 7, 2020) Georgia v. Denton, Order Granting Extraordinary Motion for New Trial (Feb. 7, 2020)
Hope for both eliminating bitemark testimony and for recognizing that new critiques of forensic methods count as “newly discovered evidence:” a 2020 court decision in Georgia.
Beacuse this is a pdf it is posted on Moodle under "Class 12."
2020 WL 7232303 (Ga.Super.) (Trial Order)
Superior Court of Georgia.
Ware County
STATE of Georgia,
v.
Sheila DENTON, Defendant.
No. 04-R-330.
February 7, 2020.
Order Granting Extraordinary Motion for New Trial
Dwayne H. Gillis, Judge.
*1 Before the Court is Sheila Denton’s Extraordinary Motion for New Trial. The Court has heard evidence, argument, and reviewed the briefs from Mrs. Denton and the State. It is noted that Mrs. Denton’s presence was waived for the hearing on the motion.
Upon conviction for the murder of Eugene Garner, the Defendant filed a motion for new trial, which was denied by the trial court. A direct appeal followed which was affirmed, Denton v. State. 286 Ga. 494: 689 S.E. 2d 322 (2010).
The issues before the Court surround the legitimacy of forensic bite mark evidence. Specific to Mrs. Denton’s case is whether advancements in scientific knowledge regarding bite mark evidence are new evidence warranting the relief sought. The advancements at issue are changes to the ABFO (American Board of Forensic Odontology) guidelines which occurred in 2016. The undisputed evidence at present is that the advancements, scientific understanding and ABFO guidelines would compel a different expert opinion if the case were tried today. Thusly, the seminal question before the Court is whether a change in the expert opinion testimony of a key trial witness, resulting from changes in scientific knowledge and understanding as expressed and contained in the ABFO guidelines is newly discovered evidence.
The Defendant in this case, Sheila Denton, was convicted of murder and sentenced to life in prison in 2006. The evidence against her at trial included testimony by forensic dentist Thomas David, who opined that a bite mark on Denton was probably caused by the victim’s teeth and that a probable bite mark on the victim was probably caused by Denton. (T. 145-80.)1 At the time of Denton’s trial, such testimony was accepted as the product of a credible forensic science discipline.
|
Citations to “T. _” refer to the 2006 trial transcript from this case. Citations to “E.M.N.T. __” refer to the transcript from the 2018 hearing on the extraordinary motion for new trial.
|
On November 8,2017, Denton filed an extraordinary motion for new trial, alleging that recent developments in the understanding of the limitations of bite mark analysis and comparison demonstrate that the testimony today would no longer be inculpatory. An evidentiary hearing was held before this Court on May 29,2018. The Court then invited post-hearing briefing.
Having carefully considered the evidence that was presented at the 2006 trial, the 2018 hearing, and the parties’ briefs, for the reasons that follow, Denton’s motion is GRANTED.
- The Legal Standard
To prevail on her extraordinary motion for new trial, Denton must satisfy the six-factor test elucidated in Timber lake v. State, 246 Ga. 488 (1980). Specifically, she must satisfy this Court “(1) that the evidence has come to [her] knowledge since the trial; (2) that it was not owing to the want of due diligence that [she] did not acquire it sooner; (3) that it is so material that it would probably produce a different verdict; (4) that it is not cumulative only; (5) that the affidavit of the witness himself should be procured or its absence accounted for; and (6) that a new trial will not be granted if the only effect of the evidence will be to impeach the credit of a witness.” Id. at 491.
*2 In granting the extraordinary motion, this Court is cognizant that motions for new trial filed pursuant to O.C.G.A. § 5-5-41(b) are an extraordinary remedy. Mitchum v. State, No. S19A0554, 2019 WL 4924049, at *2 (Ga. Oct. 7, 2019).
- Findings of Fact
At the hearing, Denton presented the live testimony of three forensic dentists-Dr. Thomas David, Dr. Cynthia Brzozowski, and Dr. Adam Freeman. All three experts are Board-Certified Diplomates of the American Board of Forensic Odontology (“ABFO”). The ABFO is the country’s only board-certifying entity for forensic dentists. (E.M.N.T. 42, 143,198.) Their testimony unanimously established that Denton is entitled to a new trial.
Dr. David, who testified against Denton at her trial, admitted that if he testified at Denton’s trial today, his testimony would be different, following advances in the scientific understanding of the limitations of bite mark evidence that are reflected in the recent change in ABFO Guidelines. (E.M.N.T. 126,129-30.) Dr. Brzozowski and Dr. Freeman agreed with that assessment. (E.M.N.T. 165-66,238.) In addition to these live witnesses, Denton also filed, in advance of the hearing, two affidavits, co-authored by a total of five Board-Certified ABFO Diplomates. The first affidavit was co-authored by Dr. Brzozowski, Dr. Anthony Cardoza, and Dr. James Wood; the other was co-authored by Dr. Freeman and Dr. Iain Pretty. Each of the experts agreed that the trial testimony would not be incriminating today based on advances in scientific understanding.
This consensus follows the 2016 change in ABFO Guidelines, which were described as “very significant” (E.M.N.T. 161), and a “seminal moment in the ABFO and a quantum leap” (E.M.N.T. 235). Thus, this Court was presented with unanimous expert opinion that the bite mark evidence presented at trial against Denton would not be incriminating today. The Court found the experts and evidence Denton presented, which challenged the reliability of the bite mark evidence presented at Denton’s trial based on new scientific developments, to be credible.
- The Evidence at Trial
On May 21,2004, the body of Eugene Garner was discovered inside of his home at 1638 ABC Avenue, in Waycross, Georgia. (T. 71, 76,270.) Observation of the scene indicated that a struggle had occurred. (T. 88, 115,129, 374.) The autopsy revealed that the cause of Garner’s death was “blunt force head trauma in conjunction with manual strangulation.” (T. 311.)
Denton was arrested for giving police officers a false name during a stop that occurred around midnight on the day Garner’s body was found. (T. 360, 366.) She was subsequently charged for his murder. At trial, forensic dentist Dr. David testified that an injury found on Denton was a bite mark “to a reasonable scientific certainty,” and an injury found on Garner was “probabl[y]” a bite mark. (T. 154.) Dr. David further opined, “to a reasonable degree of scientific certainty,” that “Denton was the probable source of the bite mark on Mr. Garner and that Mr. Garner was the probable source of the bite mark on Denton.” (T. 159,167.)
Aside from the bite mark evidence, the State presented little other incriminating evidence. An analyst from the Georgia Bureau of Investigation’s Division of Forensic Sciences testified concerning “hairs that were identified as recovered from the victim’s hand” at the scene. (T. 183.) She compared those hairs to “known head hairs” from both the decedent and Denton. (T. 183.) The analyst testified that “the questioned hair in the victim’s hand matched his own head hair.” (T. 187-88.) No hair matching Denton was found at the scene.
*3 The State presented a witness, Sharon Jones, who was an admitted drug addict and crack cocaine dealer with a criminal record. Jones testified that at approximately 4:30 a.m. one day in May of 2004, Denton came to a known crack house where Jones was guarding about $500 worth of crack cocaine. (T. 336-38, 352.) Jones admitted on the stand that she had been smoking a lot of crack cocaine that night, such that days would bleed into one another, making her unable to really say what else had happened that evening. (T. 346-47.) Jones claimed that Denton had cuts on her arms and legs, and Denton told Jones that she “just killed the man that stay in front of Walley’s.” (T. 336-38.) Jones testified that Denton gave no other information about the murder. (T. 338.) Jones testified that she did not believe Denton’s claim that she had killed someone, and instead Jones thought that Denton had been scratched after falling into some bushes. (T. 338-39, 348.) According to Jones, she then sold Denton some crack cocaine. (T. 339.)
Jones testified that when she initially spoke with police, she did not tell the questioning officers any information incriminating Denton. Instead, Jones’s account of Denton’s alleged admission came only after an interrogation during which two police officers “kept saying I did it.” (T. 342.) The interrogation, as retold by Jones at trial, then went as follows:
- They tried to say - well, [Officer] Larry Hill said he had my fingerprints at the house and he had a tape, videotape with me there, which I knew wasn’t true.
- Okay. He said you were involved?
- He said he had my fingerprints at that house.
- Right.
- And a videotape with me being there.
- Did he tell you he was going to put you in jail if you didn’t make a statement to him about it?
- Uh-huh (positive response).
- And he showed you a blank tape and told you he had a videotape of you at Mr. Garner’s house involved in something?
- Uh-huh (positive response).
- And he was going to put you in jail?
- Yes, sir.
(T. 344-45.) Jones’s trial testimony concluded with her admitting that she lied to the police three different times. (T. 354.) None of the other witnesses at trial provided testimony linking Denton to the crime.
- Dr. Cynthia Brzozowski and Dr. Adam Freeman
At the May 2018 evidentiary hearing, Denton called three experts, including Dr. Brzozowski and Dr. Freeman. The Court finds that Dr. Brzozowski and Dr. Freeman were forthcoming and credible in their testimony.
Dr. Brzozowski received her doctor of medical dentistry from the University of Boston Dental School in 1986, completed a residency program at the Veteran’s Administration the following year, has been a fellow with the American Academy of Forensic Sciences since 1994, has been a member of the American Society of Forensic Odontology since 1992, and has served on the board of American Board of Forensic Odontology for three terms. (E.M.N.T. 142-44.)
Dr. Freeman graduated from dental school at Columbia University, completed a fellowship at the University of Texas, is a consultant for the Connecticut Medical Examiner, and has consulted with the FBI. (E.M.N.T. 196-97.) He has served as president, vice-president, and secretary of the ABFO and has chaired the bite mark and metrology committees. (E.M.N.T. 198.)
Drs. Brzozowski and Freeman testified that they came to testify before this Court out of what they termed their “ethical” and “civic” duty. (E.M.N.T. 149,203.) They did so without being paid for their work, and each suffered a financial loss to be away from their respective practices to testify. (E.M.N.T. 149, 203.) They both testified that at one point in their careers in forensic dentistry they believed bite mark evidence could be probative in court. But advances in understanding, recent scientific studies, the lack of proficiency testing, and the growing number of wrongful convictions attributable to bite mark evidence led them to see the error of their past beliefs. (E.M.N.T. 149-50,246-47.) Their testimony was devoid of self-interest bias.
*4 Dr. Brzozowski candidly admitted that when she became a member of the ABFO in 2006, she believed that bite mark analysis was based on valid science; however, she no longer does. Dr. Brzozowski testified:
I think there are several contributing factors that have shaped my opinion over the last decade. The first being that we have no scientific studies to validate accuracy and reliability of bite mark methods. We actually have studies that show the unreliability of bite mark methods, but we have none to show the reliability and the accuracy.
Two, there have been three independent scientific panels who have come to the same conclusion, that being the Texas Forensic Science Commission, the PCAST and the National Academy of Sciences.
Three, the failure of the ABFO to initiate and conduct such studies to validate their methods. And lastly, the increasing number of wrongful convictions that have been largely unaddressed by the ABFO.
(E.M.N.T. 149-50.) She then testified that at the time of Denton’s trial, individualization and probabilistic testimony2 from forensic dentists was supported by the ABFO. (E.M.N.T. 153.) In 2016, however, in response to the factors that Dr. Brzozowski referenced in her testimony, the range of permissible associations changed. (E.M.N.T. 153.)
|
“Individualization” and “probabilistic association” are conclusions that “an individual is responsible or more likely than not responsible for inflicting a bite to the exclusion of all other potential sources.” (E.M.N.T. 152-53.)
|
Dr. Brzozowski testified that at a threshold level, when determining the nature of the injury, “probable bite mark”-the characterization Dr. David provided at trial for the injury to Garner-is no longer a permitted conclusion by the ABFO Guidelines. (E.M.N.T. 157-58.) Dr. David’s remaining trial testimony in this case was inconsistent with today’s scientific understanding of bite marks as well. (E.M.N.T. 165.) “He testified that Ms. Denton was the probable biter of Mr. Garner and that Mr. Garner was the probable biter of Ms. Denton. And today we would not testify as to a probabilistic association.” (E.M.N.T. 165.) Rather, the strongest conclusion that could be offered today would be “that one could not exclude that individual as being the source.” (E.M.N.T. 165.)
Dr. Brzozowski further testified that she reviewed photographs of the evidence in this case and determined that there was “insufficient information to determine that either of these injuries are human bite marks.” (E.M.N.T. 170.) Accordingly, Dr. Brzozowski would not make any comparisons between the injuries and any dentition under today’s Guidelines. (E.M.N.T. 173.)
The Court also credits Dr. Brzozowski based on her demeanor. She testified that this testimony was not easy for her to give, telling the Court that she took “pride in being a diplomate of the ABFO ... invested 12 years in [the] organization. But the failure of the ABFO to address the impact of scientifically invalidated bite mark methods on people’s lives and liberty causes [her] great concern.” (E.M.N.T. 177.) The Court notes that this statement was consistent with Dr. Brzozowski’s demeanor in court. It was clear to the Court that it was difficult for Dr. Brzozowski to testify and that she felt a professional obligation to do so. The Court finds her testimony credible.
*5 Dr. Freeman corroborated Dr. Brzozowski’s testimony. Dr. Freeman testified that he became a member of the ABFO in 2009, and eventually served as the president of the organization in 2016. (E.M.N.T. 198.) When he first began studying bite marks in 2003, he learned about individualization and probabilistic testimony, and accepted such testimony at the time as based on valid science. (E.M.N.T. 204-05.) However, Dr. Freeman no longer holds that belief today. (E.M.N.T. 237.) Dr. Freeman testified that his belief changed, in large part, after the results of a study he conducted, along with Dr. Pretty, called the Construct Validity Test. The Construct Validity Test was designed to answer the following threshold question in bite mark analysis: whether Board certified odontologists could reliably determine if an injury is a bite mark or not. (E.M.N.T. 217-19.) He testified that the study revealed that the answer is no: “there was a wide spread disagreement about what constituted a bite mark.” (E.M.N.T. 214.) Dr. Freeman testified that, in part because of his Construct Validity Study, the ABFO Guidelines changed to reject probabilistic testimony. (E.M.N.T. 235.) He classified this change as a “seminal moment” and a “quantum leap.” (E.M.N.T. 235.)
Dr. Freeman testified that he reviewed the bite mark evidence in Denton’s case, as well as Dr. David’s testimony. (E.M.N.T. 237-38, 242.) He further testified that-under the current ABFO Guidelines-Dr. David’s trial testimony concerning probability would be “a misrepresentation of what the underlying science supports.” (E.M.N.T. 238.) Dr. Freeman testified that it was his expert opinion that neither mark in Denton’s case constituted a bite mark. (E.M.N.T. 243-46.) Dr. Freeman presented credibly and answered questions in a clear and unbiased manner. His tone and demeanor indicated that he had genuine concern about errors in the use of bite mark evidence and that he felt a professional obligation to inform the Court about those concerns. The Court credits Dr. Freeman’s expert testimony as reliable, informative, and devoid of self-interest.
- Dr. Thomas David
This Court found Dr. David to be less forthcoming, with a self-interest that became notable to the Court. Dr. David testified that he wanted to “control the narrative” in this case. (E.M.N.T. 113.) He sometimes failed to answer questions directly. (See, e.g., E.M.N.T. 59-63 (testifying that giving indirect answers to questions provides the witness with “wiggle room”).) Moreover though, Dr. David noted the only thing the [new] guidelines really changed was the terminology used in his analysis [EMNT. 127]. The Court finds this statement incredulous in light of the evidence produced in this case. For instance, Dr. David’s trial testimony that an injury to the victim was a probable bite mark (EMNT. 70-71, 87) and that the Defendant was the probable source of the injury to the victim to a reasonable degree of scientific certainty (emphasis added) (EMNT. 72-73,77), yet under the new guidelines he could not even conclude that that injury was a bite mark, much less to a comparison (EMNT. 129). The defense testimony and evidence in this case is profound and leads one to conclude that bite mark evidence is no longer a valid subject of scientific inquiry unless it can exonerate one or exclude a particular person. To describe the change in the ABFO stance on bite mark evidence as one involving mere “terminology” is at best disingenuous. The Court credits Dr. David’s testimony only to the extent it is corroborated by the other credible evidence and expert testimony.
On the most critical questions in the case, Dr. David agreed with the other experts who testified at the hearing. Specifically, when questioned by the Court and Denton, Dr. David admitted that today his testimony would be different than the testimony he gave at Denton’s trial. (E.M.N.T. 126.) Dr. David admitted that, today, he would not even compare the marks on the decedent with the teeth of Denton, because it is uncertain as to what caused the injury at all:
Q: In this case when you testified, you testified that the mark on Mr. Garner was a probable bite mark.
A: Correct.
Q: Under today’s guidelines - applying today’s guidelines, you wouldn’t do a comparison.
*6 A: That is correct, under the current guidelines.
(E.M.N.T. 129.)
The Court finds that Dr. David’s testimony supports Denton’s position-that the ABFO Guidelines changed since the time of the trial in this case, and if testifying today, Dr. David would follow the Guidelines i.e., his testimony today would be different than his original trial testimony.
- Application of the Facts to the Law
- The Evidence Came to Denton’s Knowledge Since the Trial.
This Court finds that the recent changes in the field of bite mark analysis, which led to the 2016 change in the ABFO Guidelines, and the corresponding change of the trial testimony in Denton’s case, have come to Denton’s knowledge since trial. In 2016, for the first time in the forty-year history of the ABFO, the organization prohibited probabilistic or individualization testimony in the field of bite mark analysis and comparison. This change, which occurred after Denton’s trial, means that the evidence offered at trial would be altogether different today.
The evidence at trial was that Denton was the “probable biter” of the “probable bite mark” on Garner and that Garner was the “probable biter” of the “bite mark” on Denton. (T. 154, 159,167.) At trial, Dr. David rendered those opinions with a reasonable degree of scientific certainty. Now, however, Denton’s experts and, ultimately, Dr. David, agree that, under the new Guidelines, the following new expert testimony would be presented at a trial today:
1) Dr. David would no longer compare the injury to Garner to any suspected dentition;
2) Dr. David would no longer testify that the injury to Garner was a bite mark;
3) Dr. David would no longer testify that Denton was the “probable biter”;
4) The best Dr. David could say is that Garner “could not be excluded” as the source of the injury to Denton.
(See, e.g., E.M.N.T. 75-78.) Today, therefore, the most that could be offered, consistent with the best practices reflected in the ABFO Guidelines, is that any person’s given dentition cannot be excluded as one of innumerable possible dentitions to have potentially inflicted a given injury. Such an opinion will seldom, if ever, be probative of one having inflicted a particular bite mark, nor shall it likely be of any aid to a jury in reaching a decision. The future admissibility of such evidence is dubious at best.
The changes to the ABFO Guidelines constitute newly discovered evidence, as courts across the country have found. See, e.g., Ex Parte Chaney, 563 S.W.3d 239,260 (Tex. Ct. Crim. App. 2018) (“[T]he 2016 ABFO Manual has completely invalidated any population statistics, regardless of whether the population is open or closed ... the Manual no longer allows examiners to give opinions to a ‘reasonable degree of dental certainty.”’); State v. Hill, 125 N.E.3d 158,168 (Ohio Ct. App. 2018) (the “trial court’s own statements acknowledge[ed] that the [discrediting of] bite mark evidence is newly discovered evidence”); Commonwealth v. Kunco, 173 A.3d 817, 821-22 (Pa. Super. Ct. 2017) (citing post-conviction testimony by three ABFO Diplomates that acknowledged that the change in the ABFO Guidelines represents a change in scientific understanding of the limitations of bite mark evidence, and that it is impossible to know how many people might be associated with an alleged bite mark).
*7 The Texas Court of Criminal Appeals recently set aside a conviction that had been based on bite mark evidence. Ex Parte Chaney, 563 S.W.3d 239 (Tex. Ct. Crim. App. 2018). The Texas court reasoned that “the body of scientific knowledge underlying the field of bite mark comparisons evolved in a way that discredits almost all the probabilistic bite mark evidence at trial.” Id. at 257. In reaching the conclusion that there was new, material evidence before the court, the Chaney court cited to the National Academy of Sciences Report, the new ABFO Guidelines, and studies and affidavits by forensic dentists similar to those that Denton has cited to this Court.3 And the State of Texas conceded that the new “‘bite mark evidence, which once appeared proof positive of... Chaney’s guilt, no longer proves anything.”’ Id. at 258. Likewise here-the new bite mark testimony no longer proves anything.
|
Denton and the expert witnesses cited the Court to governmental reports that addressed bite marks. They confirm and corroborate the significance of the new research Denton presented at the hearing.
|
At the hearing, and in its post-hearing brief, the State advanced two counterarguments. The Court is not persuaded by either argument.
First, the State argued that Dr. David’s testimony today would be unchanged. However, the State also conceded in its post-hearing brief, that “under the new ABFO Guidelines [Dr. David’s] opinion about the injuries on the defendant would change from the [decedent] being the ‘probable biter,’ to the [decedent] ‘could not be excluded’ as the source of the bite mark.” (State’s Br. at 17.) In addition, the prosecution agreed that “the probable bite mark [Dr. David] found on the victim would be considered inconclusive under the new guidelines, and [Dr. David] should not do a comparison on that injury unless he could explain his reasons for deviating from the new guidelines.” (State’s Br. at 17.) These concessions undermine the prosecution’s position that the testimony would be unchanged. Therefore, the Court credits the testimony of the expert witnesses who testified at the hearing (including Dr. David, who was also the State’s witness at Denton’s trial), who agreed that, under the new Guidelines, the dental testimony at trial today would be fundamentally different. This constitutes new evidence.
Second, the prosecution argued that the change in the ABFO Guidelines, which led to the change in the opinion Dr. David would render in this case and the opinion of Denton’s forensic dentists, does not constitute “new” evidence. (State’s Br. at 16.) The prosecution claimed that “[c]hanges or updates in scientific guidelines that are used by experts, especially when these changes are not to be applied retroactively, are not new evidence.” (State’s Br. at 16.)4 The Court does not find the prosecution’s argument persuasive. The prosecution cited no authority in support of this contention, and the Court has found none. Rather, new scientific Guidelines that create a consensus among prosecution and defense forensic experts that the forensic evidence is no longer inculpatory, is a prime example of “new” evidence. Moreover, a focus on scientific policies, guidelines, techniques, and procedures are not key evidence upon which convictions are based, it is the “expert opinion” evidence generated from the application of such policies, guidelines, techniques, and procedures. This record is clear that the expert opinion evidence has changed. Additionally, it is clear that the ABFO guidelines changed due to the realization that bite mark evidence is inherently unreliable.
|
This Court notes that it is the role of jurist to decide how changes in testimony by expert witnesses in compliance with changes in scientific understanding, procedures, techniques, and procedures shall be applied (i.e., retroactively as newly discovered evidence, as in the present case). Of course, scientist and boards may elect to apply changes and new techniques, understanding and procedures retroactively to their field of endeavor or not and such has no bearing on the courts. It is for the courts to decide the impact of such changes.
|
- It Was Not Owing to the Want of Due Diligence that Denton Did Not Acquire the New Evidence Sooner.
*8 The Court finds that Denton acted with due diligence in filing her extraordinary motion for new trial for a number of reasons. First, she filed in 2017, following the change in the ABFO Guidelines, which occurred in 2016. The ABFO has been in existence for more than forty years, and that change marked the very first time in its history that probabilistic and individualization testimony, like that offered by Dr. David at trial, was disallowed. By filing as soon as she did after that “seminal” change, Denton acted with due diligence.
Second, Denton filed her motion in this case prior to numerous decisions that have since granted relief based on the change in the ABFO Guidelines and new understanding regarding the limitations of bite mark evidence. See, e.g., Order on Defendant’s Motion to Stay Execution of Sentence, Commonwealth v. Gary Cifizzari, No. 8385-CR-4051, at 2 (Super. Ct. Worcester Co. July 12, 2019); State v. Roden, 437 P.3d 1203 (Or. 2019); Ex Parte Chaney, 563 S.W.3d 239 (Tex. Ct. Crim. App. 2018).
Third, the Court is aware of the many, and increasing, exonerations of people wrongfully convicted based on bite mark evidence of the sort offered against Denton in this case. See Innocence Project, https://www.innocenceproject.org/wp-content/uploads/2019/01/Description-of-bite-mark-exonerations-and-statistical-analysis_UPDATED-01.28.19.pdf (last visited Nov. 12,2019). As far as this Court can tell, none of those exonerations are from Georgia, so she is among the first in the state to raise and litigate this claim.
Fourth, the Court notes that since the hearing in this case, an additional exoneration occurred as well. On July 12, 2019, Gary Cifizzari was released from prison pending further proceedings pursuant to a motion for new trial he filed challenging his 1984 murder conviction. See “Cifizzari gets out of prison; seeks new trial in “79 Milford slaying,” Telegram and Gazette (July 12, 2019), available at https://www.telegram.com/news/20190712/cifîzzari-gets-out-of-prison-seeks-new-trial-in-79-milford-slaying. The same Dr. Freeman and Dr. Pretty who offered affidavits and testimony in this case also offered affidavits in Cifizzari’s case. In granting Cifizzari’s release, the judge cited the forensic odontologists who are “condemning bite mark comparison evidence as there is no scientific basis for forensic odontologists to proffer testimony concluding that a suspect is ‘the biter’ to the exclusion of all others.” Order on Defendant’s Motion to Stay Execution of Sentence, Commonwealth v. Gary Cifizzari, No. 8385-CR-4051, at 2 (Super. Ct. Worcester Co. July 12, 2019) (emphasis added).
In light of all the above, the Court finds Denton exercised due diligence in filing an extraordinary motion for new trial in 2017, following the significant 2016 change in the ABFO Guidelines.
- The New Evidence Is So Material that It Would Probably Produce a Different Verdict.
The Court has reviewed the entire trial record in this case in light of the new evidence. After that review, the Court finds the materiality prong of Timberlake satisfied because the new evidence is so material that it would probably produce a different verdict.
First, the prosecution conceded at trial that without the bite mark evidence, Denton would have been acquitted. In his closing argument, the prosecutor argued that if not for the testimony that the decedent inflicted a bite mark on Denton, “I’d have to admit that’s reasonable doubt.” (T. 434.) As discussed above, the new evidence provided by the experts before this Court established that the bite mark evidence used at trial is now known to be unsupported by science. Rather than purportedly tying Denton to the crime of which she was convicted, the undisputed forensic evidence today would not do so. Dr. David’s trial testimony-consistent with science at the time it was given but inconsistent with science today-was powerful evidence that the prosecutor relied on in urging the jury to convict. Today, the evidence would not support the conclusions that the prosecution relied on to secure a conviction. Without such testimony, the Court agrees with the prosecutor’s statement at trial: that Denton probably would be acquitted today.
*9 In its post-hearing brief, the State argues that the new evidence is not material. In making this argument, the State relies on Sharon Jones’s trial testimony, including her statements to the police. The Court does not find this argument persuasive.
Jones’s claim about Denton’s alleged admission came after she herself was accused of murder during an interrogation in which two police officers, according to Jones, “kept saying I did it.” (T. 342.) Jones did not originally implicate Denton, she only implicated Denton after being pressured by the police, she admitted to having been under the influence of drugs around the time of the offense, and she admitted, under oath, that she had lied to the police. (T. 336-38, 346-47,352, 354.)
Significantly, the prosecutor did not rely on Jones’s statement in his closing argument to the jury. Instead, the prosecutor essentially undermined the statement’s credibility during closing argument. He stated:
Sharon Shontae Jones ... [tjhat’s who she’s smoking crack with. I guess, we could have checked the churches, but I guess we couldn’t find anybody there, so we had to get the ones who actually were with her that morning. It would have been easier to go to a church and find people who could come and testify to these things, but they’re not out smoking crack cocaine in the early morning hours.
(T. 437.) The prosecution informed the jury that it could not pick its witnesses, but had to live with the witnesses it had, despite their incredibility. Because ofthat, the prosecution explicitly asked the jury to convict based on the bite mark evidence, saying, “If that’s all we had, I’d have to admit that’s reasonable doubt-----But then we’ve got also the bite mark on Sheila Denton.” (T. 434.) Employing that logic, and applying the new ABFO Guidelines, today Denton probably would be acquitted.
The State presented no other forensic or physical evidence connecting Denton to the crime. There was no DNA evidence presented. While hair was found in the victim’s hand, the State’s trial expert determined that the hair belonged to the victim; Denton’s hair was not found on the scene. (T. 183,187-88.) The only other evidence against Denton was that she ran from the police and gave a false name when arrested. These facts do not alter the Court’s materiality analysis, however, because they provide no independent affirmative evidence that Denton committed a murder. The materiality prong is satisfied.5
|
In addition, Dr. David himself has admitted the persuasive power that bite mark evidence has on a jury once it is admitted at trial with the stamp of authority from a witness deemed to be an expert by the court. In a published article in a respected scientific journal, Dr. David wrote that without his testimony, a defendant against whom he testified would never have been convicted. (E.M.N.T. 59-61, 80.) Likewise, here.
|
In addition, as recently as March of 2019, a court, with the concession of the State, found that bite mark evidence should not have been admitted in a child murder case, and reversed the conviction. State v. Roden, 437 P.3d 1203 (Or. 2019), involved a murder and child abuse case in which testimony indicated that three children suffered various adult bite marks to several different body parts. Id. at 1208. On appeal, the prosecution conceded that admission of the bite mark testimony was erroneous. And the Oregon court agreed:
*10 [T]he state concedes that the trial court erred in admitting that evidence ... The state’s concession is appropriate ... The state’s odontologist claimed that bite mark identification had undergone testing and peer review and had an error rate of “zero,” but he admitted that he could not cite a single peer-reviewed study testing and validating the technique, and he was unable to support his assertion regarding error rates. Meanwhile, defendant cited studies highlighting concerns within the scientific community regarding the high rate of error and lack of objective, standardized results in bite mark analysis and identification.
Roden, 296 Or. App. at 615-16 (internal citations omitted). Accordingly, the court reversed convictions for felony murder, murder by abuse, and first-degree manslaughter. Id. at 1213.6
|
Although this issue is not presented before this Court and so the Court does not reach it, the evidence and expert testimony presented to the court at the evidentiary hearing was so credible that the Court doubts whether bite mark evidence would today satisfy the test articulated in Harper v. State, 249 Ga. 519 (1982).
|
The recent opinion in Ex parte Chaney is also instructive as to the question of whether the verdict today probably would be different. Last year, the Texas Court of Criminal Appeals found the following:
[The] 2016 ABFO Manual has completely invalidated any population statistics, regardless of whether the population is open or closed ... The body of scientific knowledge underlying the field of bite mark comparisons has evolved since his trial in a way that contradicts the scientific evidence relied on by the State at trial. New peer-reviewed studies discredit nearly all the testimony given by [the ABFO-certified dentists] about the mark on [the victim’s] left forearm and [Petitioner] being a “match.”
Chaney, 563 S.W.3d at 260-61. At Chaney’s trial, the prosecutor, in closing, argued:
But, most of all, we have the bite mark. I wouldn’t ask you to convict just based on the testimony of the tennis shoes, of the statements Chaney made to Westphalen, or the statements he made to Curtis Hilton. But, by golly, I’m going to ask you to convict on that dental testimony.
Id. at 262. That closing argument bears resemblance to the closing argument the prosecutor made at Denton’s trial, namely, “If that’s all we had, I’d have to admit that’s reasonable doubt. ... But then we’ve got also the bite mark on Sheila Denton.” (T. 434.) In each case, there was other evidence-including potentially inculpatory statements-that arguably tied the defendant to the crime, but in each case the “linchpin” of the case was the bite mark evidence. See id. at 262. Thus, the Chaney court granted relief, finding that the prosecution’s case would have been “incredibly weakened” had the “newly available scientific evidence been presented at his trial.” Id.
Just like in Chaney, at Denton’s trial the prosecution argued the necessity of the bite mark evidence to a conviction. Just like in Chaney, the prosecution’s case would have been incredibly weakened by the newly available scientific evidence. Accordingly, just like in Chaney, the outcome of trial today probably would be different, and Denton is entitled to a new trial.
- The New Evidence Is Not Cumulative Only.
The Court finds that the new evidence is not cumulative and further notes that the prosecution has not argued otherwise.
- Affidavits and Live Testimony Were Presented.
The Court finds that Denton submitted two affidavits and presented live testimony in support of her request for a new trial, thereby satisfying this Timberlake factor.
- The Effect of the New Evidence Is Not Only To Impeach the Credibility of a Witness.
*11 At trial, Dr. David testified that “Denton was the probable source of the probable bite mark” on Eugene Garner and that “Mr. Garner was the probable source of the bite mark on Denton.” (E.M.N.T. 72.) Dr. David further testified that he held those beliefs “to a reasonable degree of scientific certainty.” (E.M.N.T. 72-73.) The new evidence does not simply impeach those conclusions; rather, the new evidence demonstrates that Dr. David would not draw those conclusions, nor testify to those conclusions today. Instead, Dr. David would follow the ABFO Guidelines. As a result, he would not do a comparison between Denton’s dentition and the injury to Garner at all, because he would not conclude that the injury to Garner was a bite mark in the first place. (E.M.N.T. 129.) As to the mark on Denton, the most Dr. David would offer is that any person’s dentition, including Garner’s, could not be excluded as one of an unquantifiable number of possible dentitions that could have inflicted the injury. (E.M.N.T. 77.)
The following exchanges occurred at the hearing during the examination of Dr. David, and are determinative of this prong:
Q: Under today’s guidelines - applying today’s guidelines, you wouldn’t do a comparison [to the injury on Mr. Garner]?
A: That is correct, under the current guidelines.
(E.M.N.T. 129.)
Q: Today, consistent with the guidelines, the strongest conclusion you could come to would be that Mr. Garner could not be excluded as having made the bite mark [on Sheila Denton], correct?
A: That he couldn’t be - yes, correct.
Q: That would be the strongest association you could make under today’s guidelines, correct?
A: Under the current guidelines, that is correct.
(E.M.N.T. 77.)
The other experts presented at the hearing reiterated these inarguable conclusions in straightforward testimony. For example, with respect to the injury to Mr. Garner, Dr. Brzozowski testified:
Q: What if you determine that the injury is a probable bite mark?
A: You stop. No further analysis. No comparison is done.
Q: Is that even a permitted conclusion today, probable bite mark?
A: No, it’s not.
(E.M.N.T. 157-58.)
In other words, the testimony at trial was that Denton was the probable source of the probable bite mark on Garner, an opinion Dr. David held to a reasonable degree of scientific certainty. Today, all that could be said of the injury is that it is “inconclusive” as to how it was inflicted; no comparison would be done to any dentition.7 Rather than a “bite mark,” the injury to Garner, today, would be called just that, an “injury.” That is not impeachment, it is substantively different expert testimony that would not incriminate Denton.
|
The overlays that were admitted at the original trial, purporting to sketch out the work Dr. David did during the comparison phase of his analysis, are irrelevant to the question before this Court for this reason. (See, e.g., E.M.N.T. 191.) With respect to Garner’s injury, there would be no comparison to undertake.
|
Dr. Freeman analogized to another feature-comparison discipline at the hearing:
[W]e should all have some level of agreement as to whether or not the injury we are looking at, the threshold, that that injury is a bite mark. And as an example I would give, fingerprint examiners often will disagree about who has made that fingerprint, who attributed that fingerprint. They don’t disagree over whether or not it is a fingerprint.
And if... we can’t establish step one, that something is a bite mark, if you’re wrong about that step then surely any conclusion you make attributing a person to making that bite mark would be faulty.
(E.M.N.T.209,216.)
Likewise, an opinion that someone is the “probable biter” is dramatically different from an opinion that someone is in a class of an innumerable number of people who potentially could have inflicted the injury. With respect to the injury to Denton, Dr. Brzozowski testified:
Q: What would be the strongest conclusion that could be offered consistent with the guidelines today?
A: The strongest conclusion was that one could not exclude that individual as being the source.
*12 Q: You said you can not exclude as opposed to probable biter, which would be different conclusions?
A: Yes.
Q: Is that a matter of semantics or is that a difference, a significant difference?
A: Well, it is a significant difference because can not exclude may mean that there might be thousands of other people that could possibly have inflicted the injury.
(E.M.N.T. 165-66.) Dr. Freeman characterized it as follows:
Making a statement like someone is the probable biter [the testimony at trial], that is significant, versus saying I can not exclude this person as having made this bite mark [the testimony now]. That indicates that the population that could have made the bite mark is very wide spread.
(E.M.N.T. 235.)
In other words, the expert testimony at trial, to a reasonable degree of scientific certainty, was that Garner was the probable source of the bite mark on Denton. Today, even assuming that the injury is a bite mark (a conclusion not supported by the expert testimony before this Court), the most that could be said is that Garner could be one of innumerable people whose dentition could have inflicted that injury. That is not impeachment; it is qualitatively different testimony that no longer implicates Denton to the exclusion of myriad suspects.
Conclusion
When evidence which is deemed to be “scientific” or “expert”, it transcends the normal understanding of the lay juror. Courts generally admit scientific evidence and the expert opinions based thereon, if the Court affirmatively determines, “whether the technique has gained general acceptance in the scientific community which recognizes it,” Harper v. State, 249 Ga 519,524-525 (1982). The Court in Harper went on to state, “An evaluation of whether the principle has gained acceptance will often be transmitted to the trial court by members of the appropriate scientific community testifying as expert witnesses at trial.” Id. at 525. What then should the Court do when the experts testify that the scientific principle has lost acceptance in the appropriate scientific community? In the present case, the loss of acceptance within the field of forensic odontology is not disputed. The Harper test is still the law under the new evidence code. Winters v. State, 305 Ga. 229 (2019). The object of the Court’s inquiry in Harper was to determine “whether a scientific principle or technique is a phenomenon that may be verified with such certainty that it is competent evidence in a court of law...” Id. At 524. It stands to reason then, that the Harper test shall serve to determine that a scientific principle or technique has become a phenomenon that can not (emphasis added) be verified with the requisite certainty and that it is incompetent evidence in a court of law.
The ABFO’s recognition of the inadequacies of bite mark evidence is profound and undisputed. We now know the bite mark evidence presented in Denton’s trial was not competent evidence under Harper. The State’s attempt to cast this case as involving mere changes in scientific guidelines misses the point. The changes in ABFO guidelines are clearly intended to represent the state of scientific understanding in the field. These guidelines even limit the conclusions that an ABFO diplomat may testify to, presumably due to the inadequacies of bite mark evidence.
*13 An analysis or determination that a given scientific principle or technique is no longer generally accepted in the scientific community, will by definition most often involve a post-trial review. With the current understanding of bite mark evidence in the field of forensic odontology, the strongest testimony that can be rendered, in favor of the prosecution, is that the defendant cannot be excluded as the biter. Such testimony will likely be deemed highly prejudicial and of little probative value in the future. With the future of such expert opinion and testimony in serious doubt, the question arises as to whether a Harper analysis can and should be allowed retroactively post-trial, as newly discovered evidence. The Court has answered that question in the affirmative. However, another issue is whether due process would dictate that the Court undertake a reverse Harper type analysis where prior scientific evidence, principles, techniques and testimony are essentially de-ligitimized. This analysis may and should occur without the constraints of Drane v. State, and our extraordinary motion for new trial body of law. Proven unreliable scientific evidence should never serve as the basis of a conviction and should be dealt with by the Courts if and when it is found. Applying such an analysis to the facts of this case, it is uncontroverted that bite mark analysis and testimony as existed at the time of Denton’s conviction has been proven to be unreliable and not generally accepted within the scientific community of forensic odontology. Therefore, the Court grants the Defendant’s motion pursuant to Drane and a reverse Harper analysis.
WHEREFORE, Denton’s Extraordinary Motion for New Trial is hereby GRANTED. The Defendant shall be remanded to the custody of the Sheriff of Ware County, when this order becomes final.
Date: 2-7-20
<<signature>>
Dwayne H. Gillis, Chief Judge
Superior Court, Ware County
Waycross Judicial Circuit
Proposed legislation in California
SB-467 Expert witnesses: writ of habeas corpus.
According to the Bill's sponsor, Senator Scott Wiener, this law "clarifies that the definition of false testimony includes opinions based on flawed scientific research or outdated technology that is now unreliable or moot, and opinions about which a reasonable scientific dispute has emerged regarding its validity. SB 467 also clarifies that expert opinions that fail to use valid methodology, research, peer-reviewed studies, or scientifically sound data do not satisfy the requirements for admissible testimony."
Writing Reflection #13 Writing Reflection #13
Please go to our Moodle Page and under "Class 13" you will find the prompt and submission folder for Writing Reflection #13.