Forensics Under the Microscope
More courts are starting to question the facts proved by scientific evidence.
Beth Schwartzapfel
Newsweek
February 17, 2011
Warren Horinek did not murder his wife. Thats what he said, thats what the medical examiner said, thats what the homicide sergeant said. Even the district attorneys office in the Horineks hometown of Ft. Worth, Texas, agreed that he was innocentnot something a Texas prosecutor typically says. But when Bonnie Horinek died in 1995, her parents refused to believe what the evidence strongly suggestedthat Bonnie shot herselfand instead they enlisted the services of a blood-spatter analyst to prove that it was their son-in-law who had killed their daughter.
The spatter analyst zeroed in on the blood-soaked T shirt Horinek was wearing when the paramedics arrived. To him, the fine spray of blood on Horineks left shoulder was not from administering CPR, as Warren said it was, and as the 911 recording seemed to indicate, but from shooting Bonnie at close range. On the basis of that testimony, Horinek was convicted of murder and sentenced to 30 years. But did they really get their man? Horineks lawyers have filed a writ of habeas corpus to try to have him released, based in part on the National Academys report; much of the spatter analysts testimony, the lawyers argue, was contrary to known and accepted science.
In the age of CSI and Dexter, were led to believe that forensic science is a high-tech discipline, powerful and sophisticated enough to catch any criminal.
As it turns out, whether blood-spatter analysis and disciplines like it qualify as science at all is a matter of increasing debate. In a sharply critical report issued in 2009, the National Academy of Sciences said, The simple reality is that the interpretation of forensic evidence is not always based on scientific studies. Taking aim at disciplines as varied as ballistics, hair and fiber analysis, bite-mark comparisoneven fingerprintsthe report declared, This is a serious problem.
The last few years have seemed to bear out the report. Dozens of elite crime labs all over the country, from Nassau County, N.Y., to San Francisco, to Virginia, Cleveland, Oklahoma, and Baltimore, have been involved in scandals involving mishandled evidence and false or misleading forensic testimony. This past summer, a North Carolina attorney generals audit discovered that the states Bureau of Investigation had withheld or distorted evidence in more than 200 cases.
Even some of the best funded and most sophisticated crime-fighting organizations are being taken to task for their use of forensic evidence. This week, the New York Times reported that the Federal Buerau of Investigation had overstated the strenght of genetic analysis during the investigation of Bruce E. Ivins, who allegedly mailed anthrax to newsrooms and Senate offices in the wake of the 9/11 attacks.
A year-long investigation by the independent journalism nonprofit ProPublica revealed major problems in the nations coroner system: pathologists not certified in pathology, physicians who flunk their board exams, even coroners who are not physicians at all. In nearly 1,600 counties across the country, the investigation found, elected or appointed coroners who may have no qualifications beyond a high-school degree have the final say on whether fatalities are homicides, suicides, accidents or the result of natural or undetermined causes.
For his forthcoming book, Convicting the Innocent: Where Criminal Prosecutions Go Wrong (Harvard University Press, April 2011), University of Virginia law professor Brandon Garrett examined the trial transcripts and other legal documents of the first 250 people to be exonerated by DNA in this country. He discovered that in more than half these cases, trials were tainted by invalid, unreliable, concealed, or erroneous forensic evidence. The errors ranged from analysts making up statistics on the fly, implying that their methods were more scientific than they actually were, and exaggerating or distorting their findings to support the prosecution.
Peter Neufeld, a lawyer in New York and cofounder of the Innocence Project, which has helped to facilitate many of these exonerations, calls it the elastic expert: no matter what you see, I can distort it so that it would be a match.
This elasticity is possible because the tests are largely subjective. Just how much human judgment is required depends on the discipline: DNA testing is mostlythough not entirelydone by machine, for instance, whereas microscopic hair comparison is based solely on the analysts opinion. Even fingerprints, which many of us regard as foolproof tools for identifying culpritsthink Dexter feeding a print into his computer and a bad guys photo and drivers license appearing on the screenin fact rely largely on human interpretation, and therefore are subject to human error.
One of the most famous examples of the danger of fingerprints was the case of Oregon lawyer Brandon Mayfield, arrested in 2004 in the wake of the Madrid train bombings. Working from a partial print that Spanish authorities had found on a plastic bag of detonators, several top FBI analysts declared Mayfields print a match. That is, until Spanish authorities identified Ouhnane Daoud, now wanted for terrorism in connection to the crime. When it became clear that Daouds prints were a much better match, the FBI was forced to admit that its own bias and circular reasoning had led them to Mayfield, who had no involvement in the bombings.
Part of the problem is what social scientists call context bias. Most forensics labs are located within police departments, so analysts may see themselves as working for the prosecution. They also usually have information about the evidence theyre testingfor example, that the suspect has a prior record. Theres a lot of research to suggest that knowledge could have biasing effect, says Jennifer Mnookin, a professor at the UCLA School of Law.
In a recent Supreme Court case, Justice Antonin Scalia, writing for the majority, said that whether consciously or not, an analyst responding to a request from a law enforcement official may feel pressureor have an incentiveto alter the evidence in a manner favorable to the prosecution. The judges ruling means that forensic test results may be subject to the same kind of scrutiny as any other evidence, and an analyst from the lab that ran the test must be present in court to be cross-examined, just like any other witness.
Obviously, most people in this community are trying to do their jobs well and are not trying to frame innocent people, says the University of Virginias Garrett. But what weve seen come out of these exoneration cases and in additional scandals at the laboratories is that this is not a problem of a few bad apples. Who is the competent analyst that can testify about a technique thats fundamentally unreliable? Thats not a bad-apple problem. Thats a serious problem with our entire system.
At the heart of these criticisms is the issue of what scientists call validity and reliability. A test is valid if its results are factually accurate. A test is reliable if multiple tests will lead to the same conclusion. Some forensics tests, like blood typing, are very reliable: no matter how many times your doctor draws your blood, you will always have the same blood type. Occasionally there are mistakes, of course, but they are predictable: blood-typing tests have well-documented and well-understood error rates. Others, like hair comparison, are unreliable: studies have shown that multiple technicians examining the same two hairseven the same technician examining the same two hairs at different timescome to multiple conclusions. Critics say that many of forensic sciences most basic tools are neither reliable nor valid.
For example, at the trial of Jimmy Ray Bromgard, who served more than 14 years of a 40-year sentence for sexual intercourse without consent until he was exonerated in 2002, the director of the Montana State Crime Lab told the jury that hairs found on a blanket in the victims house matched hairs taken from Bromgards body. There were so many hairs that matched so well, the analyst said, that there was a one in 10,000 chance the hairs could have come from anyone else.
But no one has ever established any statistics about the microscopic characteristics of hair, so one in 10,000 odds isnt based on scientific consensus. How common is it for a person to have a particular hair color, or for a hair to crinkle or curl just so? Scientists have never answered that question systematically. And what does match mean, anyway? There are no uniform guidelines to say how many characteristics two hairs must have in common before theyre said to match. It varies entirely from one lab to the next, from one technician to the next.
Barry Fisher, who served as the crime-laboratory director for the Los Angeles Sheriffs Department for more than 20 years, was often stymied by this problem when he took the stand. How do you convey the level of certainty? Fisher asks. Do you say to the jury, Im pretty sure? Im very sure? What do these things mean?
To get around this problem, Garrett found, forensics experts too often overreach in their testimony.
When Ray Krone was convicted of murder in Arizona and sentenced to death in 1995, the testimony of a bite-mark analyst was key to the states case. This is really an excellent match, the analyst said on the stand, comparing Krones teeth with a bite mark on the murder victim. That tooth caused that injury.
In fact, in its report the National Academy of Sciences found that, among all the forensic disciplines, only DNA has proved capable of individualizationthat is, demonstrating a connection between evidence and a specific individual or source. When the DNA in the Krone case was tested year later, he was exonerated, but only after spending a decade in prison.
The report has led a small but growing number of judges to take a more skeptical approach to forensics. In addition to the Supreme Court case, Melendez-Diaz v. Massachusetts, U.S. District Court Judge Nancy Gertner announced in March that she will allow forensic evidence in her courtroom only if a lawyer first proves in a pretrial hearing that the method is scientifically sound. In the past, the admissibility of this kind of evidence was effectively presumed, largely because of the fact that it had been admitted for decades, Judge Gertner wrote in her order. The NAS report suggests a different calculus.
The National Institute of Justice has funded some preliminary studies to establish the scientific information that has so far been missing; UCLAs Mnookin and her colleagues are less than a year into a two-and-a-half-year grant to develop a more formalized and scientifically validated approach to fingerprint analysis. Its not that we know that they dont work, Mnookin says of fingerprints and other forensic methods. Its that we dont have enough evidence about when they work, how they work, when they might not work. The report also led to a series of Senate Judiciary Committee hearings. In January Senator Patrick Leahy (D-Vt. introduced a bill to address some of the major issues in the nations forensic system. The The Criminal Justice and Forensic Science Reform Act takes up many of the issues identified in the NAS report. Although the report has gotten a chilly reception from many forensics experts and prosecutors, many others in the field, like Fisher, believe reforms in the system are long overdue.
Geoffrey Mearns, a former federal prosecutor who helped try both Oklahoma City bombers Terry Nichols and Timothy McVeigh, regularly used forensics in his work. Mearns served on the committee that wrote the academys report. I had assumed that there were well-established uniform processes and procedures in place. I really had faith in the accuracy, reliability, and that it was well grounded in science, says Mearns, now provost and senior vice president for academic affairs at Cleveland State University. When I realized my faith was not well placed, I was very concerned about the damage that it was doing to the accuracy and efficiency of law-enforcement investigations. Because if the science is not accurate, and is leading us to the wrong person, its not only causing a terrible injury to the wrong person, but its leading you away from the right person.
The 265 innocent people so far exonerated by DNA are lucky. Among the hundreds, if not thousands of people that the Innocence Projects Peter Neufeld estimates were wrongfully convicted on the basis of faulty forensics, only a small percentage have DNA available to test. What is their recourse? Neufeld says his organization is counseling attorneys to submit a writ of habeas corpusthe legal systems document of last resorton the basis of newly discovered evidence: the fact that forensic science is not as scientific as it purported to be at the time of trial. However, given the reluctance of judges to ever set aside convictions with anything less than DNA, says Neufeld, I am not as optimistic as I would like to be despite the fact that theres a matter of fairness.
One of those exonerated after 15 years in prison was Roy Brown. He was convicted of murder in 1992 and sentenced to 25 years to life, partly on the basis of a bite-mark analyst who said that Browns teeth matched a wound on the victim to a reasonable degree of dental certainty. The fact that whoever had bitten the victim had six teeth on his upper jawthe wound clearly had six impressionswhereas Roy Brown had only four was inconsistent, the analyst admitted, but explainably so in my opinion.
DNA proved him innocent in 2006.
Beth Schwartzapfel is a Brooklyn freelance journalist with an interest in criminal justice issues.
LINK TO PHOTOS:
http://www.newsweek.com/photo/2011/02/07/faulty-forensics.html
April 2024 March 2024 February 2024 January 2024 December 2023 November 2023 October 2023 September 2023 August 2023 July 2023 June 2023 May 2023 April 2023 March 2023 February 2023 January 2023 December 2022 November 2022 October 2022 September 2022 August 2022 July 2022 June 2022 May 2022 April 2022 March 2022 February 2022 January 2022 December 2021 November 2021 October 2021 September 2021 August 2021 July 2021 June 2021 May 2021 April 2021 March 2021 February 2021 January 2021 December 2020 November 2020 October 2020 September 2020 August 2020 July 2020 June 2020 May 2020 April 2020 March 2020 February 2020 January 2020 December 2019 November 2019 October 2019 September 2019 August 2019 July 2019 June 2019 May 2019 April 2019 March 2019 February 2019 January 2019 December 2018 November 2018 October 2018 September 2018 August 2018 July 2018 June 2018 May 2018 April 2018 March 2018 February 2018 January 2018 December 2017 November 2017 October 2017 September 2017 August 2017 July 2017 June 2017 May 2017 April 2017 March 2017 February 2017 January 2017 December 2016 November 2016 October 2016 September 2016 August 2016 July 2016 June 2016 May 2016 April 2016 March 2016 February 2016 January 2016 December 2015 November 2015 October 2015 September 2015 August 2015 July 2015 June 2015 May 2015 April 2015 March 2015 February 2015 January 2015 December 2014 November 2014 October 2014 September 2014 August 2014 July 2014 June 2014 May 2014 April 2014 March 2014 February 2014 January 2014 December 2013 November 2013 October 2013 September 2013 August 2013 July 2013 June 2013 May 2013 April 2013 March 2013 February 2013 January 2013 December 2012 November 2012 October 2012 September 2012 August 2012 July 2012 June 2012 May 2012 April 2012 March 2012 February 2012 January 2012 December 2011 November 2011 October 2011 September 2011 August 2011 July 2011 June 2011 May 2011 April 2011 March 2011 February 2011 January 2011 December 2010 November 2010 October 2010 September 2010 August 2010 July 2010 June 2010 May 2010 April 2010 March 2010 February 2010 January 2010 December 2009 November 2009 October 2009 September 2009 August 2009 July 2009 June 2009 May 2009 April 2009 March 2009 February 2009 January 2009 December 2008