Liệu có phải đang có một cuộc khủng hoảng khả năng tái tạo trong khoa học?

Is there a reproducibility crisis in science? - Matt Anticole
play-sharp-fill

Is there a reproducibility crisis in science? - Matt Anticole

 
In 2011, a team of physicists reported a startling discovery: neutrinos traveled faster than the speed of light by 60 billionths of a second in their 730 kilometer trip from Geneva to a detector in Italy. Despite six months of double checking, the bizarre discovery refused to yield. But rather than celebrating a physics revolution, the researchers published a cautious paper arguing for continued research in an effort to explain the observed anomaly. In time, the error was tracked to a single incorrectly connected fiber optic cable. This example reminds us that real science is more than static textbooks. Instead, researchers around the world are continuously publishing their latest discoveries with each paper adding to the scientific conversation. Published studies can motivate future research, inspire new products, and inform government policy. So it's important that we have confidence in the published results. If their conclusions are wrong, we risk time, resources, and even our health in the pursuit of false leads. When findings are significant, they are frequently double-checked by other researchers, either by reanalyzing the data or by redoing the entire experiment. For example, it took repeated investigation of the CERN data before the timing error was tracked down. Unfortunately, there are currently neither the resources nor professional incentives to double check the more than 1 million scientific papers published annually. Even when papers are challenged, the results are not reassuring. Recent studies that examined dozens of published pharmaceutical papers managed to replicate the results of less than 25% of them. And similar results have been found in other scientific disciplines. There are a variety of sources for irreproducible results. Errors could hide in their original design, execution, or analysis of the data. Unknown factors, such as patients' undisclosed condition in a medical study, can produce results that are not repeatable in new test subjects. And sometimes, the second research group can't reproduce the original results simply because they don't know exactly what the original group did. However, some problems might stem from systematic decisions in how we do science. Researchers, the institutions that employ them, and the scientific journals that publish findings are expected to produce big results frequently. Important papers can advance careers, generate media interest, and secure essential funding, so there's slim motivation for researchers to challenge their own exciting results. In addition, little incentive exists to publish results unsupportive of the expected hypothesis. That results in a deluge of agreement between what was expected and what was found. In rare occasions, this can even lead to deliberate fabrication, such as in 2013, when a researcher spiked rabbit blood with human blood to give false evidence that his HIV vaccine was working. The publish or perish mindset can also compromise academic journals' traditional peer-review processes which are safety checks where experts examine submitted papers for potential shortcomings. The current system, which might involve only one or two reviewers, can be woefully ineffective. That was demonstrated in a 1998 study where eight weaknesses were deliberately inserted into papers, but only around 25% were caught upon review. Many scientists are working toward improving reproducibility in their fields. There's a push to make researchers raw data, experimental procedures, and analytical techniques more openly available in order to ease replication efforts. The peer review process can also be strengthened to more efficiently weed out weak papers prior to publication. And we could temper the pressure to find big results by publishing more papers that fail to confirm the original hypothesis, an event that happens far more than current scientific literature suggests. Science always has, and always will, encounter some false starts as part of the collective acquisition of new knowledge. Finding ways to improve the reproducibility of our results can help us weed out those false starts more effectively, keeping us moving steadily toward exciting new discoveries.

Matthew Anticole, Brett Underhill, scientific irreproducibility, peer review process, false results, scientific hoaxes, mistakes in science, how to catch a mistake in science, CERN, publish or perish, check your results, who checks experiments, teded science, teded, double check, published results, verify results, how reliable are experiments, TED, TED-Ed, Teded, Ted Education

Hide picture