That’s fake news. Real news COSTS. Please turn off your ad blocker for our web site.
Our PROMISE: Our ads will never cover up content.
Published: Wednesday, December 1, 2021 - 13:03 The origin of SARS-CoV-2 remains a mystery. One theory is that the coronavirus that causes Covid-19 was transmitted from animals to humans—a fairly common occurrence. Another is that it came from a laboratory accident—a more infrequent circumstance. Around the world, scientists conduct many kinds of biological research experiments, from basic studies exploring how living systems operate to synthesizing novel organisms. Applications range from developing medical treatments to protecting the food supply to modifying bacteria to cleaning up oil spills and much more. A subset of experiments may also involve gain-of-function research, which involves modifying an organism to gain a new property or ability. The idea that a pathogen could escape from a laboratory and infect the entire world is the stuff of horror movies. Working with biological materials does have inherent risks, and laboratory incidents will happen. The goal is to minimize risks to laboratory personnel, the community, and the environment. We are biosafety and biosecurity professionals with expertise in mitigating risks associated with biological research. Without a standardized, international framework for reporting laboratory incidents and responses, the task of mitigating such risks is quite difficult. If laboratories were more open about when things go wrong, others could learn from their mistakes and lessen the chances of a future accident. In 1984, 30 tons of a highly toxic gas were released in Bhopal, India. Considered one of the world’s worst industrial accidents, the explosion killed several thousand people. When one of Chernobyl’s radioactive nuclear cores had a meltdown in 1986, the Soviet government hid details and spread misinformation about the event, even though the heat from the reactor could be seen from space. By contrast, when an accident occurs with a biological material, it’s not a spectacular event like an explosion or meltdown. A disease caused by a biological organism takes time to appear. It may take days or weeks for symptoms to present after infection. The 1979 Sverdlovsk Anthrax Outbreak in the Soviet Union and the 2007 Pirbright Institute’s foot-and-mouth incident in the U.K. are examples where biological materials unintentionally escaped the laboratory. People got sick, and farm animals died. Laboratory-related infections are frequently tied to the breakdown of a particular laboratory safety procedure, equipment, or organizational process. Here in the United States, several well-documented laboratory errors have resulted in potential exposures, including the 2014 unintentional release of potentially viable anthrax bacteria, the 2014 potential exposure of a laboratory technician to Ebola virus and the 2015 discovery of improperly inactivated anthrax bacteria that was shipped around the globe. In each case, medical care was provided, and no one became ill. In the United States, a standardized system to report all biological incidents and potential exposures does not exist. The U.S. National Institutes of Health (NIH) has requirements for reporting any significant problems, accidents, and illnesses involving experiments with altered genetic material. If a research institution receives U.S. government funding, failure to comply with NIH rules can result in a loss of this funding, no matter where in the world the lab is located. But private, corporate or DIY biology laboratories operate with even less government oversight and fewer reporting requirements—although many have adopted their own biosafety practices and follow local requirements and best management practices. Outside the United States, the robustness of biosafety and biosecurity oversight varies significantly from country to country. Although reporting to the U.S. government is required in certain circumstances, the information contained in the reports may never reach the public. Some institutions openly publish information about their incidents, while others keep those data private. Reasons may include concerns about reputation, protection of personal health information, or even sensationalism in the media. Some fear reprisal from a disgruntled employee, a competitor, or even a nation-state. Others are concerned about the spread of misinformation by individuals who fear biological labs or those who seek to end human genetic engineering research or ban animal experiments. Even with these concerns, we believe a more transparent and comprehensive system of reporting biological incidents to a neutral third party would help reduce the number of laboratory incidents—and could improve public trust in the scientific enterprise. If this type of system had been in place prior to Covid-19, more data presumably would have been available to help evaluate the Wuhan laboratory leak hypothesis and cut down on speculation. In the United States, a possible way to do this is to expand the American Biological Safety Association’s Laboratory Acquired Infection database. Currently, it contains incident data only from published research papers. But it could be broadened to include all kinds of incident data. In our opinion, an international version of such a reporting system would also help reduce the number and severity of laboratory incidents, both locally and worldwide. The more information available about the root cause of incidents, the more it could be used to help improve training, procedures, and controls—and prevent future problems. It would also suggest how safety systems break down and what systems might be at risk. Biosafety and biosecurity professionals have been discussing this topic for a long time, including at a U.S. Trans-Federal Task Force on Optimizing Biosafety and Biocontainment and by a Federal Experts Security Advisory Panel. But to make a centralized reporting system a reality, key players will need to commit and act. They include governments, international agencies, industry partners, and the scientific community. This article is republished from The Conversation under a Creative Commons license. Read the original article. Quality Digest does not charge readers for its content. We believe that industry news is important for you to do your job, and Quality Digest supports businesses of all types. However, someone has to pay for this content. And that’s where advertising comes in. Most people consider ads a nuisance, but they do serve a useful function besides allowing media companies to stay afloat. They keep you aware of new products and services relevant to your industry. All ads in Quality Digest apply directly to products and services that most of our readers need. You won’t see automobile or health supplement ads. So please consider turning off your ad blocker for our site. Thanks, David Gillum is the assistant vice president of Environmental Health and Safety and the chief safety officer at Arizona State University. Kathleen Vogel is a professor and the interim director at the School for the Future of Innovation in Society at Arizona State University. Rebecca Moritz is the biosafety director and responsible official at Colorado State University.Reporting All Biosafety Errors Could Improve Labs Worldwide
There’s no central reporting system in the U.S. or internationally
Science and technology mishaps
Biological incident reporting
Sharing information
Our PROMISE: Quality Digest only displays static ads that never overlay or cover up content. They never get in your way. They are there for you to read, or not.
Quality Digest Discuss
About The Authors
David Gillum
Kathleen Vogel
Rebecca Moritz
© 2022 Quality Digest. Copyright on content held by Quality Digest or by individual authors. Contact Quality Digest for reprint information.
“Quality Digest" is a trademark owned by Quality Circle Institute, Inc.