Science Policy For All

Because science policy affects everyone.

Science Policy Around the Web – June 27, 2017

leave a comment »

By: Sarah Hawes, PhD

Source: pixabay

Influenza

An Arms Race with Nature

H7N9, a new bird flu emerging in China, has infected roughly 1,500 people and killed 40% of them. The virus is contracted directly from infected birds but is not yet easily transmissible between humans, however researchers at The Scripps Research Institute have evidence H7N9 could potentially become transmissible between humans fairly easily. They examined a fragment of the virus that interacts with receptors on animal cells to gain entrance, and identified three minor mutations that could cause the fragment to shift from preferentially entering avian cells to preferentially entering human cells. If these mutations were to occur, it could rapidly result in a pandemic.

Tests in a viral fragment do not prove functionality in the intact virus; that would require mutating H7N9 itself. A 2014 moratorium on mutating three types of viruses (SARS, MERS, influenza) to more dangerous forms is expected to lift when the Department of Health and Human Services finishes current work drafting a new policy establishing reviews designed to assess benefit/risk ratios before funding research.

The subject is divisive, even among scientists in the field. Stanford researcher David Relman says he would support efforts to test mutations in a weakened strain of flu, but not in the H7N9 virus.  Bioterrorism expert Thomas Inglesby opposes increasing the contagious lethality of a virus, and opposes publishing such procedures due to concern that less benevolent actors would be enabled to replicate the process. NIH funded researcher, Ron Fouchier in the Netherlands, whose alteration of H5N1 to become highly contagious between ferrets (the animal model for humans) in 2011 influenced the moratorium, believes examining dangerous virus mutations in a controlled lab environment is important to identify potential pandemic viruses.

Many of these topics were discussed at the recent Immunology and Evolution of Influenza Symposium, and are sure to be a hot topic at the July 16 – 19 Centers of Excellence for Influenza Research and Surveillance meeting. With policy guidance needed on benefit/risk, potentially safer models, security, and publication limitations, the new HHS policy will be critical. (Nell Greenfieldboyce, NPR)

Conservation

Modeling with Dough – Pick your Species

The Supreme Court found the Endangered Species Act was “intended to halt and reverse the trend toward species extinction—whatever the cost.” Today, in light of the cost, conservation policy makers are being invited to triage species extinctions. Fish and Wildlife Service representatives recently met with ecologist Dr. Leah Gerber to discuss her proposed use of an algorithm guiding conservation funding.

A self-proclaimed environmentalist, Gerber says her model suggests that defunding “costly failures,” including the spotted owl, golden-cheeked warbler and gopher tortoise, could help save about 180 other species. Gerber says policy makers may opt to continue to support species that her algorithm rejects, as was done for the koala in Australia where algorithm triage has been used. In this case, a popularity contest may determine who lives and who goes extinct.

Details of the algorithm are not explicit, but Dr. Gerber’s recent publication in PNAS is a straightforward return-on-investment calculation analyzing the mathematical relationship between funds requested, spent, and species success or decline.  Gerber finds “the cost–success curve is convex; funding surpluses were common for the species least likely and most likely to recover” so it’s not simply ‘money in – species out’. Other factors – endemism, keystone status, level of species risk – are also important, though Gerber acknowledges they are not currently included.

While proponents call use of the equation “doing the best you can with what you have,” lack of data on its predictive validity make it a frightening policy tool governing something as permanent as species extinction. What if region affects costs, population growth is slower in species reaching sexual maturity later, a break-through in understanding one species’ requirements is just around the corner or we haven’t yet discovered the significance of the niche occupied by another species? What if business or political interests conflict with a species’ needs? What if the algorithm developer seeks intellectual property legal status, as is happening now with a proprietary algorithm used in parole and sentencing situations? Algorithms impacting public policy should be vetted by multiple experts in germane disciplines, validated, and kept publicly accessible for healthy scrutiny. (Sharon Bernstein, Reuters)

 

Have an interesting science policy link?  Share it in the comments!

Written by sciencepolicyforall

June 27, 2017 at 11:42 am

Science Policy Around the Web – June 23, 2017

leave a comment »

By: Saurav Seshadri, PhD

Drug Policy

Trump’s New Policy to Tackle Sky-High Drug Prices Makes Sense — Sort Of

Tackling high prescription drug prices was a repeated promise of the Trump campaign. The Trump administration has now taken its first step towards fulfilling this pledge, outlined in a blog post by Food and Drug Administration (FDA) commissioner Scott Gottlieb. The agency will pursue a Drug Competition Action Plan, whose goal will be to eliminate obstacles to the development of cheap generic drugs – particularly those caused by loopholes in existing FDA policies, which are exploited by pharmaceutical companies to extend their patent exclusivity period and maximize profits. An example of such ‘gaming’ the system, cited in the post, is the practice of limiting access to branded products for comparative testing by generic developers. Ultimately, the FDA will work closely with the Federal Trade Commission (FTC) to address such issues, since directly regulating business practices is outside its mandate.

On its face, the FDA’s effort is a step in the right direction. Availability of generics reduces the cost of medications by over half within the first year, and according to a recent Congressional report, manufacturers state that ‘competition…is the primary driver of generic drug prices’. However, it ignores evidence that the real driver of increased drug spending is new, branded medicines, not overpriced generics. In fact, early indications are that Trump’s policies will favor the pharmaceutical companies that produce such medicines, by reducing regulations and apparently abandoning his promise to enable the government to negotiate drug pricing through Medicare. Overall, these actions signal a commitment to promoting free market mechanisms in the pharmaceutical industry; time will tell whether this approach will actually lead to more affordable drugs. (Julia Belluz, Vox)

Cancer

In a Major Shift, Cancer Drugs go ‘Tissue-Agnostic’

With the landmark approval of Keytruda in May, the Food and Drug Administration (FDA) appears to have ushered in a new era of cancer drug development.  So far, cancer treatment and drug evaluation have largely used the tumor’s tissue of origin as a starting point. Keytruda (an immune system enabling drug developed by Merck and approved for melanoma in 2014) marked the first departure from this approach, receiving priority approval to treat any solid tumor containing a mutation in the mismatch repair pathway, regardless of context. Recently released data suggests that another tissue-agnostic cancer therapy is on the way: larotrectinib (a cell growth inhibitor developed by Loxo Oncology) showed high efficacy for any tumor with a certain biomarker (TRK fusion). Several other such drugs, whose indications will be based on tumor genetics rather than location, are in the clinical pipeline.

Although these advances have generated significant excitement in the cancer community, some caveats exist. First, identifying the patients that could benefit from tissue-agnostic treatments will require individual initiative and depend on the cost of screening, particularly when considering markers that are rare for a certain tumor type. A potential solution is suggested by the NCI-MATCH trial, part of the NIH’s Precision Medicine Initiative (PMI) – in it, patients can enroll in one of several parallel clinical trials if a corresponding drug-targeted mutation is found in their tumor’s genome. If these trials prove effective, patients could eventually be regularly matched with a personalized, tissue-agnostic, biologically valid treatment, based on a standardized screen.  Second, researchers caution that tissue-agnostic studies should have a strong scientific rationale and/or breakthrough-level efficacy. Otherwise, such efforts ‘could actually slow drug development if there are differential effects across tumor types by diverting resources from enrolling patients in a predominant population or in the tumor type most likely to respond’.

Despite these concerns, the tissue-agnostic paradigm offers great promise for cancer patients. NIH-funded resources such as The Cancer Genome Atlas could be invaluable to this field moving forward. (Ken Garber, Science)

Scientific Publishing

US Court Grants Elsevier Millions in Damages from Sci-Hub

A New York district court has awarded academic publishing giant Elsevier $15 million in damages from Alexandra Elbakyan, founder of the website Sci-Hub, for copyright infringement. Elbakyan, a 27-year-old neuroscientist turned programmer, started Sci-Hub in 2011 with the goal of ‘remov[ing] all barriers in the way of science’. The site allows users to download research papers that would normally be blocked by a paywall, by obtaining credentials from subscribing institutions and using them to access publisher-run databases like ScienceDirect. Over 60 million papers are posted on Sci-Hub, and users downloaded 28 million articles in 2016.

Elbakyan’s case is reminiscent of Aaron Swartz, another high-profile champion of open access to scientific research. Faced with federal charges related to his hacking of journal archive JSTOR, Swartz tragically committed suicide in 2013. Both Elbakyan and Swartz found publishers’ ability to profit from restricting access to scientific literature, effectively withholding knowledge from anyone outside of a privileged inner circle, as well as the legal protection provided to this system, to be deeply unethical. Their willingness to act upon these convictions has earned each a sizable following in the scientific community.

For their part, publishers claim that fees go towards overhead, and point to significant efforts to expand free and open access programs. While judges have so far been sympathetic, Elsevier’s legal battle has been largely one-sided. Elbakyan has been ignoring rulings requiring her to shut down Sci-Hub since 2015, opting to simply change domains instead, and since she is currently based in Russia and has no American assets, she is unlikely to pay any damages. (Quirin Schiermeier, Nature News)

Have an interesting science policy link?  Share it in the comments!

Written by sciencepolicyforall

June 23, 2017 at 11:00 am

Science Policy Around the Web – June 20, 2017

leave a comment »

By: Eric Cheng, PhD

Source: Flickr, via Creative Commons (CC BY 2.0)

Research Funding

America is Still First in Science, but China Rose Fast as Funding Stalled in U. S. and Other Countries

American scientific groups continue to publish more biomedical research discoveries than groups from any other country, and the United States still leads the world in research and development expenditures. However, American dominance is slowly diminishing as China’s increase in funding on science over the last twenty years are starting to pay off. Chinese biomedical research now ranks fourth in the world for total number of discoveries published in six top-tier journals. This is with China spending three-fourths of the amount of money that the U.S. spent on research and development in 2015. In addition, new discoveries and advances in science are becoming more of a collaborative effort, which include researchers from around the world.

These findings come from research published in The Journal of Clinical Investigation by a group of University of Michigan researchers. The analysis comes at an important time for Congress to think about whether the annual uncertainty of the National Institutes of Health’s(NIH) budget and proposed cuts are in the nation’s best interest over the long-term. Bishr Omary, the senior author of the article commented, “If we continue on the path we’re on, it will be harder to maintain our lead and, even more importantly, we could be disenchanting the next generation of bright and passionate biomedical scientists who see a limited future in pursuing a scientist or physician-investigator career.”

The research was based on data up to 2015. During the current fiscal year of 2017, funding for NIH was proposed to be increased by 2 billion dollars, which is the second year in a row where funding was increased after 12 years of flat budgets. With this increase in funding, Omary hopes that, “our current and future investment in NIH and other federal research support agencies will rise above any branch of government to help our next generation reach their potential and dreams.” (University of Michigan, ScienceDaily)

Opioid Crisis

The Role of Science in Addressing the Opioid Crisis

Opioid addiction is an ongoing public health crisis. Millions of individuals all over the United States suffer from opioid use disorder with millions more suffering from chronic pain. Due to the urgency and scale of this crisis, innovative scientific solutions need to be developed. As part of a government-wide effort to address this crisis, the National Institutes of Health (NIH) is supplementing current research efforts with a public-private collaborative research initiative on pain and opioid abuse.

The Director of NIH, Dr. Francis Collins met with research and development leaders from biopharmaceutical companies in April 2017 to discuss new ways in which  government and industry can work together to address the opioid crisis. Dr. Collins stated how some advances such as improved formulations, opioids with abuse-deterrent properties, longer-acting overdose-reversal drugs, and repurposing of treatments approved for other conditions may be quick. Other advances such as mu-opioid receptor-based agonists, opioid vaccines, and novel overdose-reversal medications may be slower to develop. Overall, the goal for this partnership is to reduce the time typically required to develop new, safe, and effective therapeutics to half the average time. (Nora D. Volkow and Francis S. Collins, New England Journal of Medicine)

Climate Change

France is Offering US Scientists 4-year Grants to Move to the Country and do Research

Following President Donald Trump’s decision to withdraw the United States from the Paris climate agreement, France created an initiative that will allow researchers, teachers, and students to apply for a fully financed four-year grant to combat climate change. The website for the initiativesays,

“You will be able to stay in France at least for the duration of the grant, and longer if you are granted a permanent position. There is no restriction on your husband / wife working in France. If you have children, note that French public schools are free, and the tuition fees of universities and ‘grandes écoles’ are very low compared to the American system.”

Since Emmanuel Macron won the French presidential election in May, he has addressed American scientists who feel alienated by the Trump administration. Macron has promised strong funding for climate initiatives. However, some U.S. scientists like David Blockstein of the National Council for Science and the Environment see Macron’s invitation as “both a publicity stunt and a real opportunity.” Although it is not very likely that many U.S. researchers will take up the offer, it does provide a “sharp contrast to an increasingly hostile U.S. political environment for science.” (Chris Weller, Business Insider)

Have an interesting science policy link?  Share it in the comments!

Written by sciencepolicyforall

June 20, 2017 at 1:10 pm

Science Policy Around the Web – June 16, 2017

leave a comment »

By: Emily Petrus, PhD

Source: pixabay

Science and Politics

Politics in Science – It’s Not Just the U.S.!

Romania is a country in eastern Europe that joined the European Union (EU) in 2007. Scientists there are few and far between; research spending only accounts for 0.49% of GDP, the lowest in Europe (the US spent 2.7% in 2016). After joining the EU, Romanian researchers were encouraged to apply for European merit-based grants and sit on international review boards such as the National Research Council and the National Council of Ethics. It seemed that research was making slow but steady progress, but the new administration elected this year has shaken things up in all facets of government, including scientific research.

The new research minister, Serban Valeca, removed the international members appointed to government councils that oversee research funding, ethics, innovation and science policy, and replaced them with city council members, government-loyal union members, researchers from second tier Romanian institutes and even a surgeon being investigated for embezzlement. Grant review panels have been shuffled to remove international scientists and replace them with domestic researchers, but only if they have a certificate saying their university approves of their participation. These changes mark a departure from welcoming international input into Romanian proceedings and a movement towards scientific isolation.

To combat these changes, Romanian scientists have formed an organization, Ad Astra, which calls on researchers to boycott grant evaluations. Combined with the shuffling, the councils have been suspended for 3 months, which delays funding and puts already under-funded researchers in peril. The European University Association calls the policies deeply concerning, and although the current president may disagree with the research minister’s handling of the situation, his political ties ensure he won’t hold much sway over how this plays out. A computer scientist at the University Politehnica in Bucharest, Costin Raiciu, is concerned that the policies will affect talented researchers who have returned to Romania and says, “Without [merit-based] funding, people would either give up research altogether or move out of the country”. This is an all too familiar scenario in which it is apparent that policy and science must cooperate to produce ideal outcomes. (Alexandra Nistoroiu, ScienceInsider)

Mental Health

Clinical Trials Down, Basic Research Up at NIMH

Mental health is a notoriously tricky field. The development of the Diagnostic and Statistical Manual of Mental Disorders (DSM) in the 1950s has historically been a way to diagnose patients with mental health issues, and then give appropriate treatment. This has proved to be an imprecise treatment strategy, because within a category of diagnosis there is a broad spectrum of behaviors, and underlying this behavior there may be multiple causes. The NIH’s Precision Medicine Initiative (PMI) seeks to characterize 1,000,000 people by behavior, genetics, environment, and physiology. Researchers from the NIMH will send questionnaires evaluating behavior to detect mood and reward responses for this group of people. When this mental health evaluation is combined with information about their genetics, lifestyle and environment, scientists can characterize mental health disorders more specifically.

Many clinician researchers are upset by the steep decline in clinical trial research funded by NIMH, which has become higher profile with director Joshua Gordon’s arrival in 2016. NIMH seeks to route funding to study mental disorders using a basic research approach before spending time and money on costly clinical trials which too often lead to inconclusive or disappointing results. In 2011 NIMH launched the Research Domain Criteria (RDoC), which encourages research proposals to include a hunt for the mechanism underlying mental health studies. Since the initial call to include a RDoC perspective in grant applications, the incidence of RDoC appearing in funded applications has increased while mention of the DSM has decreased. Other buzzwords that are present in funded grants include biomarker, circuit, target and mechanism.

These data represent a shift in how funding decisions will proceed in mental health but may have broader reaching implications for other areas of research. In a blog post Dr. Gordon writes, “the idea that RDoC will facilitate rapid, robust and reproducible neurobiological explanations for psychopathology (as observed within and across DSM disorders) represents a hypothesis”. It remains to be seen if RDoC is an effective metric to evaluate successful grants. (Sara Reardon, Nature News)

Have an interesting science policy link?  Share it in the comments!

Science Policy Around the Web – June 13, 2017

leave a comment »

By: Nivedita Sengupta, PhD

By Mikael Häggström, used with permission. [Public domain], via Wikimedia Commons

Stem Cell Therapy

Texas on Track to Become First State to Explicitly Back Stem Cell Therapies

On 30th May, Texas passed a bill  authorizing unapproved stem cell therapies, making Texas the first state to openly recognize experimental treatments. The bill will make the use of unapproved stem cell therapies legal for patients and is currently awaiting the approval of Governor Greg Abbott, who already supports the measure. Experimental stem cell therapies for terminal and chronic conditions have struggled for years to gain support without much success. Until now, no state has provided legal validation for these kind of therapies and the current stem cell procedures are mostly done under strict regulations.

Amendments were added to the bill, which require that the treatments be delivered by doctors with the approval of an institutional review board, which deals with human research. It will also add another amendment that will allow patients to have authority to sue in case the treatments go wrong. Many scientists and advocates opposed the measure stating that unapproved stem cell therapies can be harmful rather than beneficial. They state that though the amendments add protection to the patients, there are a few aspects of the bill that make them uncomfortable. Two other bills focused on patient access to experimental therapies, also known as “right-to-try” policies, failed to pass in the Texas Senate. (Andrew Joseph, STATNews)

Research Funding

NIH Scraps Plans for Cap on Research Grants

US National Institutes of Health (NIH) decided to drop the controversial proposal of capping the number of grants that an investigator can have at a time. The initial capping attempt was suggested to gather funds for younger researchers by NIH in May. The proposal was based on studies that suggested that a lab’s productivity decreases once it holds too many grants. Younger scientists often face more difficulties in obtaining NIH RO1 grants compared to their older more experienced colleagues. As a result, many researchers applauded the NIH’s effort to provide more funding for younger scientists. Yet the capping proposal received major adverse response from the scientific community stating that the NIH’s interpretation of the productivity study data does not apply to all labs, especially to the collaborative lab groups with four or five R01s that are more productive than labs with only one. Researchers also complained that the proposed point-based scoring system will also make collaborations difficult thus hampering productivity in the long run.

NIH director Dr. Francis Collins stated that the original idea was still a work in progress and NIH is going to put a hold on it. Instead of the cap, on 8th June, NIH announced the creation of the special fund, the Next Generation Researchers Initiative (NGRI), starting with US$210 for funding young researchers. The initiative will focus on investigators with less than 10 years of experience as NIH- funded principal investigators, and on high score grant proposals that were rejected because of lack of money. The initiative will grow up to $1.1 billion over the next five years. According to NIH principal deputy director Larry Tabak, NIH will immediately start creating an inventory of investigators who meet these criteria and expects that this approach will allow more than 2,000 additional R01 grants to be funded to younger scientists compared to the cap-based plan, which would have supported only 1600 awards. Nonetheless, the current proposal is still going to generate controversy as it will affect the older researchers because of NIH’s diversion of funding. (Sara Reardon, Nature News)

Have an interesting science policy link?  Share it in the comments!

Written by sciencepolicyforall

June 13, 2017 at 7:08 pm

Science Policy Around the Web – June 10, 2017

leave a comment »

By: Allison Dennis, BS

Source: pixabay

Animal Testing

Lack of Clarity Puts Chemical Safety and Animal Welfare at Odds

In the lineup of American stereotypes, the health-nut who cares about the chemicals in his shampoo is often the same person who cares if that shampoo was tested on animals or not. However, a bill signed June 22, 2016, known as the Frank R. Lautenberg Chemical Safety for the 21st Century Act, may be placing those two views at odds. The bill requires the U.S. Environmental Protection Agency (EPA) to implement a risk-based process to evaluate the safety of chemical substances currently being used in the marketplace and approve the use of new chemicals before their introduction. The bill was passed with bipartisan support and offered EPA the new-found power to fully regulate the use of well-known carcinogens like asbestos.

Yet the pathway forward for the EPA is daunting. More than 62,000 substances find their way into and onto our bodies through the products we use and our environment. While many of these substances have become associated with disease over time, how can the EPA certify the risks associated with different exposures to varying amounts of each substance on such an extensive list? The Act itself suggested that once the EPA has evaluated the existing information on the 62,000 substances currently in use, it spend the next twelve months triaging chemicals according to their potential risk. Next, the highest priority chemicals will be evaluated on a three-year deadline to develop knowledge of their toxicity and guidelines for their regulation. Ultimately, by clearly cataloging the risk of common chemicals the Frank R. Lautenberg Chemical Safety for the 21st Century Act promises to greatly reduce the amount of animal testing needed in the long-term.

In the meantime, however, the companies that use to-be-regulated substances in their products may be inclined to undertake independent toxicity testing, collecting enough data to guarantee that their favorite substances meet the low-risk criteria and avoid a drawn-out evaluation. Defining toxicity requires careful experimentation, which can sometimes be carried out in human cells outside of the body, but often require evaluation in animals. Animal rights groups like the Human Society find concern with the lack of transparency in the pre-prioritization process. They fear the eagerness of companies to provide data without any clear guidelines about how that data will be evaluated or what substances will require extensive evaluation could result in extensive and unnecessary animal testing. Further they suggested that the EPA require any new pre-approval data obtained by companies to be collected using non-animal methods. (Maggie Koerth-Baker, FiveThirtyEight)

CRISPR

Small Study may Reveal Big Concerns over CRISPR-Based Therapy

A one-page letter published in Nature Methods last week reports unexpectedly high levels of unintended changes to the genomes of mice that underwent a CRISPR-based therapy. Since it’s renaissance as a therapeutic tool in 2012, CRISPR has occupied the imaginations of scientists, doctors, patients, investors, and ethicists. CRISPR technology provides a relatively straight-forward and reproducible means to gene editing on the cellular level, but its applications to create heritable mutations in the human germ line is on hold until more is understood about the long-term effects such treatments would have.

The original study sought to explore potential long-term effects of germline manipulation by CRISPR in a mouse model. Guide RNA along with the Cas9 enzyme were injected into mouse zygotes, which introduced a correction in a mutation in the rd1 gene of otherwise blind mice. Initiating this change before the first cell division enabled this corrected mutation to be inherited by all cells arising in the developing mouse, consequently restoring the ability for the eyes to develop normally. In a follow-up experiment described in their one page letter, the researchers looked for mutations in the genomic DNA of two CRISPR-treated adult mice compared with a control mouse, revealing over 2,000 unintended mutations following CRISPR treatments. None of these mutations appeared to affect the mice, suggesting that deep genomic sequencing may be required to reveal unanticipated changes in an outwardly healthy mouse. Further, the nature of these unintended mutations offered few clues explaining how they might have occurred.

This result stands in contrast with other reports quantifying the extent of these unintended changes, which found CRISPR to be highly specific. While the CRISPR-Cas9 system has been observed to sometimes alter off-target regions of the genome, this activity can usually be curbed through the careful design and evaluation of the guide RNA. The limitations of this small study have been discussed extensively since its publication. However, the findings have sparked the need for further investigation into the long-term-whole-animal effects of germline-editing by CRISPR. As human germline-editing creeps closer to reality, the FDA will be tasked with developing an entirely new means of evaluating the safety of such technologies (Megan Molteni, Wired)

Have an interesting science policy link?  Share it in the comments!

Written by sciencepolicyforall

June 10, 2017 at 11:33 am

Science Policy Around the Web – June 06, 2017

leave a comment »

By: Kseniya Golovnina, PhD

Source: Flickr, by USDA, via Creative Commons     (CC BY 2.0)

Food Security

What if Food Crops Failed at the Same Time?

When one group of people is fighting with climate change and another considers it “mythical”, researchers specialized in the study of social-ecological systems are developing food supply risk assessment models. Food crops are one of the most important sources of human being existence, and less than one-fourth of the planet (“breadbaskets”) produces three-fourth of the staple crops that feed the world’s population. In fact, climate change could cause crop losses in most of the breadbaskets.

Two important factors included in the models are shocks to major land crop production and economy. Shocks like droughts and heat waves in Ukraine and Russia in 2007 and 2009 almost wiped out wheat crops, and caused global wheat prices to spike. And demand assessments project that food production may have to double by 2050 to feed a growing population. Together, the potential environmental and economic stresses are making the world food production system less resilient, and will affect both rich and poor nations. To measure the fragility of the system, researchers developed scenarios of small shocks (10 percent crop loss) and large shocks (50 percent crop loss). These were then applied to corn, wheat or rice output using an integrated assessment model, the Global Change Assessment Model, which was developed by the U.S. Department of Energy.

Among the critical findings are that “breadbasket” regions respond to shocks in different ways. For example, South Asia, where most of the arable land is already in use, is quite unresponsive to shocks occurring elsewhere in the world, because the total amount of land in agricultural production cannot be changed significantly. In Brazil the situation is opposite, it has a lot of potential to bring new land into production if large shocks occur. However, cleaning Brazil’s forests requires significant effort and would add significantly to global climate change. Within the research agenda of the Pardee Center, these risks and preventive actions are discussed in more detail. The warning is clear: humankind needs to be aware and prepared for potential multiple “breadbaskets” failure if we want to reduce the potential for catastrophe. (Anthony Janetos, The Conversation)

Reproducibility in Science

Research Transparency: Open Science

Increasing amounts of scientific data, complexity of experiments, and the hidden or proprietary nature of data has given rise to the “reproducibility crisis” in science. Reproducibility studies in cancer biology have revealed that only 40 % or less peer-reviewed analyses are replicable. Another large-scale project attempting to replicate 100 recent psychology studies was successful in replicating less than 50% of the original results.

These findings are driving scientists to look for ways to increase study reliability, and make research practices more efficient and available for evaluation. A philosophy of open science, where scientists share their primary materials and data, makes analytical approaches more transparent and allows common research practices and standards to emerge more quickly. For scientific journals and associations, open science methods enable the creation of different ways to store and utilize data. Some journals are specifically dedicated to publishing data sets for reuse (Scientific DataJournal of Open Psychology Data), others require or reward open science practices like publicly posting materials and data.

The widespread use of online repositories to share study materials and data helps to store large data sets and physical materials to help mitigate the problems of reproducibility. However, open science practice is still very much in development, and faces some significant disincentives. Habits and reward structures are two major forces work against. Researchers are used to being close, and hide their data from being stolen. Journal editors tend to favor publishing papers that tell a tidy story with perfectly clear results. This causes researchers to omit “failed” studies that don’t clearly support their theories.

While efforts to overcome these obstacles are difficult, development of fully transparent science should be encouraged, as openness helps improve understanding, and acknowledges the truth that real data are often messy. (Elizabeth Gilbert and Katie Corker, The Conversation)

 

Have an interesting science policy link?  Share it in the comments!

Written by sciencepolicyforall

June 6, 2017 at 9:00 am