If you’ve just read a health-related headline that’s caused you to spit out your morning coffee (“Coffee causes cancer” usually does the trick) it’s always best to follow the Blitz slogan: “Keep Calm and Carry On.”
On reading further you’ll often find the headline has left out something important, like “Injecting five rats with really highly concentrated coffee solution caused some changes in cells that might lead to tumors eventually. (Not to mention that the study was funded by The Association of Tea Marketing)The most important rule to remember: “Don’t automatically believe the headline.” It is there to draw you into buying the paper and reading the story. Would you read an article called “Coffee pretty unlikely to cause cancer, but you never know?” Probably not. Before spraying your newspaper with coffee in the future, you need to interrogate the article to see what it says about the research it is reporting on. Bazian (the company I work for) has interrogated hundreds of articles for "Behind The Headlines" on the NHS health news Web site, and we’ve developed the following questions to help you figure out which articles you’re going to believe, and which you’re not.
Does the article support its claims with scientific research?
If an article touts a treatment or a lifestyle factor that is supposed to prevent or cause a disease, but doesn’t give any information about the scientific research behind it, or refers to research that has yet to be published, then treat it with caution. A lot of caution, like balling the article up and throwing it in the (recycling) bin.Is the article based on a conference abstract?
Another area for caution: news articles based on conference abstracts. Research presented at conferences is often at a preliminary stage and usually hasn’t been scrutinised by experts in the field. Also conference abstracts rarely provide full details about methods, making it difficult to judge how well the research was conducted. For these reasons, articles based on conference abstracts should be no cause for alarm. Don’t panic or rush off to your GP. Was the research in humans?
Quite often the “miracle cure” in the headline turns out to have only been tested on cells in the laboratory or on animals. These stories are often accompanied by pictures of humans, creating the illusion that the “miracle cure” came from human studies. Studies in cells and animals are crucial first steps and should not be undervalued. However, many drugs that show promising results in cells in laboratories don’t work in animals, and many drugs that show promising results in animals don’t work in humans. If you read a headline about a drug or food “curing” rats, there is a chance it might cure humans in the future, but unfortunately a larger chance that it won’t. So no need to start eating large amounts of the “wonder food” featured in the article.How many people did the research study include?
In general, the larger a study the more you can trust its results. Small studies may miss important differences because they lack statistical “power”, and small studies are more susceptible to finding things (including things that are wrong) purely by chance. You can visualise this by thinking about tossing a coin. We know that if we toss a coin the chance of getting a head is the same as that of getting a tail – 50/50. However, if we didn’t know this and we tossed a coin four times and got three heads and one tail, we might conclude that getting heads was more likely than tails. But this chance finding would be wrong. If we tossed the coin 500 times – gave the experiment more “power” – we’d be much more likely to get an even number of heads and tails, giving us a better idea of the true odds. When it comes to sample sizes, bigger is usually better. So when you see a study conducted in a handful of people, proceed with caution.Did the study have a control group?
There are many different types of studies, and they are appropriate for answering different types of questions. If the question being asked is about whether a treatment or exposure has an effect or not, then the study needs to have a control group. A control group allows the researchers to compare what happens to people who have the treatment/exposure with what happens to people who don’t. If the study doesn’t have a control group, then it’s difficult to attribute results to the treatment or exposure with any level of certainty.
Also, it’s important that the control group is as similar to the treated/exposed group as possible. The best way to achieve this is to randomly assign some people to be in the treated/exposed group and some people to be in the control group. This is what happens in a randomised controlled trial (RCT) which is why they are considered the “gold standard” way of testing the effects of treatments and exposures. So when reading about a drug, food or treatment that is supposed to have an effect, you want to look for evidence of a control group, and ideally evidence that the study was an RCT. Without either, retain some healthy scepticism.Did the study actually assess what’s in the headline?
This one is a bit tricky to explain without going into a lot of detail about “proxy outcomes”. To avoid doing that, here is the key thought: the research study needs to have examined what is being talked about in the headline and article. (Somewhat alarmingly, this isn’t always the case.) For example, you might read a headline that claims “Tomatoes reduce the risk of heart attacks”. What you need to look for is evidence that the study actually looked at heart attacks. You might instead see that the study found that tomatoes reduce blood pressure. This means that someone has extrapolated that tomatoes must also impact heart attacks, as high blood pressure is a risk factor for heart attacks. Sometimes these extrapolations will prove to be true, but other times they won’t. So if a news story is focusing on a health outcome that was not examined by the research, treat it with a grain of salt.Who paid for and conducted the study?
This is a somewhat cynical point, but one that’s worth making. The majority of trials today are funded by manufacturers of the product being tested – be it a drug, vitamin cream or foodstuff. This means they have a vested interest in the results of the trial which can affect what the researchers find and report in all sorts of conscious and unconscious ways. This is not to say that all manufacturer-sponsored trials are unreliable. Many are very good. But it’s worth looking to see who funded the study to sniff out a potential conflict of interest for yourself.How can I find out more?
It’s not possible to cover all the questions that need to be asked about research studies in a short article, but we’ve covered some of the major ones. For more, go to Behind the Headlines at www.nhs.uk/news for daily breakdowns of health care stories in the media.
See also: THCB review of health care journalist Robert Davis' book, "The Healthy Skeptic."
Dr. Alicia White is Bazian's Health Research Reviewer Manager. She received her doctorate from the University of Manchester and worked as a Postdoctoral Research Fellow in the Department of Pediatrics and Child Health at University College London before joining Bazian.