Thursday, 20 January 2011

Retract That Seroxat?

Should a dodgy paper on antidepressants be retracted? And what's scientific retraction for, anyway?


Read all about it in a new article in the BMJ: Rules of Retraction. It's about the efforts of two academics, Jon Jureidini and Leemon McHenry. Their mission - so far unsuccesful - is to get this 2001 paper retracted: Efficacy of paroxetine in the treatment of adolescent major depression.

Jureidini is a member of Healthy Skepticism, a fantastic Australian organization that Neuroskeptic readers have encountered before. They've got lots of detail on the ill-fated "Study 329", including internal drug company documents, here.

So what's the story? Study 329 was a placebo-controlled trial of the SSRI paroxetine (Paxil, Seroxat) in 275 depressed adolescents. The paper concluded: that "Paroxetine is generally well tolerated and effective for major depression in adolescents." It was published in the Journal of the American Academy of Child and Adolescent Psychiatry (JAACAP).

There's two issues here: whether paroxetine worked, and whether it was safe. On safety, the paper concluded that "Paroxetine was generally well tolerated...and most adverse effects were not serious." Technically true, but only because there were so many mild side effects.

In fact, 11 patients on paroxetine reported serious adverse events, including suicidal ideation or behaviour, and 7 were hospitalized. Just 2 patients in the placebo group had such events. Yet we are reassured that "Of the 11, only headache (1 patient) was considered by the treating investigator to be related to paroxetine treatment."

The drug company argue that it didn't become clear that paroxetine caused suicidal ideation in adolescents until after the paper was published. In 2002, British authorities reviewed the evidence and said that paroxetine should not be given in this age group.

That's as maybe; the fact remains that in this paper there was a strongly raised risk. However, in fairness, all that data was there in the paper, for readers to draw their own conclusions from. The paper downplays it, but the numbers are there.

*

The efficacy question is where the allegations of dodgy practices are most convincing. The paper concludes that paroxetine worked, while imipramine, an older antidepressant, didn't.

Jureidini and McHenry say that paroxetine only worked on a few of the outcomes - ways of measuring depression and how much the patients improved. On most of the outcomes, it didn't work, but the paper focusses on the ones where it did. According to the BMJ

Study 329’s results showed that paroxetine was no more effective than the placebo according to measurements of eight outcomes specified by Martin Keller, professor of psychiatry at Brown University, when he first drew up the trial.

Two of these were primary outcomes...the drug also showed no significant effect for the initial six secondary outcome measures. [it] only produced a positive result when four new secondary outcome measures, which were introduced following the initial data analysis, were used... Fifteen other new secondary outcome measures failed to throw up positive results.

Here's the worst example. In the original protocol, two "primary" endpoints were specified: the change in the total Hamilton Scale (HAMD) score, and % of patients who 'responded', defined as either an improvement of more than 50% of their starting HAMD score or a final HAMD of 8 or below.

On neither of these measures did paroxetine work better than placebo at the p=0.05 significance level. It did work if you defined 'responded' to mean only a final HAMD of 8 or below, but this was not how it was defined in the protocol. In fact, the Methods section of the paper follows the protocol faithfully. Yet in the Results section, the authors still say that:
Of the depression-related variables, paroxetine separated statistically from placebo at endpoint among four of the parameters: response (i.e., primary outcome measure)...
It may seem like a subtle point. But it's absolutely crucial. Paroxetine just did not work on either pre-defined primary outcome measure, and the paper says that it did.

Finally, there were also issues of ghostwriting. I've never been that concerned by this in itself. If the science is bad, it's bad whoever wrote it. Still, it's hardly a good thing.

*

Does any of this matter? In one sense, no. Authorities have told doctors not to use paroxetine in adolescents with depression since 2002 (in the UK) and 2003 (in the USA). So retracting this paper wouldn't change much in the real world of treatment.

But in another sense, the stakes are enormous. If this paper were retracted, it would set a precedent and send a message: this kind of p-value fishing to get positive results, is grounds for retraction.

This would be huge, because this kind of fishing is sadly very common. Retracting this paper would be saying: selective outcome reporting is a form of misconduct. So this debate is really not about Seroxat, but about science.


There are no Senates or Supreme Courts in science. However, journal editors are in a unique position to help change this. They're just about the only people (grant awarders being the others) who have the power to actually impose sanctions on scientists. They have no official power. But they have clout.

Were the JAACAP to retract this paper, which they've so far said they have no plans to do, it would go some way to making these practices unacceptable. And I think no-one can seriously disagree that they should be unacceptable, and that science and medicine would be much better off if they were. Do we want more papers like this, or do we want fewer?

So I think the question of whether to retract or not boils down to whether it's OK to punish some people "to make an example of them", even though we know of plenty of others who have done the same, or worse, and won't be punished.

My feeling is: no, it's not very fair, but we're talking about multi-billion pound companies and a list of authors whose high-flying careers are not going to crash and burn just because one paper from 10 years ago gets pulled. If this were some poor 24 year old's PhD thesis, it would be different, but these are grown-ups who can handle themselves.

So I say: retract.

ResearchBlogging.orgNewman, M. (2010). The rules of retraction BMJ, 341 (dec07 4) DOI: 10.1136/bmj.c6985

Keller MB, et al. (2001). Efficacy of paroxetine in the treatment of adolescent major depression: a randomized, controlled trial. Journal of the American Academy of Child and Adolescent Psychiatry, 40 (7), 762-72 PMID: 11437014

23 comments:

petrossa said...

If only. But then again they should retract practically all papers submitted by industry sponsored trials.

It's hard to find an SSSRI (my version: Somewhat Selective Serotonin Reuptake Inhibitors) that actually works markedly better then a placebo for the majority of depressions and it's even harder to find one that doesn't have a multitude of quality of life influencing negative side-effects.

Since that's not what most papers say they are retractable too.

But, as with smoking, there is a duality. On the one hand huge profits, on the other people suffering the consequences.

I predict a major issue on the same line within the next decade on statins.

brightpsych said...

Just wanted to say thank you very much for your great posts!

veri said...

In terms of sanctions the UN has a science body.. not sure what they do. Maybe if they collected a million signatures or something they might consider a forced retraction. Can Amnesty tackle this sort of thing? Science NGO? They should beta-test about to be published articles. Send them to specific professionals / PhD students in that field. Top journals from my understanding consult think tanks that make easy to read summaries of the relevant articles and send them as online memos or something to cabinet members, people in politics. The same service could extend to professionals in that field with a bit more detail, author/s name/s omitted. That way the article goes through a truly democratic process. Professionals need to be up to date with the latest findings anyway right? This might give an opportunity for people who want to speak up to speak up before publication. Democracy driven peer review. That would be interesting. Of course it might not fare so well for journal subscriptions particularly among professionals since they see the content beforehand but so what. I believe science should be a public discourse funded by tax payers. It shouldn't remain a private monopoly when it's clearly changing the way we live.

Anonymous said...

More praise for Neuroskeptic. Without question one of the best blogs I've ever read. You have a talent for popularizing science. Thank you.

Anonymous said...

But the problem is that a lot of these journal editors and reviewers are on the payrolls of the drug companies.And as far as psychiatry goes, all of what has been published about SSRI's over the past 20 years is crap-- bought and paid for by Big Pharma.

Mike said...

When you have 8 primary and 15 secondary outcome measures, about one of those is bound to be significant at p = .05 just by chance. Then you can revise your "hypothesis" and say that was the measure where you expected to see a difference all along. Unless they corrected for family-wise error, which by the sound of it they didn't

Zen said...

Nice job unraveling this one.

Penny said...

As I sit here sipping on a cup of coffee I just brewed on my keurig b70 I am thinking that the Prozac I have taken for years has worked for me.

From time to time when I seem to get down I consider other SSSRIs but so far just stick with what I know and wait - I always improve.

Neuroskeptic said...

Zen: Thanks, but I didn't do the unravelling, Jureidini and McHenry did. I just wrote about it.

Neuroskeptic said...

anonymous 13:19 - The BMJ article mentions a suggestion that Pharma studies should be banned from peer-reviewed journals, while they should be forced to reveal the raw data online.

It's a nice idea, but I'm not in favor, because it would lead to Pharma websites becoming de facto a respected source of information - people would cite them as if they were journals. Which is not a good idea.

At least with journals, they have to convince the peer reviewers that their stuff is worth publishing. Which is not trivial - the BMJ mention that Study 329 was submitted to JAMA, but rejected. So peer review does work, at least a little.

Anonymous said...

Perhaps I'm just missing something, but you said:

Here's the worst example. In the original protocol, two "primary" endpoints were specified: the change in the total Hamilton Scale (HAMD) score, and % of patients who 'responded', defined as either an improvement of more than 50% of their starting HAMD score or a final HAMD of 8 or below.

On neither of these measures did paroxetine work better than placebo at the p=0.05 significance level. It did work if you defined 'responded' to mean only a final HAMD of 8 or below, but this was not how it was defined in the protocol.

Explain again how a final HAMD score of 8 or below was not how the protocol defined response if what you said was "defined as either an improvement of more than 50% of their starting HAMD score or a final HAMD of 8 or below."

Neuroskeptic said...

Anonymous: Under the original definition, if you improved by 50%, but you still had a HAMD of 9 (e.g. if your baseline HAMD was 18 or more), you 'responded', under the narrower definition of 8 or less final score, you didn't.

The reason paroxetine didn't work under the original condition is that the % of 'response' on placebo was a lot higher under that definition.

Paul said...

Isn't it an example of the Texas sharpshooter fallacy?

Anonymous said...

NEURO SAYS: At least with journals, they have to convince the peer reviewers that their stuff is worth publishing. Which is not trivial - the BMJ mention that Study 329 was submitted to JAMA, but rejected. So peer review does work, at least a little.

OH PLEASE. THE STUDY WAS PUBLISHED IN THE FLAGSHIP JOURNAL OF ADOLESCENT PSYCHIATRY: JAACP!!!!!

Anonymous said...

Thank you for publicising these scandals and adding your carefully informed critique of them.

It is continually shocking what Big Pharma can get away with in small-minded psychiatry. Or maybe not when considering it's, I gather, the most profitable industry on the planet other than the arms trade. It's a kind of arms trade itself I guess against disease. Well, or unwanted mental states (brain states). Or lack of economic productivity. Or lack of an erection. Big Pharma vs the Ununited States of Suffering or Burden. The first casualty of pharma economy war is scientific method. The second and millionth are human beings.

Love the idea someone mentioned of trying to make the peer review process more transparent and accountable to more people.

Anonymous said...

Btw I see that beneath the abstract of the article on the JAACP website http://bit.ly/em6CXz it discloses that:

"This study was supported by a grant from GlaxoSmithKline".

But it doesn't disclose what has been linked to in this blog post - that the lead author Martin Keller has personally received "numerous grants from...the pharmaceutical industry".

The BMJ article also points out that "Professor Keller and some of the other 22 listed authors had worked for GSK, but this had not been declared".

And ironically, only after the huge list of people listed as actual authors, is given the name of the person who actually authored the first draft of the paper, and what it doesn't say is that she is not one of the psychiatrists or something, but an employee of Scientific Therapeutics Information Inc. which has a track record of dodgy ghostwriting and its website says: "serving members of the pharmaceutical industry and medical associations since 1985"

So I am assuming that neither the abstract webpage nor the full text of the article disclose these financial conflicts of interest of the authors.

veri said...

Anon :) you are such a bad azz. you take cheeky to a whole new dimension.. like bees. it still drives me crazy you know that? *sigh* lol.

Anonymous said...

Neuroskeptic could you rescue other post from Google's overly murderous spam filter if it's in there cheers

Anonymous said...

Very well said, as usual. As you know and write about, the practice of selective reporting is also rampant in basic science and has begun to raise questions about the whole scientific enterprise. Given that the incentives are somewhat (but not entirely) different than pharma, What do you think should be done in that domain to increase transparency?

petrossa said...

They are waking up i guyess:

January 21, 2011

The vast majority of already published and relevant clinical trials of a given drug, device or procedure are routinely ignored by scientists conducting new research on the same topic, a new Johns Hopkins study suggests.

http://www.physorg.com/news/2011-01-clinical-trials-cited-previous-relevant.html

Neuroskeptic said...

Anon 17:45 - Briefly, my view is that we should make all scientists abide by the rules currently in place in the US regarding clinical trials, notably clinicaltrials.gov.

Essentially this means:

1) All study protocols, and this includes a specification of the statistics you plan to use, should be publicly announced before data collection starts.

You are of course free to analyse your data (and anyone else's) however you want, but if you choose to analyse it in a way you didn't plan to do at the start, you have to admit that it's post hoc (which will be verifiable, see above).

2) All results, including raw data, must be made publicly available within a reasonable time-frame after the study's done.

Now there are lots of potential problems here, e.g. if you do a human genetics study, complying with 2) might mean uploading all of your volunteers' genomes to the internet, which they are unlikely to consent to. Presumably you'd "anonymize" it, but still.

However that's a matter of detail. There will be plenty of if's and but's and exceptions. But in general, that's what I'd like to see.

Anonymous said...

>If this were some poor 24 year old's PhD thesis, it would be different

I don't see why. Pusblishing intentionally misleading or outright incorrect works should be discouraged. Maybe a 24 year old would be set down a more honorable career path than shilling for pharmaceuticals - like working the Dairy Queen drive-through, for instance.

Neuroskeptic said...

Anon: 21 January 2011 03:36

Sorry just noticed this comment. Yes, it was a failure of peer review that it got into the JAACAP. Interestingly though, if you read the BMJ article, the JAACAP reviewers were critical. But it got accepted. Maybe not a peer review failure so much as an editorial one...