Jennifer Stevens asks why the babies who didn't receive *only* the assigned
treatment, be excluded from the analysis entirely.
When the original analysis by intent to treat has been carried out, you can,
later, go through and pick out only the babies who got the assigned
treatment, and see how things stack up then. But if you substituted that
kind of analysis for the one that includes all the babies enrolled and
randomized to each group in an article reporting on an RCT, you would be
guilty of something at least bordering on scientific fraud. What happens
after randomization is what you are out to report, not just what happens to
a small subset of the babies after randomization.
To illustrate this, what if only 2 of 100 babies in each group actually got
the assigned treatment as it was prescribed? Far-fetched, I know, but if
all the other babies were excluded, you would have no basis on which to do
statistical analysis. You'd also have lots of reasons to go back to the
drawing board when designing your next study, but that's another issue.
A good thing to look for as a marker of a good epidemiologic research
article is a flow chart, showing how many people were in their pool of
possible participants, and following the entire process showing clearly how
many people disappeared at each step - those who didn't get asked to
participate, those who were excluded for specified reasons, those who
declined to participate how many were randomized to each group, how many in
each group who dropped out along the way so that data were simply not
available on their outcomes, and so on, until you reach the final number,
which should be the exact number of participants included in the statistical
analysis of the results.
Also, as Susan Burger so aptly points out, there is a huge difference
between being statistically significant and being clinically significant as
well. It seems this study was designed to have sufficient statistical power
to detect differences between groups if the incidence of the outcomes they
were examining had been higher. The rarer the outcome, the more patients
you will need to enroll in each group in order to detect a STATISTICALLY
significant difference between them. NEC was less prevalent than expected
in all the groups, so that even though the numerical differences between
groups were striking, it did not achieve statistical significance. In
plainer terms, this means that the researchers could not say for sure that
the differences in NEC were due to the different treatments; it could have
been chance, based on the accepted statistical criterion for significance.
As a clinician treating individual babies, it behooves one to take a look at
the actual numbers, and to reflect on the difference between clinical and
statistical significance. Statistical significance can be entirely
uninteresting clinically too. Research doesn't do the job of reflecting and
reasoning for us - it gives us information that may enable to us to make
better clinical decisions when we reason and reflect on what to do with the
baby and mother in front of us.
Another way the press could have reported on this study was 'results
suggested a dose-dependent, beneficial effect of mother's milk and of donor
milk, but the number of babies enrolled was too small for this to reach
statistical significance'. In this world of lamentably dumbed-down news
reports, that is probably too complex a message. But if, as Marsha Walker
posted, there are already hospitals where clinicians are prepared to change
practice based on the mass media version of a single study, the problem
isn't limited to the media.
For the record, it is highly unlikely that Schanler or any of the other
authors had any influence on the media spin on the story.
Rachel Myr
Kristiansand, Norway
And don't even ASK me about p-values!
***********************************************
To temporarily stop your subscription: set lactnet nomail
To start it again: set lactnet mail (or digest)
To unsubscribe: unsubscribe lactnet
All commands go to [log in to unmask]
The LACTNET mailing list is powered by L-Soft's renowned
LISTSERV(R) list management software together with L-Soft's LSMTP(R)
mailer for lightning fast mail delivery. For more information, go to:
http://www.lsoft.com/LISTSERV-powered.html
|