↓ Skip to main content

Dove Medical Press

Imbalance p values for baseline covariates in randomized controlled trials: a last resort for the use of p values? A pro and contra debate

Overview of attention for article published in Clinical Epidemiology, May 2018
Altmetric Badge

About this Attention Score

  • Above-average Attention Score compared to outputs of the same age (64th percentile)
  • Average Attention Score compared to outputs of the same age and source

Mentioned by

twitter
6 X users

Citations

dimensions_citation
4 Dimensions

Readers on

mendeley
18 Mendeley
Title
Imbalance p values for baseline covariates in randomized controlled trials: a last resort for the use of p values? A pro and contra debate
Published in
Clinical Epidemiology, May 2018
DOI 10.2147/clep.s161508
Pubmed ID
Authors

Andreas Stang, Christopher Baethge

Abstract

Results of randomized controlled trials (RCTs) are usually accompanied by a table that compares covariates between the study groups at baseline. Sometimes, the investigators report p values for imbalanced covariates. The aim of this debate is to illustrate the pro and contra of the use of these p values in RCTs. Low p values can be a sign of biased or fraudulent randomization and can be used as a warning sign. They can be considered as a screening tool with low positive-predictive value. Low p values should prompt us to ask for the reasons and for potential consequences, especially in combination with hints of methodological problems. A fair randomization produces the expectation that the distribution of p values follows a flat distribution. It does not produce an expectation related to a single p value. The distribution of p values in RCTs can be influenced by the correlation among covariates, differential misclassification or differential mismeasurement of baseline covariates. Given only a small number of reported p values in the reports of RCTs, judging whether the realized p value distribution is, indeed, a flat distribution becomes difficult. If p values ≤0.005 or ≥0.995 were used as a sign of alarm, the false-positive rate would be 5.0% if randomization was done correctly, and five p values per RCT were reported. Use of a low p value as a warning sign that randomization is potentially biased can be considered a vague heuristic. The authors of this debate are obviously more or less enthusiastic with this heuristic and differ in the consequences they propose.

X Demographics

X Demographics

The data shown below were collected from the profiles of 6 X users who shared this research output. Click here to find out more about how the information was compiled.
Mendeley readers

Mendeley readers

The data shown below were compiled from readership statistics for 18 Mendeley readers of this research output. Click here to see the associated Mendeley record.

Geographical breakdown

Country Count As %
Unknown 18 100%

Demographic breakdown

Readers by professional status Count As %
Student > Ph. D. Student 4 22%
Student > Master 2 11%
Researcher 2 11%
Professor 1 6%
Lecturer > Senior Lecturer 1 6%
Other 2 11%
Unknown 6 33%
Readers by discipline Count As %
Medicine and Dentistry 4 22%
Biochemistry, Genetics and Molecular Biology 2 11%
Psychology 2 11%
Nursing and Health Professions 2 11%
Unknown 8 44%
Attention Score in Context

Attention Score in Context

This research output has an Altmetric Attention Score of 4. This is our high-level measure of the quality and quantity of online attention that it has received. This Attention Score, as well as the ranking and number of research outputs shown below, was calculated when the research output was last mentioned on 27 March 2022.
All research outputs
#6,628,531
of 23,426,104 outputs
Outputs from Clinical Epidemiology
#258
of 739 outputs
Outputs of similar age
#114,523
of 327,138 outputs
Outputs of similar age from Clinical Epidemiology
#12
of 24 outputs
Altmetric has tracked 23,426,104 research outputs across all sources so far. This one has received more attention than most of these and is in the 70th percentile.
So far Altmetric has tracked 739 research outputs from this source. They typically receive a lot more attention than average, with a mean Attention Score of 14.3. This one has gotten more attention than average, scoring higher than 63% of its peers.
Older research outputs will score higher simply because they've had more time to accumulate mentions. To account for age we can compare this Altmetric Attention Score to the 327,138 tracked outputs that were published within six weeks on either side of this one in any source. This one has gotten more attention than average, scoring higher than 64% of its contemporaries.
We're also able to compare this research output to 24 others from the same source and published within six weeks on either side of this one. This one has gotten more attention than average, scoring higher than 50% of its contemporaries.