Robbie van Aert and Marcel van Assen have a new article out on a method for estimating average effect size in meta-analyses subject to publication bias, called p-uniform*. The method has been around for a while (the first version of their paper is from 2018, van Aert & van Assen, 2018) and has been cited hundreds of times. The recent appearance of the peer-reviewed article prompted me to take a closer look.
The authors frame the p-uniform* method as a generalization of an earlier proposal of theirs, called p-uniform, which was limited in that it applies only to meta-analytic data where all effect size estimates are statistically significant at the conventional
P-uniform* has the advantage over 3PSM that the weight in the selection model does not need to be estimated. These weights are known to be imprecisely estimated in more complex selection models than the selection model of 3PSM where multiple weights need to be estimated (Hedges & Vevea, 1996, 2005; Vevea & Woods, 2005). P-uniform* only assumes that these weights are the same for the statistically significant and the same for the nonsignificant primary studies’ effect sizes, without requiring estimation of the weights. This makes p-uniform* a more parsimonious model than 3PSM with similar statistical properties as shown in our simulation studies (van Aert & van Assen, 2026, p. 15).
In this post, I offer a different interpretation of p-uniform*. Rather than being a more parsimonious model than the 3PSM, I think it is useful to understand p-uniform* as a different estimator of the average effect size and heterogeneity parameters in the 3PSM. I will demonstrate that the maximum likelihood estimator of p-uniform* is exactly equivalent to the maximum likelihood estimator of 3PSM under the (admittedly artificial) condition that all effect sizes in the meta-analytic data have the same sampling variance. Outside of this condition, the methods are not exactly equivalent but will tend to produce very similar point estimates of average effect size and between-study heterogeneity because both are based on unbiased estimating equations. However, confidence intervals produced by each method will differ except when based on the same method of construction and when all effect sizes have the same sampling variance.
Notation
My discussion is necessarily going to get a bit technical. Consider a meta-analysis that includes
The three-parameter selection model
Before considering how p-uniform* relates to the 3PSM, let me explain the latter model. The three-parameter selection model makes two sets of assumptions. First, it is assumed that if we could observe every effect size estimate ever produced, then those effect size estimates would follow a conventional random effects model with mean
Under these assumptions, the distribution of an observed effect size estimate is what I call a “piece-wise normal” distribution, with density
One way to compute maximum likelihood estimators is by profiling in one or more of the variables. For instance, we can consider the profile log-likelihood of
For the three-parameter selection model, the maximum likelihood estimator of
p-uniform*
Now let me consider van Aert and van Assen’s proposed method, p-uniform*. The method is an estimator of
The objective function
Using conditional distributions for inference is warranted when the conditioning information is an ancillary statistic that carries no information about the model parameters (Ghosh et al., 2010). However, the statistical significance of the effect size estimates are not ancillary for the average effect size and heterogeneity parameters of the 3PSM. van Aert & van Assen (2026) offer no justification for basing inferences about these parameters on the conditional distribution. It is not obvious to me that confidence intervals based on the conditional distribution would be valid—even asymptotically as
Equal standard errors
The main simulations reported in van Aert & van Assen (2026) are based on a data-generating process where all the effect size estimates have equal standard errors, so
In this scenario with equal standard errors, the estimators of
Several implications follow from this. First, the equivalence of the p-uniform* and 3PSM estimators explains why the performance of the point estimators for
Varying standard errors
Except when the standard errors are all equal, the objective function of p-uniform* is not exactly identical to the profile log-likelihood of the 3PSM. However, objective function of p-uniform* produces unbiased estimating equations for
My previous post gave expressions for the mean and variance of an observed effect size estimate under the step-function selection model. For the case of a single-step model, these results imply that
How different is the p-uniform* estimator from the 3PSM maximum likelihood estimator in this case of unequal variances? The degree of potential discrepancy depends on how different the objective functions are. Allowing for unequal variances
van Aert and van Assen argue that p-uniform* is advantageous because it does not entail estimation of the selection weight
Like the 3PSM MLE, p-uniform* can be defined by a set of unbiased estimating equations. Although this property is not sufficient to ensure that p-uniform* estimator will work well, it does suggest that p-uniform* might have biases similar to those of the 3PSM MLE. Furthermore, sandwich variance estimators or non-parametric bootstrapping might be useful for assessing uncertainty in model parameter estimates generated with p-uniform* (as we examine for MLEs in Pustejovsky et al., 2025). However, it seems unlikely that p-uniform* would be more efficient than the MLE of the 3PSM under the model’s assumptions—maximum likelihood estimation is usually pretty hard to improve upon, at least for low-dimensional problems. Does p-uniform* have some robustness advantage that maximum likelihood lacks? In my view, there needs to be more thorough examination of the properties of p-uniform* (be it through theoretical analysis or simulation under varied, realistic data-generating processes) before it can be recommended for conducting meta-analyses of real data.
References
Footnotes
The p-uniform* method does not explicitly define the
, but it is implicit in the assumption that the “weights are the same for the statistically significant and the same for the nonsignificant primary studies’ effect sizes.” Although the weights are constant conditional on whether the effects are affirmative, they are allowed to differ between affirmative and non-affirmative effect sizes, and is the ratio of the weights for non-affirmative effects to the weights for affirmative ones.↩︎