ASA Connect

 View Only
Expand all | Collapse all

Important Meta-analysis paper prepring

  • 1.  Important Meta-analysis paper prepring

    Posted 20 days ago

    The paper, linked below, is now In Press in Statistics in Biopharmaceutical Research. If you have any interest in meta-analysis, which is at the apex of most evidence pyramids,  or are seeking a fertile area for further research, you should pay close attention to this paper.  If you are a peer-reviewer of a report of a meta-analysis of clinical trials, I hope you prioritize the math over tradition to be sure there are evidence-based conclusions.   Start by reading the second and third paragraphs of the introduction, which motivate the importance of this article. Next, Section 2 demonstrates that the current mainstream methods (Inverse variance weighting) represent a misuse of linear combination theory, rendering these estimates potentially seriously biased.  Mainstream methodology can lead to unsupportable conclusions about a therapy. Section 3 produces an asymptotically valid methodology, but it defines its target population differently from the mainstream. The Shuster (2023) reference in the link showed in two examples from major medical journals that unsupportable public health conclusions that affected patient safety were reached.

    Mainstream Meta-Analysis of Clinical Trials produces strongly Inconsistent Estimators



    ------------------------------
    Jonathan Shuster
    ------------------------------


  • 2.  RE: Important Meta-analysis paper prepring

    Posted 20 days ago

    Thank you, thank you, thank you! There are so many issues with meta-analysis. The large data sets with 0% error, etc. I'm so glad that someone has taken the time to bring forward the many concerns with this strategy. I often think about the measurement errors (lab instruments and techniques) associated with each data set and the combined errors. Thank you for your post. Eileen



    ------------------------------
    Eileen Beachell
    ------------------------------



  • 3.  RE: Important Meta-analysis paper prepring

    Posted 17 days ago
      |   view attached

    Hi Jonathan. My distrust of overall effect sizes from meta-analysis (MA) stems not from the statistical methods but from the GIGO (garbage in, garbage out) principle. An overall effect size can be highly misleading due to selection bias and compromised research quality and integrity in the collection and presentation of the evidence (https://www.cochrane.org/evidence/why-our-evidence-trusted). In any event, unlike fixed-effects MA, random-effects MA permits the inclusion of study-level covariates to reduce the nuisance caused by the between-studies variance (see attached). Do you think your ratio estimation method can be extended to accommodate study-level covariates? 
    Eugene



    ------------------------------
    Eugene Komaroff
    komaroffeugene@gmail.com
    ------------------------------

    Attachment(s)

    pdf
    Komaroff Wolfinger MA.pdf   33 KB 1 version


  • 4.  RE: Important Meta-analysis paper prepring

    Posted 16 days ago

    Thanks for your response. Selection bias is a problem which I address (not resolve)  in the discussion.  Patient level data would be very helpful but with a lot of experience with Simpson's Paradox, I do not believe in study level covariates. But yes, these methods can be extended to apply to patient level data-I do mention stratification as well.  With mandatory registries for clinical trials, I think that the problem of missing studies has been resolved. The big drawback to all meta-analysis lies in quitting unpromising therapies.  This suggests bias given there are a large number of studies done.

    Best wishes,

    Jon

     



    ------------------------------
    Jonathan Shuster
    ------------------------------



  • 5.  RE: Important Meta-analysis paper prepring

    Posted 16 days ago

    Hi Jon. Thank you for your reply. I fully agree that patient-level meta-analysis with stratification is the optimal approach for controlling and adjusting for Simpson's paradox, which can arise from treatment-by-patient interactions. Additionally, if there were any discrepancy between the "average" treatment effect observed in a large meta-analysis and the "average" treatment effect from a substantial patient-level randomized controlled trial, I would vote or place greater trust in the patient-level results.

    Eugene



    ------------------------------
    Eugene Komaroff
    komaroffeugene@gmail.com
    ------------------------------



  • 6.  RE: Important Meta-analysis paper prepring

    Posted 14 days ago

    Hi,

    I think the issue Jon is raising basically boils down to a concern that the study size may be informative, i.e., there is correlation between the study size and the effect size. I think it would be clearer to frame the issue in terms of informative study size, as opposed to saying that the problem is that the weights are nonrandom. There are basically two main sources of randomness in the weights used in a standard meta-analysis: (a) randomness arising from the fact that the weights depend on quantities such as the between-study variance in the treatment effect and the within-study variance of the outcome variable, which are generally unknown and have to be estimated, and (b) randomness arising from variation in the sample size from study to study. The standard meta-analysis is based on the assumption that the randomness due to parameter estimation is negligible and the assumption that the study size is noninformative. If the number of studies and the sample size within studies are both large, the assumption of negligible parameter estimation error is probably reasonable, so that the main source of randomness is the variation in study size. The fact that the study sizes are random is not in and of itself the problem. If the study sizes are noninformative, the standard meta-analysis is valid despite the randomness. The problem comes in when the study sizes are informative. There is literature on methods for clustered data with informative cluster sizes that can be brought to bear here. One message that comes out of what Jon is saying is that researchers conducting a meta-analysis should examine whether there is a material degree of correlation between the study size and the treatment effect, and if there is, they should try to identify the source of the correlation.

    Regards,
    David



    ------------------------------
    David Zucker
    Department of Statistics and Data Science
    Hebrew University of Jerusalem
    ------------------------------



  • 7.  RE: Important Meta-analysis paper prepring

    Posted 11 days ago

    My previous post should not be taken as disparagement of meta-analysis, which is a subset of the systematic review methodology as practiced in medical research (Cochrane Reviews), psychosocial research (Campbell Collaboration), and educational research (What Works Clearinghouse), among others. The reviewers enhance their qualitative systematic review with a statistical meta-analysis if the data permits.

    Consider a meta-analysis of sample means where the statistic of interest is the pooled mean. The standard error (SE) in statistical theory is the standard deviation of an infinitely large theoretical sampling distribution of means that is indexed by sample size (n). Therefore, SE is actually a function [SE(n)], not a fixed number. The sample estimates of SE (n) are obtained from each study in the meta-analysis by dividing the sample standard deviation (sd) by the square root of the sample size [se = sd/sqrt(n)]. Because standard error is a ratio, the value is affected by both the numerator (sample standard deviation) and the denominator (sample size). Regarding David's post, even if all the studies in a meta-analysis had the same sample size, the estimated standard errors would still differ because the sample standard deviations vary (hopefully randomly, or only due to sampling error).

    The sample estimates of the standard error are fundamental to the concept of precision in meta-analysis. Precision (1/se2) is the weight that multiplies the corresponding sample mean when computing the pooled (overall) mean. For instance, if one study had se2 = 10 and another had se2 = 100, the mean with the smaller precision (1/se2 = 0.10) contributes more to the pooled mean than the mean that is multiplied by the larger precision (0.01). It is important not to confuse precision with accuracy, as often happens in discussions of confidence intervals. Precision and accuracy are known as reliability and validity in psychometrics. Measurements can be precise but may not be accurate. For instance, one can get the same number by repeatedly stepping on a weight scale. The measurements are precise but not accurate unless the scale is calibrated to zero. Precise measurements may not be accurate, but accurate measurements must be precise.

    What I described above was a fixed effects meta-analysis. A random-effects meta-analysis enables the incorporation of a variance component, which inflates precision. This component represents heterogeneity of the sample means in the analysis. It is problematic when it exceeds what is expected by chance alone (sampling error). This is similar to the workaround for a pooled standard deviation needed for a two-sample t-test, for example, when the assumption of homogeneity of variance is violated. The variance component is an explicit acknowledgement that the means in the meta-analysis may have been estimates of different parameters due to differences in, for example, demographics, interventions, exposures, study designs, outcomes, measurements, and statistical methods. Nevertheless, it is possible to reduce the component by adding study-level covariates to the model. Although some have criticized this as an example of adding apples and oranges, others have argued that there is something to be gained by studying the fruit salad.

    Note that, above, I stated "assume randomly," which acknowledges Jonathan's scholarly, mathematical/statistical concern about the weights not being random. However, the data are assumed to be random samples when conducting inferential analyses (e.g., t-tests, ANOVAs, regression). This assumption is clearly violated when compelling evidence establishes that the observed data could not have arisen from chance variation alone.



    ------------------------------
    Eugene Komaroff
    komaroffeugene@gmail.com
    ------------------------------



  • 8.  RE: Important Meta-analysis paper prepring

    Posted 10 days ago
    Dear Eugene:

    Thanks for this reply. But the key irrefutable mathematical facts against the inverse variance weighted random effects methods for a set of randomized clinical trials are (1) they make contradictory assumptions and (2) they apply linear weighted distribution theory in an illegitimate manner (weights are not constants to a high degree of accuracy). Any statistician, who in a situation that might involve public health policy, uses these methods in such a situation or as a reviewer who allows the use of these methods is potentially risking scientifically unsupportable inferences.

    Fixed-effects are legitimate for the narrow hypothesis that the true effect sizes are zero for all studies. The resulting confidence intervals, which today are the crux of our inferences, cannot be trusted under the more general and realistic random effects scenario. 



    ------------------------------
    Jonathan Shuster
    ------------------------------



  • 9.  RE: Important Meta-analysis paper prepring

    Posted 9 days ago
      |   view attached

    Hi Jonathan. Thank you for reinforcing your strong opposition to inverse-variance–weighted random-effects meta-analysis when high-value outcomes are at stake. Substantive scientific, clinical, or public health harm can result from misleading confidence intervals, biased pooled estimates, and misestimated uncertainty.

    Re: "The resulting confidence intervals, which today are the crux of our inferences…" 

    I have no problem with scrutinizing confidence bounds to test a null hypothesis parameter. For example, checking whether the confidence interval includes zero. But this is equivalent to evaluating a p-value for statistical significance. Both methods lead to the same binary decision: reject or fail to reject the null hypothesis. I prefer the p-value because that makes the decision easier or more direct. The problem begins after the null hypothesis is rejected. Researchers are confident that the interval contains the true value of the alternative parameter. This assumes the interval is an accurate estimator of a population parameter, whereas it reveals only the level of precision, not accuracy. The confidence interval can be narrow, but that does not mean it is accurate.

    The practice of simultaneously rejecting the null hypothesis and estimating population parameters conflates two different statistical goals. (1) Parameter estimation (estimating a true population value) is the goal of population survey research. (2) Hypothesis testing is the goal of much science and engineering, where the aim is to falsify a null hypothesis rather than estimate an alternative parameter. Hypothesis testing only justifies a decision about the existence of an effect, not confidence that the reported interval contains the true effect. Treating the interval as an accurate estimator is a conceptual error.

    Finally, I said in the previous post that some have argued that there is something to be gained from studying a "fruit salad," should be understood as an exploratory analysis only. This is far different from the high regulatory standards required for confirmatory analysis, as produced by properly planned and analyzed randomized, double-blind, placebo (comparator)-controlled clinical trials. It is noteworthy that the FDA's draft Guidance for Industry: "Meta-Analyses of Randomized Controlled Clinical Trials to Evaluate the Safety of Human Drugs or Biological Products" was posted for public comment in 2018 (attached). The comment phase closed, but 8 years later, there is still no final guidance. Perhaps the FDA has also understood that the benefits of meta-analysis for experimental pharmaceutical research do not outweigh the risks.



    ------------------------------
    Eugene Komaroff
    komaroffeugene@gmail.com
    ------------------------------



  • 10.  RE: Important Meta-analysis paper prepring

    Posted yesterday

    Dear Jonathan. First, I want to apologize for my earlier posts. I did not take the time or effort to fully understand the scope of your objection to mainstream meta-analysis. I now completely agree that applying mainstream meta-analytic methods to sparse 0/1 SAE data is dangerous. Such approaches can yield questionable results that influence FDA decisions and, ultimately, adversely affect public health. I evaluated several models for metaanalyzing the cardiovasculardeath data (Nissen & Wolski, 2007, Table3). There were 42 studies, but 19 with double zeros for mortality were excluded by all the methods. Both the mainstream and GENMOD analyses used the Peto method, so thre was no arbitrary continuity correction (0.50). The results are in the table below.

    Meta-Analyses of Rosiglitazone Mortality Risk (23 Studies)

    Model Category

    Method

    Relative Risk (95% CI)

    P-Value

    Likelihood-Based

    GENMOD (Likelihood Ratio)

    1.64 (1.13 – 2.37)*

    0.0059*

    GLIMMIX (Log Link / LR)

    1.55 (0.90 – 2.66)

    0.1102

    Robust / Exact

    Shuster Method

    1.73 (1.08 – 2.75)*

    0.0235*

    Permutation Test

    1.64 (N/A)

    0.0484*

    Bootstrap Analysis

    1.64 (1.27 – 2.93)

    < 0.05*

    Mainstream Meta

    Fixed Effect (Peto)

    1.64 (0.98 – 2.74)**

    0.0590

    Random Effects (Peto)

    1.64 (0.98 – 2.74)**

    0.0590

    *The Fixed and Random Effects RR and 95% CI are identical because the estimated betweenstudy variance (τ 2 = 0.00) reduced the Random Effects model to the Fixed Effects model.

    ** GENMOD (Fixed) and GLIMMIX (Random) both use Likelihood Ratio logic; however, GLIMMIX's estimate of τ^2 = 0.65 resulted in a different RR and a wider, non-statistically significant confidence interval.

    While the mainstream Fixed Effect (Peto and GENMOD), Random Effect, and GLIMMIX models do not reach a consistent consensus on precision and statistical significance, the Robust Shuster method, Bootstrap analysis, and Permutation test all indicate a significantly higher incidence of death (64%) in the treatment group.

    I would like to develop a manuscript for publication with these results and would be honored if you would take the lead as first author. I will be responsible for the Methods and Results sections as the second author. Having archived the code and datasets locally, I have repeated the analyses several times. I am therefore confident that the results are fully reproducible and can be expanded upon as needed.

    Best regards,

    Eugene



    ------------------------------
    Eugene Komaroff
    komaroffeugene@gmail.com
    ------------------------------



  • 11.  RE: Important Meta-analysis paper prepring

    Posted 22 hours ago
    Thank you for the invitation.  I am honored, but as a nearly 10 year retiree, I regrettably decline.  I object to any method that treats study sample sizes as fixed. They are volatile random variables and review material from 9/9 distinguished reviewers answered you to that question. In my course evaluations for my short course on meta-analysis of clinical trials, I asked that question of my class of 32. Twenty-nine answered and all agreed.   One left it blank and two did not turn in the evaluation. I had one former editor and five FDA biostatisticians in the class.

    If sample sizes or weights are correlated with effect size, these methods risk serious bias.   Tossing 25 double zeros is a problem for me.  My method is asymptotic not exact.

    Good luck with your paper.  If I am selected as a reviewer, you should know I am a user friendly editor.  Disagreement on philosophy is not a criterion I use.  But my method is a random-effects method not a fixed effects method so I do not understand the *.

    Best,

    Jon





  • 12.  RE: Important Meta-analysis paper prepring

    Posted an hour ago

    Dear Jonathan,

    I am recently retired and am enjoying the freedom to pursue my own statistical research without the pressure of a paycheck. You noted that 'if you are seeking a fertile area for further research, your article in Statistics in Biopharmaceutical Research is a good place to begin. In 2000, I was a statistician at MetaWorks.com, a firm specializing in systematic reviews and meta-analysis, where Ingram Olkin served as a consultant. Your article has given me a lot of fresh food for thought.

    Thank you for your thoughtful and candid reply to my previous post. I completely respect your decision to remain in your well-earned retirement, though your insights remain as sharp as ever.

    I take your point regarding the asymptotic nature of your ratio method, the volatility of sample sizes, and the inherently random-variable nature of the weights; ignoring sample size by excluding double-zero studies; and the mainstream meta-analysis routine use of odds ratios, which obscure the clearer interpretation offered by risk ratios. I have reanalyzed all 42 studies from Nissen's dataset (again, only RR and no arbitrary 0.5 corrections), and the results again reveal the severe mortality risk associated with Rosiglitazone (see table below).

    I have conducted statistical analyses with SAS for the past 35 years, but I know that proprietary code severely limits reproducibility and replication. When the WWW appeared in the early 90's, it was heralded as the information superhighway. In reality, it was like driving on a busy city street with many traffic lights and distractions. With agentic AI, I am truly driving on an information superhighway. Not only has the Agent helped me quickly debug cryptic SAS errors, but it has also assisted in revising the code that executed my ideas. That is why I was able to rework the analyses so quickly this morning.

    I am now working with Agent AI to convert my SAS code to R. I am confident I can include the R code in the article's appendix, enabling others to replicate the methodology with their own data. In the submitted manuscript, GLIMMIX will be GLMM, and GENMOD will be GZLM.

    Methodological Framework

    Statistical Method

    Relative Risk (95% CI)

    P-Value

    Singly-Asymptotic Ratio

    Cluster-Level Ratio Estimator

    1.73 (1.08 – 2.75)

    0.0235

    Likelihood-Based (Mixed)

    GLIMMIX (Unstructured R-Side)

    1.51 (1.08 – 2.11)

    0.0172

    Likelihood-Based (Fixed)

    GENMOD (Likelihood Ratio)

    1.40 (1.00 – 1.95)

    0.0466

    Non-Parametric / Exact

    Permutation Test

    1.64 (N/A)

    0.0484

    Bootstrap Analysis

    1.64 (1.27 – 2.93)

    < 0.05

    Mainstream (Approx.)

    Fixed Effect (Peto)

    1.64 (0.98 – 2.74)

    0.0590

    Best Regards,

    Eugene



    ------------------------------
    Eugene Komaroff
    komaroffeugene@gmail.com
    ------------------------------



  • 13.  RE: Important Meta-analysis paper prepring

    Posted 3 hours ago

    From the perspective of supporting Jonathan's arguments, the point you make are well taken.  However, there are other issues ignored in this Nissen paper that are probably examples of other reasons why Meta-analyses are flawed, such as the various control groups active and placebo considered as comparable.  The growth and formulaic mechanical approach to meta-analyses goes beyond fixed effects...but that certainly is one issue.  However, the increase in the sheer number may be problematic:  Schumacher FL, Cutter G. The ever growing incidence of Meta-Analyses and Reviews. Mult Scler Relat Disord. 2025 Dec;104:106753. doi: 10.1016/j.msard.2025.106753. Epub 2025 Sep 14. PMID: 40976201.



    ------------------------------
    [Gary] [Cutter]
    [Emeritus Professor]
    [UAB School of Public Healtth]
    ------------------------------