Discussion: View Thread

  • 1.  Unadjusted Post-Selection Inference: An Uncomfortable Example (Discussion)

    Posted 20 days ago

    Hello all, 

    I want to share a post I wrote under a pseudonym (P. Hackman, PhD). The goal was to expose a real failure mode: how model selection combined with unadjusted post-selection inference can be used to all but guarantee statistically significant results.

    Link: https://data-diction.com/posts/upsi-example/

    I'd be interested in hearing your perspectives on a few questions:

    1) How do you explain post-selection inference issues to collaborators in practice?
    2) Where do you draw the line between routine model building and analysis choices that require adjustment or reframing?
    3) Do satirical examples like this clarify the problem, or do they risk normalizing behavior we'd rather discourage?

    As someone stepping into a leadership role in the section next year, I'd like to help this forum continue to be a place where we regularly have constructive discussions about issues like this. I encourage replies here, or new threads on similar challenges that arise in practice. Our forum of practicing statisticians offers something increasingly rare: context-rich advice shaped by real collaboration, incentives, and consequences. 

    Thank you!



    ------------------------------
    Ryan Peterson
    ASA Statistical Consulting Section Chair-Elect
    Associate Professor
    University of Iowa
    ------------------------------


  • 2.  RE: Unadjusted Post-Selection Inference: An Uncomfortable Example (Discussion)

    Posted 20 days ago

    Thanks Ryan for sharing!

    I have in the past used the story of Baron Munchausen to describe this phenomenon before nailing them to the wall with a chat on  researcher degrees of freedom / the garden of forking paths. In my experience it helps the medicine go down to lead with the humor. For those unfamiliar with the Baron ~

    While traveling, Munchausen finds himself stuck in a deep, muddy swamp. He cannot climb out because the ground gives way beneath his feet.
    Remembering a legend that a man could escape a swamp by pulling on his own hair, he decides to try it. He grabs a lock of his own hair, ties it into a rope‑like knot, and begins to pull upward with all his strength.
    As he tugs, the hair somehow "holds" and he manages to hoist himself out of the mire, emerging dry and triumphant. 

    Cheers all- Happy Holidays and all the best in 2026!

    - AB



    ------------------------------
    Adam Batten
    Lead Statistician & President
    AB EVERGREEN ANALYTICS LLC
    ------------------------------



  • 3.  RE: Unadjusted Post-Selection Inference: An Uncomfortable Example (Discussion)

    Posted 13 days ago
    Ryan:
    I am an economist who frequently has to build forecast models. I learned a long time ago that predicting the future growth rate of any economic variable is a fool's game. Instead I use multiple (nonlinear) regression to generate forecast models for economic variables (GDP, company sales, productivity and hundreds of others). Multicollinearity is almost always a problem, and also almost always a problem that has to be ignored to some extent. All industries, and all of the variables associated with them, are directly connected to the larger economic cycle. I frequently find myself in a position where the best predictors of a given variable are so highly correlated with each other that I can't use more than one of them. So, when I have multiple predictor variables that consistently reach cyclical peaks and troughs prior to the dependent variable I find it difficult to justify excluding them from the model simply because of a higher than desired P-value. Reminding myself that probability is not equal to reality helps.

    What helps more is visualization. When a graph of the variables shows that one of them is a predictor of the other, but a low R squared says it isn't, I generally conclude that the method of least squares is telling me more about a disparity between long-run trend and the short-run predictive consistency at cyclical turns than it is about whether the variable is useful to me as a leading indicator. Useful leading indicators are too valuable to ignore simply because a measure of probability says they aren't useful. In other words, I let the graph tell me if a low R squared is a false negative.

    I realize my comments aren't specific to your email, but modelling is as much art as it is science. While it seems obvious to me that lying with statistics should be prevented to the extent possible, I remain just as concerned that sometimes the statistic is the lie. To be terminologically correct, P-values (and any other probabilistic measure of error) are always reliable, but sometimes they are not valid. To close, I think what you're asking is where do we draw the line. For me, the line is 5 independent variables or less.

    Happy holidays, Steve L-K

    Stephen Latin-Kasper
    Economist / CEO
    414-779-0886
    stephen@coherentforecasting.org