Stephen Wild

@stephenjwild
317 Followers
299 Following
237 Posts
Sometimes I try to put straight lines through things. Most times I fail. Try to be Bayesian when I can. Views my own. RT or like != endorsement
Websitesjwild.github.io
@kjhealy @siracusa @atpfm A nice related article from 1995 (!)
Hayes, Brian. 1995. “Debugging Myself.” American Scientist
https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.208.9970&rep=rep1&type=pdf. "Other oncoming developments are not so encouraging. As computers become more powerful, the size of both the microstate and the macrostate grows exponentially. This means there will be more pieces susceptible to failure, and quadratically more interactions among those pieces." Connecting to the network *vastly* expands the state space ...

📈 Markov Chain Monte Carlo (MCMC) played a crucial role in modeling Covid-19 spread during the pandemic 🦠

I found this to be a good free resource for learning MCMC:

https://avehtari.github.io/BDA_course_Aalto/

#MCMC #datascience

Bayesian Data Analysis course

@peter_ellis I don’t think it’s well known - my supervisor mentioned it off hand and me and the postdocs in the group weren’t aware. This blog cites what I think is the article my supervisor sent us: https://tompepinsky.com/2019/04/29/multiple-imputation-with-colliders/
Multiple Imputation with Colliders

I have found myself thinking a lot recently about multiple imputation in the presence of colliders. Proponents of MI commonly recommend that any variable available in the dataset should be included…

@cameronpat @healthstatsdude

I was forced to experience the real world. I did not like at all.

@peter_ellis @cameronpat

Stop ruining multiple imputation for me. But also, really? I did not know this.

The {marginaleffects} 📦 book is now online! 25 chapters on post-estimation analyses and interpretation with #Rstats. The 📖 is full of tutorials, case studies, tips, and technical notes. Please check it out and let us know how we can improve this resource vincentarelbundock.github.io/marginaleffects

I just noticed our review paper "Prior Knowledge Elicitation: The Past, Present, and Future" with Mikkola, Martin, Chandramouli, Hartmann, @oriolabril, Thomas, @henri_pesonen, Corander, I, @samikaski, @paul_buerkner, and Klami is now online in Bayesian Analysis https://doi.org/10.1214/23-BA1381

Supported by @FCAI

#Bayesian #PriorElicitation #newpaper

Prior Knowledge Elicitation: The Past, Present, and Future

Specification of the prior distribution for a Bayesian model is a central part of the Bayesian workflow for data analysis, but it is often difficult even for statistical experts. In principle, prior elicitation transforms domain knowledge of various kinds into well-defined prior distributions, and offers a solution to the prior specification problem. In practice, however, we are still fairly far from having usable prior elicitation tools that could significantly influence the way we build probabilistic models in academia and industry. We lack elicitation methods that integrate well into the Bayesian workflow and perform elicitation efficiently in terms of costs of time and effort. We even lack a comprehensive theoretical framework for understanding different facets of the prior elicitation problem. Why are we not widely using prior elicitation? We analyse the state of the art by identifying a range of key aspects of prior knowledge elicitation, from properties of the modelling task and the nature of the priors to the form of interaction with the expert. The existing prior elicitation literature is reviewed and categorized in these terms. This allows recognizing under-studied directions in prior elicitation research, finally leading to a proposal of several new avenues to improve prior elicitation methodology.

Project Euclid

Today on my Substack:

Around 1 in 270 people in India die from snakebites by the age of 70.

Here's why that statistic is so surprising, and why it's a sign of one of the most underrated problems in global health: missing data.
https://open.substack.com/pub/salonium/p/14-how-many-people-die-from-snakebites?utm_source=share&utm_medium=android

#14: How many people die from snakebites?

This week: Missing data – the most underrated problem in global health.

Scientific Discovery

New `loo` CRAN release https://cran.r-project.org/package=loo

- loo_predictive_metric(): with MAE, MSE and RMSE, and accuracy and balanced accuracy for binary target. (by Leevi Lindgren)

- crps(), scrps(), loo_crps(), and loo_scrps() for computing the (Scaled) Continuously Ranked Probability Score. (by LL)

- Vignette “Mixture IS leave-one-out cross-validation for high-dimensional Bayesian models.” https://mc-stan.org/loo/articles/loo2-mixis.html demonstrating https://arxiv.org/abs/2209.09190 (by Silva and Zanella)

#rstats #Stan #Bayes

loo: Efficient Leave-One-Out Cross-Validation and WAIC for Bayesian Models

Efficient approximate leave-one-out cross-validation (LOO) for Bayesian models fit using Markov chain Monte Carlo, as described in Vehtari, Gelman, and Gabry (2017) <<a href="https://doi.org/10.1007%2Fs11222-016-9696-4">doi:10.1007/s11222-016-9696-4</a>>. The approximation uses Pareto smoothed importance sampling (PSIS), a new procedure for regularizing importance weights. As a byproduct of the calculations, we also obtain approximate standard errors for estimated predictive errors and for the comparison of predictive errors between models. The package also provides methods for using stacking and other model weighting techniques to average Bayesian predictive distributions.

It is popularly claimed that long-term trends in national and international well-being are influenced by the widespread availability and use of the internet, social media, & smartphones. It's a worrying idea, but not supported by research.

In our new paper @matti and I analysed data from 2,414,294 individuals (15 years & up) across 168 countries from 2006 to 2021. We tracked eight kinds of well-being against three forms of technology.

Please read, comment, and share!

https://psyarxiv.com/jp5nd