In 1993, H. Schwabl published in @PNASNews a seminal paper: “Yolk [as] a source of maternal testosterone for developing birds”

This was the first study proposing a link between maternal egg hormones and fitness.

Our preregistered #systematicreview & #metaanalysis in Ecology Letters synthesises 438 effects from 57 studies on 19 wild 🐦species to test if & how egg hormones relate to fitness

📰 https://doi.org/10.1111/ele.70100

Data & Code https://github.com/ASanchez-Tojar/meta-analysis_egg_hormones_and_fitness

Pre-registration https://doi.org/10.17605/OSF.IO/KU47W

Our #metaanalysis tested the adaptive significance of egg hormones in birds by combining evidence published since 1993, mostly experiments (76% effects).

In short: regardless of hormone type, we found little evidence for an effect of egg hormones on fitness-related traits, but high heterogeneity.

📰https://doi.org/10.1111/ele.70100

We tested several biological and methodological hypotheses:

🔹 Offspring age
🔹 Experiments vs Observational
🔹 Experiments on eggs vs mothers
🔹 Fitness proxy
🔹 Etc.

However, they did not explain much heterogeneity.

Importantly, most heterogeneity was associated with phylogeny and within-study variation, with negligible differences among studies

That is, on average, studies didn't differ in what they found: Our results are generalizable (replicable) among studies

📰 https://doi.org/10.1111/ele.70100

In the light of a recent article (https://doi.org/10.1186/s12915-024-02101-x), showing substantial heterogeneity among analysts.

We tested the robustness of our findings to several analytical decisions. Our results were consistent across.

Also, though evidence for small-study & decline effects is widespread (https://doi.org/10.1186/s12915-022-01485-y), we did NOT find clear evidence of #publicationbias in our dataset. Presumably thanks to our efforts to obtain nonreported results directly from authors.

📰 https://doi.org/10.1111/ele.70100

Same data, different analysts: variation in effect sizes due to analytical decisions in ecology and evolutionary biology - BMC Biology

Although variation in effect sizes and predicted values among studies of similar phenomena is inevitable, such variation far exceeds what might be produced by sampling error alone. One possible explanation for variation among results is differences among researchers in the decisions they make regarding statistical analyses. A growing array of studies has explored this analytical variability in different fields and has found substantial variability among results despite analysts having the same data and research question. Many of these studies have been in the social sciences, but one small “many analyst” study found similar variability in ecology. We expanded the scope of this prior work by implementing a large-scale empirical exploration of the variation in effect sizes and model predictions generated by the analytical decisions of different researchers in ecology and evolutionary biology. We used two unpublished datasets, one from evolutionary ecology (blue tit, Cyanistes caeruleus, to compare sibling number and nestling growth) and one from conservation ecology (Eucalyptus, to compare grass cover and tree seedling recruitment). The project leaders recruited 174 analyst teams, comprising 246 analysts, to investigate the answers to prespecified research questions. Analyses conducted by these teams yielded 141 usable effects (compatible with our meta-analyses and with all necessary information provided) for the blue tit dataset, and 85 usable effects for the Eucalyptus dataset. We found substantial heterogeneity among results for both datasets, although the patterns of variation differed between them. For the blue tit analyses, the average effect was convincingly negative, with less growth for nestlings living with more siblings, but there was near continuous variation in effect size from large negative effects to effects near zero, and even effects crossing the traditional threshold of statistical significance in the opposite direction. In contrast, the average relationship between grass cover and Eucalyptus seedling number was only slightly negative and not convincingly different from zero, and most effects ranged from weakly negative to weakly positive, with about a third of effects crossing the traditional threshold of significance in one direction or the other. However, there were also several striking outliers in the Eucalyptus dataset, with effects far from zero. For both datasets, we found substantial variation in the variable selection and random effects structures among analyses, as well as in the ratings of the analytical methods by peer reviewers, but we found no strong relationship between any of these and deviation from the meta-analytic mean. In other words, analyses with results that were far from the mean were no more or less likely to have dissimilar variable sets, use random effects in their models, or receive poor peer reviews than those analyses that found results that were close to the mean. The existence of substantial variability among analysis outcomes raises important questions about how ecologists and evolutionary biologists should interpret published results, and how they should conduct analyses in the future.

BioMed Central

Indeed, something hard to digest was that:

➡️ Complete data extraction was only possible for 22/57 included studies

➡️ 20 additional studies were excluded due to preventable reporting issues

➡️ 8 authors had lost the data FOREVER 😭

📢 PLEASE, publish your data! #opendata 📢

(and if you want to help spread the word about the importance of #OpenScience, join @sortee!)

Though we didn't find clear evidence for a relationship between egg hormones and fitness proxies, this does NOT mean hormones are unimportant

The high heterogeneity found suggests context dependency

📢 We need more research and better reporting! 👇

📰 https://doi.org/10.1111/ele.70100

Thank you to all the authors who replied to our emails 🙏

Also to the Ecology Letters editors & reviewers for the most thoughtful criticism I’ve ever received during peer review. It’s not often that peer review makes a difference.

Thank you all🙏

📰 https://doi.org/10.1111/ele.70100