Proceedings for #pearc23 are up. Looks like I have a lot of reading to do over the next week. https://dl.acm.org/doi/proceedings/10.1145/3569951
A few thoughts on what I've gleaned from #PEARC23 - there's a refreshing focus on providing greater and cheaper access to HPC resources in fields that traditionally don't consider HPC as a resource, there's a push to encourage and uplift younger voices in HPC, and that at the end of the day, it's not about the hardware or software in HPC, but the human element.

The last two years of being in #HPC have been amazing - it's become obvious that it's a rapidly changing and growing field, and despite some of the challenges and setbacks we're constantly facing, the many bright minds of this community are able to overcome all of these.
Akkoma

Containers and security at #Pearc23 today.

Yes, please! Containers are fantastic for so many aspects of research computing, addressing the security aspects unique to this use will be incredibly helpful for adoption into our work

@jhc I think this might be the paper: https://www.nature.com/articles/s41598-023-29340-2

We were talking about better coding practices, reproducibility, and how this can lead to more citations at #PEARC23 yesterday. Nice to see more evidence of that

Bayesian estimation reveals that reproducible models in Systems Biology get more citations - Scientific Reports

The Systems Biology community has taken numerous actions to develop data and modeling standards towards FAIR data and model handling. Nevertheless, the debate about incentives and rewards for individual researchers to make their results reproducible is ongoing. Here, we pose the specific question of whether reproducible models have a higher impact in terms of citations. Therefore, we statistically analyze 328 published models recently classified by Tiwari et al. based on their reproducibility. For hypothesis testing, we use a flexible Bayesian approach that provides complete distributional information for all quantities of interest and can handle outliers. The results show that in the period from 2013, i.e., 10 years after the introduction of SBML, to 2020, the group of reproducible models is significantly more cited than the non-reproducible group. We show that differences in journal impact factors do not explain this effect and that this effect increases with additional standardization of data and error model integration via PEtab. Overall, our statistical analysis demonstrates the long-term merits of reproducible modeling for the individual researcher in terms of citations. Moreover, it provides evidence for the increased use of reproducible models in the scientific community.

Nature

I just gave a talk at SE4RS'23 at PEARC:

“The Changing Role of RSEs over the Lifetime of @ParslProject" by @danielskatz, @benc, Yadu Babuji, Kevin Hunter Kesling, Anna Woodard and Kyle Chard

paper: https://doi.org/10.48550/arXiv.2307.11060
slides: https://doi.org/10.5281/zenodo.8180120

#RSE #RSEng #SE4RS #SE4Science #PEARC23

The Changing Role of RSEs over the Lifetime of Parsl

This position paper describes the Parsl open source research software project and its various phases over seven years. It defines four types of research software engineers (RSEs) who have been important to the project in those phases; we believe this is also applicable to other research software projects.

arXiv.org
Less than a month before #PEARC23 Full Paper Submissions close! If your #HPC, #CompSci, or #AI research is working towards their theme of 'computing for the common good' why not submit your abstract? https://pearc.acm.org/pearc23/call-for-participation/
Tutorial and Workshop Submissions for #PEARC23 are now open! This year's theme? Computing for the Common Good! Read more and apply: https://buff.ly/3Gel3Ja