RE: https://fediscience.org/@snakemake/116523521544819091

Software provenance with #Snakemake: Using the reporter plugin for nanopublications, we can now get slightly improved nanopublications like this one: https://w3id.org/np/RAmgzfta63xx0wWc_zzQVm7kwOc4tsEOA0JJJCfsiLL1g (press on the little blue arrow on the right to see the full details). Automatically captured for this workflow: https://w3id.org/np/RAjHDlPDghZzc9ZvQ3uJQNJ9Jd_KAYzZt7dk5PXKgjRyE - again expressed a nanopub declaration. 😉

It now supports to capture the "classic" software support for #Conda and Snakemake wrappers.

There is more work to do. Let's see when and if I get to it.

#reproducibleComputing #softwareprovenance #nanopub

RE: https://fediscience.org/@snakemake/116443839004108636

Now, folks, this is my first release of a Snakemake reporter plugin!

I am pretty proud of it, but it would have been impossible without the input of @johanneskoester , @tkuhn and - most of all - @fbartusch !

What is it about?

Have you ever read a #bioinformatics paper and thought: "Hugh? How did they get there? This is absolutely not #reproducibleComputing , because I cannot apply it!"

The Snakemake Workflow Management System helps researchers by generating a self-contained HTML report with all bells and whistles (all software metadata, runtime stats and publication-ready figures). And yet, I felt it is time to easy writing the materials & methods section.

Declare a worfklow #nanopub using this template: https://w3id.org/np/RAOT7z3RA0XYlHIikne8rfUUYZrtHyrzXBD1HpI_GvcRk and get a declaration like this: https://w3id.org/np/RAjHDlPDghZzc9ZvQ3uJQNJ9Jd_KAYzZt7dk5PXKgjRyE

Use this reporter plugin and get all metadata reference in one other nano publication: https://w3id.org/np/RAK9xz_ccnu0Xhs4vX2KtqCxX44mmSt6nq-ePLeewMrFE

1/3

The #SnakemakeHackathon2026 has ended, we are still preparing our preprint release. But, our host has prepared a note on their homepage: https://go.tum.de/946236 🥳

#Snakemake #ReproducibleComputing

Snakemake Hackathon 2026 – faster and further improved reproducible data analysis

Snakemake Hackathon 2026 delivers even faster and further improved support for reproducible data analysis. The 2026 Snakemake Hackathon brought together about 40 experts in Munich to deliver major improvements, new features, and performance enhancements for reproducible data analysis. Their community driven work strengthens Snakemake’s scalability, usability, and support for modern scientific workflows.

Munich Data Science Institute

RE: https://fediscience.org/@snakemake/116295568336688286

This is a big step forward: The SLURM plugin for Snakemake now supports so-called job arrays. These are cluster jobs, with ~ equal resource requirements in terms of memory and compute resources.

The change in itself was big: The purpose of a workflow system is to make use of the vast resources of an HPC cluster. Hence, jobs are submitted to run concurrently. However, for a job array, we have to "wait" for all eligible jobs to be ready. And then we submit.

To preserve concurrent execution of other jobs which are ready to be executed, a thread pool has been introduced. In itself, I do not see job arrays as such a big feature: The LSF system profited much more from arrays than the rather lean SLURM implementation does.

BUT: the new code base will ease further development to pooling many shared memory tasks (applications which support no parallel execution or are confined to one computer by "only" supporting threading). Until then, there is more work to do.

#HPC #SLURM #Snakemake #SnakemakeHackathon2026 #ReproducibleComputing #OpenScience

Did you know? We started to gather cluster profiles for Snakemake users in a little repo: https://github.com/snakemake/snakemake-cluster-profiles

They should serve as a template for others.

#HPC #Snakemake #ReproducibleComputing #SnakemakeHackathon2026

RE: https://fediscience.org/@snakemake/116222696140712833

What a week at the #SnakemakeHackathon2026 !

What a wonderful week with wonderful people!

We were pretty productive and this #Snakemake release is just the peak of it. The list of features, bug fixes, performance improvement and additional documentation is so long — our little announcement robot cannot display it all. Even here on FediiScience with its 1500-character limit!

#ReproducibleComputing #OpenScience

The #SnakemakeHackathon2026 just started! Lots of interesting topics. If you are curious: #SnakemakHackathon or #SnakemakeHackathon2026 are your hashtags to follow. 😉

#reproducibleComputing #OpenScience #Datananalysis

What do you see here? This is an example knowledge graph describing a #Snakemake analysis workflow. You see the workflow description, a linked data set and a linked report.

All work done to boost #HPC user support for those conducting their workflows on HPC systems (you can run Snakemake on other platforms, too).

My to-do list:
- an assertion template for workflows: ✅
- another for reports: ✅ (simple datasets are already in the #nanopub verse)
- a plugin to gather software metadata and publish as a nanopub ❌ (half done: #SnakemakeHackathon2026 )

Kudos to @nanopub / @tkuhn and @johanneskoester - without them this pursuit would (have been) futile! And my feeling is that @fbartusch will play an important role in any further development ...

#OpenScience #ReproducibleComputing

I want to reach out: I have this pending release for the SLURM executor (https://github.com/snakemake/snakemake-executor-plugin-slurm/pull/412 ). It implements better error feedback (in case of hardware failures and otherwise). It would need some thorough checking, and I cannot provoke too many hardware failures. Is anyone working on an older cluster? 😉

Feedback (as issues) would be appreciated. Also nice, if you tell me it is working, here.

#Snakemake #HPC #SLURM #ReproducibleComputing

chore(main): release 2.3.0 by snakemake-bot · Pull Request #412 · snakemake/snakemake-executor-plugin-slurm

🤖 I have created a release beep boop 2.3.0 (2026-02-12) Features attempt for detailed error reporting for job-steps (#327) (79c986e) excluding broken nodes from further job submission (#411) (dddd...

GitHub

The last blog post I wrote was about Life Science Support on HPC clusters. Honestly? It was more of a rant. Not a good blog post.

So, someone suggested I delete it, which I did. It took me a long time to recover. Now, I have re-written this blog post. I think it is better. I weighted every phrase. It still has some rant flavour.

Here it is: https://blogs.fediscience.org/rupture-de-catenaire/2025/12/15/empowering-life-scientists-what-hpc-support-should-look-like/

The upcoming articles will have more of an example character. But I still will not be able to update on a regular basis.

#Bioinformatics #HPC #ReproducibleComputing #Snakemake #Nextflow

Empowering Life Scientists: What HPC Support Should Look Like – rupture de caténaire