Stephen Royle

@steveroyle@biologists.social
1.3K Followers
895 Following
4.9K Posts

I'm into chord changes and cell biology.

Professor at University of Warwick. Director of The Company of Biologists. Affiliate of bioRxiv. Author of "The Digital Cell". Views are my own and do not represent any organisation I am associated with.

#CellBiology #MembraneTrafficking #CellDivision #Microscopy #ImageAnalysis #Science #ScienceMastodon #Academia #AcademicMastodon and #Music #Running #Cycling

Lab Websitehttps://roylelab.org
ORCiDhttps://orcid.org/0000-0001-8927-6967
GitHubhttps://github.com/quantixed
Data Analysishttps://quantixed.org

Wow, in Life Sciences 14 Advanced Grants went to the UK (well England actually).

3 went to PIs in Exeter, Sheffield and Newcastle. The rest went to PIs in what we could call the “Golden Bicycle” of Oxford and Cambridge, i.e. 0 to London!

I know it depends on who applies etc. but it's remarkable that University of Cambridge alone got more Life Sciences Awards than France…

Congrats to all awardees.

https://erc.europa.eu/news-events/news/erc-2024-advanced-grants-results

ERC backs cutting-edge research with €721 million in funding

The European Research Council (ERC) today announced the winners of its latest Advanced Grant competition. The funding, worth in total €721 million, will go to 281 leading researchers across Europe. The Advanced Grant competition is one of the most prestigious and competitive funding schemes in the EU. It gives senior researchers the opportunity to pursue ambitious, curiosity-driven projects that could lead to major scientific breakthroughs. The new grants are part of the EU’s Horizon Europe programme.

ERC
Some pretty stark preliminary percentages for the cuts to different parts of science budget in the US. But one of the key messages is that even before the latest stuff, China has overtaken the US in R&D funding and by some margin.
Although industry is paying for 80% of R&D, much of that is much more applied (of course).
Basic, blue skies science in the US is likely to take a pretty serious hit. Extraordinary to watch the government whack one of the key foundations of US prosperity and power.
For Sale.
Baby Shoes.
Ready Salted.
now that the ink has dried I can finally share this with the world: from mid-August I will be the new Data Management Coordinator at @embl - working across all six sites, with all data types (not only imaging)! It's a HUGE new challenge for me and I could not be more excited about it.

Developmental Biology, From Stem Cells to Morphogenesis: an advanced course for MSc and PhD students, run by the Sorbonne and the Curie in Paris.

Open to international students, free registration (not including room and board)

https://training.institut-curie.org/courses/developmental-biology-2025

#DevBio

In the days of the Yellow Pages everything started with “a” to be listed first.
Then computers arrived and everything started with “e”.
Then Apple took over and everything started with “i”.
I look forward to the inevitable coming era of “o” words.

Are there useful or interesting ways to use LLMs other than prompting them? I feel like compressing all the text in the world via a hierarchically structured statistical model is probably useful, but that we're using it in a way that is unlikely to do what we'd hope.

Like everyone I'm impressed and amazed by what they can do, but also very frequently nonplussed by their stupidity. The amazing things convince me there's something important here, but the stupidity seems to have a consistent character that makes me think we're using them in a non-optimal way.

For example, I would expect them to be good at gathering text from their training data that is talking about the same thing in two different ways. This seems like it would be very helpful for synthesizing views on a complex question, but I think not.

I think that part of synthesizing multiple views is building a mental model of the underlying meaning, finding the points of disagreement, and putting that into a new framing. This feels like an inherently back and forth process that LLMs can't do by their very structure.

"Reasoning" models with chain of thought get a little way towards this but they feel like an overkill solution that also isn't enough to really address it. But I have to admit I don't know much about their internals and I've never really had the chance to use them myself.

Another aspect is that I'm not sure it's possible to train models to produce truth, in some sense. I feel like we learn this by living in the world and trying to use the imperfect pieces of knowledge and skills we have to achieve stuff. Without that connection, can it go beyond compression?

So that's why I'm wondering if there is another way to use LLMs that more clearly makes use of the fact that they're an incredible compression scheme? Search seems like one possibility but maintaining sources would undermine their compressing role I guess.

Maybe generating good keywords and alternative phrases that people use when talking about something, that could be the starting point of a literature search? Has anyone tried using them in this way just via prompting? Or maybe there's another way to use the core model without prompting?

Lab preprint now in PubMed:
Nonequivalence of Zfp423 premature termination codons in mice.

In which two overlapping indel variants that predict the same frameshift, stop codon, and potential reinitiation codon have very different outcomes.
https://doi.org/10.1101/2025.05.30.656936

A very honest article that outlines the problems not only with #AI in #science, but with the system of incentives in academic research in general.

The author is a physics graduate who, like many other researchers, has hopped on the AI hype wagon in search for better tools to do his research.

Just to find out that, in spite of many papers claiming that Physics-Informed Neural Networks (PINNs) could be successfully used to solve partial differential equations (PDEs), their results where actually way more underwhelming than they claimed to be, and almost as good as coin tosses when presented with PDEs outside of their carefully cherrypicked training set.

Unlike traditional numeric methods to solve differential equations (which usually operate on a tensor of points and estimate the values at each of those points), a PINN provides an analytical solution and puts the equations into its loss function.

But, for real-world problems, PINNs seem to actually perform worse than the current (non-AI) numeric estimates. The problem is that this is not what emerges from the papers on this topic - the original PINN paper has a whopping 14,000 citations, making it the most cited numerical methods paper of the 21st century.

The underlying issues don’t come to the surface if, in academia like in many other industries, there are no incentives for spotting those issues:

  • Survivorship bias: Researchers rarely write about their studies that didn’t yield statistically relevant results. Publishers are also usually incentivised to publish studies about new stuff that got discovered rather than studies that explored dead ends. And scientific journalists usually get more clicks if they write articles with titles such as “The upcoming AI revolution in physics“ rather than “This guy tried to solve Navier-Stokes equations with AI, and the results weren’t any better than the methods that have been around for the past four decades“. If only 10% of the studies that tested a certain tool in a certain field yielded some results, and the other 90% didn’t, but only that 10% gets published, then people may start to believe that that tool is much more effective than it actually is. This is also a problem in medical studies, not only in AI: if researchers fail to publish negative results, it can cause medical practitioners and the general public to overestimate the effectiveness of a certain medical treatment. For the sake of science, we should also start publishing papers that seem to lead to a dead end. If you just found out that PINNs don’t perform better than traditional numeric methods to solve PDEs, you should still publish your findings. So future researchers have a better idea of what to ignore and what to improve.

  • Arbitrary success thresholds: academia has set very arbitrary thresholds to measure the statistical significance of a given result. If everything below a given sigma is statistically insignificant, and everything above is statistically significant, then researchers are unlikely to publish potentially significant results that just so happened to yield a z-value slightly below the threshold (often because of insufficient or bad data), are instead incentivised to publish results that passed the threshold just by a hair, and may otherwise massage their data until it fits the threshold. This is eventually the biggest reason behind the replicability issue in modern science.

  • Conflict of interests: AI adoption is exploding among scientists less because it benefits science and more because it benefits scientists themselves. Studies that involve AI get big funding by names such as Google, Microsoft and Nvidia, while almost everybody else struggles for the breadcrumbs. And big companies who have invested big money in AI benefit from the narrative that AI is the tool that can solve a lot of problems in science - and this conflict of interests should be called out more often. Because of so much attention in the field, writing a paper that involves AI in some form means that your publication will receive on average 3x the numer of citations. So researchers who work on AI often end up working backwards. Instead of identifying a problem and then trying to find a solution, they start by assuming that AI will be the solution and then looking for problems to solve. Science shouldn’t be a game of hammers looking for nails.

To be clear, I believe that AI can be a very helpful tool in science.

There are fields where AI is already proving to push progress, such as protein folding and climate models.

But this doesn’t seem to be the case for many problems where an analytical solution is required. Maybe that will change in the near future, as models become better at reasoning. But that’s not the current state. And there’s nothing wrong with admitting it. It can’t be that the original paper on PINNs ended up with 14,000 citations, in spite of its obvious cherrypicking and weak baseline, and, since no incentives were present to prove its statistical irrelevance, for years hundreds if not thousands of scientists tried to make science pretending that PINNs worked - just because we’ve created a system of incentives that rewards those who believe that everything is a nail, and AI is the hammer for everything.

@ai @science

https://www.understandingai.org/p/i-got-fooled-by-ai-for-science-hypeheres

UK Science & Technology Select Committee to interview Ottoline Leyser Tues (17 June) at 09.30. Leyser has just stepped down as chair of UKRI. Will be live streamed https://committees.parliament.uk/work/9184/professor-dame-ottoline-leyser-valedictory/
×
Got a letter in The Economist on the skill of public speaking. Does this count as a publication?
@tomstafford Was it peer-reviewed?
@StephenCurry no, but the time from submission to acceptance was excellent!