How do we make sense of the body of scientific literature that is growing explosively to the point where no individual could read all the relevant papers, and is contaminated with fraudulent and LLM-generated papers? I think that science is not currently equipped to deal with this, and we need to.

I think a critical part will be post-publication peer review. With such rapid growth and time pressure on scientists, pre-publication PR cannot maintain sufficient standards so we need a way to make the informal networked review (journal clubs, conference chats etc.) more transparent and shared.

We also need ways to summarise and make connections between many related papers. I know that many people are hoping that LLMs will step up into this role, but right now that seems very risky and I don't see that changing any time soon.

LLMs are too distracted by surface level presentation, and can be manipulated at scale by iterating over multiple presentations until the LLM summarises it in the way you want it to. In addition, they're known to have problematic biases, and it's unclear if this can be fixed.

I think we need to be experimenting with ways to distribute the work of summarising and making connections between papers, and aggregating that into a collective understanding. An individual can't read all the papers, but collectively we can and already are. We just need ways to integrate that.

In principle we could do this with a post-publication peer review system that allows reviewers to annotate with connections, e.g. in reviewing paper X you create a formal link saying that part of this paper is similar to paper Y, or uses the same technique, etc.

One issue is that these annotations might become corrupted or manipulated in the same way that papers, journals and peer review have been. How do we fix that? It's not ideal, but one option might be some form of trust network: I trust X, they trust Y and thereore I (to a lesser extent) trust Y.

This would mean our summary or evaluation of the literature would depend on our individual trust network. But, this isn't a bad thing in principle. Diversity of opinion is good: there shouldn't be one definitive reference summary because that's a single point of failure and target for manipulation.

All these ideas require experimentation, and both technology development and a cultural shift towards collectively taking responsibility for doing this work. I think we need to do it and would love to hear others' ideas about how to do it and how to convince everyone that we need to.

#science #metascience #academicchatter

@neuralreckoning

I think we very simply need a trust network that works like internet certificates, except more grass roots.

So trust would go:

mentor -> young researcher

young researcher writes a paper.

mistake in the paper -> researcher get untrusted -> mentor gets untrusted -> all papers the prof participated in get untrusted. -> papers that cite it get untrusted. etc.

Then comes a new reevaluation step that checks if the papers still hold without that dependency.

@bmaxv right - it shouldn't be binary but that's the idea. I haven't seen any analysis of whether or not this would work for scientific literature. Maybe it has problems that aren't obvious to me right now, but it feels like if we want a scalable solution it's going to have to be something like this.

@neuralreckoning @bmaxv I'll propose that the gold standard is to be found more in replication than review.

I can recall precious few papers where I've implemented the method described and not had to make corrections. I've found incorrect equations in well-respected journals, key missing details in numerical method papers, horrendous errors in methods in engineering papers, and even a brutal mistake in a major reference text.

That really only happens, though, when I've implemented the method or the math in question.

It *is* more common these days to see full data sets published, which is a step in the right direction, for sure.

There's no "Journal of Replicated Results" as such, nor does replication do much for young researchers' chances to advance in academia. (Getting your interns to replicate methods is a good torture test, though 😈)

I'm most trusting of results from researchers where I've done that replication.

@TallSimon @bmaxv I've had many of those same issues and the problem is there's no way to tell people other than to individually tell them, which doesn't scale well. That's precisely why we need a highly visible and transparent system of post publication peer review.
@TallSimon @bmaxv a replication should be a particularly prominent type of review comment, in this view.

@neuralreckoning @bmaxv @dvdgc13

Perhaps replication, and reporting replication, is something we can also encourage in the nascent #OpenScienceNetwork (based on @bonfire) as well.

Edit: IIRC Royal Society papers print a lot of review responses. Theirs might be a pattern to replicate.

@TallSimon @neuralreckoning
I've seen several links to Bonfire, but I still don't understand what is it trying to do and what makes it stand out (and promising)? I would appreciate any explanations or pointers

@nsmarkov @dvdgc13 @neuralreckoning

My understanding is that @bonfire is about federated components and #OpenScienceNetwork https://openscience.network/ is in the early stages of a using those to build a federated science network, aimed at hosting citable work.

I think ORCID and DOI's have been a good step forward. Journals such as ETNA in my field have loosened the grip of the big journals, SIAM remains inexpensive and well respected. SSRN, arXiv and friends are getting worse, though. Academia.edu and ResearchGate feel too much like LinkedIn, too centralized, too much low-quality algorithmic recommendation.

Personal home pages and department pre-print archives (a la Web1) still do the trick, I source from those a lot. I think having federated components to use for those would be a win. It might make the review/feedback loop a little better.

I'm outside of academic research, kinda doing my own thing, but could see my firm setting up an #OpenScienceNetwork node for things that we can collaborate on.

Open Science Network

Reclaim scientific discourse with federated digital spaces where researchers shape their own conversations, data, and collaborations

@TallSimon @nsmarkov @dvdgc13 is it built on ActivityPub? I'd feel a bit worried about building something on a relatively niche protocol (sorry mastodon).

@neuralreckoning @nsmarkov @dvdgc13 "Niche" eh? I see ActivityPub all over the docs. I'd say "yes", but I don't have the experience to win a protocol bun fight 🙂.

The road behind us is littered with crash test dummies. Chances are we're all going to play that role at some point; I'll probably try it out, sit in the driver's seat... I'll let you know how it goes 😁

@TallSimon @nsmarkov @dvdgc13 haha, no shade on bonfire or fedi. I'd feel much less comfortable basing it on ATproto!