I've read through this paper a couple times and started looking at its data.

https://www.nature.com/articles/s41586-023-06767-1

I'm predicting it now before I get further into reviewing this: this paper is wrong for all the reasons you'd suspect and isn't going to hold up to any scrutiny.
But managers and non-scientists everywhere are going to use it to reign in remote workers and say that it's "supported by the science".

#science #remotework #innovation

Remote collaboration fuses fewer breakthrough ideas - Nature

Analysis of research articles and patent applications shows that members of teams that collaborate remotely are less likely to make breakthrough discoveries than members of on-site teams.

Nature

After reviewing "Remote collaboration fuses fewer breakthrough ideas" more thoroughly, I believe it should be retracted. Here is a write-up of why this is the case:

https://jrhawley.ca/2023/12/01/remote-work-disruption

#science #retractions #remotework #innovation #statistics

Remote collaboration fuses fewer breakthrough ideas? Probably not.

An interesting paper was published in Nature two days ago, titled 'Remote collaboration fuses fewer breakthrough ideas'. I believe it should be retracted.

I can summarize the problems with this paper as follows:

1. Confounded relationships between the explanatory and response variables that are not properly adjusted for,
2. Ill-defined response variable that is insufficiently explored and whose quantitative value does not match the interpretation the authors suggest,
3. Large amounts of data with NaNs that are not discussed,
4. A large dataset with most of its variance contained in < 1% of the overall data,
...

5. Insufficient analysis and discussion of raw data,
6. Insufficient description of methods,
7. Inappropriate statistical methods and interpretations of results,
8. Results shown do not refute a hypothesis that contradicts the authors’ original hypothesis, and
9. Overly strong conclusions that are not supported by the evidence provided.

Like I say in the blog post, the most likely explanation for the observations the authors make it Berkson's paradox:

https://en.wikipedia.org/wiki/Berkson's_paradox

The authors do not adjust for the web of confounding relationships that exist in their data, the most important relating to publication year, time since publication, and the COVID-19 pandemic.

Berkson's paradox - Wikipedia

Here is a causal diagram I drew of the variables the authors talk about in the paper, as well as variables they probably should have included.

Now I feel some sense of professional obligation to submit a "letter of concern" to the journal, but I can't find any links to do this directly.

It seems like I'm supposed to email Nature's main address to kick start the process. Does anyone have a better idea about how to do this?

#science #retractions #peerreview

Alright, I've made minor corrections to my blog post, emailed the authors expressing my concerns about this paper, and posted the concerns on @PubPeer (the comment should show up after moderator approval).

https://pubpeer.com/publications/7DD790DB75C7B6DEEDA47125241E3C

I also put my code on GitHub for easier inspection: https://github.com/jrhawley/linRemoteCollaborationFuses2023/

Let's see how this goes.

PubPeer - Remote collaboration fuses fewer breakthrough ideas

There are comments on PubPeer for publication: Remote collaboration fuses fewer breakthrough ideas (2023)

I haven't received any reply from the corresponding authors since I contacted them 2 weeks ago. I'm emailing them again to see if we can discuss these concerns.

I don't want to write to the journal unless I have to, because I know how painful that process is going to be and how much extra work it's going to take. But is that my only option if I can't get a hold of any of the authors?

#PeerReview #science

@jrhawley generally that's how it works. I follow Retraction Watch and from their postings it seems like a) authors never respond and b) journals take for*ever* to respond as well. Getting a paper retracted apparently takes years. But you can submit a commentary/rebuttal as a letter, I would think, if full-out retraction isn't worth your time.

@jrhawley When I was involved in something like this back in 2009, we submitted it as a "brief communication arising", a category that I think is now "matters arising":

https://www.nature.com/ncomms/submit/matters-arising

Matters Arising | Nature Communications

Matters Arising

Corrections, Retractions and Matters Arising | Nature

Corrections, Retractions and Matters Arising

@jrhawley thanks so much for this!! I think funding is an underdiscussed confounding factor.

IMHO: Funding is highly concentrated to a few rich universities which produce some great research, while poorer academics need to take any collaborator they can get leading to more geographic dispersion.

@PhilippBayer Right, me too. I don't study productivity and the "economics of science", but from all the PIs I've talked to about running their labs, funding is the thing that PIs spend most of their time working on because it so drastically impacts the type of research they are able to do in their labs.

How can that not affect the types of papers you can publish and how widely they'll be read/cited? How is that not an important factor in how local/remote your projects and collaborations are?

@jrhawley Exactly!!! For a long time I was in an Australian OK-funded lab, but we could never start any collaborations with US-well-funded labs.
We just couldn't offer anything they didn't already have in-house or could have easily bought, leading to tiny collaboration networks on their side...
@jrhawley @bstacey Nature prefers Matters Arising to be quite brief, <1,200 words. And this is an interesting expectation: “Comments should ideally have been sent to the authors of the paper under discussion before submission to Nature, so that disputes can be resolved directly whenever possible and points on which both parties agree removed from the submitted contribution.”

@StatGenDan @bstacey Yeah, I read that and opted to contact the authors directly, first. I haven't reached out to the journal, yet. Maybe I won't have to.

I obviously need to give the authors some time to read and think about what I've written. And that Matters Arising page says Nature will consider post-publication peer review for 18 months after publication. So I don't think there's any rush for that, yet (aside from the thousands of people accessing the article while I wait, of course)

@StatGenDan @bstacey My blog post is obviously >> 1200 words, because there's so much to talk about. So I wonder how concise I'm going to have to be if I need to go down that route
@jrhawley Repost it on pubpeer (or your blog link), it will be much more seen with more chances for the crappy magazine to act. Put te link in comment to the paper on the N. website, so they can't completly ignore it. Surprisingly, it went through probably two rounds of reviewing... and you took it apart in a few hours.

@jrhawley your point about cultural and language barriers was one of the best points in your critique.

My reading of your critique leads me to think that having an experienced humanities researcher in the team (ie. Increased diversity) would have helped.

What also would have helped would have been discussing it with researchers who are highly marginalised who are forced to do remote collaborations.

@rowlandm Completely agree. Involving domain area experts from the beginning would eliminate so many of these problems.
And you're right, the unequal impact of a paper like this on marginalized communities is so important. It is way too easy to see bad logic extending from this:

1. We need in-person work to be disruptive
2. Marginalized people need remote work to contribute fully to the workplace
3. Thus, we cannot have disabled and marginalized people on our team if we want to be disruptive

@rowlandm Is remote work perfect? No. Is it great for everyone? Also no. Are there major benefits to remote work? Yes! Let's figure out what and how they work.

I'm probably preaching to the choir, here, but just like curb cuts, remote work has huge benefits for disabled and non-disabled people, alike.

- live-transcription of talks for deaf people
- stable working conditions for new parents
- the ability to avoid bad weather and risky travel conditions
- reduced carbon emissions from commuting

@jrhawley I wouldn't still be in my current job if they forced me to work onsite.

@jrhawley and even the inverse of those questions give the same answer.

Is onsite work perfect? No. Is it great for everyone? Also no. Etc etc

@rowlandm Absolutely! It's always easier to question the new stuff than ask the same questions of the old stuff. As you said, in-person work doesn't work for so many people, and they've mostly been ignored until recently.

I don't like the idea that this paper can (and probably will) used to continue justifying the status quo when it doesn't show anything close to what it says it does.

@jrhawley @rowlandm Given the pandemic, I'd refuse to work in an office permanently. Not happening.
I'm grateful to have been able to work remotely for the last 24 years, and my working conditions in my HO far exceed what any office offered. I've managed fully distributed teams. If they're on another continent, might as well be at home rather than some office.
I do miss conferences though, so I'm definitely in favor of augmenting remote work with meetings in person, eventually again.

@jrhawley very interesting, but you lost me at "OLS assumes the response data comes from a normal distribution." Not at all. The hypothesis tests on coefficients are only valid if the _residuals_ are normally distributed.

That said, I do agree with much of your critique, and plan to add this paper to the class on "misleading statistics".

@jrhawley from the following paragraphs I reckon you refer to the Maximum Likelihood approach where you model the distribution of every y_i, but that's not a necessary assumption for OLS. (A bit nitpicking, I know, but details matter).
@JorisMeys Fair enough, you are right about that and I did gloss over that point. I can edit that to make it clearer.
@jrhawley Great piece! @ MBP biostats students - this is why you learn the assumptions of a linear regression model :)

@calderds Hey Calder, thank you! Nice to hear from you again.

Hahah, understanding linear models is important! But that is probably one of the least problematic things about this paper. The causal graphs, the confounding, the proxy variables and relationships - that's where the meat is.

If I can continue being your biostats TA for a bit longer, go watch this lecture series:

https://www.youtube.com/watch?v=FdnMWdICdRs&list=PLDcUM9US4XdPz-KxHM4XHt7uUVGWWVSus

And go read this book: http://bayes.cs.ucla.edu/BOOK-2K/

Statistical Rethinking 2023 - 01 - The Golem of Prague

YouTube
@jrhawley That's probably what it's meant for. But don't worry, in just a few years it will be retracted and in a couple of decades some managers may find out what "retracted" means. Or centuries...
@jrhawley
Excellent, well-explained analysis! How the “disruption metric” is derived indeed seems to disqualify their interpretation. Disruption or excellence invoke strong associations, but they are inherently vague - any author-defined metric centred on them is bound to run into problems.
Reference type must skew the disruption metric too: method sections often refer to specifics rather than precedence - so I would expect pure method (e.g. software) papers having inflated “disruption” values.