Signs of undeclared ChatGPT use in papers mounting
Signs of undeclared ChatGPT use in papers mounting
I’ve heard many papers are published to never be read by humans. It only makes sense that some portion of those papers aren’t written by humans either.
I wonder what the overlap is between AI assisted papers and papers with few to no readers.
The whole system should get ready for the 21st century.
Most of the scientists arent great writers. It does not make sense to still force them to be a good writer.
Let be fishes be good at swimming instead of climbing trees.
In a modern world where basically EVERYTHING is specialized and no generalist is alive anymore we should make use of language tools.
Hell Chatgpt writes an introduction which is fun to read instead or my overcomplicated bullshit that I would have brought up
One important thing is that you have potential. ChatGDP will write something alright-ish, but it's literally impossible for it to move beyond that. It doesn't have the power of creativity.
Writing is painful, but it also helps us think clearer about our work and contribution. I think it's an important part of the process of doing science, no matter which field. And one gets better at it with training.
it's literally impossible for it to move beyond that. It doesn't have the power of creativity.
This is the key - it does not create, it can only copy. Which is good enough to fool us - there's enough stuff to copy out there that you can spend your whole life copying other people and nobody will ever notice you're not actually creating anything new. What's more, you'll probably come across as pretty clever. But you're not creating anything new.
For me, this poses an existential threat to academia. It might halt development in the field without researchers even noticing: Their words look fine, as if they had thought it through, and they of course read it to make sure it's logically consistent. However, the creative force is gone. Nothing new will come under the sun - the kind of new thoughts that can only be made by creative humans thinking new thoughts that have never been put on paper before.
If we give up that, what's even the point of doing science in the first place.
There's a difference between:
Using ChatGPT to help write parts of the text in the same way you'd use a grammar- or spell-checker (e.g. if English isn't your first language) after you've finished the experiments
Using ChatGPT to write a paper without even doing any experiments
Clearly the second is academic misconduct. The first one is a lot more defensible.
Yes, absolutely. But I still think it has its dangers.
Using it to write the introduction doesn't change the substance of the paper, yet it does provide the framework for how the reader interprets it, and also often decides whether it'll be read at all.
Maybe worse, I find that it's oftem in the painful writing and rewriting of the introduction and conclusion that I truly understand my own contribution - I've done the analysis and all that, but in forcing myself to think about the relevance for the field and the reader I also bring myself to better understand what the paper means in a deeper sense. I believe this kind of deep thinking at the end of the process is incredibly valuable, and it's what I'm afraid we might be losing with AI.
This is the key - it does not create, it can only copy.
I have asked ChatGPT to write poetry on subjects that I know with great certainty have never had poems written about them.
You can of course shuffle around the meanings of "create" and "copy" to try to accommodate that, but eventually you end up with a "copying" process that's so flexible and malleable that it might as well be creativity. It's not like what comes out of human brains isn't based on stuff that went into them earlier either.