techhub.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
A hub primarily for passionate technologists, but everyone is welcome

Administered by:

Server stats:

5.4K
active users

#PeerReview

11 posts11 participants0 posts today

Sorry, aber manchmal fordern Reviewer Nachbesserungen, bei denen ich mich frage: Warum willst du jetzt, dass da eine Referenz genannt wird? Manche Dinge sind so dermaßen obvious, dass ich nicht wüsste, was irgendeine Referenz auf irgendein Paper, was irgendwer mal irgendwann geschrieben hat, beitragen würde? Ich referenziere ja auch nicht die Aussage, dass der klare Himmel tagsüber in der Regel blau aussieht? Glaubt ihr mir nicht? Soll ich referenzieren? Meine Güte... #peerreview #academia

Continued thread

Update. "Peer review is a cornerstone of academic publishing, but essentially no formal training exists at the [undergraduate] or graduate medical education levels to prepare trainees for participation in the process as authors or reviewers. This clinical research primer presents an introductory set of guidelines and pearls to empower trainee participation in the peer-review process as both authors and reviewers."
thieme-connect.de/products/ejo

www.thieme-connect.deThieme E-Journals - Journal of Neurological Surgery Reports / AbstractThieme E-Books & E-Journals

Finished the revisions today of Part 3 of the 4-Part discussion section of my paper Universal Basic Income and Value Realization: Metaphysical Insights for Implementation in the United States. The writing’s been done for months—this is tone, structure, and formatting as I prepare for peer review. One more major part to revise. #WritingProgress #UBI #ValueCreation #PostTaxation #EconomicPhilosophy #PeerReview #BasicIncome #InvestmentBasedUBI #UniversalWealth #MetaphysicsOfEconomics #SoundFinance

Back in the days I used a certain platform to post a lot about trains, well about train problems, while commuting in Germany. Now I use the Swiss train system quite a lot and could only send praise and love to #sbb.
…so, every time I come here, I am tempted to post about #peerreview and how broken academic publishing is. Well, and about politics. But then: nobody wants to read more rants about those issues. And then I leave. :wyd:

The Ethical Grey Areas of Machine Writing in Higher Education

Once we start to examine how academics actually use conversational agents in real settings, it becomes harder to draw a clear distinction between problematic and unproblematic use. To entirely substitute machine writing for your own, while presenting it under your own name, would strike most as problematic. But this often has little relationship to how machine writing is drawn upon in practice, at least by academics, not least of all because substantive direction is necessary to produce outputs which aren’t generic or vacuous.

Unless you’re willing to explain to the machine what you want it to do, its capacity to meet your needs as an academic will be limited. The specific characteristics of academic work, the extremely specialized forms of output we are expected to produce for equally precise purposes, means that at least some engagement will be necessary in the process. At which point we are faced with confusion of distinguishing between fully accepted resources we draw upon in our intellectual work and those such as machine writing which are seen as potentially contentious. This is a question which Atwell poses very succinctly:

How does collaboration with others and using all the resources we now see as legitimate (the internet, research papers, colleagues work/advice) to do the best work we can, differ from utilising GenAI tools?

https://nationalcentreforai.jiscinvolve.org/wp/2024/08/08/do-you-feel-guilty-about-using-genai/

Through this framing, Atwell draws attention to what machine writing shares with other resources we might draw upon in our intellectual work. There are weaknesses to any resource, which we will ideally always review in light of our understanding of where it is limited. In this sense, machine writing could be seen as simply another resource amongst others. Atwell argues for a “symbiotic relationship where technological efficiency is balanced with human insight, creativity and ethical judgement.”

The challenge is that detecting machine writing is fundamentally unreliable. There will always be plausible deniability and an obvious cost to making accusations on the basis of false positives. The statistical “burstiness” and “perplexity” which a detector like GPT Zero uses are features which can be found in some writers at least some of the time. The formal character of academic writing and the specialized vocabulary in use means that it will tend towards moderate perplexity, in the sense that the word choices in a sequence will tend to be unusual relative to less specialized forms of writing.

As I’ve just demonstrated, academics are prone to using complex constructions which might be edited out of monographs but are likely to figure in journal articles. Obviously these conventions vary between disciplines and fields. I’ve often been struck by how different the writing styles are between the social theory journals which are my natural home and the computational social science and linguistics journals I increasingly explore due to my interest in GenAI. There are conventions which can be found in some fields which involve a low level of burstiness, such as presenting an argument as a sequence of propositions which could lead to a text being misdiagnosed as machine written. These are just examples from my own narrow experience to illustrate the complexity involved in identifying machine writing in scholarly publishing. The problem isn’t the detectors failing to measure features of the texts but rather that what they’re measuring varies across academic writing in rather convoluted ways.

If AI detectors were fully integrated into the workflows of scholarly publishing, it simply wouldn’t be possible to reliably infer the presence of machine writing in a fully automated way. There would need to be a conversation about what the AI detectors had inferred about the writing. Given there’s a widely acknowledged crisis of review in scholarly publishing, it’s difficult to see how this labor would be integrated into a system already struggling under the weight of submissions. Would already overburdened managing editors be expected to take up this role? Would it be delegated to the editorial board? What if the author simply denies the allegation?

It is very difficult to conceive of how such a system could work in practice, at least in the context of scholarly publishing as it is currently organized. In the event it was enforced automatically, we would likely see an undesirable shift in the character of academic writing in order to minimize characteristics which lead text to be flagged by detectors. Furthermore, academics who wanted to ensure their machine writing could pass undetected could simply ask conversational agents to modify the features of the text to evade these measures. Asking ChatGPT or Claude to modify the perplexity and burstiness of a paragraph can be an effective way to grasp what these measures point to in practice, if these concepts remain abstract to you. But it also indicates how the very flexibility of conversational agents, the fact you can specify in great detail what kind of text you want to be produced, limits how effective these detectors can ever plausibly be.

In the absence of reliable detection, there has been a tendency to seize upon questionable evidence in the hope we might find some way of identifying machine writing. It was widely reported that ‘delve’ figured more heavily in machine writing by ChatGPT than in natural language. The original claim seems to come from a Search Engine Optimization (SEO) firm AI Phrase Finder, which offers a free tool to identify “common AI phrases” that might lead machine writing to be downgraded in search results. It invokes a dataset of 50,000 ChatGPT responses as evidence for this claim without providing any information about this dataset or how it was collected.

It’s odd that the ‘delve’ claim provoked so much attention online given that it only figured ninth on the list of ten most common ChatGPT words, beyond words like ‘leverage’ and ‘resonate’. I chose those words, fourth and sixth on the list respectively, because I used both of them in a short conversation before sitting down to write this section. This might reflect the peculiar conversations I often have, as well as the peculiar words I tend to use in them. But if you’re reading a blog post on academic writing and AI, then I suspect you share this peculiarity to the extent you’re more likely to write ‘captivate’ (#2), ‘dynamic’ (#7) and ‘delve’ (#9) than most ChatGPT users. The reason that ‘delve’ attracted so much attention is that tech guru Paul Graham, founder of the startup accelerator Y Combinator which launched the career of OpenAI’s Sam Altman, claimed confidently on Twitter that ‘delve’ was a sure sign of an e-mail being AI generated.

It’s a comforting idea that ChatGPT has red flags that give away its use, as if it were an initially overwhelming poker player whose tells we gradually identify as we proceed with the game. The problem is that we will never be able to infer confidently from those signs that what we are reading is machine generated. This is a state of affairs which is unsettling in its novelty, calling into question assumptions which would have barely been visible to us until recently.

These claims can still have an impact even if they lack a firm foundation. If scholarly publishers integrated AI detection into their workflows, I expect lists of words to avoid in academic writing would similarly circulate amongst academics, regardless of whether there were solid grounds to believe these would in fact be flagged. If we accept that we can’t conclusively know if a text has been machine generated, then it leaves us in an uncomfortable position. It is easy to see how paranoia could spread under these circumstances when we know that machine writing is circulating but we cannot establish where it is and who is producing it.

If we outsource ever increasing amounts of our writing to automated systems, which are by their nature epistemic blackboxes, how could we sustain trust in the knowledge we are producing? If we are unsure which academic writing is produced by other scholars and which is produced by machines, how will this change how we relate to what we read? How will we distinguish between the different forms which machine writing can take, ranging from skillful co-production through to lazy outsourcing, if we make such a distinction at all? Will we gravitate towards writing which feels authentically human, even if the markers we draw upon to inform such a judgment are liable to be intensely unreliable?

Whose writing will be imbued with the dignity of human authenticity and whose will be written off as machine generated, even when there’s no such machine at work in the process? There’s a risk that publication profiles which don’t match received expectations, particularly those by scholars who don’t match the hegemonic vision of an academic, might find themselves dismissed and repudiated on the assumption that machine writing explains the quantity or quality of what they have written.

Consider the forcefulness with which what are fundamentally intuitions of malpractice have been levied by academics, certain that a wrong has been committed, against their students. Now imagine that same forcefulness directed towards colleagues, inflected through the prevailing competitive individualism of the academy. The trawling of academic profiles, the nocturnal consulting of fundamentally unreliable AI-detectors, and the academic gossip liable to accompany existing vendettas when the legitimacy of a scholar’s writing can suddenly be called into question in a fundamentally unverifiable way.

How do you prove you’ve not used machine writing in your work? Unless you’ve effectively self-surveilled and “human marked” your work, to use GPTZero’s (2024) terminology, there’s no way to prove this negative. You might establish on the balance of probability that you have plausibly written what you claim to have written, but to even find yourself in the position where this is under issue would itself be unpleasant.

Artificial intelligence · Do you feel guilty about using GenAI? - Artificial intelligenceHave you ever felt like you were cheating when using GenAI tools like ChatGPT or Copilot 365? You’re not alone, I’ve heard this a lot recently, but let’s unpack why this might not be the case. Traditionally when I start a new piece of work, particularly when it is something I have never done before […]
Replied in thread

At Paper-Wizard, we're not here to replace peer review - we're here to strengthen it. Our pre-peer-review system helps authors improve their work before submission, and gives editors and reviewers more time to focus on scientific merit. The future of publishing is collaborative - with AI and human expertise working together to advance science.

We invite researchers to experience how AI can strengthen their submission process at paper-wizard.com

#AIinScience #PeerReview"

paper-wizard.comPaper-Wizard | Transform Your Academic ResearchGet comprehensive academic paper reviews in minutes. Trusted by 2000+ researchers at leading institutions worldwide.

Publishers trial paying peer reviewers — what did they find? nature.com/articles/d41586-025

Would like to see trials/research on free publications in exchange for peer review. APC's are too expensive for most southern hemisphere countries...

www.nature.comPublishers trial paying peer reviewers — what did they find?Two journals embarked on efforts to compensate reviewers, with different results.

Siamo tutti d'accordo: il lavoro non retribuito è sfruttamento. Eppure, nell'ambiente accademico, si ritiene accettabile che la #PeerReview sia gratuita.

Ma cosa succederebbe se le riviste pagassero i reviewer? Due esperimenti suggeriscono che le review sarebbero più veloci e di pari qualità.

Publishers trial paying peer reviewers — what did they find?
nature.com/articles/d41586-025

www.nature.comPublishers trial paying peer reviewers — what did they find?Two journals embarked on efforts to compensate reviewers, with different results.

"For scientists, imagining a world without arXiv is like the rest of us imagining one without public libraries or GPS. But a look at its inner workings reveals that it isn’t a frictionless utopia of open-access knowledge. Over the years, arXiv’s permanence has been threatened by everything from bureaucratic strife to outdated code to even, once, a spy scandal. In the words of Ginsparg, who usually redirects interview requests to an FAQ document—on arXiv, no less—and tried to talk me out of visiting him in person, arXiv is “a child I sent off to college but who keeps coming back to camp out in my living room, behaving badly.”

Ginsparg and I met over the course of several days last spring in Ithaca, New York, home of Cornell University. I’ll admit, I was apprehensive ahead of our time together. Geoffrey West, a former supervisor of Ginsparg’s at Los Alamos National Laboratory, once described him as “quite a character” who is “infamous in the community” for being “quite difficult.” He also said he was “extremely funny” and a “great guy.” In our early email exchanges, Ginsparg told me, upfront, that stories about arXiv never impress him: “So many articles, so few insights,” he wrote."

wired.com/story/inside-arxiv-m

WIRED · Inside arXiv—the Most Transformative Platform in All of ScienceBy Sheon Han

Liebe Editor*innen, Conference Chairs und akademische Verleger*innen – wenn ihr wissenschaftliche Arbeiten zur doppelblinden anonymen #PeerReview-Begutachtung verschickt, achtet doch bitte darauf, dass die PDF-Dateien keine Metadaten enthalten, aus denen man die Namen oder Zugehörigkeiten der Autor*innen ablesen kann.

Ansonsten weiß man halt, wer das Paper geschrieben hat und die ganze weitere Anoymisierung ist für die Katz.

Vielen Dank an alle.

#AcademicWriting
#ScientificWriting