Detecting LLM-assisted writing in scientific communication: Are we there yet?

📅 2024-01-30
🏛️ Journal of Data and Information Science
📈 Citations: 5
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the lack of transparency in LLM-assisted scientific writing by systematically evaluating four state-of-the-art LLM-generated text detectors under realistic research scenarios. We find that general-purpose detectors perform significantly worse than a simple time-based heuristic—namely, detecting abrupt stylistic shifts—when identifying hybrid scientific texts (human-authored + LLM-generated). Our work is the first to reveal the severe inadequacy of current detection methods in handling gradual, non-uniform LLM integration patterns prevalent in scientific writing. Consequently, we argue for the development of a domain-specific detection framework tailored to the mixed-authorship characteristics of scientific texts. A key contribution is the empirical validation that stylistic discontinuity—not just lexical or statistical anomalies—serves as a more robust detection signal. This insight establishes a new paradigm for ensuring traceability and upholding academic integrity in LLM-augmented scholarly communication.

Technology Category

Application Category

📝 Abstract
Abstract Large Language Models (LLMs), exemplified by ChatGPT, have significantly reshaped text generation, particularly in the realm of writing assistance. While ethical considerations underscore the importance of transparently acknowledging LLM use, especially in scientific communication, genuine acknowledgment remains infrequent. A potential avenue to encourage accurate acknowledging of LLM-assisted writing involves employing automated detectors. Our evaluation of four cutting-edge LLM-generated text detectors reveals their suboptimal performance compared to a simple ad-hoc detector designed to identify abrupt writing style changes around the time of LLM proliferation. We contend that the development of specialized detectors exclusively dedicated to LLM-assisted writing detection is necessary. Such detectors could play a crucial role in fostering more authentic recognition of LLM involvement in scientific communication, addressing the current challenges in acknowledgment practices.
Problem

Research questions and friction points this paper is trying to address.

Detecting LLM-assisted writing in scientific communication
Evaluating performance of existing AI-generated text detectors
Developing specialized detectors for accurate LLM use acknowledgment
Innovation

Methods, ideas, or system contributions that make the work stand out.

Evaluated four cutting-edge LLM-generated text detectors
Proposed simple ad-hoc detector for writing style changes
Advocated developing specialized detectors for LLM-assisted writing
🔎 Similar Papers
No similar papers found.
T
T. Lazebnik
Department of Mathematics, Ariel University, Ariel, Israel; Department of Cancer Biology, Cancer Institute, University College London, London, UK
Ariel Rosenfeld
Ariel Rosenfeld
Associate Professor at Bar-Ilan University
Artificial IntelligenceHuman-Agent InteractionAI for Social GoodScientometrics