🤖 AI Summary
This study investigates whether large language models (LLMs) exhibit identifiable, stable, and individualized linguistic styles enabling automatic model attribution. Focusing on GPT-3.5, GPT-4, and Bard, we propose a lightweight supervised classification framework—employing SVM and Random Forest—based on multidimensional linguistic features: part-of-speech distributions, dependency structures, sentiment polarity, and lexical statistics. We systematically demonstrate, for the first time, that LLMs possess robust, task- and prompt-invariant linguistic “fingerprints.” Extensive experiments achieve an average attribution accuracy of 88% across diverse prompts and tasks. Our work confirms that distinct LLMs exhibit significant, consistent fine-grained stylistic differences and establishes the first interpretable, high-accuracy, low-computational-overhead paradigm for LLM provenance identification—providing critical technical support for model regulation, content溯源, and accountability assignment.
📝 Abstract
Large Language Models (LLMs) are capable of generating text that is similar to or surpasses human quality. However, it is unclear whether LLMs tend to exhibit distinctive linguistic styles akin to how human authors do. Through a comprehensive linguistic analysis, we compare the vocabulary, Part-Of-Speech (POS) distribution, dependency distribution, and sentiment of texts generated by three of the most popular LLMS today (GPT-3.5, GPT-4, and Bard) to diverse inputs. The results point to significant linguistic variations which, in turn, enable us to attribute a given text to its LLM origin with a favorable 88% accuracy using a simple off-the-shelf classification model. Theoretical and practical implications of this intriguing finding are discussed.