Clickbait Detection via Large Language Models

📅 2023-06-16
🏛️ arXiv.org
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
This work investigates the effectiveness of large language models (LLMs) for headline-level clickbait detection under zero-shot and few-shot settings. We systematically evaluate mainstream LLMs—including GPT and ChatGLM—across multiple Chinese and English benchmark datasets. Our empirical study reveals, for the first time, that LLMs relying solely on headline text—without fine-tuning—achieve substantially lower performance than task-specific, fine-tuned pretrained language models (PLMs), exposing a critical structural limitation in prompt-based inference for fine-grained content safety tasks. The core contributions are threefold: (1) a cross-lingual, multi-benchmark evaluation framework for LLM-based clickbait detection; (2) empirical validation that unimodal headline input imposes a fundamental information bottleneck, constituting a primary source of performance degradation; and (3) evidence-based delineation of the applicability boundaries of LLMs in content safety, challenging the assumption that LLMs can directly replace fine-tuned PLMs in such domains.
📝 Abstract
Clickbait, which aims to induce users with some surprising and even thrilling headlines for increasing click-through rates, permeates almost all online content publishers, such as news portals and social media. Recently, Large Language Models (LLMs) have emerged as a powerful instrument and achieved tremendous success in a series of NLP downstream tasks. However, it is not yet known whether LLMs can be served as a high-quality clickbait detection system. In this paper, we analyze the performance of LLMs in the few-shot and zero-shot scenarios on several English and Chinese benchmark datasets. Experimental results show that LLMs cannot achieve the best results compared to the state-of-the-art deep and fine-tuning PLMs methods. Different from human intuition, the experiments demonstrated that LLMs cannot make satisfied clickbait detection just by the headlines.
Problem

Research questions and friction points this paper is trying to address.

Evaluating LLMs' effectiveness in clickbait detection
Comparing LLMs with state-of-the-art methods in few-shot scenarios
Assessing LLMs' performance using only headlines for detection
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLMs for clickbait detection in few-shot scenarios
Comparison with state-of-the-art deep PLMs methods
Headline-only analysis insufficient for clickbait detection
🔎 Similar Papers
No similar papers found.
Y
Yi Zhu
School of Information Engineering, Yangzhou University, Yangzhou, China
H
H. Wang
School of Information Engineering, Yangzhou University, Yangzhou, China
Y
Ye Wang
School of Information Engineering, Yangzhou University, Yangzhou, China
Y
Yun Li
School of Information Engineering, Yangzhou University, Yangzhou, China
Y
Yunhao Yuan
School of Information Engineering, Yangzhou University, Yangzhou, China
Jipeng Qiang
Jipeng Qiang
Yangzhou University
Data miningNLP