🤖 AI Summary
To address insufficient creative diversity and repetitive, monotonous outputs in large language models (LLMs) for story generation, this paper proposes an adaptive decoding strategy. The method integrates concept-level and narrative-level similarity penalties, employs a two-stage logits correction mechanism to dynamically balance diversity and coherence, and leverages neuron activation analysis to enhance creative modeling. Evaluated on multiple story generation benchmarks, our approach achieves up to a 2.6× improvement in output diversity and reduces repetition rates by an average of 30%, significantly mitigating text degeneration. The core contribution lies in the first integration of fine-grained semantic similarity modeling with neuron activation guidance, enabling controllable and interpretable multi-branch narrative generation.
📝 Abstract
Large Language Models (LLMs) often generate repetitive and monotonous outputs, especially in tasks like story generation, due to limited creative diversity when given the same input prompt. To address this challenge, we propose a novel decoding strategy, Avoidance Decoding, that modifies token logits by penalizing similarity to previously generated outputs, thereby encouraging more diverse multi-branch stories. This penalty adaptively balances two similarity measures: (1) Concept-level Similarity Penalty, which is prioritized in early stages to diversify initial story concepts, and (2) Narrative-level Similarity Penalty, which is increasingly emphasized later to ensure natural yet diverse plot development. Notably, our method achieves up to 2.6 times higher output diversity and reduces repetition by an average of 30% compared to strong baselines, while effectively mitigating text degeneration. Furthermore, we reveal that our method activates a broader range of neurons, demonstrating that it leverages the model's intrinsic creativity.