🤖 AI Summary
This study addresses the limited creativity of ChatGPT in literary translation from English into Dutch, Chinese, Catalan, and Spanish. We systematically investigate how prompt engineering, temperature settings (0.2–1.0), and text granularity (sentence- vs. paragraph-level) affect translational creativity. Methodologically, we introduce a novel multidimensional human evaluation framework and a quantitative creativity scoring formula. Results show that minimalist prompts (e.g., “Translate… creatively”) combined with temperature = 1.0 significantly enhance creativity—outperforming DeepL on Spanish, Dutch, and Chinese tasks. AI-generated translations achieve 70%–85% of the creativity level of professional human translations, yet remain systematically inferior to expert human performance. This work provides a reproducible methodology and empirical foundation for controllable creative optimization in LLM-based literary translation.
📝 Abstract
This study examines the variability of Chat-GPT machine translation (MT) outputs across six different configurations in four languages,with a focus on creativity in a literary text. We evaluate GPT translations in different text granularity levels, temperature settings and prompting strategies with a Creativity Score formula. We found that prompting ChatGPT with a minimal instruction yields the best creative translations, with"Translate the following text into [TG] creatively"at the temperature of 1.0 outperforming other configurations and DeepL in Spanish, Dutch, and Chinese. Nonetheless, ChatGPT consistently underperforms compared to human translation (HT).