Papers
arxiv:2412.01340

A 2-step Framework for Automated Literary Translation Evaluation: Its Promises and Pitfalls

Published on Dec 2, 2024
Authors:
,
,
,
,
,

Abstract

A two-stage pipeline evaluates literary machine translation from English to Korean, providing fine-grained metrics that correlate better with human judgment than traditional metrics, though still falling short in areas like Korean Honorifics.

AI-generated summary

In this work, we propose and evaluate the feasibility of a two-stage pipeline to evaluate literary machine translation, in a fine-grained manner, from English to Korean. The results show that our framework provides fine-grained, interpretable metrics suited for literary translation and obtains a higher correlation with human judgment than traditional machine translation metrics. Nonetheless, it still fails to match inter-human agreement, especially in metrics like Korean Honorifics. We also observe that LLMs tend to favor translations generated by other LLMs, and we highlight the necessity of developing more sophisticated evaluation methods to ensure accurate and culturally sensitive machine translation of literary works.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2412.01340 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2412.01340 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.