Papers
arxiv:2512.05100

Structured Document Translation via Format Reinforcement Learning

Published on Dec 4
· Submitted by Haiyue Song on Dec 9
Authors:
,
,
,
,
,
,

Abstract

Format Reinforcement Learning enhances structured text translation by optimizing structure-aware rewards and distinguishing between minor errors and major structural failures.

AI-generated summary

Recent works on structured text translation remain limited to the sentence level, as they struggle to effectively handle the complex document-level XML or HTML structures. To address this, we propose Format Reinforcement Learning (FormatRL), which employs Group Relative Policy Optimization on top of a supervised fine-tuning model to directly optimize novel structure-aware rewards: 1) TreeSim, which measures structural similarity between predicted and reference XML trees and 2) Node-chrF, which measures translation quality at the level of XML nodes. Additionally, we apply StrucAUC, a fine-grained metric distinguishing between minor errors and major structural failures. Experiments on the SAP software-documentation benchmark demonstrate improvements across six metrics and an analysis further shows how different reward functions contribute to improvements in both structural and translation quality.

Community

AACL2025_StructureDoc_Song_poster_small

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2512.05100 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2512.05100 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2512.05100 in a Space README.md to link it from this page.

Collections including this paper 1