Abstract
Fact-to-text generation can allow for the generation of high-quality, informative texts such as Wikipedia articles. Cross-lingual fact-to-text generation (XF2T) involves using facts available in a language, typically English, and generating texts in a different language based on these facts. This is particularly relevant for low and medium-resource languages, which have relatively structured informative content. This work explores the problem of XF2T for generating long text from given facts with a specific focus on generating factually grounded content. Unfortunately, previous work either focuses on cross-lingual facts to short text or monolingual graph to text generation. In this paper, we propose a novel solution to the multi-sentence XF2T task, which addresses these challenges by training multilingual Transformer-based models with coverage prompts and rebalanced beam search, and further improving the quality by defining task-specific reward functions and training on them using reinforcement learning. Keywords: XF2T, text generation, cross-lingual, NLG evaluation, low resource NLG