
Meet ALMA: A New Training Method That Boosts Translation Performance for Large Language Models
Researchers from Johns Hopkins and Microsoft propose a new 2-stage fine-tuning method that unlocks stronger translation abilities in smaller models with just 7-13 billion parameters.