SlatorPod artwork

#218 How Large Language Models Replace Neural Machine Translation with Unbabel’s João Graca

SlatorPod

English - July 12, 2024 09:00 - 44 minutes - 30.9 MB - ★★★★ - 5 ratings
Business translation localization machine translation language industry localization industry globalization language venture capital artificial intelligence machine learning Homepage Download Apple Podcasts Google Podcasts Overcast Castro Pocket Casts RSS feed


João Graça, Co-founder and CTO of language operations platform Unbabel, joins SlatorPod to talk about the present and future of large language models (LLMs) and their broad impact across all things translation and localization.

First, the CTO explains how Unbabel was founded to address language barriers for people using services like Airbnb, combining MT with human validation to improve translation quality.

João believes that LLMs are quickly replacing neural MT models as much more R&D is going into LLMs vs NMT. He highlights that LLMs can handle more complex tasks like automatic post-editing, source correction, and cultural adaptation, which were previously difficult to achieve with traditional models.

He also tells the backstory of the company's decision to develop TowerLLM. João shares how Unbabel's approach involves using open-source LLMs, fine-tuning them with multilingual data, and applying techniques like retrieval-augmented generation to improve translation quality in production settings.

Despite the advancements, João acknowledges that human intervention is still necessary for high-stakes translation tasks.

The podcast concludes with the hiring environment for AI talent and the future directions for LLM development, with João expressing optimism about the continued progress and potential of these models.