DeepL's Specialized AI: Data, Compute, and Translation Mastery
How DeepL Built a Translation Powerhouse with AI with CEO Jarek Kutylowski
Resources
Resources & Recommendations
Books
- "transformer" - Mentioned as a foundational architecture that quickly emerged and became relevant for translation models.
Tools & Software
- DGX - Generation of GPUs from Nvidia, mentioned in the context of substantial costs for training large-scale models.
People Mentioned
- Jarek Kutylowski (CEO of DeepL) - The guest on the podcast, discussing DeepL's translation powerhouse built with AI.
Organizations & Institutions
- DeepL - A company specializing in AI-powered translation, particularly for businesses.
- OpenAI - Mentioned as a competitor in the AI space, against whom DeepL aims to stay ahead.
- Meta - Mentioned in the context of publishing large language models like Llama.
- Nvidia - Manufacturer of DGX GPUs, crucial for DeepL's model training.
Websites & Online Resources
- Llama (Meta) - A large language model published by Meta, discussed in relation to DeepL's model training strategies.
Other Resources
- Neural Machine Translation - The technology that DeepL adopted in 2017, marking a significant shift in translation capabilities.
- Transformer architecture - A key development in neural machine translation, mentioned for its rapid emergence.
- Parallel corpora - Bilingual text data used for training translation models, discussed in the context of data acquisition.
- Monolingual corpora - Single-language text data, also important for training translation models, especially for languages with less bilingual data.
- Speech translation - A newer market for DeepL, introduced last year, offering real-time translation for spoken language.