Machine Translation
The use of AI to automatically translate text or speech from one language to another. Modern neural machine translation uses transformer models and achieves near-human quality for many language pairs.
Why It Matters
Machine translation breaks language barriers at scale. It enables global commerce, diplomacy, and communication that would be impossible with human translators alone.
Example
Google Translate processing 100 billion words per day across 133 languages, enabling travelers, businesses, and researchers to communicate across language barriers.
Think of it like...
Like having a universal interpreter in your pocket — not perfect, but good enough to understand and be understood in most situations worldwide.
Related Terms
Natural Language Processing
The branch of AI that deals with the interaction between computers and human language. NLP enables machines to read, understand, generate, and make sense of human language in a useful way.
Transformer
A neural network architecture introduced in 2017 that uses self-attention mechanisms to process sequential data in parallel rather than sequentially. Transformers are the foundation of modern LLMs like GPT, Claude, and Gemini.
Sequence-to-Sequence
A model architecture that transforms one sequence into another, where the input and output can be different lengths. It uses an encoder to process input and a decoder to generate output.
Multilingual AI
AI models capable of understanding and generating text in multiple languages. Modern LLMs often support 50-100+ languages, though performance varies significantly across languages.
Encoder-Decoder
An architecture where the encoder compresses input into a fixed representation and the decoder generates output from that representation. This structure is used in translation, summarization, and image captioning.