All articles
VimeoRead originalJanuary 16, 2026

Building AI-Powered Subtitles at Vimeo

10 min readIntermediateAI/MLSystem DesignLLMsNLPPipeline Architecture

Vimeo's AI subtitle system uses LLMs to translate video subtitles across 9+ languages. The core challenge: LLMs optimize for fluency and merge fragmented speech into clean sentences, breaking subtitle timing sync. Their fix is a three-phase "split-brain" pipeline that separates creative translation from structural line mapping, with a self-healing fallback chain that guarantees 100% of subtitle slots are filled.

TL;DR

  • LLMs produce fluent translations but break subtitle sync by merging lines — a "blank screen bug" where subtitles disappear mid-speech because the model consolidated multiple time slots into one.
  • Different languages have fundamentally different "geometries": German verb brackets, Hindi SOV reversal, and Japanese compression each break line-by-line sync in unique ways.
  • The fix is a three-phase split-brain pipeline: (1) smart chunking into 3-5 line thought blocks, (2) unconstrained creative translation, (3) a separate LLM call for structural line mapping.
  • ~95% of chunks map perfectly on first pass. The remaining 5% trigger a correction loop (fixes ~32%), then LLM fallback, then a deterministic rule-based splitter — guaranteeing 100% slot fill rate.
  • Total overhead: 4-8% more processing time and 6-10% more tokens vs. single-pass translation. Tradeoff: zero blank screens and ~20 hours of manual QA saved per 1,000 videos.

Why this matters for interviews

This post is a masterclass in designing ML pipelines that handle unpredictable model outputs gracefully. It covers task decomposition (splitting competing objectives into separate passes), graceful degradation (multi-tier fallback chains), and the real cost of "intelligence" in production systems. These are exactly the patterns interviewers probe when asking you to design any AI-powered feature.

Breakdown

1.The blank screen bug: when fluency breaks the product

Vimeo's subtitle system expects every source line to have a corresponding translated line — if the speaker talks from 0:05 to 0:08, the system needs text in that specific slot. But LLMs are designed to be smart: when they detect filler words ("um", "you know") or fragmented speech, they merge everything into one clean, grammatically correct sentence. The translation is linguistically excellent, but now three subtitle time slots are empty. The viewer sees subtitles disappear mid-speech — a "blank screen bug." This isn't a translation failure; it's a competence problem. The model is too good at its job.

The blank screen bug — the LLM merged both subtitle lines into one, leaving the second time slot empty while the speaker keeps talking
The blank screen bug in action: the LLM consolidated fragmented speech into one fluent line, leaving the next time slot empty. The translation is correct — the product experience is broken.

Interview angle: This is a perfect example of the gap between "model quality" and "product quality." In interviews, when designing AI features, always ask: what structural constraints does the downstream system impose? A model that scores perfectly on BLEU might still break the product. This applies broadly — code generation that ignores line limits, summarization that drops required fields, etc.

2.The geometry of language: why sync is harder than it looks

Different languages don't just use different words — they organize thought in fundamentally different structures that break linear time-syncing. Vimeo identified three patterns: (1) German verb brackets (Satzklammer) place the main verb at the end of the clause, so splitting mid-sentence creates grammatically incomplete fragments the LLM resists producing. (2) Hindi SOV order reverses English structure — the end of the English thought maps to the middle of the Hindi sentence, breaking timestamp alignment. (3) Japanese compression — four lines of English filler genuinely condense into one tight Japanese sentence, leaving three empty slots. Romance languages like Spanish and Italian had ~99% clean pass rates, while Hindi had a 28.1% mismatch rate — 3-5x worse than Romance languages.

The German verb bracket — the infinitive "auszustatten" sits at the end of the clause, creating a bracket that holds the entire thought together and makes mid-sentence splitting impossible
German verb brackets place the main verb at the end. Splitting at the English line boundary leaves the first German subtitle grammatically incomplete — the LLM resists producing this.

Interview angle: This is a data distribution problem in disguise. When an interviewer asks "how would you handle internationalization in an ML system?", the strong answer acknowledges that languages have different structural properties that affect pipeline behavior unevenly. You need per-language error budgets and testing, not uniform assumptions. The 28.1% vs 1% mismatch rate shows why.

3.The split-brain architecture: separating creativity from structure

Vimeo abandoned single-prompt translation after finding that imposing format constraints on LLMs measurably degrades reasoning quality. Their solution splits the work into three phases. Phase 1 (Smart Chunking): a chunking algorithm scans for sentence boundaries across 10+ punctuation systems and groups lines into 3-5 line thought blocks — enough context to avoid hallucination but small enough to prevent drift. Phase 2 (Creative Translation): the LLM translates each chunk for meaning only — no line count enforcement. It handles German verb brackets naturally, reorders Hindi syntax correctly, and consolidates Japanese efficiently. Phase 3 (Line Mapping): a second LLM call performs purely structural work — "here are the original four English lines with timestamps, here is the translated block, break it back into four lines matching the source rhythm." By decoupling these concerns, each pass does its job without compromise.

The split-brain architecture flowchart showing three phases (Smart Chunking, Creative Translation, Line Mapping) with correction loops and fallback paths
The full pipeline: Phase 1 chunks input, Phase 2 translates for fluency, Phase 3 maps back to source line count. ~95% succeed on first pass; the rest enter correction loops and deterministic fallback.

Interview angle: This is the single-responsibility principle applied to ML pipelines. When an interviewer asks you to design a system with competing objectives, the answer is almost never "one model that does both." Task decomposition — splitting creative from structural work — is a production ML pattern that appears everywhere: retrieval vs. ranking in search, generation vs. filtering in content moderation, extraction vs. formatting in document processing.

4.The self-healing fallback chain

~95% of chunks map perfectly on the first pass. For the ~5% that fail, the system enters a tiered fallback chain. Tier 1 — LLM Correction Loop: retries with explicit feedback about what went wrong ("you returned 1 line, we need 2"). The model often finds a valid synonym or slightly less natural phrasing. This resolves ~32% of failures, adding ~2-3 seconds and ~25% more tokens per retry. Tier 2 — LLM Simplified Fallback: strips all semantic instructions. The prompt becomes "split this text into exactly N lines" — no translation quality checks, no clause boundary awareness. Tier 3 — Deterministic Rule-Based Splitter: when LLMs still fail, a pure algorithm fills empty slots with the last valid content, pads insufficient lines by duplication, and truncates excess lines. Zero tokens, instant execution. Result: 100% of chunks reach the user in a valid state.

Interview angle: This is the canonical graceful degradation pattern. Interviewers love asking "what happens when your model fails?" The weak answer is "retry." The strong answer is a multi-tier fallback with progressively simpler strategies: first retry with feedback, then simplify the task, then fall back to deterministic rules. Each tier trades quality for reliability. This exact pattern applies to any AI feature — chatbots, recommendations, content generation.

5.Cost analysis: the infrastructure tax of intelligence

The multi-pass architecture adds 4-8% processing time and 6-10% token overhead versus single-call translation. Broken down: ~95% of chunks incur zero additional cost (first-pass success). The LLM correction loop adds ~2-3 seconds and ~25% more tokens per retry. The deterministic fallback (3-4% of chunks) is instant and free. The tradeoff: zero blank-screen bugs and approximately 20 hours of manual QA eliminated per 1,000 videos. Vimeo frames this as "the infrastructure tax of intelligence" — a dumb word-for-word translator never breaks sync, but a smart translator that merges, reorders, and condenses requires smart infrastructure around it to absorb the creative unpredictability.

Interview angle: Cost-benefit analysis of ML systems is a common interview topic. Know how to quantify the tradeoff: what does the extra infrastructure cost (tokens, latency, engineering time) vs. what does it save (manual QA hours, user experience). The "infrastructure tax" framing is powerful — smarter models don't reduce system complexity, they shift it from the model layer to the orchestration layer.

6.Production results across 9 languages

Running the architecture against thousands of chunks across nine languages confirmed the design. Romance languages (Spanish, Italian) achieved ~99% clean pass rate on the line mapping phase — their structure is close enough to English that mismatches are rare. German and Hindi were the hardest: Hindi's 28.1% mismatch rate meant heavy reliance on the correction loop and fallback chain. Japanese compression was frequent but more predictable and easier for the line mapper to handle. The key insight: error handling budget should be proportional to linguistic distance from the source language, not uniform across all targets.

Interview angle: When designing multi-language ML systems, interviewers expect you to recognize that performance varies by language pair. The strong answer includes per-language monitoring, differentiated SLAs, and proportional error handling investment. Mentioning specific metrics (99% vs 28.1%) shows you think quantitatively about language-specific behavior, not just in generalities.

Key Concepts

Split-Brain Architecture

A pipeline design that separates competing objectives into independent passes. In Vimeo's case: creative translation (optimizing for fluency) and structural mapping (optimizing for line count) run as separate LLM calls rather than a single prompt trying to do both.

Analogy: Like having a novelist write the story and then a typesetter format it for the page — asking one person to write beautifully while counting characters per line produces worse results on both dimensions.

Graceful Degradation (Fallback Chain)

A multi-tier error handling strategy where each tier trades quality for reliability. Tier 1: retry with feedback. Tier 2: simplify the task. Tier 3: deterministic algorithm. Guarantees the system always produces output, even if imperfect.

Analogy: Like a restaurant: if the chef can't make the dish, the sous chef tries a simpler version. If that fails, you get a pre-made meal from the freezer. You always eat — quality varies, but no one goes hungry.

Smart Chunking

Grouping input text into logical thought blocks (typically 3-5 lines) by scanning for sentence boundaries across multiple punctuation systems. Prevents hallucination from processing too much context while avoiding the loss of meaning from processing too little.

Line Mapping

The structural pass that takes a fluent translated block and redistributes it back into the same number of lines as the source, preserving timestamp alignment. Purely structural — no concern for meaning, only line count.

Infrastructure Tax

The additional system complexity required when using intelligent (but unpredictable) models in production. Smarter models produce better output but require more sophisticated orchestration to handle their creative unpredictability — the cost shifts from the model layer to the infrastructure layer.

Analogy: Like hiring a brilliant but unreliable artist vs. a predictable printer. The artist produces better work, but you need a project manager, revision cycles, and backup plans — the "tax" on your operations.

Language Geometry

The structural properties of a language that affect how it maps to a timeline — word order (SOV vs SVO), verb placement (German brackets), and information density (Japanese compression). These geometric differences cause systematic, predictable mismatches in subtitle sync.

Task Decomposition in ML Pipelines

Breaking a complex task with competing objectives into separate, focused steps. Each step optimizes for one dimension without compromising another. Common pattern in production ML: retrieval/ranking in search, generation/filtering in moderation, translation/formatting in subtitles.

Knowledge Check

10 questions — your answers are saved locally so you can come back anytime.

Free to start

Ready to ace your system design interview?

This article is just one piece. SWEQuiz gives you structured, interview-focused practice across every topic that comes up in senior engineering rounds.

  • 1,000+ quiz questions across system design, JS, and ML/AI
  • Spaced repetition to lock in what you learn
  • Full case study walkthroughs of real interview topics
  • Track streaks, XP, and progress over time
SWE Quiz - Master System Design & ML Interviews