Esta página se muestra en inglés. Una traducción al español está en curso.
ConfigsMay 4, 2026·2 min de lectura

KrillinAI — AI Video Translation and Dubbing in 100 Languages

An open-source tool that uses LLMs to translate and dub video content into over 100 languages with one-click deployment, optimized for YouTube, TikTok, and other platforms.

Introduction

KrillinAI is an open-source video translation and dubbing platform that leverages LLMs to produce natural-sounding translations for video content. It handles the entire pipeline from speech recognition through translation to voice synthesis, outputting platform-ready videos with synchronized subtitles and dubbed audio.

What KrillinAI Does

  • Extracts and transcribes audio from video files or URLs
  • Translates transcripts into 100+ target languages using LLMs
  • Generates dubbed audio with TTS matching the original timing
  • Burns translated subtitles into the output video
  • Exports in formats optimized for YouTube, TikTok, and Bilibili

Architecture Overview

KrillinAI orchestrates a multi-stage pipeline: FFmpeg extracts audio, a speech recognition model (Whisper or alternatives) produces timestamped transcripts, an LLM translates segments while preserving timing constraints, and a TTS engine synthesizes the dubbed audio. A Go backend coordinates these stages and serves a web UI for job management. Results are composited back into the video with synchronized subtitles.

Self-Hosting & Configuration

  • Deploy via Docker with a single command
  • Requires an LLM API key (OpenAI, or compatible providers)
  • Configure TTS provider (built-in or external like ElevenLabs)
  • Set source and target languages per job
  • GPU recommended for faster Whisper transcription but not required

Key Features

  • End-to-end pipeline from raw video to translated output
  • Supports 100+ language pairs for translation
  • Platform-aware output formats (vertical for TikTok, widescreen for YouTube)
  • Web UI for managing translation jobs and reviewing results
  • Batch processing for translating multiple videos

Comparison with Similar Tools

  • Rask.ai — cloud SaaS with subscription; KrillinAI is self-hosted and open source
  • HeyGen — avatar-focused video translation; KrillinAI preserves original video
  • Kapwing — general video editor with limited translation; KrillinAI is purpose-built
  • Whisper + manual workflow — requires manual orchestration; KrillinAI automates the full pipeline
  • VideoCaptioner — subtitle-focused; KrillinAI adds full audio dubbing

FAQ

Q: Does it support local LLMs? A: Yes. Any OpenAI-compatible API endpoint works, including Ollama and vLLM.

Q: How accurate is the translation? A: Quality depends on the LLM used. GPT-4 class models produce near-professional results for major languages.

Q: Can it handle long videos? A: Yes. It processes videos in segments and stitches the output together, handling videos of any length.

Q: Does it clone the original speaker voice? A: By default it uses standard TTS voices. Voice cloning can be configured with compatible TTS providers.

Sources

Discusión

Inicia sesión para unirte a la discusión.
Aún no hay comentarios. Sé el primero en compartir tus ideas.

Activos relacionados