remotion-production
Original:🇺🇸 English
Translated
Full video production workflow for Remotion projects. Teaches how to orchestrate MCP tools (TTS, music, SFX, stock footage, video analysis) into complete Remotion compositions. Use this skill whenever producing a video that needs audio, voiceovers, music, stock footage, or analyzing existing video files.
3installs
Added on
NPX Install
npx skill4agent add dojocodinglabs/remotion-superpowers remotion-productionTags
Translated version includes tags in frontmatterSKILL.md Content
View Translation Comparison →Remotion Production Workflow
This skill teaches how to produce complete videos with Remotion by orchestrating multiple MCP tools together. It covers the full pipeline from concept to rendered MP4.
Available MCP Tools
You have access to these MCP servers for media production:
remotion-media (via KIE)
- — Text-to-speech voiceovers (ElevenLabs TTS)
generate_tts - — Background music (Suno V3.5–V5)
generate_music - — Sound effects (ElevenLabs SFX V2)
generate_sfx - — AI images (Nano Banana Pro)
generate_image - — AI video clips (Veo 3.1)
generate_video - — Transcribe audio/video to SRT (Whisper)
generate_subtitles - — List all generated media in the project
list_assets
TwelveLabs (video understanding)
- Index and analyze video files
- Semantic search within videos ("find the part where...")
- Scene detection, object detection, speaker identification
- Video summarization
Pexels (stock footage)
- — Search free stock photos
searchPhotos - — Search free stock videos
searchVideos - /
getVideo— Get details by IDgetPhoto - — Download video to project
downloadVideo
ElevenLabs (optional — advanced voice)
- Voice cloning from audio samples
- Advanced TTS with custom voices
- Audio isolation and processing
- Transcription
Replicate (optional — 100+ AI models)
- — Run a model synchronously (images)
replicate_run - — Start async prediction (video)
replicate_create_prediction - — Poll prediction status
replicate_get_prediction - Image models: FLUX 1.1 Pro, Imagen 4, Ideogram v3, FLUX Kontext
- Video models: Wan 2.5 (T2V, I2V), Kling 2.6 Pro
Production Pipeline
Read individual rule files for detailed workflows:
- — End-to-end workflow from concept to final render
rules/production-pipeline.md - — How to integrate generated audio into Remotion compositions
rules/audio-integration.md - — Syncing TTS voiceovers with animations and captions
rules/voiceover-sync.md - — Generating and timing background music
rules/music-scoring.md - — Searching, downloading, and using stock footage in Remotion
rules/stock-footage-workflow.md - — Using TwelveLabs to analyze and select clips from existing footage
rules/video-analysis.md - — TikTok-style animated captions using @remotion/captions and Whisper
rules/captions-workflow.md - — Reusable animation patterns (fade, slide, scale, typewriter, stagger)
rules/animation-presets.md - — Three.js and React Three Fiber via @remotion/three
rules/3d-content.md - — Animated charts, dashboards, and number counters
rules/data-visualization.md - — Light leaks, Lottie, film grain, vignettes, Ken Burns
rules/visual-effects.md - — GitHub Actions workflows for automated video rendering
rules/ci-rendering.md - — Replicate MCP model catalog, usage, and decision guide
rules/replicate-models.md - — AI image prompt engineering, provider selection, Remotion integration
rules/image-generation.md - — AI video clip generation, I2V pipeline, sequencing in Remotion
rules/video-generation.md - — SFX generation, prompt engineering, timing to visual events
rules/sound-effects.md - — Voice cloning, custom TTS parameters, multi-voice scripts
rules/elevenlabs-advanced.md - — File organization, naming conventions, staticFile() reference
rules/asset-management.md
Key Principles
- Audio drives timing — Generate voiceover first, get its duration, then set composition length to match.
- Assets go in — All generated media files (audio, video, images) must be saved to the project's
public/directory so Remotion can access them viapublic/.staticFile() - Use Remotion's audio components — Always use component with
<Audio>for audio. Never use HTMLstaticFile()tags.<audio> - Frame-based timing — Remotion uses frames, not seconds. Convert with . At 30fps, 1 second = 30 frames.
fps * seconds - Progressive composition — Build the video in layers: visuals first, then voiceover, then music, then SFX.
- Preview frequently — Use to preview after each major change. The Remotion player updates live.
npm run dev