Loading...
Loading...
Cloud GPU processing via RunPod serverless. Use when setting up RunPod endpoints, deploying Docker images, managing GPU resources, troubleshooting endpoint issues, or understanding costs. Covers all 5 toolkit images (qwen-edit, realesrgan, propainter, sadtalker, qwen3-tts).
npx skill4agent add digitalsamba/claude-code-video-toolkit runpod# 1. Create account at https://runpod.io
# 2. Add API key to .env
echo "RUNPOD_API_KEY=your_key_here" >> .env
# 3. Deploy any tool with --setup
python tools/image_edit.py --setup
python tools/upscale.py --setup
python tools/dewatermark.py --setup
python tools/sadtalker.py --setup
python tools/qwen3_tts.py --setup--setup.envRUNPOD_QWEN_EDIT_ENDPOINT_ID| Tool | Docker Image | GPU | VRAM | Typical Cost |
|---|---|---|---|---|
| image_edit | | A6000/L40S | 48GB+ | ~$0.05-0.15/job |
| upscale | | RTX 3090/4090 | 24GB | ~$0.01-0.05/job |
| dewatermark | | RTX 3090/4090 | 24GB | ~$0.05-0.30/job |
| sadtalker | | RTX 4090 | 24GB | ~$0.05-0.15/job |
| qwen3_tts | | ADA 24GB | 24GB | ~$0.01-0.05/job |
Local CLI → Upload input to cloud storage → RunPod API → Poll for result → Download outputR2_ACCOUNT_IDR2_ACCESS_KEY_IDR2_SECRET_ACCESS_KEYR2_BUCKET_NAME/run/status/{job_id}workersMin: 0 — Scale to zero when idle (no cost)
workersMax: 1 — Max concurrent jobs (increase for throughput)
idleTimeout: 5 — Seconds before worker scales downworkersMax.env| Tool | Env Var |
|---|---|
| image_edit | |
| upscale | |
| dewatermark | |
| sadtalker | |
| qwen3_tts | |
workersMax=0imageName@sha256:DIGEST:latestworkersMin: 1--resize-ratio--stepsworkersMax=0docker/runpod-*/runpod/pytorchdocker buildx build --platform linux/amd64 -t ghcr.io/conalmullan/video-toolkit-<name>:latest docker/runpod-<name>/
docker push ghcr.io/conalmullan/video-toolkit-<name>:latestworkersMin: 0workersMax=0