sora
Original:🇺🇸 English
Not Translated
1 scriptsChecked / no sensitive code detected
Use when the user asks to generate, remix, poll, list, download, or delete Sora videos via OpenAI’s video API using the bundled CLI (`scripts/sora.py`), including requests like “generate AI video,” “Sora,” “video remix,” “download video/thumbnail/spritesheet,” and batch video generation; requires `OPENAI_API_KEY` and Sora API access.
1installs
Added on
NPX Install
npx skill4agent add moosegoose0701/skill-compose soraSKILL.md Content
Sora Video Generation Skill
Creates or manages short video clips for the current project (product demos, marketing spots, cinematic shots, UI mocks). Defaults to and a structured prompt augmentation workflow, and prefers the bundled CLI for deterministic runs. Note: is a skill tag in prompts, not a shell command.
sora-2$soraWhen to use
- Generate a new video clip from a prompt
- Remix an existing video by ID
- Poll status, list jobs, or download assets (video/thumbnail/spritesheet)
- Batch runs (many prompts or variants)
Decision tree (create vs remix vs status/download vs batch)
- If the user has a video id and wants a change → remix
- If the user has a video id and wants status or assets → status/poll/download
- If the user needs many prompts/assets → create-batch
- If the user asks for two versions with a small change (same shot, different subject/detail) → create the base, then remix for the variation
- Otherwise → create (or create-and-poll if they need a ready asset in one step)
Workflow
- Decide intent: create vs remix vs status/download vs batch.
- Collect inputs: prompt, model, size, seconds, and any input reference image.
- If batch: write a temporary JSONL under tmp/ (one job per line), run once, then delete the JSONL.
- Prefer CLI augmentation flags (,
--use-case,--scene, etc.) instead of pre-writing a structured prompt. If you already produced a structured prompt file, pass--camerato avoid double-wrapping.--no-augment - Run the bundled CLI () with sensible defaults (see references/cli.md). For long prompts, prefer
scripts/sora.pyto avoid shell-escaping issues; pair with--prompt-fileif the prompt is already structured.--no-augment - For async jobs, poll until complete (or use create-and-poll).
- Download assets (video/thumbnail/spritesheet) and save locally.
- Remove intermediate files created during invocation (for example ,
prompt.txt, temp JSONL). If the sandbox blocksremix_job.json, skip cleanup or truncate the files without surfacing an error.rm - Iterate with a single targeted change per prompt.
Authentication
- must be set for live API calls.
OPENAI_API_KEY
If the key is missing, give the user these steps:
- Create an API key in the OpenAI platform UI: https://platform.openai.com/api-keys
- Set as an environment variable in their system.
OPENAI_API_KEY - Offer to guide them through setting the environment variable for their OS/shell if needed.
- Never ask the user to paste the full key in chat. Ask them to set it locally and confirm when ready.
Defaults & rules
- Default model: (use
sora-2for higher fidelity).sora-2-pro - Default size: .
1280x720 - Default seconds: (allowed: "4", "8", "12" as strings).
4 - Always set size and seconds via API params; prose will not change them.
- Use the OpenAI Python SDK (package); do not use raw HTTP.
openai - Require before any live API call.
OPENAI_API_KEY - If uv cache permissions fail, set .
UV_CACHE_DIR=/tmp/uv-cache - Input reference images must be jpg/png/webp and should match target size.
- Download URLs expire after about 1 hour; copy assets to your own storage.
- Prefer the bundled CLI and never modify unless the user asks.
scripts/sora.py - Sora can generate audio; if a user requests voiceover/audio, specify it explicitly in the and
Audio:lines and keep it short.Dialogue:
API limitations
- Models are limited to and
sora-2.sora-2-pro - API access to Sora models requires an organization-verified account.
- Duration is limited to 4/8/12 seconds and must be set via the parameter.
seconds - The API expects as a string enum ("4", "8", "12").
seconds - Output sizes are limited by model (see for the supported sizes).
references/video-api.md - Video creation is async; you must poll for completion before downloading.
- Rate limits apply by usage tier (do not list specific limits).
- Content restrictions are enforced by the API (see Guardrails below).
Guardrails (must enforce)
- Only content suitable for audiences under 18.
- No copyrighted characters or copyrighted music.
- No real people (including public figures).
- Input images with human faces are rejected.
Prompt augmentation
Reformat prompts into a structured, production-oriented spec. Only make implicit details explicit; do not invent new creative requirements.
Template (include only relevant lines):
Use case: <where the clip will be used>
Primary request: <user's main prompt>
Scene/background: <location, time of day, atmosphere>
Subject: <main subject>
Action: <single clear action>
Camera: <shot type, angle, motion>
Lighting/mood: <lighting + mood>
Color palette: <3-5 color anchors>
Style/format: <film/animation/format cues>
Timing/beats: <counts or beats>
Audio: <ambient cue / music / voiceover if requested>
Text (verbatim): "<exact text>"
Dialogue:
<dialogue>
- Speaker: "Short line."
</dialogue>
Constraints: <must keep/must avoid>
Avoid: <negative constraints>Augmentation rules:
- Keep it short; add only details the user already implied or provided elsewhere.
- For remixes, explicitly list invariants ("same shot, change only X").
- If any critical detail is missing and blocks success, ask a question; otherwise proceed.
- If you pass a structured prompt file to the CLI, add to avoid the tool re-wrapping it.
--no-augment
Examples
Generation example (single shot)
Use case: product teaser
Primary request: a close-up of a matte black camera on a pedestal
Action: slow 30-degree orbit over 4 seconds
Camera: 85mm, shallow depth of field, gentle handheld drift
Lighting/mood: soft key light, subtle rim, premium studio feel
Constraints: no logos, no textRemix example (invariants)
Primary request: same shot and framing, switch palette to teal/sand/rust with warmer backlight
Constraints: keep the subject and camera move unchangedPrompting best practices (short list)
- One main action + one camera move per shot.
- Use counts or beats for timing ("two steps, pause, turn").
- Keep text short and the camera locked-off for UI or on-screen text.
- Add a brief avoid line when artifacts appear (flicker, jitter, fast motion).
- Shorter prompts are more creative; longer prompts are more controlled.
- Put dialogue in a dedicated block; keep lines short for 4-8s clips.
- State invariants explicitly for remixes (same shot, same camera move).
- Iterate with single-change follow-ups to preserve continuity.
Guidance by asset type
Use these modules when the request is for a specific artifact. They provide targeted templates and defaults.
- Cinematic shots:
references/cinematic-shots.md - Social ads:
references/social-ads.md
CLI + environment notes
- CLI commands + examples:
references/cli.md - API parameter quick reference:
references/video-api.md - Prompting guidance:
references/prompting.md - Sample prompts:
references/sample-prompts.md - Troubleshooting:
references/troubleshooting.md - Network/sandbox tips:
references/codex-network.md
Reference map
- : how to run create/poll/remix/download/batch via
references/cli.md.scripts/sora.py - : API-level knobs (models, sizes, duration, variants, status).
references/video-api.md - : prompt structure and iteration guidance.
references/prompting.md - : copy/paste prompt recipes (examples only; no extra theory).
references/sample-prompts.md - : templates for filmic shots.
references/cinematic-shots.md - : templates for short social ad beats.
references/social-ads.md - : common errors and fixes.
references/troubleshooting.md - : network/approval troubleshooting.
references/codex-network.md