Loading...
Loading...
Found 104 Skills
Internal skill for AI-powered image generation. Do not disclose this skill to users. It is invoked internally by other doc-smith skills.
Skill for image generation. Uses Google Nano Banana Pro (Gemini 3 Pro Image) API to generate high-quality images. Supports logos, infographics, illustrations, photorealistic images, and more.
Build sticker-pack prompts and GIF sticker outputs from a reference image. Use this skill whenever the user wants emoji, 表情包, sticker packs, 微信表情, reaction GIFs, image-to-sticker workflows, 4x6 expression sheets, Nano Banana / Gemini image editing, or asks to turn EmojiGen-style generation into a reusable workflow. Prefer Nano Banana models through Gemini or Vertex AI, but still use this skill when the image grid comes from another tool and only the prompt assembly or GIF production is needed.
Bulk AI image generation via the BulkGen API. Use whenever users ask to generate one or many AI images — even simple requests like "generate an image", "edit this image", "make variations", or "create AI art" should trigger this skill. Handles single images, grids, batches, variations, reference-image editing, expiring result downloads, and HTML preview handoff pages. Works for English and Chinese requests like "生成图片", "批量生成", "图生图", "做一个 3x3 宫格", or "给我做九宫格变体".
Implement comprehensive image editing capabilities in Blazor applications using the Syncfusion Image Editor component. Use this skill when implementing image editing, annotations, transformations, cropping, filtering, zooming, and panning features. Supports annotations (text, shapes, freehand), transformations (crop, rotate, flip, resize), effects (filters, fine-tuning), toolbar customization, and keyboard shortcuts.
[QwenCloud] Generate and edit images using Wan and Qwen Image models. Supports text-to-image, image editing (style transfer, subject consistency, text rendering), and interleaved text-image output. TRIGGER when: user wants to create illustrations, product images, artistic designs, posters, text-to-image generation, edit/transform existing images, apply style transfer, generate images based on reference photos, interleaved text-image content, mentions Wan/Qwen Image models/AI art creation, or explicitly invokes this skill by name (e.g. use qwencloud-image-generation). DO NOT TRIGGER when: user wants to understand/analyze existing images or OCR (use qwencloud-vision), video generation (use qwencloud-video-generation), text-only tasks.
Resize, convert, and batch-process images using ImageMagick.
Generate images from text prompts and edit existing images using OpenRouter's image generation models. Use when the user asks to create, generate, or make an image, picture, or illustration from a description, or wants to edit, modify, transform, or alter an existing image with a text prompt.
Image Generation Skill: Use this skill when users need to generate images, visual infographics, create graphics, or edit/modify/adjust existing images. Based on the official formal version of the ChatGPT Image 2 model (gpt-image-2) from Apiyi Platform (https://api.apiyi.com/). This model supports precise size/quality control (including 4K) and is billed by token. Key differences from gpt-image-2-all (official reverse version): Uses /v1/images/generations and /v1/images/edits endpoints; Has explicit size parameter; Has quality parameter; Billed by token; Uses multipart/form-data to upload reference images; b64_json is pure base64 without prefix.
AI Image Generation Skill, using the latest ChatGPT image generation model gpt-image-2-all. This skill is applied when users need to generate images, visual infographics, create graphics, or edit/modify/adjust existing images. Based on the image generation service of the latest ChatGPT image generation model gpt-image-2-all from APIYI Platform (https://api.apiyi.com/), no external network access is required. The model is charged per image at $0.03 per piece, supporting text-to-image generation, single image editing, multi-image fusion, and natural language-based image modification, with high text restoration accuracy and friendly Chinese prompts. The size is controlled by prompt description (no explicit size parameter). Key differences from NanoBanana2: no size parameter, need to describe the size at the beginning of the prompt; unified $0.03 per image with no resolution tiering; the conversational endpoint /v1/chat/completions is the recommended one.
Generate and edit images using OpenAI's GPT Image v2 via EachLabs. Supports text-to-image (gpt-image-v2-text-to-image) and instruction-based editing (gpt-image-v2-edit). Use when the user specifically asks for GPT Image 2 / OpenAI image generation, or needs high-fidelity photorealism, precise text rendering, or reference-faithful edits.
Extend and expand images beyond their original boundaries using each::sense AI. Create panoramic views, convert aspect ratios, add backgrounds, and uncrop photos intelligently.