Loading...
Loading...
Found 24 Skills
Analyze images using AI with the understand_image tool
Analyze text and images for harmful content using Azure AI Content Safety (@azure-rest/ai-content-safety). Use when moderating user-generated content, detecting hate speech, violence, sexual conten...
Analyze images using AI — segment objects, detect objects, extract text (OCR), describe images, ask questions about images. Use when the user requests "Segment image", "Detect objects", "OCR", "Extract text from image", "Describe image", "What's in this image", "Image analysis".
Analyze, describe, and extract information from images using the MiniMax vision MCP tool. Use when: user shares an image file path or URL (any message containing .jpg, .jpeg, .png, .gif, .webp, .bmp, or .svg file extension) or uses any of these words/phrases near an image: "analyze", "analyse", "describe", "explain", "understand", "look at", "review", "extract text", "OCR", "what is in", "what's in", "read this image", "see this image", "tell me about", "explain this", "interpret this", in connection with an image, screenshot, diagram, chart, mockup, wireframe, or photo. Also triggers for: UI mockup review, wireframe analysis, design critique, data extraction from charts, object detection, person/animal/activity identification. Triggers: any message with an image file extension (jpg, jpeg, png, gif, webp, bmp, svg), or any request to analyze/describ/understand/review/extract text from an image, screenshot, diagram, chart, photo, mockup, or wireframe.
Query Google Gemini 3 Pro via grsai.com API for text generation and image analysis. Use for text generation, Q&A, summarization, code generation, creative writing, image analysis/vision, complex reasoning, and structured document generation. Triggers on "ask gemini", "use gemini", "query gemini", "analyze this image with gemini", or when a second opinion from another LLM is needed. Optionally accepts an image input for vision tasks.
Analyze images using OpenAI's Vision API. Use bash command to execute the vision script like 'bash <base_dir>/scripts/vision.sh <image> <question>'. Can understand image content, objects, text, colors, and answer questions about images.
Local vision-language model for image analysis using SmolVLM-2B
Perform image understanding and analysis using GLM Vision MCP. Trigger conditions: (1) Users request image analysis, image understanding, or description of image content (2) Need to identify objects, text, or scenes in images (3) Use GLM's visual understanding capabilities
Use when visual reasoning is needed with Alibaba Cloud Model Studio QVQ models, including step-by-step image reasoning, chart analysis, and visually grounded problem solving.
Image analysis and manipulation. Use when: user wants to analyze images, extract metadata, convert formats, resize, or get image information.
Azure AI Vision integration. Manage data, records, and automate workflows. Use when the user wants to interact with Azure AI Vision data.
Audit, plan, and safely optimize Shopify image alt text for product media, collection featured images, article featured images, and article inline images. Use when a merchant wants an AI agent to scan Shopify images, test whether the active AI model can inspect images, generate concise alt text with multimodal image understanding when available or context-only fallback when it is not, review the proposed changes in batches, and apply approved Shopify Admin updates.