Loading...
Loading...
Found 85 Skills
Atlas Cloud API integration skill — quickly call 300+ AI image generation, video generation, and LLM models through a unified API. Use this skill when the user needs to integrate AI image generation (e.g., Flux, Seedream, DALL-E), AI video generation (e.g., Kling, Sora, Seedance), or call LLM APIs (OpenAI-compatible format) into their project. Applicable scenarios include: generating images, generating videos, calling large language models, using Atlas Cloud API, configuring ATLASCLOUD_API_KEY, querying available model lists, searching models by keyword, uploading local images/media files, one-step quick generation, image-to-video, text-to-image, text-to-video, AI content creation tool integration. Even if the user doesn't explicitly mention Atlas Cloud, this skill should be considered whenever AI media generation API integration development is involved.
Run agency-orchestrator YAML workflows directly in Claude Code / OpenClaw / Cursor — no API key required, using the current session's LLM as the execution engine. Triggered when the user provides a .yaml workflow file or requests multi-role collaboration to complete a task.
Grafana Cloud AI and ML features — Grafana Assistant (natural language queries, dashboard generation, incident investigations), Dynamic Alerting (ML forecasting and outlier detection), Sift (automated root cause analysis with 8 analysis types), Knowledge Graph (entity discovery and RCA Workbench), and the LLM Plugin (OpenAI/Anthropic/Azure integration). Use when setting up AI-powered alerting, using natural language to query metrics/logs, automating incident investigation, or integrating LLMs with Grafana panels and workflows.
Format prompts for different LLM providers with chat templates and HNSW-powered context retrieval
AI Agent long-term memory system with cross-session, cross-project persistence. Triggers: - /remember - Store memories - /recall - Search memories - /forget - Delete/archive memories - /memory-status - Check status - When needing to persist important conversation insights - When sharing user preferences across projects
Architecting real-time Voice AI agents.
Expert guide for configuring, customizing, and creatively leveraging OpenClaw — the self-hosted AI gateway that connects LLMs to messaging channels (Telegram, WhatsApp, Discord, Slack, iMessage, etc.). Use when the user wants to: (1) Set up or modify their openclaw.json configuration, (2) Write or edit bootstrap files (SOUL.md, USER.md, AGENTS.md, IDENTITY.md, TOOLS.md), (3) Configure messaging channels, (4) Set up models and providers, (5) Create multi-agent routing, (6) Build skills, hooks, or cron jobs, (7) Troubleshoot OpenClaw issues, (8) Get creative ideas for leveraging OpenClaw in non-obvious ways. Triggers on: openclaw, gateway, SOUL.md, USER.md, AGENTS.md, IDENTITY.md, channels setup, agent routing, heartbeat, cron jobs, openclaw hooks, openclaw skills, openclaw config, openclaw.json, personal assistant setup.
企业微信客服自动化系统。自动同意好友添加、基于知识库的智能问答、未知问题人工介入提醒。适用于企业微信客服场景的 AI 助手机器人。
Use when working with Anthropic Claude Agent SDK. Provides architecture guidance, implementation patterns, best practices, and common pitfalls.
Guide pour la création de serveurs MCP (Model Context Protocol) de qualité permettant aux LLM d'interagir avec des services externes via des outils bien conçus. À utiliser pour construire des serveurs MCP intégrant des API ou services externes, en Python (FastMCP) ou Node/TypeScript (MCP SDK).
Expert skill for building AI systems with Weft, a Rust-based programming language where LLMs, humans, APIs, and infrastructure are first-class primitives with typed connections and durable execution.
Guide for creating high-quality MCP (Model Context Protocol) servers that enable LLMs to interact with external services through well-designed tools. Use when building MCP servers to integrate external APIs or services, whether in Python (FastMCP) or Node/TypeScript (MCP SDK).