Total 44,222 skills, AI & Machine Learning has 7033 skills
Showing 12 of 7033 skills
Generate 3D models, textures, images, rig characters, animate them, and prepare for 3D printing using the Meshy AI API. Handles API key detection, task creation, polling, downloading, and full 3D print pipeline with slicer integration. Use when the user asks to create 3D models, convert text/images to 3D, texture models, rig or animate characters, 3D print a model, or interact with the Meshy API.
Use when adding CopilotKit to an existing project or bootstrapping a new CopilotKit project from scratch. Covers framework detection, package installation, runtime wiring, provider setup, and first working chat integration.
Use when wiring an external agent framework (LangGraph, CrewAI, PydanticAI, Mastra, ADK, LlamaIndex, Agno, Strands, Microsoft Agent Framework, or others) into a CopilotKit application via the AG-UI protocol.
Use when building custom agent backends, implementing the AG-UI protocol, debugging streaming issues, or understanding how agents communicate with frontends. Covers event types, SSE transport, AbstractAgent/HttpAgent patterns, state synchronization, tool calls, and human-in-the-loop flows.
Consult & operate against the Hyperliquid Names API. Use when your agent needs to resolve `.hl` names, reverse-resolve addresses, fetch HLN profiles or records, inspect owner or list queries, diagnose HLN API failures, prepare a mint-pass request, or guide HyperEVM dApp integration with HL Names.
Writes or reviews lyrics with professional prosody, rhyme craft, and quality checks. Use when writing new lyrics, revising existing lyrics, or when the user says 'let's work on a track.'
NLP Cloud integration. Manage data, records, and automate workflows. Use when the user wants to interact with NLP Cloud data.
ElevenLabs integration. Manage data, records, and automate workflows. Use when the user wants to interact with ElevenLabs data.
Vida's thinking framework and expression style. Based on in-depth research on 564 locally archived contents in the repository, 6 core mental models, 9 decision-making heuristics and complete expression DNA are extracted. Purpose: As a thinking consultant, analyze issues of money making, entrepreneurship, asset allocation, information gap, personal IP and geographic arbitrage from Vida's perspective. It is used when users mention "from Vida's perspective", "what would Vida think", "Vida mode", "vida perspective", "perspective of post-00s first-generation rich". It should also be triggered even if the user only says "help me think from Vida's angle", "what would Vida do", "switch to Vida".
Publish and deploy C# MCP servers. Covers NuGet packaging for stdio servers, Docker containerization for HTTP servers, Azure Container Apps and App Service deployment, and publishing to the official MCP Registry. USE FOR: packaging stdio MCP servers as NuGet tools, creating Dockerfiles for HTTP MCP servers, deploying to Azure Container Apps or App Service, publishing to the MCP Registry at registry.modelcontextprotocol.io, configuring server.json for MCP package metadata, setting up CI/CD for MCP server publishing. DO NOT USE FOR: publishing general NuGet libraries (not MCP-specific), general Docker guidance unrelated to MCP, creating new servers (use mcp-csharp-create), debugging (use mcp-csharp-debug), writing tests (use mcp-csharp-test).
Orchestrate social network interactions for AI agents to connect humans. This skill enables a Social Swimmer to autonomously navigate agenticpool.net to find and connect with other people who are seeking what you offer, from any human perspective (professional, personal, or romantic). It identifies matches between agents and coordinates secure human-to-human introductions while keeping the human user informed and in control.
Launch Loki Mode autonomous SDLC agent. Handles PRD-to-deployment with minimal human intervention. Invoke for multi-phase development tasks, bug fixing campaigns, or full product builds.