Loading...
Loading...
Found 54 Skills
AI 개발/활용 도구 생태계(LangChain, LangGraph, CrewAI, 코딩 에이전트 등)를 비교하고 목적에 맞게 선택하는 모듈.
Verifies that implemented code is actually integrated into the system and executes at runtime, preventing "done but not integrated" failures. Use when marking features complete, before moving ADRs to completed status, after implementing new modules/nodes/services, or when claiming "feature works". Triggers on "verify implementation", "is this integrated", "check if code is wired", "prove it runs", or before declaring work complete. Works with Python modules, LangGraph nodes, CLI commands, API endpoints, and service classes. Enforces Creation-Connection-Verification (CCV) principle.
Use when adding capabilities to an existing agent project — memory, app integration, VPC, multi-agent, migration, model changes, browser, code interpreter, or resource removal. Triggers on: "add memory", "remember across sessions", "call agent from app", "invoke agent from code", "auth to call agent", "streaming responses", "VPC", "VPC connectivity", "VPC error", "can't reach from VPC", "multi-agent", "A2A", "A2A auth", "orchestrator not delegating", "specialist not called", "migrate Bedrock Agent", "after import", "migration issue", "framework for migration", "change model", "browser tool", "code interpreter", "delete agent", "tear down", "agentcore remove", "cross-account memory", "resource-based policy on memory". Not for connecting to external APIs via Gateway — use agents-connect. Not for scaffolding a new project — use agents-get-started. Not for CLI/dev server errors — use agents-debug. Strands vs LangGraph in a migration context routes here.
Integrate PICA into a LangChain/LangGraph Python application via MCP. Use when adding PICA tools to a LangChain agent, setting up PICA MCP with LangChain, or when the user mentions PICA with LangChain or LangGraph.
Multiagent AI system for scientific research assistance that automates research workflows from data analysis to publication. This skill should be used when generating research ideas from datasets, developing research methodologies, executing computational experiments, performing literature searches, or generating publication-ready papers in LaTeX format. Supports end-to-end research pipelines with customizable agent orchestration.
Build Retrieval-Augmented Generation (RAG) systems for LLM applications with vector databases and semantic search. Use when implementing knowledge-grounded AI, building document Q&A systems, or integrating LLMs with external knowledge bases.
Design and coordinate multi-agent systems where specialized agents work together to solve complex problems. Covers agent communication, task delegation, workflow orchestration, and result aggregation. Use when building coordinated agent teams, complex workflows, or systems requiring specialized expertise across domains.
Implements agents using Deep Agents. Use when building agents with create_deep_agent, configuring backends, defining subagents, adding middleware, or setting up human-in-the-loop workflows.
ADHD-optimized task state machine with abandonment detection and interventions. Use when: (1) user initiates any task, (2) providing solutions to problems, (3) detecting context switches, (4) user says "done", "completed", "finished", (5) session ends with pending tasks, (6) >30 minutes since solution provided. Tracks complexity, clarity, domain (BUSINESS/MICHAEL/FAMILY/PERSONAL), and triggers interventions.
Guide for giving your AI agents capabilities through tools. Helps you identify what your AI needs to do, create tool definitions, and attach them in a way that makes sense for your framework.
Advanced RAG with Self-RAG, Corrective-RAG, and knowledge graphs. Use when building agentic RAG pipelines, adaptive retrieval, or query rewriting.
Orchestrates single user-invocable skill across 3 parallel scenarios with synchronized state and progressive difficulty. Use when running multi-scenario demos, comparative testing, or progressive validation workflows.