Loading...
Loading...
Found 1,288 Skills
Auto-Claude Graphiti memory system configuration and usage. Use when setting up memory persistence, configuring LLM/embedding providers, querying knowledge graph, or optimizing memory performance.
Transition from static LLM chats to autonomous agents that execute multi-step tasks. Use this when you need to automate cross-platform reports (e.g., Snowflake to Google Docs), build self-service tools for non-technical teams, or create "anticipatory" engineering workflows that draft PRs based on Slack discussions.
Opik observability for LLM agents — Agent Configuration, Local Runner (opik connect), Evaluation Suites, threads, integrations. Use for "configure my agent", "connect my agent", "evaluate my agent" or "integrate with Opik".
Guide pour la création de serveurs MCP (Model Context Protocol) de qualité permettant aux LLM d'interagir avec des services externes via des outils bien conçus. À utiliser pour construire des serveurs MCP intégrant des API ou services externes, en Python (FastMCP) ou Node/TypeScript (MCP SDK).
Design Pydantic models and LLM prompt templates for structured extraction pipelines. Use when creating, editing, or reviewing Pydantic models that serve as LLM output schemas, or when writing prompt templates that pair with those models. Trigger: "pydantic model", "structured output", "extraction schema", "LLM output model", "schema design".
AI-powered crypto trading agent, wallet API, and LLM gateway via natural language. Use when the user wants to trade crypto, check portfolio balances (with PnL and NFTs), view token prices, search tokens, transfer crypto, manage NFTs, use leverage, bet on Polymarket, deploy tokens, set up automated trading, sign and submit raw transactions, or access LLM models through the Bankr LLM gateway funded by your Bankr wallet. Supports Base, Ethereum, Polygon, Solana, and Unichain.
A prompt repetition technique for improving LLM accuracy. Achieves significant performance gains in 67% (47/70) of 70 benchmarks. Automatically applied on lightweight models (haiku, flash, mini).
Instrument, trace, evaluate, and monitor LLM applications and AI agents with LangSmith. Use when setting up observability for LLM pipelines, running offline or online evaluations, managing prompts in the Prompt Hub, creating datasets for regression testing, or deploying agent servers. Triggers on: langsmith, langchain tracing, llm tracing, llm observability, llm evaluation, trace llm calls, @traceable, wrap_openai, langsmith evaluate, langsmith dataset, langsmith feedback, langsmith prompt hub, langsmith project, llm monitoring, llm debugging, llm quality, openevals, langsmith cli, langsmith experiment, annotate llm, llm judge.
Push the LLM to reconsider, refine, and improve its recent output. Use when user asks for deeper critique or mentions a known deeper critique method, e.g. socratic, first principles, pre-mortem, red team.
Transform code, issues, or context into a detailed prompt/context for another LLM to fix or implement. Use when preparing comprehensive context for external LLM assistance, bug fixes, improvements, or feature implementations. Provides detailed context without implementation suggestions, letting the receiving LLM decide how to implement solutions. Focuses on "what" (problem, requirements, current state) not "how" (implementation approach).
Generate README documentation writing plans and tasks. Use when the user wants to create README files for packages, plan documentation writing, or generate doc tasks for manual or LLM authoring.
Run and interact with KarpathyTalk, an open markdown-based developer social network with GitHub auth, SQLite, and an LLM-friendly JSON/markdown API.