Loading...
Loading...
Found 1,288 Skills
Integrates Flowlines observability SDK into Python LLM applications. Use when adding Flowlines telemetry, instrumenting LLM providers, or setting up OpenTelemetry-based LLM monitoring.
Access real-time, continuously refreshed investment context through the Primary Logic External API under /v1. Use when asked to power Codex, Claude Code, OpenClaw, or custom agents with LLM-ranked relevance and impact signals from podcasts, articles and news, X/Twitter, Kalshi, Polymarket, earnings calls, filings, and other monitored sources across public and private companies for decision support or user-controlled trading workflows.
Diseño de prompts para LLMs: system prompts, few-shot examples, chain-of-thought, RAG, structured outputs.
Learn how to manage conversation context in AMCP to avoid LLM API errors from exceeding context windows. This skill covers SmartCompactor strategies, token estimation, configuration, and best practices.
Fine-tune LLMs using reinforcement learning with TRL - SFT for instruction tuning, DPO for preference alignment, PPO/GRPO for reward optimization, and reward model training. Use when need RLHF, align model with preferences, or train from human feedback. Works with HuggingFace Transformers.
Trains an X/Twitter account's algorithmic feed to surface niche-relevant content and positions the account as a thought leader. Browser scripts for manual operation, Persona Engine for identity management, and 24/7 Algorithm Builder with LLM-powered engagement via Puppeteer. Use when a user wants to build their algorithm, cultivate their feed for a niche, grow a fresh account, become a thought leader, or run automated engagement with AI-generated content.
Edit prose to sound more natural, direct, and engaging. Works top-down through four levels (Document → Paragraph → Sentence → Word) with human checkpoints at each stage. Fixes LLM patterns, writerly bad habits, and style deficits. Works for academic papers, reports, memos, essays, blog posts, proposals, and other nonfiction. Use when prose sounds robotic, dull, or inaccessible.
Model Context Protocol (MCP) server development and AI/ML integration patterns. Covers MCP server implementation, tool design, resource handling, and LLM integration best practices. Use when developing MCP servers, creating AI tools, integrating with LLMs, or when asking about MCP protocol, prompt engineering, or AI system architecture.
LLM deployment strategies including vLLM, TGI, and cloud inference endpoints.
Inline adversarial plan review — 3 sequential checks (Feasibility, Completeness, Scope & Alignment) performed by the calling LLM in its own context. No subagents spawned. Call after saving a plan. Returns GATE_PASS or GATE_FAIL with blocking issues.
Ready-to-use prompt templates for specialized agents. Use when building n8n workflows, AI integrations, or sales materials. Contains structured prompts for automation-architect, llm-engineer, and sales-automator agents.
AI-optimized web search using Tavily Search API. Use when you need comprehensive web research, current events lookup, domain-specific search, or AI-generated answer summaries. Tavily is optimized for LLM consumption with clean structured results, answer generation, and raw content extraction. Best for research tasks, news queries, fact-checking, and gathering authoritative sources.