Loading...
Loading...
Found 36 Skills
Automated hypothesis generation and testing using large language models. Use this skill when generating scientific hypotheses from datasets, combining literature insights with empirical data, testing hypotheses against observational data, or conducting systematic hypothesis exploration for research discovery in domains like deception detection, AI content detection, mental health analysis, or other empirical research tasks.
Builds LLM applications with LangChain including chains, agents, memory, tools, and RAG pipelines. Use when users request "LangChain setup", "LLM chain", "AI workflow", "conversational AI", or "RAG pipeline".
Comprehensive guide for building production-grade LLM applications using LangChain's chains, agents, memory systems, RAG patterns, and advanced orchestration
PocketFlow framework for building LLM applications with graph-based abstractions, design patterns, and agentic coding workflows
Guides the agent through building LLM-powered applications with LangChain and stateful agent workflows with LangGraph. Triggered when the user asks to "create an AI agent", "build a LangChain chain", "create a LangGraph workflow", "implement tool calling", "build RAG pipeline", "create a multi-agent system", "define agent state", "add human-in-the-loop", "implement streaming", or mentions LangChain, LangGraph, chains, agents, tools, retrieval augmented generation, state graphs, or LLM orchestration.
Specialized AI assistant for DSPy development with deep knowledge of predictors, optimizers, adapters, and GEPA integration. Provides session management, codebase indexing, and command-based workflows.
Engineer effective LLM prompts using zero-shot, few-shot, chain-of-thought, and structured output techniques. Use when building LLM applications requiring reliable outputs, implementing RAG systems, creating AI agents, or optimizing prompt quality and cost. Covers OpenAI, Anthropic, and open-source models with multi-language examples (Python/TypeScript).
Framework adoption decision matrix: custom vs large frameworks in the Claude Code era. Use when evaluating whether to adopt a large framework or build custom with AI.
Data engineering, machine learning, AI, and MLOps. From data pipelines to production ML systems and LLM applications.
Form a high-level investment committee consisting of three virtual experts modeled after legendary investors (Buffett, Wood, Druckenmiller) to conduct independent multi-round adversarial debates. True independent thinking is achieved through physically isolated Gemini API calls, and final resolutions are formed via voting. Use when evaluating investment decisions, reviewing stock research reports, or seeking multi-perspective analysis on public companies.
Build LLM applications using Dify's visual workflow platform. Use when creating AI chatbots, implementing RAG pipelines, developing agents with tools, managing knowledge bases, deploying LLM apps, or building workflows with drag-and-drop. Supports hundreds of LLMs, Docker/Kubernetes deployment.
LLM app development with RAG, prompt engineering, vector databases, and AI agents