Loading...
Loading...
Found 30 Skills
LLM and ML model deployment for inference. Use when serving models in production, building AI APIs, or optimizing inference. Covers vLLM (LLM serving), TensorRT-LLM (GPU optimization), Ollama (local), BentoML (ML deployment), Triton (multi-model), LangChain (orchestration), LlamaIndex (RAG), and streaming patterns.
Build AI-first applications with RAG pipelines, embeddings, vector databases, agentic workflows, and LLM integration. Master prompt engineering, function calling, streaming responses, and cost optimization for 2025+ AI development.
Document chunking implementations and benchmarking tools for RAG pipelines including fixed-size, semantic, recursive, and sentence-based strategies. Use when implementing document processing, optimizing chunk sizes, comparing chunking approaches, benchmarking retrieval performance, or when user mentions chunking, text splitting, document segmentation, RAG optimization, or chunk evaluation.
Interact with the Denser Retriever API to build and query knowledge bases. Use this skill whenever the user wants to create a knowledge base, upload documents (files or URLs), search/query a knowledge base, list or delete knowledge bases or documents, check document processing status, or check account usage/balance. Also trigger when the user mentions 'denser retriever', 'knowledge base', 'document search', 'semantic search', 'RAG pipeline', or wants to index and search their files.
Build LLM applications using Dify's visual workflow platform. Use when creating AI chatbots, implementing RAG pipelines, developing agents with tools, managing knowledge bases, deploying LLM apps, or building workflows with drag-and-drop. Supports hundreds of LLMs, Docker/Kubernetes deployment.
LLM app development with RAG, prompt engineering, vector databases, and AI agents