Loading...
Loading...
Found 277 Skills
Production-ready patterns for building LLM applications. Covers RAG pipelines, agent architectures, prompt IDEs, and LLMOps monitoring. Use when designing AI applications, implementing RAG, building agents, or setting up LLM observability.
High-performance vector similarity search engine for RAG and semantic search. Use when building production RAG systems requiring fast nearest neighbor search, hybrid search with filtering, or scalable vector storage with Rust-powered performance.
Data framework for building LLM applications with RAG. Specializes in document ingestion (300+ connectors), indexing, and querying. Features vector indices, query engines, agents, and multi-modal support. Use for document Q&A, chatbots, knowledge retrieval, or building RAG pipelines. Best for data-centric LLM applications.
Framework for state-of-the-art sentence, text, and image embeddings. Provides 5000+ pre-trained models for semantic similarity, clustering, and retrieval. Supports multilingual, domain-specific, and multimodal models. Use for generating embeddings for RAG, semantic search, or similarity tasks. Best for production embedding generation.
Production-grade AI agent patterns with MCP integration, agentic RAG, handoff orchestration, multi-layer guardrails, observability, token economics, ROI frameworks, and build-vs-not decision guidance (modern best practices)
Security guidelines for LLM applications based on OWASP Top 10 for LLM 2025. Use when building LLM apps, reviewing AI security, implementing RAG systems, or asking about LLM vulnerabilities like "prompt injection" or "check LLM security".
Expert in vector databases, embedding strategies, and semantic search implementation. Masters Pinecone, Weaviate, Qdrant, Milvus, and pgvector for RAG applications, recommendation systems, and similar
Build LLM applications with LangChain and LangGraph. Use when creating RAG pipelines, agent workflows, chains, or complex LLM orchestration. Triggers on LangChain, LangGraph, LCEL, RAG, retrieval, agent chain.
Expert in aggregating, processing, and synthesizing information from multiple sources into coherent insights. Use when building knowledge graphs, ontologies, RAG systems, or extracting insights across documents. Triggers include "knowledge graph", "ontology", "synthesize information", "GraphRAG", "insight extraction", "cross-document analysis".
Arquitecto de soluciones digitales basadas en IA. Dos modos: (1) ANALIZAR repositorios o código existente y explicar su arquitectura para cualquier audiencia, incluyendo personas sin conocimiento técnico. (2) DISEÑAR la arquitectura completa de sistemas nuevos que usan LLMs, RAG, agentes o fine-tuning. Usa este skill cuando el usuario mencione: arquitectura de IA, diseño de sistema con LLM, capas arquitectónicas, RAG architecture, tech stack para IA, vector database, diagrama de arquitectura, componentes del sistema, embedding, retrieval, pipeline de datos, MLOps, LLMOps, evaluar enfoques, RAG vs fine-tuning, diseñar solución de inteligencia artificial, explicar repositorio, explicar código, analizar proyecto, qué hace este repo, cómo funciona este sistema, explícame este proyecto, o cualquier variación de "qué componentes necesito" o "explícame cómo funciona esto". Actívalo cuando el usuario pegue código, README, estructura de archivos, o mencione un repositorio de GitHub para analizar. También cuando quiera diseñar arquitectura nueva.
Select and optimize embedding models for semantic search and RAG applications. Use when choosing embedding models, implementing chunking strategies, or optimizing embedding quality for specific domains.
Every product will be AI-powered. The question is whether you'll build it right or ship a demo that falls apart in production. This skill covers LLM integration patterns, RAG architecture, prompt engineering that scales, AI UX that users trust, and cost optimization that doesn't bankrupt you. Use when: keywords, file_patterns, code_patterns.