Loading...
Loading...
Found 32 Skills
Runs a quick overview of Clickhouse server health.
MUST USE when installing chv, setting up local ClickHouse development, or running ClickHouse locally. Contains 5 guides covering chv CLI installation, local project initialization, running a local server, executing SQL from files, and migrating to cloud. Always read relevant guide files and cite them in responses.
Comprehensive backend development guide for Langfuse's Next.js 14/tRPC/Express/TypeScript monorepo. Use when creating tRPC routers, public API endpoints, BullMQ queue processors, services, or working with tRPC procedures, Next.js API routes, Prisma database access, ClickHouse analytics queries, Redis queues, OpenTelemetry instrumentation, Zod v4 validation, env.mjs configuration, tenant isolation patterns, or async patterns. Covers layered architecture (tRPC procedures → services, queue processors → services), dual database system (PostgreSQL + ClickHouse), projectId filtering for multi-tenant isolation, traceException error handling, observability patterns, and testing strategies (Jest for web, vitest for worker).
Database operations including querying, schema exploration, and data analysis. Activates for tasks involving PostgreSQL, MySQL, MariaDB, SQLite, MongoDB, Redis, Elasticsearch, or ClickHouse databases.
Use this skill when a user wants to store, manage, or work with Goldsky secrets — the named credential objects used by pipeline sinks. This includes: creating a new secret from a connection string or credentials, listing or inspecting existing secrets, updating or rotating credentials after a password change, and deleting secrets that are no longer needed. Trigger for any query where the user mentions 'goldsky secret', wants to securely store database credentials for a pipeline, or is working with sink authentication for PostgreSQL, Neon, Supabase, ClickHouse, Kafka, S3, Elasticsearch, DynamoDB, SQS, OpenSearch, or webhooks.
Design and architect Goldsky Turbo pipelines. Use this skill for 'should I use X or Y' decisions: kafka source vs dataset source, streaming vs job mode, which resource size (xs/s/m/l/xl/xxl) for my workload, postgres vs clickhouse vs kafka sink, fan-in vs fan-out data flow, one pipeline vs many, dynamic table vs SQL join, how to handle multi-chain deployments. Also use when the user asks 'what's the best way to...' for a pipeline design problem, or is unsure how to structure their pipeline before building it.
Time-series database implementation for metrics, IoT, financial data, and observability backends. Use when building dashboards, monitoring systems, IoT platforms, or financial applications. Covers TimescaleDB (PostgreSQL), InfluxDB, ClickHouse, QuestDB, continuous aggregates, downsampling (LTTB), and retention policies.
Earthly E2E validation for YAML schema contracts. TRIGGERS - schema validation, YAML schema, schema contracts, regenerate types.