Loading...
Loading...
Found 47 Skills
Use when managing Helm values files and configuration overrides for customizing Kubernetes deployments.
Deploy Customer.io integrations to production. Use when deploying to cloud platforms, setting up production infrastructure, or automating deployments. Trigger with phrases like "deploy customer.io", "customer.io production", "customer.io cloud run", "customer.io kubernetes".
Expert guidance for authoring and maintaining Helm charts following standardized conventions, global registry support, templating best practices, and Kubernetes deployment patterns.
Deploy Juicebox integrations to production. Use when deploying to cloud platforms, configuring production environments, or setting up infrastructure for Juicebox. Trigger with phrases like "deploy juicebox", "juicebox production deploy", "juicebox infrastructure", "juicebox cloud setup".
Execute Clay production deployment checklist and rollback procedures. Use when deploying Clay integrations to production, preparing for launch, or implementing go-live procedures. Trigger with phrases like "clay production", "deploy clay", "clay go-live", "clay launch checklist".
Kubernetes Deployment 管理
CI/CD 流水线配置
Create, manage, and deploy Helm charts for Kubernetes package management. Build reusable chart templates, manage releases, configure values, and use Helm repositories. Use when packaging Kubernetes applications or managing K8s deployments with Helm.
CI/CD pipeline design with GitHub Actions, Docker, Kubernetes, Helm, and GitOps patterns
End-to-end application deployment orchestration for the Kubernetes homelab. Use when: (1) Deploying a new application to the cluster, (2) Adding a new Helm release to the platform, (3) Setting up monitoring, alerting, and health checks for a new service, (4) Research before deploying, (5) Testing deployment on dev cluster before GitOps promotion. Triggers: "deploy app", "add new application", "deploy to kubernetes", "install helm chart", "/deploy-app", "set up new service", "add monitoring for", "deploy with monitoring"
LLM deployment strategies including vLLM, TGI, and cloud inference endpoints.