Loading...
Loading...
Found 17 Skills
Deploy ML models with FastAPI, Docker, Kubernetes. Use for serving predictions, containerization, monitoring, drift detection, or encountering latency issues, health check failures, version conflicts.
Expert in building scalable ML systems, from data pipelines and model training to production deployment and monitoring.
模型自动降级与故障切换。当主模型请求失败、超时、达到速率限制或配额耗尽时,自动切换到备用模型,确保服务连续性。支持多供应商、多优先级的智能模型选择,提供健康监控、自动重试和错误恢复机制。
Expert in Machine Learning Operations bridging data science and DevOps. Use when building ML pipelines, model versioning, feature stores, or production ML serving. Triggers include "MLOps", "ML pipeline", "model deployment", "feature store", "model versioning", "ML monitoring", "Kubeflow", "MLflow".
NannyML integration. Manage data, records, and automate workflows. Use when the user wants to interact with NannyML data.
Build trading systems in the style of Two Sigma, the systematic investment manager pioneering machine learning at scale. Emphasizes alternative data, distributed computing, feature engineering, and rigorous ML infrastructure. Use when building ML pipelines for alpha research, feature stores, or large-scale backtesting systems.
Expert ML engineering covering model development, MLOps, feature engineering, model deployment, and production ML systems.
Use when establishing tests, monitoring, and incident response for analytics models.
Use when "deploying ML models", "MLOps", "model serving", "feature stores", "model monitoring", or asking about "PyTorch deployment", "TensorFlow production", "RAG systems", "LLM integration", "ML infrastructure"
Produce a long-form, shareable markdown writeup on whether Claude has regressed on this user's work. A bundled Python script scans `~/.claude/projects/`, computes every metric, and renders a markdown skeleton with tables already filled — in ~2.5s. Claude fills a dozen short narrative placeholders and saves. Writes `./cc-canary-<YYYY-MM-DD>.md` suitable for pasting into a GitHub issue or gist.
Add PostHog LLM analytics to trace AI model usage. Use after implementing LLM features or reviewing PRs to ensure all generations are captured with token counts, latency, and costs. Also handles initial PostHog SDK setup if not yet installed.
You are **Model QA Specialist**, an independent QA expert who audits machine learning and statistical models across their full lifecycle. You challenge assumptions, replicate results, dissect predi...