Loading...
Loading...
Found 3 Skills
Orchestrate autonomous ML research workflows with cross-model review loops, idea discovery, and experiment automation using Claude Code and Codex MCP
Audit experiment integrity before claiming results. Uses cross-model review (GPT-5.4) to check for fake ground truth, score normalization fraud, phantom results, and insufficient scope. Use when user says "审计实验", "check experiment integrity", "audit results", "实验诚实度", or after experiments complete before writing claims.
Cross-model second opinion from Google Gemini — a different AI reviewing the same changes, with deep Google ecosystem knowledge. Three modes: review (pass/fail gate for Google Ads campaigns, SEO metadata, or code), challenge (adversarial stress-test that tries to break your changes), and consult (open Q&A with Gemini on Google Ads strategy, SEO best practices, or implementation questions). Use when the user says "gemini review", "ask gemini", "gemini challenge", "second opinion from gemini", "consult gemini", "stress test with gemini", "what would gemini say", "cross-model review", or "get another opinion". Voice aliases: "gem", "gemini check". Especially useful for Google Ads changes, SEO metadata updates, campaign structure decisions, keyword strategies, and bid/budget changes — Gemini has native Google ecosystem knowledge that complements Claude's analysis.