Loading...
Loading...
Found 758 Skills
Deep test, analyze, and audit Claude skills. Use this skill whenever the user wants to test a skill's behavior, analyze how it uses the Claude API, inspect inputs/outputs from scripts, or run security and code review audits against skill scripts. Trigger on: "test my skill", "analyze this skill", "audit skill scripts", "review skill for security issues", "what does this skill actually do when it runs", "inspect API calls from skill", "run a skill through its paces", "check my skill for bugs or vulnerabilities". Also trigger when the user shows you a SKILL.md and asks you to evaluate, critique, or stress-test it.
Applies principles from Robert C. Martin's 'Clean Code'. Use this skill when writing, reviewing, or refactoring code to ensure high quality, readability, and maintainability. Covers naming, functions, comments, error handling, and class design.
Diff Review - analyzes code changes and provides structured feedback before commit
Comprehensive Python programming guidelines based on Google's Python Style Guide. Use when you needs to write Python code, review Python code for style issues, refactor Python code, or provide Python programming guidance. Covers language rules (imports, exceptions, type annotations), style rules (naming conventions, formatting, docstrings), and best practices for clean, maintainable Python code.
Execute CodeRabbit secondary workflow: Core Workflow B. Use when implementing secondary use case, or complementing primary workflow. Trigger with phrases like "coderabbit secondary workflow", "secondary task with coderabbit".
Guides the contribution workflow for Apache Beam, including creating PRs, issue management, code review process, and release cycles. Use when contributing code, creating PRs, or understanding the contribution process.
Review code for conceptual errors, wrong assumptions, edge cases, and overcomplication; use after medium/large changes or when risk is high.
Use this skill when validating ANY potential code review finding. Apply BEFORE classifying to verify the finding is real; can you trace incorrect behavior, is it handled elsewhere, and are you certain about framework semantics? If any answer is no, DO NOT create the finding.
LOAD THIS SKILL when: user mentions 'pr'. Covers the full PR lifecycle — create PR, then aggressively watch CI + reviews in a continuous loop, fixing failures and addressing comments immediately until the PR is fully green.
Operate OpenAI Codex CLI (terminal coding agent) to accomplish software engineering tasks. Use when the user asks to: run codex commands, use codex for coding tasks, execute codex exec for automation, do code review with codex, manage codex sessions (resume/fork), configure codex (config.toml, approval modes, sandbox), use codex cloud, set up MCP servers in codex, or any task involving the `codex` command-line tool. Triggers: codex, codex exec, codex review, codex cloud, codex mcp, codex resume, codex sandbox, openai codex.
Review code changes in Tenzir projects. Use when auditing diffs or pull requests for bugs, security issues, missing tests, documentation drift, readability problems, performance regressions, user experience issues, or when deciding how to respond to GitHub review comments. Also use this skill whenever the user says "review", "look at this PR", "check my changes", "audit this diff", "what do you think of this code", or asks for feedback on any code they've written or changed — even if they don't explicitly say "code review."
Python coding standards and best practices. Applicable to code generation, code review, and code refactoring scenarios.