Loading...
Loading...
Found 5 Skills
Add unsigned integer (uint) type support to PyTorch operators by updating AT_DISPATCH macros. Use when adding support for uint16, uint32, uint64 types to operators, kernels, or when user mentions enabling unsigned types, barebones unsigned types, or uint support.
AscendC Operator Design Completion - Assist users in completing operator architecture design, interface definition, and performance planning. Use this skill when users mention operator design, operator development, tiling strategy, memory planning, AscendC kernel design, two-level tiling, inter-core splitting, or intra-core splitting.
The basics of how to program GPUs using Mojo. Use this skill in addition to mojo-syntax when writing Mojo code that targets GPUs or other accelerators. Use targeting code to NVIDIA, AMD, Apple silicon GPUs, or others. Use this skill to overcome misconceptions about how Mojo GPU code is written.
Initialize AscendC operator project and create compilable operator skeleton. Trigger scenarios: (1) User requests to create a new operator; (2) Keywords: ascendc operator, new operator, operator directory, operator initialization; (3) Need to quickly implement based on ascend-kernel template. This skill not only creates directories, but also outputs standard files and checklists for "continuous development".
This skill should be used when the user asks about "Ascend NPU", "昇腾", "Huawei NPU", "triton-ascend", "Ascend kernel development", "NPU算子开发", "Atlas", "CANN", or mentions Ascend hardware, AI Core, Cube/Vector/Scalar units. Provides expert guidance on Ascend NPU hardware architecture, triton-ascend kernel development, and GPU to NPU migration. Always use this skill for Ascend-related questions to avoid confusion with GPU documentation and concepts.