Loading...
Loading...
Found 6 Skills
Compress large language models using knowledge distillation from teacher to student models. Use when deploying smaller models with retained performance, transferring GPT-4 capabilities to open-source models, or reducing inference costs. Covers temperature scaling, soft targets, reverse KLD, logit distillation, and MiniLLM training strategies.
Bootstraps modular Agent Skills from any repository. Clones the source to `sources/`, extracts core documentation into categorized references under `skills/`, and registers the output in the workspace `AGENTS.md`.
Perform "Rank Reduction" on any domain — start from phenomena, extract dimensions, identify constraints, find irreducible independent generators (rank), and verify through generation tests and validation. Use when the user says "Rank Reduction", "find the rank", "what is the rank of this domain", or wants to find the irreducible principles of any domain.
Extract knowledge from closed tasks and archive context
Use when reducing model size, improving inference speed, or deploying to edge devices - covers quantization, pruning, knowledge distillation, ONNX export, and TensorRT optimizationUse when ", " mentioned.
Efficient AI techniques including model compression, quantization, pruning, knowledge distillation, and hardware-aware optimization for production systems.