Loading...
Loading...
Found 18 Skills
Expert skill for Open-AutoGLM, an AI phone agent framework that controls Android/HarmonyOS/iOS devices via natural language using the AutoGLM vision-language model
Perform image understanding and analysis using GLM Vision MCP. Trigger conditions: (1) Users request image analysis, image understanding, or description of image content (2) Need to identify objects, text, or scenes in images (3) Use GLM's visual understanding capabilities
Use GLM web search MCP to perform internet searches. Trigger conditions: (1) Users request network search, online search, or information lookup (2) Need to query the latest information, news, and materials (3) Use GLM's web_search function
Text-to-speech conversion using GLM-TTS service via the `uvx zai-tts` command for generating audio from text. Use when (1) User requests audio/voice output with the "tts" trigger or keyword. (2) Content needs to be spoken rather than read (multitasking, accessibility, podcast, driving, cooking). (3) Using pre-cloned voices for speech.
Z.ai API integration for building applications with GLM models. Use when working with Z.ai/ZhipuAI APIs for: (1) Chat completions with GLM-4.7/4.6/4.5 models, (2) Vision/multimodal tasks with GLM-4.6V, (3) Image generation with GLM-Image or CogView-4, (4) Video generation with CogVideoX-3 or Vidu models, (5) Audio transcription with GLM-ASR-2512, (6) Function calling and tool use, (7) Web search integration, (8) Translation, slide/poster generation agents. Triggers: Z.ai, ZhipuAI, GLM, BigModel, Zhipu, CogVideoX, CogView, Vidu.
Delegate tasks to the cost-effective opencode/glm-5 model. Use when you need inexpensive task execution, simple research, or delegating work that doesn't require the most powerful models.
Quickly search the user's library of real experiences, opinions and case materials (over 1800 entries) to add authenticity and credibility to content. Use this skill when the user mentions "real experience", "personal case", "well-founded", "practical experience", "material library" or "looking for examples".
Systematically reduce the AI detection rate to below 30%, and add a human touch through a three-round review process (content, style, details). Use this skill when users mention phrases such as "too AI-like", "sounds written by AI", "reduce AI detection rate", "more human-like", "more natural", or "colloquial"
Quickly generate 2-3 video script outline plans, including title suggestions, thumbnail design recommendations, and complete structure design. Use this Skill when users mention "video outline", "video script", "video planning", "shooting videos", or "video content".
Access Claude, Gemini, Kimi, GLM and 100+ LLMs via inference.sh CLI using OpenRouter. Models: Claude Opus 4.5, Claude Sonnet 4.5, Claude Haiku 4.5, Gemini 3 Pro, Kimi K2, GLM-4.6, Intellect 3. One API for all models with automatic fallback and cost optimization. Use for: AI assistants, code generation, reasoning, agents, chat, content generation. Triggers: claude api, openrouter, llm api, claude sonnet, claude opus, gemini api, kimi, language model, gpt alternative, anthropic api, ai model api, llm access, chat api, claude alternative, openai alternative
Statistical models library for Python. Use when you need specific model classes (OLS, GLM, mixed models, ARIMA) with detailed diagnostics, residuals, and inference. Best for econometrics, time series, rigorous inference with coefficient tables. For guided statistical test selection with APA reporting use statistical-analysis.
Provides guidance for LLM post-training with RL using slime, a Megatron+SGLang framework. Use when training GLM models, implementing custom data generation workflows, or needing tight Megatron-LM integration for RL scaling.