Loading...
Loading...
Found 103 Skills
Use when diagnosing CopilotKit issues -- runtime connectivity failures, agent not responding, streaming errors, tool execution problems, transcription failures, version mismatches, and AG-UI event tracing.
ElevenLabs Speech-to-Text transcription workflows with Scribe v1 supporting 99 languages, speaker diarization, and Vercel AI SDK integration. Use when implementing audio transcription, building STT features, integrating speech-to-text, setting up Vercel AI SDK with ElevenLabs, or when user mentions transcription, STT, Scribe v1, audio-to-text, speaker diarization, or multi-language transcription.
Read, watch, and listen to video/audio files. Use Gemini for native video understanding, or extract key frames + Whisper transcription as fallback. Use when a user sends a video/audio and asks about its content, what's in it, what someone said, etc.
Subtitle generation and burning. Transcription → dictionary correction → review → burning. Trigger words: add subtitles, generate subtitles, subtitles
Transcribe speech to text using the Speech framework. Use when implementing live microphone transcription with AVAudioEngine, recognizing pre-recorded audio files, configuring on-device vs server-based recognition, handling authorization flows, or adopting the new SpeechAnalyzer API (iOS 26+) for modern async/await speech-to-text.
Implement speech-to-text voice input in Blazor applications using Syncfusion SpeechToText component. ALWAYS use this when users need voice input, speech recognition, audio transcription, or implementing the SpeechToText component in Blazor. Trigger for Syncfusion.Blazor.Inputs, microphone input, voice-to-text conversion, language support, transcript binding, listening states, error handling, browser speech API, or any speech recognition requirements.
Speech-to-text transcription using Whisper with word-level timestamps. Use when users ask to transcribe audio or video to text, generate subtitles, or recognize speech.
Cluely platform help — real-time AI meeting assistant with live coaching overlay, pre-call briefs, meeting notes, conversation analytics, and knowledge base RAG. Use when setting up Cluely for live AI prompts during sales calls, configuring the knowledge base with company docs for real-time RAG retrieval, connecting Cluely to HubSpot or Salesforce via Merge.dev, troubleshooting transcription accuracy or speaker attribution errors, comparing Cluely Pro vs Pro + Undetectability plans, or setting up team coaching scorecards and missed opportunity tracking. Do NOT use for choosing between AI note-takers across vendors (use /sales-note-taker) or reviewing a call for coaching (use /sales-call-review).
Omi AI wearable platform help — open-source AI necklace for all-day conversation capture (in-person + online meetings), Developer API (`api.omi.me/v1/dev`, Bearer token, 100 req/min), app marketplace with webhook integrations, memories/conversations/action-items endpoints. Use when setting up an Omi wearable for meeting capture, building a custom Omi app or integration, troubleshooting Bluetooth disconnects or transcription accuracy, connecting Omi to Slack or CRM via webhooks, comparing Omi to Plaud or Limitless for in-person recording, or accessing Omi's API to export conversations and action items. Do NOT use for choosing between software-only note-takers without wearable needs (use /sales-note-taker).
Cleft Notes platform help — Apple-native AI voice-to-notes app with on-device transcription that turns spoken thoughts into organized markdown notes with auto-headings. Use when setting up Cleft Notes for capturing voice memos and converting rambling thoughts into structured notes, configuring Obsidian or Notion sync to route Cleft notes into an existing knowledge base, troubleshooting recordings that fail after a couple minutes or produce garbled transcription output, setting up Zapier automations to send Cleft notes to project management or CRM tools, choosing between Cleft free and Plus plans, deciding whether Cleft or Voicenotes or AudioPen fits your voice capture workflow, or evaluating Cleft for ADHD-friendly voice-first note-taking on Apple devices. Do NOT use for comparing AI meeting note-takers across platforms (use /sales-note-taker) or reviewing a sales call for coaching (use /sales-call-review).
Shadow platform help — bot-free AI meeting assistant capturing audio + screen on macOS, on-device transcription, autopilot meeting detection, AI summaries/action items/follow-up emails, Skills system for custom post-meeting tasks. Use when setting up Shadow for the first time, Shadow not detecting meetings automatically, Shadow using too much CPU or memory on Mac, Shadow speaker attribution is wrong, Shadow screen capture not working, Shadow free tier ran out of AI meetings, choosing between Shadow and Granola or Jamie or Bluedot for bot-free recording, or exporting Shadow notes to Markdown or Zapier. Do NOT use for choosing between all AI note-takers (use /sales-note-taker) or reviewing a call for coaching (use /sales-call-review).
Convin platform help — AI-powered contact center QA, coaching, and conversation intelligence. Use when setting up Convin automated QA scoring, Convin Real-Time Assist not surfacing prompts, Convin transcription missing speakers or inaccurate with accents, Convin audits hanging or calls delayed on dashboard, Convin AI Phone Call agent for outbound, Convin LMS agent training, or evaluating Convin vs Observe.AI vs Cresta vs Balto vs Enthu.AI for contact center QA. Do NOT use for CCaaS platform selection (use /sales-ccaas-selection) or building a coaching program (use /sales-coaching).