Loading...
Loading...
Found 7 Skills
Apple Intelligence skills for on-device AI features including Foundation Models, Visual Intelligence, App Intents, and intelligent assistants. Use when implementing AI-powered features.
Use when debugging Foundation Models issues — context exceeded, guardrail violations, slow generation, availability problems, unsupported language, or unexpected output. Systematic diagnostics with production crisis defense.
Use when integrating App Intents for Siri, Apple Intelligence, Shortcuts, Spotlight, or system experiences - covers AppIntent, AppEntity, parameter handling, entity queries, background execution, authentication, and debugging common integration issues for iOS 16+
Use when implementing ANY Apple Intelligence or on-device AI feature. Covers Foundation Models, @Generable, LanguageModelSession, structured output, Tool protocol, iOS 26 AI integration.
Implement, review, or improve App Intents for Siri, Shortcuts, Spotlight, widgets, and Apple Intelligence. Use when creating AppIntent actions, defining AppEntity models with EntityQuery, building AppShortcutsProvider phrases, adding Spotlight indexing with IndexedEntity, integrating assistant schemas for Apple Intelligence, migrating from SiriKit to App Intents, building interactive widgets with WidgetConfigurationIntent, creating Control Center widgets, implementing SnippetIntent for visual intelligence, or wiring focus filters with SetFocusFilterIntent.
Use when implementing on-device AI with Apple's Foundation Models framework (iOS 26+), building summarization/extraction/classification features, or using @Generable for type-safe structured output.
Xcode 26 and Apple platform update notes with a focus on Liquid Glass, Foundation Models Framework, and Apple Intelligence. Use when asked to explain, summarize, compare, or implement Xcode 26 era framework changes across SwiftUI, UIKit, AppKit, Foundation, App Intents, StoreKit, MapKit, WidgetKit, visionOS, or Swift language features.