Implements advanced speech-to-text capabilities in iOS applications using the latest Apple Speech frameworks and Swift concurrency.
This skill provides comprehensive guidance for integrating speech recognition into Swift applications, covering everything from legacy SFSpeechRecognizer to the cutting-edge SpeechAnalyzer API introduced in iOS 26. It assists developers in managing complex authorization flows for both microphones and speech services, configuring on-device processing for offline use, and utilizing modern AsyncSequence patterns for real-time transcription. Whether you are building live voice-controlled interfaces or processing pre-recorded audio files, this skill ensures best practices for performance, battery efficiency, and transcription accuracy using Swift 6 concurrency.
주요 기능
01Automated handling of dual authorization for Microphone and Speech Recognition
02Configuration for on-device versus server-based recognition and asset management
03Live microphone transcription patterns using AVAudioEngine
04Implementation of advanced features like automatic punctuation and contextual strings
05409 GitHub stars
06Support for the new iOS 26+ SpeechAnalyzer actor-based API
사용 사례
01Transcribing pre-recorded audio files for content indexing or accessibility
02Implementing on-device, offline speech recognition for privacy-focused applications
03Building a real-time voice-to-text dictation feature for a productivity app