Real-time voice transcription, advanced speaker diarization, on-device AI processing, and intelligent note-taking exclusively for iOS 26 & macOS 26 and above
Uses Apple's new Foundation Model Framework and SpeechTranscriber. Requires macOS 26 to run and compile the project. The goal is to demonstrate how easy it is now to build local, AI-first apps.
Swift Scribe is a privacy-first, AI-enhanced transcription application built exclusively for iOS 26/macOS 26+ that transforms spoken words into organized, searchable notes with professional-grade speaker identification. Using Apple's latest SpeechAnalyzer and SpeechTranscriber frameworks (available only in iOS 26/macOS 26+) combined with FluidAudio's advanced speaker diarization and on-device Foundation Models, it delivers real-time speech recognition, intelligent speaker attribution, content analysis, and advanced text editing capabilities.
- iOS 26 Beta or newer (REQUIRED - will not work on iOS 25 or earlier)
- macOS 26 Beta or newer (REQUIRED - will not work on macOS 25 or earlier)
- Xcode Beta with latest Swift 6.2+ toolchain
- Swift 6.2+ programming language
- Apple Developer Account with beta access to iOS 26/macOS 26
- Microphone permissions for speech input
-
Clone the repository:
git clone https://github.com/seamlesscompute/swift-scribe cd swift-scribe
-
Open in Xcode Beta:
open SwiftScribe.xcodeproj
-
Configure deployment targets for iOS 26 Beta/macOS 26 Beta or newer
-
Build and run using Xcode Beta with Swift 6.2+ toolchain
Transform your workflow with AI-powered transcription:
- 📊 Meeting transcription with automatic speaker identification and minute generation
- 📝 Interview recording with real-time speaker diarization and attribution
- 💼 Business documentation with speaker-tagged content and report creation
- 🎯 Sales call analysis with participant tracking and follow-up automation
- 🏥 Medical dictation and clinical documentation
- 👨⚕️ Patient interview transcription with medical terminology
- 📋 Healthcare report generation and chart notes
- 🔬 Research interview analysis and coding
- 🎓 Lecture transcription with chapter segmentation
- 📚 Study note creation from audio recordings
- 🔍 Research interview analysis with theme identification
- 📖 Language learning with pronunciation feedback
- ⚖️ Court proceeding transcription with timestamp accuracy
- 📑 Deposition recording and legal documentation
- 🏛️ Legal research and case note compilation
- 📋 Compliance documentation and audit trails
- 🎙️ Podcast transcription with automatic speaker labeling and show note generation
- 🎬 Video content scripting with professional speaker diarization
- ✍️ Article writing from multi-speaker voice recordings
- 📺 Content creation workflows with speaker-attributed production notes
- 🦻 Real-time captions for hearing-impaired users
- 🗣️ Speech accessibility tools with customizable formatting
- 🌐 Multi-language accessibility support
- 🎯 Assistive technology integration
Scribe/ # Core application logic and modules
├── Audio/ # Audio capture, processing, and FluidAudio speaker diarization
├── Transcription/ # SpeechAnalyzer and SpeechTranscriber implementation
├── AI/ # Foundation Models integration and AI processing
├── Views/ # SwiftUI interface with rich text editing
├── Models/ # Data models for memos, transcription, speakers, and AI
├── Storage/ # Local data persistence and model management
└── Extensions/ # Swift extensions and utilities
- FluidAudio Integration: Industry-grade speaker identification and clustering
- Research-Grade Performance: Competitive with academic benchmarks (17.7% DER on AMI dataset)
- Real-time Processing: Live speaker identification during recording with minimal latency
- Speaker Attribution: Color-coded transcription with confidence scores and timeline mapping
- Automatic Speaker Detection: No manual configuration required
- Speaker Persistence: Consistent speaker identification across recording sessions
- Visual Attribution: Rich text formatting with speaker-specific colors and metadata
- Speaker Analytics: Detailed insights into speaking patterns and participation
- Fully On-Device: All processing happens locally - no cloud dependencies
- Zero Data Transmission: Audio and speaker data never leave your device
- Secure Storage: Speaker embeddings and models stored securely with SwiftData
- Complete Offline Operation: Works without internet connectivity
This project is licensed under the MIT License - see the LICENSE file for complete details.
- Apple WWDC 2025 sessions on SpeechAnalyzer, Foundation Models, and Rich Text editing
- Apple Developer Frameworks - SpeechAnalyzer, Foundation Models, Rich Text Editor
- FluidAudio - Professional speaker diarization and voice identification technology