diff --git a/VoiceInk/Whisper/WhisperState.swift b/VoiceInk/Whisper/WhisperState.swift index 3014ad1..60fc61f 100644 --- a/VoiceInk/Whisper/WhisperState.swift +++ b/VoiceInk/Whisper/WhisperState.swift @@ -306,9 +306,9 @@ class WhisperState: NSObject, ObservableObject { let transcriptionStart = Date() var text = try await transcriptionService.transcribe(audioURL: url, model: model) - logger.notice("📝 Raw transcript: \(text)") + logger.notice("📝 Raw transcript: \(text, privacy: .public)") text = TranscriptionOutputFilter.filter(text) - logger.notice("📝 Output filter result: \(text)") + logger.notice("📝 Output filter result: \(text, privacy: .public)") let transcriptionDuration = Date().timeIntervalSince(transcriptionStart) let powerModeManager = PowerModeManager.shared @@ -322,12 +322,12 @@ class WhisperState: NSObject, ObservableObject { if UserDefaults.standard.object(forKey: "IsTextFormattingEnabled") as? Bool ?? true { text = WhisperTextFormatter.format(text) - logger.notice("📝 Formatted transcript: \(text)") + logger.notice("📝 Formatted transcript: \(text, privacy: .public)") } if UserDefaults.standard.bool(forKey: "IsWordReplacementEnabled") { text = WordReplacementService.shared.applyReplacements(to: text) - logger.notice("📝 WordReplacement: \(text)") + logger.notice("📝 WordReplacement: \(text, privacy: .public)") } let audioAsset = AVURLAsset(url: url) @@ -357,7 +357,7 @@ class WhisperState: NSObject, ObservableObject { do { let (enhancedText, enhancementDuration, promptName) = try await enhancementService.enhance(textForAI) - logger.notice("📝 AI enhancement: \(enhancedText)") + logger.notice("📝 AI enhancement: \(enhancedText, privacy: .public)") transcription.enhancedText = enhancedText transcription.aiEnhancementModelName = enhancementService.getAIService()?.currentModel transcription.promptName = promptName