vOOice/VoiceInk/Views/Metrics/MetricsSetupView.swift
2025-06-15 19:13:57 +05:45

210 lines
7.6 KiB
Swift

import SwiftUI
import KeyboardShortcuts
struct MetricsSetupView: View {
@EnvironmentObject private var whisperState: WhisperState
@State private var isAccessibilityEnabled = AXIsProcessTrusted()
@State private var isScreenRecordingEnabled = CGPreflightScreenCaptureAccess()
var body: some View {
ScrollView {
VStack(spacing: 24) {
// Header
VStack(spacing: 12) {
AppIconView()
.frame(width: 80, height: 80)
.padding(.bottom, 20)
VStack(spacing: 4) {
Text("Welcome to VoiceInk")
.font(.system(size: 28, weight: .bold, design: .rounded))
.multilineTextAlignment(.center)
Text("Complete the setup to get started")
.font(.system(size: 16))
.foregroundColor(.secondary)
.multilineTextAlignment(.center)
}
}
.padding(.top, 20)
.padding(.bottom, 20)
// Setup Steps
VStack(alignment: .leading, spacing: 0) {
ForEach(0..<4) { index in
setupStep(for: index)
if index < 3 {
Divider().padding(.leading, 70)
}
}
}
.background(Color(NSColor.textBackgroundColor))
.clipShape(RoundedRectangle(cornerRadius: 12))
.overlay(
RoundedRectangle(cornerRadius: 12)
.stroke(Color.gray.opacity(0.2), lineWidth: 1)
)
.padding(.horizontal)
Spacer(minLength: 20)
// Action Button
actionButton
.frame(maxWidth: 400)
// Help Text
helpText
}
.padding()
}
.frame(minWidth: 500, minHeight: 600)
.background(Color(NSColor.controlBackgroundColor))
}
private func setupStep(for index: Int) -> some View {
let stepInfo: (isCompleted: Bool, icon: String, title: String, description: String)
switch index {
case 0:
stepInfo = (
isCompleted: KeyboardShortcuts.getShortcut(for: .toggleMiniRecorder) != nil,
icon: "command",
title: "Set Keyboard Shortcut",
description: "Use VoiceInk anywhere with a shortcut."
)
case 1:
stepInfo = (
isCompleted: isAccessibilityEnabled,
icon: "hand.raised.fill",
title: "Enable Accessibility",
description: "Paste transcribed text at your cursor."
)
case 2:
stepInfo = (
isCompleted: isScreenRecordingEnabled,
icon: "video.fill",
title: "Enable Screen Recording",
description: "Get better transcriptions with screen context."
)
default:
stepInfo = (
isCompleted: whisperState.currentTranscriptionModel != nil,
icon: "arrow.down.to.line",
title: "Download Model",
description: "Choose an AI model to start transcribing."
)
}
return HStack(spacing: 16) {
Image(systemName: stepInfo.icon)
.font(.system(size: 18))
.frame(width: 40, height: 40)
.background((stepInfo.isCompleted ? Color.green : Color.accentColor).opacity(0.1))
.foregroundColor(stepInfo.isCompleted ? .green : Color.accentColor)
.clipShape(Circle())
VStack(alignment: .leading, spacing: 3) {
Text(stepInfo.title)
.font(.headline)
.fontWeight(.semibold)
Text(stepInfo.description)
.font(.subheadline)
.foregroundColor(.secondary)
}
Spacer()
if stepInfo.isCompleted {
Image(systemName: "checkmark.circle.fill")
.font(.system(size: 24))
.foregroundColor(.green)
} else {
Image(systemName: "chevron.right")
.font(.system(size: 14, weight: .bold))
.foregroundColor(Color(NSColor.separatorColor))
}
}
.padding()
}
private var actionButton: some View {
Button(action: handleActionButton) {
HStack {
Text(getActionButtonTitle())
.fontWeight(.semibold)
Image(systemName: "arrow.right")
}
.frame(maxWidth: .infinity)
.padding(.vertical, 12)
.background(Color.accentColor)
.foregroundColor(.white)
.cornerRadius(12)
}
.buttonStyle(.plain)
.shadow(color: Color.accentColor.opacity(0.3), radius: 8, y: 4)
}
private func handleActionButton() {
if isShortcutAndAccessibilityGranted {
openModelManagement()
} else {
// Handle different permission requests based on which one is missing
if KeyboardShortcuts.getShortcut(for: .toggleMiniRecorder) == nil {
openSettings()
} else if !AXIsProcessTrusted() {
if let url = URL(string: "x-apple.systempreferences:com.apple.preference.security?Privacy_Accessibility") {
NSWorkspace.shared.open(url)
}
} else if !CGPreflightScreenCaptureAccess() {
CGRequestScreenCaptureAccess()
// After requesting, open system preferences as fallback
if let url = URL(string: "x-apple.systempreferences:com.apple.preference.security?Privacy_ScreenCapture") {
NSWorkspace.shared.open(url)
}
}
}
}
private func getActionButtonTitle() -> String {
if KeyboardShortcuts.getShortcut(for: .toggleMiniRecorder) == nil {
return "Configure Shortcut"
} else if !AXIsProcessTrusted() {
return "Enable Accessibility"
} else if !CGPreflightScreenCaptureAccess() {
return "Enable Screen Recording"
} else if whisperState.currentTranscriptionModel == nil {
return "Download Model"
}
return "Get Started"
}
private var helpText: some View {
Text("Need help? Check the Help menu for support options")
.font(.caption)
.foregroundColor(.secondary)
}
private var isShortcutAndAccessibilityGranted: Bool {
KeyboardShortcuts.getShortcut(for: .toggleMiniRecorder) != nil &&
AXIsProcessTrusted() &&
CGPreflightScreenCaptureAccess()
}
private func openSettings() {
NotificationCenter.default.post(
name: .navigateToDestination,
object: nil,
userInfo: ["destination": "Settings"]
)
}
private func openModelManagement() {
NotificationCenter.default.post(
name: .navigateToDestination,
object: nil,
userInfo: ["destination": "AI Models"]
)
}
}