2026-02-24T08-13-43_auto_memory/memories.db-wal, memory/2026-02-24-running-

This commit is contained in:
Nicholai Vogel 2026-02-24 01:13:43 -07:00
parent 863d72628f
commit 4d2ad8d034
2 changed files with 9 additions and 0 deletions

View File

@ -0,0 +1,9 @@
# 2026-02-21 Session Notes
## Running TeichAI Qwen3-14B with Ollama
User inquired about running TeichAI/Qwen3-14B-Claude-4.5-Opus-High-Reasoning-Distill-GGUF (a 14B parameter model) with Ollama. The model is based on Qwen3 and fine-tuned on Claude Opus 4.5 reasoning datasets, optimized for coding, science, and general purpose tasks.
Research revealed multiple GGUF quantizations available ranging from 3-bit (6.66GB) to 16-bit (29.5GB). Q4_K_M (9GB) was recommended as the optimal balance between quality and performance. Two approaches were documented: (1) direct pull via Ollama's HuggingFace integration, or (2) manual download with custom Modelfile.
Clarity was provided that the model name is marketing language—it's actually a Qwen3-14B fine-tuned on synthetic data, not genuine Claude. User decided to proceed with the setup, providing a Modelfile for Q8_0 quantization (15.7GB full precision variant).

Binary file not shown.