mirror of
https://github.com/Memo-2023/mana-monorepo.git
synced 2026-05-14 22:41:09 +02:00
The Mac Mini hasn't run mana-llm/stt/tts/image-gen for a while — those services live on the Windows GPU server now. The Mac-targeted installers, plists, and platform-checking setup scripts have been sitting in the repo as cargo-cult, suggesting Mac Mini deployment is still a real option. It isn't. Removed (Mac-Mini deployment infrastructure): services/mana-stt/ - com.mana.mana-stt.plist (LaunchAgent) - com.mana.vllm-voxtral.plist (LaunchAgent for the abandoned local Voxtral experiment) - install-service.sh (single-service launchd installer) - install-services.sh (mana-stt + vllm-voxtral installer) - setup.sh (Mac arm64 installer) - scripts/setup-vllm.sh (vLLM-Voxtral setup) - scripts/start-vllm-voxtral.sh services/mana-tts/ - com.mana.mana-tts.plist - install-service.sh - setup.sh (Mac arm64 installer) scripts/mac-mini/ - setup-image-gen.sh (Mac flux2.c launchd installer) - setup-stt.sh - setup-tts.sh - launchd/com.mana.image-gen.plist - launchd/com.mana.mana-stt.plist - launchd/com.mana.mana-tts.plist setup-tts-bot.sh stays — it's the Matrix TTS bot installer (Synapse side), not the mana-tts service. Updated: - services/mana-stt/CLAUDE.md, README.md — fully rewritten for the Windows GPU reality (CUDA WhisperX, Scheduled Task ManaSTT, .env keys matching the actual production .env on the box) - services/mana-tts/CLAUDE.md, README.md — same treatment, documenting Kokoro/Piper/F5-TTS on the Windows GPU under Scheduled Task ManaTTS - scripts/mac-mini/README.md — dropped the STT setup section, replaced with a pointer to docs/WINDOWS_GPU_SERVER_SETUP.md and the per-service CLAUDE.md files - docs/MAC_MINI_SERVER.md — expanded the "deactivated launchagents" list to mention the now-removed plists, added the full GPU service port table with public URLs, added a cleanup snippet for any old plists still installed on a Mac Mini somewhere |
||
|---|---|---|
| .. | ||
| app | ||
| .env.example | ||
| CLAUDE.md | ||
| grafana-dashboard.json | ||
| README.md | ||
| requirements-cuda.txt | ||
| requirements.txt | ||
| service.pyw | ||
Mana STT Service
Speech-to-Text API service running on the Windows GPU server (mana-server-gpu, RTX 3090). Wraps WhisperX (CUDA, large-v3 + word alignment + pyannote diarization), local Voxtral via vLLM, and the hosted Mistral Voxtral API.
For architecture, deployment, configuration, and operations see CLAUDE.md and docs/WINDOWS_GPU_SERVER_SETUP.md.
Port: 3020
Public URL
https://gpu-stt.mana.how (via Cloudflare Tunnel + Mac Mini gpu-proxy)
API Endpoints
| Endpoint | Method | Description |
|---|---|---|
/health |
GET | Health check + which backends are loaded |
/models |
GET | List available models |
/transcribe |
POST | Whisper / WhisperX transcription |
/transcribe/voxtral |
POST | Voxtral transcription (local vLLM) |
/transcribe/auto |
POST | Auto-select best backend for the input |
All endpoints (except /health) require Authorization: Bearer <token>.
Quick Test
curl -F "file=@audio.wav" -F "language=de" \
-H "Authorization: Bearer $INTERNAL_API_KEY" \
https://gpu-stt.mana.how/transcribe