Speech-To-Text with MLX (Apple Silicon) and opensource models (default GLM-ASR-Nano-2512) locally.
Installation
Details
Usage
After installing, this skill will be available to your AI coding assistant.
Verify installation:
npx agent-skills-cli listSkill Instructions
name: mlx-stt description: Speech-To-Text with MLX (Apple Silicon) and opensource models (default GLM-ASR-Nano-2512) locally. version: 1.0.7 author: guoqiao metadata: {"openclaw":{"always":true,"emoji":"🦞","homepage":"https://github.com/guoqiao/skills/blob/main/mlx-stt/mlx-stt/SKILL.md","os":["darwin"],"requires":{"bins":["brew"]}}} triggers:
- "/mlx-stt <audio>"
- "STT ..."
- "ASR ..."
- "Transcribe ..."
- "Convert audio to text ..."
MLX STT
Speech-To-Text/ASR/Transcribe with MLX (Apple Silicon) and opensource models (default GLM-ASR-Nano-2512) locally.
Free and Accurate. No api key required. No server required.
Requirements
mlx: macOS with Apple Siliconbrew: used to install deps if not available
Installation
bash ${baseDir}/install.sh
This script will use brew to install these cli tools if not available:
ffmpeg: convert audio format when neededuv: install python package and run python scriptmlx_audio: do the real job
Usage
To transcribe an audio file, run this script:
bash ${baseDir}/mlx-stt.sh <audio_file_path>
- First run could be a little slow, since it will need to download model.
- The transcript result will be printed to stdout.
More by openclaw
View allManage Things 3 via the `things` CLI on macOS (add/update projects+todos via URL scheme; read/search/list from the local Things database). Use when a user asks OpenClaw to add a task to Things, list inbox/today/upcoming, search tasks, or inspect projects/areas/tags.
Capture frames or clips from RTSP/ONVIF cameras.
Generate spectrograms and feature-panel visualizations from audio with the songsee CLI.
Send WhatsApp messages to other people or search/sync WhatsApp history via the wacli CLI (not for normal user chats).
