<p align="center"> <a href="https://www.youtube.com/watch?v=OEXxvxA0_AE"> <img src="https://img.youtube.com/vi/OEXxvxA0_AE/maxresdefault.jpg" alt="CODEC Demo" width="660"/> </a> <br/> <em>Watch the full demo</em> </p>
What Is CODEC
CODEC is a framework that turns a Mac into a voice-controlled AI workstation. Give it a brain (any LLM โ local or cloud), ears (Whisper), a voice (Kokoro), and eyes (vision model). The rest is Python.
It listens, sees the screen, speaks back, controls apps, writes code, drafts messages, manages Google Workspace, and when it doesn't know how to do something โ it writes its own plugin and learns.
No cloud dependency. No subscription. No data leaving the machine. MIT licensed.
7 Products. One System.
CODEC Core โ The Command Layer
Always-on voice assistant. Say "Hey CODEC" or press F13 to activate. F18 for voice commands. F16 for text input.
50+ skills fire instantly: Google Calendar, Gmail, Drive, Docs, Sheets, Tasks, Keep, Chrome automation, web search, Hue lights, timers, Spotify, clipboard, terminal commands, and more. Most skills bypass the LLM entirely โ direct action, zero latency.
Vision Mouse Control โ See & Click
No other open-source voice assistant does this.
Say "Hey CODEC, click the Submit button" โ CODEC screenshots the screen, sends it to a local UI-specialist vision model (UI-TARS), gets back pixel coordinates, and moves the mouse to click. Fully voice-controlled. Works on any app. No accessibility API required โ pure vision.
| Step | What happens | Speed |
|---|---|---|
| 1 | Whisper transcribes voice command | ~2s |
| 2 | Target extracted from natural speech | instant |
| 3 | Screenshot captured and downscaled | instant |
| 4 | UI-TARS locates the element by pixel coordinates | ~4s |
| 5 | pyautogui moves cursor and clicks | instant |
"I'm on Cloudflare and can't find the SSL button โ click it for me." That works. CODEC strips the conversational noise, extracts "SSL button", and finds it on screen.
CODEC Dictate โ Hold, Speak, Paste
Hold a key. Say what you mean. Release. Text appears wherever the cursor is. If CODEC detects a message draft, it refines through the LLM โ grammar fixed, tone polished, meaning preserved. Works in every app on macOS. A free, open-source SuperWhisper replacement that runs entirely local.
CODEC Instant โ One Right-Click
Select any text, anywhere. Right-click. Eight AI services system-wide: Proofread, Elevate, Explain, Translate, Reply (with :tone syntax), Prompt, Read Aloud, Save. Powered by the local LLM.
CODEC Chat โ 250K Context + 12 Agent Crews
Full conversational AI. Long context. File uploads. Image analysis via vision model. Web search. Conversation history.
Plus 12 autonomous agent crews โ not single prompts, full multi-step workflows. Say "research the latest AI agent frameworks and write a report." Minutes later there's a formatted Google Doc in Drive with sources, images, and recommendations.
| Crew | Output |
|---|---|
| Deep Research | 10,000-word illustrated report โ Google Docs |
| Daily Briefing | Morning news + calendar โ Google Docs |
| Competitor Analysis | SWOT + positioning โ Google Docs |
| Trip Planner | Full itinerary โ Google Docs |
| Email Handler | Triage inbox, draft replies |
| Social Media | Posts for Twitter, LinkedIn, Instagram |
| Code Review | Bugs + security + clean code |
| Data Analysis | Trends + insights report |
| Content Writer | Blog posts, articles, copy |
| Meeting Summarizer | Action items from transcripts |
| Invoice Generator | Professional invoices |
| Custom Agent | Define your own role, tools, task |
Schedule any crew: "Run competitor analysis every Monday at 9am"
The multi-agent framework is under 800 lines. Zero dependencies. No CrewAI. No LangChain.
CODEC Vibe โ AI Coding IDE + Skill Forge
Split-screen in the browser. Monaco editor on the left (same engine as VS Code). AI chat on the right. Describe what's needed โ CODEC writes it, click Apply, run it, live preview in browser.
Skill Forge takes it further: describe a new capability in plain English, CODEC converts it into a working plugin. The framework writes its own extensions.
CODEC Voice โ Live Voice Calls
Real-time voice-to-voice conversations with the AI. WebSocket pipeline โ no Pipecat, no external dependencies. Call CODEC from a phone, talk naturally, and mid-call say "check my screen" โ it takes a screenshot, analyzes it, and speaks the result back.
Full transcript saved to memory. Every conversation becomes searchable context for future sessions.
CODEC Overview โ Dashboard Anywhere
Private dashboard accessible from any device, anywhere. Cloudflare Tunnel or Tailscale VPN โ no port forwarding, no third-party relay. Send commands, view the screen, launch voice calls, manage agents โ all from a browser.
Screenshots
<p align="center"> <img src="docs/screenshots/quick-chat.png" alt="Quick Chat" width="720"/><br/> <em>Chat โ ask anything, drag & drop files, full conversation history</em> </p> <p align="center"> <img src="docs/screenshots/chat-analysis.png" alt="Chat with File Analysis" width="720"/><br/> <em>Deep Chat โ upload files, select agents, get structured analysis</em> </p> <p align="center"> <img src="docs/screenshots/voice-call.png" alt="Voice Call" width="720"/><br/> <em>Voice Call โ real-time conversation with live transcript</em> </p> <p align="center"> <img src="docs/screenshots/vibe-code.png" alt="Vibe Code" width="720"/><br/> <em>Vibe Code โ describe what you want, get working code with live preview</em> </p> <p align="center"> <img src="docs/screenshots/deep-research.png" alt="Deep Research Report" width="720"/><br/> <em>Deep Research โ multi-agent reports delivered to Google Docs</em> </p> <p align="center"> <img src="docs/screenshots/tasks.png" alt="Tasks & Schedules" width="720"/><br/> <em>Scheduled automations โ morning briefings, competitor analysis, on cron</em> </p> <details> <summary><strong>More screenshots</strong></summary> <br/> <p align="center"> <img src="docs/screenshots/settings.png" alt="Settings" width="720"/><br/> <em>Settings โ LLM, TTS, STT, hotkeys, wake word configuration</em> </p> <p align="center"> <img src="docs/screenshots/agent-options.png" alt="Agent Options" width="420"/><br/> <em>12 specialized agent crews</em> </p> <p align="center"> <img src="docs/screenshots/login-auth.png" alt="Authentication" width="320"/><br/> <em>Touch ID + PIN + 2FA authentication</em> </p> <p align="center"> <img src="docs/screenshots/right-click-menu.png" alt="Right-Click Menu" width="300"/><br/> <em>Right-click integration โ CODEC in every app</em> </p> <p align="center"> <img src="docs/screenshots/terminal.png" alt="Terminal" width="400"/><br/> <em>50+ skills loaded at startup</em> </p> </details>What Makes CODEC Different
| Capability | CODEC | Siri / Alexa / Google | ChatGPT / Claude |
|---|---|---|---|
| Controls the computer | Full macOS control | Limited smart home | No |
| Reads the screen | Vision model | No | No |
| Clicks UI elements by voice | Vision + mouse control | No | No (Cloud Computer Use only) |
| Runs 100% local | Yes โ all models on device | No | No |
| Voice-to-voice calls | WebSocket, real-time | Yes but cloud | Yes but cloud |
| Multi-agent workflows | 12 crews, local LLM | No | Limited |
| Right-click AI services | 8 system-wide services | No | No |
| Writes its own plugins | Skill Forge | No | No |
| Open source | MIT | No | No |
What CODEC replaced with native code:
| Before | After |
|---|---|
| Pipecat | CODEC Voice (own WebSocket pipeline) |
| CrewAI + LangChain | CODEC Agents (795 lines, zero dependencies) |
| SuperWhisper | CODEC Dictate (free, open source) |
| Cursor / Windsurf | CODEC Vibe (Monaco + AI + Skill Forge) |
| Google Assistant / Siri | CODEC Core (actually controls the computer) |
| Grammarly | CODEC Instant (right-click services via local LLM) |
| ChatGPT | CODEC Chat (250K context, fully local) |
| Cloud LLM APIs | Local stack (Qwen + Whisper + Kokoro + Vision) |
| Vector databases | FTS5 SQLite (simpler, faster, private) |
External services: DuckDuckGo for web search. Cloudflare free tier for the tunnel (or Tailscale). Everything else runs on local hardware.
Quick Start
git clone https://github.com/AVADSA25/codec.git
cd codec
./install.sh
The setup wizard handles everything in 9 steps: LLM, voice, vision, hotkeys, Google OAuth, remote access, and more.
Requirements:
- macOS Ventura or later
- Python 3.10+
- An LLM (Ollama, LM Studio, MLX, or any OpenAI-compatible API)
- Whisper for voice input, Kokoro for voice output, a vision model for screen reading
Supported LLMs
| Model | How to run |
|---|---|
| Qwen 3.5 35B (recommended) | mlx-lm.server --model mlx-community/Qwen3.5-35B-A3B-4bit |
| Llama 3.3 70B | mlx-lm.server --model mlx-community/Llama-3.3-70B-Instruct-4bit |
| Mistral 24B | mlx-lm.server --model mlx-community/Mistral-Small-3.1-24B-Instruct-2503-4bit |
| Gemma 3 27B | mlx-lm.server --model mlx-community/gemma-3-27b-it-4bit |
| GPT-4o (cloud) | "llm_url": "https://api.openai.com/v1" |
| Claude (cloud) | OpenAI-compatible proxy |
| Ollama (any model) | "llm_url": "http://localhost:11434/v1" |
Configure in ~/.codec/config.json:
{
"llm_url": "http://localhost:8081/v1",
"model": "mlx-community/Qwen3.5-35B-A3B-4bit"
}
Keyboard Shortcuts
| Key | Action |
|---|---|
| F13 | Toggle CODEC ON/OFF |
| F18 (hold) | Record voice โ release to send |
| F18 (double-tap) | PTT Lock โ hands-free recording |
| F16 | Text input dialog |
* * | Screenshot + AI analysis |
+ + | Document mode |
| Camera icon | Live webcam PIP โ drag around, snapshot anytime |
| Select text โ right-click | 8 AI services in context menu |
Laptop (F1-F12): F5 = toggle, F8 = voice, F9 = text input. Run python3 setup_codec.py โ select "Laptop / Compact" in Step 4.
Custom shortcuts in ~/.codec/config.json. Restart after changes: pm2 restart open-codec
Privacy & Security
5-layer security stack:
| Layer | Protection |
|---|---|
| Network | Cloudflare Zero Trust tunnel or Tailscale VPN, CORS restricted origins |
| Auth | Touch ID + PIN + TOTP 2FA, timing-safe token comparison |
| Encryption | AES-256-GCM + ECDH P-256 key exchange, per-session keys |
| Execution | Subprocess isolation, resource limits (512MB RAM, 120s CPU), command blocklist, human review gate |
| Data | Local SQLite, parameterized queries, FTS5 full-text search โ searchable, private, yours |
Every conversation is stored locally in SQLite with FTS5 full-text search. No cloud sync. No analytics. No telemetry.
MCP Server โ CODEC Inside Claude, Cursor, VS Code
CODEC exposes tools as an MCP server. Any MCP-compatible client can invoke CODEC skills directly:
{
"mcpServers": {
"codec": {
"command": "python3",
"args": ["/path/to/codec-repo/codec_mcp.py"]
}
}
}
Then in Claude Desktop: "Use CODEC to check my calendar for tomorrow."
Skills opt-in to MCP exposure with SKILL_MCP_EXPOSE = True.
Debugging & Development
Recommended tools:
| Tool | How it helps |
|---|---|
| Claude Code | Terminal AI โ reads the full codebase, runs commands, fixes errors in context |
| Cursor | AI IDE โ navigate CODEC's 60+ files, refactor, debug with full project awareness |
| Windsurf | AI IDE โ strong at multi-file reasoning |
| Antigravity | AI debugging assistant โ paste errors, get fixes with codebase context |
Quick debug commands:
# Check all services
pm2 list
# Check specific service logs
pm2 logs open-codec --lines 30 --nostream # Main CODEC process
pm2 logs codec-dashboard --lines 30 --nostream # Dashboard API
pm2 logs whisper-stt --lines 10 --nostream # Speech-to-text
pm2 logs kokoro-82m --lines 10 --nostream # Text-to-speech
# Verify LLM is responding
curl -s http://localhost:8081/v1/models | python3 -m json.tool
# Verify dashboard is up
curl -s http://localhost:8090/health
# Restart everything
pm2 restart all
# Full health check
python3 -c "from codec_config import *; print('Config OK')"
Common issues:
<details> <summary><strong>Keys don't work</strong></summary>- macOS stealing F-keys? System Settings โ Keyboard โ "Use F1, F2, etc. as standard function keys"
- After config change:
pm2 restart open-codec
- Check Whisper:
pm2 logs whisper-stt --lines 5 --nostream - Check mic permission: System Settings โ Privacy โ Microphone
- Say "Hey CODEC" clearly โ 3 distinct syllables
- Check Kokoro TTS:
curl http://localhost:8085/v1/models - Fallback:
"tts_engine": "say"in config.json (macOS built-in) - Disable:
"tts_engine": "none"
- Check:
curl http://localhost:8090/health - Restart:
pm2 restart codec-dashboard - Remote via Cloudflare:
pm2 logs cloudflared --lines 3 --nostream - Remote via Tailscale: access CODEC at
http://100.x.x.x:8090
- First run takes 2-5 min โ multi-step research with multiple searches
- Check:
pm2 logs codec-dashboard --lines 30 --nostream | grep -i agent - Agents run as background jobs โ no Cloudflare timeout
Project Structure
codec.py โ Entry point
codec_config.py โ Configuration + transcript cleaning
codec_keyboard.py โ Keyboard listener, PTT lock, wake word
codec_dispatch.py โ Skill matching and dispatch (with fallback)
codec_agent.py โ LLM session builder
codec_agents.py โ Multi-agent crew framework (12 crews)
codec_voice.py โ WebSocket voice pipeline
codec_voice.html โ Voice call UI
codec_dashboard.py โ Web API + dashboard (60+ endpoints)
codec_dashboard.html โ Dashboard UI
codec_chat.html โ Chat UI
codec_vibe.html โ Vibe Code IDE
codec_auth.html โ Authentication (Touch ID + PIN + TOTP 2FA)
codec_textassist.py โ 8 right-click services
codec_search.py โ DuckDuckGo + Serper search
codec_mcp.py โ MCP server
codec_memory.py โ FTS5 memory search
codec_heartbeat.py โ Health monitoring + task auto-execution
codec_scheduler.py โ Cron-like agent scheduling
codec_marketplace.py โ Skill marketplace CLI
codec_overlays.py โ AppKit overlay notifications (fullscreen compatible)
ax_bridge/ โ Swift AX accessibility bridge
swift-overlay/ โ SwiftUI status bar app
skills/ โ 50+ built-in skills (incl. vision mouse control)
tests/ โ 212+ pytest tests
install.sh โ One-line installer
setup_codec.py โ Setup wizard (9 steps)
What's Coming
- Linux support
- Windows via WSL
- Multi-machine sync (skills + memory across devices)
- iOS app (dictation + remote dashboard)
- Streaming voice responses (first token plays while rest generates)
- Multi-LLM routing (fast model for simple, strong model for complex)
Contributing
All skill contributions welcome. 50+ built-in, marketplace growing.
git clone https://github.com/AVADSA25/codec.git
cd codec && ./install.sh
python3 -m pytest # all tests must pass
See CONTRIBUTING.md.
Support the Project
If CODEC saves you time:
- Star this repo
- Donate via PayPal โ ava.dsa25@proton.me
- Enterprise setup: avadigital.ai
Professional Setup
Need CODEC configured for a business, integrated with existing tools, or deployed across a team?
Contact AVA Digital for professional setup and custom skill development.
<p align="center"> Star it. Clone it. Rip it apart. Make it yours. </p> <p align="center"> Built by <a href="https://avadigital.ai">AVA Digital LLC</a> ยท MIT License </p>