MCPHub LabRegistryAVADSA25/codec
AVADSA25

AVADSA25/codec

Built by AVADSA25 โ€ข 74 stars

What is AVADSA25/codec?

Open Source Computer Command Framework

How to use AVADSA25/codec?

1. Install a compatible MCP client (like Claude Desktop). 2. Open your configuration settings. 3. Add AVADSA25/codec using the following command: npx @modelcontextprotocol/avadsa25-codec 4. Restart the client and verify the new tools are active.
๐Ÿ›ก๏ธ Scoped (Restricted)
npx @modelcontextprotocol/avadsa25-codec --scope restricted
๐Ÿ”“ Unrestricted Access
npx @modelcontextprotocol/avadsa25-codec

Key Features

Native MCP Protocol Support
Real-time Tool Activation & Execution
Verified Standard Implementation
Secure Resource & Context Handling

Optimized Use Cases

Extending AI models with custom local capabilities
Automating system workflows via natural language
Connecting external data sources to LLM context windows

AVADSA25/codec FAQ

Q

Is AVADSA25/codec safe?

Yes, AVADSA25/codec follows the standardized Model Context Protocol security patterns and only executes tools with explicit user-granted permissions.

Q

Is AVADSA25/codec up to date?

AVADSA25/codec is currently active in the registry with 74 stars on GitHub, indicating its reliability and community support.

Q

Are there any limits for AVADSA25/codec?

Usage limits depend on the specific implementation of the MCP server and your system resources. Refer to the official documentation below for technical details.

Official Documentation

View on GitHub
<p align="center"> <img src="https://i.imgur.com/RbrQ7Bt.png" alt="CODEC" width="280"/> </p> <h1 align="center">CODEC</h1> <p align="center"><strong>Open-Source Intelligent Command Layer for macOS</strong></p> <p align="center"><em>Your voice. Your computer. Your rules. No limit.</em></p> <p align="center"> <a href="https://opencodec.org">opencodec.org</a> ยท <a href="https://avadigital.ai">AVA Digital LLC</a> ยท <a href="#quick-start">Get Started</a> ยท <a href="#support-the-project">Support</a> ยท <a href="#professional-setup">Enterprise</a> </p>
<p align="center"> <a href="https://www.youtube.com/watch?v=OEXxvxA0_AE"> <img src="https://img.youtube.com/vi/OEXxvxA0_AE/maxresdefault.jpg" alt="CODEC Demo" width="660"/> </a> <br/> <em>Watch the full demo</em> </p>

What Is CODEC

CODEC is a framework that turns a Mac into a voice-controlled AI workstation. Give it a brain (any LLM โ€” local or cloud), ears (Whisper), a voice (Kokoro), and eyes (vision model). The rest is Python.

It listens, sees the screen, speaks back, controls apps, writes code, drafts messages, manages Google Workspace, and when it doesn't know how to do something โ€” it writes its own plugin and learns.

No cloud dependency. No subscription. No data leaving the machine. MIT licensed.


7 Products. One System.

CODEC Core โ€” The Command Layer

Always-on voice assistant. Say "Hey CODEC" or press F13 to activate. F18 for voice commands. F16 for text input.

50+ skills fire instantly: Google Calendar, Gmail, Drive, Docs, Sheets, Tasks, Keep, Chrome automation, web search, Hue lights, timers, Spotify, clipboard, terminal commands, and more. Most skills bypass the LLM entirely โ€” direct action, zero latency.

Vision Mouse Control โ€” See & Click

No other open-source voice assistant does this.

Say "Hey CODEC, click the Submit button" โ€” CODEC screenshots the screen, sends it to a local UI-specialist vision model (UI-TARS), gets back pixel coordinates, and moves the mouse to click. Fully voice-controlled. Works on any app. No accessibility API required โ€” pure vision.

StepWhat happensSpeed
1Whisper transcribes voice command~2s
2Target extracted from natural speechinstant
3Screenshot captured and downscaledinstant
4UI-TARS locates the element by pixel coordinates~4s
5pyautogui moves cursor and clicksinstant

"I'm on Cloudflare and can't find the SSL button โ€” click it for me." That works. CODEC strips the conversational noise, extracts "SSL button", and finds it on screen.

CODEC Dictate โ€” Hold, Speak, Paste

Hold a key. Say what you mean. Release. Text appears wherever the cursor is. If CODEC detects a message draft, it refines through the LLM โ€” grammar fixed, tone polished, meaning preserved. Works in every app on macOS. A free, open-source SuperWhisper replacement that runs entirely local.

CODEC Instant โ€” One Right-Click

Select any text, anywhere. Right-click. Eight AI services system-wide: Proofread, Elevate, Explain, Translate, Reply (with :tone syntax), Prompt, Read Aloud, Save. Powered by the local LLM.

CODEC Chat โ€” 250K Context + 12 Agent Crews

Full conversational AI. Long context. File uploads. Image analysis via vision model. Web search. Conversation history.

Plus 12 autonomous agent crews โ€” not single prompts, full multi-step workflows. Say "research the latest AI agent frameworks and write a report." Minutes later there's a formatted Google Doc in Drive with sources, images, and recommendations.

CrewOutput
Deep Research10,000-word illustrated report โ†’ Google Docs
Daily BriefingMorning news + calendar โ†’ Google Docs
Competitor AnalysisSWOT + positioning โ†’ Google Docs
Trip PlannerFull itinerary โ†’ Google Docs
Email HandlerTriage inbox, draft replies
Social MediaPosts for Twitter, LinkedIn, Instagram
Code ReviewBugs + security + clean code
Data AnalysisTrends + insights report
Content WriterBlog posts, articles, copy
Meeting SummarizerAction items from transcripts
Invoice GeneratorProfessional invoices
Custom AgentDefine your own role, tools, task

Schedule any crew: "Run competitor analysis every Monday at 9am"

The multi-agent framework is under 800 lines. Zero dependencies. No CrewAI. No LangChain.

CODEC Vibe โ€” AI Coding IDE + Skill Forge

Split-screen in the browser. Monaco editor on the left (same engine as VS Code). AI chat on the right. Describe what's needed โ€” CODEC writes it, click Apply, run it, live preview in browser.

Skill Forge takes it further: describe a new capability in plain English, CODEC converts it into a working plugin. The framework writes its own extensions.

CODEC Voice โ€” Live Voice Calls

Real-time voice-to-voice conversations with the AI. WebSocket pipeline โ€” no Pipecat, no external dependencies. Call CODEC from a phone, talk naturally, and mid-call say "check my screen" โ€” it takes a screenshot, analyzes it, and speaks the result back.

Full transcript saved to memory. Every conversation becomes searchable context for future sessions.

CODEC Overview โ€” Dashboard Anywhere

Private dashboard accessible from any device, anywhere. Cloudflare Tunnel or Tailscale VPN โ€” no port forwarding, no third-party relay. Send commands, view the screen, launch voice calls, manage agents โ€” all from a browser.


Screenshots

<p align="center"> <img src="docs/screenshots/quick-chat.png" alt="Quick Chat" width="720"/><br/> <em>Chat โ€” ask anything, drag & drop files, full conversation history</em> </p> <p align="center"> <img src="docs/screenshots/chat-analysis.png" alt="Chat with File Analysis" width="720"/><br/> <em>Deep Chat โ€” upload files, select agents, get structured analysis</em> </p> <p align="center"> <img src="docs/screenshots/voice-call.png" alt="Voice Call" width="720"/><br/> <em>Voice Call โ€” real-time conversation with live transcript</em> </p> <p align="center"> <img src="docs/screenshots/vibe-code.png" alt="Vibe Code" width="720"/><br/> <em>Vibe Code โ€” describe what you want, get working code with live preview</em> </p> <p align="center"> <img src="docs/screenshots/deep-research.png" alt="Deep Research Report" width="720"/><br/> <em>Deep Research โ€” multi-agent reports delivered to Google Docs</em> </p> <p align="center"> <img src="docs/screenshots/tasks.png" alt="Tasks & Schedules" width="720"/><br/> <em>Scheduled automations โ€” morning briefings, competitor analysis, on cron</em> </p> <details> <summary><strong>More screenshots</strong></summary> <br/> <p align="center"> <img src="docs/screenshots/settings.png" alt="Settings" width="720"/><br/> <em>Settings โ€” LLM, TTS, STT, hotkeys, wake word configuration</em> </p> <p align="center"> <img src="docs/screenshots/agent-options.png" alt="Agent Options" width="420"/><br/> <em>12 specialized agent crews</em> </p> <p align="center"> <img src="docs/screenshots/login-auth.png" alt="Authentication" width="320"/><br/> <em>Touch ID + PIN + 2FA authentication</em> </p> <p align="center"> <img src="docs/screenshots/right-click-menu.png" alt="Right-Click Menu" width="300"/><br/> <em>Right-click integration โ€” CODEC in every app</em> </p> <p align="center"> <img src="docs/screenshots/terminal.png" alt="Terminal" width="400"/><br/> <em>50+ skills loaded at startup</em> </p> </details>

What Makes CODEC Different

CapabilityCODECSiri / Alexa / GoogleChatGPT / Claude
Controls the computerFull macOS controlLimited smart homeNo
Reads the screenVision modelNoNo
Clicks UI elements by voiceVision + mouse controlNoNo (Cloud Computer Use only)
Runs 100% localYes โ€” all models on deviceNoNo
Voice-to-voice callsWebSocket, real-timeYes but cloudYes but cloud
Multi-agent workflows12 crews, local LLMNoLimited
Right-click AI services8 system-wide servicesNoNo
Writes its own pluginsSkill ForgeNoNo
Open sourceMITNoNo

What CODEC replaced with native code:

BeforeAfter
PipecatCODEC Voice (own WebSocket pipeline)
CrewAI + LangChainCODEC Agents (795 lines, zero dependencies)
SuperWhisperCODEC Dictate (free, open source)
Cursor / WindsurfCODEC Vibe (Monaco + AI + Skill Forge)
Google Assistant / SiriCODEC Core (actually controls the computer)
GrammarlyCODEC Instant (right-click services via local LLM)
ChatGPTCODEC Chat (250K context, fully local)
Cloud LLM APIsLocal stack (Qwen + Whisper + Kokoro + Vision)
Vector databasesFTS5 SQLite (simpler, faster, private)

External services: DuckDuckGo for web search. Cloudflare free tier for the tunnel (or Tailscale). Everything else runs on local hardware.


Quick Start

git clone https://github.com/AVADSA25/codec.git
cd codec
./install.sh

The setup wizard handles everything in 9 steps: LLM, voice, vision, hotkeys, Google OAuth, remote access, and more.

Requirements:

  • macOS Ventura or later
  • Python 3.10+
  • An LLM (Ollama, LM Studio, MLX, or any OpenAI-compatible API)
  • Whisper for voice input, Kokoro for voice output, a vision model for screen reading

Supported LLMs

ModelHow to run
Qwen 3.5 35B (recommended)mlx-lm.server --model mlx-community/Qwen3.5-35B-A3B-4bit
Llama 3.3 70Bmlx-lm.server --model mlx-community/Llama-3.3-70B-Instruct-4bit
Mistral 24Bmlx-lm.server --model mlx-community/Mistral-Small-3.1-24B-Instruct-2503-4bit
Gemma 3 27Bmlx-lm.server --model mlx-community/gemma-3-27b-it-4bit
GPT-4o (cloud)"llm_url": "https://api.openai.com/v1"
Claude (cloud)OpenAI-compatible proxy
Ollama (any model)"llm_url": "http://localhost:11434/v1"

Configure in ~/.codec/config.json:

{
  "llm_url": "http://localhost:8081/v1",
  "model": "mlx-community/Qwen3.5-35B-A3B-4bit"
}

Keyboard Shortcuts

KeyAction
F13Toggle CODEC ON/OFF
F18 (hold)Record voice โ†’ release to send
F18 (double-tap)PTT Lock โ€” hands-free recording
F16Text input dialog
* *Screenshot + AI analysis
+ +Document mode
Camera iconLive webcam PIP โ€” drag around, snapshot anytime
Select text โ†’ right-click8 AI services in context menu

Laptop (F1-F12): F5 = toggle, F8 = voice, F9 = text input. Run python3 setup_codec.py โ†’ select "Laptop / Compact" in Step 4.

Custom shortcuts in ~/.codec/config.json. Restart after changes: pm2 restart open-codec


Privacy & Security

5-layer security stack:

LayerProtection
NetworkCloudflare Zero Trust tunnel or Tailscale VPN, CORS restricted origins
AuthTouch ID + PIN + TOTP 2FA, timing-safe token comparison
EncryptionAES-256-GCM + ECDH P-256 key exchange, per-session keys
ExecutionSubprocess isolation, resource limits (512MB RAM, 120s CPU), command blocklist, human review gate
DataLocal SQLite, parameterized queries, FTS5 full-text search โ€” searchable, private, yours

Every conversation is stored locally in SQLite with FTS5 full-text search. No cloud sync. No analytics. No telemetry.


MCP Server โ€” CODEC Inside Claude, Cursor, VS Code

CODEC exposes tools as an MCP server. Any MCP-compatible client can invoke CODEC skills directly:

{
  "mcpServers": {
    "codec": {
      "command": "python3",
      "args": ["/path/to/codec-repo/codec_mcp.py"]
    }
  }
}

Then in Claude Desktop: "Use CODEC to check my calendar for tomorrow."

Skills opt-in to MCP exposure with SKILL_MCP_EXPOSE = True.


Debugging & Development

Recommended tools:

ToolHow it helps
Claude CodeTerminal AI โ€” reads the full codebase, runs commands, fixes errors in context
CursorAI IDE โ€” navigate CODEC's 60+ files, refactor, debug with full project awareness
WindsurfAI IDE โ€” strong at multi-file reasoning
AntigravityAI debugging assistant โ€” paste errors, get fixes with codebase context

Quick debug commands:

# Check all services
pm2 list

# Check specific service logs
pm2 logs open-codec --lines 30 --nostream        # Main CODEC process
pm2 logs codec-dashboard --lines 30 --nostream    # Dashboard API
pm2 logs whisper-stt --lines 10 --nostream        # Speech-to-text
pm2 logs kokoro-82m --lines 10 --nostream         # Text-to-speech

# Verify LLM is responding
curl -s http://localhost:8081/v1/models | python3 -m json.tool

# Verify dashboard is up
curl -s http://localhost:8090/health

# Restart everything
pm2 restart all

# Full health check
python3 -c "from codec_config import *; print('Config OK')"

Common issues:

<details> <summary><strong>Keys don't work</strong></summary>
  • macOS stealing F-keys? System Settings โ†’ Keyboard โ†’ "Use F1, F2, etc. as standard function keys"
  • After config change: pm2 restart open-codec
</details> <details> <summary><strong>Wake word doesn't trigger</strong></summary>
  • Check Whisper: pm2 logs whisper-stt --lines 5 --nostream
  • Check mic permission: System Settings โ†’ Privacy โ†’ Microphone
  • Say "Hey CODEC" clearly โ€” 3 distinct syllables
</details> <details> <summary><strong>No voice output</strong></summary>
  • Check Kokoro TTS: curl http://localhost:8085/v1/models
  • Fallback: "tts_engine": "say" in config.json (macOS built-in)
  • Disable: "tts_engine": "none"
</details> <details> <summary><strong>Dashboard not loading</strong></summary>
  • Check: curl http://localhost:8090/health
  • Restart: pm2 restart codec-dashboard
  • Remote via Cloudflare: pm2 logs cloudflared --lines 3 --nostream
  • Remote via Tailscale: access CODEC at http://100.x.x.x:8090
</details> <details> <summary><strong>Agents timing out</strong></summary>
  • First run takes 2-5 min โ€” multi-step research with multiple searches
  • Check: pm2 logs codec-dashboard --lines 30 --nostream | grep -i agent
  • Agents run as background jobs โ€” no Cloudflare timeout
</details>

Project Structure

codec.py              โ€” Entry point
codec_config.py       โ€” Configuration + transcript cleaning
codec_keyboard.py     โ€” Keyboard listener, PTT lock, wake word
codec_dispatch.py     โ€” Skill matching and dispatch (with fallback)
codec_agent.py        โ€” LLM session builder
codec_agents.py       โ€” Multi-agent crew framework (12 crews)
codec_voice.py        โ€” WebSocket voice pipeline
codec_voice.html      โ€” Voice call UI
codec_dashboard.py    โ€” Web API + dashboard (60+ endpoints)
codec_dashboard.html  โ€” Dashboard UI
codec_chat.html       โ€” Chat UI
codec_vibe.html       โ€” Vibe Code IDE
codec_auth.html       โ€” Authentication (Touch ID + PIN + TOTP 2FA)
codec_textassist.py   โ€” 8 right-click services
codec_search.py       โ€” DuckDuckGo + Serper search
codec_mcp.py          โ€” MCP server
codec_memory.py       โ€” FTS5 memory search
codec_heartbeat.py    โ€” Health monitoring + task auto-execution
codec_scheduler.py    โ€” Cron-like agent scheduling
codec_marketplace.py  โ€” Skill marketplace CLI
codec_overlays.py     โ€” AppKit overlay notifications (fullscreen compatible)
ax_bridge/            โ€” Swift AX accessibility bridge
swift-overlay/        โ€” SwiftUI status bar app
skills/               โ€” 50+ built-in skills (incl. vision mouse control)
tests/                โ€” 212+ pytest tests
install.sh            โ€” One-line installer
setup_codec.py        โ€” Setup wizard (9 steps)

What's Coming

  • Linux support
  • Windows via WSL
  • Multi-machine sync (skills + memory across devices)
  • iOS app (dictation + remote dashboard)
  • Streaming voice responses (first token plays while rest generates)
  • Multi-LLM routing (fast model for simple, strong model for complex)

Contributing

All skill contributions welcome. 50+ built-in, marketplace growing.

git clone https://github.com/AVADSA25/codec.git
cd codec && ./install.sh
python3 -m pytest   # all tests must pass

See CONTRIBUTING.md.


Support the Project

If CODEC saves you time:


Professional Setup

Need CODEC configured for a business, integrated with existing tools, or deployed across a team?

Contact AVA Digital for professional setup and custom skill development.


<p align="center"> Star it. Clone it. Rip it apart. Make it yours. </p> <p align="center"> Built by <a href="https://avadigital.ai">AVA Digital LLC</a> ยท MIT License </p>

Global Ranking

7.4
Trust ScoreMCPHub Index

Based on codebase health & activity.

Manual Config

{ "mcpServers": { "avadsa25-codec": { "command": "npx", "args": ["avadsa25-codec"] } } }