Enhanced fork of the official ElevenLabs MCP server with additional conversational AI features including conversation history and transcript retrieval.
This enhanced version is developed and maintained by Boris Djordjevic and the 199 Longevity team.
This enhanced version adds critical conversational AI features missing from the original:
-
🎯 Smart Voice Defaults:
search_voices()
now returns common working voices instantly - 📚 Educational Error Messages: Errors guide AI agents to success with examples
-
🔧 Fixed v3 Proxy:
text_to_dialogue
now properly uses v3 proxy (no more 403 errors!) - 💡 Clear Tool Guidance: No more confusion about single vs multi-speaker tools
- 🎤 Accurate v3 Voice IDs: All 20 v3-optimized voices now have correct IDs and descriptions
- 🏯 Auto-Split Long Dialogues: Automatically splits dialogues over 3000 chars into multiple files
- 🎯 Auto-Adjust Stability: Invalid stability values auto-round to nearest valid option (0.0, 0.5, 1.0)
- 🏷️ Smart Tag Simplification: Complex tags auto-convert to valid v3 tags for better quality
- ⏱️ Dynamic Timeouts: Prevents timeouts on complex dialogues by calculating appropriate wait times
-
🎭 Enhanced Expressiveness: Use the new v3 model with
model="v3"
parameter -
🎤 Audio Tags: Add emotions and sound effects like
[thoughtful]
,[crying]
,[laughing]
,[piano]
- 👥 Multi-Speaker Dialogue: Generate natural conversations between multiple speakers
- ✨ Dialogue Enhancement: Automatically enhance your dialogue with proper formatting and tags
- 🔐 v3 Proxy Support: Use v3 even without API access by enabling the built-in proxy (see v3 Proxy section)
- Conversation History: Retrieve full conversation details including transcripts
- 📝 Transcript Access: Get conversation transcripts in multiple formats (plain, timestamps, JSON)
- ⏳ Real-time Monitoring: Wait for ongoing conversations to complete and retrieve results
- 🔍 Conversation Search: List and filter conversations by agent, status, and more
- 🎨 Improved Formatting: Consistent formatting across all list operations
This is an enhanced fork of the official ElevenLabs Model Context Protocol (MCP) server that enables interaction with powerful Text to Speech and audio processing APIs. This server allows MCP clients like Claude Desktop, Cursor, Windsurf, OpenAI Agents and others to generate speech, clone voices, transcribe audio, manage conversational AI agents, and now retrieve conversation history.
No installation required! Just use npx:
npx elevenlabs-mcp-enhanced --api-key YOUR_API_KEY
Install once, use everywhere:
npm install -g elevenlabs-mcp-enhanced
elevenlabs-mcp-enhanced --api-key YOUR_API_KEY
Set your API key once:
export ELEVENLABS_API_KEY="your-api-key"
npx elevenlabs-mcp-enhanced
- Node.js 16+ (for npm/npx)
- Python 3.11+ (automatically managed by the npm package)
- ElevenLabs API Key - Get one at elevenlabs.io
- Get your API key from ElevenLabs. There is a free tier with 10k credits per month.
- Go to Claude > Settings > Developer > Edit Config > claude_desktop_config.json to include the following:
{
"mcpServers": {
"ElevenLabs": {
"command": "npx",
"args": ["elevenlabs-mcp-enhanced"],
"env": {
"ELEVENLABS_API_KEY": "<insert-your-api-key-here>"
}
}
}
}
That's it! No installation needed - npx will automatically download and run the server.
If you prefer the original Python installation:
- Get your API key from ElevenLabs.
- Install from GitHub:
pip install git+https://github.com/199-biotechnologies/elevenlabs-mcp-enhanced.git
- Configure Claude Desktop with:
{ "mcpServers": { "ElevenLabs": { "command": "python", "args": ["-m", "elevenlabs_mcp"], "env": { "ELEVENLABS_API_KEY": "<insert-your-api-key-here>" } } } }
If you're using Windows, you will have to enable "Developer Mode" in Claude Desktop to use the MCP server. Click "Help" in the hamburger menu at the top left and select "Enable Developer Mode".
For other clients like Cursor and Windsurf, you can run the server directly:
npx elevenlabs-mcp-enhanced --api-key YOUR_API_KEY
pip install elevenlabs-mcp
-
python -m elevenlabs_mcp --api-key={{PUT_YOUR_API_KEY_HERE}} --print
to get the configuration. Paste it into appropriate configuration directory specified by your MCP client.
That's it. Your MCP client can now interact with ElevenLabs through these tools:
Try asking Claude:
- "Create an AI agent that speaks like a film noir detective and can answer questions about classic movies"
- "Generate three voice variations for a wise, ancient dragon character, then I will choose my favorite voice to add to my voice library"
- "Convert this recording of my voice to sound like a medieval knight"
- "Create a soundscape of a thunderstorm in a dense jungle with animals reacting to the weather"
- "Turn this speech into text, identify different speakers, then convert it back using unique voices for each person"
🎯 DECISION TREE:
-
Single speaker? → Use
text_to_speech
withmodel="v3"
-
Multiple speakers? → Use
text_to_dialogue
(automatically v3) -
Need tag examples? → Call
fetch_v3_tags()
first
📋 RECOMMENDED WORKFLOW FOR AI:
1. User: "Create an emotional story with sound effects"
2. AI: fetch_v3_tags() → Gets list of available tags
3. AI: search_voices("v3") → Gets v3-optimized voices
4. AI: text_to_dialogue(...) → Creates the story
Single Speaker Examples (text_to_speech):
- "Generate: '[thoughtful] The universe is vast... [piano] ...and full of mysteries.'"
- "Create narration with: '[whispering] Secret message [footsteps] [door creaking]'"
Multi-Speaker Examples (text_to_dialogue - ALWAYS v3):
# Simple conversation
inputs = [
{"text": "How are you?", "voice_name": "James"},
{"text": "I'm great!", "voice_name": "Jane"}
]
# With emotion tags
inputs = [
{"text": "[excited] I found treasure!", "voice_name": "James"},
{"text": "[skeptical] Really? [pause] Where?", "voice_name": "Jane"}
]
- Stability: MUST be 0.0, 0.5, or 1.0 (no other values!)
- Best voices: James, Jane, Sarah, Mark, etc. (search "v3" to find them)
- Always check fetch_v3_tags() for available audio tags
With the enhanced conversation tools, you can now:
- "Get the conversation transcript from conversation ID abc123" (automatically waits for completion)
- "List all conversations from my agent and show me the completed ones"
- "Get conversation xyz789 immediately without waiting" (set wait_for_completion=false)
- "Show me all conversations in JSON format with timestamps"
- "Get the conversation history including analysis data"
Note: The get_conversation
tool now waits for conversations to complete by default (up to 5 minutes), ensuring you always get the full transcript.
You can add the ELEVENLABS_MCP_BASE_PATH
environment variable to the claude_desktop_config.json
to specify the base path MCP server should look for and output files specified with relative paths.
The v3 model is currently in alpha and requires special access. If you have access through the ElevenLabs website but not through the API, you can use the built-in proxy:
- Enable the proxy by adding these environment variables to your MCP config:
{
"mcpServers": {
"ElevenLabs": {
"command": "npx",
"args": ["elevenlabs-mcp-enhanced@latest"],
"env": {
"ELEVENLABS_API_KEY": "your_api_key",
"ELEVENLABS_EMAIL": "your_email@example.com",
"ELEVENLABS_PASSWORD": "your_password",
"ELEVENLABS_V3_PROXY": "true"
}
}
}
}
-
How it works:
- The proxy automatically starts when you use
model="v3"
- It authenticates with your ElevenLabs account using Firebase
- Maintains session tokens and refreshes them automatically
- Forwards v3 requests with proper authentication
- The proxy automatically starts when you use
-
Security notes:
- Credentials are only used locally for authentication
- The proxy runs on localhost (port 8123 by default)
- Tokens are refreshed automatically before expiry
-
Limitations:
- Requires valid ElevenLabs account with v3 web access
- Proxy must authenticate through Firebase (same as web login)
- Not needed once ElevenLabs releases public v3 API access
If you want to contribute or run from source:
- Clone the repository:
git clone https://github.com/elevenlabs/elevenlabs-mcp
cd elevenlabs-mcp
- Create a virtual environment and install dependencies using uv:
uv venv
source .venv/bin/activate
uv pip install -e ".[dev]"
- Copy
.env.example
to.env
and add your ElevenLabs API key:
cp .env.example .env
# Edit .env and add your API key
- Run the tests to make sure everything is working:
./scripts/test.sh
# Or with options
./scripts/test.sh --verbose --fail-fast
-
Install the server in Claude Desktop:
mcp install elevenlabs_mcp/server.py
-
Debug and test locally with MCP Inspector:
mcp dev elevenlabs_mcp/server.py
Logs when running with Claude Desktop can be found at:
-
Windows:
%APPDATA%\Claude\logs\mcp-server-elevenlabs.log
-
macOS:
~/Library/Logs/Claude/mcp-server-elevenlabs.log
Certain ElevenLabs API operations, like voice design and audio isolation, can take a long time to resolve. When using the MCP inspector in dev mode, you might get timeout errors despite the tool completing its intended task.
This shouldn't occur when using a client like Claude.
If you encounter the error "MCP ElevenLabs: spawn uvx ENOENT", confirm its absolute path by running this command in your terminal:
which uvx
Once you obtain the absolute path (e.g., /usr/local/bin/uvx
), update your configuration to use that path (e.g., "command": "/usr/local/bin/uvx"
). This ensures that the correct executable is referenced.
- Boris Djordjevic - Lead Developer
- 199 Longevity Team - Development and Testing
- Jacek Duszenko - jacek@elevenlabs.io
- Paul Asjes - paul.asjes@elevenlabs.io
- Louis Jordan - louis@elevenlabs.io
- Luke Harries - luke@elevenlabs.io
This enhanced fork builds upon the excellent foundation created by the ElevenLabs team, adding critical conversational AI features for improved agent interaction and monitoring.
This project maintains the same MIT license as the original ElevenLabs MCP server. See LICENSE for details.