WellSaid MCP Server
Guide to walk through the end-to-end process of usin ghte WellSaid MCP server
Overview
The WellSaid MCP Server lets you connect AI clients (such as Claude Desktop) directly to the WellSaid API. Once installed, your AI client can generate speech, discover voices, and manage advanced audio features without writing any custom integration code.
This guide covers installation, configuration, and common usage patterns.
Prerequisites
Before getting started, make sure you have:
- A WellSaid API key, available in your WellSaid account
- An MCP-compatible AI client (e.g., Claude Desktop)
- uv installed — this tool manages the MCP server environment
- The WellSaid MCP Server installed in your LLM client
Installation
Option 1: From Package
If you only need the server for use with Claude Desktop, install from the published package:
uvx wellsaid-mcp-setup-claude-desktop
Option 2: Local Development
If you want to run the MCP server locally for development:
# Create and activate a virtual environment
uv venv
# Install the server in editable mode
uv pip install --editable .
Configuration with Claude Desktop
Once installed, you’ll need to register the WellSaid MCP server with Claude Desktop.
uv run wellsaid-mcp-setup-claude-desktop --dev .
When prompted, enter your WellSaid API key.
After configuration, restart Claude Desktop to apply the changes.
Available Tools
Once connected, your AI client can call WellSaid functions through natural prompts.
Voice Discovery
get_avatar_criteria
– Lists available filters for voices
Example: “Show me all available voice filtering options”get_avatars
– Search for voices matching specific traits
Example: “Find a confident, professional female voice with US accent”get_avatar_characteristics
– Lists available voice characteristics
Example: “What voice characteristics can I choose from?”
Speech Generation
text_to_speech
– Convert text into speech with a selected voice
Example: “Generate speech for ‘Welcome to our training program’ using speaker ID 145”create_multiple_clips_and_combine
– Generate multi-speaker or long-form audio
Example: “Create a dialogue where Speaker 89 says ‘Hello there’ and Speaker 76 replies ‘Good morning’”
Advanced Voice Controls (Caruso Model)
Adjust_pitch
– Modify pitch (-250 to +500)Adjust_tempo
– Change speaking speed (0.5 to 2.5)Adjust_loudness
– Control loudness (-20 to +10)Apply_respelling
– Override pronunciationsValidate_AI_Director_tags
– Validate markup formatting
Example Workflows
Training Content
“Find a clear, informative voice for eLearning, then generate speech at a slower tempo so it's easy to comprehend:
‘Today we’ll learn about machine learning algorithms.’”
Marketing
“Use an upbeat voice, increase pitch by 50, and generate a script that has a sales pitch for our new product"
Character Dialogue
“Select two characters to have a conversation the first says ‘Hello,’ and the second will reply ‘Hi there,’ with short natural pauses in the conversation.”
Multilingual
“Find a Spanish voice with Mexico accent and a French voice to say the following line in each language:
‘Welcome to our new program.’”
Conclusion
The WellSaid MCP Server makes it easy to:
- Discover and select the right voice for your project
- Generate professional AI-powered audio directly in your AI client
- Control pitch, pacing, and pronunciation with precision
With this integration, you get the full power of the WellSaid API — without having to write API calls manually.
Updated 1 day ago