Skip to main content
Glama
elevenlabs

ElevenLabs MCP Server

Official
by elevenlabs

text_to_speech

Convert text to speech using ElevenLabs' AI voices, saving audio files to your desktop with customizable voice parameters and output formats.

Instructions

Convert text to speech with a given voice. Saves output file to directory (default: $HOME/Desktop).

Only one of voice_id or voice_name can be provided. If none are provided, the default voice will be used. ⚠️ COST WARNING: This tool makes an API call to ElevenLabs which may incur costs. Only use when explicitly requested by the user. Args: text (str): The text to convert to speech. voice_name (str, optional): The name of the voice to use. model_id (str, optional): The model ID to use for speech synthesis. Options include: - eleven_multilingual_v2: High quality multilingual model (29 languages) - eleven_flash_v2_5: Fastest model with ultra-low latency (32 languages) - eleven_turbo_v2_5: Balanced quality and speed (32 languages) - eleven_flash_v2: Fast English-only model - eleven_turbo_v2: Balanced English-only model - eleven_monolingual_v1: Legacy English model Defaults to eleven_multilingual_v2 or environment variable ELEVENLABS_MODEL_ID. stability (float, optional): Stability of the generated audio. Determines how stable the voice is and the randomness between each generation. Lower values introduce broader emotional range for the voice. Higher values can result in a monotonous voice with limited emotion. Range is 0 to 1. similarity_boost (float, optional): Similarity boost of the generated audio. Determines how closely the AI should adhere to the original voice when attempting to replicate it. Range is 0 to 1. style (float, optional): Style of the generated audio. Determines the style exaggeration of the voice. This setting attempts to amplify the style of the original speaker. It does consume additional computational resources and might increase latency if set to anything other than 0. Range is 0 to 1. use_speaker_boost (bool, optional): Use speaker boost of the generated audio. This setting boosts the similarity to the original speaker. Using this setting requires a slightly higher computational load, which in turn increases latency. speed (float, optional): Speed of the generated audio. Controls the speed of the generated speech. Values range from 0.7 to 1.2, with 1.0 being the default speed. Lower values create slower, more deliberate speech while higher values produce faster-paced speech. Extreme values can impact the quality of the generated speech. Range is 0.7 to 1.2. output_directory (str, optional): Directory where files should be saved (only used when saving files). Defaults to $HOME/Desktop if not provided. language: ISO 639-1 language code for the voice. output_format (str, optional): Output format of the generated audio. Formatted as codec_sample_rate_bitrate. So an mp3 with 22.05kHz sample rate at 32kbs is represented as mp3_22050_32. MP3 with 192kbps bitrate requires you to be subscribed to Creator tier or above. PCM with 44.1kHz sample rate requires you to be subscribed to Pro tier or above. Note that the μ-law format (sometimes written mu-law, often approximated as u-law) is commonly used for Twilio audio inputs. Defaults to "mp3_44100_128". Must be one of: mp3_22050_32 mp3_44100_32 mp3_44100_64 mp3_44100_96 mp3_44100_128 mp3_44100_192 pcm_8000 pcm_16000 pcm_22050 pcm_24000 pcm_44100 ulaw_8000 alaw_8000 opus_48000_32 opus_48000_64 opus_48000_96 opus_48000_128 opus_48000_192 Returns: Text content with file path or MCP resource with audio data, depending on output mode.

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
textYes
voice_nameNo
output_directoryNo
voice_idNo
stabilityNo
similarity_boostNo
styleNo
use_speaker_boostNo
speedNo
languageNoen
output_formatNomp3_44100_128
model_idNo

Implementation Reference

  • The handler function for the 'text_to_speech' tool. Converts input text to speech audio using the ElevenLabs API, handles voice selection, model choice, voice settings, generates audio, and returns output based on configured mode (files, resources, or both).
    def text_to_speech( text: str, voice_name: str | None = None, output_directory: str | None = None, voice_id: str | None = None, stability: float = 0.5, similarity_boost: float = 0.75, style: float = 0, use_speaker_boost: bool = True, speed: float = 1.0, language: str = "en", output_format: str = "mp3_44100_128", model_id: str | None = None, ) -> Union[TextContent, EmbeddedResource]: if text == "": make_error("Text is required.") if voice_id is not None and voice_name is not None: make_error("voice_id and voice_name cannot both be provided.") voice = None if voice_id is not None: voice = client.voices.get(voice_id=voice_id) elif voice_name is not None: voices = client.voices.search(search=voice_name) if len(voices.voices) == 0: make_error("No voices found with that name.") voice = next((v for v in voices.voices if v.name == voice_name), None) if voice is None: make_error(f"Voice with name: {voice_name} does not exist.") voice_id = voice.voice_id if voice else DEFAULT_VOICE_ID output_path = make_output_path(output_directory, base_path) output_file_name = make_output_file("tts", text, "mp3") if model_id is None: model_id = ( "eleven_flash_v2_5" if language in ["hu", "no", "vi"] else "eleven_multilingual_v2" ) audio_data = client.text_to_speech.convert( text=text, voice_id=voice_id, model_id=model_id, output_format=output_format, voice_settings={ "stability": stability, "similarity_boost": similarity_boost, "style": style, "use_speaker_boost": use_speaker_boost, "speed": speed, }, ) audio_bytes = b"".join(audio_data) # Handle different output modes success_message = f"Success. File saved as: {{file_path}}. Voice used: {voice.name if voice else DEFAULT_VOICE_ID}" return handle_output_mode( audio_bytes, output_path, output_file_name, output_mode, success_message )
  • Registration of the 'text_to_speech' tool using the @mcp.tool decorator, including detailed description, parameters, and usage instructions.
    @mcp.tool( description=f"""Convert text to speech with a given voice. {get_output_mode_description(output_mode)}. Only one of voice_id or voice_name can be provided. If none are provided, the default voice will be used. ⚠️ COST WARNING: This tool makes an API call to ElevenLabs which may incur costs. Only use when explicitly requested by the user. Args: text (str): The text to convert to speech. voice_name (str, optional): The name of the voice to use. model_id (str, optional): The model ID to use for speech synthesis. Options include: - eleven_multilingual_v2: High quality multilingual model (29 languages) - eleven_flash_v2_5: Fastest model with ultra-low latency (32 languages) - eleven_turbo_v2_5: Balanced quality and speed (32 languages) - eleven_flash_v2: Fast English-only model - eleven_turbo_v2: Balanced English-only model - eleven_monolingual_v1: Legacy English model Defaults to eleven_multilingual_v2 or environment variable ELEVENLABS_MODEL_ID. stability (float, optional): Stability of the generated audio. Determines how stable the voice is and the randomness between each generation. Lower values introduce broader emotional range for the voice. Higher values can result in a monotonous voice with limited emotion. Range is 0 to 1. similarity_boost (float, optional): Similarity boost of the generated audio. Determines how closely the AI should adhere to the original voice when attempting to replicate it. Range is 0 to 1. style (float, optional): Style of the generated audio. Determines the style exaggeration of the voice. This setting attempts to amplify the style of the original speaker. It does consume additional computational resources and might increase latency if set to anything other than 0. Range is 0 to 1. use_speaker_boost (bool, optional): Use speaker boost of the generated audio. This setting boosts the similarity to the original speaker. Using this setting requires a slightly higher computational load, which in turn increases latency. speed (float, optional): Speed of the generated audio. Controls the speed of the generated speech. Values range from 0.7 to 1.2, with 1.0 being the default speed. Lower values create slower, more deliberate speech while higher values produce faster-paced speech. Extreme values can impact the quality of the generated speech. Range is 0.7 to 1.2. output_directory (str, optional): Directory where files should be saved (only used when saving files). Defaults to $HOME/Desktop if not provided. language: ISO 639-1 language code for the voice. output_format (str, optional): Output format of the generated audio. Formatted as codec_sample_rate_bitrate. So an mp3 with 22.05kHz sample rate at 32kbs is represented as mp3_22050_32. MP3 with 192kbps bitrate requires you to be subscribed to Creator tier or above. PCM with 44.1kHz sample rate requires you to be subscribed to Pro tier or above. Note that the μ-law format (sometimes written mu-law, often approximated as u-law) is commonly used for Twilio audio inputs. Defaults to "mp3_44100_128". Must be one of: mp3_22050_32 mp3_44100_32 mp3_44100_64 mp3_44100_96 mp3_44100_128 mp3_44100_192 pcm_8000 pcm_16000 pcm_22050 pcm_24000 pcm_44100 ulaw_8000 alaw_8000 opus_48000_32 opus_48000_64 opus_48000_96 opus_48000_128 opus_48000_192 Returns: Text content with file path or MCP resource with audio data, depending on output mode. """ )

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/elevenlabs/elevenlabs-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server