Skip to main content
Glama
elevenlabs

ElevenLabs MCP Server

Official
by elevenlabs

text_to_speech

Convert text to speech using ElevenLabs' AI voices, saving audio files to your desktop with customizable voice parameters and output formats.

Instructions

Convert text to speech with a given voice. Saves output file to directory (default: $HOME/Desktop).

Only one of voice_id or voice_name can be provided. If none are provided, the default voice will be used.

⚠️ COST WARNING: This tool makes an API call to ElevenLabs which may incur costs. Only use when explicitly requested by the user.

 Args:
    text (str): The text to convert to speech.
    voice_name (str, optional): The name of the voice to use.
    model_id (str, optional): The model ID to use for speech synthesis. Options include:
        - eleven_multilingual_v2: High quality multilingual model (29 languages)
        - eleven_flash_v2_5: Fastest model with ultra-low latency (32 languages)
        - eleven_turbo_v2_5: Balanced quality and speed (32 languages)
        - eleven_flash_v2: Fast English-only model
        - eleven_turbo_v2: Balanced English-only model
        - eleven_monolingual_v1: Legacy English model
        Defaults to eleven_multilingual_v2 or environment variable ELEVENLABS_MODEL_ID.
    stability (float, optional): Stability of the generated audio. Determines how stable the voice is and the randomness between each generation. Lower values introduce broader emotional range for the voice. Higher values can result in a monotonous voice with limited emotion. Range is 0 to 1.
    similarity_boost (float, optional): Similarity boost of the generated audio. Determines how closely the AI should adhere to the original voice when attempting to replicate it. Range is 0 to 1.
    style (float, optional): Style of the generated audio. Determines the style exaggeration of the voice. This setting attempts to amplify the style of the original speaker. It does consume additional computational resources and might increase latency if set to anything other than 0. Range is 0 to 1.
    use_speaker_boost (bool, optional): Use speaker boost of the generated audio. This setting boosts the similarity to the original speaker. Using this setting requires a slightly higher computational load, which in turn increases latency.
    speed (float, optional): Speed of the generated audio. Controls the speed of the generated speech. Values range from 0.7 to 1.2, with 1.0 being the default speed. Lower values create slower, more deliberate speech while higher values produce faster-paced speech. Extreme values can impact the quality of the generated speech. Range is 0.7 to 1.2.
    output_directory (str, optional): Directory where files should be saved (only used when saving files).
        Defaults to $HOME/Desktop if not provided.
    language: ISO 639-1 language code for the voice.
    output_format (str, optional): Output format of the generated audio. Formatted as codec_sample_rate_bitrate. So an mp3 with 22.05kHz sample rate at 32kbs is represented as mp3_22050_32. MP3 with 192kbps bitrate requires you to be subscribed to Creator tier or above. PCM with 44.1kHz sample rate requires you to be subscribed to Pro tier or above. Note that the μ-law format (sometimes written mu-law, often approximated as u-law) is commonly used for Twilio audio inputs.
        Defaults to "mp3_44100_128". Must be one of:
        mp3_22050_32
        mp3_44100_32
        mp3_44100_64
        mp3_44100_96
        mp3_44100_128
        mp3_44100_192
        pcm_8000
        pcm_16000
        pcm_22050
        pcm_24000
        pcm_44100
        ulaw_8000
        alaw_8000
        opus_48000_32
        opus_48000_64
        opus_48000_96
        opus_48000_128
        opus_48000_192

Returns:
    Text content with file path or MCP resource with audio data, depending on output mode.

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
textYes
voice_nameNo
output_directoryNo
voice_idNo
stabilityNo
similarity_boostNo
styleNo
use_speaker_boostNo
speedNo
languageNoen
output_formatNomp3_44100_128
model_idNo

Implementation Reference

  • The handler function for the 'text_to_speech' tool. Converts input text to speech audio using the ElevenLabs API, handles voice selection, model choice, voice settings, generates audio, and returns output based on configured mode (files, resources, or both).
    def text_to_speech(
        text: str,
        voice_name: str | None = None,
        output_directory: str | None = None,
        voice_id: str | None = None,
        stability: float = 0.5,
        similarity_boost: float = 0.75,
        style: float = 0,
        use_speaker_boost: bool = True,
        speed: float = 1.0,
        language: str = "en",
        output_format: str = "mp3_44100_128",
        model_id: str | None = None,
    ) -> Union[TextContent, EmbeddedResource]:
        if text == "":
            make_error("Text is required.")
    
        if voice_id is not None and voice_name is not None:
            make_error("voice_id and voice_name cannot both be provided.")
    
        voice = None
        if voice_id is not None:
            voice = client.voices.get(voice_id=voice_id)
        elif voice_name is not None:
            voices = client.voices.search(search=voice_name)
            if len(voices.voices) == 0:
                make_error("No voices found with that name.")
            voice = next((v for v in voices.voices if v.name == voice_name), None)
            if voice is None:
                make_error(f"Voice with name: {voice_name} does not exist.")
    
        voice_id = voice.voice_id if voice else DEFAULT_VOICE_ID
    
        output_path = make_output_path(output_directory, base_path)
        output_file_name = make_output_file("tts", text, "mp3")
    
        if model_id is None:
            model_id = (
                "eleven_flash_v2_5"
                if language in ["hu", "no", "vi"]
                else "eleven_multilingual_v2"
            )
    
        audio_data = client.text_to_speech.convert(
            text=text,
            voice_id=voice_id,
            model_id=model_id,
            output_format=output_format,
            voice_settings={
                "stability": stability,
                "similarity_boost": similarity_boost,
                "style": style,
                "use_speaker_boost": use_speaker_boost,
                "speed": speed,
            },
        )
        audio_bytes = b"".join(audio_data)
    
        # Handle different output modes
        success_message = f"Success. File saved as: {{file_path}}. Voice used: {voice.name if voice else DEFAULT_VOICE_ID}"
        return handle_output_mode(
            audio_bytes, output_path, output_file_name, output_mode, success_message
        )
  • Registration of the 'text_to_speech' tool using the @mcp.tool decorator, including detailed description, parameters, and usage instructions.
    @mcp.tool(
        description=f"""Convert text to speech with a given voice. {get_output_mode_description(output_mode)}.
        
        Only one of voice_id or voice_name can be provided. If none are provided, the default voice will be used.
    
        ⚠️ COST WARNING: This tool makes an API call to ElevenLabs which may incur costs. Only use when explicitly requested by the user.
    
         Args:
            text (str): The text to convert to speech.
            voice_name (str, optional): The name of the voice to use.
            model_id (str, optional): The model ID to use for speech synthesis. Options include:
                - eleven_multilingual_v2: High quality multilingual model (29 languages)
                - eleven_flash_v2_5: Fastest model with ultra-low latency (32 languages)
                - eleven_turbo_v2_5: Balanced quality and speed (32 languages)
                - eleven_flash_v2: Fast English-only model
                - eleven_turbo_v2: Balanced English-only model
                - eleven_monolingual_v1: Legacy English model
                Defaults to eleven_multilingual_v2 or environment variable ELEVENLABS_MODEL_ID.
            stability (float, optional): Stability of the generated audio. Determines how stable the voice is and the randomness between each generation. Lower values introduce broader emotional range for the voice. Higher values can result in a monotonous voice with limited emotion. Range is 0 to 1.
            similarity_boost (float, optional): Similarity boost of the generated audio. Determines how closely the AI should adhere to the original voice when attempting to replicate it. Range is 0 to 1.
            style (float, optional): Style of the generated audio. Determines the style exaggeration of the voice. This setting attempts to amplify the style of the original speaker. It does consume additional computational resources and might increase latency if set to anything other than 0. Range is 0 to 1.
            use_speaker_boost (bool, optional): Use speaker boost of the generated audio. This setting boosts the similarity to the original speaker. Using this setting requires a slightly higher computational load, which in turn increases latency.
            speed (float, optional): Speed of the generated audio. Controls the speed of the generated speech. Values range from 0.7 to 1.2, with 1.0 being the default speed. Lower values create slower, more deliberate speech while higher values produce faster-paced speech. Extreme values can impact the quality of the generated speech. Range is 0.7 to 1.2.
            output_directory (str, optional): Directory where files should be saved (only used when saving files).
                Defaults to $HOME/Desktop if not provided.
            language: ISO 639-1 language code for the voice.
            output_format (str, optional): Output format of the generated audio. Formatted as codec_sample_rate_bitrate. So an mp3 with 22.05kHz sample rate at 32kbs is represented as mp3_22050_32. MP3 with 192kbps bitrate requires you to be subscribed to Creator tier or above. PCM with 44.1kHz sample rate requires you to be subscribed to Pro tier or above. Note that the μ-law format (sometimes written mu-law, often approximated as u-law) is commonly used for Twilio audio inputs.
                Defaults to "mp3_44100_128". Must be one of:
                mp3_22050_32
                mp3_44100_32
                mp3_44100_64
                mp3_44100_96
                mp3_44100_128
                mp3_44100_192
                pcm_8000
                pcm_16000
                pcm_22050
                pcm_24000
                pcm_44100
                ulaw_8000
                alaw_8000
                opus_48000_32
                opus_48000_64
                opus_48000_96
                opus_48000_128
                opus_48000_192
    
        Returns:
            Text content with file path or MCP resource with audio data, depending on output mode.
        """
    )

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/elevenlabs/elevenlabs-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server