📢 Blabber-MCP 🗣️
An MCP server that gives your LLMs a voice using OpenAI's Text-to-Speech API! 🔊
✨ Features
- Text-to-Speech: Converts input text into high-quality spoken audio.
- Voice Selection: Choose from various OpenAI voices (
alloy
,echo
,fable
,onyx
,nova
,shimmer
). - Model Selection: Use standard (
tts-1
) or high-definition (tts-1-hd
) models. - Format Options: Get audio output in
mp3
,opus
,aac
, orflac
. - File Saving: Saves the generated audio to a local file.
- Optional Playback: Automatically play the generated audio using a configurable system command.
- Configurable Defaults: Set a default voice via configuration.
🔧 Configuration
To use this server, you need to add its configuration to your MCP client's settings file (e.g., mcp_settings.json
).
- Get OpenAI API Key: You need an API key from OpenAI.
- Add to MCP Settings: Add the following block to the
mcpServers
object in your settings file, replacing"YOUR_OPENAI_API_KEY"
with your actual key.
Important: Make sure the args
path points to the correct location of the build/index.js
file within your blabber-mcp
project directory. Use the full absolute path.
🚀 Usage
Once configured and running, you can use the text_to_speech
tool via your MCP client.
Tool: text_to_speech
Server: blabber-mcp
(or the key you used in the config)
Arguments:
input
(string, required): The text to synthesize.voice
(string, optional): The voice to use (alloy
,echo
,fable
,onyx
,nova
,shimmer
). Defaults to theDEFAULT_TTS_VOICE
set in config, ornova
.model
(string, optional): The model (tts-1
,tts-1-hd
). Defaults totts-1
.response_format
(string, optional): Audio format (mp3
,opus
,aac
,flac
). Defaults tomp3
.play
(boolean, optional): Set totrue
to automatically play the audio after saving. Defaults tofalse
.
Example Tool Call (with playback):
Output:
The tool saves the audio file to the output/
directory within the blabber-mcp
project folder and returns a JSON response like this:
📜 License
This project is licensed under the MIT License - see the LICENSE file for details.
🕒 Changelog
See the CHANGELOG.md file for details on version history.
This server cannot be installed
hybrid server
The server is able to function both locally and remotely, depending on the configuration or use case.
An MCP server that enables LLMs to generate spoken audio from text using OpenAI's Text-to-Speech API, supporting various voices, models, and audio formats.
Related MCP Servers
- -securityAlicense-qualityEnables recording audio from a microphone and transcribing it using OpenAI's Whisper model. Works as both a standalone MCP server and a Goose AI agent extension.Last updated -4PythonMIT License
- AsecurityAlicenseAqualityA MCP server that enables transcription of audio files using OpenAI's Speech-to-Text API, with support for multiple languages and file saving options.Last updated -12JavaScriptMIT License
- AsecurityAlicenseAqualityMCP server for Synthesizer V AI Vocal Studio, which allows LLMs to create/edit vocal tracks e.g. adding lyrics to the melody.Last updated -6Apache 2.0
ElevenLabs MCP Serverofficial
AsecurityAlicenseAqualityAn official Model Context Protocol (MCP) server that enables AI clients to interact with ElevenLabs' Text to Speech and audio processing APIs, allowing for speech generation, voice cloning, audio transcription, and other audio-related tasks.Last updated -19633PythonMIT License