Skip to main content
Glama

suno_create_persona

Create reusable artist personas from existing vocal styles to maintain consistent voice characteristics across multiple song generations.

Instructions

Create a new artist persona from an existing audio's vocal style.

This saves the vocal characteristics from a generated song so you can reuse
that same voice style in future generations. Great for maintaining consistency
across multiple songs.

Use this when:
- You generated a song and love the voice
- You want to create multiple songs with the same vocalist
- You're building an album with consistent vocal style
- You want to save a unique voice for future use

After creating a persona, use suno_generate_with_persona with the returned
persona_id to generate new songs with that voice.

Returns:
    Persona ID that can be used with suno_generate_with_persona tool.

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
audio_idYesID of the audio to use as the persona reference. This should be a previously generated song whose vocal style you want to save and reuse.
nameYesName for this persona. Use a descriptive name that helps you remember the voice style. Examples: 'My Rock Voice', 'Soft Female Singer', 'Deep Male Baritone', 'Energetic Pop Vocalist'
vox_audio_idNoOptional audio ID used to generate a new singer's style by combining with the main audio. Useful for creating hybrid vocal personas.
vocal_startNoStart time in seconds of the vocal segment to use from the audio. Useful for isolating a specific vocal section.
vocal_endNoEnd time in seconds of the vocal segment to use from the audio.
descriptionNoDescription of the singer's style. Examples: 'Warm and breathy female voice with jazz influences', 'Powerful male rock vocalist with raspy tone'

Output Schema

TableJSON Schema
NameRequiredDescriptionDefault
resultYes

Implementation Reference

  • Main tool handler function for suno_create_persona. Decorated with @mcp.tool(), it accepts audio_id, name, and optional parameters (vox_audio_id, vocal_start, vocal_end, description) to create a reusable vocal persona from an existing audio. Calls client.create_persona() and returns formatted result.
    @mcp.tool()
    async def suno_create_persona(
        audio_id: Annotated[
            str,
            Field(
                description="ID of the audio to use as the persona reference. This should be a previously generated song whose vocal style you want to save and reuse."
            ),
        ],
        name: Annotated[
            str,
            Field(
                description="Name for this persona. Use a descriptive name that helps you remember the voice style. Examples: 'My Rock Voice', 'Soft Female Singer', 'Deep Male Baritone', 'Energetic Pop Vocalist'"
            ),
        ],
        vox_audio_id: Annotated[
            str | None,
            Field(
                description="Optional audio ID used to generate a new singer's style by combining with the main audio. Useful for creating hybrid vocal personas."
            ),
        ] = None,
        vocal_start: Annotated[
            float | None,
            Field(
                description="Start time in seconds of the vocal segment to use from the audio. Useful for isolating a specific vocal section."
            ),
        ] = None,
        vocal_end: Annotated[
            float | None,
            Field(description="End time in seconds of the vocal segment to use from the audio."),
        ] = None,
        description: Annotated[
            str | None,
            Field(
                description="Description of the singer's style. Examples: 'Warm and breathy female voice with jazz influences', 'Powerful male rock vocalist with raspy tone'"
            ),
        ] = None,
    ) -> str:
        """Create a new artist persona from an existing audio's vocal style.
    
        This saves the vocal characteristics from a generated song so you can reuse
        that same voice style in future generations. Great for maintaining consistency
        across multiple songs.
    
        Use this when:
        - You generated a song and love the voice
        - You want to create multiple songs with the same vocalist
        - You're building an album with consistent vocal style
        - You want to save a unique voice for future use
    
        After creating a persona, use suno_generate_with_persona with the returned
        persona_id to generate new songs with that voice.
    
        Returns:
            Persona ID that can be used with suno_generate_with_persona tool.
        """
        payload: dict = {
            "audio_id": audio_id,
            "name": name,
        }
    
        if vox_audio_id:
            payload["vox_audio_id"] = vox_audio_id
        if vocal_start is not None:
            payload["vocal_start"] = vocal_start
        if vocal_end is not None:
            payload["vocal_end"] = vocal_end
        if description:
            payload["description"] = description
    
        result = await client.create_persona(**payload)
        return format_persona_result(result)
  • Helper function format_persona_result that formats the persona creation API response as a JSON string with submission guidance for follow-up task queries.
    def format_persona_result(data: dict[str, Any]) -> str:
        """Format persona creation result as JSON.
    
        Args:
            data: API response dictionary
    
        Returns:
            JSON string representation of the result
        """
        return json.dumps(
            _with_submission_guidance(data, "suno_get_task", "suno_get_tasks_batch"),
            ensure_ascii=False,
            indent=2,
        )
  • main.py:239-241 (registration)
    Registration of suno_create_persona tool in the main capabilities list with description 'Save a voice style for reuse'. This is part of the tools listing sent to the MCP client.
        "name": "suno_create_persona",
        "description": "Save a voice style for reuse",
    },
Behavior4/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden and successfully explains the persistent nature of the operation ('saves', 'reuse in future generations') and the return value (Persona ID). However, it omits operational details like error conditions, whether the process is synchronous, or constraints on audio ownership.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is well-structured with a clear opening statement, followed by explanatory context, a bulleted usage list, and workflow guidance. While the four bullet points are somewhat repetitive, they serve as useful trigger phrases for an LLM agent. No sentences are wasted.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness5/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the rich input schema (100% coverage) and the presence of an output schema, the description appropriately focuses on conceptual explanation and workflow integration rather than repeating parameter details. It confirms the return type (Persona ID) and explains the tool's role in the broader creation pipeline, making it complete for its complexity level.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

The input schema has 100% description coverage with detailed explanations and examples for each parameter (e.g., vocal_start, vox_audio_id). The description references 'existing audio' and 'generated song' which aligns with but does not significantly extend the schema's semantic documentation, meeting the baseline for high-coverage schemas.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description opens with a specific verb ('Create') and clearly defines the resource ('artist persona') and source material ('existing audio's vocal style'). It effectively distinguishes this tool from its sibling 'suno_generate_with_persona' by establishing the workflow: this tool saves the voice, the sibling uses it.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines5/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

Provides explicit 'Use this when:' criteria with four concrete scenarios, clearly indicating when the tool is appropriate. It also explicitly names the sibling tool 'suno_generate_with_persona' as the next step in the workflow, giving clear guidance on tool selection sequencing.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Install Server

Other Tools

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/AceDataCloud/MCPSuno'

If you have feedback or need assistance with the MCP directory API, please join our Discord server