Flyworks MCP

by Flyworks-AI
MIT License
2
  • Linux
  • Apple

Integrations

  • Supports configuration via .env files for storing the Flyworks API token and other settings

  • Provides a community support channel through their Discord server for users of the Flyworks MCP

  • Requires Python 3.8+ as the runtime environment for the MCP server

Flyworks MCP: Free & Fast Zeroshot Lipsync Tool

Overview

The Flyworks MCP is a Model Context Protocol (MCP) server that provides a convenient interface for interacting with the Flyworks API. It facilitates fast and free lipsync video creation for a wide range of digital avatars, including realistic and cartoon styles.

Demo

Input avatar video (footage):

Audio clip with TTS saying 我是一个飞影数字人。Welcome to Flyworks MCP server demo. This tool enables fast and free lipsync video creation for a wide range of digital avatars, including realistic and cartoon styles.:

Generated lipsync video:

Features

  • Create lipsynced videos using digital avatar video and audio as inputs
  • Create lipsynced videos by text (with text-to-speech)
  • Create digital human avatars from images or videos
  • Support for both async and sync modes of operation
  • More features coming soon...

Requirements

  • Python 3.8+
  • Dependencies: httpx, mcp[cli]

Installation

  1. Clone this repository:
    git clone https://github.com/yourusername/flyworks-mcp.git cd flyworks-mcp
  2. Install dependencies:
    pip install httpx "mcp[cli]>=1.6.0"
    Or using uv:
    uv pip install httpx "mcp[cli]>=1.6.0"
    To avoid timeout issues during server startup, we recommend pre-installing all dependencies:
    pip install pygments pydantic-core httpx "mcp[cli]>=1.6.0"

Configuration

Set your Flyworks API token as an environment variable:

# Linux/macOS export FLYWORKS_API_TOKEN="your_token_here" # Windows (Command Prompt) set FLYWORKS_API_TOKEN=your_token_here # Windows (PowerShell) $env:FLYWORKS_API_TOKEN="your_token_here"

Alternatively, you can create a .env file.

Note: We offer free trial access to our tool with the token 2aeda3bcefac46a3. However, please be aware that the daily quota for this free access is limited. Additionally, the generated videos will be watermarked and restricted to a duration of 45 seconds. For full access, please contact us at bd@flyworks.ai to acquire your token.

Usage

Running the Server

Run the server.py file directly:

python server.py
Integration with Claude or Other MCP Clients
Using in Claude Desktop

Go to Claude > Settings > Developer > Edit Config > claude_desktop_config.json to include the following:

{ "mcpServers": { "flyworks": { "command": "uvx", "args": [ "flyworks-mcp", "-y" ], "env": { "FLYWORKS_API_TOKEN": "your_api_token_here", "FLYWORKS_API_BASE_URL": "https://hfw-api.hifly.cc/api/v2/hifly", "FLYWORKS_MCP_BASE_PATH": "/path/to/your/output/directory" } } } }
Using in Cursor

Go to Cursor -> Preferences -> Cursor Settings -> MCP -> Add new global MCP Server to add above config.

Make sure to replace your_api_token_here with your actual API token, and update the FLYWORKS_MCP_BASE_PATH to a valid directory on your system where output files will be saved.

Issues: spawn uvx ENOENT

Please confirm its absolute path by running this command in your terminal:

which uvx

Once you obtain the absolute path (e.g., /usr/local/bin/uvx), update your configuration to use that path (e.g., "command": "/usr/local/bin/uvx").

Tool Description

1. Create Lipsync Video by Audio (create_lipsync_video_by_audio)

Create a lipsync video with audio input. Animates a digital human avatar to speak in sync with the provided audio.

Parameters:

  • avatar: Digital human avatar ID. Either this or avatar creation parameters must be provided.
  • avatar_video_url: URL of a video to create the avatar from.
  • avatar_image_url: URL of an image to create the avatar from.
  • avatar_video_file: Local path to a video file to create the avatar from.
  • avatar_image_file: Local path to an image file to create the avatar from.
  • audio_url: Remote URL of the audio file. One of audio_url or audio_file must be provided.
  • audio_file: Local path to the audio file. One of audio_url or audio_file must be provided.
  • title: Optional title for the created video.
  • async_mode: If true, returns task_id immediately. If false, waits for completion and downloads the video. Default is true.
  • output_path: Where to save the downloaded video if async_mode is false. Default is "output.mp4".

Notes:

  • For avatar creation, provide exactly ONE of avatar_video_url, avatar_image_url, avatar_video_file, or avatar_image_file.
  • If avatar ID is directly provided, these parameters will be ignored.

Returns:

  • If async_mode is true: task_id for checking status later and created_avatar (if a new avatar was created)
  • If async_mode is false: downloaded video path, task result, and created_avatar (if applicable)
2. Create Lipsync Video by Text (create_lipsync_video_by_text)

Create a lipsync video with text input. Generates audio from the text and animates a digital human avatar to speak it.

Parameters:

  • avatar: Digital human avatar ID. Either this or avatar creation parameters must be provided.
  • avatar_video_url: URL of a video to create the avatar from.
  • avatar_image_url: URL of an image to create the avatar from.
  • avatar_video_file: Local path to a video file to create the avatar from.
  • avatar_image_file: Local path to an image file to create the avatar from.
  • text: Text content to be spoken by the avatar. Required.
  • voice: Voice ID to use for text-to-speech. If not provided, a random voice will be selected automatically.
  • title: Optional title for the created video.
  • async_mode: If true, returns task_id immediately. If false, waits for completion and downloads the video. Default is true.
  • output_path: Where to save the downloaded video if async_mode is false. Default is "output.mp4".

Notes:

  • For avatar creation, provide exactly ONE of avatar_video_url, avatar_image_url, avatar_video_file, or avatar_image_file.
  • If avatar ID is directly provided, these parameters will be ignored.

Returns:

  • If async_mode is true: task_id for checking status later, selected voice ID, and created_avatar (if applicable)
  • If async_mode is false: downloaded video path, task result, selected voice ID, and created_avatar (if applicable)

Checking Task Status

For tasks run in async mode, you can check their status using the Flyworks API's /creation/task endpoint with the task_id returned by the tool.

Notes

  • Job processing may take some time, please be patient
  • Video file URLs are temporary, please download and save them promptly
  • When using local files, the server will automatically upload them to Flyworks servers
  • In sync mode, the tool will wait for the task to complete and automatically download the video
  • Maximum allowed wait time for sync mode is 10 minutes (600 seconds)
  • Avatar creation through videos usually provides better quality but takes longer
  • For quick testing, avatar creation through images is faster but may have lower quality
-
security - not tested
A
license - permissive license
-
quality - not tested

A Model Context Protocol server that enables fast and free lipsync video creation for a wide range of digital avatars, supporting both audio and text inputs to generate synchronized lip movements.

  1. Overview
    1. Demo
      1. Features
        1. Requirements
          1. Installation
            1. Configuration
              1. Usage
                1. Running the Server
                2. Integration with Claude or Other MCP Clients
              2. Tool Description
                1. 1. Create Lipsync Video by Audio (create_lipsync_video_by_audio)
                2. 2. Create Lipsync Video by Text (create_lipsync_video_by_text)
              3. Checking Task Status
                1. Notes
                  1. Related Links

                    Related MCP Servers

                    • -
                      security
                      A
                      license
                      -
                      quality
                      A Model Context Protocol server that integrates high-quality text-to-speech capabilities with Claude Desktop and other MCP-compatible clients, supporting multiple voice options and audio formats.
                      Last updated -
                      TypeScript
                      MIT License
                    • A
                      security
                      A
                      license
                      A
                      quality
                      A Model Context Protocol server that enables retrieval of transcripts from YouTube videos. This server provides direct access to video transcripts and subtitles through a simple interface, making it ideal for content analysis and processing.
                      Last updated -
                      1
                      258
                      10
                      TypeScript
                      MIT License
                      • Linux
                      • Apple
                    • -
                      security
                      F
                      license
                      -
                      quality
                      A Model Context Protocol server that provides Claude with tools to interact with YouTube, built on the mcp-framework.
                      Last updated -
                      TypeScript
                      • Apple
                    • -
                      security
                      -
                      license
                      -
                      quality
                      Model Context Protocol server that enables generating videos from text prompts and/or images using AI models (Luma Ray2 Flash and Kling v1.6 Pro) with configurable parameters like aspect ratio, resolution, and duration.
                      Last updated -
                      1
                      JavaScript
                      MIT License

                    View all related MCP servers

                    ID: y9enz2ldhd