Skip to main content
Glama

Hardware MCP Server

README.md1.31 kB
# Hardware for MCP This repository shows examples of how you can use a language model to perform tool calls for: - building a CAD part - running heat transfer simulations This repository uses the Model Context Protocol (MCP) ### Install Dependencies First, clone the current GitHub repository with the following command. ```bash git clone https://github.com/TheFloatingString/mcp-for-hardware.git cd mcp-for-hardware ``` Now, install all the Python dependencies using the `uv` package manager: ```bash pip install uv uv sync ``` Install ngrok or any similar tunelling services. ngrok can be installed from https://ngrok.com/downloads Next, if you use ngrok, run: ```bash ngrok http 8000 ``` This will expose port 8000 (which we will be using soon for the MCP server) to a specific URL. You should see a publicly visible URL after running the `ngrok` command. Keep note of this, you will need to add this in the `.env` file's environment variables. ### Configure Environment Variables Create a `.env` file and add the following environment variables ```bash SERVER_BASE_URL="" ANTHROPIC_API_KEY="" ``` ### Quickstart In one terminal, run the MCP server with the following: ```bash uv run server.py ``` In a second terminal, run the client with the following: ```bash uv run client.py ```

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/TheFloatingString/mcp-for-hardware'

If you have feedback or need assistance with the MCP directory API, please join our Discord server