Skip to main content
Glama

O'RLY MCP Server

pyproject.toml1.25 kB
[build-system] requires = ["hatchling"] build-backend = "hatchling.build" [project] name = "orly-mcp" version = "0.1.3" description = "An MCP server for the ORLY tool." authors = [ {name = "Chris Frewin", email = "hi@fullstackcraft.com"}, ] readme = "README.md" requires-python = ">=3.10" classifiers = [ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", # Or your preferred license "Operating System :: OS Independent", ] dependencies = [ "fastmcp", "requests>=2.28.0", "Pillow>=11", "fonttools>=4.0.0", ] [project.scripts] orly-mcp = "orly_mcp.server:main" [tool.hatch.build.targets.wheel] packages = ["orly_mcp", "orly_generator"] [tool.hatch.build.targets.wheel.force-include] "fonts" = "fonts" "images" = "images" [tool.hatch.build.targets.sdist] include = [ "/orly_mcp", "/orly_generator", "/fonts", "/images", "/README.md", "/LICENSE.txt", "/pyproject.toml" ] [project.urls] "Homepage" = "https://github.com/princefishthrower/orly-mcp" "Bug Tracker" = "https://github.com/princefishthrower/orly-mcp/issues" "Repository" = "https://github.com/princefishthrower/orly-mcp" [dependency-groups] dev = [ "build>=1.2.2.post1", "twine>=6.1.0", ]

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/princefishthrower/orly-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server