Skip to main content
Glama
pyproject.toml745 B
# pyproject.toml [build-system] requires = ["setuptools>=61.0", "wheel"] build-backend = "setuptools.build_meta" [project] name = "linkedmcp" version = "0.1.0" description = "MCP server to get help with Linkedin content using LLM" readme = "README.md" authors = [ {name = "harsh_buddhdev", email = "harshbuddhdev147@gmail.com"} ] license = {text = "MIT"} classifiers = [ "Programming Language :: Python :: 3", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", ] dependencies = [ "fpdf>=1.7.2", "httpx>=0.28.1", "mcp[cli]>=1.2.0", "requests", ] requires-python = ">=3.10" [project.scripts] mcp-web-extractor = "linkedmcp.server:main" [tool.setuptools] package-dir = {"" = "src"}

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/buddhdev-harsh/linkedmcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server