Skip to main content
Glama

Jenkins MCP Server

client.py1.18 kB
from langchain_mcp_adapters.client import MultiServerMCPClient from langgraph.prebuilt import create_react_agent from langchain_openai import ChatOpenAI from dotenv import load_dotenv import asyncio load_dotenv() async def main(): client = MultiServerMCPClient( { "jenkins": { "command": "python", "args": ["-m", "jenkins_mcp_server"], "transport": "stdio" } } ) # await client.connect() tools = await client.get_tools() model = ChatOpenAI(model="gpt-4.1",api_key="") agent = create_react_agent(model,tools) queryy = await agent.ainvoke({ # "messages": [{"role":"user", "content":"can you access jenkins server"}] # "messages": [{"role":"user", "content":"get jobs list in jenkins"}] # "messages": [{"role":"user", "content":"start the job demo"}] # "messages": [{"role":"user", "content":"get the console log of job demo, latest build"}] "messages": [{"role":"user", "content":"find the error of job demo, latest build"}] }) print(queryy['messages'][-1].content) if __name__ == "__main__": asyncio.run(main())

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/Anil-appari007/jenkins-mcp-langchain'

If you have feedback or need assistance with the MCP directory API, please join our Discord server