Skip to main content
Glama

mcp-server-circleci

Official
inputSchema.ts1.26 kB
import { z } from 'zod'; import { defaultModel, defaultTemperature, PromptOrigin, } from '../shared/constants.js'; export const createPromptTemplateInputSchema = z.object({ prompt: z .string() .describe( "The user's application, feature, or product requirements that will be used to generate a prompt template. Alternatively, a pre-existing prompt or prompt template can be provided if a user wants to test, evaluate, or modify it. (Can be a multi-line string.)", ), promptOrigin: z .nativeEnum(PromptOrigin) .describe( `The origin of the prompt - either "${PromptOrigin.codebase}" for existing prompts from the codebase, or "${PromptOrigin.requirements}" for new prompts from requirements.`, ), model: z .string() .default(defaultModel) .describe( `The model that the prompt template will be tested against. Explicitly specify the model if it can be inferred from the codebase. Otherwise, defaults to \`${defaultModel}\`.`, ), temperature: z .number() .default(defaultTemperature) .describe( `The temperature of the prompt template. Explicitly specify the temperature if it can be inferred from the codebase. Otherwise, defaults to ${defaultTemperature}.`, ), });

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/CircleCI-Public/mcp-server-circleci'

If you have feedback or need assistance with the MCP directory API, please join our Discord server