MCP Terminal Server

/** * Copyright 2024 Google LLC * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ import { Genkit, ModelArgument, z } from 'genkit'; import { BaseEvalDataPoint, Score } from 'genkit/evaluator'; import path from 'path'; import { getDirName, loadPromptFile, renderText } from './helper.js'; const MaliciousnessResponseSchema = z.object({ reason: z.string(), verdict: z.boolean(), }); export async function maliciousnessScore< CustomModelOptions extends z.ZodTypeAny, >( ai: Genkit, judgeLlm: ModelArgument<CustomModelOptions>, dataPoint: BaseEvalDataPoint, judgeConfig?: CustomModelOptions ): Promise<Score> { try { if (!dataPoint.input) { throw new Error('Input was not provided'); } if (!dataPoint.output) { throw new Error('Output was not provided'); } const input = typeof dataPoint.input === 'string' ? dataPoint.input : JSON.stringify(dataPoint.input); const output = typeof dataPoint.output === 'string' ? dataPoint.output : JSON.stringify(dataPoint.output); const prompt = await loadPromptFile( path.resolve(getDirName(), '../../prompts/maliciousness.prompt') ); //TODO: safetySettings are gemini specific - pull these out so they are tied to the LLM const response = await ai.generate({ model: judgeLlm, config: judgeConfig, prompt: await renderText(prompt, { input: input, submission: output, }), output: { schema: MaliciousnessResponseSchema, }, }); const parsedResponse = response.output; if (!parsedResponse) { throw new Error(`Unable to parse evaluator response: ${response.text}`); } return { score: 1.0 * (parsedResponse.verdict ? 1 : 0), details: { reasoning: parsedResponse.reason }, }; } catch (err) { console.debug( `Genkit answer relevancy evaluation failed with error ${err} for sample ${JSON.stringify(dataPoint)}` ); throw err; } }