/**
* Copyright 2024 Google LLC
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
import { Genkit, ModelArgument, z } from 'genkit';
import { BaseEvalDataPoint, Score } from 'genkit/evaluator';
import path from 'path';
import { getDirName, loadPromptFile, renderText } from './helper.js';
const MaliciousnessResponseSchema = z.object({
reason: z.string(),
verdict: z.boolean(),
});
export async function maliciousnessScore<
CustomModelOptions extends z.ZodTypeAny,
>(
ai: Genkit,
judgeLlm: ModelArgument<CustomModelOptions>,
dataPoint: BaseEvalDataPoint,
judgeConfig?: CustomModelOptions
): Promise<Score> {
try {
if (!dataPoint.input) {
throw new Error('Input was not provided');
}
if (!dataPoint.output) {
throw new Error('Output was not provided');
}
const input =
typeof dataPoint.input === 'string'
? dataPoint.input
: JSON.stringify(dataPoint.input);
const output =
typeof dataPoint.output === 'string'
? dataPoint.output
: JSON.stringify(dataPoint.output);
const prompt = await loadPromptFile(
path.resolve(getDirName(), '../../prompts/maliciousness.prompt')
);
//TODO: safetySettings are gemini specific - pull these out so they are tied to the LLM
const response = await ai.generate({
model: judgeLlm,
config: judgeConfig,
prompt: await renderText(prompt, {
input: input,
submission: output,
}),
output: {
schema: MaliciousnessResponseSchema,
},
});
const parsedResponse = response.output;
if (!parsedResponse) {
throw new Error(`Unable to parse evaluator response: ${response.text}`);
}
return {
score: 1.0 * (parsedResponse.verdict ? 1 : 0),
details: { reasoning: parsedResponse.reason },
};
} catch (err) {
console.debug(
`Genkit answer relevancy evaluation failed with error ${err} for sample ${JSON.stringify(dataPoint)}`
);
throw err;
}
}