getSecondOpinion
Obtain diverse responses from multiple LLM providers to validate or compare outputs. Ideal for cross-checking prompt results across OpenAI, Anthropic, Google, and other AI models.
Instructions
Get responses from various LLM providers
Input Schema
Name | Required | Description | Default |
---|---|---|---|
frequency_penalty | No | ||
maxTokens | No | ||
model | Yes | ||
presence_penalty | No | ||
prompt | Yes | ||
provider | Yes | ||
reasoning_effort | No | ||
stop_sequences | No | ||
stream | No | ||
systemPrompt | No | ||
temperature | No | ||
top_k | No | ||
top_p | No |