truthcheck
Analyze AI-generated content for truthfulness using seven specialized methods including linguistic signals, logical consistency, and confidence-based correction to verify accuracy.
Instructions
[TRUTHFULNESS] 7 tools: probe (linguistic truth signals), truth_direction (truth vector projection), ncb (perturbation robustness), logic (formal logical consistency), verify_first (5-dimension verification), ioe (confidence-based correction), self_critique (iterative refinement). Auto-selects or use 'check' to override. Set cascade=true for auto-correction on low scores.
Input Schema
| Name | Required | Description | Default |
|---|---|---|---|
| sessionId | No | Session identifier | default |
| check | No | Override: run a specific truthfulness check. If omitted, auto-selects based on params. probe — linguistic truth proxy signals; truth_direction — truth vector projection; ncb — perturbation robustness; logic — formal logical consistency; verify_first — 5-dimension verification; ioe — confidence-based self-correction; self_critique — iterative multi-criteria refinement | |
| cascade | No | If true, after primary checks, auto-run ioe_self_correct → self_critique when any extracted truthfulness score falls below 0.5. | |
| params | No | Parameters for the underlying tool(s), minus sessionId. probe: {assistantOutput, includeHistory?}; truth_direction: {assistantOutput, includePriorOutputs?}; ncb: {originalQuery, response}; logic: {claims[], includeGroundTruth?}; verify_first: {candidateAnswer, question, context?}; ioe: {response, question?, priorAttempts?}; self_critique: {solution, criteria?, maxIterations?, question?} |