heor-agent-mcp
The HEORAgent MCP server is an AI-powered Health Economics and Outcomes Research (HEOR) platform that automates evidence generation, economic modeling, and HTA submission workflows.
Literature & Evidence
Literature search across 44 data sources (PubMed, ClinicalTrials.gov, NICE TAs, ICER, CADTH, PBAC, Embase, Cochrane, FDA, LATAM/APAC sources, etc.) with PRISMA-style audit trail
Abstract screening with PICO-based relevance scoring and study design classification
Risk of bias assessment using Cochrane RoB 2 (RCTs), ROBINS-I (observational), and AMSTAR-2 (systematic reviews) with GRADE domain summaries
Evidence network building, NMA feasibility assessment, and indirect treatment comparisons (Bucher, frequentist NMA with consistency checks)
Population-adjusted indirect comparisons (MAIC/STC) with feasibility assessment
Survival curve fitting for 5 parametric distributions per NICE DSU TSD 14
Economic Modeling
Cost-effectiveness modeling (Markov, PartSA, decision-tree) with PSA (up to 10,000 Monte Carlo iterations), OWSA tornado diagrams, CEAC, EVPI/EVPPI, and WTP assessment against NHS, US payer, and societal thresholds
Budget impact modeling (ISPOR-compliant, year-by-year output, treatment-displacement modeling)
Utility value set analysis with EQ-5D-3L/5L references and baseline-utility-aware ICER impact estimation
HTA & Regulatory
HTA dossier drafting for NICE, EMA, FDA, IQWiG, HAS, and EU JCA with structured GRADE tables and gap analysis
Pharmacovigilance classification into EMA GVP categories with ENCePP protocol templates, PASS/PAES classification, and RMP implications
ITC feasibility assessment evaluating the 3-assumption framework
Project & Knowledge Management
Persistent project workspaces with directory skeleton and metadata
Full-text knowledge search, file read/write across
raw/andwiki/trees, Obsidian-compatible markdown with wikilinksMulti-tool workflow orchestration (e.g., end-to-end MAIC pipelines, full dossier creation)
Link validation for HTTP citation URLs
Output & Integration
Output formats: Markdown, structured JSON, or DOCX
Full audit trail per tool call (source selection, queries, inclusions/exclusions, assumptions, warnings)
Available via Claude Desktop/Code, Claude.ai, any MCP-compatible host, and a ChatGPT Custom GPT adapter
Provides access to Embase and ScienceDirect databases (requires ELSEVIER_API_KEY) for comprehensive literature review across 41 data sources with audit trails.
Hosts source code repository for the HEORAgent MCP server at github.com/neptun2000/heor-agent-mcp for development and collaboration.
Searches academic literature through Google Scholar API (requires SERPAPI_KEY) as part of comprehensive literature review across 41 data sources with audit trails.
Runs the HEORAgent MCP server as a Node.js application with support for both stdio and HTTP transport modes.
Distributes the HEORAgent MCP server as an npm package for easy installation and version management.
Writes compiled evidence to project wiki in Obsidian-compatible markdown format with wikilinks for persistent knowledge management.
Searches 35M+ biomedical citations through NCBI E-utilities as part of comprehensive literature review across 41 data sources with PRISMA-style audit trails.
Hosts MCP server deployment for web UI backend, enabling tool execution without local setup when using the companion chat interface.
Hosts companion chat interface at web-michael-ns-projects.vercel.app for interacting with HEOR tools through Claude Opus with BYOK (Bring Your Own Key) functionality.
HEORAgent MCP Server
AI-powered Health Economics and Outcomes Research (HEOR) agent as a Model Context Protocol server.
Try it now → HEORAgent on ChatGPT (ChatGPT Plus / Team) · Web UI (Claude, BYOK) ·
npx heor-agent-mcpfor Claude Desktop / Claude Code
Automates literature review across 44 data sources, risk of bias assessment (RoB 2 / ROBINS-I / AMSTAR-2), EQ-5D value set impact estimation, state-of-the-art cost-effectiveness modelling, HTA dossier preparation for NICE / EMA / FDA / IQWiG / HAS / EU JCA, and a persistent project knowledge base — all callable as MCP tools from Claude.ai, Claude Code, and any MCP-compatible host.
Built for pharmaceutical, biotech, CRO, and medical affairs teams who need rigorous, auditable HEOR workflows without building infrastructure from scratch.
What's new in v1.1.0
Pharmacovigilance + workflow orchestration:
pv_classifytool — classifies a planned study into its EMA pharmacovigilance regulatory category (PASS imposed/voluntary, PAES, RMP Annex 4, DUS, active surveillance registry, pregnancy registry, spontaneous reporting, ICH E2E plan). Returns the matching GVP module (V/VI/VIII/VIII Addendum I), ENCePP protocol template ID, RMP implications, FDA analogue, and submission obligations. Pure decision-tree per EMA GVP rev 4 + EU Regulation 1235/2010 Article 107a. <200ms response.hta_dossierPharmacovigilance Plan section — passpv_classificationfrompv_classifytohta_dossierand the dossier output now includes a PV Plan section between RoB and CEA. Without it, a one-line "PV plan not provided" note flags the gap so reviewers see what's missing.maic_workfloworchestrator (v1.0.6) — runs the full MAIC discovery+screening pipeline (ITC feasibility + parallel literature_search + screening + RoB + network) in one MCP call. Built for ChatGPT-5.3 surfaces where chaining 5+ tool calls in parallel is unreliable; works equally well from Claude.examplestool (v1.0.5) — pre-filled JSON inputs for heavy-schema tools (CEA, BIA, survival, MAIC, Bucher) plus amaic_workflow_recipemulti-step prompt template for ChatGPT users.CMS IRA awareness — when
pv_classifyis called with US jurisdiction, output explicitly notes that CMS IRA Medicare price-negotiation calculations exclude PV cost data — track those obligations in the regulatory budget, not the HEOR cost-effectiveness model.GRADE I²-based inconsistency, GRADE upgrading (Guyatt 2011), Bucher consistency check, EQ-5D 5L baseline-utility-aware impact (v1.0.4) — see CHANGELOG.md.
ChatGPT Custom GPT support (v1.0.4) — OpenAPI 3.1 adapter at
/api/openapilets you build a Custom GPT in 5 minutes. See ChatGPT Custom GPT below.Surface-tagged analytics (v1.0.4) — every
tool_callPostHog event carries asurfaceproperty (claude_anthropic_web/chatgpt_adapter/claude_desktop/direct_mcp).
See CHANGELOG.md for the full diff.
Quick Start
Claude Code
claude mcp add heor-agent -- npx heor-agent-mcpThen restart Claude Code.
Claude Desktop / claude.ai
Add to your MCP configuration:
{
"mcpServers": {
"heor-agent": {
"command": "npx",
"args": ["heor-agent-mcp"]
}
}
}Verify
> Run a literature search for semaglutide cost-effectiveness in T2D using PubMed and NICE TAsTools (22)
Tool | Purpose |
| Search 44 data sources with a full PRISMA-style audit trail |
| PICO-based relevance scoring and study design classification |
| Cochrane RoB 2 / ROBINS-I / AMSTAR-2 with GRADE RoB domain summary |
| Build treatment comparison network and assess NMA feasibility |
| Bucher and frequentist NMA with automatic consistency check vs direct h2h evidence (NICE DSU TSD 18) |
| MAIC and STC for population-adjusted indirect comparisons |
| Fit 5 parametric distributions to KM data (NICE DSU TSD 14) |
| Assess the 3-assumption ITC framework and recommend Bucher / NMA / MAIC / STC / ML-NMR |
| Markov / PartSA / decision-tree CEA with PSA, OWSA, CEAC, EVPI, EVPPI; QALY + evLYG support |
| ISPOR-compliant BIA with year-by-year output and treatment-displacement modelling |
| Draft submissions for NICE, EMA, FDA, IQWiG, HAS, and EU JCA — GRADE table uses structured RoB when |
| EQ-5D-3L / 5L value-set reference + baseline-utility-aware Biz 2026 ICER impact estimator (UK 5L transition) |
| HTTP validation of citation URLs before presentation |
| Initialize a persistent project workspace |
| Full-text search across a project's raw/ and wiki/ trees |
| Read any file from a project's knowledge base |
| Write compiled evidence to the project wiki (Obsidian-compatible) |
literature_search
Searches across 44 sources in parallel. Every call returns a source selection table showing which of the 44 sources were used and why — essential for HTA audit trails.
Example call:
{
"query": "semaglutide cardiovascular outcomes type 2 diabetes",
"sources": ["pubmed", "clinicaltrials", "nice_ta", "cadth_reviews", "icer_reports"],
"max_results": 20,
"output_format": "text"
}cost_effectiveness_model
Multi-state Markov model (default) or Partitioned Survival Analysis (oncology), following ISPOR good practice and NICE reference case (3.5% discount rate, half-cycle correction). Includes:
PSA — 1,000–10,000 Monte Carlo iterations, probability cost-effective at WTP thresholds
OWSA — one-way sensitivity analysis with tornado summary
CEAC — cost-effectiveness acceptability curve
EVPI — expected value of perfect information
WTP assessment — verdict against NHS (£25–35K/QALY, updated April 2026), US payer ($100–150K), societal thresholds
Example call:
{
"intervention": "Semaglutide 1mg SC weekly",
"comparator": "Sitagliptin 100mg daily",
"indication": "Type 2 Diabetes Mellitus",
"time_horizon": "lifetime",
"perspective": "nhs",
"model_type": "markov",
"clinical_inputs": { "efficacy_delta": 0.5, "mortality_reduction": 0.15 },
"cost_inputs": { "drug_cost_annual": 3200, "comparator_cost_annual": 480 },
"utility_inputs": { "qaly_on_treatment": 0.82, "qaly_comparator": 0.76 },
"run_psa": true,
"output_format": "docx"
}hta_dossier_prep
Drafts submission-ready sections for six HTA frameworks with gap analysis:
Body | Country | Submission types |
NICE | UK | STA, MTA, early_access |
EMA | EU | STA, MTA |
FDA | US | STA, MTA |
IQWiG | Germany | STA, MTA |
HAS | France | STA, MTA |
JCA | EU (Reg. 2021/2282) | initial, renewal, variation (with PICOs) |
Accepts piped output from literature_search and cost_effectiveness_model.
risk_of_bias
Assesses risk of bias using the appropriate Cochrane instrument, auto-detected from study_type:
Study type | Instrument |
RCT | RoB 2 (5 domains: randomization, deviations, missing data, measurement, reporting) |
Observational | ROBINS-I (7 domains: confounding, selection, classification, deviations, missing data, measurement, reporting) |
Systematic review | AMSTAR-2 (16 items, critical vs non-critical) |
Returns a rob_results object you can pass directly to hta_dossier_prep — this replaces the heuristic RoB estimate in the GRADE table with structured domain judgments.
Example call:
{
"studies": [{ "id": "pmid_1", "study_type": "RCT", "title": "...", "abstract": "..." }],
"output_format": "json"
}Pipeline:
literature_search→screen_abstracts→risk_of_bias→hta_dossier_prep
Knowledge base tools
Projects live at ~/.heor-agent/projects/{project-id}/ with:
raw/literature/— auto-populated literature search resultsraw/models/— auto-populated model runsraw/dossiers/— auto-populated dossier draftsreports/— generated DOCX fileswiki/— manually curated, Obsidian-compatible markdown with[[wikilinks]]
Pass project: "project-id" to any tool and results are saved automatically.
Examples
Copy-paste prompts to try in Claude Code, Claude Desktop, or the web UI.
Single-tool examples
Literature search
Search the literature for tirzepatide cardiovascular outcomes in type 2 diabetes. Use PubMed, ClinicalTrials.gov, and NICE TAs.
Survival curve fitting
Fit survival curves to this OS data from KEYNOTE-189: time 0 survival 1.0, time 6 survival 0.88, time 12 survival 0.72, time 18 survival 0.60, time 24 survival 0.51, time 36 survival 0.38. Use months.
Budget impact
Estimate the 5-year NHS budget impact of semaglutide for obesity. 200,000 eligible patients, drug cost £1,200/year, comparator (orlistat) £250/year, uptake 15% year 1 to 40% year 5.
Cost-effectiveness model
Build a CE model for semaglutide vs sitagliptin in T2D, NHS perspective, lifetime horizon, with PSA.
Indirect comparison (Bucher)
I have two trials: SUSTAIN-1 showed semaglutide vs placebo HR 0.74 (0.58-0.95) for HbA1c, and AWARD-5 showed dulaglutide vs placebo HR 0.78 (0.65-0.93). Run a Bucher indirect comparison between semaglutide and dulaglutide.
MAIC (population-adjusted comparison)
Run a MAIC between SUSTAIN-7 (N=300, semaglutide vs placebo, HR 0.74, CI 0.58-0.95, age 56±10, BMI 33±5) and AWARD-11 (N=600, dulaglutide vs placebo, HR 0.78, CI 0.65-0.93, age 58±9, BMI 35±6). Adjust for age and BMI.
Multi-tool workflows
Abstract screening workflow
Search PubMed for pembrolizumab in NSCLC, then screen the results with population adults with NSCLC, intervention pembrolizumab, comparator chemotherapy, outcomes overall survival and PFS.
Evidence network + NMA feasibility
Search for GLP-1 receptor agonists in T2D using PubMed, build an evidence network from the results, and assess NMA feasibility.
CE model with scenarios
Build a CE model for dapagliflozin vs placebo in heart failure, NHS perspective, lifetime horizon, with PSA. Add scenarios: "20% price reduction" with drug cost 400, "10-year horizon" with time_horizon 10yr.
End-to-end HTA workflow
Full dossier preparation
Create a project for semaglutide in obesity targeting NICE and ICER. Search literature for evidence, screen the results for adults with obesity comparing semaglutide to placebo for weight loss outcomes, assess risk of bias on the screened studies, then draft a NICE STA dossier using the screened results and rob_results.
This single prompt exercises: project_create → literature_search → screen_abstracts → risk_of_bias → hta_dossier_prep (GRADE RoB from structured assessment).
Data Sources
44 sources across 10 categories. Every literature_search call includes a source selection table showing used/not-used status and reason for each.
PubMed — 35M+ biomedical citations (NCBI E-utilities)
ClinicalTrials.gov — NIH/NLM trial registry (CT.gov v2 API)
bioRxiv / medRxiv — Life sciences and medical preprints
ChEMBL — Drug bioactivity, mechanisms, ADMET (EMBL-EBI)
Wiley Online Library — Pharmacoeconomics, Health Economics, Journal of Medical Economics, Value in Health (CrossRef, ~77% abstract coverage, no key required)
WHO GHO — WHO Global Health Observatory
World Bank — Demographics, macroeconomics, health expenditure
OECD Health — OECD health statistics (expenditure, workforce, outcomes)
IHME GBD — Global Burden of Disease (DALYs, prevalence across 204 countries)
All of Us — NIH precision medicine cohort
FDA Orange Book — Drug approvals and therapeutic equivalence
FDA Purple Book — Licensed biologics and biosimilars
NICE TAs (UK) · CADTH (Canada) · ICER (US) · PBAC (Australia)
G-BA AMNOG (Germany) · IQWiG (Germany) · HAS (France)
AIFA (Italy) · TLV (Sweden) · INESSS (Quebec, Canada)
CMS NADAC (US drug acquisition costs)
PSSRU (UK unit costs) · NHS National Cost Collection · BNF (UK drug pricing)
PBS Schedule (Australia)
DATASUS · CONITEC · ANVISA (Brazil)
PAHO (Pan American regional) · IETS (Colombia) · FONASA (Chile)
HITAP (Thailand)
Source | Env variable |
Embase |
|
ScienceDirect |
|
Cochrane Library |
|
Citeline |
|
Pharmapendium |
|
Cortellis |
|
Google Scholar |
|
ISPOR — HEOR methodology and conference abstracts
OHE (Office of Health Economics) — EQ-5D value set research and HEOR methodology
EuroQol Group — EQ-5D instruments, value sets, and registry
Output Formats
All tools support output_format:
text(default) — Markdown with formatted tables and headingsjson— Structured objects for downstream toolsdocx— Microsoft Word document, saved to disk, path returned in response
DOCX files are saved to ~/.heor-agent/projects/{project}/reports/ (when a project is set) or ~/.heor-agent/reports/ (global). The tool response contains the absolute path — ready to attach to submissions or share with stakeholders.
Audit Trail
Every tool call returns a full audit record:
Source selection table — all 44 sources with used/not-used and reason
Sources queried — queries sent, response counts, status, latency
Inclusions / exclusions — counts with reasons
Methodology — PRISMA-style for literature, ISPOR/NICE for economics
Assumptions — every assumption logged with justification
Warnings — data quality flags, missing API keys, failed sources
Suitable for inclusion in HTA submission appendices.
Configuration
# Optional — enterprise data sources
ELSEVIER_API_KEY=... # Embase + ScienceDirect
COCHRANE_API_KEY=... # Cochrane Library
CITELINE_API_KEY=... # Citeline
PHARMAPENDIUM_API_KEY=... # Pharmapendium
CORTELLIS_API_KEY=... # Cortellis
SERPAPI_KEY=... # Google Scholar
# Optional — knowledge base location
HEOR_KB_ROOT=~/.heor-agent # Default
# Optional — localhost proxy for enterprise APIs behind corporate VPN
HEOR_PROXY_URL=http://localhost:8787
# Optional — hosted tier (future)
HEOR_API_KEY=...Web UI
A companion chat interface is available at:
https://web-michael-ns-projects.vercel.app
Chat with Claude Sonnet 4.6 + all 22 HEOR tools
BYOK (Bring Your Own Key) — paste your Anthropic API key in the settings; it stays in your browser's localStorage and is never stored on our servers
Markdown rendering with styled tables, tool call cards with live progress timers, and theme-aware mermaid network diagrams
12 example prompts covering literature search, CEA, BIA, NMA, ITC feasibility, RoB, EQ-5D 5L, EU JCA dossiers
Per-request MCP sessions (no cross-user session bleed)
The web UI calls the hosted MCP server on Railway for tool execution. No setup required — just add your API key and start querying.
Self-hosting the web UI
cd web
npm install
echo "ANTHROPIC_API_KEY=sk-ant-..." > .env.local # optional server-side fallback
npm run dev -- -p 3456Set MCP_SERVER_URL to point to your own MCP server instance (default: the public Railway deployment).
ChatGPT Custom GPT
🟢 Live: HEORAgent on ChatGPT →
Open in ChatGPT (Plus / Team / Enterprise account required), pick a conversation starter, and you're querying 44 HEOR data sources.
HEORAgent is also available as a ChatGPT Custom GPT — useful when you (or your team) prefer the ChatGPT interface or have a ChatGPT Plus/Team account but no Anthropic API access.
Behind the scenes, the web tier exposes an OpenAPI 3.1 adapter at /api/openapi, with one POST endpoint per tool at /api/v1/{tool_name}. ChatGPT speaks this contract natively.
What's different from the Anthropic surface
Web UI / MCP / Claude Desktop | ChatGPT Custom GPT | |
Streaming | yes (SSE) | no (45s single response) |
| up to 10,000 | capped to 1,000 (CEA) / 500 (BIA) |
| 1–5 | capped to 1 |
| up to 100 | capped to 30 |
Auth model | BYOK Anthropic | optional |
Surface label in PostHog |
|
|
The caps exist because ChatGPT Actions hard-fail at the 45-second response timeout. PSA, multi-run literature search, and full max_results would routinely exceed it. The web UI and MCP clients are unaffected.
Build a Custom GPT (ChatGPT Plus / Team required)
Visit chatgpt.com/gpts/editor and click Create.
Configure tab — fill in name (e.g., "HEORAgent"), description, and conversation starters. Paste the system prompt from
web/lib/claude.ts(or write your own — the tool descriptions are self-documenting).Actions → Create new action → Import from URL → paste:
https://web-michael-ns-projects.vercel.app/api/openapiChatGPT auto-imports all 17 endpoints with their schemas.
Authentication — choose None for the open public endpoint, or API Key with the
CHATGPT_ADAPTER_TOKENvalue if you've configured one (recommended for prod).Privacy policy URL — required by GPT Store. Use the web UI's privacy URL or your own.
Test in the playground (right pane), then Publish → "Anyone with the link" or "GPT Store".
Securing the adapter for production
By default the /api/v1/* endpoint is open. Two layers of protection are recommended for any public-facing GPT:
# 1. Token-gate the endpoint
cd web
vercel env add CHATGPT_ADAPTER_TOKEN production # generate a long random token
# Configure the same token in your Custom GPT under Authentication → API Key
# 2. Built-in rate limit
# 60 req/min per IP is enforced automatically (lib/rateLimit.ts).
# For multi-region/high-traffic prod, swap in @upstash/ratelimit + Vercel KV.Sample call (manual, no GPT needed)
curl -X POST https://web-michael-ns-projects.vercel.app/api/v1/utility_value_set \
-H "Content-Type: application/json" \
-d '{
"action": "estimate_impact",
"indication_type": "non_cancer_qol_only",
"baseline_utility": 0.85,
"base_icer": 30000
}'Returns the Biz 2026 baseline-utility-adjusted ICER projection (the new EQ-5D 5L impact estimator).
HTTP Transport
The server supports both stdio (default, for local MCP clients) and Streamable HTTP (for hosted deployment).
# Stdio mode (default — for Claude Code, Claude Desktop)
npx heor-agent-mcp
# HTTP mode — for hosted deployment, Smithery, web UI backend
npx heor-agent-mcp --http # port 8787
MCP_HTTP_PORT=3000 npx heor-agent-mcp # custom portHTTP endpoints:
POST/GET/DELETE /mcp— MCP Streamable HTTP protocolGET /health— health checkGET /.well-known/mcp/server-card.json— Smithery discovery
Development
git clone https://github.com/neptun2000/heor-agent-mcp
cd heor-agent-mcp
npm install
npm test # 401 tests across 84 suites
npm run build # Compile TypeScript to dist/
npm run dev # Run with tsx (no build step)Requires: Node.js ≥ 20.
Architecture
┌────────────────────────────────────────────┐
│ MCP Host (Claude.ai / Claude Code / etc.) │
└────────────────┬───────────────────────────┘
│ stdio
┌────────────────▼──────────────────────────┐
│ heor-agent-mcp server │
│ ┌──────────────────────────────────────┐ │
│ │ 17 MCP tools (Zod-validated) │ │
│ ├──────────────────────────────────────┤ │
│ │ DirectProvider (default) │ │
│ │ ├─ 44 source fetchers │ │
│ │ ├─ Audit builder + PRISMA trail │ │
│ │ ├─ Markov / PartSA economic models │ │
│ │ ├─ Markdown + DOCX formatters │ │
│ │ └─ Knowledge base (YAML + MD) │ │
│ └──────────────────────────────────────┘ │
└───────────────────────────────────────────┘
│
┌────────────┴─────────────┐
▼ ▼
┌────────────┐ ┌──────────────────┐
│ ~/.heor- │ │ External APIs │
│ agent/ │ │ (PubMed, NICE, │
│ projects/ │ │ ICER, CADTH, …) │
└────────────┘ └──────────────────┘License
MIT — see LICENSE.
Trust & Transparency
HEORAgent is a research and analysis tool — not a clinical decision-support system. It is not classified as high-risk under the EU AI Act because it does not drive individual diagnosis, treatment, or monitoring; it falls under limited-risk transparency obligations only. Every output is intended for review by a qualified HEOR/HTA/PV professional before any action is taken.
EU AI Pact signatory — committed to AI governance, high-risk system mapping, and AI literacy promotion (voluntary commitments per the European Commission, ahead of the AI Act's August 2026 deadline).
PRISMA-style audit trail on every tool call (sources queried, succeeded, failed, assumptions applied).
AI commentary explicitly labelled — domain claims (ICERs, trial results, regulatory decisions) come exclusively from tool outputs, never from training-data recall.
Methodology cited inline — ISPOR, NICE DSU TSDs, NICE PMG36, Cochrane Handbook, GRADE, EMA GVP, Cope 2014, Phillippo 2016, Biz 2026.
Full statement: /ai-transparency — risk classification, human oversight model, methodological references, and reporting channel.
Disclaimer
All outputs are preliminary and for research orientation only. Results require validation by a qualified health economist before use in any HTA submission, payer negotiation, regulatory filing, or clinical decision. This tool does not replace professional HEOR expertise.
Distribution
Channel | How to use | Who pays |
npm |
| User's Claude subscription |
Smithery | User's Claude subscription | |
Web UI | User's own Anthropic API key (BYOK) | |
Hosted MCP |
| Free (tool execution only) |
Links
Latest Blog Posts
MCP directory API
We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/neptun2000/heor-agent-mcp'
If you have feedback or need assistance with the MCP directory API, please join our Discord server