Web Resource Ledger
Integrates with the Google Web Risk API for content security screening, checking submitted URLs against threat intelligence databases to identify malware, phishing, and other security threats before capture and re-screening existing captures daily.
Web Resource Ledger (WRL)
Cryptographic evidence of web content -- capture what a page looked like, when, with proof anyone can verify.
Submit a URL, get back a screenshot, rendered HTML, HTTP headers, and an Ed25519-signed WACZ bundle. The verification URL works for anyone -- no account needed. Deploy on your own infrastructure; your captures, your keys, your evidence.
WRL was built almost entirely by an LLM (Claude) with a human directing the work. Every phase -- from MVP to production -- is documented in docs/evolution/. The prompts, agent debates, decisions, and outcomes are all public.
For comprehensive guides on authentication, verification, batch captures, and MCP integration, see docs.webresourceledger.com.
Status: Production service, single-operator deployment. All three roadmap acts are complete. The API is pre-1.0 but functional, deployed, and serving real traffic.
What you get
A single API call produces:
Dual screenshots (PNG) -- before and after cookie consent dismissal, so both the banner presence and the underlying page content are preserved
Rendered HTML -- the DOM after JavaScript execution
HTTP response headers -- the server's response at capture time
Signed WACZ bundle -- all artifacts packaged, hashed, and signed with Ed25519
Verification URL -- a shareable link anyone can use to confirm authenticity
FRE 902(13) certification PDF --
GET /v1/captures/{captureId}/certificatereturns a signed PDF describing the capture process and integrity evidence, designed to support a Rule 902(13) self-authentication foundation
Content Security
WRL screens every submitted URL against threat intelligence databases before capture begins:
URLs identified as threats (malware, phishing, etc.) are rejected at submission time with a 422 response.
Existing captures are re-scanned daily. Captures that fail re-screening are quarantined: metadata remains accessible, artifact downloads return 451.
Screening degrades gracefully -- if the threat intelligence service is unavailable, captures proceed normally. The
threatCheckfield on a capture record reflects the outcome:pass,fail, orunavailable.
Configure the service via GOOGLE_WEB_RISK_API_KEY in your environment (see step 8b below). The key is optional; omitting it disables screening and all captures report threatCheck: unavailable.
Usage
Requires a running WRL instance. See Setup below.
export WRL_API_KEY=your_tenant_api_keyTenant keys are created via the admin API (see step 8a). For deployments using the legacy static key, WRL_API_KEY is your CAPTURE_API_KEY value.
The examples below use api.webresourceledger.com. Replace with localhost:8787 for local dev or your own deployment URL if self-hosting.
Step 1: Submit a capture
curl -X POST https://api.webresourceledger.com/v1/captures \
-H "Authorization: Bearer $WRL_API_KEY" \
-H "Content-Type: application/json" \
-d '{"url": "https://example.com"}'{
"id": "cap_a1b2c3d4e5f6a7b8c9d0e1f2a3b4c5d6",
"statusUrl": "https://api.webresourceledger.com/v1/captures/cap_a1b2c3d4e5f6a7b8c9d0e1f2a3b4c5d6/status",
"note": "Use GET /v1/captures to list and search your captures."
}Your captures are always accessible. Use GET /v1/captures to list them, or save the capture ID for direct access.
Step 2: Poll for completion
curl https://api.webresourceledger.com/v1/captures/cap_a1b2c3d4e5f6a7b8c9d0e1f2a3b4c5d6/statusPoll until status is complete or failed. The response includes a captureUrl when complete.
Step 3: Retrieve artifacts
curl https://api.webresourceledger.com/v1/captures/cap_a1b2c3d4e5f6a7b8c9d0e1f2a3b4c5d6Returns metadata and signed artifact URLs (screenshot, html, headers, wacz) plus a verifyUrl.
Step 4: Verify the bundle
curl https://api.webresourceledger.com/v1/verify/cap_a1b2c3d4e5f6a7b8c9d0e1f2a3b4c5d6Returns a JSON verification result with up to four checks: artifactHashes, bundleHash, signature, and (for new captures) timestamp. The timestamp check verifies an RFC 3161 independent timestamp obtained at capture time. Legacy captures return three checks. The verifyUrl from step 3 also renders as a human-readable page in browsers.
Capture URLs are directly shareable -- anyone with the capture ID can access the capture and its artifacts. No tokens needed.
Sharing: Capture URLs are inherently shareable. Anyone with the URL can access the capture record and all artifacts. For proof-of-authenticity, share the
verifyUrlwhich renders as a human-readable verification page.
Offline verification
For independent, offline verification -- including full CMS/PKCS#7 certificate chain validation -- use the CLI tool:
npx @w-r-l/verify capture.wacz --origin https://api.webresourceledger.comSee packages/verify/ for details.
Finding captures
GET /v1/captures lists your captures (requires your API key). Use it to browse and recover capture IDs.
curl https://api.webresourceledger.com/v1/captures \
-H "Authorization: Bearer $WRL_API_KEY"{
"data": [
{
"id": "cap_a1b2c3d4e5f6a7b8c9d0e1f2a3b4c5d6",
"status": "complete",
"url": "https://example.com",
"createdAt": "2024-01-15T10:30:00.000Z",
"completedAt": "2024-01-15T10:30:45.123Z"
}
],
"pagination": {
"cursor": null,
"hasMore": false,
"limit": 20
}
}Optional query parameters: limit (1-100, default 20), cursor (for paging), status (pending, complete, failed, or quarantined).
Captures are rate-limited to 10 per minute per IP. Verification is limited to 60 per minute per IP. Error responses use RFC 9457 application/problem+json format. For full details, see openapi.yaml.
Detailed guides: See Getting Started for a complete walkthrough, or browse the API Reference for the full endpoint catalog.
Setup
Prerequisites
Wrangler CLI installed and authenticated
Node.js 22+ (see
.nvmrc)Cloudflare account with R2 and Browser Rendering enabled
1. Install dependencies
npm install2. Create D1 database
wrangler d1 create wrl-metadataUpdate the database_id in wrangler.toml under [[d1_databases]] with the returned ID. Then apply migrations:
wrangler d1 migrations apply wrl-metadata2a. Create KV namespace (rate limiting)
KV is used only for rate limit counters. Create the namespace:
wrangler kv namespace create wrl-kvUpdate wrangler.toml with the returned id and preview_id.
3. Create R2 bucket
wrangler r2 bucket create wrl-captures
wrangler r2 bucket create wrl-captures-preview4. Configure capture API key (legacy fallback)
CAPTURE_API_KEY is a static bearer token that acts as a fallback when no tenant key is found in D1. For new deployments, consider setting up the admin API (step 8a) and creating tenant keys instead. For existing deployments, this key continues to work during migration.
In the usage examples above, this is $WRL_API_KEY.
Generate a key:
openssl rand -hex 32Set the production secret:
wrangler secret put CAPTURE_API_KEYFor local dev, add to .dev.vars:
CAPTURE_API_KEY=<hex string from the command above>Security: Never commit this value to version control. .dev.vars is already in .gitignore.
5. Configure signing key
WRL signs WACZ bundles with Ed25519. The signing key is optional -- if SIGNING_KEY is not set, captures complete successfully but without WACZ bundles (screenshot, HTML, and headers are still stored).
Generate a key pair:
node scripts/generate-signing-key.jsThe script prints a private key (PKCS8 DER, base64) and the corresponding public key (raw, base64). The public key is embedded automatically in every signed bundle for verification -- no separate distribution needed.
Set the production secret:
wrangler secret put SIGNING_KEY
# Paste the private key (PKCS8 DER, base64) when promptedFor local dev, add to .dev.vars:
SIGNING_KEY=<base64 string from the script>Security: Never commit the private key to version control. .dev.vars is already in .gitignore.
6. Configure IP hash seed (recommended)
IP_HASH_SEED is an HMAC seed used to hash IP addresses before they appear in logs. Without it, log entries have no IP correlation for abuse analysis.
Generate a seed:
openssl rand -hex 32Set the production secret:
wrangler secret put IP_HASH_SEEDFor local dev, add to .dev.vars:
IP_HASH_SEED=<hex string from the command above>7. Configure Coralogix log ingestion (required for production observability)
CORALOGIX_SEND_KEY is the API key for structured log ingestion to Coralogix. Without it, the Worker runs normally but logs go to console only -- no structured log ingestion occurs. For fork developers who do not use Coralogix, this key is effectively optional.
Find your send key in the Coralogix dashboard under Settings > Send Your Data > API Keys.
Set the production secret:
wrangler secret put CORALOGIX_SEND_KEYFor local dev, structured logs are emitted to the console. No key is needed for wrangler dev.
8. Configure CORS origins (optional)
CORS_ORIGINS is a comma-separated list of allowed origins for CORS preflight responses. Only needed if browser-based clients will call the API directly.
Set it as an environment variable in wrangler.toml (not a secret):
[vars]
CORS_ORIGINS = "https://app.example.com,https://www.example.com"If omitted, cross-origin requests from browsers are blocked. Server-to-server requests are unaffected.
8a. Configure admin key (required for per-tenant key management)
ADMIN_KEY is the bearer token for the admin API (/v1/admin/keys). It grants the ability to create, list, and revoke per-tenant API keys. It does not grant capture or read access.
Generate a key:
openssl rand -hex 32Set the production secret:
wrangler secret put ADMIN_KEYFor local dev, add to .dev.vars:
ADMIN_KEY=<hex string from the command above>Once deployed, create your first tenant key:
curl -X POST https://api.webresourceledger.com/v1/admin/keys \
-H "Authorization: Bearer $ADMIN_KEY" \
-H "Content-Type: application/json" \
-d '{"tenantId": "default", "scopes": ["capture"], "name": "default-key"}' | jq .Use the returned key value as $WRL_API_KEY going forward. See OPERATIONS.md for the full migration runbook.
Security: Never commit this value to version control. .dev.vars is already in .gitignore.
8b. Configure content security screening (optional)
GOOGLE_WEB_RISK_API_KEY enables URL threat screening before capture and during daily re-scans. Without it, all captures report threatCheck: unavailable and no URLs are blocked.
Obtain a key from the Google Cloud Console with the Web Risk API enabled, then set the production secret:
wrangler secret put GOOGLE_WEB_RISK_API_KEYFor local dev, add to .dev.vars:
GOOGLE_WEB_RISK_API_KEY=<your api key>Security: Never commit this value to version control. .dev.vars is already in .gitignore.
9. Deploy
wrangler deploySteps 1-9 are one-time setup. After initial deployment, the CD pipeline handles staging and production deploys automatically on every push to main. For the full deploy flow, environment configuration, rollback procedures, and how secrets map across Worker runtime, GitHub CI, and local development, see OPERATIONS.md.
MCP Server
WRL exposes all four core operations — capture, retrieve, list, and verify — as MCP tools. Any MCP-compatible agent or IDE can capture web pages and verify evidence without writing HTTP client code.
See the MCP Server Guide for setup instructions and example workflows. For local reference, docs/mcp.md covers Claude Code, Cursor, Windsurf, and generic clients along with the full tool reference.
Web UI
WRL includes a browser-based interface for submitting captures and browsing
results. Navigate to /ui on your WRL deployment to access it. Authentication
requires a WRL API key with capture and read scopes.
The UI is served directly from the Worker — no separate hosting or CORS configuration required.
Development
See CONTRIBUTING.md for local dev setup, test conventions, and contribution guidelines.
See OPERATIONS.md for deployment, rollback, and environment setup.
Staging
wrangler.toml includes an [env.staging] configuration with its own R2 bucket, D1 database, and KV namespace (rate limiting only). Before deploying, you must create those resources in your own Cloudflare account (same as steps 2-3 above, but scoped to staging).
Create the staging D1 database:
wrangler d1 create wrl-metadata-stagingUpdate the database_id under [[env.staging.d1_databases]] in wrangler.toml, then apply migrations:
wrangler d1 migrations apply wrl-metadata-staging --env stagingCreate the staging KV namespace (rate limiting only):
wrangler kv namespace create KV --env stagingUpdate the id field under [env.staging.kv_namespaces] in wrangler.toml with the returned ID.
Create the staging R2 bucket:
wrangler r2 bucket create wrl-captures-stagingThen deploy to staging:
wrangler deploy --env stagingStaging auto-deploys on merge to main via deploy-staging.yml. Secrets must be set separately for the staging environment:
wrangler secret put CAPTURE_API_KEY --env staging
wrangler secret put SIGNING_KEY --env staging
# repeat for any other secretsSmoke tests run against a live deployment. Set SMOKE_URL and SMOKE_API_KEY, then:
npm run smokeRoadmap
WRL followed a three-act development plan. All three acts are complete:
Solid Foundation (complete) -- List endpoint, key versioning, CORS, security hardening. Closed trust gaps for single-operator use.
Evidence-Grade (complete) -- RFC 3161 timestamps, per-tenant keys, audit logging. Made captures independently verifiable.
Infrastructure (complete) -- MCP server, web UI, batch capture, scheduled captures, docs site, landing page, usage metering, webhooks. Expanded WRL into a platform.
See docs/backlog.md for the ongoing backlog and GitHub issues for detailed tracking.
How WRL Was Built
WRL is a production web evidence service. It is also a documented experiment in LLM-driven software development. A human defined what to build and made product decisions; an LLM (Claude) wrote virtually all the code, tests, infrastructure, and documentation.
The full build history is in docs/evolution/ -- 60+ phases covering architecture decisions, agent disagreements, and human overrides. If you're interested in what LLM-driven development actually looks like at production scale, start there.
Reference
Key Rotation
Key rotation is safe -- old captures continue to verify after rotation. Every time a capture is signed, the signing key is archived automatically. Each key is identified by a keyId: the first 8 hex characters of the SHA-256 of the raw 32-byte public key. The keyId is stored in the WACZ bundle's signedData.signatures array (v0.2.0) or signedData directly (v0.1.0 legacy) and in the D1 capture record. During verification, the system looks up the correct historical key by keyId rather than assuming the current key.
Rotation procedure:
Generate a new key pair:
node scripts/generate-signing-key.jsUpdate the production secret:
wrangler secret put SIGNING_KEYUpdate local dev secret in
.dev.vars(if applicable)
New captures are signed with the new key. Existing captures are verified against the archived key that signed them. The /.well-known/signing-keys endpoint lists the full key archive for third-party verifiers.
Public Key Endpoint
GET /.well-known/signing-key returns the current Ed25519 public key. Third-party verifiers can fetch the key without trusting the publicKey embedded in individual WACZ bundles. Responses are cached for 1 hour at the edge.
{
"algorithm": "Ed25519",
"publicKey": "<base64-encoded raw 32-byte key>",
"keyId": "<8-char hex fingerprint>"
}keyId is the first 8 hex characters of the SHA-256 of the raw public key bytes. Use it to match against the key archive when verifying historical captures.
Key Archive Endpoint
GET /.well-known/signing-keys lists all historical signing keys. Use this endpoint to verify captures signed with any key, not just the current one.
{
"keys": [
{
"keyId": "<8-char hex fingerprint>",
"algorithm": "Ed25519",
"publicKey": "<base64-encoded raw 32-byte key>",
"archivedAt": "<ISO 8601 timestamp>"
}
]
}Third-party verifiers: match the keyId from a WACZ bundle's signedData (v0.1.0) or signedData.signatures array (v0.2.0) against this list to retrieve the correct public key for signature verification. Rate-limited at the same limit as the singular endpoint.
Health Endpoint
GET /health returns the current service status, legal document URLs, and build identity metadata.
{
"status": "ok",
"legal": { "terms": "<url>", "policy": "<url>" },
"build": { "commit": "<sha>", "version": "<semver>", "env": "production", "deployedAt": "<iso8601>" }
}The build object is present when deployed via CI (injected at build time via wrangler --define). It is absent during local development. Useful for uptime monitoring, smoke tests, and deploy verification.
Response Headers
All responses include:
Link-- points to the terms-of-service URL withrel="terms-of-service". Present on every response.Strict-Transport-Security-- includespreloadandincludeSubDomains. Present on every response.X-RateLimit-Limit-- the rate limit ceiling for the endpoint. Present on responses from rate-limited endpoints (captures, verification, signing key endpoints).
Legal
By using this API, you agree to the Terms of Service.
License
PolyForm Shield 1.0.0 -- source-available. You may use, modify, and self-host WRL for any purpose except offering a competing web capture service. See LICENSE for the full text.
Latest Blog Posts
MCP directory API
We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/benpeter/web-resource-ledger'
If you have feedback or need assistance with the MCP directory API, please join our Discord server