handbook-frames
Decompose web pages into independent segments using scoped navigation, eager/lazy loading, and custom rendering. Enhance performance with targeted navigation, cache benefits, and frame promotion to visits.
Instructions
Turbo Frames for page decomposition - covers independent page segments, scoped navigation, eager/lazy loading, cache benefits, targeting navigation, frame promotion to visits, and custom rendering
Input Schema
TableJSON Schema
| Name | Required | Description | Default |
|---|---|---|---|
No arguments | |||
Implementation Reference
- src/index.ts:17-45 (registration)Registers the 'handbook-frames' tool (along with other doc tools) using the docFiles configuration, providing name, description, and an inline async handler that reads the corresponding markdown file and returns its content as text.docFiles.forEach(({ folder, file, name, description }) => { server.tool( name, description, async () => { try { const content = await readMarkdownFile(path.join(folder, file)); return { content: [ { type: "text", text: content } ] }; } catch (error) { const errorMessage = error instanceof Error ? error.message : String(error); return { content: [ { type: "text", text: `Error reading ${file}: ${errorMessage}` } ] }; } } ); });
- src/config.ts:32-37 (schema)Schema/configuration defining the tool name 'handbook-frames', its description, source folder, and markdown file path.{ folder: 'handbook', file: '04_frames.md', name: 'handbook-frames', description: 'Turbo Frames for page decomposition - covers independent page segments, scoped navigation, eager/lazy loading, cache benefits, targeting navigation, frame promotion to visits, and custom rendering' },
- src/documentReader.ts:14-61 (helper)Helper function invoked by the tool handler to read the markdown file content, with caching, GitHub fetching, and local fallback.export async function readMarkdownFile(filename: string): Promise<string> { const filePath = path.join(docsFolder, filename); if (!filePath.startsWith(docsFolder)) { throw new Error("Invalid file path"); } // Get current commit info if we don't have it yet if (!mainBranchInfo) { try { const commitInfo = await fetchMainBranchInformation(); const cacheKey = `${commitInfo.sha.substring(0, 7)}-${commitInfo.timestamp}`; mainBranchInfo = { ...commitInfo, cacheKey }; } catch (shaError) { console.error('Failed to get GitHub commit info, falling back to direct fetch'); } } // Try to read from cache first if we have commit info if (mainBranchInfo) { const cachedFilePath = path.join(cacheFolder, mainBranchInfo.cacheKey, filename); try { const content = await fs.promises.readFile(cachedFilePath, "utf-8"); console.error(`Using cached content for ${mainBranchInfo.cacheKey}: ${filename}`); return content; } catch (cacheError) { // Cache miss, continue to fetch from GitHub } } // Fetch from GitHub try { return await fetchFromGitHub(filename, mainBranchInfo?.cacheKey); } catch (githubError) { console.error(`GitHub fetch failed: ${githubError}, attempting to read from local files...`); // Fallback: read from local files try { return await fs.promises.readFile(filePath, "utf-8"); } catch (localError) { const githubErrorMessage = githubError instanceof Error ? githubError.message : String(githubError); const localErrorMessage = localError instanceof Error ? localError.message : String(localError); throw new Error(`Failed to read file from GitHub (${githubErrorMessage}) and locally (${localErrorMessage})`); } } }