Skip to main content
Glama

handbook-frames

Decompose web pages into independent segments using scoped navigation, eager/lazy loading, and custom rendering. Enhance performance with targeted navigation, cache benefits, and frame promotion to visits.

Instructions

Turbo Frames for page decomposition - covers independent page segments, scoped navigation, eager/lazy loading, cache benefits, targeting navigation, frame promotion to visits, and custom rendering

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault

No arguments

Implementation Reference

  • src/index.ts:17-45 (registration)
    Registers the 'handbook-frames' tool (along with other doc tools) using the docFiles configuration, providing name, description, and an inline async handler that reads the corresponding markdown file and returns its content as text.
    docFiles.forEach(({ folder, file, name, description }) => { server.tool( name, description, async () => { try { const content = await readMarkdownFile(path.join(folder, file)); return { content: [ { type: "text", text: content } ] }; } catch (error) { const errorMessage = error instanceof Error ? error.message : String(error); return { content: [ { type: "text", text: `Error reading ${file}: ${errorMessage}` } ] }; } } ); });
  • Schema/configuration defining the tool name 'handbook-frames', its description, source folder, and markdown file path.
    { folder: 'handbook', file: '04_frames.md', name: 'handbook-frames', description: 'Turbo Frames for page decomposition - covers independent page segments, scoped navigation, eager/lazy loading, cache benefits, targeting navigation, frame promotion to visits, and custom rendering' },
  • Helper function invoked by the tool handler to read the markdown file content, with caching, GitHub fetching, and local fallback.
    export async function readMarkdownFile(filename: string): Promise<string> { const filePath = path.join(docsFolder, filename); if (!filePath.startsWith(docsFolder)) { throw new Error("Invalid file path"); } // Get current commit info if we don't have it yet if (!mainBranchInfo) { try { const commitInfo = await fetchMainBranchInformation(); const cacheKey = `${commitInfo.sha.substring(0, 7)}-${commitInfo.timestamp}`; mainBranchInfo = { ...commitInfo, cacheKey }; } catch (shaError) { console.error('Failed to get GitHub commit info, falling back to direct fetch'); } } // Try to read from cache first if we have commit info if (mainBranchInfo) { const cachedFilePath = path.join(cacheFolder, mainBranchInfo.cacheKey, filename); try { const content = await fs.promises.readFile(cachedFilePath, "utf-8"); console.error(`Using cached content for ${mainBranchInfo.cacheKey}: ${filename}`); return content; } catch (cacheError) { // Cache miss, continue to fetch from GitHub } } // Fetch from GitHub try { return await fetchFromGitHub(filename, mainBranchInfo?.cacheKey); } catch (githubError) { console.error(`GitHub fetch failed: ${githubError}, attempting to read from local files...`); // Fallback: read from local files try { return await fs.promises.readFile(filePath, "utf-8"); } catch (localError) { const githubErrorMessage = githubError instanceof Error ? githubError.message : String(githubError); const localErrorMessage = localError instanceof Error ? localError.message : String(localError); throw new Error(`Failed to read file from GitHub (${githubErrorMessage}) and locally (${localErrorMessage})`); } } }

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/pinzonjulian/turbo-docs-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server