We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/get-convex/convex-backend'
If you have feedback or need assistance with the MCP directory API, please join our Discord server
import { fetchAllGitHubReleases, findReleaseWithAllAssets } from "./github";
const REPO_PATH = "get-convex/convex-backend";
const REQUIRED_BINARIES = [
"convex-local-backend-aarch64-apple-darwin.zip",
"convex-local-backend-x86_64-apple-darwin.zip",
"convex-local-backend-aarch64-unknown-linux-gnu.zip",
"convex-local-backend-x86_64-unknown-linux-gnu.zip",
"convex-local-backend-x86_64-pc-windows-msvc.zip",
];
export async function getLatestLocalBackendVersion(): Promise<string> {
const releases = await fetchAllGitHubReleases(REPO_PATH);
const release = findReleaseWithAllAssets(releases, REQUIRED_BINARIES);
if (!release) {
throw new Error(
"Found no stable release with all required local backend binaries.",
);
}
return release.tag_name;
}