Skip to main content
Glama
msftnadavbh

Azure Pricing MCP Server

by msftnadavbh

azure_ptu_sizing

Estimate required Provisioned Throughput Units (PTUs) for Azure OpenAI deployments based on workload parameters like RPM and token usage, with optional cost calculation using Azure retail pricing.

Instructions

Estimate required Provisioned Throughput Units (PTUs) for Azure OpenAI / AI Foundry model deployments. Calculates PTUs based on workload shape (RPM, input/output tokens, caching) with official rounding rules. Optionally estimates hourly/monthly cost via Azure Retail Prices API. Supports Global, Data Zone, and Regional Provisioned deployment types.

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
modelYesModel identifier. Supported: gpt-5.2, gpt-5.2-codex, gpt-5.1, gpt-5.1-codex, gpt-5, gpt-5-mini, gpt-4.1, gpt-4.1-mini, gpt-4.1-nano, o3, o4-mini, gpt-4o, gpt-4o-mini, o3-mini, o1, Llama-3.3-70B-Instruct, DeepSeek-R1, DeepSeek-V3-0324, DeepSeek-R1-0528
deployment_typeYesProvisioned deployment type
rpmYesRequests per minute at peak workload
avg_input_tokensYesAverage input (prompt) tokens per request
avg_output_tokensYesAverage output (completion) tokens per request
cached_tokens_per_requestNoAverage cached tokens per request (deducted 100%% from utilization). Default: 0
include_costNoFetch live $/PTU/hr pricing from Azure Retail Prices API. Default: false
regionNoAzure region for cost lookup (e.g., 'eastus', 'westeurope'). Default: 'eastus'eastus
currency_codeNoCurrency code for pricing (default: 'USD')USD
Behavior3/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

With no annotations provided, the description carries the full burden of behavioral disclosure. It describes the calculation method ('official rounding rules'), mentions cost estimation as optional, and specifies supported deployment types. However, it doesn't address important behavioral aspects like whether this is a read-only operation, potential rate limits, authentication requirements, or what the output format looks like (though there's no output schema).

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness5/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is efficiently structured in three sentences: first states the core purpose, second explains the calculation method and optional features, third specifies deployment type support. Every sentence adds value with zero wasted words, and the most important information (what the tool does) comes first.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness3/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

For a tool with 9 parameters, no annotations, and no output schema, the description provides adequate but incomplete context. It covers the purpose, calculation method, and optional features well, but doesn't address output format, error conditions, or authentication requirements. Given the complexity and lack of structured metadata, the description should do more to compensate for these gaps.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters3/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 100%, so the schema already documents all 9 parameters thoroughly. The description adds some context about workload shape parameters (RPM, tokens, caching) and cost estimation, but doesn't provide additional semantic meaning beyond what's in the parameter descriptions. The baseline of 3 is appropriate when the schema does the heavy lifting.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose5/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the specific action ('Estimate required Provisioned Throughput Units'), identifies the target resource ('Azure OpenAI / AI Foundry model deployments'), and distinguishes from siblings by focusing on PTU sizing rather than cost analysis, SKU discovery, or other Azure operations. It specifies the calculation method ('based on workload shape with official rounding rules') and optional cost estimation.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines4/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides clear context about when to use this tool: for estimating PTU requirements for Azure OpenAI/AI Foundry deployments with specific workload parameters. It mentions optional cost estimation via Azure Retail Prices API, which helps differentiate from pure cost tools. However, it doesn't explicitly state when NOT to use it or name specific alternatives among the sibling tools.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Install Server

Other Tools

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/msftnadavbh/AzurePricingMCP'

If you have feedback or need assistance with the MCP directory API, please join our Discord server