Skip to main content
Glama
Teradata

Teradata MCP Server

Official
by Teradata

qlty_columnSummary

Analyze table columns to generate summary statistics for data quality assessment in Teradata databases.

Instructions

Get the column summary statistics for a table.

Arguments: database_name - name of the database table_name - table name to analyze

Returns: ResponseType: formatted response with query results + metadata

Input Schema

TableJSON Schema
NameRequiredDescriptionDefault
database_nameYes
table_nameYes

Implementation Reference

  • The handler function that implements the core logic of the qlty_columnSummary tool. It connects to Teradata, runs TD_ColumnSummary on the table, converts results to JSON, adds metadata, and returns a formatted response.
    def handle_qlty_columnSummary(conn: TeradataConnection, database_name: str | None, table_name: str, *args, **kwargs):
        """
        Get the column summary statistics for a table.
    
        Arguments:
          database_name - name of the database
          table_name - table name to analyze
    
        Returns:
          ResponseType: formatted response with query results + metadata
        """
        logger.debug(f"Tool: handle_qlty_columnSummary: Args: table_name: {database_name}.{table_name}")
    
        if database_name is not None:
                table_name = f"{database_name}.{table_name}"
        with conn.cursor() as cur:
            rows = cur.execute(f"select * from TD_ColumnSummary ( on {table_name} as InputTable using TargetColumns ('[:]')) as dt")
            data = rows_to_json(cur.description, rows.fetchall())
            metadata = {
                "tool_name": "qlty_columnSummary",
                "database_name": database_name,
                "table_name": table_name,
                "rows": len(data)
            }
            logger.debug(f"Tool: handle_qlty_columnSummary: Metadata: {metadata}")
            return create_response(data, metadata)
Behavior2/5

Does the description disclose side effects, auth requirements, rate limits, or destructive behavior?

No annotations are provided, so the description carries the full burden of behavioral disclosure. It mentions the return type ('formatted response with query results + metadata'), which adds some value, but fails to describe critical behaviors such as whether this is a read-only operation, potential performance impacts, error conditions, or data access permissions. For a tool with no annotation coverage, this leaves significant gaps.

Agents need to know what a tool does to the world before calling it. Descriptions should go beyond structured annotations to explain consequences.

Conciseness4/5

Is the description appropriately sized, front-loaded, and free of redundancy?

The description is appropriately concise and well-structured: it starts with the purpose, then lists arguments and returns in separate sections. Every sentence adds value, with no redundant information. However, the lack of detail in parameter explanations slightly reduces efficiency.

Shorter descriptions cost fewer tokens and are easier for agents to parse. Every sentence should earn its place.

Completeness2/5

Given the tool's complexity, does the description cover enough for an agent to succeed on first attempt?

Given the complexity of statistical analysis tools, no annotations, 0% schema coverage, and no output schema, the description is incomplete. It doesn't explain what 'column summary statistics' include, how results are formatted, or any limitations. The return statement is vague, and critical context for safe and effective use is missing.

Complex tools with many parameters or behaviors need more documentation. Simple tools need less. This dimension scales expectations accordingly.

Parameters2/5

Does the description clarify parameter syntax, constraints, interactions, or defaults beyond what the schema provides?

Schema description coverage is 0%, meaning the input schema provides no descriptions for parameters. The description lists the parameters ('database_name', 'table_name') but only repeats their names without explaining what they represent, valid formats, or examples. It doesn't compensate for the schema's lack of documentation, leaving parameters poorly understood.

Input schemas describe structure but not intent. Descriptions should explain non-obvious parameter relationships and valid value ranges.

Purpose4/5

Does the description clearly state what the tool does and how it differs from similar tools?

The description clearly states the tool's purpose: 'Get the column summary statistics for a table.' This specifies the verb ('Get'), resource ('column summary statistics'), and target ('table'), making it easy to understand what the tool does. However, it doesn't explicitly differentiate from sibling tools like 'qlty_univariateStatistics' or 'qlty_distinctCategories', which may also provide statistical analyses.

Agents choose between tools based on descriptions. A clear purpose with a specific verb and resource helps agents select the right tool.

Usage Guidelines2/5

Does the description explain when to use this tool, when not to, or what alternatives exist?

The description provides no guidance on when to use this tool versus alternatives. It doesn't mention sibling tools, prerequisites, or specific contexts for application. The agent must infer usage from the purpose alone, which is insufficient for optimal tool selection.

Agents often have multiple tools that could apply. Explicit usage guidance like "use X instead of Y when Z" prevents misuse.

Install Server

Other Tools

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/Teradata/teradata-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server