Skip to main content
Glama

MCP Toolbox for Databases

by googleapis
Apache 2.0
11,060
  • Linux
bigqueryforecast.go12.1 kB
// Copyright 2025 Google LLC // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package bigqueryforecast import ( "context" "fmt" "strings" bigqueryapi "cloud.google.com/go/bigquery" yaml "github.com/goccy/go-yaml" "github.com/googleapis/genai-toolbox/internal/sources" bigqueryds "github.com/googleapis/genai-toolbox/internal/sources/bigquery" "github.com/googleapis/genai-toolbox/internal/tools" bqutil "github.com/googleapis/genai-toolbox/internal/tools/bigquery/bigquerycommon" "github.com/googleapis/genai-toolbox/internal/util" bigqueryrestapi "google.golang.org/api/bigquery/v2" "google.golang.org/api/iterator" ) const kind string = "bigquery-forecast" func init() { if !tools.Register(kind, newConfig) { panic(fmt.Sprintf("tool kind %q already registered", kind)) } } func newConfig(ctx context.Context, name string, decoder *yaml.Decoder) (tools.ToolConfig, error) { actual := Config{Name: name} if err := decoder.DecodeContext(ctx, &actual); err != nil { return nil, err } return actual, nil } type compatibleSource interface { BigQueryClient() *bigqueryapi.Client BigQueryRestService() *bigqueryrestapi.Service BigQueryClientCreator() bigqueryds.BigqueryClientCreator UseClientAuthorization() bool IsDatasetAllowed(projectID, datasetID string) bool BigQueryAllowedDatasets() []string BigQuerySession() bigqueryds.BigQuerySessionProvider } // validate compatible sources are still compatible var _ compatibleSource = &bigqueryds.Source{} var compatibleSources = [...]string{bigqueryds.SourceKind} type Config struct { Name string `yaml:"name" validate:"required"` Kind string `yaml:"kind" validate:"required"` Source string `yaml:"source" validate:"required"` Description string `yaml:"description" validate:"required"` AuthRequired []string `yaml:"authRequired"` } // validate interface var _ tools.ToolConfig = Config{} func (cfg Config) ToolConfigKind() string { return kind } func (cfg Config) Initialize(srcs map[string]sources.Source) (tools.Tool, error) { // verify source exists rawS, ok := srcs[cfg.Source] if !ok { return nil, fmt.Errorf("no source named %q configured", cfg.Source) } // verify the source is compatible s, ok := rawS.(compatibleSource) if !ok { return nil, fmt.Errorf("invalid source for %q tool: source kind must be one of %q", kind, compatibleSources) } allowedDatasets := s.BigQueryAllowedDatasets() historyDataDescription := "The table id or the query of the history time series data." if len(allowedDatasets) > 0 { datasetIDs := []string{} for _, ds := range allowedDatasets { datasetIDs = append(datasetIDs, fmt.Sprintf("`%s`", ds)) } historyDataDescription += fmt.Sprintf(" The query or table must only access datasets from the following list: %s.", strings.Join(datasetIDs, ", ")) } historyDataParameter := tools.NewStringParameter("history_data", historyDataDescription) timestampColumnNameParameter := tools.NewStringParameter("timestamp_col", "The name of the time series timestamp column.") dataColumnNameParameter := tools.NewStringParameter("data_col", "The name of the time series data column.") idColumnNameParameter := tools.NewArrayParameterWithDefault("id_cols", []any{}, "An array of the time series id column names.", tools.NewStringParameter("id_col", "The name of time series id column.")) horizonParameter := tools.NewIntParameterWithDefault("horizon", 10, "The number of forecasting steps.") parameters := tools.Parameters{historyDataParameter, timestampColumnNameParameter, dataColumnNameParameter, idColumnNameParameter, horizonParameter} mcpManifest := tools.GetMcpManifest(cfg.Name, cfg.Description, cfg.AuthRequired, parameters) // finish tool setup t := Tool{ Name: cfg.Name, Kind: kind, Parameters: parameters, AuthRequired: cfg.AuthRequired, UseClientOAuth: s.UseClientAuthorization(), ClientCreator: s.BigQueryClientCreator(), Client: s.BigQueryClient(), RestService: s.BigQueryRestService(), IsDatasetAllowed: s.IsDatasetAllowed, SessionProvider: s.BigQuerySession(), AllowedDatasets: allowedDatasets, manifest: tools.Manifest{Description: cfg.Description, Parameters: parameters.Manifest(), AuthRequired: cfg.AuthRequired}, mcpManifest: mcpManifest, } return t, nil } // validate interface var _ tools.Tool = Tool{} type Tool struct { Name string `yaml:"name"` Kind string `yaml:"kind"` AuthRequired []string `yaml:"authRequired"` UseClientOAuth bool `yaml:"useClientOAuth"` Parameters tools.Parameters `yaml:"parameters"` Client *bigqueryapi.Client RestService *bigqueryrestapi.Service ClientCreator bigqueryds.BigqueryClientCreator IsDatasetAllowed func(projectID, datasetID string) bool AllowedDatasets []string SessionProvider bigqueryds.BigQuerySessionProvider manifest tools.Manifest mcpManifest tools.McpManifest } func (t Tool) Invoke(ctx context.Context, params tools.ParamValues, accessToken tools.AccessToken) (any, error) { paramsMap := params.AsMap() historyData, ok := paramsMap["history_data"].(string) if !ok { return nil, fmt.Errorf("unable to cast history_data parameter %v", paramsMap["history_data"]) } timestampCol, ok := paramsMap["timestamp_col"].(string) if !ok { return nil, fmt.Errorf("unable to cast timestamp_col parameter %v", paramsMap["timestamp_col"]) } dataCol, ok := paramsMap["data_col"].(string) if !ok { return nil, fmt.Errorf("unable to cast data_col parameter %v", paramsMap["data_col"]) } idColsRaw, ok := paramsMap["id_cols"].([]any) if !ok { return nil, fmt.Errorf("unable to cast id_cols parameter %v", paramsMap["id_cols"]) } var idCols []string for _, v := range idColsRaw { s, ok := v.(string) if !ok { return nil, fmt.Errorf("id_cols contains non-string value: %v", v) } idCols = append(idCols, s) } horizon, ok := paramsMap["horizon"].(int) if !ok { if h, ok := paramsMap["horizon"].(float64); ok { horizon = int(h) } else { return nil, fmt.Errorf("unable to cast horizon parameter %v", paramsMap["horizon"]) } } bqClient := t.Client restService := t.RestService var err error // Initialize new client if using user OAuth token if t.UseClientOAuth { tokenStr, err := accessToken.ParseBearerToken() if err != nil { return nil, fmt.Errorf("error parsing access token: %w", err) } bqClient, restService, err = t.ClientCreator(tokenStr, false) if err != nil { return nil, fmt.Errorf("error creating client from OAuth access token: %w", err) } } var historyDataSource string trimmedUpperHistoryData := strings.TrimSpace(strings.ToUpper(historyData)) if strings.HasPrefix(trimmedUpperHistoryData, "SELECT") || strings.HasPrefix(trimmedUpperHistoryData, "WITH") { if len(t.AllowedDatasets) > 0 { var connProps []*bigqueryapi.ConnectionProperty session, err := t.SessionProvider(ctx) if err != nil { return nil, fmt.Errorf("failed to get BigQuery session: %w", err) } if session != nil { connProps = []*bigqueryapi.ConnectionProperty{ {Key: "session_id", Value: session.ID}, } } dryRunJob, err := bqutil.DryRunQuery(ctx, restService, t.Client.Project(), t.Client.Location, historyData, nil, connProps) if err != nil { return nil, fmt.Errorf("query validation failed: %w", err) } statementType := dryRunJob.Statistics.Query.StatementType if statementType != "SELECT" { return nil, fmt.Errorf("the 'history_data' parameter only supports a table ID or a SELECT query. The provided query has statement type '%s'", statementType) } queryStats := dryRunJob.Statistics.Query if queryStats != nil { for _, tableRef := range queryStats.ReferencedTables { if !t.IsDatasetAllowed(tableRef.ProjectId, tableRef.DatasetId) { return nil, fmt.Errorf("query in history_data accesses dataset '%s.%s', which is not in the allowed list", tableRef.ProjectId, tableRef.DatasetId) } } } else { return nil, fmt.Errorf("could not analyze query in history_data to validate against allowed datasets") } } historyDataSource = fmt.Sprintf("(%s)", historyData) } else { if len(t.AllowedDatasets) > 0 { parts := strings.Split(historyData, ".") var projectID, datasetID string switch len(parts) { case 3: // project.dataset.table projectID = parts[0] datasetID = parts[1] case 2: // dataset.table projectID = t.Client.Project() datasetID = parts[0] default: return nil, fmt.Errorf("invalid table ID format for 'history_data': %q. Expected 'dataset.table' or 'project.dataset.table'", historyData) } if !t.IsDatasetAllowed(projectID, datasetID) { return nil, fmt.Errorf("access to dataset '%s.%s' (from table '%s') is not allowed", projectID, datasetID, historyData) } } historyDataSource = fmt.Sprintf("TABLE `%s`", historyData) } idColsArg := "" if len(idCols) > 0 { idColsFormatted := fmt.Sprintf("['%s']", strings.Join(idCols, "', '")) idColsArg = fmt.Sprintf(", id_cols => %s", idColsFormatted) } sql := fmt.Sprintf(`SELECT * FROM AI.FORECAST( %s, data_col => '%s', timestamp_col => '%s', horizon => %d%s)`, historyDataSource, dataCol, timestampCol, horizon, idColsArg) // JobStatistics.QueryStatistics.StatementType query := bqClient.Query(sql) query.Location = bqClient.Location session, err := t.SessionProvider(ctx) if err != nil { return nil, fmt.Errorf("failed to get BigQuery session: %w", err) } if session != nil { // Add session ID to the connection properties for subsequent calls. query.ConnectionProperties = []*bigqueryapi.ConnectionProperty{ {Key: "session_id", Value: session.ID}, } } // Log the query executed for debugging. logger, err := util.LoggerFromContext(ctx) if err != nil { return nil, fmt.Errorf("error getting logger: %s", err) } logger.DebugContext(ctx, "executing `%s` tool query: %s", kind, sql) // This block handles SELECT statements, which return a row set. // We iterate through the results, convert each row into a map of // column names to values, and return the collection of rows. var out []any job, err := query.Run(ctx) if err != nil { return nil, fmt.Errorf("unable to execute query: %w", err) } it, err := job.Read(ctx) if err != nil { return nil, fmt.Errorf("unable to read query results: %w", err) } for { var row map[string]bigqueryapi.Value err = it.Next(&row) if err == iterator.Done { break } if err != nil { return nil, fmt.Errorf("unable to iterate through query results: %w", err) } vMap := make(map[string]any) for key, value := range row { vMap[key] = value } out = append(out, vMap) } // If the query returned any rows, return them directly. if len(out) > 0 { return out, nil } // This handles the standard case for a SELECT query that successfully return "The query returned 0 rows.", nil } func (t Tool) ParseParams(data map[string]any, claims map[string]map[string]any) (tools.ParamValues, error) { return tools.ParseParams(t.Parameters, data, claims) } func (t Tool) Manifest() tools.Manifest { return t.manifest } func (t Tool) McpManifest() tools.McpManifest { return t.mcpManifest } func (t Tool) Authorized(verifiedAuthServices []string) bool { return tools.IsAuthorized(t.AuthRequired, verifiedAuthServices) } func (t Tool) RequiresClientAuthorization() bool { return t.UseClientOAuth }

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/googleapis/genai-toolbox'

If you have feedback or need assistance with the MCP directory API, please join our Discord server