Skip to main content
Glama

Genkit MCP

Official
by firebase
cache_test.go4.11 kB
// Copyright 2025 Google LLC // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // // SPDX-License-Identifier: Apache-2.0 package googlegenai import ( "strings" "testing" "github.com/firebase/genkit/go/ai" ) func TestGetContentForCache_NoCacheMetadata(t *testing.T) { req := &ai.ModelRequest{ Messages: []*ai.Message{ { Role: ai.RoleUser, Content: []*ai.Part{ {Text: "Hello"}, }, }, }, } gotContent, err := findCacheMarker(req) if err != nil { t.Fatalf("unexpected error: %v", err) } if gotContent != nil { t.Errorf("expected nil content when no cache metadata, got: %#v", gotContent) } } func TestGetContentForCache_NoContentToCache(t *testing.T) { req := &ai.ModelRequest{ Messages: []*ai.Message{ { Role: ai.RoleUser, Metadata: map[string]any{"cache": map[string]any{ "ttlSeconds": int(160), }}, // No text content }, }, } _, err := findCacheMarker(req) if err == nil { t.Fatalf("should fail due no text in message") } } func TestGetContentForCache_Invalid(t *testing.T) { req := &ai.ModelRequest{ Messages: []*ai.Message{ { Role: ai.RoleSystem, Content: []*ai.Part{ {Text: "System instructions"}, }, }, { Role: ai.RoleUser, Content: []*ai.Part{{Text: "Hello user"}}, Metadata: map[string]any{"cache": map[string]any{ "ttlSeconds": 160, }}, }, }, } err := validateContextCacheRequest(req, "gemini-1.5-fash-001") if err == nil { t.Fatal("expecting error, system instructions are not supported with Context Cache") } } func TestValidateContextCacheRequest_HasTools(t *testing.T) { req := &ai.ModelRequest{ Tools: []*ai.ToolDefinition{{Name: "someTool"}}, } err := validateContextCacheRequest(req, "gemini-1.5-flash-001") if err == nil { t.Fatal("expected error if Tools are present") } if !strings.Contains(err.Error(), invalidArgMessages.tools) { t.Errorf("expected error to contain %q, got %v", invalidArgMessages.tools, err) } } func TestValidateContextCacheRequest_Valid(t *testing.T) { req := &ai.ModelRequest{} err := validateContextCacheRequest(req, "gemini-1.5-flash-001") if err != nil { t.Fatalf("did not expect error, got: %v", err) } } func TestExtractCacheConfig_NoMetadata(t *testing.T) { req := &ai.ModelRequest{ Messages: []*ai.Message{ {Role: ai.RoleUser, Content: []*ai.Part{{Text: "Hello"}}}, }, } cs, err := findCacheMarker(req) if err != nil { t.Errorf("unexpected error: %v", err) } if cs != nil { t.Fatalf("expecting cache settings to be nil, got %#v", cs) } } func TestExtractCacheConfig_MapTTL(t *testing.T) { req := &ai.ModelRequest{ Messages: []*ai.Message{ { Role: ai.RoleUser, Content: []*ai.Part{ {Text: "Hello"}, }, Metadata: map[string]any{ "cache": map[string]any{ "ttlSeconds": int(123), }, }, }, }, } cs, err := findCacheMarker(req) if err != nil { t.Fatalf("unexpected error: %v", err) } if cs.endIndex != 0 { t.Errorf("expected endIndex=0, got %d", cs.endIndex) } if cs.ttl != 123 { t.Errorf("expected TTLSeconds=123, got %v", cs.ttl) } } func TestExtractCacheConfig_InvalidCacheType(t *testing.T) { req := &ai.ModelRequest{ Messages: []*ai.Message{ { Role: ai.RoleUser, Content: []*ai.Part{ {Text: "Hello"}, }, Metadata: map[string]any{ "cache": []string{"not valid"}, }, }, }, } cs, err := findCacheMarker(req) if err == nil { t.Fatal("expected error for invalid cache type") } if cs != nil { t.Fatalf("expecting empty cache settings but got: %v", cs) } }

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/firebase/genkit'

If you have feedback or need assistance with the MCP directory API, please join our Discord server