Skip to main content
Glama
environment.go5.54 kB
/* Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ package common import ( "context" "fmt" "log" "os" "strconv" "testing" "time" "github.com/awslabs/operatorpkg/object" "github.com/onsi/gomega" "github.com/samber/lo" corev1 "k8s.io/api/core/v1" "k8s.io/apimachinery/pkg/api/resource" "k8s.io/client-go/kubernetes" "k8s.io/client-go/kubernetes/scheme" "k8s.io/client-go/rest" . "sigs.k8s.io/karpenter/pkg/utils/testing" //nolint:stylecheck,staticcheck controllerruntime "sigs.k8s.io/controller-runtime" "sigs.k8s.io/controller-runtime/pkg/cache" "sigs.k8s.io/controller-runtime/pkg/client" karpv1 "sigs.k8s.io/karpenter/pkg/apis/v1" "sigs.k8s.io/karpenter/pkg/operator" coretest "sigs.k8s.io/karpenter/pkg/test" v1 "github.com/aws/karpenter-provider-aws/pkg/apis/v1" ) type ContextKey string const ( GitRefContextKey = ContextKey("gitRef") ) type Environment struct { context.Context cancel context.CancelFunc Client client.Client Config *rest.Config KubeClient kubernetes.Interface Monitor *Monitor StartingNodeCount int } func NewEnvironment(t *testing.T) *Environment { ctx := TestContextWithLogger(t) ctx, cancel := context.WithCancel(ctx) config := NewConfig() client := NewClient(ctx, config) if val, ok := os.LookupEnv("GIT_REF"); ok { ctx = context.WithValue(ctx, GitRefContextKey, val) } gomega.SetDefaultEventuallyTimeout(16 * time.Minute) gomega.SetDefaultEventuallyPollingInterval(1 * time.Second) return &Environment{ Context: ctx, cancel: cancel, Config: config, Client: client, KubeClient: kubernetes.NewForConfigOrDie(config), Monitor: NewMonitor(ctx, client), } } func (env *Environment) Stop() { env.cancel() } func NewConfig() *rest.Config { config := controllerruntime.GetConfigOrDie() config.UserAgent = fmt.Sprintf("testing-%s", operator.Version) config.QPS = 1e6 config.Burst = 1e6 return config } func NewClient(ctx context.Context, config *rest.Config) client.Client { cache := lo.Must(cache.New(config, cache.Options{Scheme: scheme.Scheme})) lo.Must0(cache.IndexField(ctx, &corev1.Pod{}, "spec.nodeName", func(o client.Object) []string { pod := o.(*corev1.Pod) return []string{pod.Spec.NodeName} })) lo.Must0(cache.IndexField(ctx, &corev1.Event{}, "involvedObject.kind", func(o client.Object) []string { evt := o.(*corev1.Event) return []string{evt.InvolvedObject.Kind} })) lo.Must0(cache.IndexField(ctx, &corev1.Node{}, "spec.unschedulable", func(o client.Object) []string { node := o.(*corev1.Node) return []string{strconv.FormatBool(node.Spec.Unschedulable)} })) lo.Must0(cache.IndexField(ctx, &corev1.Node{}, "spec.taints[*].karpenter.sh/disrupted", func(o client.Object) []string { node := o.(*corev1.Node) _, found := lo.Find(node.Spec.Taints, func(t corev1.Taint) bool { return t.Key == karpv1.DisruptedTaintKey }) return []string{lo.Ternary(found, "true", "false")} })) c := lo.Must(client.New(config, client.Options{Scheme: scheme.Scheme, Cache: &client.CacheOptions{Reader: cache}})) go func() { lo.Must0(cache.Start(ctx)) }() if !cache.WaitForCacheSync(ctx) { log.Fatalf("cache failed to sync") } return c } func (env *Environment) DefaultNodePool(nodeClass *v1.EC2NodeClass) *karpv1.NodePool { nodePool := coretest.NodePool() nodePool.Spec.Template.Spec.NodeClassRef = &karpv1.NodeClassReference{ Group: object.GVK(nodeClass).Group, Kind: object.GVK(nodeClass).Kind, Name: nodeClass.Name, } nodePool.Spec.Template.Spec.Requirements = []karpv1.NodeSelectorRequirementWithMinValues{ { NodeSelectorRequirement: corev1.NodeSelectorRequirement{ Key: corev1.LabelOSStable, Operator: corev1.NodeSelectorOpIn, Values: []string{string(corev1.Linux)}, }, }, { NodeSelectorRequirement: corev1.NodeSelectorRequirement{ Key: karpv1.CapacityTypeLabelKey, Operator: corev1.NodeSelectorOpIn, Values: []string{karpv1.CapacityTypeOnDemand}, }, }, { NodeSelectorRequirement: corev1.NodeSelectorRequirement{ Key: v1.LabelInstanceCategory, Operator: corev1.NodeSelectorOpIn, Values: []string{"c", "m", "r"}, }, }, { NodeSelectorRequirement: corev1.NodeSelectorRequirement{ Key: v1.LabelInstanceGeneration, Operator: corev1.NodeSelectorOpGt, Values: []string{"4"}, }, }, // Filter out a1 instance types, which are incompatible with AL2023 AMIs { NodeSelectorRequirement: corev1.NodeSelectorRequirement{ Key: v1.LabelInstanceFamily, Operator: corev1.NodeSelectorOpNotIn, Values: []string{"a1"}, }, }, } nodePool.Spec.Disruption.ConsolidationPolicy = karpv1.ConsolidationPolicyWhenEmptyOrUnderutilized nodePool.Spec.Disruption.ConsolidateAfter = karpv1.MustParseNillableDuration("Never") nodePool.Spec.Template.Spec.ExpireAfter.Duration = nil nodePool.Spec.Limits = karpv1.Limits(corev1.ResourceList{ corev1.ResourceCPU: resource.MustParse("1000"), corev1.ResourceMemory: resource.MustParse("1000Gi"), }) return nodePool }

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/mengfwan/test-mcp-glama'

If you have feedback or need assistance with the MCP directory API, please join our Discord server