Tuning Engines
Server Configuration
Describes the environment variables required to run the server.
| Name | Required | Description | Default |
|---|---|---|---|
| TE_API_KEY | Yes | Your Tuning Engines API key, used for authenticating with the server. | |
| TE_API_URL | No | The API URL for the Tuning Engines service. | https://app.tuningengines.com |
Capabilities
Features and capabilities supported by this server
| Capability | Details |
|---|---|
| tools | {} |
Tools
Functions exposed to the LLM to take actions
| Name | Description |
|---|---|
| list_jobsA | List fine-tuning training jobs on Tuning Engines. Returns recent jobs with status, base model, agent type, GPU usage, and cost. Use this to check on existing training runs or find a job ID. |
| show_jobA | Get full details of a specific fine-tuning job including status, base model, agent type, GPU minutes, cost, error messages, and whether it can be retried from checkpoint. |
| create_jobA | Fine-tune an LLM on a GitHub repository using Tuning Engines. This trains a custom model that learns from the code patterns, style, and conventions in the repo. Choose an agent to control the training approach: AVAILABLE AGENTS:
SUPPORTED BASE MODELS (by size):
TYPICAL WORKFLOW: estimate_job first to check cost, then create_job, then job_status to monitor progress. |
| cancel_jobA | Cancel a running or queued fine-tuning job. The job will be charged for any GPU time already used. |
| job_statusA | Get live status of a fine-tuning job including current status, GPU minutes used, estimated charges, remaining balance, and delivery progress. Use this to monitor a running job. |
| retry_jobA | Retry a failed fine-tuning job from its last checkpoint. Creates a new job that resumes training where the failed one stopped, saving GPU time. Each retry is billed separately. IMPORTANT: This tool fetches a cost estimate and includes it in the response. You MUST show the estimate to the user and get their explicit approval before considering the retry confirmed. The retry is submitted automatically (the server validates balance), but always present the cost to the user. |
| estimate_jobA | Get a cost estimate for a fine-tuning job before submitting it. Returns estimated cost, cost range, current balance, and whether balance is sufficient. Always estimate before creating a job. |
| validate_s3A | Validate S3 credentials by testing read/write access to the specified bucket. Use before submitting a job with S3 export. |
| list_modelsB | List your trained and imported models on Tuning Engines. |
| show_modelB | Get details of a specific trained model. |
| delete_modelB | Delete a trained model from cloud storage. |
| get_balanceB | Check your Tuning Engines account balance and recent transactions. |
| get_accountA | Get your Tuning Engines account details and settings. |
| list_supported_modelsA | List the supported base HuggingFace models available for fine-tuning on Tuning Engines. Optionally filter by agent to see only compatible models. |
| import_modelA | Import a model from S3 into Tuning Engines cloud storage so it can be used as a base for future fine-tuning jobs. |
| export_modelA | Export a trained model from Tuning Engines cloud storage to your S3 bucket. |
| model_statusB | Check the status of a model import or export operation. |
| list_catalog_modelsA | List available pre-built models and datasets from the Tuning Engines Marketplace. These are platform-owned, ready-to-use assets that can be exported to your S3 bucket. Returns name, description, base model, size, export price, and category. |
| get_catalog_modelA | Get detailed information about a specific pre-built model or dataset from the Marketplace including description, pricing, and export options. |
| export_catalog_modelA | Export a pre-built model or dataset from the Marketplace to your S3 bucket. Credits will be charged based on the export price upon successful completion. |
| catalog_export_statusA | Check the status of a Marketplace export operation. Returns status, charge info, and any error messages. |
| list_datasetsB | List datasets available for training and evaluation. Datasets can be uploaded from S3 and used for fine-tuning or model evaluation. |
| show_datasetA | Get details of a specific dataset including status, source, and metadata. |
| create_datasetA | Create a new dataset by importing from S3. Datasets can be used for fine-tuning or model evaluation. |
| delete_datasetC | Delete a dataset from the platform. |
| dataset_statusC | Check the status of a dataset import or processing operation. |
| list_evaluationsB | List model evaluations. Evaluations run your trained models against benchmark datasets using various evaluators to measure quality. |
| show_evaluationB | Get full details of a specific evaluation including status, scores, metrics, and comparison data. |
| create_evaluationA | Create a new model evaluation. Run your trained model or a base model against a dataset using selected evaluators. Use list_evaluators to see available evaluators (e.g. code_execution, similarity, llm_judge). |
| cancel_evaluationB | Cancel a running or queued evaluation. |
| evaluation_statusB | Get live status of an evaluation including progress and current metrics. |
| list_evaluatorsA | List available evaluators for model evaluation. Evaluators measure different aspects of model quality like code execution, similarity, or LLM-based judgment. |
| estimate_evaluationB | Get a cost estimate for an evaluation before running it. |
| list_inference_modelsA | List models available for inference through the Tuning Engines inference API. Includes both platform models and your deployed trained models. |
| inference_usageB | Get inference API usage statistics including request counts, token usage, and costs. |
| get_inference_jwtA | Get a JWT token for authenticating with the Tuning Engines inference API. Use this to make direct API calls to the inference endpoint. |
| list_agentsA | List available agents configured for your organization. Agents are AI assistants with specific capabilities and tool access. |
| show_agentA | Get details of a specific agent including capabilities, tools, and configuration. |
Prompts
Interactive templates invoked by user choice
| Name | Description |
|---|---|
No prompts | |
Resources
Contextual data attached and managed by the client
| Name | Description |
|---|---|
No resources | |
Latest Blog Posts
MCP directory API
We provide all the information about MCP servers via our MCP API.
curl -X GET 'https://glama.ai/api/mcp/v1/servers/cerebrixos-org/tuning-engines-cli'
If you have feedback or need assistance with the MCP directory API, please join our Discord server