# FastAgent Configuration File
# Default Model Configuration:
#
# Takes format:
# <provider>.<model_string>.<reasoning_effort?> (e.g. anthropic.claude-3-5-sonnet-20241022 or openai.o3-mini.low)
# Accepts aliases for Anthropic Models: haiku, haiku3, sonnet, sonnet35, opus, opus3
# and OpenAI Models: gpt-4.1, gpt-4.1-mini, o1, o1-mini, o3-mini
#
# If not specified, defaults to "gpt-5-mini.low".
# Can be overriden with a command line switch --model=<model>, or within the Agent constructor.
default_model: haiku
# mcp-ui support: disabled, enabled or auto. "auto" opens the web browser on the asset automatically
# mcp_ui_output_dir: ".fast-agent/ui" # Where to write MCP-UI HTML files (relative to CWD if not absolute)
# mcp_ui_mode: enabled
# MCP timeline display (adjust activity window/intervals in MCP UI + fast-agent check)
mcp_timeline:
steps: 20 # number of timeline buckets to render
step_seconds: 15 # seconds per bucket (accepts values like "45s", "2m")
#shell_execution:
# length of time before terminating subprocess
# timeout_seconds: 20
# warning interval if no output seen
# warning_seconds: 5
# Logging and Console Configuration:
logger:
# level: "debug" # | "info" | "warning" | "error"
# type: "none" | "console" | "file" | "http"
# path: "/path/to/logfile.jsonl"
# Switch the progress display on or off
progress_display: true
# Show chat User/Assistant messages on the console
show_chat: true
# Show tool calls on the console
show_tools: true
# Truncate long tool responses on the console
truncate_tools: true
# Streaming renderer for assistant responses: "markdown", "plain", or "none"
streaming: markdown
# MCP Servers
mcp:
servers:
chemspace:
command: "uv"
args: ["run", "chemspace-mcp"]