Skip to main content
Glama

@arizeai/phoenix-mcp

Official
by Arize-ai
how-do-i-resolve-phoenix-evals-showing-not_parsable.md926 B
# How do I resolve Phoenix Evals showing NOT\_PARSABLE? `NOT_PARSABLE` errors often occur when LLM responses exceed the `max_tokens` limit or produce incomplete JSON. Here's how to fix it: 1. Increase `max_tokens`: Update the model configuration as follows: ```python pythonCopy codellm_judge_model = OpenAIModel( api_key=getpass("Enter your OpenAI API key..."), model="gpt-4o-2024-08-06", temperature=0.2, max_tokens=1000, # Increase token limit ) ``` 2. Update Phoenix: Use version ≥0.17.4, which removes token limits for OpenAI and increases defaults for other APIs. 3. Check Logs: Look for `finish_reason="length"` to confirm token limits caused the issue. 4. If the above doesn't work, it's possible the llm-as-a-judge output might not fit into the defined rails for that particular custom Phoenix eval. Double check the prompt output matches the rail expectations.

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/Arize-ai/phoenix'

If you have feedback or need assistance with the MCP directory API, please join our Discord server