Skip to main content
Glama

@arizeai/phoenix-mcp

Official
by Arize-ai
annotating-in-the-ui.md3.69 kB
--- description: How to annotate traces in the UI for analysis and dataset curation --- # Annotating in the UI ## Configuring Annotations To annotate data in the UI, you first will want to setup a rubric for how to annotate. Navigate to `Settings` and create annotation configs (e.g. a rubric) for your data. You can create various different types of annotations: **Categorical**, **Continuous**, and **Freeform**. <details> <summary>Annotation Types</summary> * Annotation Type: \ \- **Categorical:** Predefined labels for selection. (e.x. 👍 or 👎) \ \- **Continuous:** a score across a specified range. (e.g. confidence score 0-100) \ \- **Freeform:** Open-ended text comments. (e.g. "correct") - Optimize the direction based on your goal:\ \- **Maximize:** higher scores are better. (e.g. confidence) \ \- **Minimize:** lower scores are better. (e.g. hallucinations) \ \- **None:** direction optimization does not apply. (e.g. tone) <figure><img src="https://storage.googleapis.com/arize-phoenix-assets/assets/images/annotation_types.png" alt=""><figcaption><p>Different types of annotations change the way human annotators provide feedback</p></figcaption></figure> </details> <figure><img src="https://storage.googleapis.com/arize-phoenix-assets/assets/images/annotation_configuration.png" alt=""><figcaption><p>Configure an annotation to guide how a user should input an annotation</p></figcaption></figure> ## Adding Annotations <figure><img src="https://storage.googleapis.com/arize-phoenix-assets/assets/images/annotate_trace.png" alt=""><figcaption><p>Once annotations are configured, you can add them to your project to build out a custom annotation form</p></figcaption></figure> Once you have annotations configured, you can associate annotations to the data that you have traced. Click on the `Annotate` button and fill out the form to rate different steps in your AI application. \ \ You can also take notes as you go by either clicking on the `explain` link or by adding your notes to the bottom messages UI.\ \ You can always come back and edit / and delete your annotations. Annotations can be deleted from the table view under the `Annotations` tab. {% hint style="success" %} Once an annotation has been provided, you can also add a reason to explain why this particular label or score was provided. This is useful to add additional context to the annotation. {% endhint %} ## Viewing Annotations As annotations come in from various sources (annotators, evals), the entire list of annotations can be found under the `Annotations` tab. Here you can see the author, the annotator kind (e.g. was the annotation performed by a human, llm, or code), and so on. This can be particularly useful if you want to see if different annotators disagree. <figure><img src="https://storage.googleapis.com/arize-phoenix-assets/assets/images/annotation_table.png" alt=""><figcaption><p>You can view the annotations by different users, llms, and annotators</p></figcaption></figure> ## Exporting Traces with specific Annotation Values Once you have collected feedback in the form of annotations, you can filter your traces by the annotation values to narrow down to interesting samples (e.x. llm spans that are incorrect). Once filtered down to a sample of spans, you can export your selection to a dataset, which in turn can be used for things like experimentation, fine-tuning, or building a human-aligned eval. <figure><img src="https://storage.googleapis.com/arize-phoenix-assets/assets/images/annotation_filters.png" alt=""><figcaption><p>Narrow down your data to areas that need more attention or refinement</p></figcaption></figure>

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/Arize-ai/phoenix'

If you have feedback or need assistance with the MCP directory API, please join our Discord server