Human Annotations

Human annotations are an important step in any AI evaluation system. Human reviewed scores on AI system outputs can come from engineers, domain experts (either in-house or external), or even the product's end users. Our human annotations enable you to do the following

  • Human review AI system outputs in experiments or live monitoring
  • Validate agreement of automated metrics, such as LLM judges or Patronus evaluators
  • Store comments on logs
  • Create more challenging test datasets

Create Annotation Criteria through the Annotations Tab

Navigate to the Annotations tab.


Create Annotation Criteria from the Logs tab


Note 📝:

The project of the log where you create a human annotation from will be passed along and pre-populate your selected project field!


Human Annotation Criteria Fields

Human Annotation Criteria FieldRequired
Project
Name
Description
2+ Categories(only for "Categorical" and "Discrete")

Tip 👀:

Use the "Preview" section at the bottom to see how the human annotation criteria would look on the logs page!


Start Adding Human Annotations

Now to finally start adding human annotations, navigate to the "Logs" tab and select a log. On the right side you should see a section titled "Annotations" that you can use to provide human annotations.