DocumentationAPI ReferenceRelease Notes
DocumentationLog In
Documentation

Identify Problems Using Properties, Estimated Annotations and Insights

Find Problematic Samples based on Built-in Properties

In this section we'll identify problematic behaviors in our evaluation sets using a few of Deepchecks’ Built-in calculated properties: "Grounded in Context" and "Retrieval Relevance".
In addition, we'll have a look at the "Avoided Answer" property as well, as it is sometimes related to problematic findings, and thus filtering and sorting using this property can help pinpoint undesired outputs in our application.
More info about each of these properties can be seen upon hovering on the property in the UI:

To explore bad behaviors in our Evaluation set, follow these steps:

  1. Go to the Data page in version v1_gpt3.5, and sort according to "bad" and "estimated bad". This can be done by selecting the thumbs down filters on the left side of the page.

    Notice we have a user supplied annotation reason for each sample that was annotated “bad”. The User Reason can help us highlight the major factors we gave a specific sample its annotation.

  2. Click on the property column and choose the "Grounded in Context" property. Sort the interactions from low to high.

    Click on the samples to inspect how the LLM's answers align with the context provided in the Information Retrieval. Higher scores in the "Grounded in Context" property should show better context alignment.

  3. Repeat this process with the “Retrieval Relevance” and "Avoided Answer" properties. These properties will help you assess the relevance of the retrieved information and indicate whether the LLM may have avoided answering the input question.

  4. You can add up to three property score columns in the data table for side-by-side comparison.


  5. Use property filters on the left side of the samples table to view subgroup scores alongside the overall score.

Find Problematic Samples based on Estimated Annotations

To identify problematic samples, check the auto-annotations in version v2_improved_IR. These annotations are based on properties, similarity, and the Deepchecks evaluator. For detailed criteria, refer to the Automatic Annotations page.

Note: In the GIF above, you can see a sample with a low “Retrieval Relevance” score and a high “Avoided Answer” score is automatically annotated as “good.” This configuration is designed to identify cases where the LLM avoids answering due to insufficient relevant information. Such configurations can be adjusted in the “Annotation Config” page.


Find Problematic Samples based on Insights

Insights in the Deepchecks system help identify groups of interactions with low scores and shared characteristics. They include Recommendations for improving LLM-based applications and Weak Segments showing subgroups with consistently poor annotations. For more details, see the Root Cause Analysis (RCA).

Go back to our GVHD demo and inspect the Insights calculated by the Deepchecks system for versions v1_gpt3.5 and v2_improved_IR.
For example, we can see that while both v1_gpt3.5 and v2_improved_IR versions contain samples that triggered the “Irrelevant Retrieval” Recommendations, version v2_improved_IR alone contains output answers that activated 2 more rule Recommendations - “Hallucination” and “Irrelevant Retrieval”.