DocumentationAPI ReferenceRelease Notes
DocumentationLog In
Documentation

Data Fields Reference

Complete reference for every data field Deepchecks supports - content fields, metadata fields, and span fields - with descriptions and usage guidance.

This page is a complete reference for every field you can include when sending data to Deepchecks - whether via the Python SDK, CSV upload, or auto-instrumentation. Use it to understand what each field does and what evaluation or observability features it enables.

Using auto-instrumentation? If you are sending data via framework integrations (LangGraph, CrewAI, Google ADK, LangChain), most fields are filled in automatically from parsed span attributes. This reference is most useful when uploading data via the SDK or CSV, where you control exactly which fields to include.


Interaction content fields

Content fields contain the actual text from the interaction and are used for property calculations. Which fields are relevant depends on your interaction type.

FieldRequiredTypeDescriptionWhat it enables
inputRequiredstrThe input to the LLM pipeline - the user's question, request, or source textRequired for most properties. Without it, Deepchecks cannot evaluate the interaction.
outputRequiredstrThe pipeline's response or generated contentRequired for most properties. Without it, Deepchecks cannot evaluate the interaction.
full_promptOptionalstrThe complete prompt sent to the LLM, including system instructionsRequired for LLM-based properties. Enables Instruction Following and Instruction Fulfillment. Displayed in the interaction detail view.
information_retrievalOptionallist of strDocuments or data retrieved as context for the LLM (e.g., RAG results, database query results)Enables retrieval-based properties: Grounded in Context, Retrieval Relevance, Retrieval Coverage
historyOptionallist of strAdditional context relevant to the interaction that was not retrieved from a knowledge base - for example, chat conversation historyEnables history-aware evaluation in Chat and multi-turn interaction types
expected_outputOptionalstrA reference output serving as ground truth - typically created by a human annotator or deemed high-quality through reviewEnables similarity-based evaluation and reference comparison
stepsOptionallist of StepIntermediate processing steps within the interaction (e.g., query rephrasing, routing, PII removal). Can be used in property calculations.Displayed in the interaction detail view. Provides visibility into intermediate processing. See Python SDK Integration.

Interaction metadata fields

Metadata fields are used for organization, filtering, annotations, system metrics, and version comparison.

FieldRequiredTypeDescriptionWhat it enables
user_interaction_idOptionalstrUnique identifier for the interaction within a version. Auto-generated if not provided. Providing your own stable ID is recommended when you need to match interactions across versions or update them after upload.Matching interactions across versions for comparison. Updating annotations or properties after upload.
session_idOptionalstrGroups related interactions into a session (e.g., a conversation, a multi-step workflow). Auto-generated if not provided.Session-level evaluation, Sessions view in the UI, grouping multi-turn conversations
interaction_typeOptionalstrThe type of interaction (e.g., Q&A, Summarization, Agent). Defaults to the application's default type if not provided.Determines which built-in properties are enabled, which auto-annotation rules apply, and how the interaction is grouped in the UI
started_atOptionaltimestampWhen the interaction started. Accepts ISO 8601 format (e.g., 2025-01-01T00:00:01+00:00) or Unix epoch (e.g., 1742742893).Latency calculation (with finished_at), production monitoring timeline
finished_atOptionaltimestampWhen the interaction ended. Same format as started_at.Latency calculation, production monitoring
modelOptionalstrThe model name used for the LLM call (e.g., gpt-4o, claude-3-5-sonnet).Cost tracking (combined with model pricing configuration), filtering by model
model_providerOptionalstrThe model provider (e.g., openai, anthropic).Cost tracking, filtering by provider
input_tokensOptionalintNumber of prompt tokens consumed by the LLM.Token usage metrics, cost calculation
output_tokensOptionalintNumber of completion tokens generated by the LLM.Token usage metrics, cost calculation
tokensOptionalintTotal token count. Auto-calculated from input_tokens + output_tokens if not provided.Token usage metrics
annotationOptionalstrHuman annotation label: Good, Bad, or UnknownAppears as a filled badge in the UI. Takes precedence over estimated annotations.
annotation_reasonOptionalstrTextual reasoning for the annotationDisplayed alongside the annotation for context

Span-specific fields

These fields are used when uploading hierarchical agentic data via the Span class and log_spans function. See Upload Agentic Data for full details.

Required span fields

FieldTypeDescription
span_idstrUnique identifier for the span
span_namestrDescriptive name of the operation
trace_idstrIdentifier grouping spans into a single trace
span_kindSpanKindType of span: CHAIN, AGENT, TOOL, LLM, RETRIEVAL
parent_idstr or Nonespan_id of the parent span (None for the Root span)
started_atfloatStart timestamp (Unix epoch)
finished_atfloatEnd timestamp (Unix epoch)
inputstrData passed into the operation. Required for property calculation - without it, Deepchecks cannot evaluate this span.
outputstrData returned by the operation. Required for property calculation - without it, Deepchecks cannot evaluate this span.
full_promptstrThe complete prompt sent to the LLM. Required for LLM spans - enables Instruction Following and Instruction Fulfillment properties.

Optional span fields

FieldTypeDescription
expected_outputstrExpected result for evaluation
information_retrievallistRetrieved documents or data
modelstrModel name (e.g., gpt-4o)
model_providerstrModel provider (e.g., openai)
input_tokensintNumber of prompt tokens
output_tokensintNumber of completion tokens
tokensintTotal tokens (auto-calculated from input + output if not provided)
status_codestrExecution status (e.g., OK, ERROR)
status_descriptionstrAdditional context about the status
graph_parent_namestrLogical parent span name (used for graph visualization)
session_idstrGroups related traces into a session
metadatadictCustom key-value properties. Raw attributes stored here are parsed by Deepchecks into structured fields.
stepslist of StepIntermediate processing steps within the span
user_value_propertieslist of UserValuePropertyCustom numeric or categorical properties

System metrics

System metrics are computed from metadata fields and provide operational observability alongside quality evaluation:

MetricSourceDescription
Latencystarted_at + finished_atTime to process the interaction
Input tokensinput_tokens field or auto-instrumentationNumber of prompt tokens consumed
Output tokensoutput_tokens field or auto-instrumentationNumber of completion tokens generated
Total tokenstokens field or sum of input + outputTotal token consumption
CostToken counts + model pricing configurationMonetary cost per interaction (configure model pricing in Workspace Settings). Broken into input_cost and output_cost per span.
Run statusstatus_code field or auto-instrumentationWhether the span executed successfully

In agentic pipelines, system metrics are aggregated across child spans - so you can see total token usage and cost for an entire agent trace, not just individual LLM calls.

When using auto-instrumentation, most system metrics are captured automatically from parsed framework attributes. For manual uploads via the SDK or CSV, include timestamps, model info, and token counts to get a complete observability picture.


Field requirements by use case

Not every interaction type needs every field. Here is a quick guide to what matters most for common use cases:

Use caseEssential fieldsRecommended fields
Q&A / RAGinput, output, information_retrievalfull_prompt, expected_output, timestamps, token fields
Summarizationinput, outputfull_prompt, timestamps, token fields
Generationinput, outputfull_prompt, information_retrieval, timestamps
Classificationinput, outputexpected_output, full_prompt
Chatinput, output, historyfull_prompt, session_id, timestamps
Agenticspan_id, trace_id, span_kind, parent_id, timestamps, input, output, full_prompt (LLM spans)model, input_tokens, output_tokens, status_code

See Supported Use Cases for a full breakdown of interaction types and their properties.