Skip to main content
Galtea allows you to run inferences against your AI system and evaluate its responses directly from the Dashboard, without writing any SDK code. This is made possible by Endpoint Connections, which tell Galtea how to call your API, extract the response, and manage session state across turns.
This guide covers the platform-based workflow. If you prefer to generate inferences programmatically (e.g., in a CI/CD pipeline or custom script), see the SDK tutorials instead.

Prerequisites

Before you begin, make sure you have the following set up in the Galtea Dashboard:
  • A Product representing your AI system
  • A Test with at least one Test Case to run against your endpoint

Workflow Overview

1

Create an Endpoint Connection

Define how Galtea should call your AI endpoint — URL, authentication, request format, and response extraction.
2

Create a Version with the Endpoint Connection

Create a new version of your product and attach the endpoint connection to it.
3

Run a Test from the Dashboard

Select a test and run it against the version. Galtea calls your endpoint for each test case and records the inference results.
4

Evaluate the Results

Once inferences are generated, trigger evaluations with the metrics of your choice to assess your AI’s performance.

Step 1: Create an Endpoint Connection

Navigate to your product in the Dashboard and go to the Endpoint Connections section. Click New Endpoint Connection and configure the following:
  1. Name — A descriptive name (e.g., “Production Chat API”).
  2. Type — Select CONVERSATION for the primary request/response endpoint.
  3. URL — The full URL of your AI endpoint (e.g., https://api.company.com/v1/chat).
  4. HTTP Method — Typically POST.
  5. Authentication — Choose the auth type (Bearer, API_KEY, Basic, or None) and provide the token.
  6. Input Template — A Jinja2 template that defines the request body Galtea will send.
  7. Output Mapping — JSONPath expressions that tell Galtea how to extract values from the response.

Input Template

The input template uses Jinja2 syntax with placeholders that Galtea fills automatically. At minimum, use {{ input }} to inject the test case input:
{
  "model": "gpt-4",
  "messages": [
    {"role": "system", "content": "You are a helpful assistant."},
    {"role": "user", "content": "{{ input }}"}
  ]
}
For multi-turn conversations, use past_turns to include conversation history:
{
  "model": "gpt-4",
  "messages": [
    {"role": "system", "content": "You are a helpful assistant."},
    {% for turn in past_turns %}
    {"role": "user", "content": "{{ turn.input }}"},
    {"role": "assistant", "content": "{{ turn.output }}"},
    {% endfor %}
    {"role": "user", "content": "{{ input }}"}
  ]
}
See Endpoint Connection — Input Template for the full list of available placeholders and advanced template examples.

Output Mapping

The output mapping tells Galtea how to extract values from the API response using JSONPath expressions. The output key is required:
{
  "output": "$.choices[0].message.content"
}
You can also extract additional values to store as session metadata:
{
  "output": "$.choices[0].message.content",
  "retrieval_context": "$.choices[0].retrieval_context",
  "session_id": "$.metadata.session_id"
}
Any extra keys beyond output and retrieval_context are saved to the session metadata and become available as {{ key }} placeholders in subsequent turns.
See Version — Special keys in Output Mapping for a complete reference of how extracted values are stored and reused.

Step 2: Create a Version with the Endpoint Connection

Navigate to your product and create a new Version. When configuring the version:
  1. Fill in the version name, model, and any other relevant properties.
  2. In the Conversation Endpoint Connection field, select the endpoint connection you created in Step 1.
The Conversation Endpoint Connection is the only required endpoint connection. For most integrations, this single endpoint handles the entire interaction lifecycle.
If your AI system requires separate endpoints for session initialization or cleanup, you can optionally configure Initialization and Finalization endpoint connections. See Version — Multi-Step Session Lifecycle for details.

Step 3: Run a Test

Once your version is set up with an endpoint connection, you can run tests directly from the Dashboard:
  1. Navigate to your product’s Tests section.
  2. Select the test you want to run.
  3. Choose the version with the configured endpoint connection.
  4. Start the test run.
Galtea will iterate through each test case, call your endpoint using the configured endpoint connection, and record the resulting Inference Results. Each test case produces a session with one or more inference results depending on whether it’s a single-turn or multi-turn test.

Step 4: Evaluate the Results

After the inferences have been generated, you can trigger evaluations:
  1. Navigate to the session results in the Dashboard.
  2. Select the Metrics you want to use for the evaluation.
  3. Run the evaluation.
Galtea will assess each inference result using the selected metrics and provide scores and explanations.
For single-turn tests, metrics like Factual Accuracy and Answer Relevancy work well. For multi-turn conversations, consider Knowledge Retention, Role Adherence, and Conversation Completeness.

Collecting Traces During Direct Inference

There are two ways to collect traces during Direct Inference:
  1. Output Mapping (no code) — Extract traces from the API response using a traces key in your output mapping.
  2. SDK set_context (in your handler) — Pass {{ inference_result_id }} to your endpoint and use the SDK to create traces from within the handler.

Option 1: Extract Traces via Output Mapping

If your endpoint returns trace data in its response, you can extract it using the traces key in the output mapping. Galtea will store each trace object linked to the inference result automatically. Example API response:
{
  "response": "The answer is 42.",
  "traces": [
    {
      "name": "retrieve_context",
      "type": "RETRIEVER",
      "latencyMs": 120,
      "inputData": { "query": "meaning of life" },
      "outputData": { "documents": ["..."] }
    },
    {
      "name": "generate_response",
      "type": "GENERATION",
      "latencyMs": 350,
      "inputData": { "context": "..." },
      "outputData": { "text": "The answer is 42." }
    }
  ]
}
Output Mapping:
{
  "output": "$.response",
  "traces": "$.traces"
}
Galtea extracts the traces array and creates Trace entities linked to the inference result. Each object in the array must contain at least a name field and can include any Trace properties:
PropertyTypeRequiredDescription
namestringYesName of the traced operation
typestringNoOne of: SPAN, GENERATION, EVENT, AGENT, TOOL, CHAIN, RETRIEVER, EVALUATOR, EMBEDDING, GUARDRAIL
descriptionstringNoHuman-readable description of the operation
inputDataobjectNoInput parameters passed to the operation
outputDataobjectNoResult returned by the operation
errorstringNoError message if the operation failed
latencyMsnumberNoExecution time in milliseconds
metadataobjectNoAdditional custom metadata
startTimestringNoISO 8601 timestamp when the operation started
endTimestringNoISO 8601 timestamp when the operation completed
parentTraceIdstringNoID of the parent trace for hierarchical relationships
This approach requires no SDK code in your endpoint handler — it works purely through configuration.

Option 2: Use set_context in Your Endpoint Handler

When running evaluations via Direct Inference, you can collect traces from your endpoint handler by passing the {{ inference_result_id }} placeholder in your input template. This lets your endpoint know which inference result the call belongs to, so it can link traces back to Galtea.

1. Add {{ inference_result_id }} to your Input Template

Include the placeholder in your endpoint connection’s input template so your handler receives the ID:
{
  "model": "gpt-4",
  "messages": [
    {"role": "user", "content": "{{ input }}"}
  ],
  "metadata": {
    "inference_result_id": "{{ inference_result_id }}"
  }
}

2. Use set_context in Your Endpoint Handler

In your API endpoint, extract the inference_result_id from the request and use the SDK’s set_context / clear_context to associate traces with it:
@trace(type=TraceType.AGENT)
def run_agent(query: str) -> str:
    # Your agent logic here — all nested @trace calls
    # will be linked to the inference result automatically
    return "Agent response to: " + query


def my_endpoint_handler(request):
    """Your API endpoint that Galtea calls during Direct Inference."""
    body = request.json()
    user_input = body["messages"][-1]["content"]
    inference_result_id = body["metadata"]["inference_result_id"]

    # Set trace context so all @trace calls are linked to this inference result
    token = set_context(inference_result_id=inference_result_id)
    try:
        response = run_agent(user_input)
    finally:
        # Flush traces to Galtea and clear context
        clear_context(token)

    return {"choices": [{"message": {"content": response}}]}
All @trace-decorated functions called while the context is active will be automatically linked to the inference result in Galtea.
For a complete guide on tracing setup, decorators, and context managers, see the Tracing Agent Operations tutorial.

Learn More

Endpoint Connection

Full reference for configuring endpoint connections

Version

Learn about versions and how endpoint connections integrate with them

Evaluations

Understand how evaluations work

Metrics

Browse available metrics for evaluating your AI

Tracing Agent Operations

Capture and analyze your agent’s internal operations