SEAL Framework

Evaluating Augmented Language Models with Process-Based Assessment

LLM Evaluation Augmented Models Tool Use Benchmarking AI Agents

Introduce & Analogize

Let's grasp the core idea with a simple, powerful analogy.

The Master Chef's Kitchen

Imagine an Augmented Language Model (ALM) is a world-class chef. The core Language Model (like GPT-4) is the chef's brilliant mind—full of creativity, knowledge, and culinary intuition. But a chef's mind alone can't create a feast. They need a kitchen full of tools.

🧠

The Chef's Brain

The core LLM, planning and reasoning.

🍳

Kitchen Tools

APIs, search engines, code interpreters.

📜

The Complex Recipe

The user's complex query or task.

Asking a standard LLM a complex, real-world question is like asking the chef to make a soufflé without an oven. They can describe it perfectly but can't produce it. An ALM, our chef, can use tools: a Search Engine is their pantry for fresh facts, a Code Interpreter is a precision sous-vide machine for complex calculations, and other APIs are specialized knives.

The SEAL framework, then, is like a Michelin inspector. It doesn't just taste the final dish (the outcome); it goes into the kitchen to evaluate the chef's process. Did they use the right knife for the fish? Was the oven at the correct temperature? This process-based evaluation is the key to truly understanding how good these AI "chefs" are.

Deconstruct & Rebuild

From first principles, how does an ALM actually work?

From Prompt to Action: The "Reason-Act" Cycle

An ALM follows a "Reason-Act" cycle. The LLM's 'brain' first thinks about the problem, then chooses and uses a tool, observes the result, and repeats until the task is complete. This section explains this fundamental process and how these tools are defined in code.

User Query
LLM Planner (Brain)
Select & Use Tool
Synthesize & Respond

Code Snippet: Defining a Tool

For an ALM to use a tool, a developer must define it, typically as a function with a very clear description (the "docstring"). This docstring is crucial, as it's what the LLM reads to understand what the tool does and when to use it. Below is a Python example for a simple calculator tool, inspired by code from the Continual-Intelligence repository.

def calculator(expression: str) -> str:
    """
    A calculator that can evaluate mathematical expressions.
    Use this tool to compute numerical answers.

    Args:
        expression (str): The mathematical expression to evaluate. 
                          Example: "2 * (3 + 4)"
    Returns:
        str: The result of the calculation.
    """
    try:
        result = eval(expression)
        return str(result)
    except Exception as e:
        return f"Error: {e}"

Synthesize & "Chunk"

Breaking down the SEAL Framework into memorable chunks.

The SEAL framework provides the "Michelin inspector's toolkit" for evaluating our AI chefs. Instead of one giant, complex test, it breaks down the evaluation into two core, manageable concepts: Skills and Scenarios. By testing these fundamental building blocks, we get a much clearer picture of what a model can and cannot do. Explore the cards below to see how they work.

The core idea to "chunk" is: SEAL evaluates ALMs by testing atomic Skills in isolation and then combining them in complex Scenarios to assess both process and outcome.

Atomic Skills: The Basic Techniques

These are the fundamental actions an ALM can take, like individual cooking techniques. SEAL tests these in isolation to see if the model has mastered the basics. Click on a skill to learn more.

Complex Scenarios: The Full Recipe

These are multi-step problems that require combining multiple skills in the correct order to solve, just like a complete recipe. This tests the model's ability to plan and execute a complex strategy.

The Science of SEAL

The data science and scientific methodology behind the evaluation.

From "Recipe" to "Graph"

This is the core technical insight. A "Scenario" isn't just a prompt; it's a "white-box" test case. From a data science perspective, each scenario is structured as a Directed Acyclic Graph (DAG). Each node in the graph is a specific skill (a tool call), and the edges define the dependencies. This graph *is* the "correct recipe."

Why a DAG? Because it allows for a rigorous, scientific evaluation of the model's *process*. The model isn't just given a final exam; it's graded on every step of its homework.

Code Snippet: A Scenario as a DAG

This JSON-like snippet shows how the "AI Stock Analyst" scenario is defined. Notice the `depends_on` field, which creates the graph structure. The model's job is to replicate this logical flow.

{
  "scenario": "AI Stock Analyst",
  "steps": [
    { 
      "id": 1, 
      "skill": "Search", 
      "prompt": "Find recent news for $STOCK",
      "depends_on": [] 
    },
    { 
      "id": 2, 
      "skill": "Search", 
      "prompt": "Find financial data for $STOCK",
      "depends_on": [] 
    },
    { 
      "id": 3, 
      "skill": "Code", 
      "prompt": "Analyze sentiment of news (step 1) and financials (step 2)",
      "depends_on": [1, 2] 
    },
    { 
      "id": 4, 
      "skill": "Logic", 
      "prompt": "Synthesize analysis (step 3) into a recommendation",
      "depends_on": [3] 
    }
  ]
}

Process vs. Outcome Metrics

This graph structure allows for two types of scientific evaluation, which is what makes SEAL so powerful:

1. Outcome-Based Metric

This is the traditional "final answer" test. Does the model's final response match the ground truth? It's a simple binary (Correct/Incorrect) score. It tells us *if* the chef made a good dish.

2. Process-Based Metric

This is the "Michelin inspection." The model's entire sequence of tool calls (its "execution trace") is compared against the ground-truth DAG. This metric measures:

  • Tool Accuracy: Did it use the right tool?
  • Argument Accuracy: Did it use the tool with the right inputs?
  • Order: Did it follow the correct logical dependencies?
This tells us *how* the chef made the dish, which is a much richer and more reliable signal of true capability.

The Model Showdown

An interactive comparison of how different models perform.

Here, you can explore the performance data from the SEAL evaluation. The chart below visualizes how leading models compare across different evaluation criteria. Use the buttons to switch between viewing overall performance, breaking it down by specific skills, or analyzing performance in complex scenarios. This interactive view allows you to see the nuanced strengths and weaknesses of each model, something a static report can't show.

Select a view to see insights.

Analyze Critically

No framework is perfect. Let's discuss the trade-offs.

While SEAL's process-based, DAG-driven approach is a significant scientific advancement, it's essential to understand its limitations and the broader challenges facing ALMs. A critical perspective helps us see the full picture.

Limitations of SEAL

  • Graph Rigidity: The "correct" DAG is predefined. This tests reliability but may penalize models that find a *different* but equally valid path to the solution.
  • Skill Coverage: The defined set of skills isn't exhaustive. Real-world tasks may require skills not yet included in the framework.
  • Evaluation Complexity: Generating and validating these complex ground-truth graphs is labor-intensive and a significant data science challenge in itself.

Broader Challenges for ALMs

  • Tool Selection Errors: Models can choose the wrong tool for the job, or use the right tool with incorrect inputs, leading to faulty results.
  • Error Propagation: An error in an early step (e.g., a bad web search result) can cascade through the entire reasoning process, making the final answer incorrect.
  • "Over-Tooling": Models might sometimes try to use a tool for a simple task their own "brain" could have handled, adding unnecessary complexity and failure points.

Engage & Reflect

Prompt for Active Recall

To help solidify your understanding, try to answer this in your own words (you don't need to write it down, just think it through): From a data science perspective, why is evaluating a "process" (like a DAG) a more robust method than just evaluating the "outcome" (the final answer)?

A Question for Deeper Thought

If a model achieves the correct *outcome* but follows a completely different *process* than the ground-truth DAG, should it be penalized? What does this tell us about the challenge of "creativity" vs. "reliability" in AI evaluation?