LogoCua Documentation
Supported Agents

Human-In-The-Loop

Use humans as agents for evaluation, demonstrations, and interactive control

The Agent SDK provides a human tool, with native support for using a human-in-the-loop as a way to evaluate your environment, tools, or to create demonstrations. You can use it by doing grounding_model+human/human or human/human directly.

Getting Started

To start the human agent tool, simply run:

python -m agent.human_tool

The UI will show you pending completions. Select a completion to take control of the agent.

Usage Examples

Direct Human Agent

from agent import ComputerAgent
from agent.computer import computer

agent = ComputerAgent(
    "human/human",
    tools=[computer]
)

async for _ in agent.run("Take a screenshot, analyze the UI, and click on the most prominent button"):
    pass

Composed with Grounding Model

agent = ComputerAgent(
    "huggingface-local/HelloKKMe/GTA1-7B+human/human",
    tools=[computer]
)

async for _ in agent.run("Navigate to the settings page and enable dark mode"):
    pass

Features

The human-in-the-loop interface provides:

  • Interactive UI: Web-based interface for reviewing and responding to agent requests
  • Image Display: Screenshots with click handlers for direct interaction
  • Action Accordions: Support for various computer actions (click, type, keypress, etc.)
  • Tool Calls: Full OpenAI-compatible tool call support
  • Real-time Updates: Smart polling for responsive UI updates

Use Cases

  • Evaluation: Have humans evaluate agent performance and provide ground truth responses
  • Demonstrations: Create training data by having humans demonstrate tasks
  • Interactive Control: Take manual control when automated agents need human guidance
  • Testing: Validate agent, tool, and environment behavior manually

For more details on the human tool implementation, see the Human Tool Documentation.