LogoCua Documentation

LLM Integrations

LiteLLM Integration

This MCP server features comprehensive liteLLM integration, allowing you to use any supported LLM provider with a simple model string configuration.

  • Unified Configuration: Use a single CUA_MODEL_NAME environment variable with a model string
  • Automatic Provider Detection: The agent automatically detects the provider and capabilities from the model string
  • Extensive Provider Support: Works with Anthropic, OpenAI, local models, and any liteLLM-compatible provider

Model String Examples:

  • Anthropic: "anthropic/claude-3-5-sonnet-20241022"
  • OpenAI: "openai/computer-use-preview"
  • UI-TARS: "huggingface-local/ByteDance-Seed/UI-TARS-1.5-7B"
  • Omni + Any LiteLLM: "omniparser+litellm/gpt-4o", "omniparser+litellm/claude-3-haiku", "omniparser+ollama_chat/gemma3"