Skip to content

OpenAdapt

AI-First Process Automation with Large Multimodal Models (LMMs)

OpenAdapt is the open source software **adapt**er between Large Multimodal Models (LMMs) and traditional desktop and web GUIs.

Record GUI demonstrations, train ML models, and evaluate agents - all from a unified CLI.

Join Discord View on GitHub


What is OpenAdapt?

OpenAdapt bridges the gap between powerful AI models and everyday software automation. Instead of writing complex scripts or learning APIs, you simply:

  1. Record - Demonstrate a task by doing it yourself
  2. Train - Let OpenAdapt learn from your demonstration
  3. Deploy - Run your trained agent to automate the task
  4. Evaluate - Measure performance on standardized benchmarks
flowchart LR
    subgraph Record["1. Record"]
        A[User Demo] --> B[Capture]
    end

    subgraph Train["2. Train"]
        B --> C[ML Model]
    end

    subgraph Deploy["3. Deploy"]
        C --> D[Agent Policy]
        D --> E[Action Replay]
    end

    subgraph Evaluate["4. Evaluate"]
        D --> F[Benchmark]
        F --> G[Metrics]
    end

    GROUND[Grounding] -.-> E
    RETRIEVE[Retrieval] -.-> C
    PRIV[Privacy] -.-> B

Key Features

Model Agnostic

Works with any Large Multimodal Model - Claude, GPT-4V, Gemini, Qwen-VL, or your own fine-tuned models.

Learn from Demonstration

No prompting required. OpenAdapt learns directly from how you perform tasks, automatically generating the right prompts.

Universal GUI Support

Works with all desktop GUIs including native applications, web browsers, and virtualized environments.

Open Source

MIT licensed. Full transparency, community-driven development, and no vendor lock-in.


Quick Start

Install OpenAdapt with the features you need:

pip install openadapt[all]  # Everything

Record a demonstration:

openadapt capture start --name my-task
# Perform your task, then press Ctrl+C

Train a model:

openadapt train start --capture my-task --model qwen3vl-2b

Evaluate:

openadapt eval run --checkpoint training_output/model.pt --benchmark waa

See the Installation Guide for detailed setup instructions.


Architecture

OpenAdapt v1.0+ uses a modular meta-package architecture. The main openadapt package provides a unified CLI and depends on focused sub-packages:

Package Description
openadapt-capture Event recording and storage
openadapt-ml ML engine, training, inference
openadapt-evals Benchmark evaluation
openadapt-viewer HTML visualization
openadapt-grounding UI element localization
openadapt-retrieval Multimodal demo retrieval
openadapt-privacy PII/PHI scrubbing

See the full Architecture Documentation for detailed diagrams.


Demos


Community


License

OpenAdapt is released under the MIT License.