Report Abuse

Basic Information

LaVague is an open-source framework for developers who want to build AI Web Agents that automate multi-step tasks on websites. It implements a Large Action Model approach where agents take a high-level objective and the current web page state, use a World Model to produce instructions, and use an Action Engine to compile those instructions into executable automation code such as Selenium or Playwright commands. The project includes a purpose-built LaVague QA tool that converts Gherkin specifications into web tests, a demo and interactive Gradio interface, driver integrations for Selenium, Playwright and a Chrome extension, and examples that show how to instantiate agents and run objectives. The README notes that LLM calls are used by default (OpenAI gpt4-o in examples) and that an OPENAI_API_KEY is required for those demos and examples. Telemetry and data collection are documented with an option to disable it via an environment variable.

Links

Categorization

App Details

Features
LaVague provides built-in contexts and customizable configurations to tailor agent behavior. It includes a test runner for benchmarking and validating agents, a token counter for estimating LLM token usage and costs, and logging and debugging tools to inspect agent execution. An optional interactive Gradio interface and a Chrome extension let users try and visualize agents. Supported drivers include Selenium, Playwright and a Chrome extension driver with varying feature support such as headless operation, iframe handling, multiple tab control and element highlighting. The repository ships examples and a quick-tour notebook showing code that instantiates WorldModel, ActionEngine and WebAgent objects, and a dedicated LaVague QA extension focused on automating test creation from Gherkin specs.
Use Cases
LaVague helps teams automate complex web workflows by converting high-level objectives into actionable sequences and executing them across supported browser drivers. Developers can prototype agents that inspect page state, generate instructions via a World Model, and execute browser actions through an Action Engine, reducing manual UI interaction and accelerating tasks like documentation walkthroughs, testing flows or repetitive web operations. QA engineers benefit from LaVague QA which speeds test generation from human-readable Gherkin scenarios. Built-in token accounting and logging help manage LLM costs and debug runs. The framework also collects anonymized telemetry to build community datasets, with a documented option to turn off telemetry if required for privacy or compliance.

Please fill the required fields*