TaskFlow
DashboardFreewriteWhiteboardsProjectsCRMTasksNotificationsSettingsAgent TowerAPI Docs
OpenClaw Docs
?

User

Member

Caricamento in corso...

Home
Progetti
Task
Notifiche
CRM

    OpenClaw

    Documentation Mirror

    Documentation Overview

    Docs

    Auth credential semantics
    Scheduled tasks
    Hooks
    Automation & tasks
    Standing orders
    Task flow
    Background tasks
    BlueBubbles
    Broadcast groups
    Channel routing
    Discord
    Feishu
    Google Chat
    Group messages
    Groups
    iMessage
    Chat channels
    IRC
    LINE
    Channel location parsing
    Matrix
    Matrix migration
    Matrix push rules for quiet previews
    Mattermost
    Microsoft Teams
    Nextcloud Talk
    Nostr
    Pairing
    QA channel
    QQ bot
    Signal
    Slack
    Synology Chat
    Telegram
    Tlon
    Channel troubleshooting
    Twitch
    WeChat
    WhatsApp
    Yuanbao
    Zalo
    Zalo personal
    CI pipeline
    ACP
    Agent
    Agents
    Approvals
    Backup
    Browser
    Channels
    Clawbot
    `openclaw commitments`
    Completion
    Config
    Configure
    Cron
    Daemon
    Dashboard
    Devices
    Directory
    DNS
    Docs
    Doctor
    Flows (redirect)
    Gateway
    Health
    Hooks
    CLI reference
    Inference CLI
    Logs
    MCP
    Memory
    Message
    Migrate
    Models
    Node
    Nodes
    Onboard
    Pairing
    Plugins
    Proxy
    QR
    Reset
    Sandbox CLI
    Secrets
    Security
    Sessions
    Setup
    Skills
    Status
    System
    `openclaw tasks`
    TUI
    Uninstall
    Update
    Voicecall
    Webhooks
    Wiki
    Active memory
    Agent runtime
    Agent loop
    Agent runtimes
    Agent workspace
    Gateway architecture
    Channel docking
    Inferred commitments
    Compaction
    Context
    Context engine
    Delegate architecture
    Dreaming
    Experimental features
    Features
    Markdown formatting
    Memory overview
    Builtin memory engine
    Honcho memory
    QMD memory engine
    Memory search
    Messages
    Model failover
    Model providers
    Models CLI
    Multi-agent routing
    OAuth
    OpenClaw App SDK
    Presence
    QA overview
    Matrix QA
    Command queue
    Steering queue
    Retry policy
    Session management
    Session pruning
    Session tools
    SOUL.md personality guide
    Streaming and chunking
    System prompt
    Timezones
    TypeBox
    Typing indicators
    Usage tracking
    Date and time
    Node + tsx crash
    Diagnostics flags
    Authentication
    Background exec and process tool
    Bonjour discovery
    Bridge protocol
    CLI backends
    Configuration — agents
    Configuration — channels
    Configuration — tools and custom providers
    Configuration
    Configuration examples
    Configuration reference
    Diagnostics export
    Discovery and transports
    Doctor
    Gateway lock
    Health checks
    Heartbeat
    Gateway runbook
    Local models
    Gateway logging
    Multiple gateways
    Network model
    OpenAI chat completions
    OpenResponses API
    OpenShell
    OpenTelemetry export
    Gateway-owned pairing
    Prometheus metrics
    Gateway protocol
    Remote access
    Remote gateway setup
    Sandbox vs tool policy vs elevated
    Sandboxing
    Secrets management
    Secrets apply plan contract
    Security audit checks
    Security
    Tailscale
    Tools invoke API
    Troubleshooting
    Trusted proxy auth
    Debugging
    Environment variables
    FAQ
    FAQ: first-run setup
    FAQ: models and auth
    GPT-5.5 / Codex agentic parity
    GPT-5.5 / Codex parity maintainer notes
    Help
    Scripts
    Testing
    Testing: live suites
    General troubleshooting
    OpenClaw
    Ansible
    Azure
    Bun (experimental)
    ClawDock
    Release channels
    DigitalOcean
    Docker
    Docker VM runtime
    exe.dev
    Fly.io
    GCP
    Hetzner
    Hostinger
    Install
    Installer internals
    Kubernetes
    macOS VMs
    Migration guide
    Migrating from Claude
    Migrating from Hermes
    Nix
    Node.js
    Northflank
    Oracle Cloud
    Podman
    Railway
    Raspberry Pi
    Render
    Uninstall
    Updating
    Logging
    Network
    Audio and voice notes
    Camera capture
    Image and media support
    Nodes
    Location command
    Media understanding
    Talk mode
    Node troubleshooting
    Voice wake
    Pi integration architecture
    Pi development workflow
    Android app
    Platforms
    iOS app
    Linux app
    Gateway on macOS
    Canvas
    Gateway lifecycle
    macOS dev setup
    Health checks (macOS)
    Menu bar icon
    macOS logging
    Menu bar
    Peekaboo bridge
    macOS permissions
    Remote control
    macOS signing
    Skills (macOS)
    Voice overlay
    Voice wake (macOS)
    WebChat (macOS)
    macOS IPC
    macOS app
    Windows
    Plugin internals
    Plugin architecture internals
    Building plugins
    Plugin bundles
    Codex Computer Use
    Codex harness
    Community plugins
    Plugin compatibility
    Google Meet plugin
    Plugin hooks
    Plugin manifest
    Memory LanceDB
    Memory wiki
    Message presentation
    Agent harness plugins
    Building channel plugins
    Channel turn kernel
    Plugin entry points
    Plugin SDK migration
    Plugin SDK overview
    Building provider plugins
    Plugin runtime helpers
    Plugin setup and config
    Plugin SDK subpaths
    Plugin testing
    Skill workshop plugin
    Voice call plugin
    Webhooks plugin
    Zalo personal plugin
    OpenProse
    Alibaba Model Studio
    Anthropic
    Arcee AI
    Azure Speech
    Amazon Bedrock
    Amazon Bedrock Mantle
    Chutes
    Claude Max API proxy
    Cloudflare AI gateway
    ComfyUI
    Deepgram
    Deepinfra
    DeepSeek
    ElevenLabs
    Fal
    Fireworks
    GitHub Copilot
    GLM (Zhipu)
    Google (Gemini)
    Gradium
    Groq
    Hugging Face (inference)
    Provider directory
    Inferrs
    Inworld
    Kilocode
    LiteLLM
    LM Studio
    MiniMax
    Mistral
    Model provider quickstart
    Moonshot AI
    NVIDIA
    Ollama
    OpenAI
    OpenCode
    OpenCode Go
    OpenRouter
    Perplexity
    Qianfan
    Qwen
    Runway
    SGLang
    StepFun
    Synthetic
    Tencent Cloud (TokenHub)
    Together AI
    Venice AI
    Vercel AI gateway
    vLLM
    Volcengine (Doubao)
    Vydra
    xAI
    Xiaomi MiMo
    Z.AI
    Default AGENTS.md
    Release policy
    API usage and costs
    Credits
    Device model database
    Full release validation
    Memory configuration reference
    OpenClaw App SDK API design
    Prompt caching
    Rich output protocol
    RPC adapters
    SecretRef credential surface
    Session management deep dive
    AGENTS.md template
    BOOT.md template
    BOOTSTRAP.md template
    HEARTBEAT.md template
    IDENTITY template
    SOUL.md template
    TOOLS.md template
    USER template
    Tests
    Token use and costs
    Transcript hygiene
    Onboarding reference
    Contributing to the threat model
    Threat model (MITRE ATLAS)
    Formal verification (security models)
    Network proxy
    Agent bootstrapping
    Docs directory
    Getting started
    Docs hubs
    OpenClaw lore
    Onboarding (macOS app)
    Onboarding overview
    Personal assistant setup
    Setup
    Showcase
    Onboarding (CLI)
    CLI automation
    CLI setup reference
    ACP agents
    ACP agents — setup
    Agent send
    apply_patch tool
    Brave search
    Browser (OpenClaw-managed)
    Browser control API
    Browser troubleshooting
    Browser login
    WSL2 + Windows + remote Chrome CDP troubleshooting
    BTW side questions
    ClawHub
    Code execution
    Creating skills
    Diffs
    DuckDuckGo search
    Elevated mode
    Exa search
    Exec tool
    Exec approvals
    Exec approvals — advanced
    Firecrawl
    Gemini search
    Grok search
    Image generation
    Tools and plugins
    Kimi search
    LLM task
    Lobster
    Tool-loop detection
    Media overview
    MiniMax search
    Multi-agent sandbox and tools
    Music generation
    Ollama web search
    PDF tool
    Perplexity search
    Plugins
    Reactions
    SearXNG search
    Skills
    Skills config
    Slash commands
    Sub-agents
    Tavily
    Thinking levels
    Tokenjuice
    Trajectory bundles
    Text-to-speech
    Video generation
    Web search
    Web fetch
    Linux server
    Control UI
    Dashboard
    Web
    TUI
    WebChat

    OpenAPI Specs

    openapi
    TaskFlow
    docs/openclaw
    Original Docs

    Real-time Synchronized Documentation

    Last sync: 01/05/2026 07:05:21

    Note: This content is mirrored from docs.openclaw.ai and is subject to their terms and conditions.

    OpenClaw Docs

    v2.4.0 Production

    Last synced: Today, 22:00

    Technical reference for the OpenClaw framework. Real-time synchronization with the official documentation engine.

    Use this file to discover all available pages before exploring further.

    QA overview

    The private QA stack is meant to exercise OpenClaw in a more realistic, channel-shaped way than a single unit test can.

    Current pieces:

    • text
      extensions/qa-channel
      : synthetic message channel with DM, channel, thread, reaction, edit, and delete surfaces.
    • text
      extensions/qa-lab
      : debugger UI and QA bus for observing the transcript, injecting inbound messages, and exporting a Markdown report.
    • text
      extensions/qa-matrix
      , future runner plugins: live-transport adapters that drive a real channel inside a child QA gateway.
    • text
      qa/
      : repo-backed seed assets for the kickoff task and baseline QA scenarios.

    Command surface

    Every QA flow runs under

    text
    pnpm openclaw qa <subcommand>
    . Many have
    text
    pnpm qa:*
    script aliases; both forms are supported.

    CommandPurpose
    text
    qa run
    Bundled QA self-check; writes a Markdown report.
    text
    qa suite
    Run repo-backed scenarios against the QA gateway lane. Aliases:
    text
    pnpm openclaw qa suite --runner multipass
    for a disposable Linux VM.
    text
    qa coverage
    Print the markdown scenario-coverage inventory (
    text
    --json
    for machine output).
    text
    qa parity-report
    Compare two
    text
    qa-suite-summary.json
    files and write the agentic parity-gate report.
    text
    qa character-eval
    Run the character QA scenario across multiple live models with a judged report. See Reporting.
    text
    qa manual
    Run a one-off prompt against the selected provider/model lane.
    text
    qa ui
    Start the QA debugger UI and local QA bus (alias:
    text
    pnpm qa:lab:ui
    ).
    text
    qa docker-build-image
    Build the prebaked QA Docker image.
    text
    qa docker-scaffold
    Write a docker-compose scaffold for the QA dashboard + gateway lane.
    text
    qa up
    Build the QA site, start the Docker-backed stack, print the URL (alias:
    text
    pnpm qa:lab:up
    ;
    text
    :fast
    variant adds
    text
    --use-prebuilt-image --bind-ui-dist --skip-ui-build
    ).
    text
    qa aimock
    Start only the AIMock provider server.
    text
    qa mock-openai
    Start only the scenario-aware
    text
    mock-openai
    provider server.
    text
    qa credentials doctor
    /
    text
    add
    /
    text
    list
    /
    text
    remove
    Manage the shared Convex credential pool.
    text
    qa matrix
    Live transport lane against a disposable Tuwunel homeserver. See Matrix QA.
    text
    qa telegram
    Live transport lane against a real private Telegram group.
    text
    qa discord
    Live transport lane against a real private Discord guild channel.

    Operator flow

    The current QA operator flow is a two-pane QA site:

    • Left: Gateway dashboard (Control UI) with the agent.
    • Right: QA Lab, showing the Slack-ish transcript and scenario plan.

    Run it with:

    bash
    pnpm qa:lab:up

    That builds the QA site, starts the Docker-backed gateway lane, and exposes the QA Lab page where an operator or automation loop can give the agent a QA mission, observe real channel behavior, and record what worked, failed, or stayed blocked.

    For faster QA Lab UI iteration without rebuilding the Docker image each time, start the stack with a bind-mounted QA Lab bundle:

    bash
    pnpm openclaw qa docker-build-image pnpm qa:lab:build pnpm qa:lab:up:fast pnpm qa:lab:watch

    text
    qa:lab:up:fast
    keeps the Docker services on a prebuilt image and bind-mounts
    text
    extensions/qa-lab/web/dist
    into the
    text
    qa-lab
    container.
    text
    qa:lab:watch
    rebuilds that bundle on change, and the browser auto-reloads when the QA Lab asset hash changes.

    For a local OpenTelemetry trace smoke, run:

    bash
    pnpm qa:otel:smoke

    That script starts a local OTLP/HTTP trace receiver, runs the

    text
    otel-trace-smoke
    QA scenario with the
    text
    diagnostics-otel
    plugin enabled, then decodes the exported protobuf spans and asserts the release-critical shape:
    text
    openclaw.run
    ,
    text
    openclaw.harness.run
    ,
    text
    openclaw.model.call
    ,
    text
    openclaw.context.assembled
    , and
    text
    openclaw.message.delivery
    must be present; model calls must not export
    text
    StreamAbandoned
    on successful turns; raw diagnostic IDs and
    text
    openclaw.content.*
    attributes must stay out of the trace. It writes
    text
    otel-smoke-summary.json
    next to the QA suite artifacts.

    Observability QA stays source-checkout only. The npm tarball intentionally omits QA Lab, so package Docker release lanes do not run

    text
    qa
    commands. Use
    text
    pnpm qa:otel:smoke
    from a built source checkout when changing diagnostics instrumentation.

    For a transport-real Matrix smoke lane, run:

    bash
    pnpm openclaw qa matrix --profile fast --fail-fast

    The full CLI reference, profile/scenario catalog, env vars, and artifact layout for this lane live in Matrix QA. At a glance: it provisions a disposable Tuwunel homeserver in Docker, registers temporary driver/SUT/observer users, runs the real Matrix plugin inside a child QA gateway scoped to that transport (no

    text
    qa-channel
    ), then writes a Markdown report, JSON summary, observed-events artifact, and combined output log under
    text
    .artifacts/qa-e2e/matrix-<timestamp>/
    .

    For transport-real Telegram and Discord smoke lanes:

    bash
    pnpm openclaw qa telegram pnpm openclaw qa discord

    Both target a pre-existing real channel with two bots (driver + SUT). Required env vars, scenario lists, output artifacts, and the Convex credential pool are documented in Telegram and Discord QA reference below.

    Before using pooled live credentials, run:

    bash
    pnpm openclaw qa credentials doctor

    The doctor checks Convex broker env, validates endpoint settings, and verifies admin/list reachability when the maintainer secret is present. It reports only set/missing status for secrets.

    Live transport coverage

    Live transport lanes share one contract instead of each inventing their own scenario list shape.

    text
    qa-channel
    is the broad synthetic product-behavior suite and is not part of the live transport coverage matrix.

    LaneCanaryMention gatingBot-to-botAllowlist blockTop-level replyRestart resumeThread follow-upThread isolationReaction observationHelp commandNative command registration
    Matrixxxxxxxxxx
    Telegramxxxx
    Discordxxxx

    This keeps

    text
    qa-channel
    as the broad product-behavior suite while Matrix, Telegram, and future live transports share one explicit transport-contract checklist.

    For a disposable Linux VM lane without bringing Docker into the QA path, run:

    bash
    pnpm openclaw qa suite --runner multipass --scenario channel-chat-baseline

    This boots a fresh Multipass guest, installs dependencies, builds OpenClaw inside the guest, runs

    text
    qa suite
    , then copies the normal QA report and summary back into
    text
    .artifacts/qa-e2e/...
    on the host. It reuses the same scenario-selection behavior as
    text
    qa suite
    on the host. Host and Multipass suite runs execute multiple selected scenarios in parallel with isolated gateway workers by default.
    text
    qa-channel
    defaults to concurrency 4, capped by the selected scenario count. Use
    text
    --concurrency <count>
    to tune the worker count, or
    text
    --concurrency 1
    for serial execution. The command exits non-zero when any scenario fails. Use
    text
    --allow-failures
    when you want artifacts without a failing exit code. Live runs forward the supported QA auth inputs that are practical for the guest: env-based provider keys, the QA live provider config path, and
    text
    CODEX_HOME
    when present. Keep
    text
    --output-dir
    under the repo root so the guest can write back through the mounted workspace.

    Telegram and Discord QA reference

    Matrix has a dedicated page because of its scenario count and Docker-backed homeserver provisioning. Telegram and Discord are smaller — a handful of scenarios each, no profile system, against pre-existing real channels — so their reference lives here.

    Shared CLI flags

    Both lanes register through

    text
    extensions/qa-lab/src/live-transports/shared/live-transport-cli.ts
    and accept the same flags:

    FlagDefaultDescription
    text
    --scenario <id>
    —Run only this scenario. Repeatable.
    text
    --output-dir <path>
    text
    <repo>/.artifacts/qa-e2e/{telegram,discord}-<timestamp>
    Where reports/summary/observed messages and the output log are written. Relative paths resolve against
    text
    --repo-root
    .
    text
    --repo-root <path>
    text
    process.cwd()
    Repository root when invoking from a neutral cwd.
    text
    --sut-account <id>
    text
    sut
    Temporary account id inside the QA gateway config.
    text
    --provider-mode <mode>
    text
    live-frontier
    text
    mock-openai
    or
    text
    live-frontier
    (legacy
    text
    live-openai
    still works).
    text
    --model <ref>
    /
    text
    --alt-model <ref>
    provider defaultPrimary/alternate model refs.
    text
    --fast
    offProvider fast mode where supported.
    text
    --credential-source <env|convex>
    text
    env
    See Convex credential pool.
    text
    --credential-role <maintainer|ci>
    text
    ci
    in CI,
    text
    maintainer
    otherwise
    Role used when
    text
    --credential-source convex
    .

    Both exit non-zero on any failed scenario.

    text
    --allow-failures
    writes artifacts without setting a failing exit code.

    Telegram QA

    bash
    pnpm openclaw qa telegram

    Targets one real private Telegram group with two distinct bots (driver + SUT). The SUT bot must have a Telegram username; bot-to-bot observation works best when both bots have Bot-to-Bot Communication Mode enabled in

    text
    @BotFather
    .

    Required env when

    text
    --credential-source env
    :

    • text
      OPENCLAW_QA_TELEGRAM_GROUP_ID
      — numeric chat id (string).
    • text
      OPENCLAW_QA_TELEGRAM_DRIVER_BOT_TOKEN
    • text
      OPENCLAW_QA_TELEGRAM_SUT_BOT_TOKEN

    Optional:

    • text
      OPENCLAW_QA_TELEGRAM_CAPTURE_CONTENT=1
      keeps message bodies in observed-message artifacts (default redacts).

    Scenarios (

    text
    extensions/qa-lab/src/live-transports/telegram/telegram-live.runtime.ts:44
    ):

    • text
      telegram-canary
    • text
      telegram-mention-gating
    • text
      telegram-mentioned-message-reply
    • text
      telegram-help-command
    • text
      telegram-commands-command
    • text
      telegram-tools-compact-command
    • text
      telegram-whoami-command
    • text
      telegram-context-command

    Output artifacts:

    • text
      telegram-qa-report.md
    • text
      telegram-qa-summary.json
      — includes per-reply RTT (driver send → observed SUT reply) starting with the canary.
    • text
      telegram-qa-observed-messages.json
      — bodies redacted unless
      text
      OPENCLAW_QA_TELEGRAM_CAPTURE_CONTENT=1
      .

    Discord QA

    bash
    pnpm openclaw qa discord

    Targets one real private Discord guild channel with two bots: a driver bot controlled by the harness and a SUT bot started by the child OpenClaw gateway through the bundled Discord plugin. Verifies channel mention handling and that the SUT bot has registered the native

    text
    /help
    command with Discord.

    Required env when

    text
    --credential-source env
    :

    • text
      OPENCLAW_QA_DISCORD_GUILD_ID
    • text
      OPENCLAW_QA_DISCORD_CHANNEL_ID
    • text
      OPENCLAW_QA_DISCORD_DRIVER_BOT_TOKEN
    • text
      OPENCLAW_QA_DISCORD_SUT_BOT_TOKEN
    • text
      OPENCLAW_QA_DISCORD_SUT_APPLICATION_ID
      — must match the SUT bot user id returned by Discord (the lane fails fast otherwise).

    Optional:

    • text
      OPENCLAW_QA_DISCORD_CAPTURE_CONTENT=1
      keeps message bodies in observed-message artifacts.

    Scenarios (

    text
    extensions/qa-lab/src/live-transports/discord/discord-live.runtime.ts:36
    ):

    • text
      discord-canary
    • text
      discord-mention-gating
    • text
      discord-native-help-command-registration

    Output artifacts:

    • text
      discord-qa-report.md
    • text
      discord-qa-summary.json
    • text
      discord-qa-observed-messages.json
      — bodies redacted unless
      text
      OPENCLAW_QA_DISCORD_CAPTURE_CONTENT=1
      .

    Convex credential pool

    Both Telegram and Discord lanes can lease credentials from a shared Convex pool instead of reading the env vars above. Pass

    text
    --credential-source convex
    (or set
    text
    OPENCLAW_QA_CREDENTIAL_SOURCE=convex
    ); QA Lab acquires an exclusive lease, heartbeats it for the duration of the run, and releases it on shutdown. Pool kinds are
    text
    "telegram"
    and
    text
    "discord"
    .

    Payload shapes the broker validates on

    text
    admin/add
    :

    • Telegram (
      text
      kind: "telegram"
      ):
      text
      { groupId: string, driverToken: string, sutToken: string }
      —
      text
      groupId
      must be a numeric chat-id string.
    • Discord (
      text
      kind: "discord"
      ):
      text
      { guildId: string, channelId: string, driverBotToken: string, sutBotToken: string, sutApplicationId: string }
      .

    Operational env vars and the Convex broker endpoint contract live in Testing → Shared Telegram credentials via Convex (the section name predates Discord support; the broker semantics are identical for both kinds).

    Repo-backed seeds

    Seed assets live in

    text
    qa/
    :

    • text
      qa/scenarios/index.md
    • text
      qa/scenarios/<theme>/*.md

    These are intentionally in git so the QA plan is visible to both humans and the agent.

    text
    qa-lab
    should stay a generic markdown runner. Each scenario markdown file is the source of truth for one test run and should define:

    • scenario metadata
    • optional category, capability, lane, and risk metadata
    • docs and code refs
    • optional plugin requirements
    • optional gateway config patch
    • the executable
      text
      qa-flow

    The reusable runtime surface that backs

    text
    qa-flow
    is allowed to stay generic and cross-cutting. For example, markdown scenarios can combine transport-side helpers with browser-side helpers that drive the embedded Control UI through the Gateway
    text
    browser.request
    seam without adding a special-case runner.

    Scenario files should be grouped by product capability rather than source tree folder. Keep scenario IDs stable when files move; use

    text
    docsRefs
    and
    text
    codeRefs
    for implementation traceability.

    The baseline list should stay broad enough to cover:

    • DM and channel chat
    • thread behavior
    • message action lifecycle
    • cron callbacks
    • memory recall
    • model switching
    • subagent handoff
    • repo-reading and docs-reading
    • one small build task such as Lobster Invaders

    Provider mock lanes

    text
    qa suite
    has two local provider mock lanes:

    • text
      mock-openai
      is the scenario-aware OpenClaw mock. It remains the default deterministic mock lane for repo-backed QA and parity gates.
    • text
      aimock
      starts an AIMock-backed provider server for experimental protocol, fixture, record/replay, and chaos coverage. It is additive and does not replace the
      text
      mock-openai
      scenario dispatcher.

    Provider-lane implementation lives under

    text
    extensions/qa-lab/src/providers/
    . Each provider owns its defaults, local server startup, gateway model config, auth-profile staging needs, and live/mock capability flags. Shared suite and gateway code should route through the provider registry instead of branching on provider names.

    Transport adapters

    text
    qa-lab
    owns a generic transport seam for markdown QA scenarios.
    text
    qa-channel
    is the first adapter on that seam, but the design target is wider: future real or synthetic channels should plug into the same suite runner instead of adding a transport-specific QA runner.

    At the architecture level, the split is:

    • text
      qa-lab
      owns generic scenario execution, worker concurrency, artifact writing, and reporting.
    • The transport adapter owns gateway config, readiness, inbound and outbound observation, transport actions, and normalized transport state.
    • Markdown scenario files under
      text
      qa/scenarios/
      define the test run;
      text
      qa-lab
      provides the reusable runtime surface that executes them.

    Adding a channel

    Adding a channel to the markdown QA system requires exactly two things:

    1. A transport adapter for the channel.
    2. A scenario pack that exercises the channel contract.

    Do not add a new top-level QA command root when the shared

    text
    qa-lab
    host can own the flow.

    text
    qa-lab
    owns the shared host mechanics:

    • the
      text
      openclaw qa
      command root
    • suite startup and teardown
    • worker concurrency
    • artifact writing
    • report generation
    • scenario execution
    • compatibility aliases for older
      text
      qa-channel
      scenarios

    Runner plugins own the transport contract:

    • how
      text
      openclaw qa <runner>
      is mounted beneath the shared
      text
      qa
      root
    • how the gateway is configured for that transport
    • how readiness is checked
    • how inbound events are injected
    • how outbound messages are observed
    • how transcripts and normalized transport state are exposed
    • how transport-backed actions are executed
    • how transport-specific reset or cleanup is handled

    The minimum adoption bar for a new channel:

    1. Keep
      text
      qa-lab
      as the owner of the shared
      text
      qa
      root.
    2. Implement the transport runner on the shared
      text
      qa-lab
      host seam.
    3. Keep transport-specific mechanics inside the runner plugin or channel harness.
    4. Mount the runner as
      text
      openclaw qa <runner>
      instead of registering a competing root command. Runner plugins should declare
      text
      qaRunners
      in
      text
      openclaw.plugin.json
      and export a matching
      text
      qaRunnerCliRegistrations
      array from
      text
      runtime-api.ts
      . Keep
      text
      runtime-api.ts
      light; lazy CLI and runner execution should stay behind separate entrypoints.
    5. Author or adapt markdown scenarios under the themed
      text
      qa/scenarios/
      directories.
    6. Use the generic scenario helpers for new scenarios.
    7. Keep existing compatibility aliases working unless the repo is doing an intentional migration.

    The decision rule is strict:

    • If behavior can be expressed once in
      text
      qa-lab
      , put it in
      text
      qa-lab
      .
    • If behavior depends on one channel transport, keep it in that runner plugin or plugin harness.
    • If a scenario needs a new capability that more than one channel can use, add a generic helper instead of a channel-specific branch in
      text
      suite.ts
      .
    • If a behavior is only meaningful for one transport, keep the scenario transport-specific and make that explicit in the scenario contract.

    Scenario helper names

    Preferred generic helpers for new scenarios:

    • text
      waitForTransportReady
    • text
      waitForChannelReady
    • text
      injectInboundMessage
    • text
      injectOutboundMessage
    • text
      waitForTransportOutboundMessage
    • text
      waitForChannelOutboundMessage
    • text
      waitForNoTransportOutbound
    • text
      getTransportSnapshot
    • text
      readTransportMessage
    • text
      readTransportTranscript
    • text
      formatTransportTranscript
    • text
      resetTransport

    Compatibility aliases remain available for existing scenarios —

    text
    waitForQaChannelReady
    ,
    text
    waitForOutboundMessage
    ,
    text
    waitForNoOutbound
    ,
    text
    formatConversationTranscript
    ,
    text
    resetBus
    — but new scenario authoring should use the generic names. The aliases exist to avoid a flag-day migration, not as the model going forward.

    Reporting

    text
    qa-lab
    exports a Markdown protocol report from the observed bus timeline. The report should answer:

    • What worked
    • What failed
    • What stayed blocked
    • What follow-up scenarios are worth adding

    For the inventory of available scenarios — useful when sizing follow-up work or wiring a new transport — run

    text
    pnpm openclaw qa coverage
    (add
    text
    --json
    for machine-readable output).

    For character and style checks, run the same scenario across multiple live model refs and write a judged Markdown report:

    bash
    pnpm openclaw qa character-eval \ --model openai/gpt-5.5,thinking=medium,fast \ --model openai/gpt-5.2,thinking=xhigh \ --model openai/gpt-5,thinking=xhigh \ --model anthropic/claude-opus-4-6,thinking=high \ --model anthropic/claude-sonnet-4-6,thinking=high \ --model zai/glm-5.1,thinking=high \ --model moonshot/kimi-k2.5,thinking=high \ --model google/gemini-3.1-pro-preview,thinking=high \ --judge-model openai/gpt-5.5,thinking=xhigh,fast \ --judge-model anthropic/claude-opus-4-6,thinking=high \ --blind-judge-models \ --concurrency 16 \ --judge-concurrency 16

    The command runs local QA gateway child processes, not Docker. Character eval scenarios should set the persona through

    text
    SOUL.md
    , then run ordinary user turns such as chat, workspace help, and small file tasks. The candidate model should not be told that it is being evaluated. The command preserves each full transcript, records basic run stats, then asks the judge models in fast mode with
    text
    xhigh
    reasoning where supported to rank the runs by naturalness, vibe, and humor. Use
    text
    --blind-judge-models
    when comparing providers: the judge prompt still gets every transcript and run status, but candidate refs are replaced with neutral labels such as
    text
    candidate-01
    ; the report maps rankings back to real refs after parsing. Candidate runs default to
    text
    high
    thinking, with
    text
    medium
    for GPT-5.5 and
    text
    xhigh
    for older OpenAI eval refs that support it. Override a specific candidate inline with
    text
    --model provider/model,thinking=<level>
    .
    text
    --thinking <level>
    still sets a global fallback, and the older
    text
    --model-thinking <provider/model=level>
    form is kept for compatibility. OpenAI candidate refs default to fast mode so priority processing is used where the provider supports it. Add
    text
    ,fast
    ,
    text
    ,no-fast
    , or
    text
    ,fast=false
    inline when a single candidate or judge needs an override. Pass
    text
    --fast
    only when you want to force fast mode on for every candidate model. Candidate and judge durations are recorded in the report for benchmark analysis, but judge prompts explicitly say not to rank by speed. Candidate and judge model runs both default to concurrency 16. Lower
    text
    --concurrency
    or
    text
    --judge-concurrency
    when provider limits or local gateway pressure make a run too noisy. When no candidate
    text
    --model
    is passed, the character eval defaults to
    text
    openai/gpt-5.5
    ,
    text
    openai/gpt-5.2
    ,
    text
    openai/gpt-5
    ,
    text
    anthropic/claude-opus-4-6
    ,
    text
    anthropic/claude-sonnet-4-6
    ,
    text
    zai/glm-5.1
    ,
    text
    moonshot/kimi-k2.5
    , and
    text
    google/gemini-3.1-pro-preview
    when no
    text
    --model
    is passed. When no
    text
    --judge-model
    is passed, the judges default to
    text
    openai/gpt-5.5,thinking=xhigh,fast
    and
    text
    anthropic/claude-opus-4-6,thinking=high
    .

    Related docs

    • Matrix QA
    • QA Channel
    • Testing
    • Dashboard

    © 2024 TaskFlow Mirror

    Powered by TaskFlow Sync Engine