TaskFlow
DashboardFreewriteWhiteboardsProjectsCRMTasksNotificationsSettingsAgent TowerAPI Docs
OpenClaw Docs
?

User

Member

Caricamento in corso...

Home
Progetti
Task
Notifiche
CRM

    OpenClaw

    Documentation Mirror

    Documentation Overview

    Docs

    Auth credential semantics
    Scheduled tasks
    Hooks
    Automation & tasks
    Standing orders
    Task flow
    Background tasks
    BlueBubbles
    Broadcast groups
    Channel routing
    Discord
    Feishu
    Google Chat
    Group messages
    Groups
    iMessage
    Chat channels
    IRC
    LINE
    Channel location parsing
    Matrix
    Matrix migration
    Matrix push rules for quiet previews
    Mattermost
    Microsoft Teams
    Nextcloud Talk
    Nostr
    Pairing
    QA channel
    QQ bot
    Signal
    Slack
    Synology Chat
    Telegram
    Tlon
    Channel troubleshooting
    Twitch
    WeChat
    WhatsApp
    Yuanbao
    Zalo
    Zalo personal
    CI pipeline
    ACP
    Agent
    Agents
    Approvals
    Backup
    Browser
    Channels
    Clawbot
    `openclaw commitments`
    Completion
    Config
    Configure
    Cron
    Daemon
    Dashboard
    Devices
    Directory
    DNS
    Docs
    Doctor
    Flows (redirect)
    Gateway
    Health
    Hooks
    CLI reference
    Inference CLI
    Logs
    MCP
    Memory
    Message
    Migrate
    Models
    Node
    Nodes
    Onboard
    Pairing
    Plugins
    Proxy
    QR
    Reset
    Sandbox CLI
    Secrets
    Security
    Sessions
    Setup
    Skills
    Status
    System
    `openclaw tasks`
    TUI
    Uninstall
    Update
    Voicecall
    Webhooks
    Wiki
    Active memory
    Agent runtime
    Agent loop
    Agent runtimes
    Agent workspace
    Gateway architecture
    Channel docking
    Inferred commitments
    Compaction
    Context
    Context engine
    Delegate architecture
    Dreaming
    Experimental features
    Features
    Markdown formatting
    Memory overview
    Builtin memory engine
    Honcho memory
    QMD memory engine
    Memory search
    Messages
    Model failover
    Model providers
    Models CLI
    Multi-agent routing
    OAuth
    OpenClaw App SDK
    Presence
    QA overview
    Matrix QA
    Command queue
    Steering queue
    Retry policy
    Session management
    Session pruning
    Session tools
    SOUL.md personality guide
    Streaming and chunking
    System prompt
    Timezones
    TypeBox
    Typing indicators
    Usage tracking
    Date and time
    Node + tsx crash
    Diagnostics flags
    Authentication
    Background exec and process tool
    Bonjour discovery
    Bridge protocol
    CLI backends
    Configuration — agents
    Configuration — channels
    Configuration — tools and custom providers
    Configuration
    Configuration examples
    Configuration reference
    Diagnostics export
    Discovery and transports
    Doctor
    Gateway lock
    Health checks
    Heartbeat
    Gateway runbook
    Local models
    Gateway logging
    Multiple gateways
    Network model
    OpenAI chat completions
    OpenResponses API
    OpenShell
    OpenTelemetry export
    Gateway-owned pairing
    Prometheus metrics
    Gateway protocol
    Remote access
    Remote gateway setup
    Sandbox vs tool policy vs elevated
    Sandboxing
    Secrets management
    Secrets apply plan contract
    Security audit checks
    Security
    Tailscale
    Tools invoke API
    Troubleshooting
    Trusted proxy auth
    Debugging
    Environment variables
    FAQ
    FAQ: first-run setup
    FAQ: models and auth
    GPT-5.5 / Codex agentic parity
    GPT-5.5 / Codex parity maintainer notes
    Help
    Scripts
    Testing
    Testing: live suites
    General troubleshooting
    OpenClaw
    Ansible
    Azure
    Bun (experimental)
    ClawDock
    Release channels
    DigitalOcean
    Docker
    Docker VM runtime
    exe.dev
    Fly.io
    GCP
    Hetzner
    Hostinger
    Install
    Installer internals
    Kubernetes
    macOS VMs
    Migration guide
    Migrating from Claude
    Migrating from Hermes
    Nix
    Node.js
    Northflank
    Oracle Cloud
    Podman
    Railway
    Raspberry Pi
    Render
    Uninstall
    Updating
    Logging
    Network
    Audio and voice notes
    Camera capture
    Image and media support
    Nodes
    Location command
    Media understanding
    Talk mode
    Node troubleshooting
    Voice wake
    Pi integration architecture
    Pi development workflow
    Android app
    Platforms
    iOS app
    Linux app
    Gateway on macOS
    Canvas
    Gateway lifecycle
    macOS dev setup
    Health checks (macOS)
    Menu bar icon
    macOS logging
    Menu bar
    Peekaboo bridge
    macOS permissions
    Remote control
    macOS signing
    Skills (macOS)
    Voice overlay
    Voice wake (macOS)
    WebChat (macOS)
    macOS IPC
    macOS app
    Windows
    Plugin internals
    Plugin architecture internals
    Building plugins
    Plugin bundles
    Codex Computer Use
    Codex harness
    Community plugins
    Plugin compatibility
    Google Meet plugin
    Plugin hooks
    Plugin manifest
    Memory LanceDB
    Memory wiki
    Message presentation
    Agent harness plugins
    Building channel plugins
    Channel turn kernel
    Plugin entry points
    Plugin SDK migration
    Plugin SDK overview
    Building provider plugins
    Plugin runtime helpers
    Plugin setup and config
    Plugin SDK subpaths
    Plugin testing
    Skill workshop plugin
    Voice call plugin
    Webhooks plugin
    Zalo personal plugin
    OpenProse
    Alibaba Model Studio
    Anthropic
    Arcee AI
    Azure Speech
    Amazon Bedrock
    Amazon Bedrock Mantle
    Chutes
    Claude Max API proxy
    Cloudflare AI gateway
    ComfyUI
    Deepgram
    Deepinfra
    DeepSeek
    ElevenLabs
    Fal
    Fireworks
    GitHub Copilot
    GLM (Zhipu)
    Google (Gemini)
    Gradium
    Groq
    Hugging Face (inference)
    Provider directory
    Inferrs
    Inworld
    Kilocode
    LiteLLM
    LM Studio
    MiniMax
    Mistral
    Model provider quickstart
    Moonshot AI
    NVIDIA
    Ollama
    OpenAI
    OpenCode
    OpenCode Go
    OpenRouter
    Perplexity
    Qianfan
    Qwen
    Runway
    SGLang
    StepFun
    Synthetic
    Tencent Cloud (TokenHub)
    Together AI
    Venice AI
    Vercel AI gateway
    vLLM
    Volcengine (Doubao)
    Vydra
    xAI
    Xiaomi MiMo
    Z.AI
    Default AGENTS.md
    Release policy
    API usage and costs
    Credits
    Device model database
    Full release validation
    Memory configuration reference
    OpenClaw App SDK API design
    Prompt caching
    Rich output protocol
    RPC adapters
    SecretRef credential surface
    Session management deep dive
    AGENTS.md template
    BOOT.md template
    BOOTSTRAP.md template
    HEARTBEAT.md template
    IDENTITY template
    SOUL.md template
    TOOLS.md template
    USER template
    Tests
    Token use and costs
    Transcript hygiene
    Onboarding reference
    Contributing to the threat model
    Threat model (MITRE ATLAS)
    Formal verification (security models)
    Network proxy
    Agent bootstrapping
    Docs directory
    Getting started
    Docs hubs
    OpenClaw lore
    Onboarding (macOS app)
    Onboarding overview
    Personal assistant setup
    Setup
    Showcase
    Onboarding (CLI)
    CLI automation
    CLI setup reference
    ACP agents
    ACP agents — setup
    Agent send
    apply_patch tool
    Brave search
    Browser (OpenClaw-managed)
    Browser control API
    Browser troubleshooting
    Browser login
    WSL2 + Windows + remote Chrome CDP troubleshooting
    BTW side questions
    ClawHub
    Code execution
    Creating skills
    Diffs
    DuckDuckGo search
    Elevated mode
    Exa search
    Exec tool
    Exec approvals
    Exec approvals — advanced
    Firecrawl
    Gemini search
    Grok search
    Image generation
    Tools and plugins
    Kimi search
    LLM task
    Lobster
    Tool-loop detection
    Media overview
    MiniMax search
    Multi-agent sandbox and tools
    Music generation
    Ollama web search
    PDF tool
    Perplexity search
    Plugins
    Reactions
    SearXNG search
    Skills
    Skills config
    Slash commands
    Sub-agents
    Tavily
    Thinking levels
    Tokenjuice
    Trajectory bundles
    Text-to-speech
    Video generation
    Web search
    Web fetch
    Linux server
    Control UI
    Dashboard
    Web
    TUI
    WebChat

    OpenAPI Specs

    openapi
    TaskFlow
    docs/openclaw
    Original Docs

    Real-time Synchronized Documentation

    Last sync: 01/05/2026 07:01:49

    Note: This content is mirrored from docs.openclaw.ai and is subject to their terms and conditions.

    OpenClaw Docs

    v2.4.0 Production

    Last synced: Today, 22:00

    Technical reference for the OpenClaw framework. Real-time synchronization with the official documentation engine.

    Use this file to discover all available pages before exploring further.

    Prompt caching

    Prompt caching means the model provider can reuse unchanged prompt prefixes (usually system/developer instructions and other stable context) across turns instead of re-processing them every time. OpenClaw normalizes provider usage into

    text
    cacheRead
    and
    text
    cacheWrite
    where the upstream API exposes those counters directly.

    Status surfaces can also recover cache counters from the most recent transcript usage log when the live session snapshot is missing them, so

    text
    /status
    can keep showing a cache line after partial session metadata loss. Existing nonzero live cache values still take precedence over transcript fallback values.

    Why this matters: lower token cost, faster responses, and more predictable performance for long-running sessions. Without caching, repeated prompts pay the full prompt cost on every turn even when most input did not change.

    The sections below cover every cache-related knob that affects prompt reuse and token cost.

    Provider references:

    • Anthropic prompt caching: https://platform.claude.com/docs/en/build-with-claude/prompt-caching
    • OpenAI prompt caching: https://developers.openai.com/api/docs/guides/prompt-caching
    • OpenAI API headers and request IDs: https://developers.openai.com/api/reference/overview
    • Anthropic request IDs and errors: https://platform.claude.com/docs/en/api/errors

    Primary knobs

    text
    cacheRetention
    (global default, model, and per-agent)

    Set cache retention as a global default for all models:

    yaml
    agents: defaults: params: cacheRetention: "long" # none | short | long

    Override per-model:

    yaml
    agents: defaults: models: "anthropic/claude-opus-4-6": params: cacheRetention: "short" # none | short | long

    Per-agent override:

    yaml
    agents: list: - id: "alerts" params: cacheRetention: "none"

    Config merge order:

    1. text
      agents.defaults.params
      (global default — applies to all models)
    2. text
      agents.defaults.models["provider/model"].params
      (per-model override)
    3. text
      agents.list[].params
      (matching agent id; overrides by key)

    text
    contextPruning.mode: "cache-ttl"

    Prunes old tool-result context after cache TTL windows so post-idle requests do not re-cache oversized history.

    yaml
    agents: defaults: contextPruning: mode: "cache-ttl" ttl: "1h"

    See Session Pruning for full behavior.

    Heartbeat keep-warm

    Heartbeat can keep cache windows warm and reduce repeated cache writes after idle gaps.

    yaml
    agents: defaults: heartbeat: every: "55m"

    Per-agent heartbeat is supported at

    text
    agents.list[].heartbeat
    .

    Provider behavior

    Anthropic (direct API)

    • text
      cacheRetention
      is supported.
    • With Anthropic API-key auth profiles, OpenClaw seeds
      text
      cacheRetention: "short"
      for Anthropic model refs when unset.
    • Anthropic native Messages responses expose both
      text
      cache_read_input_tokens
      and
      text
      cache_creation_input_tokens
      , so OpenClaw can show both
      text
      cacheRead
      and
      text
      cacheWrite
      .
    • For native Anthropic requests,
      text
      cacheRetention: "short"
      maps to the default 5-minute ephemeral cache, and
      text
      cacheRetention: "long"
      upgrades to the 1-hour TTL only on direct
      text
      api.anthropic.com
      hosts.

    OpenAI (direct API)

    • Prompt caching is automatic on supported recent models. OpenClaw does not need to inject block-level cache markers.
    • OpenClaw uses
      text
      prompt_cache_key
      to keep cache routing stable across turns and uses
      text
      prompt_cache_retention: "24h"
      only when
      text
      cacheRetention: "long"
      is selected on direct OpenAI hosts.
    • OpenAI-compatible Completions providers receive
      text
      prompt_cache_key
      only when their model config explicitly sets
      text
      compat.supportsPromptCacheKey: true
      ;
      text
      cacheRetention: "none"
      still suppresses it.
    • OpenAI responses expose cached prompt tokens via
      text
      usage.prompt_tokens_details.cached_tokens
      (or
      text
      input_tokens_details.cached_tokens
      on Responses API events). OpenClaw maps that to
      text
      cacheRead
      .
    • OpenAI does not expose a separate cache-write token counter, so
      text
      cacheWrite
      stays
      text
      0
      on OpenAI paths even when the provider is warming a cache.
    • OpenAI returns useful tracing and rate-limit headers such as
      text
      x-request-id
      ,
      text
      openai-processing-ms
      , and
      text
      x-ratelimit-*
      , but cache-hit accounting should come from the usage payload, not from headers.
    • In practice, OpenAI often behaves like an initial-prefix cache rather than Anthropic-style moving full-history reuse. Stable long-prefix text turns can land near a
      text
      4864
      cached-token plateau in current live probes, while tool-heavy or MCP-style transcripts often plateau near
      text
      4608
      cached tokens even on exact repeats.

    Anthropic Vertex

    • Anthropic models on Vertex AI (
      text
      anthropic-vertex/*
      ) support
      text
      cacheRetention
      the same way as direct Anthropic.
    • text
      cacheRetention: "long"
      maps to the real 1-hour prompt-cache TTL on Vertex AI endpoints.
    • Default cache retention for
      text
      anthropic-vertex
      matches direct Anthropic defaults.
    • Vertex requests are routed through boundary-aware cache shaping so cache reuse stays aligned with what providers actually receive.

    Amazon Bedrock

    • Anthropic Claude model refs (
      text
      amazon-bedrock/*anthropic.claude*
      ) support explicit
      text
      cacheRetention
      pass-through.
    • Non-Anthropic Bedrock models are forced to
      text
      cacheRetention: "none"
      at runtime.

    OpenRouter models

    For

    text
    openrouter/anthropic/*
    model refs, OpenClaw injects Anthropic
    text
    cache_control
    on system/developer prompt blocks to improve prompt-cache reuse only when the request is still targeting a verified OpenRouter route (
    text
    openrouter
    on its default endpoint, or any provider/base URL that resolves to
    text
    openrouter.ai
    ).

    For

    text
    openrouter/deepseek/*
    ,
    text
    openrouter/moonshot*/*
    , and
    text
    openrouter/zai/*
    model refs,
    text
    contextPruning.mode: "cache-ttl"
    is allowed because OpenRouter handles provider-side prompt caching automatically. OpenClaw does not inject Anthropic
    text
    cache_control
    markers into those requests.

    DeepSeek cache construction is best-effort and can take a few seconds. An immediate follow-up may still show

    text
    cached_tokens: 0
    ; verify with a repeated same-prefix request after a short delay and use
    text
    usage.prompt_tokens_details.cached_tokens
    as the cache-hit signal.

    If you repoint the model at an arbitrary OpenAI-compatible proxy URL, OpenClaw stops injecting those OpenRouter-specific Anthropic cache markers.

    Other providers

    If the provider does not support this cache mode,

    text
    cacheRetention
    has no effect.

    Google Gemini direct API

    • Direct Gemini transport (
      text
      api: "google-generative-ai"
      ) reports cache hits through upstream
      text
      cachedContentTokenCount
      ; OpenClaw maps that to
      text
      cacheRead
      .
    • When
      text
      cacheRetention
      is set on a direct Gemini model, OpenClaw automatically creates, reuses, and refreshes
      text
      cachedContents
      resources for system prompts on Google AI Studio runs. This means you no longer need to pre-create a cached-content handle manually.
    • You can still pass a pre-existing Gemini cached-content handle through as
      text
      params.cachedContent
      (or legacy
      text
      params.cached_content
      ) on the configured model.
    • This is separate from Anthropic/OpenAI prompt-prefix caching. For Gemini, OpenClaw manages a provider-native
      text
      cachedContents
      resource rather than injecting cache markers into the request.

    Gemini CLI JSON usage

    • Gemini CLI JSON output can also surface cache hits through
      text
      stats.cached
      ; OpenClaw maps that to
      text
      cacheRead
      .
    • If the CLI omits a direct
      text
      stats.input
      value, OpenClaw derives input tokens from
      text
      stats.input_tokens - stats.cached
      .
    • This is usage normalization only. It does not mean OpenClaw is creating Anthropic/OpenAI-style prompt-cache markers for Gemini CLI.

    System-prompt cache boundary

    OpenClaw splits the system prompt into a stable prefix and a volatile suffix separated by an internal cache-prefix boundary. Content above the boundary (tool definitions, skills metadata, workspace files, and other relatively static context) is ordered so it stays byte-identical across turns. Content below the boundary (for example

    text
    HEARTBEAT.md
    , runtime timestamps, and other per-turn metadata) is allowed to change without invalidating the cached prefix.

    Key design choices:

    • Stable workspace project-context files are ordered before
      text
      HEARTBEAT.md
      so heartbeat churn does not bust the stable prefix.
    • The boundary is applied across Anthropic-family, OpenAI-family, Google, and CLI transport shaping so all supported providers benefit from the same prefix stability.
    • Codex Responses and Anthropic Vertex requests are routed through boundary-aware cache shaping so cache reuse stays aligned with what providers actually receive.
    • System-prompt fingerprints are normalized (whitespace, line endings, hook-added context, runtime capability ordering) so semantically unchanged prompts share KV/cache across turns.

    If you see unexpected

    text
    cacheWrite
    spikes after a config or workspace change, check whether the change lands above or below the cache boundary. Moving volatile content below the boundary (or stabilizing it) often resolves the issue.

    OpenClaw cache-stability guards

    OpenClaw also keeps several cache-sensitive payload shapes deterministic before the request reaches the provider:

    • Bundle MCP tool catalogs are sorted deterministically before tool registration, so
      text
      listTools()
      order changes do not churn the tools block and bust prompt-cache prefixes.
    • Legacy sessions with persisted image blocks keep the 3 most recent completed turns intact; older already-processed image blocks may be replaced with a marker so image-heavy follow-ups do not keep re-sending large stale payloads.

    Tuning patterns

    Mixed traffic (recommended default)

    Keep a long-lived baseline on your main agent, disable caching on bursty notifier agents:

    yaml
    agents: defaults: model: primary: "anthropic/claude-opus-4-6" models: "anthropic/claude-opus-4-6": params: cacheRetention: "long" list: - id: "research" default: true heartbeat: every: "55m" - id: "alerts" params: cacheRetention: "none"

    Cost-first baseline

    • Set baseline
      text
      cacheRetention: "short"
      .
    • Enable
      text
      contextPruning.mode: "cache-ttl"
      .
    • Keep heartbeat below your TTL only for agents that benefit from warm caches.

    Cache diagnostics

    OpenClaw exposes dedicated cache-trace diagnostics for embedded agent runs.

    For normal user-facing diagnostics,

    text
    /status
    and other usage summaries can use the latest transcript usage entry as a fallback source for
    text
    cacheRead
    /
    text
    cacheWrite
    when the live session entry does not have those counters.

    Live regression tests

    OpenClaw keeps one combined live cache regression gate for repeated prefixes, tool turns, image turns, MCP-style tool transcripts, and an Anthropic no-cache control.

    • text
      src/agents/live-cache-regression.live.test.ts
    • text
      src/agents/live-cache-regression-baseline.ts

    Run the narrow live gate with:

    sh
    OPENCLAW_LIVE_TEST=1 OPENCLAW_LIVE_CACHE_TEST=1 pnpm test:live:cache

    The baseline file stores the most recent observed live numbers plus the provider-specific regression floors used by the test. The runner also uses fresh per-run session IDs and prompt namespaces so previous cache state does not pollute the current regression sample.

    These tests intentionally do not use identical success criteria across providers.

    Anthropic live expectations

    • Expect explicit warmup writes via
      text
      cacheWrite
      .
    • Expect near-full history reuse on repeated turns because Anthropic cache control advances the cache breakpoint through the conversation.
    • Current live assertions still use high hit-rate thresholds for stable, tool, and image paths.

    OpenAI live expectations

    • Expect
      text
      cacheRead
      only.
      text
      cacheWrite
      remains
      text
      0
      .
    • Treat repeated-turn cache reuse as a provider-specific plateau, not as Anthropic-style moving full-history reuse.
    • Current live assertions use conservative floor checks derived from observed live behavior on
      text
      gpt-5.4-mini
      :
      • stable prefix:
        text
        cacheRead >= 4608
        , hit rate
        text
        >= 0.90
      • tool transcript:
        text
        cacheRead >= 4096
        , hit rate
        text
        >= 0.85
      • image transcript:
        text
        cacheRead >= 3840
        , hit rate
        text
        >= 0.82
      • MCP-style transcript:
        text
        cacheRead >= 4096
        , hit rate
        text
        >= 0.85

    Fresh combined live verification on 2026-04-04 landed at:

    • stable prefix:
      text
      cacheRead=4864
      , hit rate
      text
      0.966
    • tool transcript:
      text
      cacheRead=4608
      , hit rate
      text
      0.896
    • image transcript:
      text
      cacheRead=4864
      , hit rate
      text
      0.954
    • MCP-style transcript:
      text
      cacheRead=4608
      , hit rate
      text
      0.891

    Recent local wall-clock time for the combined gate was about

    text
    88s
    .

    Why the assertions differ:

    • Anthropic exposes explicit cache breakpoints and moving conversation-history reuse.
    • OpenAI prompt caching is still exact-prefix sensitive, but the effective reusable prefix in live Responses traffic can plateau earlier than the full prompt.
    • Because of that, comparing Anthropic and OpenAI by a single cross-provider percentage threshold creates false regressions.

    text
    diagnostics.cacheTrace
    config

    yaml
    diagnostics: cacheTrace: enabled: true filePath: "~/.openclaw/logs/cache-trace.jsonl" # optional includeMessages: false # default true includePrompt: false # default true includeSystem: false # default true

    Defaults:

    • text
      filePath
      :
      text
      $OPENCLAW_STATE_DIR/logs/cache-trace.jsonl
    • text
      includeMessages
      :
      text
      true
    • text
      includePrompt
      :
      text
      true
    • text
      includeSystem
      :
      text
      true

    Env toggles (one-off debugging)

    • text
      OPENCLAW_CACHE_TRACE=1
      enables cache tracing.
    • text
      OPENCLAW_CACHE_TRACE_FILE=/path/to/cache-trace.jsonl
      overrides output path.
    • text
      OPENCLAW_CACHE_TRACE_MESSAGES=0|1
      toggles full message payload capture.
    • text
      OPENCLAW_CACHE_TRACE_PROMPT=0|1
      toggles prompt text capture.
    • text
      OPENCLAW_CACHE_TRACE_SYSTEM=0|1
      toggles system prompt capture.

    What to inspect

    • Cache trace events are JSONL and include staged snapshots like
      text
      session:loaded
      ,
      text
      prompt:before
      ,
      text
      stream:context
      , and
      text
      session:after
      .
    • Per-turn cache token impact is visible in normal usage surfaces via
      text
      cacheRead
      and
      text
      cacheWrite
      (for example
      text
      /usage full
      and session usage summaries).
    • For Anthropic, expect both
      text
      cacheRead
      and
      text
      cacheWrite
      when caching is active.
    • For OpenAI, expect
      text
      cacheRead
      on cache hits and
      text
      cacheWrite
      to remain
      text
      0
      ; OpenAI does not publish a separate cache-write token field.
    • If you need request tracing, log request IDs and rate-limit headers separately from cache metrics. OpenClaw's current cache-trace output is focused on prompt/session shape and normalized token usage rather than raw provider response headers.

    Quick troubleshooting

    • High
      text
      cacheWrite
      on most turns: check for volatile system-prompt inputs and verify model/provider supports your cache settings.
    • High
      text
      cacheWrite
      on Anthropic: often means the cache breakpoint is landing on content that changes every request.
    • Low OpenAI
      text
      cacheRead
      : verify the stable prefix is at the front, the repeated prefix is at least 1024 tokens, and the same
      text
      prompt_cache_key
      is reused for turns that should share a cache.
    • No effect from
      text
      cacheRetention
      : confirm model key matches
      text
      agents.defaults.models["provider/model"]
      .
    • Bedrock Nova/Mistral requests with cache settings: expected runtime force to
      text
      none
      .

    Related docs:

    • Anthropic
    • Token use and costs
    • Session pruning
    • Gateway configuration reference

    Related

    • Token use and costs
    • API usage and costs

    © 2024 TaskFlow Mirror

    Powered by TaskFlow Sync Engine