Caricamento in corso...
Caricamento in corso...
Last synced: Today, 22:00
Technical reference for the OpenClaw framework. Real-time synchronization with the official documentation engine.
Use this file to discover all available pages before exploring further.
openclaw inferIt intentionally exposes capability families, not raw gateway RPC names and not raw agent tool ids.
Copy and paste this to an agent:
textRead https://docs.openclaw.ai/cli/infer, then create a skill that routes my common workflows to `openclaw infer`. Focus on model runs, image generation, video generation, audio transcription, TTS, web search, and embeddings.
A good infer-based skill should:
openclaw infer ...Typical infer-focused skill coverage:
openclaw infer model runopenclaw infer image generateopenclaw infer audio transcribeopenclaw infer tts convertopenclaw infer web searchopenclaw infer embedding createopenclaw inferBenefits:
--jsonFor end-to-end provider checks, prefer
openclaw infer ...textopenclaw infer list inspect model run list inspect providers auth login auth logout auth status image generate edit describe describe-many providers audio transcribe providers tts convert voices providers status enable disable set-provider video generate describe providers web search fetch providers embedding create providers
This table maps common inference tasks to the corresponding infer command.
| Task | Command | Notes |
|---|---|---|
| Run a text/model prompt | text openclaw infer model run --prompt "..." --json | Uses the normal local path by default |
| Run a model prompt on images | text openclaw infer model run --prompt "Describe this" --file ./image.png --model provider/model | Repeat text --file |
| Generate an image | text openclaw infer image generate --prompt "..." --json | Use text image edit |
| Describe an image file | text openclaw infer image describe --file ./image.png --prompt "..." --json | text --modeltext <provider/model> |
| Transcribe audio | text openclaw infer audio transcribe --file ./memo.m4a --json | text --modeltext <provider/model> |
| Synthesize speech | text openclaw infer tts convert --text "..." --output ./speech.mp3 --json | text tts status |
| Generate a video | text openclaw infer video generate --prompt "..." --json | Supports provider hints such as text --resolution |
| Describe a video file | text openclaw infer video describe --file ./clip.mp4 --json | text --modeltext <provider/model> |
| Search the web | text openclaw infer web search --query "..." --json | |
| Fetch a web page | text openclaw infer web fetch --url https://example.com --json | |
| Create embeddings | text openclaw infer embedding create --text "..." --json |
openclaw infer ...--json--provider--model provider/modelimage describeaudio transcribevideo describe--model<provider/model>image describe--modelcodex/<model>openai-codex/<model>model runmodel run --file--filemodel run --fileinfer audio transcribeinfer video describemodel run --gatewaymodel run --gateway --model <provider/model>Use
modelbashopenclaw infer model run --prompt "Reply with exactly: smoke-ok" --json openclaw infer model run --prompt "Summarize this changelog entry" --model openai/gpt-5.4 --json openclaw infer model run --prompt "Describe this image in one sentence" --file ./photo.jpg --model google/gemini-2.5-flash --json openclaw infer model providers --json openclaw infer model inspect --name gpt-5.5 --json
Use full
<provider/model>bashopenclaw infer model run --local --model anthropic/claude-sonnet-4-6 --prompt "Reply with exactly: pong" --json openclaw infer model run --local --model cerebras/zai-glm-4.7 --prompt "Reply with exactly: pong" --json openclaw infer model run --local --model google/gemini-2.5-flash --prompt "Reply with exactly: pong" --json openclaw infer model run --local --model groq/llama-3.1-8b-instant --prompt "Reply with exactly: pong" --json openclaw infer model run --local --model mistral/mistral-small-latest --prompt "Reply with exactly: pong" --json openclaw infer model run --local --model openai/gpt-4.1 --prompt "Reply with exactly: pong" --json openclaw infer model run --local --model ollama/qwen2.5vl:7b --prompt "Describe this image." --file ./photo.jpg --json
Notes:
model runmodel run --fileimage/*model run --fileinfer image describemodel run --promptmodel runmodel run --gatewayopenclaw agentmodel auth loginmodel auth logoutmodel auth statusUse
imagebashopenclaw infer image generate --prompt "friendly lobster illustration" --json openclaw infer image generate --prompt "cinematic product photo of headphones" --json openclaw infer image generate --model openai/gpt-image-1.5 --output-format png --background transparent --prompt "simple red circle sticker on a transparent background" --json openclaw infer image generate --prompt "slow image backend" --timeout-ms 180000 --json openclaw infer image edit --file ./logo.png --model openai/gpt-image-1.5 --output-format png --background transparent --prompt "keep the logo, remove the background" --json openclaw infer image edit --file ./poster.png --prompt "make this a vertical story ad" --size 2160x3840 --aspect-ratio 9:16 --resolution 4K --json openclaw infer image describe --file ./photo.jpg --json openclaw infer image describe --file ./receipt.jpg --prompt "Extract the merchant, date, and total" --json openclaw infer image describe-many --file ./before.png --file ./after.png --prompt "Compare the screenshots and list visible UI changes" --json openclaw infer image describe --file ./ui-screenshot.png --model openai/gpt-4.1-mini --json openclaw infer image describe --file ./photo.jpg --model ollama/qwen2.5vl:7b --prompt "Describe the image in one sentence" --timeout-ms 300000 --json
Notes:
Use
image editUse
--size--aspect-ratio--resolutionimage editUse
--output-format png --background transparent--model openai/gpt-image-1.5--openai-backgroundUse
image providers --jsonUse
image generate --model <provider/model> --jsonbashopenclaw infer image providers --json openclaw infer image generate \ --model google/gemini-3.1-flash-image-preview \ --prompt "Minimal flat test image: one blue square on a white background, no text." \ --output ./openclaw-infer-image-smoke.png \ --json
The JSON response reports
okprovidermodelattempts--outputFor
image describeimage describe-many--promptUse
--timeout-msFor
image describe--model<provider/model>For local Ollama vision models, pull the model first and set
OLLAMA_API_KEYollama-localUse
audiobashopenclaw infer audio transcribe --file ./memo.m4a --json openclaw infer audio transcribe --file ./team-sync.m4a --language en --prompt "Focus on names and action items" --json openclaw infer audio transcribe --file ./memo.m4a --model openai/whisper-1 --json
Notes:
audio transcribe--model<provider/model>Use
ttsbashopenclaw infer tts convert --text "hello from openclaw" --output ./hello.mp3 --json openclaw infer tts convert --text "Your build is complete" --output ./build-complete.mp3 --json openclaw infer tts providers --json openclaw infer tts status --json
Notes:
tts statustts providerstts voicestts set-providerUse
videobashopenclaw infer video generate --prompt "cinematic sunset over the ocean" --json openclaw infer video generate --prompt "slow drone shot over a forest lake" --resolution 768P --duration 6 --json openclaw infer video describe --file ./clip.mp4 --json openclaw infer video describe --file ./clip.mp4 --model openai/gpt-4.1-mini --json
Notes:
video generate--size--aspect-ratio--resolution--duration--audio--watermark--timeout-ms--model<provider/model>video describeUse
webbashopenclaw infer web search --query "OpenClaw docs" --json openclaw infer web search --query "OpenClaw infer web providers" --json openclaw infer web fetch --url https://docs.openclaw.ai/cli/infer --json openclaw infer web providers --json
Notes:
web providersUse
embeddingbashopenclaw infer embedding create --text "friendly lobster" --json openclaw infer embedding create --text "customer support ticket: delayed shipment" --model openai/text-embedding-3-large --json openclaw infer embedding providers --json
Infer commands normalize JSON output under a shared envelope:
json{ "ok": true, "capability": "image.generate", "transport": "local", "provider": "openai", "model": "gpt-image-2", "attempts": [], "outputs": [] }
Top-level fields are stable:
okcapabilitytransportprovidermodelattemptsoutputserrorFor generated media commands,
outputspathmimeTypesizebash# Bad openclaw infer media image generate --prompt "friendly lobster" # Good openclaw infer image generate --prompt "friendly lobster"
bash# Bad openclaw infer audio transcribe --file ./memo.m4a --model whisper-1 --json # Good openclaw infer audio transcribe --file ./memo.m4a --model openai/whisper-1 --json
openclaw capability ...openclaw infer ...© 2024 TaskFlow Mirror
Powered by TaskFlow Sync Engine