name: qwen-image-skill description: | Generate images through a configurable Qwen-compatible image generation API. This skill lets OpenClaw turn natural-language image requests into structured parameters, call the configured Qwen image model, and return downloaded image files.
Use this skill when: (1) The user wants to generate an image from text. (2) The user describes a scene, character, style, or composition and expects a rendered picture.
As an OpenClaw agent equipped with this skill, your objective is to convert the user's natural-language image request into a structured argument payload, then hand it to the local Python client to generate images through the configured Qwen-compatible API.
Before generating, inspect the currently exposed workflow and parameter surface:
python ./scripts/registry.py list --agent
Rules:
prompt as required.size, n, quality, or negative_prompt, infer sensible defaults from the request.Create a JSON object for the workflow qwen/text-to-image.
Expected arguments:
prompt: fully written image prompt.size: optional output size, such as 1024x1024, 1280x720, or 720x1280.n: optional image count.quality: optional quality hint such as standard or hd.negative_prompt: optional list of unwanted elements.seed: optional deterministic seed.If the user intent is vague, ask for the missing art direction only when necessary. Otherwise, refine the request yourself into a production-ready prompt.
Run the local client:
python ./scripts/qwen_image_client.py --workflow qwen/text-to-image --args '<JSON_ARGS>'
Requirements:
--args.config.json from config.example.json.On success, return the generated local image paths and a concise summary of what was generated.
On failure:
provider.api_key, provider.base_url, and provider.model in config.json.This skill reads runtime settings from config.json in the repository root. If that file does not exist yet, copy config.example.json to config.json and fill in the Qwen settings.