Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Route Alibaba Cloud Model Studio requests to the right local skill (Qwen Image, Qwen Image Edit, Wan Video, Wan R2V, Qwen TTS and advanced TTS variants). Use when the user asks for Model Studio without specifying a capability.
Route Alibaba Cloud Model Studio requests to the right local skill (Qwen Image, Qwen Image Edit, Wan Video, Wan R2V, Qwen TTS and advanced TTS variants). Use when the user asks for Model Studio without specifying a capability.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Route requests to existing local skills to avoid duplicating model/parameter details.
Install SDK (virtual environment recommended to avoid PEP 668 restrictions): python3 -m venv .venv . .venv/bin/activate python -m pip install dashscope Configure DASHSCOPE_API_KEY (environment variable preferred; or dashscope_api_key in ~/.alibabacloud/credentials).
NeedTarget skillText-to-image / image generationskills/ai/image/alicloud-ai-image-qwen-image/Image editingskills/ai/image/alicloud-ai-image-qwen-image-edit/Text-to-video / image-to-video (i2v)skills/ai/video/alicloud-ai-video-wan-video/Reference-to-video (r2v)skills/ai/video/alicloud-ai-video-wan-r2v/Text-to-speech (TTS)skills/ai/audio/alicloud-ai-audio-tts/Speech recognition/transcription (ASR)skills/ai/audio/alicloud-ai-audio-asr/Realtime speech recognitionskills/ai/audio/alicloud-ai-audio-asr-realtime/Realtime TTSskills/ai/audio/alicloud-ai-audio-tts-realtime/Live speech translationskills/ai/audio/alicloud-ai-audio-livetranslate/CosyVoice voice cloneskills/ai/audio/alicloud-ai-audio-cosyvoice-voice-clone/CosyVoice voice designskills/ai/audio/alicloud-ai-audio-cosyvoice-voice-design/Voice cloneskills/ai/audio/alicloud-ai-audio-tts-voice-clone/Voice designskills/ai/audio/alicloud-ai-audio-tts-voice-design/Omni multimodal interactionskills/ai/multimodal/alicloud-ai-multimodal-qwen-omni/Visual reasoningskills/ai/multimodal/alicloud-ai-multimodal-qvq/Text embeddingsskills/ai/search/alicloud-ai-search-text-embedding/Rerankskills/ai/search/alicloud-ai-search-rerank/Vector retrievalskills/ai/search/alicloud-ai-search-dashvector/ or skills/ai/search/alicloud-ai-search-opensearch/ or skills/ai/search/alicloud-ai-search-milvus/Document understandingskills/ai/text/alicloud-ai-text-document-mind/Video editingskills/ai/video/alicloud-ai-video-wan-edit/Model list crawl/updateskills/ai/misc/alicloud-ai-misc-crawl-and-skill/
Clarify model capability and input/output type first. If capability is missing in repo, add a new skill first.
text generation/chat (LLM) multimodal embeddings OCR-specialized extraction and image translation virtual try-on / digital human / advanced video personas For multimodal/ASR download failures, prefer public URLs listed above. For ASR parameter errors, use data URI in input_audio.data. For multimodal embedding 400, ensure input.contents is an array.
When X-DashScope-Async: enable returns task_id, poll as follows: GET https://dashscope.aliyuncs.com/api/v1/tasks/<task_id> Authorization: Bearer $DASHSCOPE_API_KEY Example result fields (success): { "output": { "task_status": "SUCCEEDED", "video_url": "https://..." } } Notes: Recommended polling interval: 15-20 seconds, max 10 attempts. After success, download output.video_url.
Are you working with text, image, audio, or video? Is this generation, editing/understanding, or retrieval? Do you need speech (TTS/ASR/live translate) or retrieval (embedding/rerank/vector DB)? Do you want runnable SDK scripts or just API/parameter guidance?
Model list and links:output/alicloud-model-studio-models-summary.md API/parameters/examples: see target sub-skill SKILL.md and references/*.md Official source list:references/sources.md
mkdir -p output/alicloud-ai-entry-modelstudio echo "validation_placeholder" > output/alicloud-ai-entry-modelstudio/validate.txt Pass criteria: command exits 0 and output/alicloud-ai-entry-modelstudio/validate.txt is generated.
Save artifacts, command outputs, and API response summaries under output/alicloud-ai-entry-modelstudio/. Include key parameters (region/resource id/time range) in evidence files for reproducibility.
Confirm user intent, region, identifiers, and whether the operation is read-only or mutating. Run one minimal read-only query first to verify connectivity and permissions. Execute the target operation with explicit parameters and bounded scope. Verify results and save output/evidence files.
Agent frameworks, memory systems, reasoning layers, and model-native orchestration.
Largest current source with strong distribution and engagement signals.