返回顶部
a

aliyun-wan-digital-human阿里云数字人

Use when generating talking, singing, or presentation videos from a single character image and audio with Alibaba Cloud Model Studio digital-human model `wan2.2-s2v`. Use when creating narrated avatar videos, singing portraits, or broadcast-style talking-head clips.

作者: admin | 来源: ClawHub
源自
ClawHub
版本
V 1.0.0
安全检测
已通过
76
下载量
免费
免费
0
收藏
概述
安装方式
版本历史

aliyun-wan-digital-human

Category: provider

Model Studio Digital Human

Validation

CODEBLOCK0

Pass criteria: command exits 0 and output/aliyun-wan-digital-human/validate.txt is generated.

Output And Evidence

  • - Save normalized request payloads, chosen resolution, and task polling snapshots under output/aliyun-wan-digital-human/.
  • Record image/audio URLs and whether the input image passed detection.

Use this skill for image + audio driven speaking, singing, or presenting characters.

Critical model names

Use these exact model strings:

  • - INLINECODE2
  • INLINECODE3

Selection guidance:

  • - Run wan2.2-s2v-detect first to validate the image.
  • Use wan2.2-s2v for the actual video generation job.

Prerequisites

  • - China mainland (Beijing) only.
  • Set DASHSCOPE_API_KEY in your environment, or add dashscope_api_key to ~/.alibabacloud/credentials.
  • Input audio should contain clear speech or singing, and input image should depict a clear subject.

Normalized interface (video.digital_human)

Detect Request

  • - model (string, optional): default INLINECODE10
  • INLINECODE11 (string, required)

Generate Request

  • - model (string, optional): default INLINECODE13
  • INLINECODE14 (string, required)
  • INLINECODE15 (string, required)
  • INLINECODE16 (string, optional): 480P or INLINECODE18
  • INLINECODE19 (string, optional): talk, sing, or INLINECODE22

Response

  • - task_id (string)
  • INLINECODE24 (string)
  • INLINECODE25 (string, when finished)

Quick start

CODEBLOCK1

Operational guidance

  • - Use a portrait, half-body, or full-body image with a clear face and stable framing.
  • Match audio length to the desired output duration; the output follows the audio length up to the model limit.
  • Keep image and audio as public HTTP/HTTPS URLs.
  • If the image fails detection, do not proceed directly to video generation.

Output location

  • - Default output: INLINECODE26
  • Override base dir with OUTPUT_DIR.

References

  • - INLINECODE28

标签

skill ai

通过对话安装

该技能支持在以下平台通过对话安装:

OpenClaw WorkBuddy QClaw Kimi Claude

方式一:安装 SkillHub 和技能

帮我安装 SkillHub 和 aliyun-wan-digital-human-1775882172 技能

方式二:设置 SkillHub 为优先技能安装源

设置 SkillHub 为我的优先技能安装源,然后帮我安装 aliyun-wan-digital-human-1775882172 技能

通过命令行安装

skillhub install aliyun-wan-digital-human-1775882172

下载

⬇ 下载 aliyun-wan-digital-human v1.0.0(免费)

文件大小: 3.42 KB | 发布时间: 2026-4-12 08:58

v1.0.0 最新 2026-4-12 08:58
Initial release of aliyun-wan-digital-human skill.

- Enables generation of talking, singing, or presentation videos from a character image and audio using Alibaba Cloud Model Studio digital-human models.
- Supports image validation and video generation workflows with distinct model names: `wan2.2-s2v-detect` for validation and `wan2.2-s2v` for video.
- Exposes a normalized interface for detection and video creation requests.
- Requires API key setup and China (Beijing) region.
- Outputs all requests, responses, and task snapshots to a dedicated directory for traceability.

Archiver·手机版·闲社网·闲社论坛·羊毛社区· 多链控股集团有限公司 · 苏ICP备2025199260号-1

Powered by Discuz! X5.0   © 2024-2025 闲社网·线报更新论坛·羊毛分享社区·http://xianshe.com

p2p_official_large
返回顶部