Reference to Video

WAN 2.7 Reference-to-Video

Character-consistent video from reference images/videos.

Schema
Add images (0/5)

Upload up to 5 images

Add videos (0/3)

Upload up to 3 videos

Hint: You can drag and drop a file or click to upload

Example

Sample output shown above. The prompt and settings are pre-filled on the left.

Your request will cost $0.50 per run. For $10 you can run this model approximately 20 times.

One more thing:

README

Alibaba / WAN 2.7 Reference-to-Video β€” Reference to Video (wan-2.7-r2v)

WAN 2.7 Reference-to-Video creates character-consistent videos from reference images and videos while maintaining appearance, voice tone, and style.

Highlights

  • β€’Reference images – Provide up to 5 images for character/object consistency.
  • β€’Reference videos – Provide up to 3 video clips for motion and style reference.
  • β€’Reference voice – Add voice audio for character voice consistency.
  • β€’Multi-character – Support distinct references for multiple characters in a scene.

Parameters

  • β€’prompt* – Text description to guide video generation. Reference characters with "Image 1", "Video 1" labels.
  • β€’negative_prompt – Things to avoid in the video
  • β€’reference_images – Reference images (up to 5)
  • β€’reference_videos – Reference videos (up to 3)
  • β€’reference_voice_url – Audio for voice reference consistency
  • β€’resolution – Output video resolution
    • β€’ 720p
    • β€’ 1080p
  • β€’duration – Video duration in seconds (2-10)
  • β€’enable_prompt_expansion – Enable LLM prompt rewriting for better quality
  • β€’seed – Random seed for reproducibility (0-2147483647)

Pricing

$1.00 per generation

ResolutionPrice
720p$0.12
1080p$0.18

How to Use

  1. 1.Provide reference images and/or videos as URL arrays.
  2. 2.Use labels like "Image 1", "Video 1" in the prompt to reference specific inputs.
  3. 3.Set resolution (720P/1080P) and duration (2-10s).

Pro Tips

  • β€’Clear, front-facing reference images produce the best character consistency.
  • β€’Combined total of images + videos cannot exceed 5.
  • β€’Use reference_voice_url for lip-sync and voice tone consistency.

More Models to Try

β€’Creating consistent character animations from photos.
β€’Generating videos with specific people, objects, or styles.
β€’Multi-character storytelling with consistent appearances.

Frequently Asked Questions

What is the WAN 2.7 Reference-to-Video API?
Character-consistent video from reference images/videos.
How much does WAN 2.7 Reference-to-Video cost via API?
WAN 2.7 Reference-to-Video costs $1.0000 per generation through Renderful's API. No subscription required β€” pay only for what you use.
How do I use WAN 2.7 Reference-to-Video via API?
Sign up for a free Renderful API key, then send a POST request to the /v1/predictions endpoint with model "wan-2.7-r2v". See the documentation at renderful.ai/docs for code examples in Python, JavaScript, and cURL.
What type of content does WAN 2.7 Reference-to-Video generate?
WAN 2.7 Reference-to-Video is a reference to video model by Alibaba. Key features include: Up to 5 reference images, Up to 3 reference videos, Reference voice.
Is the WAN 2.7 Reference-to-Video API fast?
WAN 2.7 Reference-to-Video has medium generation speed. Results are delivered via polling or webhook callback for seamless integration.