Reference to Video
WAN 2.7 Reference-to-Video
Character-consistent video from reference images/videos.
Add images (0/5)
Upload up to 5 images
Add videos (0/3)
Upload up to 3 videos
Hint: You can drag and drop a file or click to upload
Example
Sample output shown above. The prompt and settings are pre-filled on the left.
Your request will cost $0.50 per run. For $10 you can run this model approximately 20 times.
One more thing:
Related Models
README
Alibaba / WAN 2.7 Reference-to-Video β Reference to Video (wan-2.7-r2v)
WAN 2.7 Reference-to-Video creates character-consistent videos from reference images and videos while maintaining appearance, voice tone, and style.
Highlights
- β’Reference images β Provide up to 5 images for character/object consistency.
- β’Reference videos β Provide up to 3 video clips for motion and style reference.
- β’Reference voice β Add voice audio for character voice consistency.
- β’Multi-character β Support distinct references for multiple characters in a scene.
Parameters
- β’prompt* β Text description to guide video generation. Reference characters with "Image 1", "Video 1" labels.
- β’negative_prompt β Things to avoid in the video
- β’reference_images β Reference images (up to 5)
- β’reference_videos β Reference videos (up to 3)
- β’reference_voice_url β Audio for voice reference consistency
- β’resolution β Output video resolution
- β’ 720p
- β’ 1080p
- β’duration β Video duration in seconds (2-10)
- β’enable_prompt_expansion β Enable LLM prompt rewriting for better quality
- β’seed β Random seed for reproducibility (0-2147483647)
Pricing
$1.00 per generation
| Resolution | Price |
|---|---|
| 720p | $0.12 |
| 1080p | $0.18 |
How to Use
- 1.Provide reference images and/or videos as URL arrays.
- 2.Use labels like "Image 1", "Video 1" in the prompt to reference specific inputs.
- 3.Set resolution (720P/1080P) and duration (2-10s).
Pro Tips
- β’Clear, front-facing reference images produce the best character consistency.
- β’Combined total of images + videos cannot exceed 5.
- β’Use reference_voice_url for lip-sync and voice tone consistency.
More Models to Try
β’Creating consistent character animations from photos.
β’Generating videos with specific people, objects, or styles.
β’Multi-character storytelling with consistent appearances.
Related Guides
Frequently Asked Questions
What is the WAN 2.7 Reference-to-Video API?
Character-consistent video from reference images/videos.
How much does WAN 2.7 Reference-to-Video cost via API?
WAN 2.7 Reference-to-Video costs $1.0000 per generation through Renderful's API. No subscription required β pay only for what you use.
How do I use WAN 2.7 Reference-to-Video via API?
Sign up for a free Renderful API key, then send a POST request to the /v1/predictions endpoint with model "wan-2.7-r2v". See the documentation at renderful.ai/docs for code examples in Python, JavaScript, and cURL.
What type of content does WAN 2.7 Reference-to-Video generate?
WAN 2.7 Reference-to-Video is a reference to video model by Alibaba. Key features include: Up to 5 reference images, Up to 3 reference videos, Reference voice.
Is the WAN 2.7 Reference-to-Video API fast?
WAN 2.7 Reference-to-Video has medium generation speed. Results are delivered via polling or webhook callback for seamless integration.
