lllyasviel/framepack
Framepack is an efficient Image-to-video model that autoregressively generates videos.
Pricing
lllyasviel/
framepack
Pricing for Synexa AI models works differently from other providers. Instead of being billed by time, you are billed by input and output, making pricing more predictable.
For example, generating 100 videos should cost around $10.00.
Check out our docs for more information about how per-request pricing works on Synexa.
| Provider | Price ($) | Saving (%) |
|---|---|---|
| Synexa | $0.1000 | - |
| fal | $0.1700 | 41.2% |
Readme
FramePack - Progressive Video Generation
FramePack is a powerful image-to-video generation model that builds videos progressively, starting from a single image and a text prompt. Based on the research paper "Packing Input Frame Context in Next-Frame Prediction Models for Video Generation."
How It Works
FramePack generates videos progressively rather than all at once. It builds the video in sections, cleverly "packing" the context of previously generated frames into a fixed-size summary. This approach prevents quality degradation over longer sequences and maintains temporal coherence throughout the video.
Key Features
- Progressive Generation: Video develops section by section from your initial image
- Image-to-Video with Text Control: Start with a still image and use a text prompt to describe the motion and story
- Efficient Memory Use: Runs effectively on different GPUs by intelligently managing model loading and VRAM usage
- Smooth Video Output: Uses blending techniques to seamlessly merge newly generated video sections with previous ones
- Configurable Length: Generate videos from 1 to 60 seconds
Input Parameters
- input_image (required): Starting image for the video
- prompt (required): Text description of the desired video content
- negative_prompt: Things to avoid in the video
- total_video_length_seconds: Duration of the output video (1-60 seconds, default: 3)
- steps: Inference steps - more steps give higher quality but take longer (default: 25)
- cfg_scale: Controls how closely the output follows the prompt (default: 1)
- seed: Set for reproducible results or leave empty for random
Use Cases
- Bring still images to life: Animate a photograph or illustration based on a narrative prompt
- Create short dynamic clips: Generate videos of characters, objects, or scenes in motion
- Visualize evolving scenes: Show a scene changing over time, like a sunset or a character's expression shifting
- Storytelling and concept art: Quickly prototype visual ideas that unfold over time
Tips for Best Results
- Use clear, descriptive prompts that specify the motion you want
- Start with shorter videos and increase length once you find good settings
- Higher step counts improve quality but increase generation time
- The distilled_cfg_scale parameter (default: 10) helps maintain prompt adherence
- Enable use_teacache for faster generation with minimal quality impact