How to use from the
Use from the
Diffusers library
pip install -U diffusers transformers accelerate
import torch
from diffusers import DiffusionPipeline

# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("vdo/animov-0.1.1", dtype=torch.bfloat16, device_map="cuda")

prompt = "Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"
image = pipe(prompt).images[0]

This is a text2video model for diffusers, fine-tuned with a modelscope to have an anime-style appearance.
It was trained at 448x384 resolution.
The usage is the same as with the original modelscope model.

The main difference from version 0.1 is only the resolution.

Downloads last month
84
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support