π DAMO Text-to-Video MS-1.7B (Rehosted by Nachiketh)
This model is a rehosted and learner-accessible version of the original damo-vilab/text-to-video-ms-1.7b from Alibaba DAMO.
It enables text-to-video generation using diffusion models and is ideal for researchers, learners, and creators building cutting-edge Generative AI experiences.
π Model Summary
- Name: DAMO Text-to-Video MS-1.7B
- Architecture: Multi-stage latent diffusion
- Inputs: Text prompt
- Outputs: 16-frame videos (resolution varies)
- Framework: Hugging Face π€ Diffusers
- Precision: FP16 recommended
- License: CreativeML OpenRAIL-M
π§ͺ Example Usage (Python)
Install the Libraries
$ pip install diffusers transformers accelerate torch
from diffusers import DiffusionPipeline
import torch
from diffusers.utils import export_to_video
pipe = DiffusionPipeline.from_pretrained(
"nachikethmurthy666/text-to-video-ms-1.7b",
torch_dtype=torch.float16,
variant="fp16"
).to("cuda")
pipe.enable_model_cpu_offload()
prompt = "An astronaut riding a horse on Mars at sunset"
video = pipe(prompt, num_inference_steps=25).frames[0]
video_path = export_to_video(video)
- Downloads last month
- 42
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support