Generazione video

Modelli multi-modali con Hugging Face

James Chapman

Curriculum Manager, DataCamp

Generazione video

Fasi della generazione video

1 https://link.springer.com/article/10.1007/s11263-024-02271-9
Modelli multi-modali con Hugging Face

Generazione video

import torch
from diffusers import CogVideoXPipeline


pipe = CogVideoXPipeline.from_pretrained( "THUDM/CogVideoX-2b", torch_dtype=torch.float16 )
pipe.enable_model_cpu_offload() pipe.enable_sequential_cpu_offload()
pipe.vae.enable_slicing() pipe.vae.enable_tiling()

Video di esempio dal modello CogVideoX con prompt che mostra un astronauta che stringe la mano a un extraterrestre

1 https://huggingface.co/THUDM/CogVideoX-2b
Modelli multi-modali con Hugging Face

Generazione video

prompt = "A majestic lion in a sunlit African savanna, sitting regally 
on a rock formation. Golden sunlight illuminates its magnificent mane, 
then a big smile appears on its face"

video = pipe(
    prompt=prompt,

num_inference_steps=20,
num_frames=20,
guidance_scale=6,
generator=torch.Generator(device="cuda").manual_seed(42), ).frames[0]
Modelli multi-modali con Hugging Face

Generazione video

from diffusers.utils import export_to_video
from moviepy.editor import VideoFileClip


video_path = export_to_video(video, "output.mp4", fps=8) video = VideoFileClip(video_path)
video.write_gif("video.gif")

GIF del leone che sorride dal prompt precedente

Modelli multi-modali con Hugging Face

Analisi quantitativa

  • Aderenza al prompt difficile nei video
  • CLIP offre una possibile strategia:

Schema del confronto dei punteggi video CLIP

Modelli multi-modali con Hugging Face

Analisi quantitativa

from diffusers.utils import load_video
from torchmetrics.functional.multimodal import clip_score
from functools import partial

frames = load_video(video_path) clip_score_fn = partial(clip_score, model_name_or_path="openai/clip-vit-base-patch16")
scores = [] for frame in frames: frame_int = np.array(frame).astype("uint8") frame_tensor = torch.from_numpy(frame_int).unsqueeze(0).permute(0, 3, 1, 2)
score = clip_score_fn(frame_tensor, [prompt]).detach() scores.append(float(score)) avg_clip_score = round(np.mean(scores), 4) print(f"Average CLIP score: {avg_clip_score}")
Average CLIP score: 30.6274
Modelli multi-modali con Hugging Face

Passons à la pratique !

Modelli multi-modali con Hugging Face

Preparing Video For Download...