mochi-lora / README.md
sayakpaul's picture
sayakpaul HF staff
End of training
ebefd78 verified
---
base_model: genmo/mochi-1-preview
library_name: diffusers
license: apache-2.0
instance_prompt: A black and white animated scene unfolds with an anthropomorphic
goat surrounded by musical notes and symbols, suggesting a playful environment.
Mickey Mouse appears, leaning forward in curiosity as the goat remains still. The
goat then engages with Mickey, who bends down to converse or react. The dynamics
shift as Mickey grabs the goat, potentially in surprise or playfulness, amidst a
minimalistic background. The scene captures the evolving relationship between the
two characters in a whimsical, animated setting, emphasizing their interactions
and emotions
widget:
- text: A black and white animated scene unfolds with an anthropomorphic goat surrounded
by musical notes and symbols, suggesting a playful environment. Mickey Mouse appears,
leaning forward in curiosity as the goat remains still. The goat then engages
with Mickey, who bends down to converse or react. The dynamics shift as Mickey
grabs the goat, potentially in surprise or playfulness, amidst a minimalistic
background. The scene captures the evolving relationship between the two characters
in a whimsical, animated setting, emphasizing their interactions and emotions
output:
url: final_video_0.mp4
tags:
- text-to-video
- diffusers-training
- diffusers
- lora
- mochi-1-preview
- mochi-1-preview-diffusers
- template:sd-lora
---
<!-- This model card has been generated automatically according to the information the training script had access to. You
should probably proofread and complete it, then remove this comment. -->
# Mochi-1 Preview LoRA Finetune
<Gallery />
## Model description
This is a lora finetune of the Mochi-1 preview model `genmo/mochi-1-preview`.
The model was trained using [CogVideoX Factory](https://github.com/a-r-r-o-w/cogvideox-factory) - a repository containing memory-optimized training scripts for the CogVideoX and Mochi family of models using [TorchAO](https://github.com/pytorch/ao) and [DeepSpeed](https://github.com/microsoft/DeepSpeed). The scripts were adopted from [CogVideoX Diffusers trainer](https://github.com/huggingface/diffusers/blob/main/examples/cogvideo/train_cogvideox_lora.py).
## Download model
[Download LoRA](sayakpaul/mochi-lora/tree/main) in the Files & Versions tab.
## Usage
Requires the [🧨 Diffusers library](https://github.com/huggingface/diffusers) installed.
```py
from diffusers import MochiPipeline
from diffusers.utils import export_to_video
import torch
pipe = MochiPipeline.from_pretrained("genmo/mochi-1-preview")
pipe.load_lora_weights("CHANGE_ME")
pipe.enable_model_cpu_offload()
with torch.autocast("cuda", torch.bfloat16):
video = pipe(
prompt="CHANGE_ME",
guidance_scale=6.0,
num_inference_steps=64,
height=480,
width=848,
max_sequence_length=256,
output_type="np"
).frames[0]
export_to_video(video)
```
For more details, including weighting, merging and fusing LoRAs, check the [documentation](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters) on loading LoRAs in diffusers.
## Intended uses & limitations
#### How to use
```python
# TODO: add an example code snippet for running this diffusion pipeline
```
#### Limitations and bias
[TODO: provide examples of latent issues and potential remediations]
## Training details
[TODO: describe the data used to train the model]