torchao-diffusers / vae-compilation /benchmark_pixart.py
sayakpaul's picture
sayakpaul HF staff
Upload folder using huggingface_hub
7a15650 verified
import torch
torch.set_float32_matmul_precision("high")
import torch.utils.benchmark as benchmark
from diffusers import DiffusionPipeline
import gc
from torchao.quantization import (
int4_weight_only,
int8_weight_only,
int8_dynamic_activation_int8_weight,
quantize_,
autoquant,
)
from torchao.float8.inference import ActivationCasting, QuantConfig, quantize_to_float8
from torchao.prototype.quant_llm import fp6_llm_weight_only
from torchao.sparsity import sparsify_, int8_dynamic_activation_int8_semi_sparse_weight
from tabulate import tabulate
import argparse
import json
PROMPT = "Eiffel Tower was Made up of more than 2 million translucent straws to look like a cloud, with the bell tower at the top of the building, Michel installed huge foam-making machines in the forest to blow huge amounts of unpredictable wet clouds in the building's classic architecture."
PREFIXES = {
"stabilityai/stable-diffusion-3-medium-diffusers": "sd3",
"PixArt-alpha/PixArt-Sigma-XL-2-1024-MS": "pixart",
"fal/AuraFlow": "auraflow",
}
def flush():
gc.collect()
torch.cuda.empty_cache()
torch.cuda.reset_max_memory_allocated()
torch.cuda.reset_peak_memory_stats()
def bytes_to_giga_bytes(bytes):
return f"{(bytes / 1024 / 1024 / 1024):.3f}"
def benchmark_fn(f, *args, **kwargs):
t0 = benchmark.Timer(
stmt="f(*args, **kwargs)",
globals={"args": args, "kwargs": kwargs, "f": f},
num_threads=torch.get_num_threads(),
)
return f"{(t0.blocked_autorange().mean):.3f}"
def load_pipeline(
ckpt_id: str,
fuse_attn_projections: bool,
compile: bool,
quantization: str,
sparsify: bool,
) -> DiffusionPipeline:
pipeline = DiffusionPipeline.from_pretrained(ckpt_id, torch_dtype=torch.bfloat16).to("cuda")
if fuse_attn_projections:
pipeline.transformer.fuse_qkv_projections()
pipeline.vae.fuse_qkv_projections()
if quantization == "autoquant" and compile:
pipeline.transformer.to(memory_format=torch.channels_last)
pipeline.transformer = torch.compile(pipeline.transformer, mode="max-autotune", fullgraph=True)
pipeline.vae.to(memory_format=torch.channels_last)
pipeline.vae.decode = torch.compile(pipeline.vae.decode, mode="max-autotune", fullgraph=True)
if not sparsify:
if quantization == "int8dq":
quantize_(pipeline.transformer, int8_dynamic_activation_int8_weight())
quantize_(pipeline.vae, int8_dynamic_activation_int8_weight())
elif quantization == "int8wo":
quantize_(pipeline.transformer, int8_weight_only())
quantize_(pipeline.vae, int8_weight_only())
elif quantization == "int4wo":
quantize_(pipeline.transformer, int4_weight_only())
quantize_(pipeline.vae, int4_weight_only())
elif quantization == "fp6":
quantize_(pipeline.transformer, fp6_llm_weight_only())
quantize_(pipeline.vae, fp6_llm_weight_only())
elif quantization == "fp8":
pipeline.transformer = quantize_to_float8(pipeline.transformer, QuantConfig(ActivationCasting.DYNAMIC))
pipeline.vae = quantize_to_float8(pipeline.vae, QuantConfig(ActivationCasting.DYNAMIC))
elif quantization == "autoquant":
pipeline.transformer = autoquant(pipeline.transformer)
pipeline.vae = autoquant(pipeline.vae)
if sparsify:
sparsify_(pipeline.transformer, int8_dynamic_activation_int8_semi_sparse_weight())
sparsify_(pipeline.vae, int8_dynamic_activation_int8_semi_sparse_weight())
if quantization != "autoquant" and compile:
pipeline.transformer.to(memory_format=torch.channels_last)
pipeline.transformer = torch.compile(pipeline.transformer, mode="max-autotune", fullgraph=True)
pipeline.vae.to(memory_format=torch.channels_last)
pipeline.vae.decode = torch.compile(pipeline.vae.decode, mode="max-autotune", fullgraph=True)
pipeline.set_progress_bar_config(disable=True)
return pipeline
def run_inference(pipe, batch_size):
_ = pipe(
prompt=PROMPT,
num_images_per_prompt=batch_size,
generator=torch.manual_seed(2024),
)
def pretty_print_results(results, precision: int = 6):
def format_value(value):
if isinstance(value, float):
return f"{value:.{precision}f}"
return value
filtered_table = {k: format_value(v) for k, v in results.items()}
print(tabulate([filtered_table], headers="keys", tablefmt="pipe", stralign="center"))
def run_benchmark(pipeline, args):
for _ in range(5):
run_inference(pipeline, batch_size=args.batch_size)
time = benchmark_fn(run_inference, pipeline, args.batch_size)
torch.cuda.empty_cache()
memory = bytes_to_giga_bytes(torch.cuda.memory_allocated()) # in GBs.
info = dict(
ckpt_id=args.ckpt_id,
batch_size=args.batch_size,
fuse=args.fuse_attn_projections,
compile=args.compile,
quantization=args.quantization,
sparsify=args.sparsify,
memory=memory,
time=time,
)
pretty_print_results(info)
return info
def serialize_artifacts(info: dict, pipeline, args):
ckpt_id = PREFIXES[args.ckpt_id]
prefix = f"ckpt@{ckpt_id}-bs@{args.batch_size}-fuse@{args.fuse_attn_projections}-compile@{args.compile}-quant@{args.quantization}-sparsify@{args.sparsify}"
info_file = f"{prefix}_info.json"
with open(info_file, "w") as f:
json.dump(info, f)
image = pipeline(
prompt=PROMPT,
num_images_per_prompt=args.batch_size,
generator=torch.manual_seed(0),
).images[0]
image.save(f"{prefix}.png")
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("--ckpt_id", default="PixArt-alpha/PixArt-Sigma-XL-2-1024-MS", type=str)
parser.add_argument("--fuse_attn_projections", action="store_true")
parser.add_argument("--compile", action="store_true")
parser.add_argument(
"--quantization",
default="None",
choices=["int8dq", "int8wo", "int4wo", "autoquant", "fp6", "fp8", "None"],
help="Which quantization technique to apply",
)
parser.add_argument("--sparsify", action="store_true")
parser.add_argument("--batch_size", default=1, type=int, choices=[1, 4, 8])
args = parser.parse_args()
flush()
pipeline = load_pipeline(
ckpt_id=args.ckpt_id,
fuse_attn_projections=args.fuse_attn_projections,
compile=args.compile,
quantization=args.quantization,
sparsify=args.sparsify,
)
info = run_benchmark(pipeline, args)
serialize_artifacts(info, pipeline, args)