jschen
commited on
Commit
·
534bd5f
1
Parent(s):
0b40b5f
update README.md
Browse files
README.md
CHANGED
@@ -24,11 +24,12 @@ tags:
|
|
24 |
## 🔥 Why Need PixArt-LCM
|
25 |
Following [LCM LoRA](https://huggingface.co/blog/lcm_lora), we illustrative of the generation speed we achieve on various computers. Let us stress again how liberating it is to explore image generation so easily with PixArt-LCM.
|
26 |
|
27 |
-
| Hardware | PixArt-LCM (4 steps)
|
28 |
-
|
29 |
-
| T4 (Google Colab Free Tier) | 3.3s
|
30 |
-
| A100 (80 GB) | 0.51s
|
31 |
-
| V100 (32 GB) |
|
|
|
32 |
These tests were run with a batch size of 1 in all cases.
|
33 |
|
34 |
For cards with a lot of capacity, such as A100, performance increases significantly when generating multiple images at once, which is usually the case for production workloads.
|
|
|
24 |
## 🔥 Why Need PixArt-LCM
|
25 |
Following [LCM LoRA](https://huggingface.co/blog/lcm_lora), we illustrative of the generation speed we achieve on various computers. Let us stress again how liberating it is to explore image generation so easily with PixArt-LCM.
|
26 |
|
27 |
+
| Hardware | PixArt-LCM (4 steps) | SDXL LoRA LCM (4 steps) | PixArt standard (14 steps) | SDXL standard (25 steps) |
|
28 |
+
|-----------------------------|----------------------|-------------------------|----------------------------|---------------------------|
|
29 |
+
| T4 (Google Colab Free Tier) | 3.3s | 8.4s | 16.0s | 26.5s |
|
30 |
+
| A100 (80 GB) | 0.51s | 1.2s | 2.2s | 3.8s |
|
31 |
+
| V100 (32 GB) | 0.8s | 1.2s | 5.5s | 7.7s |
|
32 |
+
|
33 |
These tests were run with a batch size of 1 in all cases.
|
34 |
|
35 |
For cards with a lot of capacity, such as A100, performance increases significantly when generating multiple images at once, which is usually the case for production workloads.
|