Upload folder using huggingface_hub
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ pipeline_tag: text-to-image
|
|
10 |
|
11 |
## Description
|
12 |
|
13 |
-
Flex.1 alpha is a pre-trained base 8 billion parameter rectified flow transformer capable of generating images from text descriptions. It has a similar architecture to FLUX.1-dev, but with fewer double transformer blocks (8 vs 19). It began as a finetune of FLUX.1-schnell which allows
|
14 |
|
15 |
## Model Specs
|
16 |
|
@@ -19,6 +19,7 @@ Flex.1 alpha is a pre-trained base 8 billion parameter rectified flow transforme
|
|
19 |
- True CFG capable
|
20 |
- Fine tunable
|
21 |
- OSI compliant license (Apache 2.0)
|
|
|
22 |
|
23 |
## Support Needed
|
24 |
|
@@ -28,9 +29,9 @@ I am just a solo Machine Learning Engineer doing this in my free time with my ow
|
|
28 |
|
29 |
## Usage
|
30 |
|
31 |
-
The model can be used almost identically to FLUX.1-dev and will work out of the box with most inference engines that support that. (Diffusers, ComfyUI etc.)
|
32 |
|
33 |
-
For ComfyUI, there is an all in one file called `Flex.1-alpha.safetensors`. Put this in your checkpoints folder and use like you would FLUX.1-dev.
|
34 |
|
35 |
More detailed instructions coming soon.
|
36 |
|
@@ -38,13 +39,13 @@ More detailed instructions coming soon.
|
|
38 |
|
39 |
<img src="https://i0.wp.com/ostris.com/wp-content/uploads/2025/01/openflux_is_now_flex1.jpg?resize=1024%2C328&ssl=1" style="max-width: 100%; height: auto;">
|
40 |
|
41 |
-
Flex.1 started as the FLUX.1-schnell-training-adapter to make training LoRAs on FLUX.1-schnell possible. The original goal was to train a LoRA that can be activated during training to allow for fine tuning on the step compressed model. I merged this adapter into FLUX.1-schnell and continued to train it on images generated by the FLUX.1-schnell model to further break down the compression, without injecting any new data, with the goal of making a stand-alone base model. This became OpenFLUX.1, which was continuously trained for months, resulting in 10 version releases. After the final release of OpenFLUX.1, I began training the model on new data and began experimenting with pruning. I ended up with pruned versions of OpenFLUX.1 that were 7B, and 4B parameters (unreleased). Around this time, flux.1-lite-8B-alpha was released and produced very good results. I decided to follow their pruning strategy and ended up with a 8B parameter version. I continued to train the model
|
42 |
|
43 |
-
At this point, the model still required CFG in order to generate images. I decided the model needed a guidance embedder similar to FLUX.1-dev, but I wanted it to be bypassable to make the model more flexible and trainable so I trained a new guidance embedder for the model independently of the model weights so that it behaves like an optional adapter leaving the model capable of being trained and inferenced without it.
|
44 |
|
45 |
## Fine Tuning
|
46 |
|
47 |
-
Flex.1 is designed to be fine tunable. It will finetune very similar to FLUX.1
|
48 |
|
49 |
Day 1 LoRA training support is in [AI-Toolkit](https://github.com/ostris/ai-toolkit). You can use the [example config](https://github.com/ostris/ai-toolkit/blob/main/config/examples/train_lora_flex_24gb.yaml) to get started.
|
50 |
|
|
|
10 |
|
11 |
## Description
|
12 |
|
13 |
+
Flex.1 alpha is a pre-trained base 8 billion parameter rectified flow transformer capable of generating images from text descriptions. It has a similar architecture to [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev), but with fewer double transformer blocks (8 vs 19). It began as a finetune of [FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) which allows the model to retain the Apache 2.0 license. A guidance embedder has been trained for it so that it no longer requires CFG to generate images.
|
14 |
|
15 |
## Model Specs
|
16 |
|
|
|
19 |
- True CFG capable
|
20 |
- Fine tunable
|
21 |
- OSI compliant license (Apache 2.0)
|
22 |
+
- 512 token length input
|
23 |
|
24 |
## Support Needed
|
25 |
|
|
|
29 |
|
30 |
## Usage
|
31 |
|
32 |
+
The model can be used almost identically to [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev) and will work out of the box with most inference engines that support that. (Diffusers, ComfyUI etc.)
|
33 |
|
34 |
+
For ComfyUI, there is an all in one file called `Flex.1-alpha.safetensors`. Put this in your checkpoints folder and use like you would [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev).
|
35 |
|
36 |
More detailed instructions coming soon.
|
37 |
|
|
|
39 |
|
40 |
<img src="https://i0.wp.com/ostris.com/wp-content/uploads/2025/01/openflux_is_now_flex1.jpg?resize=1024%2C328&ssl=1" style="max-width: 100%; height: auto;">
|
41 |
|
42 |
+
Flex.1 started as the [FLUX.1-schnell-training-adapter](https://huggingface.co/ostris/FLUX.1-schnell-training-adapter) to make training LoRAs on [FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) possible. The original goal was to train a LoRA that can be activated during training to allow for fine tuning on the step compressed model. I merged this adapter into [FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) and continued to train it on images generated by the [FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) model to further break down the compression, without injecting any new data, with the goal of making a stand-alone base model. This became [OpenFLUX.1](https://huggingface.co/ostris/OpenFLUX.1), which was continuously trained for months, resulting in 10 version releases. After the final release of [OpenFLUX.1](https://huggingface.co/ostris/OpenFLUX.1), I began training the model on new data and began experimenting with pruning. I ended up with pruned versions of [OpenFLUX.1](https://huggingface.co/ostris/OpenFLUX.1) that were 7B, and 4B parameters (unreleased). Around this time, [flux.1-lite-8B-alpha](https://huggingface.co/Freepik/flux.1-lite-8B-alpha) was released and produced very good results. I decided to follow their pruning strategy and ended up with a 8B parameter version. I continued to train the model, adding new datasets and doing various experimental training tricks to improve the quality of the model.
|
43 |
|
44 |
+
At this point, the model still required CFG in order to generate images. I decided the model needed a guidance embedder similar to [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev), but I wanted it to be bypassable to make the model more flexible and trainable so I trained a new guidance embedder for the model independently of the model weights so that it behaves like an optional adapter leaving the model capable of being trained and inferenced without it.
|
45 |
|
46 |
## Fine Tuning
|
47 |
|
48 |
+
Flex.1 is designed to be fine tunable. It will finetune very similar to [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev), with the exception of the guidance embedder. With [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev), it is best to fine tune with a guidance of 1. However, With Flex.1, it is best to fine tune with the guidance embedder completely bypassed.
|
49 |
|
50 |
Day 1 LoRA training support is in [AI-Toolkit](https://github.com/ostris/ai-toolkit). You can use the [example config](https://github.com/ostris/ai-toolkit/blob/main/config/examples/train_lora_flex_24gb.yaml) to get started.
|
51 |
|