ostris commited on
Commit
5dd0b1b
·
verified ·
1 Parent(s): 5061fd6

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +7 -6
README.md CHANGED
@@ -10,7 +10,7 @@ pipeline_tag: text-to-image
10
 
11
  ## Description
12
 
13
- Flex.1 alpha is a pre-trained base 8 billion parameter rectified flow transformer capable of generating images from text descriptions. It has a similar architecture to FLUX.1-dev, but with fewer double transformer blocks (8 vs 19). It began as a finetune of FLUX.1-schnell which allows us to retain the Apache 2.0 licensing. A guidance embedder has been trained for it so that it no longer requires CFG to generate images.
14
 
15
  ## Model Specs
16
 
@@ -19,6 +19,7 @@ Flex.1 alpha is a pre-trained base 8 billion parameter rectified flow transforme
19
  - True CFG capable
20
  - Fine tunable
21
  - OSI compliant license (Apache 2.0)
 
22
 
23
  ## Support Needed
24
 
@@ -28,9 +29,9 @@ I am just a solo Machine Learning Engineer doing this in my free time with my ow
28
 
29
  ## Usage
30
 
31
- The model can be used almost identically to FLUX.1-dev and will work out of the box with most inference engines that support that. (Diffusers, ComfyUI etc.)
32
 
33
- For ComfyUI, there is an all in one file called `Flex.1-alpha.safetensors`. Put this in your checkpoints folder and use like you would FLUX.1-dev.
34
 
35
  More detailed instructions coming soon.
36
 
@@ -38,13 +39,13 @@ More detailed instructions coming soon.
38
 
39
  <img src="https://i0.wp.com/ostris.com/wp-content/uploads/2025/01/openflux_is_now_flex1.jpg?resize=1024%2C328&amp;ssl=1" style="max-width: 100%; height: auto;">
40
 
41
- Flex.1 started as the FLUX.1-schnell-training-adapter to make training LoRAs on FLUX.1-schnell possible. The original goal was to train a LoRA that can be activated during training to allow for fine tuning on the step compressed model. I merged this adapter into FLUX.1-schnell and continued to train it on images generated by the FLUX.1-schnell model to further break down the compression, without injecting any new data, with the goal of making a stand-alone base model. This became OpenFLUX.1, which was continuously trained for months, resulting in 10 version releases. After the final release of OpenFLUX.1, I began training the model on new data and began experimenting with pruning. I ended up with pruned versions of OpenFLUX.1 that were 7B, and 4B parameters (unreleased). Around this time, flux.1-lite-8B-alpha was released and produced very good results. I decided to follow their pruning strategy and ended up with a 8B parameter version. I continued to train the model for months, adding new datasets and doing various experimental training tricks to improve the quality of the model.
42
 
43
- At this point, the model still required CFG in order to generate images. I decided the model needed a guidance embedder similar to FLUX.1-dev, but I wanted it to be bypassable to make the model more flexible and trainable so I trained a new guidance embedder for the model independently of the model weights so that it behaves like an optional adapter leaving the model capable of being trained and inferenced without it.
44
 
45
  ## Fine Tuning
46
 
47
- Flex.1 is designed to be fine tunable. It will finetune very similar to FLUX.1 dev, with the exception of the guidance embedder. With FLUX.1 dev, it is best to fine tune with a guidance of 1. However, With Flex.1, it is best to fine tune with the guidance embedder completely bypassed.
48
 
49
  Day 1 LoRA training support is in [AI-Toolkit](https://github.com/ostris/ai-toolkit). You can use the [example config](https://github.com/ostris/ai-toolkit/blob/main/config/examples/train_lora_flex_24gb.yaml) to get started.
50
 
 
10
 
11
  ## Description
12
 
13
+ Flex.1 alpha is a pre-trained base 8 billion parameter rectified flow transformer capable of generating images from text descriptions. It has a similar architecture to [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev), but with fewer double transformer blocks (8 vs 19). It began as a finetune of [FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) which allows the model to retain the Apache 2.0 license. A guidance embedder has been trained for it so that it no longer requires CFG to generate images.
14
 
15
  ## Model Specs
16
 
 
19
  - True CFG capable
20
  - Fine tunable
21
  - OSI compliant license (Apache 2.0)
22
+ - 512 token length input
23
 
24
  ## Support Needed
25
 
 
29
 
30
  ## Usage
31
 
32
+ The model can be used almost identically to [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev) and will work out of the box with most inference engines that support that. (Diffusers, ComfyUI etc.)
33
 
34
+ For ComfyUI, there is an all in one file called `Flex.1-alpha.safetensors`. Put this in your checkpoints folder and use like you would [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev).
35
 
36
  More detailed instructions coming soon.
37
 
 
39
 
40
  <img src="https://i0.wp.com/ostris.com/wp-content/uploads/2025/01/openflux_is_now_flex1.jpg?resize=1024%2C328&amp;ssl=1" style="max-width: 100%; height: auto;">
41
 
42
+ Flex.1 started as the [FLUX.1-schnell-training-adapter](https://huggingface.co/ostris/FLUX.1-schnell-training-adapter) to make training LoRAs on [FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) possible. The original goal was to train a LoRA that can be activated during training to allow for fine tuning on the step compressed model. I merged this adapter into [FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) and continued to train it on images generated by the [FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) model to further break down the compression, without injecting any new data, with the goal of making a stand-alone base model. This became [OpenFLUX.1](https://huggingface.co/ostris/OpenFLUX.1), which was continuously trained for months, resulting in 10 version releases. After the final release of [OpenFLUX.1](https://huggingface.co/ostris/OpenFLUX.1), I began training the model on new data and began experimenting with pruning. I ended up with pruned versions of [OpenFLUX.1](https://huggingface.co/ostris/OpenFLUX.1) that were 7B, and 4B parameters (unreleased). Around this time, [flux.1-lite-8B-alpha](https://huggingface.co/Freepik/flux.1-lite-8B-alpha) was released and produced very good results. I decided to follow their pruning strategy and ended up with a 8B parameter version. I continued to train the model, adding new datasets and doing various experimental training tricks to improve the quality of the model.
43
 
44
+ At this point, the model still required CFG in order to generate images. I decided the model needed a guidance embedder similar to [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev), but I wanted it to be bypassable to make the model more flexible and trainable so I trained a new guidance embedder for the model independently of the model weights so that it behaves like an optional adapter leaving the model capable of being trained and inferenced without it.
45
 
46
  ## Fine Tuning
47
 
48
+ Flex.1 is designed to be fine tunable. It will finetune very similar to [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev), with the exception of the guidance embedder. With [FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev), it is best to fine tune with a guidance of 1. However, With Flex.1, it is best to fine tune with the guidance embedder completely bypassed.
49
 
50
  Day 1 LoRA training support is in [AI-Toolkit](https://github.com/ostris/ai-toolkit). You can use the [example config](https://github.com/ostris/ai-toolkit/blob/main/config/examples/train_lora_flex_24gb.yaml) to get started.
51