-
Notifications
You must be signed in to change notification settings - Fork 196
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #714 from bghira/main
fixed flux training
- Loading branch information
Showing
21 changed files
with
1,692 additions
and
1,119 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,59 @@ | ||
# Midjourney v6 520k | ||
|
||
## Details | ||
|
||
- **Hub link**: [terminusresearch/midjourney-v6-520k-raw](https://huggingface.co/datasets/terminusresearch/midjourney-v6-520k-raw) | ||
- **Description**: ~520,000 high quality outputs where any Japanese user prompts have been re-captioned with GPT-3.5-Turbo. | ||
- **Caption format(s)**: Parquet | ||
|
||
## Required storage | ||
|
||
This dataset contains all image data, and as such, it will be difficult to extract without adequate disk space. **Ensure you have at least 1.5TB of disk space available to extract it.** | ||
|
||
T5-XXL text embeds for this model will consume ~520GB even with `--compress_disk_cache` enabled. | ||
The VAE embeds will consume just under 80 to 100GB of space, depending on the model being trained and the resolution of the embeds. | ||
|
||
|
||
## Download | ||
|
||
```bash | ||
huggingface-cli download --repo-type=dataset terminusresearch/midjourney-v6-520k-raw --local-dir=midjourney-v6-520k-raw | ||
``` | ||
|
||
This will simultaneously download the chunked tar segments from Hugging Face Hub. | ||
|
||
## Extract | ||
|
||
```bash | ||
cd midjourney-v6-520k-raw | ||
cat *.tar | tar x | ||
``` | ||
|
||
This will create a folder containing all of the samples inside the current directory. | ||
|
||
## Dataloader configuration example | ||
|
||
```json | ||
{ | ||
"id": "midjourney-v6-520k-raw", | ||
"type": "local", | ||
"cache_dir_vae": "cache/vae-mj-520k/", | ||
"crop": true, | ||
"crop_aspect": "square", | ||
"resolution": 1.0, | ||
"maximum_image_size": 1.0, | ||
"minimum_image_size": 0.75, | ||
"target_downsample_size": 1.00, | ||
"resolution_type": "area", | ||
"caption_strategy": "parquet", | ||
"metadata_backend": "parquet", | ||
"parquet": { | ||
"path": "/path/to/midjourney-v6-520k-raw/train.parquet", | ||
"caption_column": "gpt_caption", | ||
"filename_column": "id", | ||
"width_column": "width", | ||
"height_column": "height", | ||
"identifier_includes_extension": false | ||
} | ||
} | ||
``` |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,58 @@ | ||
# Niji v6 520k | ||
|
||
## Details | ||
|
||
- **Hub link**: [terminusresearch/nijijourney-v6-520k-raw](https://huggingface.co/datasets/terminusresearch/nijijourney-v6-520k-raw) | ||
- **Description**: ~520,000 high quality outputs where any Japanese user prompts have been re-captioned with GPT-3.5-Turbo. | ||
- **Caption format(s)**: Parquet | ||
|
||
## Required storage | ||
|
||
This dataset contains all image data, and as such, it will be difficult to extract without adequate disk space. **Ensure you have at least 1.5TB of disk space available to extract it.** | ||
|
||
T5-XXL text embeds for this model will consume ~520GB even with `--compress_disk_cache` enabled. | ||
The VAE embeds will consume just under 80 to 100GB of space, depending on the model being trained and the resolution of the embeds. | ||
|
||
## Download | ||
|
||
```bash | ||
huggingface-cli download --repo-type=dataset terminusresearch/nijijourney-v6-520k-raw --local-dir=nijijourney-v6-520k-raw | ||
``` | ||
|
||
This will simultaneously download the chunked tar segments from Hugging Face Hub. | ||
|
||
## Extract | ||
|
||
```bash | ||
cd nijijourney-v6-520k-raw | ||
cat *.tar | tar x | ||
``` | ||
|
||
This will create a folder containing all of the samples inside the current directory. | ||
|
||
## Dataloader configuration example | ||
|
||
```json | ||
{ | ||
"id": "nijijourney-v6-520k-raw", | ||
"type": "local", | ||
"cache_dir_vae": "cache/vae-nj-520k/", | ||
"crop": true, | ||
"crop_aspect": "square", | ||
"resolution": 1.0, | ||
"maximum_image_size": 1.0, | ||
"minimum_image_size": 0.75, | ||
"target_downsample_size": 1.00, | ||
"resolution_type": "area", | ||
"caption_strategy": "parquet", | ||
"metadata_backend": "parquet", | ||
"parquet": { | ||
"path": "/path/to/nijijourney-v6-520k-raw/train.parquet", | ||
"caption_column": "gpt_caption", | ||
"filename_column": "id", | ||
"width_column": "width", | ||
"height_column": "height", | ||
"identifier_includes_extension": false | ||
} | ||
} | ||
``` |
Oops, something went wrong.