Kohya S
6d2d8dfd2f
add zero_terminal_snr option
2023-07-18 23:17:23 +09:00
Kohya S
0ec7166098
make crop top/left same as stabilityai's prep
2023-07-18 21:39:36 +09:00
Kohya S
41d195715d
fix scheduler steps with gradient accumulation
2023-07-16 15:56:29 +09:00
Kohya S
516f64f4d9
add caching to disk for text encoder outputs
2023-07-16 14:53:47 +09:00
Kohya S
94c151aea3
refactor caching latents (flip in same npz, etc)
2023-07-15 18:28:33 +09:00
Kohya S
81fa54837f
fix sampling in multi GPU training
2023-07-15 11:21:14 +09:00
Kohya S
814996b14f
fix NaN in sampling image
2023-07-11 23:18:35 +09:00
Kohya S
b6e328ea8f
don't hold latent on memory for finetuning dataset
2023-07-10 08:46:15 +09:00
Kohya S
3579b4570f
Merge pull request #628 from KohakuBlueleaf/full_bf16
...
Full bf16 support
2023-07-09 14:22:44 +09:00
Kohaku-Blueleaf
d974959738
Update train_util.py for full_bf16 support
2023-07-09 12:47:26 +08:00
ddPn08
d599394f60
support avif
2023-07-08 15:47:56 +09:00
Kohya S
cc3d40ca44
support sdxl in prepare scipt
2023-07-07 21:16:41 +09:00
Kohya S
3060eb5baf
remove debug print
2023-07-05 21:44:46 +09:00
Kohya S
2febbfe4b0
add error message for old npz
2023-07-03 20:58:35 +09:00
Kohya S
ea182461d3
add min/max_timestep
2023-07-03 20:44:42 +09:00
Kohya S
64cf922841
add feature to sample images during sdxl training
2023-07-02 16:42:19 +09:00
Kohya S
d395bc0647
fix max_token_length not works for sdxl
2023-06-29 13:02:19 +09:00
Kohya S
2c461e4ad3
Add no_half_vae for SDXL training, add nan check
2023-06-26 20:38:09 +09:00
Kohya S
747af145ed
add sdxl fine-tuning and LoRA
2023-06-26 08:07:24 +09:00
Kohya S
9e9df2b501
update dataset to return size, refactor ctrlnet ds
2023-06-24 17:56:02 +09:00
Kohya S
92e50133f8
Merge branch 'original-u-net' into dev
2023-06-17 21:57:08 +09:00
Kohya S
19dfa24abb
Merge branch 'main' into original-u-net
2023-06-16 20:59:34 +09:00
Kohya S
5845de7d7c
common lr checking for dadaptation and prodigy
2023-06-15 21:47:37 +09:00
青龍聖者@bdsqlsz
e97d67a681
Support for Prodigy(Dadapt variety for Dylora) ( #585 )
...
* Update train_util.py for DAdaptLion
* Update train_README-zh.md for dadaptlion
* Update train_README-ja.md for DAdaptLion
* add DAdatpt V3
* Alignment
* Update train_util.py for experimental
* Update train_util.py V3
* Update train_README-zh.md
* Update train_README-ja.md
* Update train_util.py fix
* Update train_util.py
* support Prodigy
* add lower
2023-06-15 21:12:53 +09:00
Kohya S
9806b00f74
add arbitrary dataset feature to each script
2023-06-15 20:39:39 +09:00
Kohya S
449ad7502c
use original unet for HF models, don't download TE
2023-06-14 22:26:05 +09:00
Kohya S
9aee793078
support arbitrary dataset for train_network.py
2023-06-14 12:49:12 +09:00
ykume
9e1683cf2b
support sdpa
2023-06-11 21:26:15 +09:00
ykume
0315611b11
remove workaround for accelerator=0.15, fix XTI
2023-06-11 18:32:14 +09:00
ykume
035dd3a900
fix mem_eff_attn does not work
2023-06-11 17:08:21 +09:00
ykume
4e25c8f78e
fix to work with Diffusers 0.17.0
2023-06-11 16:57:17 +09:00
Kohya S
c0a7df9ee1
fix eps value, enable xformers, etc.
2023-06-03 21:29:27 +09:00
Kohya S
ec2efe52e4
scale v-pred loss like noise pred
2023-06-03 10:52:22 +09:00
ddPn08
1e3daa247b
fix bucketing
2023-06-01 21:58:45 +09:00
ddPn08
3bd00b88c2
support for controlnet in sample output
2023-06-01 20:48:30 +09:00
ddPn08
62d00b4520
add controlnet training
2023-06-01 20:48:25 +09:00
ddPn08
c8d209d36c
update diffusers to 1.16 | train_network
2023-06-01 20:39:26 +09:00
AI-Casanova
9c7237157d
Dropout and Max Norm Regularization for LoRA training ( #545 )
...
* Instantiate max_norm
* minor
* Move to end of step
* argparse
* metadata
* phrasing
* Sqrt ratio and logging
* fix logging
* Dropout test
* Dropout Args
* Dropout changed to affect LoRA only
---------
Co-authored-by: Kohya S <52813779+kohya-ss@users.noreply.github.com >
2023-06-01 14:58:38 +09:00
Kohya S
3a06968332
warn and continue if huggingface uploading failed
2023-05-31 20:48:33 +09:00
Kohya S
990ceddd14
show warning if no caption and no class token
2023-05-30 22:53:50 +09:00
Kohya S
2429ac73b2
Merge pull request #533 from TingTingin/main
...
Added warning on training without captions
2023-05-29 08:37:33 +09:00
TingTingin
db756e9a34
Update train_util.py
...
I removed the sleep since it triggers per subset and if someone had a lot of subsets it would trigger multiple times
2023-05-26 08:08:34 -04:00
青龍聖者@bdsqlsz
5cdf4e34a1
support for dadapaption V3 ( #530 )
...
* Update train_util.py for DAdaptLion
* Update train_README-zh.md for dadaptlion
* Update train_README-ja.md for DAdaptLion
* add DAdatpt V3
* Alignment
* Update train_util.py for experimental
* Update train_util.py V3
* Update train_README-zh.md
* Update train_README-ja.md
* Update train_util.py fix
* Update train_util.py
---------
Co-authored-by: Kohya S <52813779+kohya-ss@users.noreply.github.com >
2023-05-25 21:52:36 +09:00
TingTingin
061e157191
Update train_util.py
2023-05-23 02:02:39 -04:00
TingTingin
d859a3a925
Update train_util.py
...
fix mistake
2023-05-23 02:00:33 -04:00
TingTingin
5a1a14f9fc
Update train_util.py
...
Added feature to add "." if missing in caption_extension
Added warning on training without captions
2023-05-23 01:57:35 -04:00
Kohya S
02bb8e0ac3
use xformers in VAE in gen script
2023-05-21 12:59:01 +09:00
Kohya S
bc909e8359
Merge pull request #521 from akshaal/fix/save_state
...
fix: don't save state if no --save-state arg given
2023-05-21 08:48:48 +09:00
Evgeny Chukreev
0c942106bf
fix: don't save state if no --save-state arg given
2023-05-18 20:09:06 +02:00
Fair
c0c4d4ddc6
new line with print "generating sample images"
2023-05-17 10:59:06 +08:00