Isotr0py
eb68892ab1
add lr_scheduler_type etc
2023-03-09 16:51:22 +08:00
Kohya S
5602e0e5fc
change dataset config option to dataset_config
2023-03-02 21:51:58 +09:00
Kohya S
ed19a92bbe
fix typos
2023-03-01 21:01:10 +09:00
fur0ut0
8abb8645ae
add detail dataset config feature by extra config file ( #227 )
...
* add config file schema
* change config file specification
* refactor config utility
* unify batch_size to train_batch_size
* fix indent size
* use batch_size instead of train_batch_size
* make cache_latents configurable on subset
* rename options
* bucket_repo_range
* shuffle_keep_tokens
* update readme
* revert to min_bucket_reso & max_bucket_reso
* use subset structure in dataset
* format import lines
* split mode specific options
* use only valid subset
* change valid subsets name
* manage multiple datasets by dataset group
* update config file sanitizer
* prune redundant validation
* add comments
* update type annotation
* rename json_file_name to metadata_file
* ignore when image dir is invalid
* fix tag shuffle and dropout
* ignore duplicated subset
* add method to check latent cachability
* fix format
* fix bug
* update caption dropout default values
* update annotation
* fix bug
* add option to enable bucket shuffle across dataset
* update blueprint generate function
* use blueprint generator for dataset initialization
* delete duplicated function
* update config readme
* delete debug print
* print dataset and subset info as info
* enable bucket_shuffle_across_dataset option
* update config readme for clarification
* compensate quotes for string option example
* fix bug of bad usage of join
* conserve trained metadata backward compatibility
* enable shuffle in data loader by default
* delete resolved TODO
* add comment for image data handling
* fix reference bug
* fix undefined variable bug
* prevent raise overwriting
* assert image_dir and metadata_file validity
* add debug message for ignoring subset
* fix inconsistent import statement
* loosen too strict validation on float value
* sanitize argument parser separately
* make image_dir optional for fine tuning dataset
* fix import
* fix trailing characters in print
* parse flexible dataset config deterministically
* use relative import
* print supplementary message for parsing error
* add note about different methods
* add note of benefit of separate dataset
* add error example
* add note for english readme plan
---------
Co-authored-by: Kohya S <52813779+kohya-ss@users.noreply.github.com >
2023-03-01 20:58:08 +09:00
Kohya S
dd523c94ff
sample images in training (not fully tested)
2023-02-27 17:48:32 +09:00
Kohya S
9ab964d0b8
Add Adafactor optimzier
2023-02-22 21:09:47 +09:00
Kohya S
663aad2b0d
refactor get_scheduler etc.
2023-02-20 22:47:43 +09:00
mgz-dev
b29c5a750c
expand optimizer options and refactor
...
Refactor code to make it easier to add new optimizers, and support alternate optimizer parameters
-move redundant code to train_util for initializing optimizers
- add SGD Nesterov optimizers as option (since they are already available)
- add new parameters which may be helpful for tuning existing and new optimizers
2023-02-19 17:45:09 -06:00
Kohya S
048e7cd428
add lion optimizer support
2023-02-19 15:26:14 +09:00
Kohya S
43c0a69843
Add noise_offset
2023-02-14 21:15:48 +09:00
Kohya S
3a72e6f003
add tag dropout
2023-02-09 21:35:27 +09:00
Kohya S
e42b2f7aa9
conditional caption dropout (in progress)
2023-02-07 22:28:56 +09:00
forestsource
7db98baa86
Add dropout options
2023-02-07 00:01:30 +09:00
Kohya S
ea2dfd09ef
update bucketing features
2023-02-05 21:37:46 +09:00
hitomi
26a81d075c
add --persistent_data_loader_workers option
2023-02-01 16:02:15 +08:00
forestsource
5e817e4343
Add save_n_epoch_ratio
2023-01-22 03:00:28 +09:00
Kohya S
aa40cb9345
Add train epochs and max workers option to train
2023-01-15 13:07:47 +09:00
Kohya S
6b62c44022
fix errors in fine tuning
2023-01-08 21:40:40 +09:00
Kohya S
82e585cf01
Fix full_fp16 and clip_skip==2 is not working
2023-01-08 18:49:34 +09:00
Kohya S
f56988b252
unify dataset and save functions
2023-01-05 08:10:22 +09:00
Kohya S
4c35006731
split common function from train_network to util
2023-01-03 20:22:25 +09:00
Kohya S
6b522b34c1
move code for xformers to train_util
2023-01-02 16:08:21 +09:00
Yuta Hayashibe
0f20453997
Fix typos
2022-12-31 14:39:36 +09:00
bmaltais
c5cca931ab
Proposed file structure rework and required file changes
2022-12-18 15:24:46 -05:00
Kohya S
7a04196e66
Add scripts.
2022-12-18 14:55:34 +09:00