TobiGeth commited on
Commit
0ab1be3
·
verified ·
1 Parent(s): fe0a0e1

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. config.yaml +39 -3
  2. lora.safetensors +2 -2
config.yaml CHANGED
@@ -12,11 +12,47 @@ config:
12
  linear_alpha: 128
13
  network_kwargs:
14
  only_if_contains:
 
 
 
 
 
 
 
15
  - transformer.single_transformer_blocks.7.proj_out
 
 
 
 
 
 
 
 
 
 
 
 
16
  - transformer.single_transformer_blocks.20.proj_out
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  save:
18
  dtype: float16
19
- save_every: 2001
20
  max_step_saves_to_keep: 1
21
  datasets:
22
  - folder_path: input_images
@@ -31,7 +67,7 @@ config:
31
  - 1024
32
  train:
33
  batch_size: 1
34
- steps: 2000
35
  gradient_accumulation_steps: 1
36
  train_unet: true
37
  train_text_encoder: false
@@ -50,7 +86,7 @@ config:
50
  quantize: false
51
  sample:
52
  sampler: flowmatch
53
- sample_every: 2001
54
  width: 1024
55
  height: 1024
56
  prompts: []
 
12
  linear_alpha: 128
13
  network_kwargs:
14
  only_if_contains:
15
+ - transformer.single_transformer_blocks.0.proj_out
16
+ - transformer.single_transformer_blocks.1.proj_out
17
+ - transformer.single_transformer_blocks.2.proj_out
18
+ - transformer.single_transformer_blocks.3.proj_out
19
+ - transformer.single_transformer_blocks.4.proj_out
20
+ - transformer.single_transformer_blocks.5.proj_out
21
+ - transformer.single_transformer_blocks.6.proj_out
22
  - transformer.single_transformer_blocks.7.proj_out
23
+ - transformer.single_transformer_blocks.8.proj_out
24
+ - transformer.single_transformer_blocks.9.proj_out
25
+ - transformer.single_transformer_blocks.10.proj_out
26
+ - transformer.single_transformer_blocks.11.proj_out
27
+ - transformer.single_transformer_blocks.12.proj_out
28
+ - transformer.single_transformer_blocks.13.proj_out
29
+ - transformer.single_transformer_blocks.14.proj_out
30
+ - transformer.single_transformer_blocks.15.proj_out
31
+ - transformer.single_transformer_blocks.16.proj_out
32
+ - transformer.single_transformer_blocks.17.proj_out
33
+ - transformer.single_transformer_blocks.18.proj_out
34
+ - transformer.single_transformer_blocks.19.proj_out
35
  - transformer.single_transformer_blocks.20.proj_out
36
+ - transformer.single_transformer_blocks.21.proj_out
37
+ - transformer.single_transformer_blocks.22.proj_out
38
+ - transformer.single_transformer_blocks.23.proj_out
39
+ - transformer.single_transformer_blocks.24.proj_out
40
+ - transformer.single_transformer_blocks.25.proj_out
41
+ - transformer.single_transformer_blocks.26.proj_out
42
+ - transformer.single_transformer_blocks.27.proj_out
43
+ - transformer.single_transformer_blocks.28.proj_out
44
+ - transformer.single_transformer_blocks.29.proj_out
45
+ - transformer.single_transformer_blocks.30.proj_out
46
+ - transformer.single_transformer_blocks.31.proj_out
47
+ - transformer.single_transformer_blocks.32.proj_out
48
+ - transformer.single_transformer_blocks.33.proj_out
49
+ - transformer.single_transformer_blocks.34.proj_out
50
+ - transformer.single_transformer_blocks.35.proj_out
51
+ - transformer.single_transformer_blocks.36.proj_out
52
+ - transformer.single_transformer_blocks.37.proj_out
53
  save:
54
  dtype: float16
55
+ save_every: 1201
56
  max_step_saves_to_keep: 1
57
  datasets:
58
  - folder_path: input_images
 
67
  - 1024
68
  train:
69
  batch_size: 1
70
+ steps: 1200
71
  gradient_accumulation_steps: 1
72
  train_unet: true
73
  train_text_encoder: false
 
86
  quantize: false
87
  sample:
88
  sampler: flowmatch
89
+ sample_every: 1201
90
  width: 1024
91
  height: 1024
92
  prompts: []
lora.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c25dece5488bc630f22ca587ca619eb804699890ec5d6be657353c480e8c96f0
3
- size 9438224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:793a018fff3c3161c0600a563c91486447765648abec4af2122b39b9784a7165
3
+ size 179317416