weijiawu commited on
Commit
634a418
·
1 Parent(s): 20a74c0
FrozenII/combined_model_base.log ADDED
@@ -0,0 +1,386 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-02-18 10:01:13,607 INFO: Namespace(concept_cfg='datasets/data_cfgs/MixofShow/multi-concept/MovieGen/FrozenII/Elsa+Olaf+Mattias+Kristoff+Anna.json', optimize_textenc_iters=1000, optimize_unet_iters=100, pretrained_models='experiments/pretrained_models/chilloutmix', save_path='experiments/composed_edlora/chilloutmix/FrozenII', suffix='base')
2
+ 2025-02-18 10:01:13,607 INFO: ------Step 1: load stable diffusion checkpoint------
3
+ 2025-02-18 10:01:18,406 INFO: ------Step 2: load new concepts checkpoints------
4
+ 2025-02-18 10:01:18,450 INFO: ------Step 3: merge token embedding------
5
+ 2025-02-18 10:01:18,680 INFO: concept <Elsa1> is bind with token_id: [49408, 49423]
6
+ 2025-02-18 10:01:18,902 INFO: concept <Elsa2> is bind with token_id: [49424, 49439]
7
+ 2025-02-18 10:01:19,126 INFO: concept <Olaf1> is bind with token_id: [49440, 49455]
8
+ 2025-02-18 10:01:19,350 INFO: concept <Olaf2> is bind with token_id: [49456, 49471]
9
+ 2025-02-18 10:01:19,575 INFO: concept <Mattias1> is bind with token_id: [49472, 49487]
10
+ 2025-02-18 10:01:19,801 INFO: concept <Mattias2> is bind with token_id: [49488, 49503]
11
+ 2025-02-18 10:01:20,028 INFO: concept <Kristoff1> is bind with token_id: [49504, 49519]
12
+ 2025-02-18 10:01:20,257 INFO: concept <Kristoff2> is bind with token_id: [49520, 49535]
13
+ 2025-02-18 10:01:20,485 INFO: concept <Anna1> is bind with token_id: [49536, 49551]
14
+ 2025-02-18 10:01:20,712 INFO: concept <Anna2> is bind with token_id: [49552, 49567]
15
+ 2025-02-18 10:01:20,712 INFO: ------Step 4: merge text encoder------
16
+ 2025-02-18 10:01:20,831 INFO: text_encoder have 48 linear layer need to optimize
17
+ 2025-02-18 10:01:20,832 INFO: add 48 hooker to text_encoder
18
+ 2025-02-18 10:01:20,859 INFO: load 48 LoRAs of text_encoder
19
+ 2025-02-18 10:01:21,132 INFO: load 48 LoRAs of text_encoder
20
+ 2025-02-18 10:01:21,372 INFO: load 48 LoRAs of text_encoder
21
+ 2025-02-18 10:01:21,613 INFO: load 48 LoRAs of text_encoder
22
+ 2025-02-18 10:01:21,850 INFO: load 48 LoRAs of text_encoder
23
+ 2025-02-18 10:01:22,103 INFO: [1/48] optimizing text_model.encoder.layers.4.self_attn.q_proj.weight
24
+ 2025-02-18 10:01:24,861 INFO: new_concept loss: 1.378319e-02
25
+ 2025-02-18 10:01:24,861 INFO: [2/48] optimizing text_model.encoder.layers.4.self_attn.k_proj.weight
26
+ 2025-02-18 10:01:27,617 INFO: new_concept loss: 1.130946e-02
27
+ 2025-02-18 10:01:27,617 INFO: [3/48] optimizing text_model.encoder.layers.10.self_attn.k_proj.weight
28
+ 2025-02-18 10:01:30,514 INFO: new_concept loss: 6.835580e-03
29
+ 2025-02-18 10:01:30,514 INFO: [4/48] optimizing text_model.encoder.layers.7.self_attn.k_proj.weight
30
+ 2025-02-18 10:01:33,060 INFO: new_concept loss: 1.043152e-02
31
+ 2025-02-18 10:01:33,060 INFO: [5/48] optimizing text_model.encoder.layers.6.self_attn.k_proj.weight
32
+ 2025-02-18 10:01:35,603 INFO: new_concept loss: 1.229318e-02
33
+ 2025-02-18 10:01:35,603 INFO: [6/48] optimizing text_model.encoder.layers.11.self_attn.k_proj.weight
34
+ 2025-02-18 10:01:38,154 INFO: new_concept loss: 2.929659e-02
35
+ 2025-02-18 10:01:38,154 INFO: [7/48] optimizing text_model.encoder.layers.2.self_attn.q_proj.weight
36
+ 2025-02-18 10:01:40,728 INFO: new_concept loss: 1.220211e-02
37
+ 2025-02-18 10:01:40,728 INFO: [8/48] optimizing text_model.encoder.layers.11.self_attn.v_proj.weight
38
+ 2025-02-18 10:01:43,589 INFO: new_concept loss: 1.071226e-02
39
+ 2025-02-18 10:01:43,589 INFO: [9/48] optimizing text_model.encoder.layers.6.self_attn.out_proj.weight
40
+ 2025-02-18 10:01:46,495 INFO: new_concept loss: 2.268766e-03
41
+ 2025-02-18 10:01:46,495 INFO: [10/48] optimizing text_model.encoder.layers.5.self_attn.out_proj.weight
42
+ 2025-02-18 10:01:48,888 INFO: new_concept loss: 2.712738e-03
43
+ 2025-02-18 10:01:48,889 INFO: [11/48] optimizing text_model.encoder.layers.7.self_attn.out_proj.weight
44
+ 2025-02-18 10:01:51,311 INFO: new_concept loss: 2.453873e-03
45
+ 2025-02-18 10:01:51,311 INFO: [12/48] optimizing text_model.encoder.layers.9.self_attn.k_proj.weight
46
+ 2025-02-18 10:01:54,139 INFO: new_concept loss: 1.146064e-02
47
+ 2025-02-18 10:01:54,139 INFO: [13/48] optimizing text_model.encoder.layers.2.self_attn.k_proj.weight
48
+ 2025-02-18 10:01:56,883 INFO: new_concept loss: 1.995232e-02
49
+ 2025-02-18 10:01:56,884 INFO: [14/48] optimizing text_model.encoder.layers.3.self_attn.k_proj.weight
50
+ 2025-02-18 10:01:59,640 INFO: new_concept loss: 2.060306e-02
51
+ 2025-02-18 10:01:59,641 INFO: [15/48] optimizing text_model.encoder.layers.4.self_attn.v_proj.weight
52
+ 2025-02-18 10:02:02,303 INFO: new_concept loss: 1.727206e-02
53
+ 2025-02-18 10:02:02,303 INFO: [16/48] optimizing text_model.encoder.layers.11.self_attn.out_proj.weight
54
+ 2025-02-18 10:02:05,097 INFO: new_concept loss: 4.675179e-03
55
+ 2025-02-18 10:02:05,098 INFO: [17/48] optimizing text_model.encoder.layers.3.self_attn.v_proj.weight
56
+ 2025-02-18 10:02:07,841 INFO: new_concept loss: 2.080345e-02
57
+ 2025-02-18 10:02:07,841 INFO: [18/48] optimizing text_model.encoder.layers.9.self_attn.out_proj.weight
58
+ 2025-02-18 10:02:10,796 INFO: new_concept loss: 3.901713e-03
59
+ 2025-02-18 10:02:10,796 INFO: [19/48] optimizing text_model.encoder.layers.9.self_attn.q_proj.weight
60
+ 2025-02-18 10:02:13,761 INFO: new_concept loss: 1.188236e-02
61
+ 2025-02-18 10:02:13,762 INFO: [20/48] optimizing text_model.encoder.layers.0.self_attn.k_proj.weight
62
+ 2025-02-18 10:02:14,587 INFO: new_concept loss: 3.598483e-02
63
+ 2025-02-18 10:02:14,587 INFO: [21/48] optimizing text_model.encoder.layers.8.self_attn.q_proj.weight
64
+ 2025-02-18 10:02:17,432 INFO: new_concept loss: 9.705781e-03
65
+ 2025-02-18 10:02:17,432 INFO: [22/48] optimizing text_model.encoder.layers.3.self_attn.q_proj.weight
66
+ 2025-02-18 10:02:20,256 INFO: new_concept loss: 1.718284e-02
67
+ 2025-02-18 10:02:20,256 INFO: [23/48] optimizing text_model.encoder.layers.5.self_attn.v_proj.weight
68
+ 2025-02-18 10:02:23,243 INFO: new_concept loss: 1.262343e-02
69
+ 2025-02-18 10:02:23,243 INFO: [24/48] optimizing text_model.encoder.layers.1.self_attn.out_proj.weight
70
+ 2025-02-18 10:02:26,025 INFO: new_concept loss: 2.885660e-03
71
+ 2025-02-18 10:02:26,025 INFO: [25/48] optimizing text_model.encoder.layers.8.self_attn.v_proj.weight
72
+ 2025-02-18 10:02:28,986 INFO: new_concept loss: 1.011055e-02
73
+ 2025-02-18 10:02:28,986 INFO: [26/48] optimizing text_model.encoder.layers.1.self_attn.v_proj.weight
74
+ 2025-02-18 10:02:31,798 INFO: new_concept loss: 2.109258e-02
75
+ 2025-02-18 10:02:31,798 INFO: [27/48] optimizing text_model.encoder.layers.8.self_attn.out_proj.weight
76
+ 2025-02-18 10:02:34,334 INFO: new_concept loss: 2.203831e-03
77
+ 2025-02-18 10:02:34,334 INFO: [28/48] optimizing text_model.encoder.layers.3.self_attn.out_proj.weight
78
+ 2025-02-18 10:02:37,145 INFO: new_concept loss: 1.957642e-03
79
+ 2025-02-18 10:02:37,145 INFO: [29/48] optimizing text_model.encoder.layers.7.self_attn.q_proj.weight
80
+ 2025-02-18 10:02:39,995 INFO: new_concept loss: 1.165296e-02
81
+ 2025-02-18 10:02:39,996 INFO: [30/48] optimizing text_model.encoder.layers.5.self_attn.q_proj.weight
82
+ 2025-02-18 10:02:42,915 INFO: new_concept loss: 7.787009e-03
83
+ 2025-02-18 10:02:42,916 INFO: [31/48] optimizing text_model.encoder.layers.2.self_attn.v_proj.weight
84
+ 2025-02-18 10:02:45,746 INFO: new_concept loss: 1.318311e-02
85
+ 2025-02-18 10:02:45,746 INFO: [32/48] optimizing text_model.encoder.layers.5.self_attn.k_proj.weight
86
+ 2025-02-18 10:02:48,589 INFO: new_concept loss: 9.512860e-03
87
+ 2025-02-18 10:02:48,590 INFO: [33/48] optimizing text_model.encoder.layers.6.self_attn.v_proj.weight
88
+ 2025-02-18 10:02:51,572 INFO: new_concept loss: 1.058318e-02
89
+ 2025-02-18 10:02:51,572 INFO: [34/48] optimizing text_model.encoder.layers.10.self_attn.out_proj.weight
90
+ 2025-02-18 10:02:54,468 INFO: new_concept loss: 2.815726e-03
91
+ 2025-02-18 10:02:54,469 INFO: [35/48] optimizing text_model.encoder.layers.10.self_attn.q_proj.weight
92
+ 2025-02-18 10:02:57,526 INFO: new_concept loss: 8.984621e-03
93
+ 2025-02-18 10:02:57,526 INFO: [36/48] optimizing text_model.encoder.layers.7.self_attn.v_proj.weight
94
+ 2025-02-18 10:03:00,487 INFO: new_concept loss: 1.155254e-02
95
+ 2025-02-18 10:03:00,487 INFO: [37/48] optimizing text_model.encoder.layers.8.self_attn.k_proj.weight
96
+ 2025-02-18 10:03:03,208 INFO: new_concept loss: 1.002240e-02
97
+ 2025-02-18 10:03:03,208 INFO: [38/48] optimizing text_model.encoder.layers.6.self_attn.q_proj.weight
98
+ 2025-02-18 10:03:05,931 INFO: new_concept loss: 9.551291e-03
99
+ 2025-02-18 10:03:05,932 INFO: [39/48] optimizing text_model.encoder.layers.1.self_attn.k_proj.weight
100
+ 2025-02-18 10:03:08,649 INFO: new_concept loss: 2.144185e-02
101
+ 2025-02-18 10:03:08,650 INFO: [40/48] optimizing text_model.encoder.layers.0.self_attn.out_proj.weight
102
+ 2025-02-18 10:03:11,224 INFO: new_concept loss: 4.528776e-03
103
+ 2025-02-18 10:03:11,224 INFO: [41/48] optimizing text_model.encoder.layers.2.self_attn.out_proj.weight
104
+ 2025-02-18 10:03:13,646 INFO: new_concept loss: 2.965961e-03
105
+ 2025-02-18 10:03:13,646 INFO: [42/48] optimizing text_model.encoder.layers.4.self_attn.out_proj.weight
106
+ 2025-02-18 10:03:16,621 INFO: new_concept loss: 2.300192e-03
107
+ 2025-02-18 10:03:16,621 INFO: [43/48] optimizing text_model.encoder.layers.11.self_attn.q_proj.weight
108
+ 2025-02-18 10:03:19,347 INFO: new_concept loss: 1.009758e-02
109
+ 2025-02-18 10:03:19,347 INFO: [44/48] optimizing text_model.encoder.layers.9.self_attn.v_proj.weight
110
+ 2025-02-18 10:03:22,184 INFO: new_concept loss: 8.896633e-03
111
+ 2025-02-18 10:03:22,184 INFO: [45/48] optimizing text_model.encoder.layers.0.self_attn.q_proj.weight
112
+ 2025-02-18 10:03:22,510 INFO: new_concept loss: 4.487977e-02
113
+ 2025-02-18 10:03:22,510 INFO: [46/48] optimizing text_model.encoder.layers.0.self_attn.v_proj.weight
114
+ 2025-02-18 10:03:23,257 INFO: new_concept loss: 2.651430e-02
115
+ 2025-02-18 10:03:23,257 INFO: [47/48] optimizing text_model.encoder.layers.1.self_attn.q_proj.weight
116
+ 2025-02-18 10:03:26,068 INFO: new_concept loss: 1.082598e-02
117
+ 2025-02-18 10:03:26,068 INFO: [48/48] optimizing text_model.encoder.layers.10.self_attn.v_proj.weight
118
+ 2025-02-18 10:03:28,910 INFO: new_concept loss: 1.108888e-02
119
+ 2025-02-18 10:03:28,911 INFO: remove 48 hooker from text_encoder
120
+ 2025-02-18 10:03:28,947 INFO: ------Step 5: merge kv of cross-attention in unet------
121
+ 2025-02-18 10:03:28,950 INFO: Unet have 32 linear layer (related to text feature) need to optimize
122
+ 2025-02-18 10:03:30,392 INFO: [1/32] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k.weight
123
+ 2025-02-18 10:03:31,275 INFO: new_concept loss: 4.427530e-05
124
+ 2025-02-18 10:03:31,275 INFO: [2/32] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v.weight
125
+ 2025-02-18 10:03:33,597 INFO: new_concept loss: 2.170046e-05
126
+ 2025-02-18 10:03:33,597 INFO: [3/32] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k.weight
127
+ 2025-02-18 10:03:35,695 INFO: new_concept loss: 2.655753e-05
128
+ 2025-02-18 10:03:35,695 INFO: [4/32] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v.weight
129
+ 2025-02-18 10:03:37,830 INFO: new_concept loss: 2.436045e-05
130
+ 2025-02-18 10:03:37,830 INFO: [5/32] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight
131
+ 2025-02-18 10:03:38,959 INFO: new_concept loss: 2.507451e-05
132
+ 2025-02-18 10:03:38,959 INFO: [6/32] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight
133
+ 2025-02-18 10:03:40,767 INFO: new_concept loss: 1.935738e-05
134
+ 2025-02-18 10:03:40,768 INFO: [7/32] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight
135
+ 2025-02-18 10:03:42,580 INFO: new_concept loss: 6.680089e-05
136
+ 2025-02-18 10:03:42,580 INFO: [8/32] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight
137
+ 2025-02-18 10:03:44,709 INFO: new_concept loss: 1.519853e-05
138
+ 2025-02-18 10:03:44,710 INFO: [9/32] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight
139
+ 2025-02-18 10:03:46,676 INFO: new_concept loss: 3.263517e-05
140
+ 2025-02-18 10:03:46,677 INFO: [10/32] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight
141
+ 2025-02-18 10:03:50,298 INFO: new_concept loss: 1.466394e-05
142
+ 2025-02-18 10:03:50,299 INFO: [11/32] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight
143
+ 2025-02-18 10:03:53,285 INFO: new_concept loss: 4.690791e-05
144
+ 2025-02-18 10:03:53,285 INFO: [12/32] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight
145
+ 2025-02-18 10:03:55,808 INFO: new_concept loss: 2.317002e-05
146
+ 2025-02-18 10:03:55,808 INFO: [13/32] optimizing mid_block.attentions.0.transformer_blocks.0.attn2.to_k.weight
147
+ 2025-02-18 10:03:58,787 INFO: new_concept loss: 2.382438e-05
148
+ 2025-02-18 10:03:58,787 INFO: [14/32] optimizing mid_block.attentions.0.transformer_blocks.0.attn2.to_v.weight
149
+ 2025-02-18 10:04:00,413 INFO: new_concept loss: 2.085945e-05
150
+ 2025-02-18 10:04:00,413 INFO: [15/32] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight
151
+ 2025-02-18 10:04:03,919 INFO: new_concept loss: 8.338261e-05
152
+ 2025-02-18 10:04:03,919 INFO: [16/32] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight
153
+ 2025-02-18 10:04:07,465 INFO: new_concept loss: 5.506125e-04
154
+ 2025-02-18 10:04:07,465 INFO: [17/32] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight
155
+ 2025-02-18 10:04:10,213 INFO: new_concept loss: 6.924390e-05
156
+ 2025-02-18 10:04:10,214 INFO: [18/32] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight
157
+ 2025-02-18 10:04:13,788 INFO: new_concept loss: 9.049270e-05
158
+ 2025-02-18 10:04:13,789 INFO: [19/32] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k.weight
159
+ 2025-02-18 10:04:17,357 INFO: new_concept loss: 2.786910e-04
160
+ 2025-02-18 10:04:17,358 INFO: [20/32] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v.weight
161
+ 2025-02-18 10:04:21,001 INFO: new_concept loss: 1.661567e-04
162
+ 2025-02-18 10:04:21,001 INFO: [21/32] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight
163
+ 2025-02-18 10:04:23,301 INFO: new_concept loss: 1.153841e-04
164
+ 2025-02-18 10:04:23,301 INFO: [22/32] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight
165
+ 2025-02-18 10:04:25,677 INFO: new_concept loss: 1.139642e-04
166
+ 2025-02-18 10:04:25,677 INFO: [23/32] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight
167
+ 2025-02-18 10:04:27,179 INFO: new_concept loss: 2.700988e-05
168
+ 2025-02-18 10:04:27,180 INFO: [24/32] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight
169
+ 2025-02-18 10:04:29,088 INFO: new_concept loss: 1.945955e-05
170
+ 2025-02-18 10:04:29,088 INFO: [25/32] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k.weight
171
+ 2025-02-18 10:04:30,532 INFO: new_concept loss: 6.514074e-05
172
+ 2025-02-18 10:04:30,532 INFO: [26/32] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v.weight
173
+ 2025-02-18 10:04:32,974 INFO: new_concept loss: 2.861637e-05
174
+ 2025-02-18 10:04:32,975 INFO: [27/32] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k.weight
175
+ 2025-02-18 10:04:33,947 INFO: new_concept loss: 2.005374e-05
176
+ 2025-02-18 10:04:33,947 INFO: [28/32] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v.weight
177
+ 2025-02-18 10:04:35,233 INFO: new_concept loss: 5.705929e-06
178
+ 2025-02-18 10:04:35,233 INFO: [29/32] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k.weight
179
+ 2025-02-18 10:04:37,089 INFO: new_concept loss: 3.739028e-05
180
+ 2025-02-18 10:04:37,090 INFO: [30/32] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v.weight
181
+ 2025-02-18 10:04:39,365 INFO: new_concept loss: 1.504165e-05
182
+ 2025-02-18 10:04:39,365 INFO: [31/32] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k.weight
183
+ 2025-02-18 10:04:40,939 INFO: new_concept loss: 4.594039e-05
184
+ 2025-02-18 10:04:40,940 INFO: [32/32] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v.weight
185
+ 2025-02-18 10:04:42,949 INFO: new_concept loss: 4.181080e-05
186
+ 2025-02-18 10:04:42,984 INFO: ------Step 6: merge spatial attention (q in cross-attention, qkv in self-attention) in unet------
187
+ 2025-02-18 10:04:42,985 INFO: unet have 96 linear layer need to optimize
188
+ 2025-02-18 10:04:42,986 INFO: add 96 hooker to unet
189
+ 2025-02-18 10:04:43,061 INFO: load 96 LoRAs of unet
190
+ 2025-02-18 10:04:47,441 INFO: load 96 LoRAs of unet
191
+ 2025-02-18 10:04:49,403 INFO: load 96 LoRAs of unet
192
+ 2025-02-18 10:04:51,305 INFO: load 96 LoRAs of unet
193
+ 2025-02-18 10:04:53,195 INFO: load 96 LoRAs of unet
194
+ 2025-02-18 10:04:57,664 INFO: [1/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_q.weight
195
+ 2025-02-18 10:05:05,839 INFO: new_concept loss: 1.593050e-01
196
+ 2025-02-18 10:05:05,915 INFO: [2/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn2.to_q.weight
197
+ 2025-02-18 10:05:06,917 INFO: new_concept loss: 1.371175e-01
198
+ 2025-02-18 10:05:06,918 INFO: [3/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
199
+ 2025-02-18 10:05:09,711 INFO: new_concept loss: 9.162319e-02
200
+ 2025-02-18 10:05:09,731 INFO: [4/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k.weight
201
+ 2025-02-18 10:05:13,925 INFO: new_concept loss: 6.829867e-01
202
+ 2025-02-18 10:05:13,969 INFO: [5/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q.weight
203
+ 2025-02-18 10:05:22,288 INFO: new_concept loss: 5.970761e-01
204
+ 2025-02-18 10:05:22,368 INFO: [6/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn1.to_k.weight
205
+ 2025-02-18 10:05:23,396 INFO: new_concept loss: 2.691664e-01
206
+ 2025-02-18 10:05:23,397 INFO: [7/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight
207
+ 2025-02-18 10:05:26,176 INFO: new_concept loss: 2.757275e-01
208
+ 2025-02-18 10:05:26,197 INFO: [8/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
209
+ 2025-02-18 10:05:28,994 INFO: new_concept loss: 2.462351e-01
210
+ 2025-02-18 10:05:29,016 INFO: [9/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
211
+ 2025-02-18 10:05:31,810 INFO: new_concept loss: 2.362856e-01
212
+ 2025-02-18 10:05:31,833 INFO: [10/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
213
+ 2025-02-18 10:05:39,922 INFO: new_concept loss: 5.065306e-02
214
+ 2025-02-18 10:05:40,004 INFO: [11/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight
215
+ 2025-02-18 10:05:42,750 INFO: new_concept loss: 1.462675e-01
216
+ 2025-02-18 10:05:42,772 INFO: [12/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
217
+ 2025-02-18 10:05:45,533 INFO: new_concept loss: 1.328609e-01
218
+ 2025-02-18 10:05:45,556 INFO: [13/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_q.weight
219
+ 2025-02-18 10:05:53,582 INFO: new_concept loss: 7.257976e-02
220
+ 2025-02-18 10:05:53,667 INFO: [14/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
221
+ 2025-02-18 10:06:01,915 INFO: new_concept loss: 3.954707e-02
222
+ 2025-02-18 10:06:01,997 INFO: [15/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_v.weight
223
+ 2025-02-18 10:06:10,340 INFO: new_concept loss: 1.240187e-01
224
+ 2025-02-18 10:06:10,422 INFO: [16/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out.0.weight
225
+ 2025-02-18 10:06:14,620 INFO: new_concept loss: 1.048966e-01
226
+ 2025-02-18 10:06:14,663 INFO: [17/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q.weight
227
+ 2025-02-18 10:06:18,971 INFO: new_concept loss: 4.386465e-01
228
+ 2025-02-18 10:06:19,017 INFO: [18/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight
229
+ 2025-02-18 10:06:21,835 INFO: new_concept loss: 7.685150e-01
230
+ 2025-02-18 10:06:21,859 INFO: [19/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q.weight
231
+ 2025-02-18 10:06:30,067 INFO: new_concept loss: 6.341255e-02
232
+ 2025-02-18 10:06:30,153 INFO: [20/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k.weight
233
+ 2025-02-18 10:06:32,957 INFO: new_concept loss: 7.451130e-01
234
+ 2025-02-18 10:06:32,980 INFO: [21/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
235
+ 2025-02-18 10:06:37,251 INFO: new_concept loss: 1.556838e-01
236
+ 2025-02-18 10:06:37,298 INFO: [22/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_q.weight
237
+ 2025-02-18 10:06:45,501 INFO: new_concept loss: 1.710174e-01
238
+ 2025-02-18 10:06:45,589 INFO: [23/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k.weight
239
+ 2025-02-18 10:06:53,892 INFO: new_concept loss: 3.234306e-01
240
+ 2025-02-18 10:06:53,982 INFO: [24/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
241
+ 2025-02-18 10:06:56,738 INFO: new_concept loss: 1.068963e-01
242
+ 2025-02-18 10:06:56,760 INFO: [25/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight
243
+ 2025-02-18 10:07:01,106 INFO: new_concept loss: 7.557544e-01
244
+ 2025-02-18 10:07:01,155 INFO: [26/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k.weight
245
+ 2025-02-18 10:07:03,984 INFO: new_concept loss: 7.542757e-01
246
+ 2025-02-18 10:07:04,009 INFO: [27/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q.weight
247
+ 2025-02-18 10:07:12,461 INFO: new_concept loss: 2.984946e-01
248
+ 2025-02-18 10:07:12,555 INFO: [28/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out.0.weight
249
+ 2025-02-18 10:07:16,900 INFO: new_concept loss: 1.445462e-01
250
+ 2025-02-18 10:07:16,949 INFO: [29/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight
251
+ 2025-02-18 10:07:21,145 INFO: new_concept loss: 5.463246e-01
252
+ 2025-02-18 10:07:21,194 INFO: [30/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v.weight
253
+ 2025-02-18 10:07:25,358 INFO: new_concept loss: 2.672614e-01
254
+ 2025-02-18 10:07:25,408 INFO: [31/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_v.weight
255
+ 2025-02-18 10:07:33,601 INFO: new_concept loss: 8.787098e-02
256
+ 2025-02-18 10:07:33,698 INFO: [32/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k.weight
257
+ 2025-02-18 10:07:42,035 INFO: new_concept loss: 2.288823e-01
258
+ 2025-02-18 10:07:42,131 INFO: [33/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q.weight
259
+ 2025-02-18 10:07:44,988 INFO: new_concept loss: 4.186683e-01
260
+ 2025-02-18 10:07:45,011 INFO: [34/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
261
+ 2025-02-18 10:07:53,232 INFO: new_concept loss: 6.477418e-02
262
+ 2025-02-18 10:07:53,332 INFO: [35/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out.0.weight
263
+ 2025-02-18 10:08:01,636 INFO: new_concept loss: 2.791651e-02
264
+ 2025-02-18 10:08:01,733 INFO: [36/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out.0.weight
265
+ 2025-02-18 10:08:09,930 INFO: new_concept loss: 2.724535e-02
266
+ 2025-02-18 10:08:10,024 INFO: [37/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
267
+ 2025-02-18 10:08:14,286 INFO: new_concept loss: 1.612051e-01
268
+ 2025-02-18 10:08:14,332 INFO: [38/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q.weight
269
+ 2025-02-18 10:08:18,730 INFO: new_concept loss: 1.113115e-01
270
+ 2025-02-18 10:08:18,779 INFO: [39/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v.weight
271
+ 2025-02-18 10:08:21,522 INFO: new_concept loss: 2.153144e-01
272
+ 2025-02-18 10:08:21,545 INFO: [40/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight
273
+ 2025-02-18 10:08:24,325 INFO: new_concept loss: 4.681203e-01
274
+ 2025-02-18 10:08:24,350 INFO: [41/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_k.weight
275
+ 2025-02-18 10:08:32,284 INFO: new_concept loss: 1.354147e-01
276
+ 2025-02-18 10:08:32,357 INFO: [42/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q.weight
277
+ 2025-02-18 10:08:36,701 INFO: new_concept loss: 2.082129e-01
278
+ 2025-02-18 10:08:36,743 INFO: [43/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q.weight
279
+ 2025-02-18 10:08:45,053 INFO: new_concept loss: 6.789617e-02
280
+ 2025-02-18 10:08:45,130 INFO: [44/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_q.weight
281
+ 2025-02-18 10:08:53,543 INFO: new_concept loss: 6.113157e-02
282
+ 2025-02-18 10:08:53,623 INFO: [45/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight
283
+ 2025-02-18 10:08:57,785 INFO: new_concept loss: 1.470006e-01
284
+ 2025-02-18 10:08:57,831 INFO: [46/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight
285
+ 2025-02-18 10:09:02,148 INFO: new_concept loss: 7.435892e-01
286
+ 2025-02-18 10:09:02,195 INFO: [47/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight
287
+ 2025-02-18 10:09:05,011 INFO: new_concept loss: 2.135441e-01
288
+ 2025-02-18 10:09:05,035 INFO: [48/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v.weight
289
+ 2025-02-18 10:09:13,424 INFO: new_concept loss: 1.767161e-01
290
+ 2025-02-18 10:09:13,508 INFO: [49/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
291
+ 2025-02-18 10:09:17,712 INFO: new_concept loss: 9.232481e-02
292
+ 2025-02-18 10:09:17,759 INFO: [50/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
293
+ 2025-02-18 10:09:18,797 INFO: new_concept loss: 1.938113e-01
294
+ 2025-02-18 10:09:18,798 INFO: [51/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight
295
+ 2025-02-18 10:09:21,654 INFO: new_concept loss: 1.931212e-01
296
+ 2025-02-18 10:09:21,680 INFO: [52/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight
297
+ 2025-02-18 10:09:25,985 INFO: new_concept loss: 6.031875e-01
298
+ 2025-02-18 10:09:26,030 INFO: [53/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
299
+ 2025-02-18 10:09:30,128 INFO: new_concept loss: 6.304939e-02
300
+ 2025-02-18 10:09:30,180 INFO: [54/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight
301
+ 2025-02-18 10:09:32,960 INFO: new_concept loss: 3.609971e-01
302
+ 2025-02-18 10:09:32,985 INFO: [55/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
303
+ 2025-02-18 10:09:37,153 INFO: new_concept loss: 6.763718e-01
304
+ 2025-02-18 10:09:37,205 INFO: [56/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q.weight
305
+ 2025-02-18 10:09:41,528 INFO: new_concept loss: 3.740048e-01
306
+ 2025-02-18 10:09:41,577 INFO: [57/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k.weight
307
+ 2025-02-18 10:09:45,957 INFO: new_concept loss: 2.385424e-01
308
+ 2025-02-18 10:09:46,006 INFO: [58/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
309
+ 2025-02-18 10:09:54,013 INFO: new_concept loss: 8.359801e-02
310
+ 2025-02-18 10:09:54,111 INFO: [59/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k.weight
311
+ 2025-02-18 10:10:02,516 INFO: new_concept loss: 2.154270e-01
312
+ 2025-02-18 10:10:02,621 INFO: [60/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn1.to_v.weight
313
+ 2025-02-18 10:10:03,635 INFO: new_concept loss: 1.350740e-01
314
+ 2025-02-18 10:10:03,636 INFO: [61/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
315
+ 2025-02-18 10:10:04,650 INFO: new_concept loss: 7.636815e-02
316
+ 2025-02-18 10:10:04,651 INFO: [62/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight
317
+ 2025-02-18 10:10:08,821 INFO: new_concept loss: 2.006273e-01
318
+ 2025-02-18 10:10:08,867 INFO: [63/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_k.weight
319
+ 2025-02-18 10:10:16,971 INFO: new_concept loss: 1.642606e-01
320
+ 2025-02-18 10:10:17,059 INFO: [64/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q.weight
321
+ 2025-02-18 10:10:19,851 INFO: new_concept loss: 2.283725e-01
322
+ 2025-02-18 10:10:19,873 INFO: [65/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q.weight
323
+ 2025-02-18 10:10:22,741 INFO: new_concept loss: 2.470787e-01
324
+ 2025-02-18 10:10:22,768 INFO: [66/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn1.to_q.weight
325
+ 2025-02-18 10:10:23,788 INFO: new_concept loss: 1.714881e-01
326
+ 2025-02-18 10:10:23,788 INFO: [67/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
327
+ 2025-02-18 10:10:31,935 INFO: new_concept loss: 2.508941e-02
328
+ 2025-02-18 10:10:32,031 INFO: [68/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
329
+ 2025-02-18 10:10:34,881 INFO: new_concept loss: 1.466412e-01
330
+ 2025-02-18 10:10:34,905 INFO: [69/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0.weight
331
+ 2025-02-18 10:10:37,773 INFO: new_concept loss: 1.901781e-01
332
+ 2025-02-18 10:10:37,800 INFO: [70/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight
333
+ 2025-02-18 10:10:42,327 INFO: new_concept loss: 3.197155e-01
334
+ 2025-02-18 10:10:42,381 INFO: [71/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q.weight
335
+ 2025-02-18 10:10:45,243 INFO: new_concept loss: 3.977647e-01
336
+ 2025-02-18 10:10:45,267 INFO: [72/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight
337
+ 2025-02-18 10:10:48,090 INFO: new_concept loss: 5.833027e-01
338
+ 2025-02-18 10:10:48,118 INFO: [73/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v.weight
339
+ 2025-02-18 10:10:50,951 INFO: new_concept loss: 2.237324e-01
340
+ 2025-02-18 10:10:50,978 INFO: [74/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
341
+ 2025-02-18 10:10:55,080 INFO: new_concept loss: 4.140265e-01
342
+ 2025-02-18 10:10:55,131 INFO: [75/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q.weight
343
+ 2025-02-18 10:11:03,581 INFO: new_concept loss: 2.084374e-01
344
+ 2025-02-18 10:11:03,675 INFO: [76/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight
345
+ 2025-02-18 10:11:06,508 INFO: new_concept loss: 1.883232e-01
346
+ 2025-02-18 10:11:06,534 INFO: [77/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight
347
+ 2025-02-18 10:11:10,817 INFO: new_concept loss: 4.924987e-01
348
+ 2025-02-18 10:11:10,869 INFO: [78/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q.weight
349
+ 2025-02-18 10:11:18,874 INFO: new_concept loss: 9.835833e-02
350
+ 2025-02-18 10:11:18,977 INFO: [79/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v.weight
351
+ 2025-02-18 10:11:23,223 INFO: new_concept loss: 1.199531e-01
352
+ 2025-02-18 10:11:23,273 INFO: [80/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight
353
+ 2025-02-18 10:11:27,451 INFO: new_concept loss: 2.096862e-01
354
+ 2025-02-18 10:11:27,502 INFO: [81/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
355
+ 2025-02-18 10:11:30,343 INFO: new_concept loss: 1.246875e-01
356
+ 2025-02-18 10:11:30,369 INFO: [82/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight
357
+ 2025-02-18 10:11:33,070 INFO: new_concept loss: 2.763811e-01
358
+ 2025-02-18 10:11:33,096 INFO: [83/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
359
+ 2025-02-18 10:11:35,921 INFO: new_concept loss: 3.674817e-01
360
+ 2025-02-18 10:11:35,947 INFO: [84/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
361
+ 2025-02-18 10:11:44,012 INFO: new_concept loss: 3.000834e-02
362
+ 2025-02-18 10:11:44,112 INFO: [85/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
363
+ 2025-02-18 10:11:48,387 INFO: new_concept loss: 1.812962e-02
364
+ 2025-02-18 10:11:48,436 INFO: [86/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
365
+ 2025-02-18 10:11:52,674 INFO: new_concept loss: 7.571394e-01
366
+ 2025-02-18 10:11:52,725 INFO: [87/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight
367
+ 2025-02-18 10:11:56,823 INFO: new_concept loss: 1.991518e-01
368
+ 2025-02-18 10:11:56,873 INFO: [88/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight
369
+ 2025-02-18 10:12:01,104 INFO: new_concept loss: 3.178474e-01
370
+ 2025-02-18 10:12:01,155 INFO: [89/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight
371
+ 2025-02-18 10:12:03,957 INFO: new_concept loss: 1.593530e-01
372
+ 2025-02-18 10:12:03,980 INFO: [90/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0.weight
373
+ 2025-02-18 10:12:06,736 INFO: new_concept loss: 1.556669e-01
374
+ 2025-02-18 10:12:06,755 INFO: [91/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v.weight
375
+ 2025-02-18 10:12:15,035 INFO: new_concept loss: 3.353187e-01
376
+ 2025-02-18 10:12:15,111 INFO: [92/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
377
+ 2025-02-18 10:12:23,691 INFO: new_concept loss: 6.571595e-03
378
+ 2025-02-18 10:12:23,771 INFO: [93/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight
379
+ 2025-02-18 10:12:26,686 INFO: new_concept loss: 3.799535e-01
380
+ 2025-02-18 10:12:26,707 INFO: [94/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v.weight
381
+ 2025-02-18 10:12:34,861 INFO: new_concept loss: 1.101350e-01
382
+ 2025-02-18 10:12:34,941 INFO: [95/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
383
+ 2025-02-18 10:12:43,608 INFO: new_concept loss: 1.892852e-01
384
+ 2025-02-18 10:12:43,701 INFO: [96/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight
385
+ 2025-02-18 10:12:48,229 INFO: new_concept loss: 5.540602e-01
386
+ 2025-02-18 10:12:48,276 INFO: remove 96 hooker from unet
FrozenII/combined_model_base/feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "feature_extractor_type": "CLIPFeatureExtractor",
12
+ "image_mean": [
13
+ 0.48145466,
14
+ 0.4578275,
15
+ 0.40821073
16
+ ],
17
+ "image_processor_type": "CLIPFeatureExtractor",
18
+ "image_std": [
19
+ 0.26862954,
20
+ 0.26130258,
21
+ 0.27577711
22
+ ],
23
+ "resample": 3,
24
+ "rescale_factor": 0.00392156862745098,
25
+ "size": {
26
+ "shortest_edge": 224
27
+ }
28
+ }
FrozenII/combined_model_base/model_index.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.20.0",
4
+ "_name_or_path": "experiments/pretrained_models/chilloutmix",
5
+ "feature_extractor": [
6
+ "transformers",
7
+ "CLIPFeatureExtractor"
8
+ ],
9
+ "requires_safety_checker": true,
10
+ "safety_checker": [
11
+ null,
12
+ null
13
+ ],
14
+ "scheduler": [
15
+ "diffusers",
16
+ "DPMSolverMultistepScheduler"
17
+ ],
18
+ "text_encoder": [
19
+ "transformers",
20
+ "CLIPTextModel"
21
+ ],
22
+ "tokenizer": [
23
+ "transformers",
24
+ "CLIPTokenizer"
25
+ ],
26
+ "unet": [
27
+ "diffusers",
28
+ "UNet2DConditionModel"
29
+ ],
30
+ "vae": [
31
+ "diffusers",
32
+ "AutoencoderKL"
33
+ ]
34
+ }
FrozenII/combined_model_base/new_concept_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<Elsa1>": {"concept_token_ids": [49408, 49409, 49410, 49411, 49412, 49413, 49414, 49415, 49416, 49417, 49418, 49419, 49420, 49421, 49422, 49423], "concept_token_names": ["<new0>", "<new1>", "<new2>", "<new3>", "<new4>", "<new5>", "<new6>", "<new7>", "<new8>", "<new9>", "<new10>", "<new11>", "<new12>", "<new13>", "<new14>", "<new15>"]}, "<Elsa2>": {"concept_token_ids": [49424, 49425, 49426, 49427, 49428, 49429, 49430, 49431, 49432, 49433, 49434, 49435, 49436, 49437, 49438, 49439], "concept_token_names": ["<new16>", "<new17>", "<new18>", "<new19>", "<new20>", "<new21>", "<new22>", "<new23>", "<new24>", "<new25>", "<new26>", "<new27>", "<new28>", "<new29>", "<new30>", "<new31>"]}, "<Olaf1>": {"concept_token_ids": [49440, 49441, 49442, 49443, 49444, 49445, 49446, 49447, 49448, 49449, 49450, 49451, 49452, 49453, 49454, 49455], "concept_token_names": ["<new32>", "<new33>", "<new34>", "<new35>", "<new36>", "<new37>", "<new38>", "<new39>", "<new40>", "<new41>", "<new42>", "<new43>", "<new44>", "<new45>", "<new46>", "<new47>"]}, "<Olaf2>": {"concept_token_ids": [49456, 49457, 49458, 49459, 49460, 49461, 49462, 49463, 49464, 49465, 49466, 49467, 49468, 49469, 49470, 49471], "concept_token_names": ["<new48>", "<new49>", "<new50>", "<new51>", "<new52>", "<new53>", "<new54>", "<new55>", "<new56>", "<new57>", "<new58>", "<new59>", "<new60>", "<new61>", "<new62>", "<new63>"]}, "<Mattias1>": {"concept_token_ids": [49472, 49473, 49474, 49475, 49476, 49477, 49478, 49479, 49480, 49481, 49482, 49483, 49484, 49485, 49486, 49487], "concept_token_names": ["<new64>", "<new65>", "<new66>", "<new67>", "<new68>", "<new69>", "<new70>", "<new71>", "<new72>", "<new73>", "<new74>", "<new75>", "<new76>", "<new77>", "<new78>", "<new79>"]}, "<Mattias2>": {"concept_token_ids": [49488, 49489, 49490, 49491, 49492, 49493, 49494, 49495, 49496, 49497, 49498, 49499, 49500, 49501, 49502, 49503], "concept_token_names": ["<new80>", "<new81>", "<new82>", "<new83>", "<new84>", "<new85>", "<new86>", "<new87>", "<new88>", "<new89>", "<new90>", "<new91>", "<new92>", "<new93>", "<new94>", "<new95>"]}, "<Kristoff1>": {"concept_token_ids": [49504, 49505, 49506, 49507, 49508, 49509, 49510, 49511, 49512, 49513, 49514, 49515, 49516, 49517, 49518, 49519], "concept_token_names": ["<new96>", "<new97>", "<new98>", "<new99>", "<new100>", "<new101>", "<new102>", "<new103>", "<new104>", "<new105>", "<new106>", "<new107>", "<new108>", "<new109>", "<new110>", "<new111>"]}, "<Kristoff2>": {"concept_token_ids": [49520, 49521, 49522, 49523, 49524, 49525, 49526, 49527, 49528, 49529, 49530, 49531, 49532, 49533, 49534, 49535], "concept_token_names": ["<new112>", "<new113>", "<new114>", "<new115>", "<new116>", "<new117>", "<new118>", "<new119>", "<new120>", "<new121>", "<new122>", "<new123>", "<new124>", "<new125>", "<new126>", "<new127>"]}, "<Anna1>": {"concept_token_ids": [49536, 49537, 49538, 49539, 49540, 49541, 49542, 49543, 49544, 49545, 49546, 49547, 49548, 49549, 49550, 49551], "concept_token_names": ["<new128>", "<new129>", "<new130>", "<new131>", "<new132>", "<new133>", "<new134>", "<new135>", "<new136>", "<new137>", "<new138>", "<new139>", "<new140>", "<new141>", "<new142>", "<new143>"]}, "<Anna2>": {"concept_token_ids": [49552, 49553, 49554, 49555, 49556, 49557, 49558, 49559, 49560, 49561, 49562, 49563, 49564, 49565, 49566, 49567], "concept_token_names": ["<new144>", "<new145>", "<new146>", "<new147>", "<new148>", "<new149>", "<new150>", "<new151>", "<new152>", "<new153>", "<new154>", "<new155>", "<new156>", "<new157>", "<new158>", "<new159>"]}}
FrozenII/combined_model_base/scheduler/scheduler_config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "DPMSolverMultistepScheduler",
3
+ "_diffusers_version": "0.20.0",
4
+ "algorithm_type": "dpmsolver++",
5
+ "beta_end": 0.012,
6
+ "beta_schedule": "scaled_linear",
7
+ "beta_start": 0.00085,
8
+ "clip_sample": false,
9
+ "dynamic_thresholding_ratio": 0.995,
10
+ "lambda_min_clipped": -Infinity,
11
+ "lower_order_final": true,
12
+ "num_train_timesteps": 1000,
13
+ "prediction_type": "epsilon",
14
+ "sample_max_value": 1.0,
15
+ "set_alpha_to_one": false,
16
+ "skip_prk_steps": true,
17
+ "solver_order": 2,
18
+ "solver_type": "midpoint",
19
+ "steps_offset": 1,
20
+ "thresholding": false,
21
+ "timestep_spacing": "linspace",
22
+ "trained_betas": null,
23
+ "use_karras_sigmas": false,
24
+ "variance_type": null
25
+ }
FrozenII/combined_model_base/text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "experiments/pretrained_models/chilloutmix/text_encoder",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 768,
22
+ "torch_dtype": "float16",
23
+ "transformers_version": "4.28.0",
24
+ "vocab_size": 49568
25
+ }
FrozenII/combined_model_base/text_encoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b0fe01461cd3e5445bd05cb66d6f0d58747dc9f144428da8d225bd2d1b00741
3
+ size 246390624
FrozenII/combined_model_base/tokenizer/added_tokens.json ADDED
@@ -0,0 +1,162 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<new0>": 49408,
3
+ "<new100>": 49508,
4
+ "<new101>": 49509,
5
+ "<new102>": 49510,
6
+ "<new103>": 49511,
7
+ "<new104>": 49512,
8
+ "<new105>": 49513,
9
+ "<new106>": 49514,
10
+ "<new107>": 49515,
11
+ "<new108>": 49516,
12
+ "<new109>": 49517,
13
+ "<new10>": 49418,
14
+ "<new110>": 49518,
15
+ "<new111>": 49519,
16
+ "<new112>": 49520,
17
+ "<new113>": 49521,
18
+ "<new114>": 49522,
19
+ "<new115>": 49523,
20
+ "<new116>": 49524,
21
+ "<new117>": 49525,
22
+ "<new118>": 49526,
23
+ "<new119>": 49527,
24
+ "<new11>": 49419,
25
+ "<new120>": 49528,
26
+ "<new121>": 49529,
27
+ "<new122>": 49530,
28
+ "<new123>": 49531,
29
+ "<new124>": 49532,
30
+ "<new125>": 49533,
31
+ "<new126>": 49534,
32
+ "<new127>": 49535,
33
+ "<new128>": 49536,
34
+ "<new129>": 49537,
35
+ "<new12>": 49420,
36
+ "<new130>": 49538,
37
+ "<new131>": 49539,
38
+ "<new132>": 49540,
39
+ "<new133>": 49541,
40
+ "<new134>": 49542,
41
+ "<new135>": 49543,
42
+ "<new136>": 49544,
43
+ "<new137>": 49545,
44
+ "<new138>": 49546,
45
+ "<new139>": 49547,
46
+ "<new13>": 49421,
47
+ "<new140>": 49548,
48
+ "<new141>": 49549,
49
+ "<new142>": 49550,
50
+ "<new143>": 49551,
51
+ "<new144>": 49552,
52
+ "<new145>": 49553,
53
+ "<new146>": 49554,
54
+ "<new147>": 49555,
55
+ "<new148>": 49556,
56
+ "<new149>": 49557,
57
+ "<new14>": 49422,
58
+ "<new150>": 49558,
59
+ "<new151>": 49559,
60
+ "<new152>": 49560,
61
+ "<new153>": 49561,
62
+ "<new154>": 49562,
63
+ "<new155>": 49563,
64
+ "<new156>": 49564,
65
+ "<new157>": 49565,
66
+ "<new158>": 49566,
67
+ "<new159>": 49567,
68
+ "<new15>": 49423,
69
+ "<new16>": 49424,
70
+ "<new17>": 49425,
71
+ "<new18>": 49426,
72
+ "<new19>": 49427,
73
+ "<new1>": 49409,
74
+ "<new20>": 49428,
75
+ "<new21>": 49429,
76
+ "<new22>": 49430,
77
+ "<new23>": 49431,
78
+ "<new24>": 49432,
79
+ "<new25>": 49433,
80
+ "<new26>": 49434,
81
+ "<new27>": 49435,
82
+ "<new28>": 49436,
83
+ "<new29>": 49437,
84
+ "<new2>": 49410,
85
+ "<new30>": 49438,
86
+ "<new31>": 49439,
87
+ "<new32>": 49440,
88
+ "<new33>": 49441,
89
+ "<new34>": 49442,
90
+ "<new35>": 49443,
91
+ "<new36>": 49444,
92
+ "<new37>": 49445,
93
+ "<new38>": 49446,
94
+ "<new39>": 49447,
95
+ "<new3>": 49411,
96
+ "<new40>": 49448,
97
+ "<new41>": 49449,
98
+ "<new42>": 49450,
99
+ "<new43>": 49451,
100
+ "<new44>": 49452,
101
+ "<new45>": 49453,
102
+ "<new46>": 49454,
103
+ "<new47>": 49455,
104
+ "<new48>": 49456,
105
+ "<new49>": 49457,
106
+ "<new4>": 49412,
107
+ "<new50>": 49458,
108
+ "<new51>": 49459,
109
+ "<new52>": 49460,
110
+ "<new53>": 49461,
111
+ "<new54>": 49462,
112
+ "<new55>": 49463,
113
+ "<new56>": 49464,
114
+ "<new57>": 49465,
115
+ "<new58>": 49466,
116
+ "<new59>": 49467,
117
+ "<new5>": 49413,
118
+ "<new60>": 49468,
119
+ "<new61>": 49469,
120
+ "<new62>": 49470,
121
+ "<new63>": 49471,
122
+ "<new64>": 49472,
123
+ "<new65>": 49473,
124
+ "<new66>": 49474,
125
+ "<new67>": 49475,
126
+ "<new68>": 49476,
127
+ "<new69>": 49477,
128
+ "<new6>": 49414,
129
+ "<new70>": 49478,
130
+ "<new71>": 49479,
131
+ "<new72>": 49480,
132
+ "<new73>": 49481,
133
+ "<new74>": 49482,
134
+ "<new75>": 49483,
135
+ "<new76>": 49484,
136
+ "<new77>": 49485,
137
+ "<new78>": 49486,
138
+ "<new79>": 49487,
139
+ "<new7>": 49415,
140
+ "<new80>": 49488,
141
+ "<new81>": 49489,
142
+ "<new82>": 49490,
143
+ "<new83>": 49491,
144
+ "<new84>": 49492,
145
+ "<new85>": 49493,
146
+ "<new86>": 49494,
147
+ "<new87>": 49495,
148
+ "<new88>": 49496,
149
+ "<new89>": 49497,
150
+ "<new8>": 49416,
151
+ "<new90>": 49498,
152
+ "<new91>": 49499,
153
+ "<new92>": 49500,
154
+ "<new93>": 49501,
155
+ "<new94>": 49502,
156
+ "<new95>": 49503,
157
+ "<new96>": 49504,
158
+ "<new97>": 49505,
159
+ "<new98>": 49506,
160
+ "<new99>": 49507,
161
+ "<new9>": 49417
162
+ }
FrozenII/combined_model_base/tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
FrozenII/combined_model_base/tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
FrozenII/combined_model_base/tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "clean_up_tokenization_spaces": true,
12
+ "do_lower_case": true,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "<|endoftext|>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "errors": "replace",
22
+ "model_max_length": 77,
23
+ "pad_token": "<|endoftext|>",
24
+ "tokenizer_class": "CLIPTokenizer",
25
+ "unk_token": {
26
+ "__type": "AddedToken",
27
+ "content": "<|endoftext|>",
28
+ "lstrip": false,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
+ }
FrozenII/combined_model_base/tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
FrozenII/combined_model_base/unet/config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.20.0",
4
+ "_name_or_path": "experiments/pretrained_models/chilloutmix/unet",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": null,
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": null,
9
+ "attention_head_dim": 8,
10
+ "attention_type": "default",
11
+ "block_out_channels": [
12
+ 320,
13
+ 640,
14
+ 1280,
15
+ 1280
16
+ ],
17
+ "center_input_sample": false,
18
+ "class_embed_type": null,
19
+ "class_embeddings_concat": false,
20
+ "conv_in_kernel": 3,
21
+ "conv_out_kernel": 3,
22
+ "cross_attention_dim": 768,
23
+ "cross_attention_norm": null,
24
+ "down_block_types": [
25
+ "CrossAttnDownBlock2D",
26
+ "CrossAttnDownBlock2D",
27
+ "CrossAttnDownBlock2D",
28
+ "DownBlock2D"
29
+ ],
30
+ "downsample_padding": 1,
31
+ "dual_cross_attention": false,
32
+ "encoder_hid_dim": null,
33
+ "encoder_hid_dim_type": null,
34
+ "flip_sin_to_cos": true,
35
+ "freq_shift": 0,
36
+ "in_channels": 4,
37
+ "layers_per_block": 2,
38
+ "mid_block_only_cross_attention": null,
39
+ "mid_block_scale_factor": 1,
40
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
41
+ "norm_eps": 1e-05,
42
+ "norm_num_groups": 32,
43
+ "num_attention_heads": null,
44
+ "num_class_embeds": null,
45
+ "only_cross_attention": false,
46
+ "out_channels": 4,
47
+ "projection_class_embeddings_input_dim": null,
48
+ "resnet_out_scale_factor": 1.0,
49
+ "resnet_skip_time_act": false,
50
+ "resnet_time_scale_shift": "default",
51
+ "sample_size": 64,
52
+ "time_cond_proj_dim": null,
53
+ "time_embedding_act_fn": null,
54
+ "time_embedding_dim": null,
55
+ "time_embedding_type": "positional",
56
+ "timestep_post_act": null,
57
+ "transformer_layers_per_block": 1,
58
+ "up_block_types": [
59
+ "UpBlock2D",
60
+ "CrossAttnUpBlock2D",
61
+ "CrossAttnUpBlock2D",
62
+ "CrossAttnUpBlock2D"
63
+ ],
64
+ "upcast_attention": false,
65
+ "use_linear_projection": false
66
+ }
FrozenII/combined_model_base/unet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58b099a1e1435fcb21ed9d04e6ad6c3c7f5aaeaa207b3226d4203fe51f801217
3
+ size 1719125304
FrozenII/combined_model_base/vae/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.20.0",
4
+ "_name_or_path": "experiments/pretrained_models/chilloutmix/vae",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "force_upcast": true,
19
+ "in_channels": 3,
20
+ "latent_channels": 4,
21
+ "layers_per_block": 2,
22
+ "norm_num_groups": 32,
23
+ "out_channels": 3,
24
+ "sample_size": 512,
25
+ "scaling_factor": 0.18215,
26
+ "up_block_types": [
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D",
29
+ "UpDecoderBlock2D",
30
+ "UpDecoderBlock2D"
31
+ ]
32
+ }
FrozenII/combined_model_base/vae/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e4c08995484ee61270175e9e7a072b66a6e4eeb5f0c266667fe1f45b90daf9a
3
+ size 167335342