weijiawu
commited on
Commit
·
634a418
1
Parent(s):
20a74c0
weight
Browse files- FrozenII/combined_model_base.log +386 -0
- FrozenII/combined_model_base/feature_extractor/preprocessor_config.json +28 -0
- FrozenII/combined_model_base/model_index.json +34 -0
- FrozenII/combined_model_base/new_concept_cfg.json +1 -0
- FrozenII/combined_model_base/scheduler/scheduler_config.json +25 -0
- FrozenII/combined_model_base/text_encoder/config.json +25 -0
- FrozenII/combined_model_base/text_encoder/model.safetensors +3 -0
- FrozenII/combined_model_base/tokenizer/added_tokens.json +162 -0
- FrozenII/combined_model_base/tokenizer/merges.txt +0 -0
- FrozenII/combined_model_base/tokenizer/special_tokens_map.json +24 -0
- FrozenII/combined_model_base/tokenizer/tokenizer_config.json +33 -0
- FrozenII/combined_model_base/tokenizer/vocab.json +0 -0
- FrozenII/combined_model_base/unet/config.json +66 -0
- FrozenII/combined_model_base/unet/diffusion_pytorch_model.safetensors +3 -0
- FrozenII/combined_model_base/vae/config.json +32 -0
- FrozenII/combined_model_base/vae/diffusion_pytorch_model.safetensors +3 -0
FrozenII/combined_model_base.log
ADDED
@@ -0,0 +1,386 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2025-02-18 10:01:13,607 INFO: Namespace(concept_cfg='datasets/data_cfgs/MixofShow/multi-concept/MovieGen/FrozenII/Elsa+Olaf+Mattias+Kristoff+Anna.json', optimize_textenc_iters=1000, optimize_unet_iters=100, pretrained_models='experiments/pretrained_models/chilloutmix', save_path='experiments/composed_edlora/chilloutmix/FrozenII', suffix='base')
|
2 |
+
2025-02-18 10:01:13,607 INFO: ------Step 1: load stable diffusion checkpoint------
|
3 |
+
2025-02-18 10:01:18,406 INFO: ------Step 2: load new concepts checkpoints------
|
4 |
+
2025-02-18 10:01:18,450 INFO: ------Step 3: merge token embedding------
|
5 |
+
2025-02-18 10:01:18,680 INFO: concept <Elsa1> is bind with token_id: [49408, 49423]
|
6 |
+
2025-02-18 10:01:18,902 INFO: concept <Elsa2> is bind with token_id: [49424, 49439]
|
7 |
+
2025-02-18 10:01:19,126 INFO: concept <Olaf1> is bind with token_id: [49440, 49455]
|
8 |
+
2025-02-18 10:01:19,350 INFO: concept <Olaf2> is bind with token_id: [49456, 49471]
|
9 |
+
2025-02-18 10:01:19,575 INFO: concept <Mattias1> is bind with token_id: [49472, 49487]
|
10 |
+
2025-02-18 10:01:19,801 INFO: concept <Mattias2> is bind with token_id: [49488, 49503]
|
11 |
+
2025-02-18 10:01:20,028 INFO: concept <Kristoff1> is bind with token_id: [49504, 49519]
|
12 |
+
2025-02-18 10:01:20,257 INFO: concept <Kristoff2> is bind with token_id: [49520, 49535]
|
13 |
+
2025-02-18 10:01:20,485 INFO: concept <Anna1> is bind with token_id: [49536, 49551]
|
14 |
+
2025-02-18 10:01:20,712 INFO: concept <Anna2> is bind with token_id: [49552, 49567]
|
15 |
+
2025-02-18 10:01:20,712 INFO: ------Step 4: merge text encoder------
|
16 |
+
2025-02-18 10:01:20,831 INFO: text_encoder have 48 linear layer need to optimize
|
17 |
+
2025-02-18 10:01:20,832 INFO: add 48 hooker to text_encoder
|
18 |
+
2025-02-18 10:01:20,859 INFO: load 48 LoRAs of text_encoder
|
19 |
+
2025-02-18 10:01:21,132 INFO: load 48 LoRAs of text_encoder
|
20 |
+
2025-02-18 10:01:21,372 INFO: load 48 LoRAs of text_encoder
|
21 |
+
2025-02-18 10:01:21,613 INFO: load 48 LoRAs of text_encoder
|
22 |
+
2025-02-18 10:01:21,850 INFO: load 48 LoRAs of text_encoder
|
23 |
+
2025-02-18 10:01:22,103 INFO: [1/48] optimizing text_model.encoder.layers.4.self_attn.q_proj.weight
|
24 |
+
2025-02-18 10:01:24,861 INFO: new_concept loss: 1.378319e-02
|
25 |
+
2025-02-18 10:01:24,861 INFO: [2/48] optimizing text_model.encoder.layers.4.self_attn.k_proj.weight
|
26 |
+
2025-02-18 10:01:27,617 INFO: new_concept loss: 1.130946e-02
|
27 |
+
2025-02-18 10:01:27,617 INFO: [3/48] optimizing text_model.encoder.layers.10.self_attn.k_proj.weight
|
28 |
+
2025-02-18 10:01:30,514 INFO: new_concept loss: 6.835580e-03
|
29 |
+
2025-02-18 10:01:30,514 INFO: [4/48] optimizing text_model.encoder.layers.7.self_attn.k_proj.weight
|
30 |
+
2025-02-18 10:01:33,060 INFO: new_concept loss: 1.043152e-02
|
31 |
+
2025-02-18 10:01:33,060 INFO: [5/48] optimizing text_model.encoder.layers.6.self_attn.k_proj.weight
|
32 |
+
2025-02-18 10:01:35,603 INFO: new_concept loss: 1.229318e-02
|
33 |
+
2025-02-18 10:01:35,603 INFO: [6/48] optimizing text_model.encoder.layers.11.self_attn.k_proj.weight
|
34 |
+
2025-02-18 10:01:38,154 INFO: new_concept loss: 2.929659e-02
|
35 |
+
2025-02-18 10:01:38,154 INFO: [7/48] optimizing text_model.encoder.layers.2.self_attn.q_proj.weight
|
36 |
+
2025-02-18 10:01:40,728 INFO: new_concept loss: 1.220211e-02
|
37 |
+
2025-02-18 10:01:40,728 INFO: [8/48] optimizing text_model.encoder.layers.11.self_attn.v_proj.weight
|
38 |
+
2025-02-18 10:01:43,589 INFO: new_concept loss: 1.071226e-02
|
39 |
+
2025-02-18 10:01:43,589 INFO: [9/48] optimizing text_model.encoder.layers.6.self_attn.out_proj.weight
|
40 |
+
2025-02-18 10:01:46,495 INFO: new_concept loss: 2.268766e-03
|
41 |
+
2025-02-18 10:01:46,495 INFO: [10/48] optimizing text_model.encoder.layers.5.self_attn.out_proj.weight
|
42 |
+
2025-02-18 10:01:48,888 INFO: new_concept loss: 2.712738e-03
|
43 |
+
2025-02-18 10:01:48,889 INFO: [11/48] optimizing text_model.encoder.layers.7.self_attn.out_proj.weight
|
44 |
+
2025-02-18 10:01:51,311 INFO: new_concept loss: 2.453873e-03
|
45 |
+
2025-02-18 10:01:51,311 INFO: [12/48] optimizing text_model.encoder.layers.9.self_attn.k_proj.weight
|
46 |
+
2025-02-18 10:01:54,139 INFO: new_concept loss: 1.146064e-02
|
47 |
+
2025-02-18 10:01:54,139 INFO: [13/48] optimizing text_model.encoder.layers.2.self_attn.k_proj.weight
|
48 |
+
2025-02-18 10:01:56,883 INFO: new_concept loss: 1.995232e-02
|
49 |
+
2025-02-18 10:01:56,884 INFO: [14/48] optimizing text_model.encoder.layers.3.self_attn.k_proj.weight
|
50 |
+
2025-02-18 10:01:59,640 INFO: new_concept loss: 2.060306e-02
|
51 |
+
2025-02-18 10:01:59,641 INFO: [15/48] optimizing text_model.encoder.layers.4.self_attn.v_proj.weight
|
52 |
+
2025-02-18 10:02:02,303 INFO: new_concept loss: 1.727206e-02
|
53 |
+
2025-02-18 10:02:02,303 INFO: [16/48] optimizing text_model.encoder.layers.11.self_attn.out_proj.weight
|
54 |
+
2025-02-18 10:02:05,097 INFO: new_concept loss: 4.675179e-03
|
55 |
+
2025-02-18 10:02:05,098 INFO: [17/48] optimizing text_model.encoder.layers.3.self_attn.v_proj.weight
|
56 |
+
2025-02-18 10:02:07,841 INFO: new_concept loss: 2.080345e-02
|
57 |
+
2025-02-18 10:02:07,841 INFO: [18/48] optimizing text_model.encoder.layers.9.self_attn.out_proj.weight
|
58 |
+
2025-02-18 10:02:10,796 INFO: new_concept loss: 3.901713e-03
|
59 |
+
2025-02-18 10:02:10,796 INFO: [19/48] optimizing text_model.encoder.layers.9.self_attn.q_proj.weight
|
60 |
+
2025-02-18 10:02:13,761 INFO: new_concept loss: 1.188236e-02
|
61 |
+
2025-02-18 10:02:13,762 INFO: [20/48] optimizing text_model.encoder.layers.0.self_attn.k_proj.weight
|
62 |
+
2025-02-18 10:02:14,587 INFO: new_concept loss: 3.598483e-02
|
63 |
+
2025-02-18 10:02:14,587 INFO: [21/48] optimizing text_model.encoder.layers.8.self_attn.q_proj.weight
|
64 |
+
2025-02-18 10:02:17,432 INFO: new_concept loss: 9.705781e-03
|
65 |
+
2025-02-18 10:02:17,432 INFO: [22/48] optimizing text_model.encoder.layers.3.self_attn.q_proj.weight
|
66 |
+
2025-02-18 10:02:20,256 INFO: new_concept loss: 1.718284e-02
|
67 |
+
2025-02-18 10:02:20,256 INFO: [23/48] optimizing text_model.encoder.layers.5.self_attn.v_proj.weight
|
68 |
+
2025-02-18 10:02:23,243 INFO: new_concept loss: 1.262343e-02
|
69 |
+
2025-02-18 10:02:23,243 INFO: [24/48] optimizing text_model.encoder.layers.1.self_attn.out_proj.weight
|
70 |
+
2025-02-18 10:02:26,025 INFO: new_concept loss: 2.885660e-03
|
71 |
+
2025-02-18 10:02:26,025 INFO: [25/48] optimizing text_model.encoder.layers.8.self_attn.v_proj.weight
|
72 |
+
2025-02-18 10:02:28,986 INFO: new_concept loss: 1.011055e-02
|
73 |
+
2025-02-18 10:02:28,986 INFO: [26/48] optimizing text_model.encoder.layers.1.self_attn.v_proj.weight
|
74 |
+
2025-02-18 10:02:31,798 INFO: new_concept loss: 2.109258e-02
|
75 |
+
2025-02-18 10:02:31,798 INFO: [27/48] optimizing text_model.encoder.layers.8.self_attn.out_proj.weight
|
76 |
+
2025-02-18 10:02:34,334 INFO: new_concept loss: 2.203831e-03
|
77 |
+
2025-02-18 10:02:34,334 INFO: [28/48] optimizing text_model.encoder.layers.3.self_attn.out_proj.weight
|
78 |
+
2025-02-18 10:02:37,145 INFO: new_concept loss: 1.957642e-03
|
79 |
+
2025-02-18 10:02:37,145 INFO: [29/48] optimizing text_model.encoder.layers.7.self_attn.q_proj.weight
|
80 |
+
2025-02-18 10:02:39,995 INFO: new_concept loss: 1.165296e-02
|
81 |
+
2025-02-18 10:02:39,996 INFO: [30/48] optimizing text_model.encoder.layers.5.self_attn.q_proj.weight
|
82 |
+
2025-02-18 10:02:42,915 INFO: new_concept loss: 7.787009e-03
|
83 |
+
2025-02-18 10:02:42,916 INFO: [31/48] optimizing text_model.encoder.layers.2.self_attn.v_proj.weight
|
84 |
+
2025-02-18 10:02:45,746 INFO: new_concept loss: 1.318311e-02
|
85 |
+
2025-02-18 10:02:45,746 INFO: [32/48] optimizing text_model.encoder.layers.5.self_attn.k_proj.weight
|
86 |
+
2025-02-18 10:02:48,589 INFO: new_concept loss: 9.512860e-03
|
87 |
+
2025-02-18 10:02:48,590 INFO: [33/48] optimizing text_model.encoder.layers.6.self_attn.v_proj.weight
|
88 |
+
2025-02-18 10:02:51,572 INFO: new_concept loss: 1.058318e-02
|
89 |
+
2025-02-18 10:02:51,572 INFO: [34/48] optimizing text_model.encoder.layers.10.self_attn.out_proj.weight
|
90 |
+
2025-02-18 10:02:54,468 INFO: new_concept loss: 2.815726e-03
|
91 |
+
2025-02-18 10:02:54,469 INFO: [35/48] optimizing text_model.encoder.layers.10.self_attn.q_proj.weight
|
92 |
+
2025-02-18 10:02:57,526 INFO: new_concept loss: 8.984621e-03
|
93 |
+
2025-02-18 10:02:57,526 INFO: [36/48] optimizing text_model.encoder.layers.7.self_attn.v_proj.weight
|
94 |
+
2025-02-18 10:03:00,487 INFO: new_concept loss: 1.155254e-02
|
95 |
+
2025-02-18 10:03:00,487 INFO: [37/48] optimizing text_model.encoder.layers.8.self_attn.k_proj.weight
|
96 |
+
2025-02-18 10:03:03,208 INFO: new_concept loss: 1.002240e-02
|
97 |
+
2025-02-18 10:03:03,208 INFO: [38/48] optimizing text_model.encoder.layers.6.self_attn.q_proj.weight
|
98 |
+
2025-02-18 10:03:05,931 INFO: new_concept loss: 9.551291e-03
|
99 |
+
2025-02-18 10:03:05,932 INFO: [39/48] optimizing text_model.encoder.layers.1.self_attn.k_proj.weight
|
100 |
+
2025-02-18 10:03:08,649 INFO: new_concept loss: 2.144185e-02
|
101 |
+
2025-02-18 10:03:08,650 INFO: [40/48] optimizing text_model.encoder.layers.0.self_attn.out_proj.weight
|
102 |
+
2025-02-18 10:03:11,224 INFO: new_concept loss: 4.528776e-03
|
103 |
+
2025-02-18 10:03:11,224 INFO: [41/48] optimizing text_model.encoder.layers.2.self_attn.out_proj.weight
|
104 |
+
2025-02-18 10:03:13,646 INFO: new_concept loss: 2.965961e-03
|
105 |
+
2025-02-18 10:03:13,646 INFO: [42/48] optimizing text_model.encoder.layers.4.self_attn.out_proj.weight
|
106 |
+
2025-02-18 10:03:16,621 INFO: new_concept loss: 2.300192e-03
|
107 |
+
2025-02-18 10:03:16,621 INFO: [43/48] optimizing text_model.encoder.layers.11.self_attn.q_proj.weight
|
108 |
+
2025-02-18 10:03:19,347 INFO: new_concept loss: 1.009758e-02
|
109 |
+
2025-02-18 10:03:19,347 INFO: [44/48] optimizing text_model.encoder.layers.9.self_attn.v_proj.weight
|
110 |
+
2025-02-18 10:03:22,184 INFO: new_concept loss: 8.896633e-03
|
111 |
+
2025-02-18 10:03:22,184 INFO: [45/48] optimizing text_model.encoder.layers.0.self_attn.q_proj.weight
|
112 |
+
2025-02-18 10:03:22,510 INFO: new_concept loss: 4.487977e-02
|
113 |
+
2025-02-18 10:03:22,510 INFO: [46/48] optimizing text_model.encoder.layers.0.self_attn.v_proj.weight
|
114 |
+
2025-02-18 10:03:23,257 INFO: new_concept loss: 2.651430e-02
|
115 |
+
2025-02-18 10:03:23,257 INFO: [47/48] optimizing text_model.encoder.layers.1.self_attn.q_proj.weight
|
116 |
+
2025-02-18 10:03:26,068 INFO: new_concept loss: 1.082598e-02
|
117 |
+
2025-02-18 10:03:26,068 INFO: [48/48] optimizing text_model.encoder.layers.10.self_attn.v_proj.weight
|
118 |
+
2025-02-18 10:03:28,910 INFO: new_concept loss: 1.108888e-02
|
119 |
+
2025-02-18 10:03:28,911 INFO: remove 48 hooker from text_encoder
|
120 |
+
2025-02-18 10:03:28,947 INFO: ------Step 5: merge kv of cross-attention in unet------
|
121 |
+
2025-02-18 10:03:28,950 INFO: Unet have 32 linear layer (related to text feature) need to optimize
|
122 |
+
2025-02-18 10:03:30,392 INFO: [1/32] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k.weight
|
123 |
+
2025-02-18 10:03:31,275 INFO: new_concept loss: 4.427530e-05
|
124 |
+
2025-02-18 10:03:31,275 INFO: [2/32] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v.weight
|
125 |
+
2025-02-18 10:03:33,597 INFO: new_concept loss: 2.170046e-05
|
126 |
+
2025-02-18 10:03:33,597 INFO: [3/32] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k.weight
|
127 |
+
2025-02-18 10:03:35,695 INFO: new_concept loss: 2.655753e-05
|
128 |
+
2025-02-18 10:03:35,695 INFO: [4/32] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v.weight
|
129 |
+
2025-02-18 10:03:37,830 INFO: new_concept loss: 2.436045e-05
|
130 |
+
2025-02-18 10:03:37,830 INFO: [5/32] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight
|
131 |
+
2025-02-18 10:03:38,959 INFO: new_concept loss: 2.507451e-05
|
132 |
+
2025-02-18 10:03:38,959 INFO: [6/32] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight
|
133 |
+
2025-02-18 10:03:40,767 INFO: new_concept loss: 1.935738e-05
|
134 |
+
2025-02-18 10:03:40,768 INFO: [7/32] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight
|
135 |
+
2025-02-18 10:03:42,580 INFO: new_concept loss: 6.680089e-05
|
136 |
+
2025-02-18 10:03:42,580 INFO: [8/32] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight
|
137 |
+
2025-02-18 10:03:44,709 INFO: new_concept loss: 1.519853e-05
|
138 |
+
2025-02-18 10:03:44,710 INFO: [9/32] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight
|
139 |
+
2025-02-18 10:03:46,676 INFO: new_concept loss: 3.263517e-05
|
140 |
+
2025-02-18 10:03:46,677 INFO: [10/32] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight
|
141 |
+
2025-02-18 10:03:50,298 INFO: new_concept loss: 1.466394e-05
|
142 |
+
2025-02-18 10:03:50,299 INFO: [11/32] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight
|
143 |
+
2025-02-18 10:03:53,285 INFO: new_concept loss: 4.690791e-05
|
144 |
+
2025-02-18 10:03:53,285 INFO: [12/32] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight
|
145 |
+
2025-02-18 10:03:55,808 INFO: new_concept loss: 2.317002e-05
|
146 |
+
2025-02-18 10:03:55,808 INFO: [13/32] optimizing mid_block.attentions.0.transformer_blocks.0.attn2.to_k.weight
|
147 |
+
2025-02-18 10:03:58,787 INFO: new_concept loss: 2.382438e-05
|
148 |
+
2025-02-18 10:03:58,787 INFO: [14/32] optimizing mid_block.attentions.0.transformer_blocks.0.attn2.to_v.weight
|
149 |
+
2025-02-18 10:04:00,413 INFO: new_concept loss: 2.085945e-05
|
150 |
+
2025-02-18 10:04:00,413 INFO: [15/32] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight
|
151 |
+
2025-02-18 10:04:03,919 INFO: new_concept loss: 8.338261e-05
|
152 |
+
2025-02-18 10:04:03,919 INFO: [16/32] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight
|
153 |
+
2025-02-18 10:04:07,465 INFO: new_concept loss: 5.506125e-04
|
154 |
+
2025-02-18 10:04:07,465 INFO: [17/32] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight
|
155 |
+
2025-02-18 10:04:10,213 INFO: new_concept loss: 6.924390e-05
|
156 |
+
2025-02-18 10:04:10,214 INFO: [18/32] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight
|
157 |
+
2025-02-18 10:04:13,788 INFO: new_concept loss: 9.049270e-05
|
158 |
+
2025-02-18 10:04:13,789 INFO: [19/32] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k.weight
|
159 |
+
2025-02-18 10:04:17,357 INFO: new_concept loss: 2.786910e-04
|
160 |
+
2025-02-18 10:04:17,358 INFO: [20/32] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v.weight
|
161 |
+
2025-02-18 10:04:21,001 INFO: new_concept loss: 1.661567e-04
|
162 |
+
2025-02-18 10:04:21,001 INFO: [21/32] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight
|
163 |
+
2025-02-18 10:04:23,301 INFO: new_concept loss: 1.153841e-04
|
164 |
+
2025-02-18 10:04:23,301 INFO: [22/32] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight
|
165 |
+
2025-02-18 10:04:25,677 INFO: new_concept loss: 1.139642e-04
|
166 |
+
2025-02-18 10:04:25,677 INFO: [23/32] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight
|
167 |
+
2025-02-18 10:04:27,179 INFO: new_concept loss: 2.700988e-05
|
168 |
+
2025-02-18 10:04:27,180 INFO: [24/32] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight
|
169 |
+
2025-02-18 10:04:29,088 INFO: new_concept loss: 1.945955e-05
|
170 |
+
2025-02-18 10:04:29,088 INFO: [25/32] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k.weight
|
171 |
+
2025-02-18 10:04:30,532 INFO: new_concept loss: 6.514074e-05
|
172 |
+
2025-02-18 10:04:30,532 INFO: [26/32] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v.weight
|
173 |
+
2025-02-18 10:04:32,974 INFO: new_concept loss: 2.861637e-05
|
174 |
+
2025-02-18 10:04:32,975 INFO: [27/32] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k.weight
|
175 |
+
2025-02-18 10:04:33,947 INFO: new_concept loss: 2.005374e-05
|
176 |
+
2025-02-18 10:04:33,947 INFO: [28/32] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v.weight
|
177 |
+
2025-02-18 10:04:35,233 INFO: new_concept loss: 5.705929e-06
|
178 |
+
2025-02-18 10:04:35,233 INFO: [29/32] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k.weight
|
179 |
+
2025-02-18 10:04:37,089 INFO: new_concept loss: 3.739028e-05
|
180 |
+
2025-02-18 10:04:37,090 INFO: [30/32] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v.weight
|
181 |
+
2025-02-18 10:04:39,365 INFO: new_concept loss: 1.504165e-05
|
182 |
+
2025-02-18 10:04:39,365 INFO: [31/32] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k.weight
|
183 |
+
2025-02-18 10:04:40,939 INFO: new_concept loss: 4.594039e-05
|
184 |
+
2025-02-18 10:04:40,940 INFO: [32/32] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v.weight
|
185 |
+
2025-02-18 10:04:42,949 INFO: new_concept loss: 4.181080e-05
|
186 |
+
2025-02-18 10:04:42,984 INFO: ------Step 6: merge spatial attention (q in cross-attention, qkv in self-attention) in unet------
|
187 |
+
2025-02-18 10:04:42,985 INFO: unet have 96 linear layer need to optimize
|
188 |
+
2025-02-18 10:04:42,986 INFO: add 96 hooker to unet
|
189 |
+
2025-02-18 10:04:43,061 INFO: load 96 LoRAs of unet
|
190 |
+
2025-02-18 10:04:47,441 INFO: load 96 LoRAs of unet
|
191 |
+
2025-02-18 10:04:49,403 INFO: load 96 LoRAs of unet
|
192 |
+
2025-02-18 10:04:51,305 INFO: load 96 LoRAs of unet
|
193 |
+
2025-02-18 10:04:53,195 INFO: load 96 LoRAs of unet
|
194 |
+
2025-02-18 10:04:57,664 INFO: [1/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_q.weight
|
195 |
+
2025-02-18 10:05:05,839 INFO: new_concept loss: 1.593050e-01
|
196 |
+
2025-02-18 10:05:05,915 INFO: [2/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn2.to_q.weight
|
197 |
+
2025-02-18 10:05:06,917 INFO: new_concept loss: 1.371175e-01
|
198 |
+
2025-02-18 10:05:06,918 INFO: [3/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
|
199 |
+
2025-02-18 10:05:09,711 INFO: new_concept loss: 9.162319e-02
|
200 |
+
2025-02-18 10:05:09,731 INFO: [4/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k.weight
|
201 |
+
2025-02-18 10:05:13,925 INFO: new_concept loss: 6.829867e-01
|
202 |
+
2025-02-18 10:05:13,969 INFO: [5/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q.weight
|
203 |
+
2025-02-18 10:05:22,288 INFO: new_concept loss: 5.970761e-01
|
204 |
+
2025-02-18 10:05:22,368 INFO: [6/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn1.to_k.weight
|
205 |
+
2025-02-18 10:05:23,396 INFO: new_concept loss: 2.691664e-01
|
206 |
+
2025-02-18 10:05:23,397 INFO: [7/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight
|
207 |
+
2025-02-18 10:05:26,176 INFO: new_concept loss: 2.757275e-01
|
208 |
+
2025-02-18 10:05:26,197 INFO: [8/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
|
209 |
+
2025-02-18 10:05:28,994 INFO: new_concept loss: 2.462351e-01
|
210 |
+
2025-02-18 10:05:29,016 INFO: [9/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
|
211 |
+
2025-02-18 10:05:31,810 INFO: new_concept loss: 2.362856e-01
|
212 |
+
2025-02-18 10:05:31,833 INFO: [10/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
|
213 |
+
2025-02-18 10:05:39,922 INFO: new_concept loss: 5.065306e-02
|
214 |
+
2025-02-18 10:05:40,004 INFO: [11/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight
|
215 |
+
2025-02-18 10:05:42,750 INFO: new_concept loss: 1.462675e-01
|
216 |
+
2025-02-18 10:05:42,772 INFO: [12/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
|
217 |
+
2025-02-18 10:05:45,533 INFO: new_concept loss: 1.328609e-01
|
218 |
+
2025-02-18 10:05:45,556 INFO: [13/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_q.weight
|
219 |
+
2025-02-18 10:05:53,582 INFO: new_concept loss: 7.257976e-02
|
220 |
+
2025-02-18 10:05:53,667 INFO: [14/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
|
221 |
+
2025-02-18 10:06:01,915 INFO: new_concept loss: 3.954707e-02
|
222 |
+
2025-02-18 10:06:01,997 INFO: [15/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_v.weight
|
223 |
+
2025-02-18 10:06:10,340 INFO: new_concept loss: 1.240187e-01
|
224 |
+
2025-02-18 10:06:10,422 INFO: [16/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out.0.weight
|
225 |
+
2025-02-18 10:06:14,620 INFO: new_concept loss: 1.048966e-01
|
226 |
+
2025-02-18 10:06:14,663 INFO: [17/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q.weight
|
227 |
+
2025-02-18 10:06:18,971 INFO: new_concept loss: 4.386465e-01
|
228 |
+
2025-02-18 10:06:19,017 INFO: [18/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight
|
229 |
+
2025-02-18 10:06:21,835 INFO: new_concept loss: 7.685150e-01
|
230 |
+
2025-02-18 10:06:21,859 INFO: [19/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q.weight
|
231 |
+
2025-02-18 10:06:30,067 INFO: new_concept loss: 6.341255e-02
|
232 |
+
2025-02-18 10:06:30,153 INFO: [20/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k.weight
|
233 |
+
2025-02-18 10:06:32,957 INFO: new_concept loss: 7.451130e-01
|
234 |
+
2025-02-18 10:06:32,980 INFO: [21/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
|
235 |
+
2025-02-18 10:06:37,251 INFO: new_concept loss: 1.556838e-01
|
236 |
+
2025-02-18 10:06:37,298 INFO: [22/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_q.weight
|
237 |
+
2025-02-18 10:06:45,501 INFO: new_concept loss: 1.710174e-01
|
238 |
+
2025-02-18 10:06:45,589 INFO: [23/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k.weight
|
239 |
+
2025-02-18 10:06:53,892 INFO: new_concept loss: 3.234306e-01
|
240 |
+
2025-02-18 10:06:53,982 INFO: [24/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
|
241 |
+
2025-02-18 10:06:56,738 INFO: new_concept loss: 1.068963e-01
|
242 |
+
2025-02-18 10:06:56,760 INFO: [25/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight
|
243 |
+
2025-02-18 10:07:01,106 INFO: new_concept loss: 7.557544e-01
|
244 |
+
2025-02-18 10:07:01,155 INFO: [26/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k.weight
|
245 |
+
2025-02-18 10:07:03,984 INFO: new_concept loss: 7.542757e-01
|
246 |
+
2025-02-18 10:07:04,009 INFO: [27/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q.weight
|
247 |
+
2025-02-18 10:07:12,461 INFO: new_concept loss: 2.984946e-01
|
248 |
+
2025-02-18 10:07:12,555 INFO: [28/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out.0.weight
|
249 |
+
2025-02-18 10:07:16,900 INFO: new_concept loss: 1.445462e-01
|
250 |
+
2025-02-18 10:07:16,949 INFO: [29/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight
|
251 |
+
2025-02-18 10:07:21,145 INFO: new_concept loss: 5.463246e-01
|
252 |
+
2025-02-18 10:07:21,194 INFO: [30/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v.weight
|
253 |
+
2025-02-18 10:07:25,358 INFO: new_concept loss: 2.672614e-01
|
254 |
+
2025-02-18 10:07:25,408 INFO: [31/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_v.weight
|
255 |
+
2025-02-18 10:07:33,601 INFO: new_concept loss: 8.787098e-02
|
256 |
+
2025-02-18 10:07:33,698 INFO: [32/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k.weight
|
257 |
+
2025-02-18 10:07:42,035 INFO: new_concept loss: 2.288823e-01
|
258 |
+
2025-02-18 10:07:42,131 INFO: [33/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q.weight
|
259 |
+
2025-02-18 10:07:44,988 INFO: new_concept loss: 4.186683e-01
|
260 |
+
2025-02-18 10:07:45,011 INFO: [34/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
|
261 |
+
2025-02-18 10:07:53,232 INFO: new_concept loss: 6.477418e-02
|
262 |
+
2025-02-18 10:07:53,332 INFO: [35/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out.0.weight
|
263 |
+
2025-02-18 10:08:01,636 INFO: new_concept loss: 2.791651e-02
|
264 |
+
2025-02-18 10:08:01,733 INFO: [36/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out.0.weight
|
265 |
+
2025-02-18 10:08:09,930 INFO: new_concept loss: 2.724535e-02
|
266 |
+
2025-02-18 10:08:10,024 INFO: [37/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
|
267 |
+
2025-02-18 10:08:14,286 INFO: new_concept loss: 1.612051e-01
|
268 |
+
2025-02-18 10:08:14,332 INFO: [38/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q.weight
|
269 |
+
2025-02-18 10:08:18,730 INFO: new_concept loss: 1.113115e-01
|
270 |
+
2025-02-18 10:08:18,779 INFO: [39/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v.weight
|
271 |
+
2025-02-18 10:08:21,522 INFO: new_concept loss: 2.153144e-01
|
272 |
+
2025-02-18 10:08:21,545 INFO: [40/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight
|
273 |
+
2025-02-18 10:08:24,325 INFO: new_concept loss: 4.681203e-01
|
274 |
+
2025-02-18 10:08:24,350 INFO: [41/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_k.weight
|
275 |
+
2025-02-18 10:08:32,284 INFO: new_concept loss: 1.354147e-01
|
276 |
+
2025-02-18 10:08:32,357 INFO: [42/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q.weight
|
277 |
+
2025-02-18 10:08:36,701 INFO: new_concept loss: 2.082129e-01
|
278 |
+
2025-02-18 10:08:36,743 INFO: [43/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q.weight
|
279 |
+
2025-02-18 10:08:45,053 INFO: new_concept loss: 6.789617e-02
|
280 |
+
2025-02-18 10:08:45,130 INFO: [44/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_q.weight
|
281 |
+
2025-02-18 10:08:53,543 INFO: new_concept loss: 6.113157e-02
|
282 |
+
2025-02-18 10:08:53,623 INFO: [45/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight
|
283 |
+
2025-02-18 10:08:57,785 INFO: new_concept loss: 1.470006e-01
|
284 |
+
2025-02-18 10:08:57,831 INFO: [46/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight
|
285 |
+
2025-02-18 10:09:02,148 INFO: new_concept loss: 7.435892e-01
|
286 |
+
2025-02-18 10:09:02,195 INFO: [47/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight
|
287 |
+
2025-02-18 10:09:05,011 INFO: new_concept loss: 2.135441e-01
|
288 |
+
2025-02-18 10:09:05,035 INFO: [48/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v.weight
|
289 |
+
2025-02-18 10:09:13,424 INFO: new_concept loss: 1.767161e-01
|
290 |
+
2025-02-18 10:09:13,508 INFO: [49/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
|
291 |
+
2025-02-18 10:09:17,712 INFO: new_concept loss: 9.232481e-02
|
292 |
+
2025-02-18 10:09:17,759 INFO: [50/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
|
293 |
+
2025-02-18 10:09:18,797 INFO: new_concept loss: 1.938113e-01
|
294 |
+
2025-02-18 10:09:18,798 INFO: [51/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight
|
295 |
+
2025-02-18 10:09:21,654 INFO: new_concept loss: 1.931212e-01
|
296 |
+
2025-02-18 10:09:21,680 INFO: [52/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight
|
297 |
+
2025-02-18 10:09:25,985 INFO: new_concept loss: 6.031875e-01
|
298 |
+
2025-02-18 10:09:26,030 INFO: [53/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
|
299 |
+
2025-02-18 10:09:30,128 INFO: new_concept loss: 6.304939e-02
|
300 |
+
2025-02-18 10:09:30,180 INFO: [54/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight
|
301 |
+
2025-02-18 10:09:32,960 INFO: new_concept loss: 3.609971e-01
|
302 |
+
2025-02-18 10:09:32,985 INFO: [55/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
|
303 |
+
2025-02-18 10:09:37,153 INFO: new_concept loss: 6.763718e-01
|
304 |
+
2025-02-18 10:09:37,205 INFO: [56/96] optimizing up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q.weight
|
305 |
+
2025-02-18 10:09:41,528 INFO: new_concept loss: 3.740048e-01
|
306 |
+
2025-02-18 10:09:41,577 INFO: [57/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k.weight
|
307 |
+
2025-02-18 10:09:45,957 INFO: new_concept loss: 2.385424e-01
|
308 |
+
2025-02-18 10:09:46,006 INFO: [58/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
|
309 |
+
2025-02-18 10:09:54,013 INFO: new_concept loss: 8.359801e-02
|
310 |
+
2025-02-18 10:09:54,111 INFO: [59/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k.weight
|
311 |
+
2025-02-18 10:10:02,516 INFO: new_concept loss: 2.154270e-01
|
312 |
+
2025-02-18 10:10:02,621 INFO: [60/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn1.to_v.weight
|
313 |
+
2025-02-18 10:10:03,635 INFO: new_concept loss: 1.350740e-01
|
314 |
+
2025-02-18 10:10:03,636 INFO: [61/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
|
315 |
+
2025-02-18 10:10:04,650 INFO: new_concept loss: 7.636815e-02
|
316 |
+
2025-02-18 10:10:04,651 INFO: [62/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight
|
317 |
+
2025-02-18 10:10:08,821 INFO: new_concept loss: 2.006273e-01
|
318 |
+
2025-02-18 10:10:08,867 INFO: [63/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_k.weight
|
319 |
+
2025-02-18 10:10:16,971 INFO: new_concept loss: 1.642606e-01
|
320 |
+
2025-02-18 10:10:17,059 INFO: [64/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q.weight
|
321 |
+
2025-02-18 10:10:19,851 INFO: new_concept loss: 2.283725e-01
|
322 |
+
2025-02-18 10:10:19,873 INFO: [65/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q.weight
|
323 |
+
2025-02-18 10:10:22,741 INFO: new_concept loss: 2.470787e-01
|
324 |
+
2025-02-18 10:10:22,768 INFO: [66/96] optimizing mid_block.attentions.0.transformer_blocks.0.attn1.to_q.weight
|
325 |
+
2025-02-18 10:10:23,788 INFO: new_concept loss: 1.714881e-01
|
326 |
+
2025-02-18 10:10:23,788 INFO: [67/96] optimizing down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
|
327 |
+
2025-02-18 10:10:31,935 INFO: new_concept loss: 2.508941e-02
|
328 |
+
2025-02-18 10:10:32,031 INFO: [68/96] optimizing up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
|
329 |
+
2025-02-18 10:10:34,881 INFO: new_concept loss: 1.466412e-01
|
330 |
+
2025-02-18 10:10:34,905 INFO: [69/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0.weight
|
331 |
+
2025-02-18 10:10:37,773 INFO: new_concept loss: 1.901781e-01
|
332 |
+
2025-02-18 10:10:37,800 INFO: [70/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight
|
333 |
+
2025-02-18 10:10:42,327 INFO: new_concept loss: 3.197155e-01
|
334 |
+
2025-02-18 10:10:42,381 INFO: [71/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q.weight
|
335 |
+
2025-02-18 10:10:45,243 INFO: new_concept loss: 3.977647e-01
|
336 |
+
2025-02-18 10:10:45,267 INFO: [72/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight
|
337 |
+
2025-02-18 10:10:48,090 INFO: new_concept loss: 5.833027e-01
|
338 |
+
2025-02-18 10:10:48,118 INFO: [73/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v.weight
|
339 |
+
2025-02-18 10:10:50,951 INFO: new_concept loss: 2.237324e-01
|
340 |
+
2025-02-18 10:10:50,978 INFO: [74/96] optimizing up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
|
341 |
+
2025-02-18 10:10:55,080 INFO: new_concept loss: 4.140265e-01
|
342 |
+
2025-02-18 10:10:55,131 INFO: [75/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q.weight
|
343 |
+
2025-02-18 10:11:03,581 INFO: new_concept loss: 2.084374e-01
|
344 |
+
2025-02-18 10:11:03,675 INFO: [76/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight
|
345 |
+
2025-02-18 10:11:06,508 INFO: new_concept loss: 1.883232e-01
|
346 |
+
2025-02-18 10:11:06,534 INFO: [77/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight
|
347 |
+
2025-02-18 10:11:10,817 INFO: new_concept loss: 4.924987e-01
|
348 |
+
2025-02-18 10:11:10,869 INFO: [78/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q.weight
|
349 |
+
2025-02-18 10:11:18,874 INFO: new_concept loss: 9.835833e-02
|
350 |
+
2025-02-18 10:11:18,977 INFO: [79/96] optimizing down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v.weight
|
351 |
+
2025-02-18 10:11:23,223 INFO: new_concept loss: 1.199531e-01
|
352 |
+
2025-02-18 10:11:23,273 INFO: [80/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight
|
353 |
+
2025-02-18 10:11:27,451 INFO: new_concept loss: 2.096862e-01
|
354 |
+
2025-02-18 10:11:27,502 INFO: [81/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
|
355 |
+
2025-02-18 10:11:30,343 INFO: new_concept loss: 1.246875e-01
|
356 |
+
2025-02-18 10:11:30,369 INFO: [82/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight
|
357 |
+
2025-02-18 10:11:33,070 INFO: new_concept loss: 2.763811e-01
|
358 |
+
2025-02-18 10:11:33,096 INFO: [83/96] optimizing up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight
|
359 |
+
2025-02-18 10:11:35,921 INFO: new_concept loss: 3.674817e-01
|
360 |
+
2025-02-18 10:11:35,947 INFO: [84/96] optimizing up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
|
361 |
+
2025-02-18 10:11:44,012 INFO: new_concept loss: 3.000834e-02
|
362 |
+
2025-02-18 10:11:44,112 INFO: [85/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight
|
363 |
+
2025-02-18 10:11:48,387 INFO: new_concept loss: 1.812962e-02
|
364 |
+
2025-02-18 10:11:48,436 INFO: [86/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
|
365 |
+
2025-02-18 10:11:52,674 INFO: new_concept loss: 7.571394e-01
|
366 |
+
2025-02-18 10:11:52,725 INFO: [87/96] optimizing down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight
|
367 |
+
2025-02-18 10:11:56,823 INFO: new_concept loss: 1.991518e-01
|
368 |
+
2025-02-18 10:11:56,873 INFO: [88/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight
|
369 |
+
2025-02-18 10:12:01,104 INFO: new_concept loss: 3.178474e-01
|
370 |
+
2025-02-18 10:12:01,155 INFO: [89/96] optimizing down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight
|
371 |
+
2025-02-18 10:12:03,957 INFO: new_concept loss: 1.593530e-01
|
372 |
+
2025-02-18 10:12:03,980 INFO: [90/96] optimizing up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0.weight
|
373 |
+
2025-02-18 10:12:06,736 INFO: new_concept loss: 1.556669e-01
|
374 |
+
2025-02-18 10:12:06,755 INFO: [91/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v.weight
|
375 |
+
2025-02-18 10:12:15,035 INFO: new_concept loss: 3.353187e-01
|
376 |
+
2025-02-18 10:12:15,111 INFO: [92/96] optimizing down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0.weight
|
377 |
+
2025-02-18 10:12:23,691 INFO: new_concept loss: 6.571595e-03
|
378 |
+
2025-02-18 10:12:23,771 INFO: [93/96] optimizing down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight
|
379 |
+
2025-02-18 10:12:26,686 INFO: new_concept loss: 3.799535e-01
|
380 |
+
2025-02-18 10:12:26,707 INFO: [94/96] optimizing up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v.weight
|
381 |
+
2025-02-18 10:12:34,861 INFO: new_concept loss: 1.101350e-01
|
382 |
+
2025-02-18 10:12:34,941 INFO: [95/96] optimizing up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out.0.weight
|
383 |
+
2025-02-18 10:12:43,608 INFO: new_concept loss: 1.892852e-01
|
384 |
+
2025-02-18 10:12:43,701 INFO: [96/96] optimizing up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight
|
385 |
+
2025-02-18 10:12:48,229 INFO: new_concept loss: 5.540602e-01
|
386 |
+
2025-02-18 10:12:48,276 INFO: remove 96 hooker from unet
|
FrozenII/combined_model_base/feature_extractor/preprocessor_config.json
ADDED
@@ -0,0 +1,28 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"crop_size": {
|
3 |
+
"height": 224,
|
4 |
+
"width": 224
|
5 |
+
},
|
6 |
+
"do_center_crop": true,
|
7 |
+
"do_convert_rgb": true,
|
8 |
+
"do_normalize": true,
|
9 |
+
"do_rescale": true,
|
10 |
+
"do_resize": true,
|
11 |
+
"feature_extractor_type": "CLIPFeatureExtractor",
|
12 |
+
"image_mean": [
|
13 |
+
0.48145466,
|
14 |
+
0.4578275,
|
15 |
+
0.40821073
|
16 |
+
],
|
17 |
+
"image_processor_type": "CLIPFeatureExtractor",
|
18 |
+
"image_std": [
|
19 |
+
0.26862954,
|
20 |
+
0.26130258,
|
21 |
+
0.27577711
|
22 |
+
],
|
23 |
+
"resample": 3,
|
24 |
+
"rescale_factor": 0.00392156862745098,
|
25 |
+
"size": {
|
26 |
+
"shortest_edge": 224
|
27 |
+
}
|
28 |
+
}
|
FrozenII/combined_model_base/model_index.json
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.20.0",
|
4 |
+
"_name_or_path": "experiments/pretrained_models/chilloutmix",
|
5 |
+
"feature_extractor": [
|
6 |
+
"transformers",
|
7 |
+
"CLIPFeatureExtractor"
|
8 |
+
],
|
9 |
+
"requires_safety_checker": true,
|
10 |
+
"safety_checker": [
|
11 |
+
null,
|
12 |
+
null
|
13 |
+
],
|
14 |
+
"scheduler": [
|
15 |
+
"diffusers",
|
16 |
+
"DPMSolverMultistepScheduler"
|
17 |
+
],
|
18 |
+
"text_encoder": [
|
19 |
+
"transformers",
|
20 |
+
"CLIPTextModel"
|
21 |
+
],
|
22 |
+
"tokenizer": [
|
23 |
+
"transformers",
|
24 |
+
"CLIPTokenizer"
|
25 |
+
],
|
26 |
+
"unet": [
|
27 |
+
"diffusers",
|
28 |
+
"UNet2DConditionModel"
|
29 |
+
],
|
30 |
+
"vae": [
|
31 |
+
"diffusers",
|
32 |
+
"AutoencoderKL"
|
33 |
+
]
|
34 |
+
}
|
FrozenII/combined_model_base/new_concept_cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"<Elsa1>": {"concept_token_ids": [49408, 49409, 49410, 49411, 49412, 49413, 49414, 49415, 49416, 49417, 49418, 49419, 49420, 49421, 49422, 49423], "concept_token_names": ["<new0>", "<new1>", "<new2>", "<new3>", "<new4>", "<new5>", "<new6>", "<new7>", "<new8>", "<new9>", "<new10>", "<new11>", "<new12>", "<new13>", "<new14>", "<new15>"]}, "<Elsa2>": {"concept_token_ids": [49424, 49425, 49426, 49427, 49428, 49429, 49430, 49431, 49432, 49433, 49434, 49435, 49436, 49437, 49438, 49439], "concept_token_names": ["<new16>", "<new17>", "<new18>", "<new19>", "<new20>", "<new21>", "<new22>", "<new23>", "<new24>", "<new25>", "<new26>", "<new27>", "<new28>", "<new29>", "<new30>", "<new31>"]}, "<Olaf1>": {"concept_token_ids": [49440, 49441, 49442, 49443, 49444, 49445, 49446, 49447, 49448, 49449, 49450, 49451, 49452, 49453, 49454, 49455], "concept_token_names": ["<new32>", "<new33>", "<new34>", "<new35>", "<new36>", "<new37>", "<new38>", "<new39>", "<new40>", "<new41>", "<new42>", "<new43>", "<new44>", "<new45>", "<new46>", "<new47>"]}, "<Olaf2>": {"concept_token_ids": [49456, 49457, 49458, 49459, 49460, 49461, 49462, 49463, 49464, 49465, 49466, 49467, 49468, 49469, 49470, 49471], "concept_token_names": ["<new48>", "<new49>", "<new50>", "<new51>", "<new52>", "<new53>", "<new54>", "<new55>", "<new56>", "<new57>", "<new58>", "<new59>", "<new60>", "<new61>", "<new62>", "<new63>"]}, "<Mattias1>": {"concept_token_ids": [49472, 49473, 49474, 49475, 49476, 49477, 49478, 49479, 49480, 49481, 49482, 49483, 49484, 49485, 49486, 49487], "concept_token_names": ["<new64>", "<new65>", "<new66>", "<new67>", "<new68>", "<new69>", "<new70>", "<new71>", "<new72>", "<new73>", "<new74>", "<new75>", "<new76>", "<new77>", "<new78>", "<new79>"]}, "<Mattias2>": {"concept_token_ids": [49488, 49489, 49490, 49491, 49492, 49493, 49494, 49495, 49496, 49497, 49498, 49499, 49500, 49501, 49502, 49503], "concept_token_names": ["<new80>", "<new81>", "<new82>", "<new83>", "<new84>", "<new85>", "<new86>", "<new87>", "<new88>", "<new89>", "<new90>", "<new91>", "<new92>", "<new93>", "<new94>", "<new95>"]}, "<Kristoff1>": {"concept_token_ids": [49504, 49505, 49506, 49507, 49508, 49509, 49510, 49511, 49512, 49513, 49514, 49515, 49516, 49517, 49518, 49519], "concept_token_names": ["<new96>", "<new97>", "<new98>", "<new99>", "<new100>", "<new101>", "<new102>", "<new103>", "<new104>", "<new105>", "<new106>", "<new107>", "<new108>", "<new109>", "<new110>", "<new111>"]}, "<Kristoff2>": {"concept_token_ids": [49520, 49521, 49522, 49523, 49524, 49525, 49526, 49527, 49528, 49529, 49530, 49531, 49532, 49533, 49534, 49535], "concept_token_names": ["<new112>", "<new113>", "<new114>", "<new115>", "<new116>", "<new117>", "<new118>", "<new119>", "<new120>", "<new121>", "<new122>", "<new123>", "<new124>", "<new125>", "<new126>", "<new127>"]}, "<Anna1>": {"concept_token_ids": [49536, 49537, 49538, 49539, 49540, 49541, 49542, 49543, 49544, 49545, 49546, 49547, 49548, 49549, 49550, 49551], "concept_token_names": ["<new128>", "<new129>", "<new130>", "<new131>", "<new132>", "<new133>", "<new134>", "<new135>", "<new136>", "<new137>", "<new138>", "<new139>", "<new140>", "<new141>", "<new142>", "<new143>"]}, "<Anna2>": {"concept_token_ids": [49552, 49553, 49554, 49555, 49556, 49557, 49558, 49559, 49560, 49561, 49562, 49563, 49564, 49565, 49566, 49567], "concept_token_names": ["<new144>", "<new145>", "<new146>", "<new147>", "<new148>", "<new149>", "<new150>", "<new151>", "<new152>", "<new153>", "<new154>", "<new155>", "<new156>", "<new157>", "<new158>", "<new159>"]}}
|
FrozenII/combined_model_base/scheduler/scheduler_config.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "DPMSolverMultistepScheduler",
|
3 |
+
"_diffusers_version": "0.20.0",
|
4 |
+
"algorithm_type": "dpmsolver++",
|
5 |
+
"beta_end": 0.012,
|
6 |
+
"beta_schedule": "scaled_linear",
|
7 |
+
"beta_start": 0.00085,
|
8 |
+
"clip_sample": false,
|
9 |
+
"dynamic_thresholding_ratio": 0.995,
|
10 |
+
"lambda_min_clipped": -Infinity,
|
11 |
+
"lower_order_final": true,
|
12 |
+
"num_train_timesteps": 1000,
|
13 |
+
"prediction_type": "epsilon",
|
14 |
+
"sample_max_value": 1.0,
|
15 |
+
"set_alpha_to_one": false,
|
16 |
+
"skip_prk_steps": true,
|
17 |
+
"solver_order": 2,
|
18 |
+
"solver_type": "midpoint",
|
19 |
+
"steps_offset": 1,
|
20 |
+
"thresholding": false,
|
21 |
+
"timestep_spacing": "linspace",
|
22 |
+
"trained_betas": null,
|
23 |
+
"use_karras_sigmas": false,
|
24 |
+
"variance_type": null
|
25 |
+
}
|
FrozenII/combined_model_base/text_encoder/config.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "experiments/pretrained_models/chilloutmix/text_encoder",
|
3 |
+
"architectures": [
|
4 |
+
"CLIPTextModel"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"dropout": 0.0,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "quick_gelu",
|
11 |
+
"hidden_size": 768,
|
12 |
+
"initializer_factor": 1.0,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 77,
|
17 |
+
"model_type": "clip_text_model",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 12,
|
20 |
+
"pad_token_id": 1,
|
21 |
+
"projection_dim": 768,
|
22 |
+
"torch_dtype": "float16",
|
23 |
+
"transformers_version": "4.28.0",
|
24 |
+
"vocab_size": 49568
|
25 |
+
}
|
FrozenII/combined_model_base/text_encoder/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4b0fe01461cd3e5445bd05cb66d6f0d58747dc9f144428da8d225bd2d1b00741
|
3 |
+
size 246390624
|
FrozenII/combined_model_base/tokenizer/added_tokens.json
ADDED
@@ -0,0 +1,162 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"<new0>": 49408,
|
3 |
+
"<new100>": 49508,
|
4 |
+
"<new101>": 49509,
|
5 |
+
"<new102>": 49510,
|
6 |
+
"<new103>": 49511,
|
7 |
+
"<new104>": 49512,
|
8 |
+
"<new105>": 49513,
|
9 |
+
"<new106>": 49514,
|
10 |
+
"<new107>": 49515,
|
11 |
+
"<new108>": 49516,
|
12 |
+
"<new109>": 49517,
|
13 |
+
"<new10>": 49418,
|
14 |
+
"<new110>": 49518,
|
15 |
+
"<new111>": 49519,
|
16 |
+
"<new112>": 49520,
|
17 |
+
"<new113>": 49521,
|
18 |
+
"<new114>": 49522,
|
19 |
+
"<new115>": 49523,
|
20 |
+
"<new116>": 49524,
|
21 |
+
"<new117>": 49525,
|
22 |
+
"<new118>": 49526,
|
23 |
+
"<new119>": 49527,
|
24 |
+
"<new11>": 49419,
|
25 |
+
"<new120>": 49528,
|
26 |
+
"<new121>": 49529,
|
27 |
+
"<new122>": 49530,
|
28 |
+
"<new123>": 49531,
|
29 |
+
"<new124>": 49532,
|
30 |
+
"<new125>": 49533,
|
31 |
+
"<new126>": 49534,
|
32 |
+
"<new127>": 49535,
|
33 |
+
"<new128>": 49536,
|
34 |
+
"<new129>": 49537,
|
35 |
+
"<new12>": 49420,
|
36 |
+
"<new130>": 49538,
|
37 |
+
"<new131>": 49539,
|
38 |
+
"<new132>": 49540,
|
39 |
+
"<new133>": 49541,
|
40 |
+
"<new134>": 49542,
|
41 |
+
"<new135>": 49543,
|
42 |
+
"<new136>": 49544,
|
43 |
+
"<new137>": 49545,
|
44 |
+
"<new138>": 49546,
|
45 |
+
"<new139>": 49547,
|
46 |
+
"<new13>": 49421,
|
47 |
+
"<new140>": 49548,
|
48 |
+
"<new141>": 49549,
|
49 |
+
"<new142>": 49550,
|
50 |
+
"<new143>": 49551,
|
51 |
+
"<new144>": 49552,
|
52 |
+
"<new145>": 49553,
|
53 |
+
"<new146>": 49554,
|
54 |
+
"<new147>": 49555,
|
55 |
+
"<new148>": 49556,
|
56 |
+
"<new149>": 49557,
|
57 |
+
"<new14>": 49422,
|
58 |
+
"<new150>": 49558,
|
59 |
+
"<new151>": 49559,
|
60 |
+
"<new152>": 49560,
|
61 |
+
"<new153>": 49561,
|
62 |
+
"<new154>": 49562,
|
63 |
+
"<new155>": 49563,
|
64 |
+
"<new156>": 49564,
|
65 |
+
"<new157>": 49565,
|
66 |
+
"<new158>": 49566,
|
67 |
+
"<new159>": 49567,
|
68 |
+
"<new15>": 49423,
|
69 |
+
"<new16>": 49424,
|
70 |
+
"<new17>": 49425,
|
71 |
+
"<new18>": 49426,
|
72 |
+
"<new19>": 49427,
|
73 |
+
"<new1>": 49409,
|
74 |
+
"<new20>": 49428,
|
75 |
+
"<new21>": 49429,
|
76 |
+
"<new22>": 49430,
|
77 |
+
"<new23>": 49431,
|
78 |
+
"<new24>": 49432,
|
79 |
+
"<new25>": 49433,
|
80 |
+
"<new26>": 49434,
|
81 |
+
"<new27>": 49435,
|
82 |
+
"<new28>": 49436,
|
83 |
+
"<new29>": 49437,
|
84 |
+
"<new2>": 49410,
|
85 |
+
"<new30>": 49438,
|
86 |
+
"<new31>": 49439,
|
87 |
+
"<new32>": 49440,
|
88 |
+
"<new33>": 49441,
|
89 |
+
"<new34>": 49442,
|
90 |
+
"<new35>": 49443,
|
91 |
+
"<new36>": 49444,
|
92 |
+
"<new37>": 49445,
|
93 |
+
"<new38>": 49446,
|
94 |
+
"<new39>": 49447,
|
95 |
+
"<new3>": 49411,
|
96 |
+
"<new40>": 49448,
|
97 |
+
"<new41>": 49449,
|
98 |
+
"<new42>": 49450,
|
99 |
+
"<new43>": 49451,
|
100 |
+
"<new44>": 49452,
|
101 |
+
"<new45>": 49453,
|
102 |
+
"<new46>": 49454,
|
103 |
+
"<new47>": 49455,
|
104 |
+
"<new48>": 49456,
|
105 |
+
"<new49>": 49457,
|
106 |
+
"<new4>": 49412,
|
107 |
+
"<new50>": 49458,
|
108 |
+
"<new51>": 49459,
|
109 |
+
"<new52>": 49460,
|
110 |
+
"<new53>": 49461,
|
111 |
+
"<new54>": 49462,
|
112 |
+
"<new55>": 49463,
|
113 |
+
"<new56>": 49464,
|
114 |
+
"<new57>": 49465,
|
115 |
+
"<new58>": 49466,
|
116 |
+
"<new59>": 49467,
|
117 |
+
"<new5>": 49413,
|
118 |
+
"<new60>": 49468,
|
119 |
+
"<new61>": 49469,
|
120 |
+
"<new62>": 49470,
|
121 |
+
"<new63>": 49471,
|
122 |
+
"<new64>": 49472,
|
123 |
+
"<new65>": 49473,
|
124 |
+
"<new66>": 49474,
|
125 |
+
"<new67>": 49475,
|
126 |
+
"<new68>": 49476,
|
127 |
+
"<new69>": 49477,
|
128 |
+
"<new6>": 49414,
|
129 |
+
"<new70>": 49478,
|
130 |
+
"<new71>": 49479,
|
131 |
+
"<new72>": 49480,
|
132 |
+
"<new73>": 49481,
|
133 |
+
"<new74>": 49482,
|
134 |
+
"<new75>": 49483,
|
135 |
+
"<new76>": 49484,
|
136 |
+
"<new77>": 49485,
|
137 |
+
"<new78>": 49486,
|
138 |
+
"<new79>": 49487,
|
139 |
+
"<new7>": 49415,
|
140 |
+
"<new80>": 49488,
|
141 |
+
"<new81>": 49489,
|
142 |
+
"<new82>": 49490,
|
143 |
+
"<new83>": 49491,
|
144 |
+
"<new84>": 49492,
|
145 |
+
"<new85>": 49493,
|
146 |
+
"<new86>": 49494,
|
147 |
+
"<new87>": 49495,
|
148 |
+
"<new88>": 49496,
|
149 |
+
"<new89>": 49497,
|
150 |
+
"<new8>": 49416,
|
151 |
+
"<new90>": 49498,
|
152 |
+
"<new91>": 49499,
|
153 |
+
"<new92>": 49500,
|
154 |
+
"<new93>": 49501,
|
155 |
+
"<new94>": 49502,
|
156 |
+
"<new95>": 49503,
|
157 |
+
"<new96>": 49504,
|
158 |
+
"<new97>": 49505,
|
159 |
+
"<new98>": 49506,
|
160 |
+
"<new99>": 49507,
|
161 |
+
"<new9>": 49417
|
162 |
+
}
|
FrozenII/combined_model_base/tokenizer/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
FrozenII/combined_model_base/tokenizer/special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<|startoftext|>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": true,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "<|endoftext|>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": true,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "<|endoftext|>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<|endoftext|>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": true,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
FrozenII/combined_model_base/tokenizer/tokenizer_config.json
ADDED
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"bos_token": {
|
4 |
+
"__type": "AddedToken",
|
5 |
+
"content": "<|startoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false
|
10 |
+
},
|
11 |
+
"clean_up_tokenization_spaces": true,
|
12 |
+
"do_lower_case": true,
|
13 |
+
"eos_token": {
|
14 |
+
"__type": "AddedToken",
|
15 |
+
"content": "<|endoftext|>",
|
16 |
+
"lstrip": false,
|
17 |
+
"normalized": true,
|
18 |
+
"rstrip": false,
|
19 |
+
"single_word": false
|
20 |
+
},
|
21 |
+
"errors": "replace",
|
22 |
+
"model_max_length": 77,
|
23 |
+
"pad_token": "<|endoftext|>",
|
24 |
+
"tokenizer_class": "CLIPTokenizer",
|
25 |
+
"unk_token": {
|
26 |
+
"__type": "AddedToken",
|
27 |
+
"content": "<|endoftext|>",
|
28 |
+
"lstrip": false,
|
29 |
+
"normalized": true,
|
30 |
+
"rstrip": false,
|
31 |
+
"single_word": false
|
32 |
+
}
|
33 |
+
}
|
FrozenII/combined_model_base/tokenizer/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
FrozenII/combined_model_base/unet/config.json
ADDED
@@ -0,0 +1,66 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "UNet2DConditionModel",
|
3 |
+
"_diffusers_version": "0.20.0",
|
4 |
+
"_name_or_path": "experiments/pretrained_models/chilloutmix/unet",
|
5 |
+
"act_fn": "silu",
|
6 |
+
"addition_embed_type": null,
|
7 |
+
"addition_embed_type_num_heads": 64,
|
8 |
+
"addition_time_embed_dim": null,
|
9 |
+
"attention_head_dim": 8,
|
10 |
+
"attention_type": "default",
|
11 |
+
"block_out_channels": [
|
12 |
+
320,
|
13 |
+
640,
|
14 |
+
1280,
|
15 |
+
1280
|
16 |
+
],
|
17 |
+
"center_input_sample": false,
|
18 |
+
"class_embed_type": null,
|
19 |
+
"class_embeddings_concat": false,
|
20 |
+
"conv_in_kernel": 3,
|
21 |
+
"conv_out_kernel": 3,
|
22 |
+
"cross_attention_dim": 768,
|
23 |
+
"cross_attention_norm": null,
|
24 |
+
"down_block_types": [
|
25 |
+
"CrossAttnDownBlock2D",
|
26 |
+
"CrossAttnDownBlock2D",
|
27 |
+
"CrossAttnDownBlock2D",
|
28 |
+
"DownBlock2D"
|
29 |
+
],
|
30 |
+
"downsample_padding": 1,
|
31 |
+
"dual_cross_attention": false,
|
32 |
+
"encoder_hid_dim": null,
|
33 |
+
"encoder_hid_dim_type": null,
|
34 |
+
"flip_sin_to_cos": true,
|
35 |
+
"freq_shift": 0,
|
36 |
+
"in_channels": 4,
|
37 |
+
"layers_per_block": 2,
|
38 |
+
"mid_block_only_cross_attention": null,
|
39 |
+
"mid_block_scale_factor": 1,
|
40 |
+
"mid_block_type": "UNetMidBlock2DCrossAttn",
|
41 |
+
"norm_eps": 1e-05,
|
42 |
+
"norm_num_groups": 32,
|
43 |
+
"num_attention_heads": null,
|
44 |
+
"num_class_embeds": null,
|
45 |
+
"only_cross_attention": false,
|
46 |
+
"out_channels": 4,
|
47 |
+
"projection_class_embeddings_input_dim": null,
|
48 |
+
"resnet_out_scale_factor": 1.0,
|
49 |
+
"resnet_skip_time_act": false,
|
50 |
+
"resnet_time_scale_shift": "default",
|
51 |
+
"sample_size": 64,
|
52 |
+
"time_cond_proj_dim": null,
|
53 |
+
"time_embedding_act_fn": null,
|
54 |
+
"time_embedding_dim": null,
|
55 |
+
"time_embedding_type": "positional",
|
56 |
+
"timestep_post_act": null,
|
57 |
+
"transformer_layers_per_block": 1,
|
58 |
+
"up_block_types": [
|
59 |
+
"UpBlock2D",
|
60 |
+
"CrossAttnUpBlock2D",
|
61 |
+
"CrossAttnUpBlock2D",
|
62 |
+
"CrossAttnUpBlock2D"
|
63 |
+
],
|
64 |
+
"upcast_attention": false,
|
65 |
+
"use_linear_projection": false
|
66 |
+
}
|
FrozenII/combined_model_base/unet/diffusion_pytorch_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:58b099a1e1435fcb21ed9d04e6ad6c3c7f5aaeaa207b3226d4203fe51f801217
|
3 |
+
size 1719125304
|
FrozenII/combined_model_base/vae/config.json
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "AutoencoderKL",
|
3 |
+
"_diffusers_version": "0.20.0",
|
4 |
+
"_name_or_path": "experiments/pretrained_models/chilloutmix/vae",
|
5 |
+
"act_fn": "silu",
|
6 |
+
"block_out_channels": [
|
7 |
+
128,
|
8 |
+
256,
|
9 |
+
512,
|
10 |
+
512
|
11 |
+
],
|
12 |
+
"down_block_types": [
|
13 |
+
"DownEncoderBlock2D",
|
14 |
+
"DownEncoderBlock2D",
|
15 |
+
"DownEncoderBlock2D",
|
16 |
+
"DownEncoderBlock2D"
|
17 |
+
],
|
18 |
+
"force_upcast": true,
|
19 |
+
"in_channels": 3,
|
20 |
+
"latent_channels": 4,
|
21 |
+
"layers_per_block": 2,
|
22 |
+
"norm_num_groups": 32,
|
23 |
+
"out_channels": 3,
|
24 |
+
"sample_size": 512,
|
25 |
+
"scaling_factor": 0.18215,
|
26 |
+
"up_block_types": [
|
27 |
+
"UpDecoderBlock2D",
|
28 |
+
"UpDecoderBlock2D",
|
29 |
+
"UpDecoderBlock2D",
|
30 |
+
"UpDecoderBlock2D"
|
31 |
+
]
|
32 |
+
}
|
FrozenII/combined_model_base/vae/diffusion_pytorch_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e4c08995484ee61270175e9e7a072b66a6e4eeb5f0c266667fe1f45b90daf9a
|
3 |
+
size 167335342
|