exdysa commited on
Commit
c164bbf
·
verified ·
1 Parent(s): 204c72d

Delete files ** with huggingface_hub

Browse files
.DS_Store DELETED
Binary file (10.2 kB)
 
README.md DELETED
File without changes
model_index.json DELETED
@@ -1,34 +0,0 @@
1
- {
2
- "_class_name": "StableDiffusionXLPipeline",
3
- "_diffusers_version": "0.19.0.dev0",
4
- "force_zeros_for_empty_prompt": true,
5
- "add_watermarker": null,
6
- "scheduler": [
7
- "diffusers",
8
- "EulerDiscreteScheduler"
9
- ],
10
- "text_encoder": [
11
- "transformers",
12
- "CLIPTextModel"
13
- ],
14
- "text_encoder_2": [
15
- "transformers",
16
- "CLIPTextModelWithProjection"
17
- ],
18
- "tokenizer": [
19
- "transformers",
20
- "CLIPTokenizer"
21
- ],
22
- "tokenizer_2": [
23
- "transformers",
24
- "CLIPTokenizer"
25
- ],
26
- "unet": [
27
- "diffusers",
28
- "UNet2DConditionModel"
29
- ],
30
- "vae": [
31
- "diffusers",
32
- "AutoencoderKL"
33
- ]
34
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
scheduler/scheduler_config.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "_class_name": "EulerDiscreteScheduler",
3
- "_diffusers_version": "0.19.0.dev0",
4
- "beta_end": 0.012,
5
- "beta_schedule": "scaled_linear",
6
- "beta_start": 0.00085,
7
- "clip_sample": false,
8
- "interpolation_type": "linear",
9
- "num_train_timesteps": 1000,
10
- "prediction_type": "epsilon",
11
- "sample_max_value": 1.0,
12
- "set_alpha_to_one": false,
13
- "skip_prk_steps": true,
14
- "steps_offset": 1,
15
- "timestep_spacing": "leading",
16
- "trained_betas": null,
17
- "use_karras_sigmas": false
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
text_encoder/.DS_Store DELETED
Binary file (8.2 kB)
 
text_encoder/config.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "architectures": [
3
- "CLIPTextModel"
4
- ],
5
- "attention_dropout": 0.0,
6
- "bos_token_id": 0,
7
- "dropout": 0.0,
8
- "eos_token_id": 2,
9
- "hidden_act": "quick_gelu",
10
- "hidden_size": 768,
11
- "initializer_factor": 1.0,
12
- "initializer_range": 0.02,
13
- "intermediate_size": 3072,
14
- "layer_norm_eps": 1e-05,
15
- "max_position_embeddings": 77,
16
- "model_type": "clip_text_model",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 12,
19
- "pad_token_id": 1,
20
- "projection_dim": 768,
21
- "torch_dtype": "float16",
22
- "transformers_version": "4.32.0.dev0",
23
- "vocab_size": 49408
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
text_encoder_2/.DS_Store DELETED
Binary file (6.15 kB)
 
text_encoder_2/config.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "architectures": [
3
- "CLIPTextModelWithProjection"
4
- ],
5
- "attention_dropout": 0.0,
6
- "bos_token_id": 0,
7
- "dropout": 0.0,
8
- "eos_token_id": 2,
9
- "hidden_act": "gelu",
10
- "hidden_size": 1280,
11
- "initializer_factor": 1.0,
12
- "initializer_range": 0.02,
13
- "intermediate_size": 5120,
14
- "layer_norm_eps": 1e-05,
15
- "max_position_embeddings": 77,
16
- "model_type": "clip_text_model",
17
- "num_attention_heads": 20,
18
- "num_hidden_layers": 32,
19
- "pad_token_id": 1,
20
- "projection_dim": 1280,
21
- "torch_dtype": "float16",
22
- "transformers_version": "4.32.0.dev0",
23
- "vocab_size": 49408
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|startoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "<|endoftext|>",
17
- "unk_token": {
18
- "content": "<|endoftext|>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/tokenizer_config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "bos_token": {
4
- "__type": "AddedToken",
5
- "content": "<|startoftext|>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false
10
- },
11
- "clean_up_tokenization_spaces": true,
12
- "do_lower_case": true,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "<|endoftext|>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "errors": "replace",
22
- "model_max_length": 77,
23
- "pad_token": "<|endoftext|>",
24
- "tokenizer_class": "CLIPTokenizer",
25
- "unk_token": {
26
- "__type": "AddedToken",
27
- "content": "<|endoftext|>",
28
- "lstrip": false,
29
- "normalized": true,
30
- "rstrip": false,
31
- "single_word": false
32
- }
33
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer_2/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer_2/special_tokens_map.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|startoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "!",
17
- "unk_token": {
18
- "content": "<|endoftext|>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer_2/tokenizer_config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "bos_token": {
4
- "__type": "AddedToken",
5
- "content": "<|startoftext|>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false
10
- },
11
- "clean_up_tokenization_spaces": true,
12
- "do_lower_case": true,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "<|endoftext|>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "errors": "replace",
22
- "model_max_length": 77,
23
- "pad_token": "!",
24
- "tokenizer_class": "CLIPTokenizer",
25
- "unk_token": {
26
- "__type": "AddedToken",
27
- "content": "<|endoftext|>",
28
- "lstrip": false,
29
- "normalized": true,
30
- "rstrip": false,
31
- "single_word": false
32
- }
33
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer_2/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
unet/.DS_Store DELETED
Binary file (6.15 kB)
 
unet/config.json DELETED
@@ -1,69 +0,0 @@
1
- {
2
- "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.19.0.dev0",
4
- "act_fn": "silu",
5
- "addition_embed_type": "text_time",
6
- "addition_embed_type_num_heads": 64,
7
- "addition_time_embed_dim": 256,
8
- "attention_head_dim": [
9
- 5,
10
- 10,
11
- 20
12
- ],
13
- "block_out_channels": [
14
- 320,
15
- 640,
16
- 1280
17
- ],
18
- "center_input_sample": false,
19
- "class_embed_type": null,
20
- "class_embeddings_concat": false,
21
- "conv_in_kernel": 3,
22
- "conv_out_kernel": 3,
23
- "cross_attention_dim": 2048,
24
- "cross_attention_norm": null,
25
- "down_block_types": [
26
- "DownBlock2D",
27
- "CrossAttnDownBlock2D",
28
- "CrossAttnDownBlock2D"
29
- ],
30
- "downsample_padding": 1,
31
- "dual_cross_attention": false,
32
- "encoder_hid_dim": null,
33
- "encoder_hid_dim_type": null,
34
- "flip_sin_to_cos": true,
35
- "freq_shift": 0,
36
- "in_channels": 4,
37
- "layers_per_block": 2,
38
- "mid_block_only_cross_attention": null,
39
- "mid_block_scale_factor": 1,
40
- "mid_block_type": "UNetMidBlock2DCrossAttn",
41
- "norm_eps": 1e-05,
42
- "norm_num_groups": 32,
43
- "num_attention_heads": null,
44
- "num_class_embeds": null,
45
- "only_cross_attention": false,
46
- "out_channels": 4,
47
- "projection_class_embeddings_input_dim": 2816,
48
- "resnet_out_scale_factor": 1.0,
49
- "resnet_skip_time_act": false,
50
- "resnet_time_scale_shift": "default",
51
- "sample_size": 128,
52
- "time_cond_proj_dim": null,
53
- "time_embedding_act_fn": null,
54
- "time_embedding_dim": null,
55
- "time_embedding_type": "positional",
56
- "timestep_post_act": null,
57
- "transformer_layers_per_block": [
58
- 1,
59
- 2,
60
- 10
61
- ],
62
- "up_block_types": [
63
- "CrossAttnUpBlock2D",
64
- "CrossAttnUpBlock2D",
65
- "UpBlock2D"
66
- ],
67
- "upcast_attention": null,
68
- "use_linear_projection": true
69
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
vae/.DS_Store DELETED
Binary file (6.15 kB)
 
vae/config.json DELETED
@@ -1,31 +0,0 @@
1
- {
2
- "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.20.0.dev0",
4
- "act_fn": "silu",
5
- "block_out_channels": [
6
- 128,
7
- 256,
8
- 512,
9
- 512
10
- ],
11
- "down_block_types": [
12
- "DownEncoderBlock2D",
13
- "DownEncoderBlock2D",
14
- "DownEncoderBlock2D",
15
- "DownEncoderBlock2D"
16
- ],
17
- "force_upcast": true,
18
- "in_channels": 3,
19
- "latent_channels": 4,
20
- "layers_per_block": 2,
21
- "norm_num_groups": 32,
22
- "out_channels": 3,
23
- "sample_size": 1024,
24
- "scaling_factor": 0.13025,
25
- "up_block_types": [
26
- "UpDecoderBlock2D",
27
- "UpDecoderBlock2D",
28
- "UpDecoderBlock2D",
29
- "UpDecoderBlock2D"
30
- ]
31
- }