multimodalart HF Staff commited on
Commit
1ec084c
·
verified ·
1 Parent(s): 19222fc

End of training

Browse files
README.md ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen-Image
3
+ library_name: diffusers
4
+ license: apache-2.0
5
+ instance_prompt: a trtcrd of a mecha robot
6
+ widget: []
7
+ tags:
8
+ - text-to-image
9
+ - diffusers-training
10
+ - diffusers
11
+ - lora
12
+ - qwen-image
13
+ - qwen-image-diffusers
14
+ - template:sd-lora
15
+ ---
16
+
17
+ <!-- This model card has been generated automatically according to the information the training script had access to. You
18
+ should probably proofread and complete it, then remove this comment. -->
19
+
20
+
21
+ # HiDream Image DreamBooth LoRA - multimodalart/qwen-tarot
22
+
23
+ <Gallery />
24
+
25
+ ## Model description
26
+
27
+ These are multimodalart/qwen-tarot DreamBooth LoRA weights for Qwen/Qwen-Image.
28
+
29
+ The weights were trained using [DreamBooth](https://dreambooth.github.io/) with the [Qwen Image diffusers trainer](https://github.com/huggingface/diffusers/blob/main/examples/dreambooth/README_qwen.md).
30
+
31
+ ## Trigger words
32
+
33
+ You should use `a trtcrd of a mecha robot` to trigger the image generation.
34
+
35
+ ## Download model
36
+
37
+ [Download the *.safetensors LoRA](multimodalart/qwen-tarot/tree/main) in the Files & versions tab.
38
+
39
+ ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
40
+
41
+ ```py
42
+ >>> import torch
43
+ >>> from diffusers import QwenImagePipeline
44
+
45
+ >>> pipe = QwenImagePipeline.from_pretrained(
46
+ ... "Qwen/Qwen-Image",
47
+ ... torch_dtype=torch.bfloat16,
48
+ ... )
49
+ >>> pipe.enable_model_cpu_offload()
50
+ >>> pipe.load_lora_weights(f"multimodalart/qwen-tarot")
51
+ >>> image = pipe(f"a trtcrd of a mecha robot").images[0]
52
+
53
+
54
+ ```
55
+
56
+ For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
57
+
58
+
59
+ ## Intended uses & limitations
60
+
61
+ #### How to use
62
+
63
+ ```python
64
+ # TODO: add an example code snippet for running this diffusion pipeline
65
+ ```
66
+
67
+ #### Limitations and bias
68
+
69
+ [TODO: provide examples of latent issues and potential remediations]
70
+
71
+ ## Training details
72
+
73
+ [TODO: describe the data used to train the model]
checkpoint-1000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9078c3bbc121a3333e59eebf328acbbf41291cc045ae9d5d853b173ee4f545d
3
+ size 12542309
checkpoint-1000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdeed50e557a9a0593631edd4ec7e458782f0e6e84a8f6a02c6f4337e4f0a7ed
3
+ size 11859320
checkpoint-1000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:594a854cecc44c27c776d7ad6cfb99db85f4d8c27b9a2c55a470a6e4b0e0955e
3
+ size 14821
checkpoint-1000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5ca5fd1f3d50dfbd88c5f854e399c1c22a842d2231d41415da16bf924479e76
3
+ size 1401
checkpoint-500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea8e462dfe2cf38941a46a078a64d26a84bb00ba1eb7a5945ff64f450b7e04ba
3
+ size 12542309
checkpoint-500/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6dc14740c392744bc053fa6217a2eea50f4429cefeaff16d4c18f5705796c26
3
+ size 11859320
checkpoint-500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8d45c7f9dc18c2dba7df0a8d2dd0ebb419075a4d82307b1b915c64229456f32
3
+ size 14821
checkpoint-500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8baa45cc00e6da1eefbbbc2d50fb568432ef1e20241f0a3571b49dcb8d62f7f2
3
+ size 1401
logs/dreambooth-qwen-image-lora/1756501852.1546826/events.out.tfevents.1756501852.j-multimodalart-68b216c752ee3070efb12fd7-0a252klw-18c8c-1h9t6.132.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce74c8e4432f79439b0fc9ab98cc02383c6d9325fcb224772f71b61edfbd9b43
3
+ size 3520
logs/dreambooth-qwen-image-lora/1756501852.156232/hparams.yml ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.0001
5
+ allow_tf32: false
6
+ bnb_quantization_config_path: null
7
+ cache_dir: null
8
+ cache_latents: true
9
+ caption_column: caption
10
+ center_crop: false
11
+ checkpointing_steps: 500
12
+ checkpoints_total_limit: null
13
+ class_data_dir: null
14
+ class_prompt: null
15
+ dataloader_num_workers: 0
16
+ dataset_config_name: null
17
+ dataset_name: multimodalart/1920-raider-waite-tarot-public-domain
18
+ final_validation_prompt: null
19
+ gradient_accumulation_steps: 1
20
+ gradient_checkpointing: true
21
+ hub_model_id: multimodalart/qwen-tarot
22
+ hub_token: null
23
+ image_column: image
24
+ instance_data_dir: null
25
+ instance_prompt: a trtcrd of a mecha robot
26
+ learning_rate: 0.0001
27
+ local_rank: -1
28
+ logging_dir: logs
29
+ logit_mean: 0.0
30
+ logit_std: 1.0
31
+ lora_alpha: 4
32
+ lora_dropout: 0.0
33
+ lora_layers: null
34
+ lr_num_cycles: 1
35
+ lr_power: 1.0
36
+ lr_scheduler: constant
37
+ lr_warmup_steps: 0
38
+ max_grad_norm: 1.0
39
+ max_sequence_length: 512
40
+ max_train_steps: 1000
41
+ mixed_precision: bf16
42
+ mode_scale: 1.29
43
+ num_class_images: 100
44
+ num_train_epochs: 50
45
+ num_validation_images: 4
46
+ offload: false
47
+ optimizer: AdamW
48
+ output_dir: ./output
49
+ pretrained_model_name_or_path: Qwen/Qwen-Image
50
+ pretrained_text_encoder_4_name_or_path: meta-llama/Meta-Llama-3.1-8B-Instruct
51
+ pretrained_tokenizer_4_name_or_path: meta-llama/Meta-Llama-3.1-8B-Instruct
52
+ prior_loss_weight: 1.0
53
+ prodigy_beta3: null
54
+ prodigy_decouple: true
55
+ prodigy_safeguard_warmup: true
56
+ prodigy_use_bias_correction: true
57
+ push_to_hub: true
58
+ random_flip: false
59
+ rank: 4
60
+ repeats: 1
61
+ report_to: tensorboard
62
+ resolution: 1024
63
+ resume_from_checkpoint: null
64
+ revision: null
65
+ sample_batch_size: 4
66
+ scale_lr: false
67
+ seed: null
68
+ skip_final_inference: false
69
+ train_batch_size: 4
70
+ upcast_before_saving: false
71
+ use_8bit_adam: true
72
+ validation_epochs: 1000
73
+ validation_prompt: null
74
+ variant: null
75
+ weighting_scheme: none
76
+ with_prior_preservation: false
logs/dreambooth-qwen-image-lora/events.out.tfevents.1756501852.j-multimodalart-68b216c752ee3070efb12fd7-0a252klw-18c8c-1h9t6.132.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:265be75438baa128b4a6e58de6c48fdfb83bdacfc2a52a0efe2dc7422c901493
3
+ size 83834
pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:400f055683af628eb8a085158c2de15ed54ff4d35a7b8abd85738850fc957e91
3
+ size 11859320