MiaoshouAI commited on
Commit
79ed3c5
·
verified ·
1 Parent(s): dfb68da

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -10
README.md CHANGED
@@ -2,8 +2,8 @@
2
  license: mit
3
  ---
4
  # Florence-2-base-PromptGen v.15
5
- This is a major version update for PromptGen. In this new version, two new caption instructions are added: \<GENERATE_TAGS\> and \<MIXED_CAPTION\>
6
- Also, you will notice a much improved accuracy improved by a new set of training dataset for this version. This version no longer uses Civitai Data due to lora trigger words and inaccurate tags caused for misinterpretation.
7
 
8
  # About PromptGen
9
  Florence-2-base-PromptGen is a model trained for [MiaoshouAI Tagger for ComfyUI](https://github.com/miaoshouai/ComfyUI-Miaoshouai-Tagger).
@@ -14,19 +14,29 @@ Most vision models today are trained mainly for general vision recognition purpo
14
  Florence-2-base-PromptGen is trained on such a purpose as aiming to improve the tagging experience and accuracy of the prompt and tagging job. The model is trained based on images and cleaned tags from Civitai so that the end result for tagging the images are the prompts you use to generate these images.
15
  You won't get annoying captions like "This is image is about a girl..." or
16
 
 
 
 
 
 
 
 
 
 
 
17
  ## Instruction prompt:
18
- \<GENERATE_TAGS\> generate prompt as danbooru style tags
19
- \<CAPTION\> a one line caption for the image
20
- \<DETAILED_CAPTION\> a structured caption format which detects the position of the subjects in the image
21
- \<MORE_DETAILED_CAPTION\> a very detailed description for the image
22
- \<MIXED_CAPTION\> a Mixed caption style of more detailed caption and tags, this is extremely useful for FLUX model when using T5XXL and CLIP_L together. A new node in MiaoshouTagger ComfyUI is added to support this instruction.
23
 
24
  ## Version History:
25
  For version 1.5, you will notice the following
26
  1. \<GENERATE_PROMPT\> is deprecated and replace by \<GENERATE_TAGS\>
27
- 2. A new mode for \<MIXED_CAPTION\>
28
- 2. A much improve accuracy for \<DETAILED_CAPTION\> and \<MORE_DETAILED_CAPTION\>
29
- 3. Improved ability for recognizing watermarks on images.
30
 
31
 
32
  ## How to use:
 
2
  license: mit
3
  ---
4
  # Florence-2-base-PromptGen v.15
5
+ This is a major version upgrade for PromptGen. In this new version, two new caption instructions are added: \<GENERATE_TAGS\> and \<MIXED_CAPTION\>
6
+ You'll also notice significantly improved accuracy with this version, thanks to a new training dataset. It no longer relies on Civitai Data, avoiding the issues of lora trigger words and inaccurate tags from misinterpretation.
7
 
8
  # About PromptGen
9
  Florence-2-base-PromptGen is a model trained for [MiaoshouAI Tagger for ComfyUI](https://github.com/miaoshouai/ComfyUI-Miaoshouai-Tagger).
 
14
  Florence-2-base-PromptGen is trained on such a purpose as aiming to improve the tagging experience and accuracy of the prompt and tagging job. The model is trained based on images and cleaned tags from Civitai so that the end result for tagging the images are the prompts you use to generate these images.
15
  You won't get annoying captions like "This is image is about a girl..." or
16
 
17
+ ## Features:
18
+ * Describes image in much detail when using \<MORE_DETAILED_CAPTION\> instruction.
19
+ <img style="width:70%; hight:70%" src="https://msdn.miaoshouai.com/miaoshou/bo/2024-09-05_12-40-34.png" />
20
+ * When using \<DETAILED_CAPTION\> instruction, it creates a structured caption with infomation on subject's position and also reads the text from the image, which can be super useful when recreate a scene.
21
+ <img style="width:70%; hight:70%" src="https://msdn.miaoshouai.com/miaoshou/bo/2024-09-05_13-07-54.png" />
22
+ * Memory efficient compare to other models! This is a really light weight caption model that allows you to use a little more than 1G of VRAM and produce lightening fast and high quality image captions.
23
+ <img style="width:70%; hight:70%" src="https://msdn.miaoshouai.com/miaoshou/bo/2024-09-05_12-56-39.png" />
24
+ * Designed to handle image captions for Flux model for both T5XXL CLIP and CLIP_L, the Miaoshou Tagger new node called "Flux CLIP Text Encode" which eliminates the need to run two separate tagger tools for caption creation. You can easily populate both CLIPs in a single generation, significantly boosting speed when working with Flux models.
25
+ <img style="width:70%; hight:70%" src="https://msdn.miaoshouai.com/miaoshou/bo/2024-09-05_14-11-02.png" />
26
+
27
  ## Instruction prompt:
28
+ \<GENERATE_TAGS\> generate prompt as danbooru style tags<br>
29
+ \<CAPTION\> a one line caption for the image<br>
30
+ \<DETAILED_CAPTION\> a structured caption format which detects the position of the subjects in the image<br>
31
+ \<MORE_DETAILED_CAPTION\> a very detailed description for the image<br>
32
+ \<MIXED_CAPTION\> a mixed caption style of more detailed caption and tags, this is extremely useful for FLUX model when using T5XXL and CLIP_L together. A new node in MiaoshouTagger ComfyUI is added to support this instruction.<br>
33
 
34
  ## Version History:
35
  For version 1.5, you will notice the following
36
  1. \<GENERATE_PROMPT\> is deprecated and replace by \<GENERATE_TAGS\>
37
+ 2. A new instruction for \<MIXED_CAPTION\>
38
+ 3. A much improve accuracy for \<DETAILED_CAPTION\> and \<MORE_DETAILED_CAPTION\>
39
+ 4. Improved ability for recognizing watermarks on images.
40
 
41
 
42
  ## How to use: