Update README.md
Browse files
README.md
CHANGED
@@ -22,11 +22,14 @@ Or check out our Spaces demo! [![Open in Spaces](https://huggingface.co/datasets
|
|
22 |
LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data.
|
23 |
It is an auto-regressive language model, based on the transformer architecture.
|
24 |
|
|
|
|
|
|
|
25 |
**Model date:**
|
26 |
-
|
27 |
|
28 |
**Paper or resources for more information:**
|
29 |
-
https://llava
|
30 |
|
31 |
## How to use the model
|
32 |
|
|
|
22 |
LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data.
|
23 |
It is an auto-regressive language model, based on the transformer architecture.
|
24 |
|
25 |
+
Vip-LlaVa enhances the training protocol of Llava by marking images and interact with the model using natural cues like a
|
26 |
+
“red bounding box” or “pointed arrow” during training.
|
27 |
+
|
28 |
**Model date:**
|
29 |
+
ViP-LLaVa was released in December 2023.
|
30 |
|
31 |
**Paper or resources for more information:**
|
32 |
+
https://vip-llava.github.io/
|
33 |
|
34 |
## How to use the model
|
35 |
|