praeclarumjj3 commited on
Commit
4d5a268
1 Parent(s): 8e57e4d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -0
README.md CHANGED
@@ -1,3 +1,22 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ # VCoder-DS LLaVA-1.5-7b
6
+
7
+ VCoder-DS LLaVA-1.5-7b was trained on COST training dataset in December 2023. It uses the pretrained [LLaVA-1.5-7b](https://huggingface.co/liuhaotian/llava-v1.5-7b) model weights. It was introduced by Jain et al. in [this repository](https://github.com/SHI-Labs/VCoder).
8
+
9
+ VCoder is an adapter for improving existing Vision LLMs at object-level perception tasks with the use of perception modalities as control inputs while retaining performance on other tasks.
10
+
11
+ ![img](https://praeclarumjj3.github.io/vcoder/vcoder.svg)
12
+
13
+ ### Citation
14
+
15
+ ```bibtex
16
+ @article{jain2023vcoder,
17
+ title={{VCoder: Versatile Visual Encoder for Accurate Object-Level Perception with Large Language Models}},
18
+ author={Jitesh Jain and Jianwei Yang and Humphrey Shi},
19
+ journal={arXiv},
20
+ year={2023}
21
+ }
22
+ ```