Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
singhsidhukuldeepย 
posted an update May 30
Post
1877
๐Ÿฆ… Falcon has landed... again!
And now it not just reads but sees as well ๐Ÿ“–๐Ÿ‘€

Here is a summary of the Falcon-11B-VLM model:

Model Type: Causal decoder-only model ๐Ÿ”„.

Parameters: 11 billion ๐ŸŒŒ.

Vision Integration: Uses the pretrained CLIP ViT-L/14 vision encoder with the recently released Falcon2-11B chat-finetuned model and trained with image-text data ๐Ÿ–ผ๏ธ๐Ÿ“š.

Training: Pretrained on over 5,000 billion tokens from RefinedWeb with curated corpora ๐Ÿ“Š.

Dynamic Encoding: Enhances perception of fine-grained details in images ๐Ÿ”.

Training Hardware: 16 A100 80GB GPUs with ZeRO and Flash-Attention 2 ๐Ÿ–ฅ๏ธ.

Tokenizer: Falcon-7B/11B tokenizer ๐Ÿงฉ.

Languages Supported: ๐ŸŒ Primarily English, with capabilities in German ๐Ÿ‡ฉ๐Ÿ‡ช, Spanish ๐Ÿ‡ช๐Ÿ‡ธ, French ๐Ÿ‡ซ๐Ÿ‡ท, Italian ๐Ÿ‡ฎ๐Ÿ‡น, Dutch ๐Ÿ‡ณ๐Ÿ‡ฑ, Romanian ๐Ÿ‡ท๐Ÿ‡ด, Czech ๐Ÿ‡จ๐Ÿ‡ฟ, Swedish ๐Ÿ‡ธ๐Ÿ‡ช, and more. ๐Ÿ—ฃ๏ธ๐ŸŒ.

License: Open Source - TII Falcon License 2.0, based on Apache 2.0 ๐Ÿ“œ.

Model: tiiuae/falcon-11B-vlm