🐮cow architecture gguf encoder

  • don't need to rebuild tokenizer from metadata
  • don't need separate tokenizer file
  • no more oom issues (possibly)

eligible model example

screenshot

the example workflow above is from wan-s2v-gguf; cow encoder is a special designed clip, even the lowest q2 quant still working very good; upgrade your node for cow-encoder support🥛🐮 and do drink more milk

screenshot

Prompt
a conversation between cgg and connector

reference

Downloads last month
426
GGUF
Model size
2.62B params
Architecture
cow
Hardware compatibility
Log In to view the estimation

2-bit

4-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support