embedder_collection / README.md
kalle07's picture
Update README.md
66bf69c verified
|
raw
history blame
985 Bytes
metadata
library_name: sentence-transformers
pipeline_tag: sentence-similarity
tags:
  - sentence-transformers
  - sentence-similarity
  - feature-extraction
  - embedder
  - embedding
  - moedels
  - GGUF
  - text-embeddings-inference
misc:
  - text-embeddings-inference
language:
  - en
  - de

All models tested with ALLM(AnythingLLM) with LM as server

they work more or less

my short impression:

  • nomic-embed-text
  • mxbai-embed-large
  • mug-b-1.6

working well, all other its up to you!

short hints for using: set your (Max Tokens)context-lenght 16000t main-model, set your embedder-model (Max Embedding Chunk Length) 1024t,set (Max Context Snippets) 14

-> ok what that mean!

you can receive 14-snippets a 1024t (14336t) from your document ~10000words and 1600t left for the answer ~1000words

you can play and set for your needs, eg 8-snippets a 2048t, or 28-snippets a 512t ...

16000t ~1GB VRAM usage

...

...

...

(ALL Licenses and terms of use go to original author)