Update README.md
Browse files
README.md
CHANGED
@@ -36,13 +36,13 @@ BTW embedder is only a part of a good RAG<br>
|
|
36 |
<br>
|
37 |
<b>My short impression:</b>
|
38 |
<ul style="line-height: 1;">
|
39 |
-
<li>nomic-embed-text</li>
|
40 |
<li>mxbai-embed-large</li>
|
41 |
<li>mug-b-1.6</li>
|
42 |
-
<li>snowflake-arctic-embed-l-v2.0</li>
|
43 |
-
<li>Ger-RAG-BGE-M3 (german)</li>
|
44 |
<li>german-roberta</li>
|
45 |
-
<li>bge-m3
|
46 |
</ul>
|
47 |
Working well, all other its up to you! Some models are very similar! (jina and qwen based not yet supported by LM)<br>
|
48 |
With the same setting, these embedders found same 6-7 snippets out of 10 from a book. This means that only 3-4 snippets were different, but I didn't test it extensively.
|
@@ -84,7 +84,7 @@ This text snippet is then used for your answer. <br>
|
|
84 |
<li>If, for example, the word “XYZ” occurs 100 times in one file, not all 100 are found.</li>
|
85 |
<li>If only one snippet corresponds to your question all other snippets can negatively influence your answer because they do not fit the topic (usually 4 to 32 snippet are fine)</li>
|
86 |
<li>If you expect multible search results in your docs try 16-snippets or more, if you expect only 2 than dont use more!</li>
|
87 |
-
<li>If you use chunk-length ~1024t you receive more content, if you use ~256t you receive more facts.</li>
|
88 |
<li>A question for "summary of the document" is most time not useful, if the document has an introduction or summaries its searching there if you have luck.</li>
|
89 |
<li>If a book has a table of contents or a bibliography, I would delete these pages as they often contain relevant search terms but do not help answer your question.</li>
|
90 |
<li>If the documents small like 10-20 Pages, its better you copy the whole text inside the prompt, some options called "pin".</li>
|
@@ -124,7 +124,7 @@ Your aim is to share delicious recipes, cooking tips and the stories behind diff
|
|
124 |
...
|
125 |
<br>
|
126 |
# usual models works well:<br>
|
127 |
-
llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, gemma-3,
|
128 |
(llama3 or phi3.5 are not working well) <br><br>
|
129 |
<b>⇨</b> best models for english and german:<br>
|
130 |
granit3.2-8b (2b version also) - https://huggingface.co/ibm-research/granite-3.2-8b-instruct-GGUF<br>
|
|
|
36 |
<br>
|
37 |
<b>My short impression:</b>
|
38 |
<ul style="line-height: 1;">
|
39 |
+
<li>nomic-embed-text (up to 2048t context length)</li>
|
40 |
<li>mxbai-embed-large</li>
|
41 |
<li>mug-b-1.6</li>
|
42 |
+
<li>snowflake-arctic-embed-l-v2.0 (up to 8192t context length)</li>
|
43 |
+
<li>Ger-RAG-BGE-M3 (german, up to 8192t context length)</li>
|
44 |
<li>german-roberta</li>
|
45 |
+
<li>bge-m3 (up to 8192t context length)</li>
|
46 |
</ul>
|
47 |
Working well, all other its up to you! Some models are very similar! (jina and qwen based not yet supported by LM)<br>
|
48 |
With the same setting, these embedders found same 6-7 snippets out of 10 from a book. This means that only 3-4 snippets were different, but I didn't test it extensively.
|
|
|
84 |
<li>If, for example, the word “XYZ” occurs 100 times in one file, not all 100 are found.</li>
|
85 |
<li>If only one snippet corresponds to your question all other snippets can negatively influence your answer because they do not fit the topic (usually 4 to 32 snippet are fine)</li>
|
86 |
<li>If you expect multible search results in your docs try 16-snippets or more, if you expect only 2 than dont use more!</li>
|
87 |
+
<li>If you use chunk-length ~1024t you receive more content, if you use ~256t you receive more facts BUT lower chunk-length are more chunks and need much longer time.</li>
|
88 |
<li>A question for "summary of the document" is most time not useful, if the document has an introduction or summaries its searching there if you have luck.</li>
|
89 |
<li>If a book has a table of contents or a bibliography, I would delete these pages as they often contain relevant search terms but do not help answer your question.</li>
|
90 |
<li>If the documents small like 10-20 Pages, its better you copy the whole text inside the prompt, some options called "pin".</li>
|
|
|
124 |
...
|
125 |
<br>
|
126 |
# usual models works well:<br>
|
127 |
+
llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, gemma-3, granite, SauerkrautLM-Nemo(german) ... <br>
|
128 |
(llama3 or phi3.5 are not working well) <br><br>
|
129 |
<b>⇨</b> best models for english and german:<br>
|
130 |
granit3.2-8b (2b version also) - https://huggingface.co/ibm-research/granite-3.2-8b-instruct-GGUF<br>
|