rasyosef commited on
Commit
de3ab2c
·
verified ·
1 Parent(s): 5361683

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -27
README.md CHANGED
@@ -151,33 +151,6 @@ datasets:
151
  # SPLADE-BERT-Tiny-Distil
152
 
153
  This is a [SPLADE Sparse Encoder](https://www.sbert.net/docs/sparse_encoder/usage/usage.html) model finetuned from [prajjwal1/bert-tiny](https://huggingface.co/prajjwal1/bert-tiny) using the [sentence-transformers](https://www.SBERT.net) library. It maps sentences & paragraphs to a 30522-dimensional sparse vector space and can be used for semantic search and sparse retrieval.
154
- ## Model Details
155
-
156
- ### Model Description
157
- - **Model Type:** SPLADE Sparse Encoder
158
- - **Base model:** [prajjwal1/bert-tiny](https://huggingface.co/prajjwal1/bert-tiny) <!-- at revision 6f75de8b60a9f8a2fdf7b69cbd86d9e64bcb3837 -->
159
- - **Maximum Sequence Length:** 512 tokens
160
- - **Output Dimensionality:** 30522 dimensions
161
- - **Similarity Function:** Dot Product
162
- <!-- - **Training Dataset:** Unknown -->
163
- - **Language:** en
164
- - **License:** mit
165
-
166
- ### Model Sources
167
-
168
- - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
169
- - **Documentation:** [Sparse Encoder Documentation](https://www.sbert.net/docs/sparse_encoder/usage/usage.html)
170
- - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
171
- - **Hugging Face:** [Sparse Encoders on Hugging Face](https://huggingface.co/models?library=sentence-transformers&other=sparse-encoder)
172
-
173
- ### Full Model Architecture
174
-
175
- ```
176
- SparseEncoder(
177
- (0): MLMTransformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'BertForMaskedLM'})
178
- (1): SpladePooling({'pooling_strategy': 'max', 'activation_function': 'relu', 'word_embedding_dimension': 30522})
179
- )
180
- ```
181
 
182
  ## Usage
183
 
@@ -238,6 +211,35 @@ You can finetune this model on your own dataset.
238
 
239
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
240
  -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
241
  ## More
242
  <details><summary>Click to expand</summary>
243
 
 
151
  # SPLADE-BERT-Tiny-Distil
152
 
153
  This is a [SPLADE Sparse Encoder](https://www.sbert.net/docs/sparse_encoder/usage/usage.html) model finetuned from [prajjwal1/bert-tiny](https://huggingface.co/prajjwal1/bert-tiny) using the [sentence-transformers](https://www.SBERT.net) library. It maps sentences & paragraphs to a 30522-dimensional sparse vector space and can be used for semantic search and sparse retrieval.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
154
 
155
  ## Usage
156
 
 
211
 
212
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
213
  -->
214
+
215
+ ## Model Details
216
+
217
+ ### Model Description
218
+ - **Model Type:** SPLADE Sparse Encoder
219
+ - **Base model:** [prajjwal1/bert-tiny](https://huggingface.co/prajjwal1/bert-tiny) <!-- at revision 6f75de8b60a9f8a2fdf7b69cbd86d9e64bcb3837 -->
220
+ - **Maximum Sequence Length:** 512 tokens
221
+ - **Output Dimensionality:** 30522 dimensions
222
+ - **Similarity Function:** Dot Product
223
+ <!-- - **Training Dataset:** Unknown -->
224
+ - **Language:** en
225
+ - **License:** mit
226
+
227
+ ### Model Sources
228
+
229
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
230
+ - **Documentation:** [Sparse Encoder Documentation](https://www.sbert.net/docs/sparse_encoder/usage/usage.html)
231
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
232
+ - **Hugging Face:** [Sparse Encoders on Hugging Face](https://huggingface.co/models?library=sentence-transformers&other=sparse-encoder)
233
+
234
+ ### Full Model Architecture
235
+
236
+ ```
237
+ SparseEncoder(
238
+ (0): MLMTransformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'BertForMaskedLM'})
239
+ (1): SpladePooling({'pooling_strategy': 'max', 'activation_function': 'relu', 'word_embedding_dimension': 30522})
240
+ )
241
+ ```
242
+
243
  ## More
244
  <details><summary>Click to expand</summary>
245