Updates README.md
Browse files
README.md
CHANGED
@@ -5,14 +5,13 @@ tags:
|
|
5 |
- sentence-transformers
|
6 |
- feature-extraction
|
7 |
- sentence-similarity
|
8 |
-
|
|
|
9 |
---
|
10 |
|
11 |
# {MODEL_NAME}
|
12 |
|
13 |
-
This is a [
|
14 |
-
|
15 |
-
<!--- Describe your model here -->
|
16 |
|
17 |
## Usage (Sentence-Transformers)
|
18 |
|
@@ -33,15 +32,6 @@ embeddings = model.encode(sentences)
|
|
33 |
print(embeddings)
|
34 |
```
|
35 |
|
36 |
-
|
37 |
-
|
38 |
-
## Evaluation Results
|
39 |
-
|
40 |
-
<!--- Describe how your model was evaluated -->
|
41 |
-
|
42 |
-
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
|
43 |
-
|
44 |
-
|
45 |
## Training
|
46 |
The model was trained with the parameters:
|
47 |
|
@@ -88,5 +78,4 @@ SentenceTransformer(
|
|
88 |
```
|
89 |
|
90 |
## Citing & Authors
|
91 |
-
|
92 |
-
<!--- Describe where people can find more information -->
|
|
|
5 |
- sentence-transformers
|
6 |
- feature-extraction
|
7 |
- sentence-similarity
|
8 |
+
- embeddings
|
9 |
+
license: mit
|
10 |
---
|
11 |
|
12 |
# {MODEL_NAME}
|
13 |
|
14 |
+
This is a fine-tuned embedding model of [bge-large-en-v1.5](https://huggingface.co/BAAI/bge-large-en-v1.5). It was fine-tuned on a dataset based on an ISO 27001 text corpus consisting of text chunks (1024 characters) and associated questions. A total of 2.000 chunk and question pairs were generated. The fine-tuning process is specialized on an Information Retrieval task in which the generated questions are used to find the relevant chunks. The effectiveness of the model is evaluated on whether the correct chunk was retrieved, and the loss is calculated with the multiple negative ranking loss.
|
|
|
|
|
15 |
|
16 |
## Usage (Sentence-Transformers)
|
17 |
|
|
|
32 |
print(embeddings)
|
33 |
```
|
34 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
## Training
|
36 |
The model was trained with the parameters:
|
37 |
|
|
|
78 |
```
|
79 |
|
80 |
## Citing & Authors
|
81 |
+
Based on https://huggingface.co/BAAI/bge-large-en-v1.5 from Xiao et al. (2023) (C-Pack: Packaged Resources To Advance General Chinese Embedding)
|
|