dwulff commited on
Commit
45a04ee
·
verified ·
1 Parent(s): ff11158

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -13
README.md CHANGED
@@ -10,37 +10,59 @@ tags:
10
 
11
  # dwulff/mpnet-personality
12
 
13
- This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
14
 
15
- <!--- Describe your model here -->
16
 
17
- ## Usage (Sentence-Transformers)
18
 
19
- Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
 
 
20
 
21
  ```
 
22
  pip install -U sentence-transformers
 
 
 
23
  ```
24
 
25
- Then you can use the model like this:
26
 
27
  ```python
28
  from sentence_transformers import SentenceTransformer
29
- sentences = ["This is an example sentence", "Each sentence is converted"]
30
 
31
- model = SentenceTransformer('dwulff/mpnet-personality')
 
 
 
 
 
 
32
  embeddings = model.encode(sentences)
33
  print(embeddings)
34
  ```
35
 
 
36
 
 
37
 
38
- ## Evaluation Results
39
 
40
- <!--- Describe how your model was evaluated -->
41
 
42
- For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=dwulff/mpnet-personality)
43
 
 
 
 
 
 
 
 
 
 
44
 
45
  ## Training
46
  The model was trained with the parameters:
@@ -84,6 +106,3 @@ SentenceTransformer(
84
  )
85
  ```
86
 
87
- ## Citing & Authors
88
-
89
- <!--- Describe where people can find more information -->
 
10
 
11
  # dwulff/mpnet-personality
12
 
13
+ This is a [sentence-transformers](https://www.SBERT.net) model that maps personality-related items or texts into a 768-dimensional dense vector space and can be used for many tasks in personality psychology.
14
 
15
+ The model has been generated by fine-tuning [all-mpnet-base-v2](https://huggingface.co/sentence-transformers/all-mpnet-base-v2) using unsigned empirical correlations of 200k pairs of personality items. As a result, the model encodes the content of personality-related texts independent of the direction (e.g., negation).
16
 
17
+ See [Wulff & Mata (2024)](https://osf.io/preprints/psyarxiv/9h7aw) (see [Supplement](https://osf.io/z47qs/)) for details. For a similar approach, see [Hommel & Arslan (2024)](https://osf.io/preprints/psyarxiv/kjuce).
18
 
19
+ ## Usage
20
+
21
+ Make sure [sentence-transformers](https://www.SBERT.net) is installed:
22
 
23
  ```
24
+ # latest version
25
  pip install -U sentence-transformers
26
+
27
+ # latest dev version
28
+ pip install git+https://github.com/UKPLab/sentence-transformers.git
29
  ```
30
 
31
+ You can extract embeddings in the following way:
32
 
33
  ```python
34
  from sentence_transformers import SentenceTransformer
 
35
 
36
+ # personality sentences
37
+ sentences = ["Rarely think about how I feel.", "Make decisions quickly."]
38
+
39
+ # load model
40
+ model = SentenceTransformer('Fissionships/mpnet-personality')
41
+
42
+ # extract embeddings
43
  embeddings = model.encode(sentences)
44
  print(embeddings)
45
  ```
46
 
47
+ ## Evaluation Results
48
 
49
+ The model has been evaluated on public personality data. For standard personality inventories, such as the BIG5 or HEXACO inventories, the model predicts the empirical correlations between personality items at Pearson r ~ .6 and empirical correlations between scales at Pearson r ~ .8.
50
 
51
+ Performance can be higher on the many common personality items it has been trained on due to memorization. Performance will be worse for more specialized personality assessments and texts beyond personality items.
52
 
53
+ See [Wulff & Mata (2024)](https://osf.io/preprints/psyarxiv/9h7aw) (see [Supplement](https://osf.io/z47qs/)) for details.
54
 
55
+ ## Citing
56
 
57
+ ```
58
+ @article{wulff2024jinglejangle,
59
+ author = {Wulff, Dirk U. and Mata, Rui},
60
+ title = {Automated jingle–jangle detection: Using embeddings to tackle taxonomic incommensurability},
61
+ journal = {PsyArViv},
62
+ doi = {https://doi.org/10.31234/osf.io/9h7aw}
63
+ }
64
+
65
+ ```
66
 
67
  ## Training
68
  The model was trained with the parameters:
 
106
  )
107
  ```
108