MeDa-WE / README.md
jannikskytt's picture
Update README.md
217c1fa
|
raw
history blame
663 Bytes
metadata
license: cc-by-nc-3.0
language:
  - da
tags:
  - word embeddings
  - Danish

Danish medical word embeddings

MeDa-We was trained on a Danish medical corpus of 123M tokens. The word embeddings are 300-dimensional and are trained using FastText.

The embeddings were trained for 10 epochs using a window size of 5 and 10 negative samples.

The development of the corpus and word embeddings is described further in our paper.

We also trained a transformer model on the developed corpus which can be found here.

Citing

@article{
}