Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,52 @@
|
|
1 |
-
---
|
2 |
-
license: cc-by-4.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-4.0
|
3 |
+
tags:
|
4 |
+
- autrainer
|
5 |
+
---
|
6 |
+
|
7 |
+
# Baseline model for audio classification of orthopera and hemiptera
|
8 |
+
|
9 |
+
The baseline model which was used in the ECOSoundSet-paper, in order to tag audio files, with one or more of 86 species belonging to the orthoptera or hemiptera insect orders.
|
10 |
+
|
11 |
+
## Installation
|
12 |
+
|
13 |
+
To use the model, you have to install autrainer, e.g. via pip:
|
14 |
+
|
15 |
+
`pip install autrainer`
|
16 |
+
|
17 |
+
## Usage
|
18 |
+
|
19 |
+
The model can be applied on all wav files present in a folder (`<data-root>`) and stored in another folder (`<output-root>`):
|
20 |
+
|
21 |
+
`autrainer inference hf:autrainer/edansa-2019-cnn10-32k-t <data-root> <output-root>`
|
22 |
+
|
23 |
+
## Training
|
24 |
+
|
25 |
+
### Pretraining
|
26 |
+
|
27 |
+
The model has been originally trained on AudioSet by Kong et. al.
|
28 |
+
|
29 |
+
### Dataset
|
30 |
+
|
31 |
+
The model has been further trained (finetuned) on audio segments of the ECOSoundSet dataset, which is soon to be submitted,
|
32 |
+
and will referenced here as soon as it is publicly available.
|
33 |
+
|
34 |
+
|
35 |
+
### Features
|
36 |
+
|
37 |
+
The audio recordings were resampled to 96kHz, as we wanted to avoid losing too much frequency information from the species. Log-Mel spectrograms were then extracted using torchlibrosa.
|
38 |
+
|
39 |
+
### Training process
|
40 |
+
|
41 |
+
The model has been trained for 30 epochs. At the end of each epoch, the model was evaluated on our validation set. We release the state that achieved the best performance on this validation set. All training hyperparameters can be found inside `conf/config.yaml` inside the model folder. The train, dev, and test set can be accessed in the `splits`-folder.
|
42 |
+
|
43 |
+
|
44 |
+
## Evaluation
|
45 |
+
|
46 |
+
The performance on the test set reached a 0.56 (macro) f1-score.
|
47 |
+
|
48 |
+
|
49 |
+
## Acknowledgments
|
50 |
+
|
51 |
+
Please acknowledge the work which produced the original model and the ECOSoundSet dataset. We would also appreciate an acknowledgment to autrainer.
|
52 |
+
|