• data process

  • Place the WAV and JSON files in dev_data.

To distinguish the recognition performance of each part, the file names of the training audio for Part One need to be prefixed with fold1-d-, those for Part Two need to be prefixed with fold1-a-, fold1-b-, fold1-c-, and those for Part Three need to be prefixed with fold1-e-. If the training audio file names for Part One and Part Three do not have the prefixes fold1-d- and fold1-e-, you will need to add them yourself. For example, if the file name of the training audio for Part One is 5402400A, then add the prefix to make it fold1-d-5402400A. If the file name of the audio for Part Three is audio_0001405, then add the prefix to make it fold1-e-audio_0001405. The process for the development set is the same, except that fold1 should be changed to fold2.

Download the pre-trained Sentence-BERT model and tokenizer from the following URL and Place the downloaded pre-trained model and tokenizer inside the ../../qwen2_audio_baseline/Bert_pretrain

  • Example commands
git clone https://huggingface.co/PeacefulData/2025_DCASE_AudioQA_Baselines
cd 2025_DCASE_AudioQA_Baselines
mkdir Bert_pretrain
cd Bert_pretrain
git clone https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2/tree/main
  • Environment
cd ../qwen2_audio_baseline
pip install -r requirements.txt

You can also use a mirror source to speed up the process. pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple

  • Run Audio QA Inference Baseline
sh qwen_audio_test.sh
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for PeacefulData/2025_DCASE_AudioQA_Baselines

Finetuned
(5)
this model