BeepBank-500: A Synthetic Earcon Mini-Corpus for UI Sound Research and Psychoacoustics Research
Abstract
BeepBank-500 is a synthetic earcon/alert dataset for audio machine learning, featuring parametrically generated clips with various waveform families and reverberation settings.
We introduce BeepBank-500, a compact, fully synthetic earcon/alert dataset (300-500 clips) designed for rapid, rights-clean experimentation in human-computer interaction and audio machine learning. Each clip is generated from a parametric recipe controlling waveform family (sine, square, triangle, FM), fundamental frequency, duration, amplitude envelope, amplitude modulation (AM), and lightweight Schroeder-style reverberation. We use three reverberation settings: dry, and two synthetic rooms denoted 'rir small' ('small') and 'rir medium' ('medium') throughout the paper and in the metadata. We release mono 48 kHz WAV audio (16-bit), a rich metadata table (signal/spectral features), and tiny reproducible baselines for (i) waveform-family classification and (ii) f0 regression on single tones. The corpus targets tasks such as earcon classification, timbre analyses, and onset detection, with clearly stated licensing and limitations. Audio is dedicated to the public domain via CC0-1.0; code is under MIT. Data DOI: https://doi.org/10.5281/zenodo.17172015. Code: https://github.com/mandip42/earcons-mini-500.
Community
We introduce BeepBank, a compact synthetic dataset (300–500 clips) of earcons and alerts, designed for rapid, rights-clean experimentation in human–computer interaction and audio machine learning.
🔹 Fully synthetic waveforms (sine, square, triangle, FM)
🔹 Rich metadata (signal/spectral features)
🔹 Baselines for waveform classification & f₀ regression
🔹 Public domain audio (CC0-1.0), MIT-licensed code
📄Paper: [doi.org/10.48550/arXiv.2509.17277]
📦 Dataset: [10.5281/zenodo.17172015]
💻 Code: [https://github.com/mandip42/earcons-mini-500]
Use it for timbre analysis, earcon classification, onset detection, and more. Contributions welcome!
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Automotive Sound Quality for EVs: Psychoacoustic Metrics with Reproducible AI/ML Baselines (2025)
- Exploring Procedural Data Generation for Automatic Acoustic Guitar Fingerpicking Transcription (2025)
- SonicMaster: Towards Controllable All-in-One Music Restoration and Mastering (2025)
- Vevo2: Bridging Controllable Speech and Singing Voice Generation via Unified Prosody Learning (2025)
- SwiftF0: Fast and Accurate Monophonic Pitch Detection (2025)
- FCPE: A Fast Context-based Pitch Estimation Model (2025)
- The CCF AATC 2025: Speech Restoration Challenge (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper