pith. machine review for the scientific record. sign in

arxiv: 1609.06026 · v3 · submitted 2016-09-20 · 💻 cs.SD · cs.LG· cs.MM

Recognition: unknown

An Approach for Self-Training Audio Event Detectors Using Web Data

Authors on Pith no claims yet
classification 💻 cs.SD cs.LGcs.MM
keywords audiodetectorseventunlabeledannotateddatasetslabeledsounds
0
0 comments X
read the original abstract

Audio Event Detection (AED) aims to recognize sounds within audio and video recordings. AED employs machine learning algorithms commonly trained and tested on annotated datasets. However, available datasets are limited in number of samples and hence it is difficult to model acoustic diversity. Therefore, we propose combining labeled audio from a dataset and unlabeled audio from the web to improve the sound models. The audio event detectors are trained on the labeled audio and ran on the unlabeled audio downloaded from YouTube. Whenever the detectors recognized any of the known sounds with high confidence, the unlabeled audio was use to re-train the detectors. The performance of the re-trained detectors is compared to the one from the original detectors using the annotated test set. Results showed an improvement of the AED, and uncovered challenges of using web audio from videos.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.