SoundDet: Polyphonic Moving Sound Event Detection and Localization from Raw Waveform
Abstract
We present a new framework SoundDet, which is an end-to-end trainable and light-weight framework, for polyphonic moving sound event detection and localization. Prior methods typically approach this problem by preprocessing raw waveform into time-frequency representations, which is more amenable to process with well-established image processing pipelines. Prior methods also detect in segment-wise manner, leading to incomplete and partial detections. SoundDet takes a novel approach and directly consumes the raw, multichannel waveform and treats the spatio-temporal sound event as a complete “sound-object” to be detected. Specifically, SoundDet consists of a backbone neural network and two parallel heads for temporal detection and spatial localization, respectively. Given the large sampling rate of raw waveform, the backbone network first learns a set of phase-sensitive and frequency-selective bank of filters to explicitly retain direction-of-arrival information, whilst being highly computationally and parametrically efficient than standard 1D/2D convolution. A dense sound event proposal map is then constructed to handle the challenges of predicting events with large varying temporal duration. Accompanying the dense proposal map are a temporal overlapness map and a motion smoothness map that measure a proposal’s confidence to be an event from temporal detection accuracy and movement consistency perspective. Involving the two maps guarantees SoundDet to be trained in a spatio-temporally unified manner. Experimental results on the public DCASE dataset show the advantage of SoundDet on both segment-based and our newly proposed event-based evaluation system.
1 Introduction
Acoustic source detection, classification and localization is a key task for a wide variety of applications, such as smart speakers/assistants, interactive robotics and scene visualization. Broadly, a microphone array (typically with four microphones) receives sound events from the environment, which are corrupted by background noise. The task is to firstly detect events, classify them, and finally estimate the physical location or direction-of-arrival angle (DoA).
Early work has demonstrated how deep learning techniques can be used to provide good performance for sound event detection and localization. However, the majority of these works treat event detection and localization as separate problems(Grondin et al., 2019)(Tho Nguyen et al., 2020). For example, a framewise classifier and regressor are used separately (Tho et al., 2014)(Grondin & Glass, 2018) to obtain event class and spatial location. In addition, they also rely heavily on handcrafted pre-processing to convert the raw waveforms into time-frequency representations (e.g. log-mel spectrograms, gcc-phat(Brandstein & Silverman, 1997)) that are more amenable to process with mature 2-D CNN networks like ResNet(He et al., 2016) followed by LSTM(Hochreiter & Jug̈en, 1997) or GRU(Chung et al., 2014) network for handling temporal dependencies. Lastly, sound event detection is typically achieved following a segment-based approach (e.g. cut the waveform into second-long snippet) which inevitably leads to incomplete or partial detection that span segment boundaries.
The relatively unexplored area becomes more challenging when it considers polyphonic event detection where events with different DoAs overlap temporally. In addition, rather than necessarily being stationary, these events undergo a motion. This increased realism leads to poor performance or even incapability of existing methods(Mesaros et al., 2019)(Grondin et al., 2019) that make a strong assumption that only a single and stationary event exists at a time slot.
In this paper, we rethink the polyphonic sound event detection and localization problem. We draw inspiration from the success of object detection in 2-D images and 3-D point clouds, and think of an event as being analogous to a “sound-object” that has a specific location (spatial and temporal onset/offset), size (duration) and class (semantic category).
Unlike computer vision where the primary challenge is occlusion of overlapping objects leading to partial views, the physics of sound are such that sound-objects superimpose and additively mix with one another. Sound objects thus have to rely on frequency-selective and phase-sensitive approaches so as to be distinguished from background or ambient noise. Rather than detecting in several separate and independent steps, a more desirable event detector needs to be end-to-end that directly consumes raw waveform and outputs predictions, whilst generates minimal computation cost. To this end, we propose a unified and light weight polyphonic sound event detection and localization framework, dubbed SoundDet, that naturally meets these requirements.
To the best of our knowledge, SoundDet is the first approach to directly consume the raw waveform and not rely on any spectral transformation pre-processing. To achieve this, we propose to learn a set of channel correlation aware and frequency sensitive bank of filters, these parametric filters naturally captures the phase difference between channels and frequency bandpass for different sound events. Comparing with standard 1D/2D convolution, learning such filter bank requires minimal computation cost and parameters, making it ideal for processing the raw waveform with large sampling rates. SoundDet takes a novel view and adopts a dense event proposal strategy to directly estimate an event’s location, size and category information, unlike current approaches that simply treat it as a segment-wise prediction problem. To better model a sound event’s completeness and continuity, we introduce concept like motion smoothness, boundary sensitive temporal overlapness.
To comprehensively evaluate the performance, we further propose an event-based evaluation system. Unlike previous segment-based evaluation, event-based metric takes sound event’s confidence score, length and motion consistency into account and accumulates performance under different score threshold. We embrace 2D image object detection metric(Lin et al., 2014) and propose to calculate average precision/recall, mean average precision (mAP), mean average recall (mAR) for each category. Experiment on 14 sound category DoA estimation DCASE dataset(Adavanne et al., 2018) shows the advantage and efficacy of SoundDet.
In summary, our main contributions are three fold: First, We propose the first framework for polyphonic moving sound event detection and localization that directly consumes raw waveform and estimate sound event from “event-object” perspective. Second, We propose a novel phase and frequency sensitive MaxCorr filter bank that is parameter and computation lightweight to process raw waveform. Third, An event-based evaluation system that avoids arbitrarily setting threshold is proposed to better evaluate the framework.
2 Related Work
Existing approaches(Grondin et al., 2019)(Pham et al., 2018)(Tho Nguyen et al., 2020) on sound event detection and localization heavily rely on 2D convolutional neural networks to process the large, multichannel waveforms. In order to use 2D CNN frameworks such as ResNet(Wang et al., 2020), the 1D waveforms are converted 2D time-frequency representations by using hard-coded methods. Typical method include short-time Fourier Transforms (FFT)(Kothinti et al., 2019; Krause & Kowalczyk, 2019), Generalized Cross-Correlation (GCC-PHAT)(Adavanne et al., 2016; Brandstein & Silverman, 1997), Mel-Spectrograms(Davis & Mermelstein, 1980; Cakir et al., 2016; Hayashi et al., 2017), Log-Mel Spectrograms(Adavanne et al., 2016; Xia et al., 2019) and pitch estimators(Adavanne et al., 2016). Contrary to these methods, SoundDet is an end-to-end and unified framework, it directly consumes raw waveform and jointly predicts a potential event by fully considering it as a holistic “sound-object”, embedded in time and space. It does not require any hand-selection and tweaking of preprocessing algorithms. Moreover, existing approaches usually chain 2D CNNs and recurrent networks (i.e. GRU(Chung et al., 2014), LSTM(Hochreiter & Jug̈en, 1997)) to incorporate temporal information. SoundDet consists of light-weight CNNs and hence has few parameters and is computationally efficient.
Fair and comprehensive evaluation of the sound event detection and localization is still an under-defined problem. Existing methods adopt segment-based evaluation, in which they segment a sound track into small chunks (i.e. 1s) and compare the result just within a segment, ignoring a sound event’s duration and continuity. In addition, the segment-based evaluation treats each detected sound event as either true positive or false positive by arbitrarily setting a threshold. It does not represent the performance under various thresholds. We draw inspiration from object detection in 2D images(Lin et al., 2014)(Ren et al., 2015)(Liu et al., 2016) and propose a new event-based evaluation metric, which instead computes mean average precision (mAP) and mean average recall (mAR) for each category accumulatively.
3 SoundDet
3.1 Problem Formulation
We have a multi-channel sound recording (aka sound waveform) recorded by one microphone array station or multiple microphone array stations at a fixed sampling rate, contains a sound event set 111In this paper, a sound event is restricted to be a temporally and spatially continuous and semantically meaningful. Two events with exactly the same temporal location belong to different class. , where , are the -th sound event’s start time and end time, respectively, indicates the sound event’s semantic information (i.e. category, baby cry, phone ring). indicates the sound event’s spatial location, it can either be a fixed location if the sound event is stationary or a spatial trajectory if it undergoes a motion. The task is to learn a model to accurately recover these sound events directly from raw sound waveform,
(1) |
These sound events are polyphonic, which means they may overlap in the time dimension.
3.2 MaxCorr Band-Pass Filter Bank and Backbone

Sound waveform usually has high temporal resolution, a typical 1 minute recording of 30kHz has 1.8 million data points. This leads to unbearable computation burden (FLOPs) for standard 1D/2D convolution. To counteract the large data point size issue and further to enforce the neural network to learn meaningful sound event spatio-temporal representation, we propose to learn a set of rectangular band-pass filters in frequency domain, each of which contains a lower learnable frequency cuttoff and a higher learnable frequency cutoff . After converting the rectangular band-pass filter to time domain, we obtain convolution kernels ,
(2) |
where . Note that the band-pass filter is a parametric filter that the learnable parameter number is independent of the kernel length. Moreover, the kernel length is usually much larger than standard 1D/2D convolution kernel length. A typical band-pass filter kernel length can be set as a large number (in our case 481), whereas the standard 1D/2D convolution kernel size is usually 3 or 5. These two properties guarantee these band-pass filters are highly efficient at both FLOPs computation and model parameter number. The SincNet filter has been successfully applied for speech recognition(Ravanelli & Bengio, 2018).

To further enforce the above designed filter banks to be able to dynamically decouple the sound event’s temporal information and spatial information. We further expand the sinc filter to multiple channels to model the internal correlation between different channels. We draw inspiration from canonical time difference of arrival (TDoA) estimation to explicitly encode sound events’ geometrical information recorded by microphone array. TDoA is a well-established technique for sound source geolocalization. It exploits the arrival time difference among microphones to recover the sound event spatial information (i.e. DoA or physical location) by computing the maximum time delay in either frequency domain or time domain(Van Den Broeck et al., 2012). We involve TDoA estimation into our band-pass filter bank design and propose a new filter bank that is capable of learning the correlation between different channels. Learning such filter bank helps the neural network to dynamically learn how to integrate between-channel correlation and event frequency band pass to best recover sound events, we thus call it MaxCorr filter bank,
(3) |
where is the -th channel time shift parameter, , , are learnable parameters and in our case as there are four channels. To make the time shift parameter to be differentiable, we adopt sigmoid like soft truncation to approximate the round operation, like (Fu et al., 2017) do. Please see Fig.2 for MaxCorr filter illustration and its comparison with standard convolution.
Following the MaxCorr filters, we add an encoder-decoder like 1D convolution network to learn framewise representations. In the encoder, we gradually reduce the temporal length but the increase the filter number, resulting in a compressed representation. In the decoder, it goes the opposite way and finally forms a “bottleneck”-like backbone neural network. To mitigate the model training degradation dilemma, we add skip connection between the encoder and decoder. The MaxCorr filter bank and encoder-decoder like neural network together constitute the Backbone neural network, which learns a framewise representation , where each representation’s temporal duration equals to ground truth labelling resolution, like 100ms.
3.3 Dense Sound Event Proposal
Given the representation learned by the backbone, the next step is to generate sound event proposal. Potential sound events freely span in the temporal and spatial space, which means sound events are largely varying in their temporal length and spatial location. An efficient sound event proposal generation module thus should be: 1. able to handle sound events with large varying temporal length and spatial location. 2. computationally efficient enough (i.e. generating potential event proposals at parallel). To this end, we propose a dense sound event proposal generation module, which organizes all potential sound events into a compact organization and all the potential events can be generated in parallelization.
Specifically, we compactly organize all potential sound events into a matrix-like representation , where the row indicates the sound event’s temporal length and the column represents its start time. Each cell in corresponds to a sound event with certain start time and end time . A careful selection of the size of matrix guarantees all potential sound events have their unique “cell” in . The matrix-like compact organization enables fast computation due to matrix parallelization computation supported by most GPU/CPU engines. Similar idea has been used for video-based activity recognition(Lin et al., 2018).
Accompanying the dense proposal map are two score maps of the same size: measuring a sound event proposal’s temporal overlap with ground truth, and sound event proposal motion smoothness map. The two score maps jointly help to measure sound event proposals’ confidence to be a true positive sound event. Specifically, the potential proposal indicated by cell is associated with a event wise feature representation , a temporal overlap confidence score from and a motion smoothness score from . derives from the averaging features learned by Backbone network spanning its temporal duration , .

3.4 Temporal Overlapness Score Map
Temporal overlap score map measures two sound events’ temporal overlap degree. Here we adopt temporal intersection over union (tIoU). Given two sound events at cell and its ground truth temporal location , their tIoU at is defined as,
(4) |
The higher of tIoU, the more likely the proposal at is a true positive proposal. Note that is class-agnostic and its responsibility is to generate all potential event proposals (also can be thought as eventness proposal). In training, we train a map regressor to learn the mapping from event wise feature representation to tIoU score. During test, we partially use the regressor to generate raw sound event proposals by choosing the large scores. To further enhance to regressor to fast localize a sound event, we propose a boundary sensitive tIoU which explicitly penalizes even a small temporal location alteration by multiplying an exponential term,
(5) |
where is a decay weight controlling tIoU decay rate, which we set 2. Our observation is that as a sound event temporal length increases, slight temporal alteration results in small tIoU score change. For example, given the ground truth sound event resides in frames [20,80] and the proposal is [22,82], the original tIoU will be very high (0.94 in this case) as there is a high degree of overlap. This does not give sufficient steering to the network to correctly align the event boundaries. Under our new metric, however, , the misalignment is prominently magnified so that the network receives enough clue to improve its localization capability.
3.5 Spatial Motion Smoothness Map
An independent sound event’s spatial trajectory should be consistent and smooth. In other words, there should be no abrupt “jump” along the sound event’s trajectory, regardless of its motion status. We call this property as motion smoothness. Specifically, we model motion smoothness as the maximum neighboring location displacement along the sound event’s motion path. Mathematically, for the sound event ’s sequential spatial location , the motion smoothness is defined as the maximum neighboring spatial location displacement ,
(6) |
is modelled as squared Euclidean distance. Small motion smoothness score serves as a strong indicator of the existence of a potential complete sound event because an inaccurate sound event localization inevitably introduces large “motion jump”. We hereby jointly use and to efficiently generate proposals. The introduction of motion smoothness map has two main advantages. First, it integrates temporal/spatial head so that the whole framework can be trained in a unified manner. Second, it explicitly adds temporal regularization to spatial location head, because spatial location head alone is a framewise regressor and no temporal priors are involved. Please see Fig.3 for dense event proposal illustration and its connection with and .
3.6 Framewise Spatial Location Regression
The final component of SoundDet is the spatial location head. As the the sound source physical location or DoA can change over the duration of an event, the spatial location is computed in a framewise manner, using the framewise features extracted by the backbone network. We regress per-class spatial location for each frame and all inactive sound classes’ spatial location is set 0.
3.7 Training Pipeline
The overall SoundDet pipeline is shown in Fig.1. The multi-channel raw waveform input is fed to the backbone neural network to efficiently learn framewise representation. The spatial location recovery head and dense event proposal generation head directly build on the learned framewise representation to learn framewise per-class sound event spatial locations and densely generate event proposals (per-event representation), respectively. The densely generated event proposals are fed to temporal recovery head to learn proposals’ category label.
In practice, the temporal location head builds on event wise feature. It consists of three sub-heads: a multi-label classification head deciding the proposal’s category label (); A binary classification head deciding whether a proposal is a foreground or background event (we can call it eventness prediction for better understanding). A tIoU regression head to learn to regress tIoU score based on the event wise feature input. The three heads consist of several full connection layers (FC). For multi-label classification and binary classification, we adopt the standard cross entropy loss. For the tIoU regression head, we adopt mean squared error (MSE) loss. The computation of requires no specific learning process because it directly derives from spatial location head, MSE loss is adopted again to reinforce event’s smoothness.
Data Imbalance The dense sound event proposal map presented above covers all possible events, it inevitably leads to the data imbalance problem because most pre-generated proposals are negative and just very few proposals are positive. We adopt two strategies to mitigate this problem, 1) increase positive proposal number, in which we set a tIoU threshold and any event with the tIoU larger than the threshold is treated as positive proposal. 2) negative proposal random dropout, with the remaining negative samples, we randomly drop part of them with a probability . By carefully setting and value, we can roughly keep positive and negative ratio to be 1:1.
3.8 Inference
During inference, the calculated and are jointly used to generate sound event proposals. The sound event proposal at cell is treated as positive proposal only if it passes two test: 1) so that it satisfies motion smoothness rule; 2) so that it resides in the designated temporal location with a high confidence. and are predefined spatial motion smoothness and temporal overlapness threshold. Once it passes the two tests, the final event detection score derives from the multiplication of its motion smoothness score, temporal overlapness score and the corresponding multilabel classification score (() discussed earlier). The class label comes from the same multilabel classification head and the spatial location derives from spatial location head.

Event Refinement The raw event candidates obtained above overlap in the temporal dimension. A post-processing is thus required to remove the redundancy (see Fig.4). The post-processing consists of two main parts: intra-class temporal non-maximum suppression (tNMS) and max-event clip. In intra-class tNMS, we compare within the same event class, for two events of the same class with temporal overlap (tNMS) above a predefined threshold, we delete the event with smaller confidence score. In max-event clip, we restrain the maximum number events that can happen at the same time. Max-event clip is a class-agnostic operation, for a set of events that happen at the same time, we sort them according the confidence score in descending order and only keep the top-M events. is a predefined max-event threshold. The post-processing is an iterative process, after which we get the final detected sound events from rw proposals .
4 Experiment
Methods | (∘) | |||||||||||||||
All | Sma | Mid | Lar | All | Sma | Med | Lar | All | Sma | Med | Lar | All | Sma | Med | Lar | |
SELDnet(foa) | 0.63 | 0.64 | 0.63 | 0.79 | 0.46 | 0.48 | 0.48 | 0.42 | 23.1 | 22.8 | 23.2 | 20.5 | 0.69 | 0.70 | 0.70 | 0.64 |
SELDnet(mic) | 0.66 | 0.68 | 0.66 | 0.79 | 0.43 | 0.44 | 0.45 | 0.42 | 24.2 | 23.9 | 23.6 | 22.8 | 0.66 | 0.65 | 0.67 | 0.66 |
EIN(Cao et al., 2020) | 0.25 | 0.30 | 0.25 | 0.29 | 0.82 | 0.80 | 0.82 | 0.83 | 8.0 | 8.1 | 8.0 | 8.5 | 0.86 | 0.85 | 0.86 | 0.84 |
SoundDet_backbone(foa) | 0.74 | 0.79 | 0.75 | 0.74 | 0.38 | 0.37 | 0.40 | 0.38 | 21.7 | 22.2 | 16.7 | 21.7 | 0.57 | 0.57 | 0.51 | 0.57 |
SoundDet_backbone(mic) | 0.80 | 0.85 | 0.82 | 0.90 | 0.30 | 0.29 | 0.31 | 0.34 | 28.6 | 29.5 | 28.6 | 21.0 | 0.56 | 0.56 | 0.57 | 0.54 |
SoundDet_nomaxcorr(foa) | 0.90 | 0.91 | 0.90 | 0.90 | 0.05 | 0.05 | 0.06 | 0.06 | 79.4 | 79.3 | 79.8 | 79.0 | 0.50 | 0.51 | 0.53 | 0.48 |
SoundDet_nomaxcorr(mic) | 0.90 | 0.95 | 0.95 | 0.90 | 0.04 | 0.04 | 0.04 | 0.04 | 87.3 | 85.9 | 87.4 | 91.3 | 0.48 | 0.48 | 0.49 | 0.49 |
SoundDet_nomots(foa) | 0.31 | 0.37 | 0.38 | 0.35 | 0.77 | 0.72 | 0.74 | 0.72 | 9.0 | 10.1 | 9.8 | 7.4 | 0.77 | 0.72 | 0.74 | 0.76 |
SoundDet(foa) | 0.25 | 0.31 | 0.24 | 0.26 | 0.81 | 0.79 | 0.83 | 0.86 | 8.3 | 8.5 | 7.7 | 8.1 | 0.82 | 0.80 | 0.89 | 0.85 |
SoundDet is capable of estimating 2D/3D DoA, physical location, distance and motion of various sound events in polyphonic and moving scenario. In this experiment, we focus on indoor 3D DoA estimation tasks.
Dataset We evaluate SoundDet on TAU-NIGENS DCASE sound event detection and localization (SELD)(Adavanne et al., 2018) dataset. It is an indoor synthetic sound recording collected by an Eigenmike spherical microphone array and available in two data formats: first-order ambisonics (FOA) and tetrahedral microphone array (MIC). MIC indicates four microphones with different orientations, usually in spherical coordinates. FoA is known as B-format, consisting of omni-directional and direction components. The possible azimuths of the sound recording span the whole range and the elevations lie in range . Recorded sound events are either stationary or moving, and a maximum of two sound events overlap spatially and temporally. In total, 14 sound event categories are generated: alarm, crying baby, crash, barking dog, running engine, female scream, female speech, burning fire, footsteps, knocking on door, male scream, male speech, ringing phone and piano. Each sound recording is 1 minute long with sampling rate 24 kHz. For more details about this dataset, please refer to (Adavanne et al., 2018). There are 8 folds recordings, each of which has 100 1-min format recording. We follow the official splits and use 1-6 folds for train and the remaining 7-8 folds (200 1-min recordings) for test.
4.1 Evaluation Metrics
Segment-based Metric is the standard evaluation metric for existing methods (Grondin et al., 2019),(Tho et al., 2014),(Mesaros et al., 2019). It compares prediction and ground truth within non-overlapping short temporal segments (one sec.) and evaluates each frame’s prediction result by jointly considering its predicted class and spatial location. For event detection evaluation, it calculates -score () and Error Rate (), where a true positive prediction has to be within a spatial distance threshold with the corresponding ground truth spatial location (typically with an angular threshold ). For event localization evaluation, we compute class dependent localization error which measures the average location distance between prediction and ground truth, and localization recall measures the ratio of how many such localizations were estimated within a class. For details, please refer to (Mesaros et al., 2019).
Event-based Metric draws inspiration from evaluation on 2D image object detection. It treats a sound event as an independent instance with specific start time, end time, framewise spatial location and confidence score belonging to one class (see Sec. 3.8). Rather than simply binarizing a detection as false or true detection with an arbitrary threshold, event-based metric comprehensively evaluates performance under various event confidence score and tIoU thresholds, calculating average precision (AP) and average recall (AR) for each class separately. Finally, mean average precision (mAP) and mean average recall (mAR) is accumulated by averaging AP and AR throughout classes respectively.
Specifically, given detected sound events and ground truth sound events of the same class, we compute the average precision and average recall under tIoU in range , where is the stepsize. For a particular tIoU, the AP and AP is computed by averaging precision scores and recall scores obtained under various confidence score in range . Please note that we don’t take any predefined tIoU and confidence score threshold because arbitrary chosen thresholds inevitably introduces human bias e.g. the choice of an entirely arbitrary angular threshold in the segment based approach. Our proposed event-based metric instead provides a more objective and comprehensive metric.
4.2 Comparing Methods
We compare SoundDet with two existing methods: SELDNet(Grondin et al., 2019) and EIN(Cao et al., 2020). Since we focus on polyphonic and moving scenario, other relevant methods(Tho Nguyen et al., 2020)(Cao et al., 2019) that merely work on stationary and monophonic sound events are not discussed here. SELDNet(Grondin et al., 2019) jointly trains sound event detection and localization with CRNN network. It extracts Log-mel, GCC-PHAT, Intensity features as neural network input. Bidirectional GRU network is involved to model temporal dependency. SELDNet is treated as the baseline. EIN(Cao et al., 2020) is a very recent work which models sound event detection and localization with two identical but separate neural networks as a multi-task learning format. It uses a log-mel spectrogram as input for event detection, and a GCC-PHAT approach for DoA estimation. The two parallel networks are independent but with soft parameter sharing. In addition, multi-head self-attention(Vaswani et al., 2017) is applied. EIN has large parameter size and is currently the state of the art approach under segment-based evaluation metric.
The above two methods generate framewise predictions, in which each frame is associated with class classification score and predicted DoA value. In order to generate a score that metrics event classification score and DoA closeness between predicted location and ground truth location, we propose to integrate mean classification score and mean DoA Euclidean distance together in the formula . For fair comparison, we don’t involve any data augmentation methods. We report the result for both FOA and MIC data format if possible.
Within SoundDet, we want to answer four main questions: 1. If our proposed MaxCorr band-pass filters is capable of learning useful representation for sound event spatio-temporal information recovery? 2. Event length impact on SoundDet and other methods. 3. Efficiency comparison in terms of parameters and inference time. 4. Quantitative comparison between SoundDet and other methods. To this end, we test four SoundDet variants: backbone only SoundDet with MaxCorr filter (SoundDet_backbone) and without MaxCorr filter (SoundDet_nomaxcorr). SoundDet without motion smoothness (SoundDet_nomots). Moreover, in addition to report the overall metric, we further divide the events into Small (0-2s), Medium (2s-7s) and Large (7s), three categories and report result for them separately.
SoundDet network architecture is shown Table 4. Please note that the network might need to be slightly changed to process waveform with different input length, sample rate and label interval length. and indicate dense proposal map height and width, respectively, in our experiment and , they can be directly modified to fit other input waveform temporal length or ground truth labelling resolution.
Training Detail We adopt multi-stage training strategy. First, we train the SoundDet backbone network in a framewise DoA regression and multilabel classification manner, like SELDNet(Grondin et al., 2019) and EIN(Cao et al., 2020) do. Training SoundDet backbone neural network at first provides two advantages: 1) it helps to test if our proposed MaxCorr filter bank helps to learn essential representation for sound event temporal detection and spatial localization (as we compare in the experiment), 2) it guarantees to learn a reasonable framewise representation first so that the later joint training (combine the backbone and two heads) converges much faster. For the backbone training, we use SGD optimizer with an initial learning rate 0.5, the learning rate decays every 30 epochs with decay rate 0.7. With the pretrained Backbone network, we continue to train the whole SoundDet network with the same optimizer.


4.3 Experiment Results
Methods | Params | Input | mAP | mAR | ||||||
Overall | Small | Medium | Large | Overall | Small | Medium | Large | |||
SELDNet(foa) | 0.5M | 1,3 | 0.087 | 0.038 | 0.092 | 0.157 | 0.152 | 0.079 | 0.096 | 0.097 |
SELDNet(mic) | 0.5M | 1,2 | 0.079 | 0.035 | 0.081 | 0.143 | 0.140 | 0.067 | 0.099 | 0.086 |
EIN(Cao et al., 2020) | 26.0M | 1,2 | 0.134 | 0.088 | 0.187 | 0.183 | 0.256 | 0.175 | 0.186 | 0.257 |
SoundDet_backbone(foa) | 10.0 M | 0 | 0.040 | 0.025 | 0.055 | 0.090 | 0.096 | 0.066 | 0.051 | 0.033 |
SoundDet_backbone(mic) | 10.0 M | 0 | 0.035 | 0.021 | 0.091 | 0.090 | 0.094 | 0.065 | 0.049 | 0.027 |
SoundDet_nomaxcorr(foa) | 10.0 M | 0 | 0.025 | 0.015 | 0.034 | 0.080 | 0.063 | 0.040 | 0.032 | 0.014 |
SoundDet_nomaxcorr(mic) | 10.0 M | 0 | 0.017 | 0.059 | 0.016 | 0.044 | 0.041 | 0.017 | 0.079 | 0.016 |
SoundDet_nomots(foa) | 13.0 M | 0 | 0.117 | 0.062 | 0.173 | 0.152 | 0.247 | 0.174 | 0.159 | 0.232 |
SoundDet(foa) | 13.0 M | 0 | 0.197 | 0.098 | 0.201 | 0.216 | 0.294 | 0.189 | 0.223 | 0.289 |
The segment-based evaluation result is shown in Table 1. We can see that, by involving MaxCorr filter banks, our backbone-only SoundDet achieves comparable performance comparing with SELDNet(Grondin et al., 2019). SELDNet extracts Log-mel, GCC-PHAT and Intensity features (see Table 2) as its neural network input, replacing these hand-crafted features with learnable MaxCorr filter bank helps to achieve similar performance, even with 1D convolution. If disabling SoundDet to learn between-channel correlation (SoundDet_nomaxcorr), we can witness a huge performance drop in both FOA and MIC format, especially the high event detection error rate and large DoA error angle. This shows that our proposed MaxCorr filter bank is capable of learning essential representations for recovering temporal and spatial information. Actually, we observed a fast convergence of the between-channel time shift parameters (see Eqn.3), SoundDet quickly updates time shift parameters in the very first couple of training iterations and swiftly achieves a relatively stable state. Two learned MaxCorr filters are shown in Fig.5, from which we can observe that different MaxCorr filters have learned different time shifts and frequency cutoffs. These learned frequency-selective and phase-sensitive MaxCorr filter bank is of vital importance for sound event representation learning.
Both SoundDet and SELDNet work better for FOA format than MIC format. Framewise based learning (both SoundDet, SELDNet and EIN) shows no obvious difference in estimating events with different temporal length, which is reasonable because of their framewise processing property. The whole SoundDet (combine backbone and two heads) far outperforms SELDNet baseline and achieves comparable performance with EIN with much less parameter number. Specifically, SoundDet shows advantage on longer sound event estimation. It thus attests the necessity of directly modelling “sound-object” when it comes to events with longer temporal length or more complex motion trajectory. Excluding motion smoothness largely reduces the performance, we observed involving motion smoothness map greatly helps the neural network to accurate localize sound events.
The event-based evaluation result is shown in Table 2, from which we can observe that SoundDet outperforms all existing methods in both the mAP and mAR metric. This shows that the segment-based metrics in current use do not comprehensively reflect an algorithm’s performance. Rather than focusing on a local segment evaluation, the event-based metric highlights an event’s completeness and continuity and SoundDet is naturally designed to meet these requirements. This advantage is echoed by the qualitative comparison in Fig. 6. We can clearly see that SELDNet and EIN inevitably contain many mixed sound events and frequently cut complete events into small disconnected segments. The situation becomes more serious when faced with overlapping situation. However, SoundDet better avoids this dilemma because by design it treats events as complete instances, rather than discrete segments.
SELDNet | EIN | SoundDet |
1.20 s | 2.20 s | 1.25 s |
We further report the average inference time of different methods to process a one-minute long audio in Table. 3, showing that it is almost twice as fast as EIN, and comparable to SELDNet. In summary, SoundDet is capable of learning a sound event’s spatial, temporal and class information from raw waveforms, achieving competing performance under existing segment-based metrics and leading results on the proposed event-based metrics.
5 Conclusion
We have introduced a number of innovations in this paper that attempt to unify the currently disparate fields of object detection in computer vision, with event detection and localization in audio settings. We also abandon prior approaches that resort to hand-crafted pre-processing and transformation of the multi-channel waveforms, and instead directly consume raw data. To this end, we propose a more object-centric approach to measuring the performance of sound event detection and localization through metrics of mAP and mAR. It is our hope that this alternative view will seed further development in this space. One potential future research direction is to design more elegant learnable filter banks.
6 Acknowledgement
We thank Qi Ou from Department of Earth Science, University of Oxford for helpful discussion. We also appreciate reviewers’ constructive feedback to improve the work.
layer | filter num | output size |
Input: [B,4,T] | ||
SoundDet Backbone Network | ||
MaxCorr@251,75 | 256 | [B, T/75,256] |
conv1d@3,2 | 128 | [B, T/150,128] |
conv1d@3,2 | 128 | [B, T/300,128] |
conv1d@3,2 | 256 | [B, T/600,256] |
conv1d@3,2 | 256 | [B, T/1200,512] |
conv1d@3,2 | 512 | [B, T/2400,512] |
conv1d@3,2 | 512 | [B, T/4800,512] |
conv1d@3,2 | 1024 | [B, T/9600,1024] |
deconv1d@3,2 | 512 | [B, T/4800,512] |
deconv1d@3,2 | 512 | [B, T/2400, 512] |
biGRU | 512 | [B, T/2400,512] |
conv1d@3,1 | 512 | [B, T/2400,512] |
Event Multilabel Classify | ||
FC | 512 | [B, H, W, 512] |
FC | 256 | [B, H, W, 256] |
FC | class num | [B, H, W, cn] |
Eventness Classify | ||
FC | 256 | [B, H, W, 256] |
FC | 1 | [B, H, W, 1] |
tIoU Map Regression | ||
FC | 256 | [B, H, W, 256] |
FC | 1 | [B, H, W, 1] |
Spatial Location Head | ||
FC | 512 | [B, T/2400, 512] |
FC | 512 | [B, T/2400, 512] |
FC | 256 | [B, T/2400, 256] |
FC | classnum3 | [B,T/2400,cn3] |
References
- Adavanne et al. (2016) Adavanne, S., Parascandolo, G., Pertilä, P., Heittola, T., and Virtanen, T. Sound event detection in multichannel audio using spatial and harmonic features. In Scenes and Events 2016 Workshop (DCASE2016), 2016.
- Adavanne et al. (2018) Adavanne, S., Politis, A., Nikunen, J., and Virtanen, T. Sound event localization and detection of overlapping sources using convolutional recurrent neural networks. In IEEE Journal of Selected Topics in Signal Processing, 2018.
- Brandstein & Silverman (1997) Brandstein, M. S. and Silverman, H. F. A robust method for speech signal time-delay estimation in reverberant rooms. In IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 1997.
- Cakir et al. (2016) Cakir, E., Ozan, E. C., and Virtanen, T. Filterbank learning for deep neural network based polyphonic sound event detection. In 2016 International Joint Conference on Neural Networks (IJCNN), pp. 3399–3406. IEEE, 2016.
- Cao et al. (2019) Cao, Y., Kong, Q., Iqbal, T., An, F., Wang, W., and Plumbley, M. Polyphonic sound event detection and localization using a two-stage strategy. In DCASE workshop, 2019.
- Cao et al. (2020) Cao, Y., Iqbal, T., Kong, Q., An, F., Wang, W., and Plumbley, M. D. An improved event-independent network for polyphonic sound event localization and detection. In arXiv: 2010.13092, 2020.
- Chung et al. (2014) Chung, J., Gulcehre, C., Cho, K., and Bengio, Y. Empirical evaluation of gated recurrent neural ntworks on sequence modelling. In Adv. Neural Inform. Process. Syst., 2014.
- Davis & Mermelstein (1980) Davis, S. and Mermelstein, P. Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences. In In IEEE Transactions on Acoustics, Speech, and Signal Processing, 1980.
- Fu et al. (2017) Fu, J., Zheng, H., and Mei, T. Look closer to see better: Recurrent attention convolutional neural network for fine-grained image recognition. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- Grondin & Glass (2018) Grondin, F. and Glass, J. A study of the complexity and accuracy of direction of arrival estimation methods based on gcc-phat for a pair of close microphones. In arxiv:1811.11787, 2018.
- Grondin et al. (2019) Grondin, F., Glass, J., Sobieraj, I., and Mark D., P. A study of the complexity and accuracy of direction of arrival estimation methods based on gcc-phat for a pair of close microphones. In Proceedings of the Detection and Classification of Acoustic Scenes and Events Workshop, 2019.
- Hayashi et al. (2017) Hayashi, T., Watanabe, S., Toda, T., Hori, T., Le Roux, J., and Takeda, K. Duration-controlled lstm for polyphonic sound event detection. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 25(11):2059–2070, 2017.
- He et al. (2016) He, K., Zhang, X., Ren, S., and Sun, J. Identity mappings in deep residual networks. In European Conference on Computer Vision (ECCV), 2016.
- Hochreiter & Jug̈en (1997) Hochreiter, S. and Jug̈en, S. Long short-term memory. In Neural Computation, 1997.
- Kothinti et al. (2019) Kothinti, S., Imoto, K., Chakrabarty, D., Sell, G., Watanabe, S., and Elhilali, M. Joint acoustic and class inference for weakly supervised sound event detection. In ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 36–40. IEEE, 2019.
- Krause & Kowalczyk (2019) Krause, D. and Kowalczyk, K. Arborescent neural network architectures for sound event detection and localization. Technical report, DCASE2019 Challenge, Tech. Rep, 2019.
- Lin et al. (2018) Lin, T., Zhao, X., Su, H., Wang, C., and Yang, M. Bsn: Boundary sensitive network for temporal action proposal generation. In European Conference on Computer Vision (ECCV), 2018.
- Lin et al. (2014) Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., and Zitnick, C. L. Microsoft coco: Common objects in context. In Fleet, D., Pajdla, T., Schiele, B., and Tuytelaars, T. (eds.), European Conference on Computer Vision (ECCV), 2014.
- Liu et al. (2016) Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., and Berg, A. Ssd: Single shot multibox detector. In European Conference on Computer Vision (ECCV), 2016.
- Mesaros et al. (2019) Mesaros, A., Adavanne, S., Politis, A., Heittola, T., and Virtanen, T. Joint measurement of localization and of detection of sound event. In 9 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, 2019.
- Pham et al. (2018) Pham, P., Li, J., Szurley, J., and Das, S. Eventness: Object detection on spectrograms for temporal localization of audio events. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2018.
- Ravanelli & Bengio (2018) Ravanelli, M. and Bengio, Y. Speaker recognition from raw waveform with sincnet. In In IEEE Workshop on Spoken Language Technology (SLT), 2018.
- Ren et al. (2015) Ren, S., He, K., Girshich, R., and Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems (NeurIPS), 2015.
- Tho et al. (2014) Tho, N., Zhao, S., and Jones, D. Robust doa estimation of multiple speech sources. In IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP, pp. 2287–2291, 2014.
- Tho Nguyen et al. (2020) Tho Nguyen, T. N., Jones, D. L., and Gan, W. A sequence matching network for polyphonic sound event localization and detection. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020.
- Van Den Broeck et al. (2012) Van Den Broeck, B., Bertrand, A., Karsmakers, P., Vanrumste, B., Van hamme, H., and Moonen, M. Time-domain generalized cross correlation phase transform sound source localization for small microphone arrays. In the 5th European DSP in Education and Research Conference, 2012.
- Vaswani et al. (2017) Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., and Polosukhin, I. Attention is all you need. In Adv. Neural Inform. Process. Syst., 2017.
- Wang et al. (2020) Wang, Q., Wu, H., Jing, Z., Ma, F., Fang, Y., Wang, Y., Chen, T., Pan, J., Du, J., and Lee, C.-h. The ustc-iflytek system for sound event localization and detection of dcase2020 challenge. In DCASE workshop, 2020.
- Xia et al. (2019) Xia, X., Togneri, R., Sohel, F., Zhao, Y., and Huang, D. Multi-task learning for acoustic event detection using event and frame position information. IEEE Transactions on Multimedia, 22(3):569–578, 2019.