This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

SICRN: Advancing Speech Enhancement through State Space Model and Inplace Convolution Techniques

Abstract

Speech enhancement aims to improve speech quality and intelligibility, especially in noisy environments where background noise degrades speech signals. Currently, deep learning methods achieve great success in speech enhancement, e.g. the representative convolutional recurrent neural network (CRN) and its variants. However, CRN typically employs consecutive downsampling and upsampling convolution for frequency modeling, which destroys the inherent structure of the signal over frequency. Additionally, convolutional layers lacks of temporal modelling abilities. To address these issues, we propose an innovative module combing a State space model and Inplace Convolution (SIC), and to replace the conventional convolution in CRN, called SICRN. Specifically, a dual-path multidimensional State space model captures the global frequencies dependency and long-term temporal dependencies. Meanwhile, the 2D-inplace convolution is used to capture the local structure, which abandons the downsampling and upsampling. Systematic evaluations on the public INTERSPEECH 2020 DNS challenge dataset demonstrate SICRN’s efficacy. Compared to strong baselines, SICRN achieves performance close to state-of-the-art while having advantages in model parameters, computations, and algorithmic delay. The proposed SICRN shows great promise for improved speech enhancement.

Index Terms—  Speech enhancement, state space model, deep multidimensional state space model, inplace convolution

1 Introduction

In recent years, the incorporation of deep learning techniques into single-channel speech enhancement methods leads to noteworthy improvements in the speech quality and intelligibility of enhancement systems. The time domain speech enhancement methods [1, 2] utilize neural networks to map noisy speech waveforms and directly enhance speech waveforms. Frequency domain enhancement techniques [3] typically employ noise spectral characteristics (e.g., complex spectrum, magnitude spectrum, cepstrum [4], etc.) as inputs for neural models. The learning target is generally clean speech or a mask (eg, ideal ratio masks[5], complex ideal ratio masks [6], etc.). Generally speaking, owing to the substantial computational demands of time domain signals and the uncertain representation of feature dimensions, frequency domain methods continue to dominate the landscape of speech enhancement techniques.

In the conventional CRN [7] structure, frequency domain enhancement techniques are employed. The stride of convolutional operation in the frequency dimension is normally set to 2, which shrinks the feature in the frequency dimension. By stacking the convolutional layers several times, the patterns lying in the frequency dimension are encoded into the channel dimension. This downsampling operation compromises the inherent feature structure of the original speech, consequently constraining performance. Therefore, we propose inplace convolution [8, 9, 10] method has a significant effect for speech enhancement and acoustic echo cancellation, essentially setting the convolution kernel stride to 1. Because it does not necessitate the downsampling and upsampling of speech features, It can extract the inherent characteristic information of the original speech. This extraction occurs without detriment to the amplitude spectrum harmonics and spatial position information. However, the absence of downsampling operations in inplace convolution makes it challenging to obtain full-band correlations.

Hao et al.[11] propose FullSubNet, a method that achieves effective integration of full-band and sub-band information without downsampling operations, showing remarkable performance. However, due to the introduction of the full-band model, these methods lead to a large number of overall model parameters, increasing the complexity of the model. Moreover, while FullSubNet incorporates future frame information and adheres to real-time constraints, it does not strictly adhere to the principles of causal networks.

In this study, we propose an innovative module that combine multidimensional state space model [12] and inplace convolutions for speech enhancement. While reducing the amount of parameters and calculations, it solves the problem of speech downsampling destroying the original features and better extracting and fusing local information and global information to improve speech quality and intelligibility. Specifically, inplace convolution excels in extracting local features and reconstructing speech signals without compromising the integrity of the original feature information. However, it lacks full-band correlation. On the other hand, S4ND proves beneficial for capturing global features while also preserving the integrity of the original feature information. But S4ND lacks detailed sub-band feature information. As a result, we propose combining S4ND and inplace convolution to leverage their respective strengths and compensate for each other’s weaknesses. The experimental results demonstrate notably high evaluation scores. Notably, this achievement is attained with less than 1/2 the parameters and 1/7 the computational complexity of FullSubNet. Furthermore, SICRN operates without reliance on future frames for enhancement, adhering to the principles of causal networks. The contributions are as follows:

1. We propose an innovative module called SIC for speech enhancement by combining S4ND and inplace convolution.

2. SICRN attains a remarkable level of performance while utilizing merely 2.16 M parameters and 4.24 G/s MACs.

2 Related Work

2.1 S4- State Space Model

The recently proposed deep neural state-space model(SSM) [13] advances speech tasks by combining the properties of both CNNs and RNNs. The SSM is defined in continuous time using the following equations:

h(t)=Ah(t)+Bx(t)h^{\prime}(t)=Ah(t)+Bx(t) (1)
y(t)=Ch(t)+Dx(t)y(t)=Ch(t)+Dx(t) (2)

To be applied on a discrete input sequence (u0, u1, . . .) instead of continuous function u(t), (1) must be discretized by a step size Δ\operatorname{\Delta} that represents the resolution of the input. The discrete SSM is

xk=𝑨¯xk1+𝑩¯ukyk=𝑪¯xkx_{k}=\overline{\boldsymbol{A}}x_{k-1}+\overline{\boldsymbol{B}}u_{k}\quad y_{k}=\overline{\boldsymbol{C}}x_{k} (3)
𝑨¯=(𝑰Δ/2𝑨)1(𝑰+Δ/2𝑨)\overline{\boldsymbol{A}}=(\boldsymbol{I}-\Delta/2\cdot\boldsymbol{A})^{-1}(\boldsymbol{I}+\Delta/2\cdot\boldsymbol{A}) (4)

where 𝑨¯\overline{\boldsymbol{A}},𝑩¯\overline{\boldsymbol{B}},𝑪¯\overline{\boldsymbol{C}} are the discretized state matrices.According to the conclusion in [13], it can be seen that:

yk=𝑪𝑨¯k𝑩¯u0+𝑪𝑨¯k1𝑩¯u1++𝑪𝑨𝑩¯uk1+𝑪𝑩¯uky_{k}=\overline{\boldsymbol{CA}}^{k}\overline{\boldsymbol{B}}u_{0}+\overline{\boldsymbol{CA}}^{k-1}\overline{\boldsymbol{B}}u_{1}+\cdots+\overline{\boldsymbol{CA}\boldsymbol{B}}u_{k-1}+\overline{\boldsymbol{C}\boldsymbol{B}}u_{k} (5)
y=𝑲¯uy=\overline{\boldsymbol{K}}*u (6)
𝑲¯=(𝑪𝑩¯,𝑪𝑨𝑩¯,,𝑪𝑨¯L1𝑩¯)\overline{\boldsymbol{K}}=\left(\overline{\boldsymbol{C}\boldsymbol{B}},\overline{\boldsymbol{CAB}},\ldots,\overline{\boldsymbol{CA}}^{L-1}\overline{\boldsymbol{B}}\right) (7)

In other words, (5) is a single (non-circular) convolution and can be computed very efficiently with FFTs, provided that 𝑲¯\overline{\boldsymbol{K}} is known. For the specific details of SSM, you can refer to [13, 14] to understand.

2.2 S4ND- Multidimensional State Space Model

The S4 layer was developed for 1-D inputs, which limits its applicability. In [13, 14], the input dimension is 2-D, the shape is (H, T), and the S4 layer is designed as H independent parallel calculations. Since there are no correlations in the H dimension, this is limited in speech signal processing. So S4ND compensates for the correlation in the frequency dimension. In [12], the conventional S4 layer was extended to multidimensional signals by turning the standard SSM (1-D ODEs) into multidimensional partial differential equations (PDEs) governed by an independent SSM in each dimension.

Let u=u(t(1),t(2))u=u(t^{(1)},t^{(2)}) and y=y(t(1),t(2))y=y(t^{(1)},t^{(2)}) be the input and output which are signals 2\mathbb{R}^{2}\rightarrow\mathbb{C}, and x=(x(1)(t(1),t(2))x=(x^{(1)}(t^{(1)},t^{(2)}), x(2)(t(1),t(2)))N(1)×N(2)x^{(2)}(t^{(1)},t^{(2)}))\in\mathbb{C}^{N^{(1)}\times N^{(2)}} be the SSM state of dimension N(1)×N(2)N^{(1)}\times N^{(2)}, where x(τ):2N(τ)x^{(\tau)}:\mathbb{R}^{2}\rightarrow\mathbb{C}^{N^{(\tau)}}. The 2D SSM is the map u \mapsto y defined by the linear PDE with initial condition x(0, 0) = 0 :

t(1)x(t(1),t(2))\displaystyle\frac{\partial}{\partial t^{(1)}}x(t^{(1)},t^{(2)}) =(𝑨(1)x(1)(t(1),t(2)),x(2)(t(1),t(2)))+\displaystyle=(\boldsymbol{A}^{(1)}x^{(1)}(t^{(1)},t^{(2)}),x^{(2)}(t^{(1)},t^{(2)}))+ (8)
𝑩(1)u(t(1),t(2))\displaystyle\phantom{=}\ \ \boldsymbol{B}^{(1)}u(t^{(1)},t^{(2)})
t(2)x(t(1),t(2))\displaystyle\frac{\partial}{\partial t^{(2)}}x(t^{(1)},t^{(2)}) =(x(1)(t(1),t(2)),𝑨(2)x(2)(t(1),t(2)))+\displaystyle=(x^{(1)}(t^{(1)},t^{(2)}),\boldsymbol{A}^{(2)}x^{(2)}(t^{(1)},t^{(2)}))+
𝑩(2)u(t(1),t(2))\displaystyle\phantom{=}\ \ \boldsymbol{B}^{(2)}u(t^{(1)},t^{(2)})
y(t(1),t(2))\displaystyle y(t^{(1)},t^{(2)}) =𝑪,x(t(1),t(2))\displaystyle=\langle\boldsymbol{C},x(t^{(1)},t^{(2)})\rangle

Note that (8) differs from the usual notion of multidimensional SSM, which is simply a map from u(t)ny(t)mu(t)\in\mathbb{C}^{n}\mapsto y(t)\in\mathbb{C}^{m} for higher-dimensional n, m >> 1 but still with 1 time axis. However, (8) is a map from u(t1,t2)1y(t1,t2)1u\left(t_{1},t_{2}\right)\in\mathbb{C}^{1}\mapsto y\left(t_{1},t_{2}\right)\in\mathbb{C}^{1} for scalar input/outputs but over multiple time axes. When thinking of the input u(t(1),t(2))u(t^{(1)},t^{(2)}) as a function over a 2D grid, (8) can be thought of as a simple linear PDE that just runs a standard 1D SSM over each axis independently For details, please refer to [12]. S4ND can be regarded as a convolution kernel with infinite receptive fields in N dimensions. In the frequency domain signal, the enhanced performance of S4ND-U-Net [15] is significant. Furthermore, both the parameter count and computational load are remarkably low. Importantly, S4ND adheres to the principles of causal networks and ensures real-time performance.

3 METHODOLOGY

Refer to caption
Fig. 1: Overview of the proposed SICRN system

3.1 SICRN

SICRN is shown in figure 1(a). The comprehensive network architecture adheres to the overall-detailed framework, employing the complex spectrum as its input. Initially, the feature channels are modified through inplace convolution. Subsequently, the SIC block undertakes preliminary extraction and integration of both global and local features from the real and imaginary components. Subsequently, the features undergo temporal modeling via a 2-layer LSTM. In the final step, the SIC block is employed for individual extraction and reconstruction of features from the real and imaginary components. Additionally, it estimates a complex mask, which is then applied through multiplication with the original real and imaginary parts to yield the enhanced feature spectrum.

3.2 SIC block

The SIC block can function as a convolution kernel, effectively replacing standard convolution kernels, and excels at extracting both local and global features. Importantly, the absence of downsampling operations within the entire module ensures the preservation of the original features.

The SIC block is shown in figure 1(b). The input channel is bifurcated into two segments. The initial 1/2 of the channel employs the inplace convolution kernel to capture local information. Following 1D convolution, they serve as the original features and local attention features, respectively. The latter 1/2 of the channel feeds into the S4ND layer for global information extraction, which is then harmonized with local attention features. The attention map is derived via the sigmoid activation function and subsequently multiplied with the original feature, thereby facilitating the extraction and fusion of local and global features. The specific calculation process is as follows:

X0c2L=C1d(IC(X0c2)\begin{split}X_{0\sim\frac{c}{2}}^{L}=C_{1d}(\operatorname{IC}(X_{0\sim\frac{c}{2}})\end{split} (9)
Xc2cR=S(Xc2c)\begin{split}X_{\frac{c}{2}\sim c}^{R}=\operatorname{S}(X_{\frac{c}{2}\sim c})\end{split} (10)
ATmap=σ(C1d(IC(X0c2))+Xc2cR)\begin{split}ATmap=\sigma(C_{1d}(\operatorname{IC}(X_{0\sim\frac{c}{2}}))+X_{\frac{c}{2}\sim c}^{R})\end{split} (11)
X=X0c2LATmap\begin{split}\mathrm{X}=X_{0\sim\frac{c}{2}}^{L}\cdot ATmap\end{split} (12)

Where, X0c2LX_{0\sim\frac{c}{2}}^{L} denotes the local features, while Xc2cRX_{\frac{c}{2}\sim c}^{R} represents the global features. X0c2X_{0\sim\frac{c}{2}} corresponds to the features of the first half of the channels, and Xc2cX_{\frac{c}{2}\sim c} pertains to the features of the latter half of the channels. IC()\operatorname{IC}(\cdot) denotes the inplace convolution operation, and S()\operatorname{S}(\cdot) signifies the convolution kernel used in the S4ND layer. The symbol σ\sigma represents the sigmoid activation function. ’ATmapATmap’ denotes the attention map and C1d()C_{1d}(\cdot) denotes the conv1d.

3.3 S4ND block

Figure 1(c) shows the S4ND block. Initially, the S4ND is employed to extract global features. This extraction process is followed by passage through an ELU activation function, and output via a linear layer. Subsequently, a residual connection is implemented to address potential problem related to gradient vanishing or exploding. Finally, a batch normalization layer is applied to generate the final output. The rationale behind choosing S4ND for global feature modeling is as follows:

1. The global modeling capacity of S4ND surpasses that of LSTM, while maintaining a smaller parameter count and computational load.

2. Given that S4 processes elements independently within the frequency dimension, it isn’t ideally suited for processing frequency domain information. This limitation is addressed by the utilization of S4ND.

Furthermore, as demonstrated in [12], experimental comparisons between S4ND and 2D convolutions reveal that S4ND outperforms the latter. Consequently, we opt for S4ND as the method to extract global feature information.

3.4 Loss function

We apply a scale-invariant signal-to-noise ratio (SI-SNR) [16] loss, which is a time domain loss function as follows:

𝐬target =𝐬^,𝐬𝐬𝐬2\mathbf{s}_{\text{target }}=\frac{\langle\hat{\mathbf{s}},\mathbf{s}\rangle\mathbf{s}}{\|\mathbf{s}\|^{2}} (13)
𝐞noise =𝐬^𝐬target \mathbf{e}_{\text{noise }}=\hat{\mathbf{s}}-\mathbf{s}_{\text{target }} (14)
si-snr =10log10𝐬target 2𝐞noise 2\mathcal{L}_{\text{si-snr }}=10\log_{10}\frac{\left\|\mathbf{s}_{\text{target }}\right\|^{2}}{\left\|\mathbf{e}_{\text{noise }}\right\|^{2}} (15)

where 𝐬^1×T\hat{\mathbf{s}}\in\mathbb{R}^{1\times T} and 𝐬1×T\mathbf{s}\in\mathbb{R}^{1\times T} refer to the estimated and clean sources, respectively, and 𝐬2=𝐬,𝐬\|\mathbf{s}\|^{2}=\langle\mathbf{s},\mathbf{s}\rangle denotes the signal power.

4 EXPERIMENTAL Setup

4.1 Datasets

We evaluated the SICRN on the DNS Challenge (INTERSPEECH 2020) dataset [17]. The clean speech set includes over 500 hours of clips from 2150 speakers. The noise dataset includes over 180 hours of clips from 150 classes. To make full use of the dataset, we simulate the speech-noise mixture with dynamic mixing during model training. In detail, before the start of each training epoch, 75% of the clean speeches are mixed with randomly selected room impulse responses (RIR) from (1) the Multichannel Impulse Response Database [18] with three reverberation times (T60) 0.16s, 0.36s, and 0.61 s. (2) the Reverb Challenge dataset [19] with three reverberation times 0.3 s, 0.6 s and 0.7 s. After that, the speech-noise mixtures are dynamically generated by mixing the clean speech (75% of them are reverberant) and noise with a random SNR in between -5 and 20 dB. The DNS Challenge provides a publicly available test dataset, including two categories of synthetic clips, i.e., without and with reverberations. Each category has 150 noisy clips with SNR levels distributed in between 0 dB to 20 dB. We use this test dataset for evaluation.

4.2 Configuration

Table 1: The performance in terms of WB-PESQ [MOS], NB-PESQ [MOS], STOI [%], and SI-SDR [dB] on the DNS challenge test dataset.
Method #Para (M) Look Ahead (ms) With Reverb Without Reverb
WB-PESQ NB-PESQ STOI SI-SDR WB-PESQ NB-PESQ STOI SI-SDR
Noisy - - 1.822 2.753 86.62 9.033 1.582 2.454 91.52 9.071
NSNet [20] 5.1 0 2.365 3.076 90.43 14.721 2.145 2.873 94.47 15.613
DTLN [21] 1.0 - - 2.700 84.68 10.530 - 3.040 94.76 16.340
Conv-TasNet [22] 5.08 33 2.750 - - - 2.730 - - -
DCCRN-E [23] 3.7 37.5 - 3.077 - - - 3.266 - -
PoCoNet [24] 50 - 2.832 - - - 2.748 - - -
FullSubNet [11] 5.64 32 2.969 3.473 92.62 15.750 2.777 3.305 96.11 17.290
SICRN 2.16 0 2.891 3.433 92.59 15.137 2.624 3.233 95.83 15.998

For STFT ,we adopt 510/160 for win-length/hop-length,the analysis is Hanning window. We use 510-point discrete Fourier transform (DFT) to extract 256-dimensional complex spectra for 16 kHz sampling rate. The model is optimized by Adam. The initial learning rate is 0.0002 and halved when the validation loss of four consecutive epochs no longer decreased. When training the model, the specific settings are as follows: there are 2 layers of SIC blocks, with channel sizes of 16 and 32, respectively, for each layer. Additionally, there are 2 LSTM layers in the middle. Within the SIC block, there are 3 inplace convolution layers, and 4 layers of S4ND blocks. For the integrity of the experiment, the 4-layer S4ND blocks in SIC block are replaced with 4-layer inplace convolution, which is named IICRN.

In order to verify the effectiveness of the proposed method, we select FullSubNet as the baseline models. At the same time, compare according to the same evaluation standard of FullSubNet. In addition, we also compared with the topranked methods in the DNS challenge (INTERSPEECH 2020), including NSNet [20], DTLN [21], Conv-TasNet [22], DCCRN [23] and PoCoNet [24].

5 EXPERIMENT RESULTS and Analysis

In this section, We compare the proposed SICRN network with other baseline systems across various metrics, including SI-SDR, STOI, NB-PESQ, WB-PESQ, “#Para” and “Look Ahead” , and show the results on the DNS challege test dataset. “#Para” and “Look Ahead” in the table respectively represent the parameter amount of the model and the length of used future information. “With Reverb” means that the noisy speeches in the test dataset have not only noise but also a certain degree of reverberation. “Without Reverb” means that the noisy speeches in the test dataset have only noise.

Table 1 provides a comprehensive evaluation of several methods. In the “With Reverb” column, SICRN outperforms the majority of models, ranking second only to FullSubNet, with only a negligible 0.03% difference in the STOI score. In the “Without Reverb” column, SICRN’s evaluation score also stands impressively high, with only a slight decrease compared to FullSubNet. Furthermore, as demonstrated in “#Para” and “Look Ahead”, SICRN achieves this high level of performance with a relatively small number of parameters and without relying on future frames for enhancement.

Table 2: Comprehensive comparison with FullSubNet.
Method #Para(M) MACs(G/s) Look Ahead(ms)
FullSubNet 5.64 30.84 32
SICRN 2.16 4.24 0

Table 2 provides a detailed comparison of parameter count, computational complexity, and Look Ahead between SICRN and FullSubNet, highlighting the best scores of each case in bold. It’s evident that SICRN utilizes only 2.16 M parameters and 4.24 G/s MACs. In stark contrast, FullSubNet employs 5.64 M parameters and exhibits a computational complexity of 30.84 G/s. Therefore, SICRN stands out for its significantly lower parameter count and computational complexity while delivering remarkable performance. Most notably, SICRN leverages solely the current frame and the past frame for enhancing the current frame, a strategy that presents distinct advantages and untapped potential when compared to FullSubNet.

Table 3: Ablation experiment(With Reverb).
Method WB-PESQ NB-PESQ STOI SI-SDR
mixture 1.822 2.753 86.62 9.033
IICRN 2.797 3.378 91.71 14.929
SICRN 2.891 3.433 92.59 15.137
Table 4: Ablation experiment(Without Reverb).
Method WB-PESQ NB-PESQ STOI SI-SDR
Noisy 1.582 2.454 91.52 9.071
IICRN 2.596 3.218 95.56 15.795
SICRN 2.624 3.233 95.83 15.998

To elucidate the role of S4ND, we conducted an ablation experiment to substantiate the benefits arising from the synergy between S4ND and inplace convolution. Displayed in Table 3 and Table 4, they represent the test outcomes for the datasets with reverberation and without reverberation, respectively. The experiments demonstrate a substantial performance enhancement upon the integration of S4ND, affirming the efficacy of combining S4ND and inplace convolution for feature extraction. Notably, the enhancement effect is particularly pronounced in the test set with reverberation. While the performance of IICRN is slightly inferior to SICRN, it still demonstrates a commendable level of effectiveness. This observation further underscores the advantages of employing inplace convolution to avoid downsampling speech features.

6 CONCLUSIONS

we propose an innovative model combing a state space model and inplace convolution, called SICRN. This network avoids downsampling operations throughout its architecture and combines multidimensional state space model and inplace convolution techniques to extract and integrate both global and local features. The experimental results demonstrate superior performance achieved with fewer parameters and computational resources, all without the need for future frame information.

Acknowledgments: This research was partly supported by the China National Nature Science Foundation (No. 61876214).

References

  • [1] Ashutosh Pandey and DeLiang Wang, “Tcnn: Temporal convolutional neural network for real-time speech enhancement in the time domain,” in ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2019, pp. 6875–6879.
  • [2] Ashutosh Pandey and DeLiang Wang, “Densely connected neural network with dilated convolutions for real-time speech enhancement in the time domain,” in ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020, pp. 6629–6633.
  • [3] Yanxin Hu, Yun Liu, Shubo Lv, Mengtao Xing, Shimin Zhang, Yihui Fu, Jian Wu, Bihong Zhang, and Lei Xie, “Dccrn: Deep complex convolution recurrent network for phase-aware speech enhancement,” arXiv preprint arXiv:2008.00264, 2020.
  • [4] Ivan Shchekotov, Pavel Andreev, Oleg Ivanov, Aibek Alanov, and Dmitry Vetrov, “Ffc-se: Fast fourier convolution for speech enhancement,” 2022.
  • [5] DeLiang Wang, “Time-frequency masking for speech separation and its potential for hearing aid design,” Trends in amplification, vol. 12, no. 4, pp. 332–353, 2008.
  • [6] Jun Chen, Zilin Wang, Deyi Tuo, Zhiyong Wu, Shiyin Kang, and Helen Meng, “Fullsubnet+: Channel attention fullsubnet with complex spectrograms for speech enhancement,” in ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 7857–7861.
  • [7] Ke Tan and DeLiang Wang, “A convolutional recurrent neural network for real-time speech enhancement.,” in Interspeech, 2018, vol. 2018, pp. 3229–3233.
  • [8] Jinjiang Liu and Xueliang Zhang, “Inplace gated convolutional recurrent neural network for dual-channel speech enhancement,” arXiv preprint arXiv:2107.11968, 2021.
  • [9] Jinjiang Liu and Xueliang Zhang, “Iccrn: Inplace cepstral convolutional recurrent neural network for monaural speech enhancement,” in ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2023, pp. 1–5.
  • [10] Chenggang Zhang, Jinjiang Liu, and Xueliang Zhang, “A complex spectral mapping with inplace convolution recurrent neural networks for acoustic echo cancellation,” in ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2022, pp. 751–755.
  • [11] Xiang Hao, Xiangdong Su, Radu Horaud, and Xiaofei Li, “Fullsubnet: A full-band and sub-band fusion model for real-time single-channel speech enhancement,” in ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021, pp. 6633–6637.
  • [12] Eric Nguyen, Karan Goel, Albert Gu, Gordon Downs, Preey Shah, Tri Dao, Stephen Baccus, and Christopher Ré, “S4nd: Modeling images and videos as multidimensional signals with state spaces,” Advances in neural information processing systems, vol. 35, pp. 2846–2861, 2022.
  • [13] Albert Gu, Karan Goel, and Christopher Ré, “Efficiently modeling long sequences with structured state spaces,” arXiv preprint arXiv:2111.00396, 2021.
  • [14] Karan Goel, Albert Gu, Chris Donahue, and Christopher Ré, “It’s raw! audio generation with state-space models,” in International Conference on Machine Learning. PMLR, 2022, pp. 7616–7633.
  • [15] Pin-Jui Ku, Chao-Han Huck Yang, Sabato Marco Siniscalchi, and Chin-Hui Lee, “A multi-dimensional deep structured state space approach to speech enhancement using small-footprint models,” arXiv preprint arXiv:2306.00331, 2023.
  • [16] Yi Luo and Nima Mesgarani, “Conv-tasnet: Surpassing ideal time–frequency magnitude masking for speech separation,” IEEE/ACM transactions on audio, speech, and language processing, vol. 27, no. 8, pp. 1256–1266, 2019.
  • [17] Chandan KA Reddy, Vishak Gopal, Ross Cutler, Ebrahim Beyrami, Roger Cheng, Harishchandra Dubey, Sergiy Matusevych, Robert Aichner, Ashkan Aazami, Sebastian Braun, et al., “The interspeech 2020 deep noise suppression challenge: Datasets, subjective testing framework, and challenge results,” arXiv preprint arXiv:2005.13981, 2020.
  • [18] Elior Hadad, Florian Heese, Peter Vary, and Sharon Gannot, “Multichannel audio database in various acoustic environments,” in 2014 14th International Workshop on Acoustic Signal Enhancement (IWAENC). IEEE, 2014, pp. 313–317.
  • [19] Keisuke Kinoshita, Marc Delcroix, Sharon Gannot, Emanuël A P. Habets, Reinhold Haeb-Umbach, Walter Kellermann, Volker Leutnant, Roland Maas, Tomohiro Nakatani, Bhiksha Raj, et al., “A summary of the reverb challenge: state-of-the-art and remaining challenges in reverberant speech processing research,” EURASIP Journal on Advances in Signal Processing, vol. 2016, pp. 1–19, 2016.
  • [20] Yangyang Xia, Sebastian Braun, Chandan KA Reddy, Harishchandra Dubey, Ross Cutler, and Ivan Tashev, “Weighted speech distortion losses for neural-network-based real-time speech enhancement,” in ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020, pp. 871–875.
  • [21] Nils L Westhausen and Bernd T Meyer, “Dual-signal transformation lstm network for real-time noise suppression,” arXiv preprint arXiv:2005.07551, 2020.
  • [22] Yuichiro Koyama, Tyler Vuong, Stefan Uhlich, and Bhiksha Raj, “Exploring the best loss function for dnn-based low-latency speech enhancement with temporal convolutional networks,” arXiv preprint arXiv:2005.11611, 2020.
  • [23] Yanxin Hu, Yun Liu, Shubo Lv, Mengtao Xing, Shimin Zhang, Yihui Fu, Jian Wu, Bihong Zhang, and Lei Xie, “Dccrn: Deep complex convolution recurrent network for phase-aware speech enhancement,” arXiv preprint arXiv:2008.00264, 2020.
  • [24] Umut Isik, Ritwik Giri, Neerad Phansalkar, Jean-Marc Valin, Karim Helwani, and Arvindh Krishnaswamy, “Poconet: Better speech enhancement with frequency-positional embeddings, semi-supervised conversational data, and biased loss,” arXiv preprint arXiv:2008.04470, 2020.