8
Non-stationary Noise Cancellation Using Deep Autoencoder Based on Adversarial Learning Kyung-Hyun Lim, Jin-Young Kim, and Sung-Bae Cho (&) Department of Computer Science, Yonsei University, Seoul, South Korea {lkh1075,seago0828,sbcho}@yonsei.ac.kr Abstract. Studies have been conducted to get a clean data from non-stationary noisy signal, which is one of the areas in speech enhancement. Since conven- tional methods rely on rst-order statistics, the effort to eliminate noise using deep learning method is intensive. In the real environment, many types of noises are mixed with the target sound, resulting in dif culty to remove only noises. However, most of previous works modeled a small amount of non-stationary noise, which is hard to be applied in real world. To cope with this problem, we propose a novel deep learning model to enhance the auditory signal with adversarial learning of two types of discriminators. One discriminator learns to distinguish a clean signal from the enhanced one by the generator, and the other is trained to recognize the difference between eliminated noise signal and real noise signal. In other words, the second discriminator learns the waveform of noise. Besides, a novel learning method is proposed to stabilize the unstable adversarial learning process. Compared with the previous works, to verify the performance of the propose model, we use 100 kinds of noise. The experimental results show that the proposed model has better performance than other con- ventional methods including the state-of-the-art model in removing non- stationary noise. To evaluate the performance of our model, the scale-invariant source-to-noise ratio is used as an objective evaluation metric. The proposed model shows a statistically signicant performance of 5.91 compared with other methods in t-test. Keywords: Speech enhancement Noise cancellation Non-stationary noise Generative adversarial networks Autoencoder 1 Introduction Speech enhancement can be divided into several specic problems such as noise elimination and quality improvement. Since noise has a very diverse form, it is too dif cult to remove it without prior knowledge. In particular, the problem becomes more dif cult for non-stationary noises. As shown in Fig. 1, it is dif cult to recognize the non-stationary noise in the given signal [1]. That is why modeling it is an essential issue for the noise cancellation. Several works to eliminate stationary noise have been conducted for a long time. Traditional methods have many limitations because they use ltering based on threshold [1]. If signal data is processed in the frequency domain through Fourier transform (FT) as shown in Fig. 2, the general noise does not overlap much with the © Springer Nature Switzerland AG 2019 H. Yin et al. (Eds.): IDEAL 2019, LNCS 11871, pp. 367374, 2019. https://doi.org/10.1007/978-3-030-33607-3_40

Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

  • Upload
    others

  • View
    14

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

Non-stationary Noise Cancellation Using DeepAutoencoder Based on Adversarial Learning

Kyung-Hyun Lim, Jin-Young Kim, and Sung-Bae Cho(&)

Department of Computer Science, Yonsei University, Seoul, South Korea{lkh1075,seago0828,sbcho}@yonsei.ac.kr

Abstract. Studies have been conducted to get a clean data from non-stationarynoisy signal, which is one of the areas in speech enhancement. Since conven-tional methods rely on first-order statistics, the effort to eliminate noise usingdeep learning method is intensive. In the real environment, many types of noisesare mixed with the target sound, resulting in difficulty to remove only noises.However, most of previous works modeled a small amount of non-stationarynoise, which is hard to be applied in real world. To cope with this problem, wepropose a novel deep learning model to enhance the auditory signal withadversarial learning of two types of discriminators. One discriminator learns todistinguish a clean signal from the enhanced one by the generator, and the otheris trained to recognize the difference between eliminated noise signal and realnoise signal. In other words, the second discriminator learns the waveform ofnoise. Besides, a novel learning method is proposed to stabilize the unstableadversarial learning process. Compared with the previous works, to verify theperformance of the propose model, we use 100 kinds of noise. The experimentalresults show that the proposed model has better performance than other con-ventional methods including the state-of-the-art model in removing non-stationary noise. To evaluate the performance of our model, the scale-invariantsource-to-noise ratio is used as an objective evaluation metric. The proposedmodel shows a statistically significant performance of 5.91 compared with othermethods in t-test.

Keywords: Speech enhancement � Noise cancellation � Non-stationary noise �Generative adversarial networks � Autoencoder

1 Introduction

Speech enhancement can be divided into several specific problems such as noiseelimination and quality improvement. Since noise has a very diverse form, it is toodifficult to remove it without prior knowledge. In particular, the problem becomes moredifficult for non-stationary noises. As shown in Fig. 1, it is difficult to recognize thenon-stationary noise in the given signal [1]. That is why modeling it is an essentialissue for the noise cancellation.

Several works to eliminate stationary noise have been conducted for a long time.Traditional methods have many limitations because they use filtering based onthreshold [1]. If signal data is processed in the frequency domain through Fouriertransform (FT) as shown in Fig. 2, the general noise does not overlap much with the

© Springer Nature Switzerland AG 2019H. Yin et al. (Eds.): IDEAL 2019, LNCS 11871, pp. 367–374, 2019.https://doi.org/10.1007/978-3-030-33607-3_40

Page 2: Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

clean signal frequency domain, so filtering can work to some extent. However, it can beconfirmed that it is difficult to remove the non-stationary noise by filtering because thefrequency domain of the voice signal and the noise overlaps or the area is not constant.

Many researchers have studied to eliminate noise by parameterizing it using deeplearning [1–4]. The models based on recurrent neural network (RNN) are used to modelsignal data which is time series data [5, 6]. Various conversions of sound data areapplied into model as input, resulting in high cost. Other models include the generativeadversarial network (GAN) method [6]. Pascual et al. proposed a model which hadexcellent performance in noise removal, but it has a limitation that it is vulnerable tonon-stationary noise [6].

To overcome the limitations of the previous works, we propose a model that learnsthe form of noise itself. The proposed model is trained using two discriminators andtakes the adversarial learning structure. In addition, a novel method is proposed tostabilize a process with adversarial learning. The proposed model reduces the pre-processing cost using raw data as input and generalizes various noises throughadversarial learning with two discriminators. Besides, non-stationary noise can beremoved by additional noise-aware discriminator.

Fig. 1. Example of (top) general noise and (bottom) non-stationary noise cancellation.

Fig. 2. Comparison of general noise with non-stationary noise in frequency domain.

368 K.-H. Lim et al.

Page 3: Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

2 Related Works

There are two main approaches to noise cancellation: filtering and deep learning. Thefiltering methods previously applied in the noise cancellation are the Wiener filteringand spectral subtraction [7, 8]. Both methods basically eliminate noise by removing aspecific signal with high or low frequency based on threshold. However, no matter howwell we set the threshold, there can be several types of noise in the real world. It isimpossible to set a threshold that can eliminate all these noises. To address the limi-tation that filtering alone makes it difficult to construct a function to map the complexrelationship between noisy and clean signal, a deep learning method is introduced fornoise cancellation [1].

Tamura and Waibel used the basic multilayer perceptron [1]. They argued that theneural network can find the relationship between two signals. However, the data usedin the experiment were strongly conditioned that only one speaker was used, and it wasmade up of not many words. Since then, various attempts have been made to create amodel that can eliminate independent and diverse noises. Especially, since signal alsohas the form of time series data, RNN is used as a basic model [9–11]. Parveen andGreen succeeded in showing that the noise could be removed independently from thetopic with speech data including 55 speakers, but it had the limitation that there wereonly eleven words included in the signal and experimented with less amount of noise[11]. Another study attempted to remove noise using auto-encoder [12]. This modelalso showed good performance for general noise but did consider small amount ofnoise. Since using raw data reduces the information loss and preprocessing costs,Pascual et al. proposed an end-to-end model that first used adversarial learning in thefield of noise cancellation [6]. However, it had also limitations that non-stationary noisestill became a vulnerability.

In the real environment, since the general noise can also be non-stationary due tovarious environmental changes, solving this problem is an important issue [13]. Theproposed model includes a powerful noise-aware module to eliminate non-stationarynoise, and its performance will be evaluated in Sect. 4.

3 The Proposed Method

3.1 Overview

The process of the proposed method is divided into three parts: making powerfuldiscriminators, training proposed generator, and removing the noise with shadowgenerator. While training the model with adversarial process, there is unstable learningprocedure [14, 15]. To prevent this problem, we use two generators having the samestructure. Only the first generator is trained in adversarial. In the first phase, originalgenerator and discriminators are learned through adversarial learning, and only theshadow generator is learned using discriminators learned in the second phase. It isintended to create a generator that can learn steadily with a strong teacher. In the thirdphase, noise is removed using the learned shadow generator.

Non-stationary Noise Cancellation Using Deep Autoencoder 369

Page 4: Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

3.2 Pretraining Discriminators

Figure 3 represents the adversarial learning structure of the original generator and thetwo discriminators in the first phase. These discriminators are constructed with thesame structure. It is composed of the fully-convolution layer to extract the feature fromdata and learn. One is trained to classify clean and noise as true and classify origin-generated and noisy signal as false. The task of it is to distinguish between the cleansignal and the generated signal. The other uses true-pair with the difference of the noisysignal and clean signal and uses false-pair with the difference between the noisy signaland generated signal, which means that it learns the difference between the two signalsin the form of noise, residual between two signals. It helps the generator learn toproduce a clean signal more closely. The generator is structured like auto-encoder. Theencoder and decoder are composed of fully-convolution layer and the U-NET structureis added to prevent the information loss between the encoder and the decoder [16].

After the discriminators are learned, the generator is trained by using discrimina-tors. Since Dx learns true-pair as true and false-pair as false, the objective function canbe expressed as Eq. (1). Likewise, Dx0 has similar learning objectives, so that the lossfunction can be defined as in Eq. (2).

maxGh

minDx

½ Dx T x;~xð Þð Þ � 1f g2 þ Dx T Gh xð Þ;~xð Þð Þf g2� þ Gh Z;~xð Þ � xk k1 ð1Þ

maxGh

minDx0

½ Dx0 ~x� xð Þ � 1f g2 þ Dx0 ~x� Gh ~xð Þð Þf g2� þ fGh Z;~xð Þ � xg � ~x� xð Þk k1 ð2Þ

where x means clean signal and is the target that the model should make. ~x means noisysignal, and Gh means one of the generators. T means concatenation. Noisy signal isused as an input of an encoder, and the output of the encoder and Z, which are priordistributions, are used as inputs of the decoder. The generator adapts its parameterssuch that discriminators distinguish the outputs of the generator as ‘True’. Besides,additional L1 loss causes the generator to produce a signal closer to the target [17].Equation (1) is for minimizing the difference between the signal generated and the

Fig. 3. The overall structure of the proposed method. Real noise data mean noise itself.

370 K.-H. Lim et al.

Page 5: Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

clean signal. Equation (2) can produce a signal closer to reality by adding the differencebetween the noise signal and the generated signal difference, which minimizes the noisedata [15].

3.3 Training Shadow Generator

Adversarial learning method is difficult to learn in stable [14, 15]. Generators that are incharge of complex tasks compared to discriminators often diverge during learning [13].To prevent this, we propose a shadow generator Gh0 . It is trained using pretraineddiscriminators that are powerful instructors to produce clean signal. Gh0 just learns withdiscriminators what is the target distribution, but it is not trained in adversarial. Asshown in Fig. 4, Gh that does adversarial learning is not able to converge easily and itsloss oscillates, while Gh0 continues to learn in stable. Loss functions are defined asEqs. (3) and (4) in a form like Gh.

minGh0

Dx Gh0 Z;~xð Þð Þ � 1f g2 þ Gh0 Z;Nð Þ � Ck k1 ð3Þ

minGh0

Dx0 Gh0 Z;~xð Þ � xð Þ � 1f g2 þ fGh0 Z;~xð Þ � xg � ~x� xð Þk k1 ð4Þ

4 Experiments

4.1 Dataset and Evaluation Metric

To verify the proposed method, we use noise sounds provided by Hu and Wang [18].Noise data has 20 categories and has several noises in each category. We use datarecorded by 30 native English speakers producing 400 sentences each as clean signal[19]. The dataset for learning is constructed by combining noise data and clean signaldata as shown in Fig. 5. Non-stationary noises are selected from the noise data anddivided into training data and test data at 8:2 ratio. We use SI-SNR as the evaluationmetric to compare the similarity of signal [20]. The SI-SNR value is calculated usingEq. (5).

Ctarget ¼ x;~xh ixxk k2 ;

Enoise ¼ ~x� Ctarget;

SI � SNR ¼ 10 log10Ctarget

��

��2

Enoisek k2

ð5Þ

Non-stationary Noise Cancellation Using Deep Autoencoder 371

Page 6: Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

4.2 Results of Noise Cancellation

The left figures in Figs. 6 and 7 represent the waveforms of the signals in time andfrequency domain. Comparing the waveforms of clean signal and noisy signal, it can beconfirmed that non-stationary noise is applied to clean signal. The waveform of cleansignal and generated signal can be seen to be almost similar. Four comparison groupsare used to compare the performance of the proposed model. In the first group,autoencoder is used for learning only the generator without adversarial learning. Weconduct the experiment using the traditional low pass filtering method and the model[6] with the second group. The right in Fig. 6 shows the SI-SNR values of each modelfor the test data in a box plot. Table 1 shows the result of t-test with SI-SNR values.

Fig. 4. The change of two generators loss according to epoch.

Fig. 5. Process of making noisy signal using noise data and clean signal data.

Fig. 6. (Left) Waveform of input and output signals. (Right) Box plot using SI-SNR values ofeach model.

372 K.-H. Lim et al.

Page 7: Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

The t-test shows that the experiment result is statistically significant. The experimentalresults demonstrate that our model outperforms other conventional methods incanceling the non-stationary noise.

5 Conclusions

In this paper, we propose a model that can eliminate more non-stationary noise andunseen noise than other studies. We create a powerful instructor through adversariallearning of the discriminators and the generator. In order to solve the problem that thegenerator easily diverges during the adversarial learning process, Gh0 with the samestructure is provided to facilitate stable learning. Gh0 learns from powerful instructorsand generates enhanced signal from a given noisy signal. Several experiments showthat the proposed method shows promising results of the best performance with 5.91 interms of SI-SNR. This means that our model can remove not only non-stationary noisebut also various types of noises.

We will add a comparative model to demonstrate the objective performance of themodel. Also, several metrics used in the noise reduction domain will be added toevaluate the model. Finally, we will build and release a web demo system to test ourmodel.

Acknowledgment. This work was supported by grant funded by 2019 IT promotion fund(Development of AI based Precision Medicine Emergency System) of the Korean government(Ministry of Science and ICT).

Fig. 7. Spectrogram of signals.

Table 1. Result of t-test on SI-SNR value.

Ours Autoencoder Noisy signal Pascual et al. [6] Filtering

Mean 5.971 4.176 1.121 −9.428 −16.982Variance 11.356 13.174 35.173 11.356 65.659p-value – 2.29 � 10−137 0. 0. 0.

Non-stationary Noise Cancellation Using Deep Autoencoder 373

Page 8: Non-stationary Noise Cancellation Using Deep Autoencoder ...sclab.yonsei.ac.kr/publications/Papers/IC/2019_IDEAL_KHL.pdfNon-stationary Noise Cancellation Using Deep Autoencoder Based

References

1. Tamura, S., Waibel, A.: Noise reduction using connectionist models. In: IEEE InternationalConference on Acoustics, Speech and Signal Processing, pp. 553–556 (1988)

2. Xu, Y., Du, J., Dai, L.R., Lee, C.H.: A regression approach to speech enhancement based ondeep neural networks. In: IEEE/ACM Trans. on Audio, Speech, and Language Processing,vol. 23, no. 1, pp. 7–19 (2014)

3. Kumar, A., Florencio, D.: Speech Enhancement in Multiple-noise Conditions Using DeepNeural Networks. arXiv:1605.02427 (2016)

4. Weninger, F., et al.: Speech enhancement with LSTM recurrent neural networks and itsapplication to noise-robust ASR. In: International Conference on Latent Variable Analysisand Signal Separation, pp. 91–99 (2015)

5. Kim, J.Y., Bu, S.J., Cho, S.B.: Hybrid deep learning based on GAN for classifying BSRnoises from invehicle sensors. In: International Conference on Hybrid Artificial IntelligentSystems, pp. 561–572 (2018)

6. Pascual, S., Bonafonte, A., Serra, J.: SEGAN: Speech Enhancement Generative AdversarialNetwork. arXiv:1703.09452 (2017)

7. Lim, J., Oppenheim, A.: All-pole modeling of degraded speech. In: IEEE Transactions onAcoustics Speech and Signal Processing, vol. 26, no. 3, pp. 197–210 (1978)

8. Berouti, M., Schwartz, R., Makhoul, J.: Enhancement of speech corrupted by acoustic noise.In: International Conference on Acoustics Speech and Signal Processing, vol. 4, pp. 208–211(1979)

9. Maas, A.L., Le, Q.V., O’Neil, T.M., Vinyals, O., Nguyen, P., Nguyen, A.Y.: Recurrentneural networks for noise reduction in robust ASR. In: InterSpeech, pp. 22–25 (2012)

10. Sun, L., Du, J., Dai, L.R., Lee, C.H.: Multiple-target deep learning for LSTM-RNN basedspeech enhancement. In: IEEE Hands-free Speech Communications and Microphone Arrays,pp. 136–140 (2017)

11. Parveen, S., Green, P.: Speech enhancement with missing data techniques using recurrentneural networks. In: IEEE International Conference on Acoustics Speech and SignalProcessing, pp. 733–736 (2004)

12. Lu, X., Tsao, Y., Matsuda, S., Hori, C.: Speech enhancement based on deep denoisingautoencoder. In: InterSpeech, pp. 436–440 (2013)

13. Cohen, I.: Multichannel post-filtering in nonstationary noise environments. In: IEEETransactions on Signal Processing, vol. 52, no. 5, pp. 1149–1160 (2004)

14. Goodfellow, I., et al.: Generative adversarial nets. In: Neural Information ProcessingSystems, pp. 2672–2680 (2014)

15. Kim, J.Y., Bu, S.J., Cho, S.B.: Zero-day malware detection using transferred generativeadversarial networks based on deep autoencoders. In: Information Sciences, vol. 460–461,pp. 83–102 (2018)

16. Isola, P., Zhu, J.Y., Zhou, T., Efros, A.: Image-to-Image Translation with ConditionalAdversarial Networks. arXiv:1611.07004 (2016)

17. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: IEEEConference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)

18. Hu, G., Wang, D.L.: A tandem algorithm for pitch estimation and voiced speech segregation.In: IEEE Trans. on Audio Speech and Language Processing, vol. 18, pp. 2067–2079 (2010)

19. Valentini, C.: Noisy speech database for training speech enhancement algorithms and TTSmodels. In: University of Edinburgh, School of Informatics, Centre for Speech TechnologyResearch (2016)

20. Luo, Y., Mesgarani, N.: TasNet: Surpassing Ideal Time-frequency Masking for SpeechSeparation. arXiv:1809.07454 (2018)

374 K.-H. Lim et al.