5
Real-time Streaming Wave-U-Net with Temporal Convolutions for Multichannel Speech Enhancement Vasiliy Kuzmin 1 , Fyodor Kravchenko 1 , Artem Sokolov 1,3 , Jie Geng 2 1 Huawei Russia Research Institute 2 Huawei Nanjing Research Institute, China 3 HSE University, Nizhniy Novgorod, Russia [email protected], [email protected], [email protected], [email protected] Abstract In this paper we describe our work that we have done to par- ticipate in Task1 of ConferencingSpeech2021 challenge. This task set a goal to develop the solution for multi-channel speech enhancement in a real-time manner. We propose a novel system for streaming speech enhancement. We employ Wave-U-Net architecture with temporal convolutions in encoder and decoder. We incorporate self-attention in decoder to apply attention mask retrieved from skip-connection on features from down-blocks. We explore history cache mechanisms that work like hidden states in recurrent networks and implemented them in proposal solution. It helps us to run inference with chunks length 40 ms and Real-Time Factor 0.4 with the same precision. Index Terms: Multi-channel speech enhancement, speech en- hancement, wave-u-net, self-attention, deep learning 1. Introduction Speech enhancement (SE) tries to separate out individual audio sources from an input mixture of clean audio and noise. Multi- channel approaches, generally, do it better than single-channel ones considering that they could process spatial information taken from time differences between signal reach separate items of microphone arrays (MA). Conventional approaches like multi- channel Wiener filter [1] or beamforming [2] are progressively displaced by deep learning (DL) based techniques [3] or by mix- ture with them [4]. Deep learning approaches with U-Net [5] based architectures was very successfully on various computer vision tasks connected to image and video processing where we need to transform one image to another or extract topological information that we need. Furthermore, this architecture was transferred to speech related tasks close to SE as voice separation [6] and speech enhancement itself [7]. This architecture works well as with signal waveform as with its Short Time Fourier Transform (STFT) [8]. The version adopted for processing time domain signal called Wave-U-Net is also actively exploited nowadays for the same tasks [9]. To obtain real-time streaming speech enhance- ment the network should ignore the future frames and provide the Real Time Factor (RTF) < 1. Recurrent networks like unidi- rectional Long-Short Time Memory [10] (LSTM) and Temporal Convolution Networks (TCN) could satisfy the first restrictions as they are based on casual convolutions which have zero look ahead [11]. Moreover, convolutions are very fast and allow to build real-time systems. LSTM mechanisms provide hidden states retrieving and sequences could be handled step-by-step by passing previous states for new iteration. But for convolution this trick is not supported from-the-box. In [12] the authors showed that history cache with hidden states could be saved and reused. Thus, it is also possible to fulfill inference with small pieces of stream, but it requires to "manually" fed on hidden states from previous iteration with the input. This paper pro- poses a neural network architecture based on Wave-U-Net with temporal convolutions (TC Wave-U-Net) which uses cache with hidden states for the inference. To make prediction more precise various self-attention mech- anisms are incorporated in Wave-U-Net systems [13]. We also do it in the paper as casual convolution gives in to convolu- tion with future context and extra elements are required to yield acceptable precise. This method was submitted to the ConferencingSpeech 2021 challenge 1 in INTERSPEECH 2021 conference. Summarily, our primary contributions are the follows: We developed the novel model for the multi-channel speech enhancement task that can be inferences real-time manner, have zero look ahead and based on successful U-Net architecture. This model have only 8.31 millions parameters and can be run on any device. We applied historical context cache which allowed us to decrease receptive field during inference as well as decrease total number of float points operation. The rest of the paper is organised as follows. First, in the follow- ing section, we briefly review the background of the problem. In Section 3 we describe our Wave-U-Net based Temporal Convo- lution network with attentions. The details of of our experiments and training procedure are then presented in Section 4. We present the results and comparison with the baseline system in Section 5. Finally, in Section 6 we provide some conjectures as why and how to configure historical cache, further research in this direction followed by conclusion. 2. Background During participating ConferencingSpeech 2021 challenge we aimed to solve Task1 which is Multi-channel speech enhance- ment with single microphone array. This task expects the noisy audio processing from single linear array with non-uniform dis- tributed microphones. Real time factor considering less than 1.0 while running on device as well as participant should use frame of length 40 ms. A multi-channel speech enhancement problem could be de- scribed the following way: Let take C - channel signal on the t time step as Yt = [y 1 t , ..., y c t ] 1 https://tea-lab.qq.com/conferencingspeech-2021 arXiv:2104.01923v1 [eess.SP] 5 Apr 2021

Real-time Streaming Wave-U-Net with Temporal Convolutions

  • Upload
    others

  • View
    8

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Real-time Streaming Wave-U-Net with Temporal Convolutions

Real-time Streaming Wave-U-Net with Temporal Convolutions forMultichannel Speech Enhancement

Vasiliy Kuzmin1, Fyodor Kravchenko1, Artem Sokolov1,3, Jie Geng2

1Huawei Russia Research Institute2Huawei Nanjing Research Institute, China3HSE University, Nizhniy Novgorod, Russia

[email protected], [email protected], [email protected],[email protected]

AbstractIn this paper we describe our work that we have done to par-ticipate in Task1 of ConferencingSpeech2021 challenge. Thistask set a goal to develop the solution for multi-channel speechenhancement in a real-time manner. We propose a novel systemfor streaming speech enhancement. We employ Wave-U-Netarchitecture with temporal convolutions in encoder and decoder.We incorporate self-attention in decoder to apply attention maskretrieved from skip-connection on features from down-blocks.We explore history cache mechanisms that work like hiddenstates in recurrent networks and implemented them in proposalsolution. It helps us to run inference with chunks length 40 msand Real-Time Factor 0.4 with the same precision.Index Terms: Multi-channel speech enhancement, speech en-hancement, wave-u-net, self-attention, deep learning

1. IntroductionSpeech enhancement (SE) tries to separate out individual audiosources from an input mixture of clean audio and noise. Multi-channel approaches, generally, do it better than single-channelones considering that they could process spatial informationtaken from time differences between signal reach separate itemsof microphone arrays (MA). Conventional approaches like multi-channel Wiener filter [1] or beamforming [2] are progressivelydisplaced by deep learning (DL) based techniques [3] or by mix-ture with them [4]. Deep learning approaches with U-Net [5]based architectures was very successfully on various computervision tasks connected to image and video processing where weneed to transform one image to another or extract topologicalinformation that we need. Furthermore, this architecture wastransferred to speech related tasks close to SE as voice separation[6] and speech enhancement itself [7]. This architecture workswell as with signal waveform as with its Short Time FourierTransform (STFT) [8].

The version adopted for processing time domain signalcalled Wave-U-Net is also actively exploited nowadays for thesame tasks [9]. To obtain real-time streaming speech enhance-ment the network should ignore the future frames and providethe Real Time Factor (RTF) < 1. Recurrent networks like unidi-rectional Long-Short Time Memory [10] (LSTM) and TemporalConvolution Networks (TCN) could satisfy the first restrictionsas they are based on casual convolutions which have zero lookahead [11]. Moreover, convolutions are very fast and allowto build real-time systems. LSTM mechanisms provide hiddenstates retrieving and sequences could be handled step-by-step bypassing previous states for new iteration. But for convolutionthis trick is not supported from-the-box. In [12] the authors

showed that history cache with hidden states could be saved andreused. Thus, it is also possible to fulfill inference with smallpieces of stream, but it requires to "manually" fed on hiddenstates from previous iteration with the input. This paper pro-poses a neural network architecture based on Wave-U-Net withtemporal convolutions (TC Wave-U-Net) which uses cache withhidden states for the inference.

To make prediction more precise various self-attention mech-anisms are incorporated in Wave-U-Net systems [13]. We alsodo it in the paper as casual convolution gives in to convolu-tion with future context and extra elements are required to yieldacceptable precise.

This method was submitted to the ConferencingSpeech 2021challenge1 in INTERSPEECH 2021 conference. Summarily, ourprimary contributions are the follows:

• We developed the novel model for the multi-channelspeech enhancement task that can be inferences real-timemanner, have zero look ahead and based on successfulU-Net architecture. This model have only 8.31 millionsparameters and can be run on any device.

• We applied historical context cache which allowed usto decrease receptive field during inference as well asdecrease total number of float points operation.

The rest of the paper is organised as follows. First, in the follow-ing section, we briefly review the background of the problem. InSection 3 we describe our Wave-U-Net based Temporal Convo-lution network with attentions. The details of of our experimentsand training procedure are then presented in Section 4. Wepresent the results and comparison with the baseline system inSection 5. Finally, in Section 6 we provide some conjectures aswhy and how to configure historical cache, further research inthis direction followed by conclusion.

2. BackgroundDuring participating ConferencingSpeech 2021 challenge weaimed to solve Task1 which is Multi-channel speech enhance-ment with single microphone array. This task expects the noisyaudio processing from single linear array with non-uniform dis-tributed microphones. Real time factor considering less than 1.0while running on device as well as participant should use frameof length 40 ms.

A multi-channel speech enhancement problem could be de-scribed the following way:

Let take C - channel signal on the t time step as Yt =[y1t , ..., y

ct ]

1https://tea-lab.qq.com/conferencingspeech-2021

arX

iv:2

104.

0192

3v1

[ee

ss.S

P] 5

Apr

202

1

Page 2: Real-time Streaming Wave-U-Net with Temporal Convolutions

Figure 1: Schematic view of TC Wave-U-Net.

We assume that yct is given as the following mathematicalexpression:

yct = xct + nct = hc ~ xct + nc

t (1)

where xct the clean signal recorded by the C-th microphone, nct

is the additive noise signal and hc is room impulse response(RIR). The convolution of RIR and dry signal xct called rever-berant signal xct . The aim of Task1 is to estimate signal xorigt ,where orig ∈ {1, ..., C}, by removing additive noise and bydereverberation for whole utterance t = 1, ..., T . The xorigt

represents the original mono channel audio from multi-channelmixture Yt.

3. Architecture: Wave-U-Net with temporalconvolutions

In this section, we explain the TC Wave-U-Net architecturewith attentions and history cache we implemented for streaminginference.

3.1. Structure of TC Wave-U-Net

We follow the Wave-U-Net structure for our multi-channelspeech enhancement solution. Our framework consists of anencoder, middle part (bottleneck) and decoder. We use historypadding to avoid the non zero-look ahead. Schematically systemis depicted on Figure 1. We note that our model takes C-channelwaveform as its input and predicts an cleaned audio.

Encoder and decoder consist of TC blocks. The block in-ternals are visualised on Figure 2 (a). The block is a sequenceof dilated casual 1d convolution, batch normalization, paramet-ric non-linearity, dropout and one more convolution. A finalParametric Rectified Linear Unit [14] (PReLU) is applied af-ter additive operation of result with residual connection. Weexperimentally established that such architecture provides bet-ter convergence. We keep the SAME padding everywhere andafter each TC block fulfill down-operation - reducing the timedimension on half just by removing of each second element.Our experiments showed that this kind of down sampling ismore effective for our architecture than, for example, stride 2in the convolution. The decoder is organized symmetricallybut it composed a bit more complicated as comprises attentionmechanism (Figure 2(b)). We use linear interpolation for 2×up-sampling. Skip connection between encoder and decoderblocks is organized as usual 1D convolution.

Dilated Causal Conv

BatchNorm

PReLU

Dropout

Dilated Causal Conv

1x1 Conv

+

PReLU

Up-sampling

SA

Concat

TC block

skip

(a) (b)

Figure 2: (a) TC Block, (b) Up Block. The skip connection is theoutput from the corresponding down-block.

k(D)

q(U)

v(D)

+ PReLU A(P) PReLU Concat

key

query

value

Di

Ul-i

Di

Pi Wi

Figure 3: Proposed Self-Attention

The papers [13], [15], [16] inspired us to integrate a self-attention to our architecture. We fed on self-attention an up blockand skip connection tensors from the same hierarchical level.Afterwards, we concatenate Up Block output with self-attentionoutput and result goes to TC block.

Self-attention mechanism is intended to capture the globaldependencies. It is successfully exploit, for instance, in suchfields of deep learning as speech recognition [17] and machinetranslation [18]. In our architecture we employed attention bythe similar way with [13] to find out relevant features camewith skip connections by multiplying it with attention mask. Thesame approach is applied before the output of the network whenoriginal noisy signal and output from up-block are passed toattention. Visual explanation of proposed self-attention is shownFigure 3.

Let’s give l the number of TC blocks in encoder/decoder.When Di, i = 1, ...l be output of down-blocks came as skip-connection and Ul−i be the output of linear interpolation. At-tention uses k, q and v as 1D convolution operations for therepresentation of input to embedding space. The products keyand query are then summarised and, after parametric linearity,as product Pi follows to attention mask calculationA(Pi) whichis a convolution with kernel size 1. The output of attention is aterm-wise product ofWi = A(Pi)Di. In our experiments, atten-tion block helps the network to converge deeper, while withoutit, a model converges faster, however it gets stuck in some space.

3.2. Streaming Inference

For streaming inference we employed idea from [12]. Paper rep-resents how to extend the history context length which taken bya network into account if the network was being trained on fixed-size segments of data. The authors applied the mechanismsof hidden states reusing on the Transformer architecture andachieved considerable gain in the length of context remembered

Page 3: Real-time Streaming Wave-U-Net with Temporal Convolutions

(a) (b)

Figure 4: (a) Historical padding for training stage. (b) Historicalcontext cache for inference stage.

Drop

Conv1d

Concat

Input

Buffer_in Slice Buffer_out

Figure 5: Context cache processing on inference time.

by the model. Intuitively it emulates the behavior of recurrentnetworks that save information between time steps. We testedthis idea for TC Wave-U-Net in such a way that we do not try toextend, we try to keep the context size for inference as networkused for training but for smaller chunk of data. It provokes theinvolved in processing the only part of the receptive field. More-over, it leads to speed up of inference as calculations with floatsshould be dropped. In the order to keep the accuracy the sameas for inference with whole shape we should "manually" reusehidden states of each layer. As mentioned before, the proposedmethod try to align conditions between training and inferencemodes. For training we use padding for each temporal convolu-tion which could be calculated as (kernel_size−1)∗dilation.This situation is depicted on Figure 4(a). The network goesthrough a fixed chunk of data passed by training framework, thesize of which is bigger than the receptive field of the network andall layers are correctly trained. Hidden states here are passed inusual conditions from layer to layer automatically. For inferencewe are going to engage just a part of receptive field (Figure 4(b)).If we pass cropped inputs with padding it will most likely leadto dramatic accuracy reduction, as weights did not learned tohandle such situation. Instead of that we enable cache mecha-nism. For the first step we pass initiated by zeroes caches withthe input. Cache size for each convolution layer is equal paddingwe used for training. It is concatenated with input and bufferfor next iteration copied. Afterwards the whole tensor includedinput and history cache is convolved. Visual explanation of thissequence is presented on Figure 5. At the end of the networkwe will have a stack of buffers with the length equal to the totalnumber of convolution layers in encoder, bottleneck and decoder.Down and up blocks provoke shifting of starting point for thenext layer where new cache history is captured.

As mentioned previously, the down-sampling mechanism isequal to the convolution with stride 2 and we have to take it intoaccount when calculate the position for copy beginning. Thisleads to shape reduction.

4. ExperimentsWe conducted various experiments with our streaming TC Wave-U-Net model to challenge its speech enhancement quality anddemonstrate efficiency of cache mechanism.

4.1. Datasets

The for training clean data we used audio speech datasets ofChineese language AISHELL-1, AISHELL-3, sets Librispeech(train-clean-360) and VCTK for the English language. Theorganizers were provided the lists permitted audio files from thatdatasets with loudness larger than 15dB that could be used fortraining procedure. The total duration of clean training speechwas around 550 hours. The noise set was composed from twoparts. Part 1 is selected from public noise datasets MUSANand Audioset with total duration about 120 hours. The Part 2was a real meeting room noise recordings. The total amountof provided clips was 98 items. Also we got simulated RoomImpulse Responses (RIR) for the reverberation affect obtaining.Provided framework allowed to augment data on-the-fly duringtraining. It considerably slow-down the training procedure. Wegenerated and saved data locally.

4.2. Evaluation Items

Thus, the following denoising systems were taken to evaluatethe efficacy of our proposed approach:

• Baseline. LSTM-based solution, that was provided byorganizers itself. Basically the model process audio intime and frequency domains. The baseline has 8 channelsraw audio input. It calculates inter-channel phase dif-ference between pairs of microphone channels, complexSTFT for input signal and pass he concatenated tensorthrough 3 LSTMs. Each recurrent layer has 512 hiddenstates and, finally, they followed by a projection layer.Resulted mask is applied on channel 1 of original signaland inverse STFT returns the output of baseline solution.

• Wave-U-Net. The implementation which we took fromopen GitHub repository2. It exploits 12 convolution lay-ers in encoder and decoder. We just made changes incode to establish the ability of C-channel raw waveformprocessing instead of original mono channel input.

• TC Wave-U-Net. Our proposed model with time domaininput, casual convolutions in encoder decoder and bottle-neck and self-attentions in decoder.

4.3. Experimental Setup

We fed to network multi-channel audio recordings with additiverandom noise and convoluted with random RIRs. For trainingwe capture random piece with 16384 samples from 16 kHzraw waveforms and used that data for input. Our network isconstructed from 9 TC blocks in encoder and decoder, 1 casualconvolution layer in bottleneck. Filter size for convolutions inencoder is 15 and 5 in decoder accordingly, as mentioned inpaper [19]. All channel sizes for encoder are 8, 24, 48, 72, 96,120, 144, 168, 192, 216 (and 240 in bottleneck). For decoderthey are repeated in reverse order. Dilation parameter for a eachblock is changed by the following way 1, 1, 1, 2, 4, 5, 16, 32, 64.The receptive field of the encoder is 1807 samples or 112 ms.

4.4. Learning Target

For the learning objective we used weighted signal-to-distortionloss (WSDR) [20]. This is a time-domain loss function thatcould be defined by the following formula:

LwSDR(x, y, y) := αLSDR(y, y) + (1−α)LSDR(z, z) (2)

2https://github.com/haoxiangsnr/Wave-U-Net-for-Speech-Enhancement

Page 4: Real-time Streaming Wave-U-Net with Temporal Convolutions

where LSDR is a conventional signal-to-distortion (SDR) loss,x is a mixture signal is assumed as linear sum of clean speechsignal y and original noise z. Estimated noise z given as z =x− y. The y is an enhanced audio. Taken this into account theenergy ratio α between dry speech y and noise z is defined asα =‖ y ‖2 /(‖ y ‖2 + ‖ z ‖2).

For the baseline system, we trained mean square error(MSE), signal-to-distortion and WSDR. Our experiments showthat SDR performed better that MSE and WSDR was slightlybetter than SDR. In long training scenarios with SDR loss themodel wasn’t able to continue to converge while with WSDRloss even after 300 epoch the model steadily improved. We trainall our models using 6×Tesla V100.

For the final training we choose WSDR loss and Adamoptimizer with starting learning rate = 10−3. The decayed learn-ing scheduler was applied with minimum value 10−8 after 250epochs. Batch size is equal 1000.

The baseline was trained using default configuration param-eters, SDR objective loss, Adam optimizer with learning rate =10−4 and with reduce on plateau scheduler. Chunks of 4 secondswas chooses by default as well. The dataset was identical toprimary experiment.

4.5. Streaming Measurements Setup

We evaluated the performance of Baseline, Wave-U-Net and ourTC Wave-U-Net in streaming mode. Wave-U-Net consist ofsimple convolution layers and does not satisfy the requirementof challenge for zero-look ahead. We changed code to conductpseudo stream tensor processing. The chunk with 16384 samples(∼1 sec) is fed to the model with shift 640 samples (40 ms) and40 ms of predicted audio was taken from output. For the firsttimes we provide chunk 40 ms, 80 ms, 120 ms etc. with zero-padded left part till it reached full size 16384. For TC Wave-U-Net with cache we pass chunk window with 1024 samples (64ms), and move the window on 640 samples (40 ms).

5. Results5.1. Signal Quality

According to the conference rules, we push test dataset enhancedby our system to organizers that evaluate our resulting audio.Moreover, they provide mean opinion score (MOS) - measureof the human-judged overall quality of an audio. Each raterdetermines MOS, subjective speech MOS (S-MOS) and sub-jective noise MOS (N-MOS) for each cleaned recording. Next,confidence interval (CI) of MOS score is calculated. Each file islistened by more that 20 raters. In Table 1 we denoted the resultof comparing of our enhanced audio with noised raw waveforms.

5.2. Real Time Factor

In Table 3, we report Real Time Factor (RTF) (processing timedivided by audio duration) in relative scale where lower valuesindicate faster processing and lower user-perceived latency.

We can see that the network with cache enabled outperformsthe models working with whole receptive field. It a bit slowerthan other networks but it compensated by better precise. Inthe addition to mentioned above, in Table 2 we demonstrate theevaluation results of our model in streaming and non streamingmodes. Baseline and vanilla Wave-U-Net with pseudo streaming(see 4.5) are inferring close to the speed of TC Wave-U-Netwith cache but with the lower quality than these models fornon-streaming mode. PESQ for Baseline and Wave-U-Net with

Table 1: Results of evaluation test set enhanced with TC Wave-U-Net.

Audio Metrics

MOS S-MOS N-MOS CI

NOISY 2.56 2.93 3.03 0.02ENHANCED 2.90 3.05 3.05 0.04

Table 2: Evaluation results of multichannel speech enhancementon Single Linear Nonuniform MA.

SetUp Audio Metrics

PESQ STOI E-STOI SI-SNR

BASELINE Noisy 1.515 0.823 0.690 4.474BASELINE Enhanced 1.999 0.888 0.783 9.248WAVE-U-NET Enhanced 2.132 0.888 0.797 9.528TC WAVE-U-NET Enhanced 2.181 0.892 0.799 9.61TC WAVE-U-NET CACHE ENABLED (STREAMING) Enhanced 2.192 0.895 0.802 9.64

Table 3: Real Time Factor for Streaming mode and Model sizes.Number of parameters (Millions).

Model RTF PESQ Size

BASELINE (STREAMING) 0.7X 1.76 8.68WAVE-U-NET (STREAMING) 0.9X 1.93 10.14TC WAVE-U-NET (STREAMING) 3.3X 2.18 8.31TC WAVE-U-NET CACHE ENABLED (STREAMING) X 2.19 8.31

streaming 12% and 8% less, respectively, than non-streaming.Also Wave-U-Net in streaming mode don’t meet the requirementof challenge for zero-look ahead.

Finally we converted TC Wave-U-Net to ONNX format tospeed up the solution. It gave us 0.37 RTF on Intel Core i5clocked at 2.4GHz.

6. ConclusionsThis paper proposed our speech enhancement solution for Task1on ConferencingChallenge2021. We provided the Wave-U-Netbased network that outputs cleaned audio for passed raw wave-form with reverberations and additive noise. Our system em-ployed casual dilated convolutions for encoder, decoder and abottleneck parts. It also involved self-attentions in decoder forbetter precise. We implemented historical cache and obtain faststreaming inference.

Our evaluation showed that adding cache mechanism forthe model with large receptive field not only can reduce it forthe expected one, but also reduce floating-point calculations,thereby improving the inference speed. Even more we haveshown that compared to pure pseudo streaming, our proposedmethod provides the same quality as non-streaming model ordoes it a bit better. Due to the low time constraints, we aren’table to do the full research of the evaluation of calculation ofthe beginning of the cache, and have leaved it for the furtherresearch. Moreover, our final model is still training and we arewaiting for final metrics.

Future directions of this work could include experimentswith deeper versions of this architecture, new versions of TCblocks, incorporation of Channel Attention and using complexratio masking for signal enhancement.

7. AcknowledgementsThe work of Artem Sokolov is partially supported by RSF (Rus-sian Science Foundation) grant 20-71-10010.

Page 5: Real-time Streaming Wave-U-Net with Temporal Convolutions

8. References[1] M. C. Simmer KU, Bitzer J, “Post-filtering techniques. in micro-

phone arrays.” Springer, Berlin, Heidelberg, pp. 39–60, 2001.

[2] M. Brandstein, “Arrays, microphone. signal processing techniquesand applications.” Springer Science & Business Media, 2001.

[3] Wang, DeLiang, and Jitong Chen., “Supervised speech separationbased on deep learning: An overview.” IEEE/ACM Transactionson Audio, Speech, and Language Processing, pp. 1702–1726, May2018.

[4] Erdogan, Hakan, John R. Hershey, Shinji Watanabe, Michael I.Mandel, and Jonathan Le Roux., “Improved mvdr beamformingusing single-channel mask prediction networks.” Interspeech, pp.1981–1985, Sep. 2016.

[5] Ronneberger, O., Fischer, P. and Brox, T, “U-net: Convolu-tional networks for biomedical image segmentation,” InternationalConference on Medical image computing and computer-assistedintervention, pp. 234–241, Oct. 2015.

[6] Jansson, Andreas, Eric Humphrey, Nicola Montecchio, RachelBittner, Aparna Kumar, and Tillman Weyde, “Singing voice sepa-ration with deep u-net convolutional networks,” 18th InternationalSociety for Music Information Retrieval Conference, pp. 23–27,2017.

[7] Ernst, O., Chazan, S.E., Gannot, S. and Goldberger, J., “Improvedspeech enhancement with the wave-u-net.” 26th European SignalProcessing Conference (EUSIPCO), pp. 390–394, Sep. 2018.

[8] Soni, Meet H., Neil Shah, and Hemant A. Patil., “Time-frequencymasking-based speech enhancement using generative adversarialnetwork.” IEEE international conference on acoustics, speech andsignal processing (ICASSP), pp. 5039–5043, 2018.

[9] Macartney, C. and Weyde, T., “Speech dereverberation using fullyconvolutional networks.” arXiv preprint arXiv:1811.11307, 2018.

[10] Hochreiter, Sepp, and Jürgen Schmidhuber., “Long short-termmemory.” Neural computation, pp. 1735–1780, Nov. 1997.

[11] Tawara, N., Kobayashi, T. and Ogawa, T., “Multi-channel speechenhancement using time-domain convolutional denoising autoen-coder.” Interspeech, pp. 86–90, 2019.

[12] Dai, Z., Yang, Z., Yang, Y., Carbonell, J., Le, Q.V. and Salakhut-dinov, R., “Transformer-xl: Attentive language models beyond afixed-length context,” arXiv preprint arXiv:1901.02860, 2019.

[13] Giri, R., Isik, U., Krishnaswamy, A., “Attention wave-u-net forspeech enhancement,” IEEE Transactions on Acoustics, Speechand Signal Processing, vol. 28, pp. 249–253, Oct. 2019.

[14] He K, Zhang X, Ren S, Sun J., “Delving deep into rectifiers:Surpassing human-level performance on imagenet classification.”IEEE international conference on computer vision, pp. 1026–1034, 2015.

[15] Jansson, Andreas, Eric Humphrey, Nicola Montecchio, RachelBittner, Aparna Kumar, and Tillman Weyde, “Channel-attentiondense u-net for multichannel speech enhancement.” InternationalConference on Acoustics, Speech and Signal Processing, pp. 836–840, 2020.

[16] Ho, M.T., Lee, J., Lee, B.K., Yi, D.H. and Kang, H.G, “A cross-channel attention-based wave-u-net for multi-channel speech en-hancement.” Proceedings of Interspeech, pp. 4049–4053, 2020.

[17] Kriman, Samuel, Stanislav Beliaev, Boris Ginsburg, Joce-lyn Huang, Oleksii Kuchaiev, Vitaly Lavrukhin, Ryan Leary,Jason Li, and Yang Zhang, “Quartznet: Deep automaticspeech recognition with 1d time-channel separable convolu-tions.” International Conference on Acoustics, Speech and SignalProcessing (ICASSP), pp. 6124–6128, 2020.

[18] Bahdanau, D., Cho, K., Bengio, Y., “Neural machine transla-tion by jointly learning to align and translate.” arXiv preprintarXiv:1409.0473, 2014.

[19] D. Stoller, S. Ewert, and S. Dixon, “Wave-u-net: A multi-scaleneural network for end-to-end audio source separation,” arXivpreprint arXiv:1806.03185, 2019.

[20] H.-S. Choi, J.-H. Kim, J. Huh, A. Kim, J.-W. Ha, and K. Lee,“Phase-aware speech enhancement with deep complex u-net,” arXivpreprint arXiv:1912.09582, 2019.