Upload
brati-sundar-nanda
View
844
Download
2
Embed Size (px)
DESCRIPTION
Citation preview
NOISE CANCELLATION USING ADAPTIVE FILTERS
Sweta Mohanty -1011016060Anwesha Samal -1011016057Brati Sundar Nanda -1011016238 Abhilash Mishra -1011016237
Guided By:- P.SHIVANI SAHOO
CONTENTS What Is Noise And Noise Cancellation? Adaptive Filter Basic Adaptive Filters Applications Of Adaptive Filters Problem Statement Various Adaptive Algorithms For Noise Cancellation LMS Algorithm NLMS Algorithm RLS Algorithm Affine Projection Algorithm SNRI Table Outputs Comparison Conclusion References
WHAT IS NOISE AND NOISE CANCELLATION?
• Noise consists of unwanted waveforms that can interfere with communication.• Noise can be internal or external to the system.• Sound noise: interferes with your normal hearing • Colored Noise• Impulsive Noise• White noise (AWGN)
• NOISE CANCELLATION: Noise cancellation is a method to reduce or cancel out undesirable components of the signal
ADAPTIVE FILTERS
A filter which adapts itself to the input signal given to it.
It is Non Linear And Time Variant.
Best suited when signal conditions are slowly changing.
Relies on recursive algorithm.
It has Adaptation algorithm for adjusting parameters for improved performance
It is meant to monitor the environment and varies the filter transfer function accordingly.
CONTINUED…
The basic operation of adaptive filter involves two
processes :
Filtering process• produces an output signal in response to a given input signal.
Adaptation process• aims to adjust the filter parameters to the environment.
BASIC ADAPTIVE FILTER
It contains 4 signals: Reference Signal [x(n)] Input Signal[d(n)] Filter output Signal[y(n)] Error Signal[e(n)]
Applications of Adaptive Filters:
NOISE CANCELLATION :Subtracts Noise from Received
Signal adaptively to improve SNR.
SIGNAL PREDICTION :
Used to provide a prediction of
the present value of a
random signal
SYSTEM IDENTIFICATION:
Designs an adaptive filter
that provides an approximation for an unknown
system.
ECHO CANCELLATION: Used to cancel
unknown interference
from a primary signal
VARIOUS ADAPTIVE ALGORITHMS FOR NOISE CANCELLATION
Properties of an ideal algorithm:Practical to implementAdapt to coefficient quickly to minimize errorProvide the Desired Performance
Different algorithms used are: Least Mean Squares (LMS) algorithm The Normalized Least Mean Squares(NLMS) algorithm The Recursive Least Squares (RLS) algorithm Affine Projection Algorithm(APA)
PROBLEM STATEMENT
We have taken an input random signal of N samples as reference signal.
We have taken random noise or adaptive Gaussian noise.
Then we are adding the noise signal with the input signal.
So the problem is to extract the input signal from output signal by eliminating the noise.
LMS ALGORITHM:
Adjusts the weight w(n) of the filter Adaptively adjusts the filter tap weights according to equation:
w(n+1)=w(n)+µe(n)x(n) Acts as negative feedback to minimize error signal It is robust in nature. Slow in convergence and sensitive to variations in step size
parameter. Requires number of iterations equals to dimensionality of the
input.
LMS ADAPTIVE FILTER BLOCK DIAGRAM
LMS ALGORITHM STEPS:
nenxnwnw 1
nyndne
1
0
][N
n
nwnxnyFilter output :
Estimation error:
Tap-weight adaptation:
• Each iteration of LMS involves three steps:
STABILITY:• Condition for stability is:
• Larger values for step size:-• Increases adaptation rate (faster adaptation)• Increases residual mean-squared error
power signalinput
20
ADVANTAGES AND DISADVANTAGES:
• It is simple in implementation.• Stable and robust performance against
different signal condition.ADVANTAGES
• Slow Convergence.DISADVANTAGES
OUTPUT:MEAN SQUARE ERROR FOR
LMS-----
VARIATION OF MSE WITH RESPECT TO Μ:
NLMS ALGORITHM:-
• In structural terms both NLMS filter is exactly same as a standard LMS filter.
• From one iteration to the next, the weight of an adaptive filter should be changed in a minimal manner.
NLMS CONTINUED..
One of the drawback of LMS is selection of step size parameter µ.
In order to solve this difficulty, we can use the NLMS (Normalized Least Mean Square) algorithm. Here the step size parameter is normalised.
So the NLMS algorithm is a time varying step-size algorithm, calculating the convergence factor μ as :
NLMS PARAMETERS:
Where || x(n) ||² is the squared Euclidean norm of the input x(n).
Here α is the adaption constant, which optimizes the convergence rate of the algorithm
Range of alpha is: 0<α<2 c is the constant term for normalization and always c<1. The updated filter weight is:
ADVANTAGES AND DISADVANTAGES:• As here µ is normalized this algorithm
converges faster than LMS.• Here estimated error value between the
desired signal and filter output is less than LMS.ADVANTAGES
• But LMS is less complex than NLMS and more stable.
DISADVANTAGES
OUTPUT: MEAN SQUARE ERROR FOR
NLMS-----
RLS ALGORITHM Recursively finds the filter coefficients that minimize a
weighted linear least squares cost function relating to the input signals.
In this algorithm the filter tap weight vector is updated using:
CONTINUED…
Whitens the input data by using inverse correlation matrix of data.
The Cost function C(n) should be minimized.
C(n)=
e(i)=d(i)-wH(n)u(i)where,
β(n,i) is weighting vector 0<β(n,i)<=1 i=1,2,3……,n β(n,i)=λn-i , where λ=forgetting factor
CONTINUED… REGULARISATION:
C(n)=
The sum of weighted error squares:
A regularizing term:
CONTINUED… Let Φ(n) is the correlation matrix of input u(i)
Φ(n)= λn-i u(i) uH(i)+ δλnI
Then the average cross correlation vector z(n) is given by-
z(n)=Φ(n)ŵ(n) , n=1,2………
Using the matrix inversion lemma, we can find the inverse of correlation matrix,
Φ-1(n) = P(n) (let)
Cost function is always expressed in terms of gain where ,K(n) is the
gain vector.
k(n) = P(n)u(n) = Φ-1(n)u(n)
CONTINUED… The tap weight vector ŵ(n)
ŵ(n)= Φ-1(n)z(n)
From the above equations, we summarize the RLS Algorithm as-
k(n) =
π(n) = P(n-1)u(n)
ξ(n) = d(n) – ŵH(n-1)u(n)
ŵ(n) = ŵ(n-1) + k(n)ξ*(n)
P(n) = λ-1 P(n-1) – λ-1 k(n) uH(n) P(n-1)
ADVANTAGES AND DISADVANTAGES OF RLS
•RLS converges faster than LMS and NLMS.
•Its noise cancellation capacity is the most.
ADVANTAGES
•This is the most complex algorithm of all the three algorithms.
DISADVANTAGES
MEAN SQUARE ERROR FORRLS-----
AFFINE PROJECTION ALGORITHM
Generalization of the well known normalized least mean square (NLMS) adaptive filtering algorithm.
Fast convergence compared to NLMS.
Computational complexity increases.
Convergence gets better with increase in filter order N.
Faster tracking capabilities than NLMS.
Better performance in steady state mean square error (MSE) than other algorithms.
APA MATHEMATICAL IMPLEMENTATION…
A(n) = input data matrix [N*N]
AH(n) =input data matrix in hermitian transpose[N*N]
d(n)= desired response [N*1]
Error can be computed as-
e(n)=d(n)-A(n)ŵ(n)
The updated tap weight vector can be calculated as-
ŵ(n+1)=ŵ(n)+μ AH(n)(A(n) AH(n))-1e(n)
CONVERGENCE & STABILITY OF APA
The learning curve of an APA consists of the sum of exponential terms.
It converges at a rate faster than that of a NLMS filter.
As more delayed versions of tap input vector is used, the rate of convergence improves, so does the computational complexity.
APA is less stable than LMS and NLMS algorithms, whereas it is more stable than RLS algorithm.
OUTPUT:MEAN SQUARE ERROR FOR APA-
SNRI TABLE:-• Signal to Noise ratio improvement= Final SNR-Original SNR
ALGORITHM SNRI
LMS 13.69
NLMS 18.009
APA 20.39
RLS 29.09
COMPARISION FOR CONVERGENCE FOR DIFFERENT ALGORITHMS:
COMPARISON OF MSE FOR DIFFERENT ALGORITHMS:
COMPARISON OF LMS,NLMS,APA AND RLS-----
•RLS converges faster than APA, APA converges faster than NLMS and NLMS converges faster than LMS.
CONVERGENCE:
•RLS is the most complex algorithm among the four algorithms. Hence , complexity is inversely proportional to convergence.
COMPLEXITY:
•Difference between final and initial SNR is highest in case of RLS then APA then NLMS then LMS.
SNR IMPROVEMENT:
CONCLUSION We studied the behavior of LMS, NLMS, APA and RLS algorithms
by implementing them in the adaptive filter for noise cancellation.
LMS was the simplest and easiest to implement but it converges at the slowest rate.
NLMS has a normalized step size making it converge faster than LMS but complexity also increases along with convergence rate.
APA is the improved version of NLMS with increasing convergence rate.
CONTINUED….
RLS is the fastest converging algorithm with maximum computational complexity. But it cancels maximum noise by minimizing error with the rapidest rate.
So we are making a tradeoff between computational complexity and convergence rate here to get the most noise free signal.
RLS is the best algorithm as it is faster than the other three.
REFERENCES Adaptive Filter Theory by Simon Haykin: 3rd edition, Pearson Education
Asia.LPE. Adaptive Signal Processing by John G Proakis, 3rd edition, Perntice Hall
of India. B. Widow, "Adaptive noise canceling: principles and applications",
Proceedings of the IEEE, vol. 63, pp. 1692-1716, 1975. A Family of Adaptive Filter Algorithms in Noise cancellation for Speech
Enhancement By Sayed. A. Hadei, Student Member IEEE and M. lotfizad. Steven L. Gay and Sanjeev Tavathia, “The Fast Affine Projection
Algorithm”, Acoustics Research Department, AT&T Bell Laboratories. Sundar G. Sankaran, Student Member, IEEE, and A. A. (Louis) Beex,
Senior Member, IEEE” Convergence Behavior of Affine Projection Algorithms”.