From: maury on 14 Jul 2010 14:59 On Jul 14, 9:35 am, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote: > maury wrote: > > If you look at the normalized LMS (NLMS), then the maximum mu is > > limited to the sum of the number of filter coefficients. > > IMO what does matter is optimim rather then maximum. You need to > minimize the total error of the process. This error is because of > imperfect adaptation, noisy gradients, nonlinearity, ambient noise and > numeric artifacts. It also depends on the statistics and power of the > reference signal. So the algorithm must adapt Mu in pseudo Kalman way. > > Vladimir Vassilevsky > DSP and Mixed Signal Design Consultanthttp://www.abvolt.com Hi Vlad, With respect to the NLMS, the effect of the power of the reference depends on the implmentation. For example, in a system idemtification implementation, the power of the reference is not important since the normalization of the LMS is performed over the signal power. The algorithm is self-adjusting for imput power. However!!! The performance is very dependent on the model assumed. By nonlinearity, I take you to mean that a linear model will have errors if nonlinearities are present in the reference signal (the reference being some nonlinear inner-product of the signal and system vector). This just means the wrong model was chosen. If your model matches the system, then the squared error will not be great. No matter what the model, the algorithm will give the minimum squared error for the model. This doesn't mean that the coefficients you end up with are close to the desired coefficients. The other important parameter is the condition of the input auto- correlation matrix. An ill-conditioned matrix will result in not being able to determine an optimum estimate of the system impulse response. This is the problem with the example used by the OP. An input of 50 Hz (single signal) will result in an ill-conditioned input auto- correlartion matrix. There is NO optimum solution in this case. In fact, there are an infinite number of solutions to this problem. But, your point is well taken that the maximum mu used will be based on the optimum performance required from the problem. That performance may be minimum misadjustment noise, or maximum speed of convergence. Each will ditate a different "maximum" mu. What I think the OP was looking for, is how to determine the maximum mu for convergence (often equated with stability). This value is different for LMS versus NLMS. Maurice
From: Vladimir Vassilevsky on 14 Jul 2010 15:07 maury wrote: > On Jul 14, 9:35 am, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote: > >>maury wrote: >> >>>If you look at the normalized LMS (NLMS), then the maximum mu is >>>limited to the sum of the number of filter coefficients. >> >>IMO what does matter is optimim rather then maximum. You need to >>minimize the total error of the process. This error is because of >>imperfect adaptation, noisy gradients, nonlinearity, ambient noise and >>numeric artifacts. It also depends on the statistics and power of the >>reference signal. So the algorithm must adapt Mu in pseudo Kalman way. >> >>Vladimir Vassilevsky >>DSP and Mixed Signal Design Consultanthttp://www.abvolt.com > > > Hi Vlad, > With respect to the NLMS, the effect of the power of the reference > depends on the implmentation. For example, in a system idemtification > implementation, the power of the reference is not important since the > normalization of the LMS is performed over the signal power. The > algorithm is self-adjusting for imput power. Yes, but the input is catching the ambient noise and double talk besides the reflected signal. Hence if the input power is low, the algorithm should not crank Mu all the way up to infinity, as simple NLMS would do, but do the exactly the opposite. > However!!! The performance is very dependent on the model assumed. By > nonlinearity, I take you to mean that a linear model will have errors > if nonlinearities are present in the reference signal (the reference > being some nonlinear inner-product of the signal and system vector). > This just means the wrong model was chosen. If your model matches the > system, then the squared error will not be great. System is given to you. Your algorithm has to work with whatever you got. Vladimir Vassilevsky DSP and Mixed Signal Design Consultant http://www.abvolt.com
From: maury on 14 Jul 2010 16:38 On Jul 14, 2:07 pm, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote: > System is given to you. Your algorithm has to work with whatever you got. Kobayashi maru solution!
From: Jerry Avins on 14 Jul 2010 16:57 On 7/14/2010 4:38 PM, maury wrote: > On Jul 14, 2:07 pm, Vladimir Vassilevsky<nos...(a)nowhere.com> wrote: > >> System is given to you. Your algorithm has to work with whatever you got. > > Kobayashi maru solution! Hah! another Trekke! -- Engineering is the art of making what you want from things you can get.
From: Manny on 14 Jul 2010 18:09 On Jul 14, 8:07 pm, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote: > maury wrote: > > On Jul 14, 9:35 am, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote: > > >>maury wrote: > > >>>If you look at the normalized LMS (NLMS), then the maximum mu is > >>>limited to the sum of the number of filter coefficients. > > >>IMO what does matter is optimim rather then maximum. You need to > >>minimize the total error of the process. This error is because of > >>imperfect adaptation, noisy gradients, nonlinearity, ambient noise and > >>numeric artifacts. It also depends on the statistics and power of the > >>reference signal. So the algorithm must adapt Mu in pseudo Kalman way. > > >>Vladimir Vassilevsky > >>DSP and Mixed Signal Design Consultanthttp://www.abvolt.com > > > Hi Vlad, > > With respect to the NLMS, the effect of the power of the reference > > depends on the implmentation. For example, in a system idemtification > > implementation, the power of the reference is not important since the > > normalization of the LMS is performed over the signal power. The > > algorithm is self-adjusting for imput power. > > Yes, but the input is catching the ambient noise and double talk besides > the reflected signal. Hence if the input power is low, the algorithm > should not crank Mu all the way up to infinity, as simple NLMS would do, > but do the exactly the opposite. > > > However!!! The performance is very dependent on the model assumed. By > > nonlinearity, I take you to mean that a linear model will have errors > > if nonlinearities are present in the reference signal (the reference > > being some nonlinear inner-product of the signal and system vector). > > This just means the wrong model was chosen. If your model matches the > > system, then the squared error will not be great. > > System is given to you. Your algorithm has to work with whatever you got. What if you discovered by sheer luck that doing something wanky like skipping a beat works wonders to your signal? Sure this is uncool, inelegant, and can't be analyzed, only characterized. Do you scrap all this and wake up next day pretending that nothing has happened? -Momo
First
|
Prev
|
Next
|
Last
Pages: 1 2 3 Prev: Time-Varying Systems Next: Experience all New Kenlighten and grow your contacts and knowledge |