WO2011100491A2 - Adaptive systems using correntropy - Google Patents

Adaptive systems using correntropy Download PDF

Info

Publication number
WO2011100491A2
WO2011100491A2 PCT/US2011/024435 US2011024435W WO2011100491A2 WO 2011100491 A2 WO2011100491 A2 WO 2011100491A2 US 2011024435 W US2011024435 W US 2011024435W WO 2011100491 A2 WO2011100491 A2 WO 2011100491A2
Authority
WO
WIPO (PCT)
Prior art keywords
signal
cost function
correntropy
processing device
adaptive
Prior art date
Application number
PCT/US2011/024435
Other languages
French (fr)
Other versions
WO2011100491A3 (en
Inventor
Jose Carlos Principe
Abhishek Singh
Weifeng Liu
Original Assignee
University Of Florida Research Foundation Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University Of Florida Research Foundation Inc. filed Critical University Of Florida Research Foundation Inc.
Priority to US13/578,358 priority Critical patent/US9269371B2/en
Publication of WO2011100491A2 publication Critical patent/WO2011100491A2/en
Publication of WO2011100491A3 publication Critical patent/WO2011100491A3/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0264Noise filtering characterised by the type of parameter measurement, e.g. correlation techniques, zero crossing techniques or predictive techniques
    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03HIMPEDANCE NETWORKS, e.g. RESONANT CIRCUITS; RESONATORS
    • H03H21/00Adaptive networks
    • H03H21/0012Digital adaptive filters
    • H03H21/0043Adaptive algorithms

Definitions

  • Adaptive filters are used in a wide range of signal processing applications. While a least means square (LMS) adaptive algorithm can be used to adapt a filter by minimizing the mean squared value (second order moment) of the error between the output of the adaptive filter and the desired response, the performance of such adaptive filters can be limited when operating under impulsive noise conditions.
  • LMS least means square
  • FIGS. 1 and 2 are block diagrams of examples of adaptive systems in accordance with various embodiments of the present disclosure.
  • FIG. 3 is a graphical representation illustrating a non-limiting example of a comparison of the weight signal-to-noise ratio (SNR) of an adaptive system of FIG. 2 using mean squared error (MSE) and maximum correntropy criterion (MCC) in the presence of impulsive measurement noise in accordance with various embodiments of the present disclosure.
  • SNR weight signal-to-noise ratio
  • MSE mean squared error
  • MCC maximum correntropy criterion
  • FIG. 4 is a graphical representation illustrating a non-limiting example of a weight SNR of an adaptive system of FIG. 2 using MSE and MCC while tracking a time varying system in the presence of impulsive observation noise in accordance with various embodiments of the present disclosure.
  • FIG. 5 is a block diagram of another example of an adaptive system in accordance with various embodiments of the present disclosure.
  • FIGS. 6-8 are graphical representations illustrating non-limiting examples of a speech signal, a noise signal, and an acoustic transfer function of Fig. 5 in accordance with various embodiments of the present disclosure.
  • FIG. 9 is a graphical representation illustrating a non-limiting example of the weight tracks of one of the filter weights of an adaptive system of FIG. 5 based upon the conditions illustrated in FIGS. 6-8 in accordance with various embodiments of the present disclosure.
  • FIG. 10 is a graphical representation illustrating a non-limiting example of the SNR between an original speech signal and a cleaned signal with MSE and MCC in accordance with various embodiments of the present disclosure.
  • FIG. 1 1 is a graphical representation illustrating a non-limiting example of ERLE values of the MCC and MSE techniques in accordance with various embodiments of the present disclosure. Docket: 222106-2960
  • FIG. 12 is a block diagram of another example of an adaptive system including an adaptive kernel width in accordance with various embodiments of the present disclosure.
  • FIG. 13 is a graphical representation illustrating a non-limiting example of the errors and kernel size that are produced in the adaptive system of FIG. 12 in accordance with various embodiments of the present disclosure.
  • FIG. 14 is a graphical representation illustrating a non-limiting example of the Weight SNR plots obtained using an MCC cost function with different values of the kernel size in accordance with various embodiments of the present disclosure.
  • FIG. 15 is a flow chart illustrating an example of adaptive filtering using correntropy in accordance with various embodiments of the present disclosure.
  • FIG. 16 is a graphical representation illustrating an example of a signal processing device in accordance with various embodiments of the present disclosure.
  • the present disclosure describes various embodiments using a correntropy cost function for training of adaptive systems.
  • Adaptive systems are useful in a variety of signal processing applications such as channel equalization, noise cancellation, system modeling, etc.
  • adaptive systems may be used in communication devices such as, but not limited to, cellular telephones, headphones, and other devices utilizing noise cancellation and/or equalization.
  • An adaptive system that implements a correntropy cost function may be useful in reducing the detrimental effects of outliers and impulsive noise.
  • correntropy is a localized measure as Docket: 222 06-2960 opposed to a global measure
  • a correntropy cost function in an adaptive system is useful for reducing the detrimental effects of non-Gaussian noise.
  • Correntropy is a localized similarity measure between two arbitrary scalar random variables Zand 7, and correntropy is a function of higher order moments of the difference between the random variables. Correntropy is defined by:
  • V (X, Y) E[K (X - Y) ⁇ EQ. (1 ) where K ⁇ J ( ) is a positive definite kernel. Accordingly, correntropy is a measure of the similarity of two random variables within a small neighborhood determined by the kernel width (or size).
  • ⁇ (-) is assumed to be a Gaussian function (also called the kernel) with a standard deviation ⁇ , which is also called the kernel bandwidth or kernel size.
  • is assumed to be a Gaussian function (also called the kernel) with a standard deviation ⁇ , which is also called the kernel bandwidth or kernel size.
  • correntropy may be viewed as a correlation function between two random variables, containing higher (even) order moments of the error Docket: 222106-2960
  • adaptive systems implementing a mean squared error (MSE) technique can provide a global measure.
  • MSE mean squared error
  • second-order statistical techniques are applicable only when the distribution of error is Gaussian.
  • errors may be non-Gaussian such as filtering under impulsive noise conditions, or noise
  • FIG. 1 shown is a block diagram of an embodiment of an adaptive system 100.
  • the adaptive system 100 is configured to learn the parameters of a filter 102, which are unknown, by exploiting a correntropy measure between a primary input (or the desired signal) and the output of the filter 102.
  • the adaptive system 100 implements a cost function 106 (i.e. , criterion function), a parameter learning 108, and an adaptive filter 104 including a set of filter weights W.
  • the filter 102 and the adaptive filter 104 each receive a reference signal X.
  • the filter 102 outputs a filtered reference signal s wh which is based at least in part on the reference signal X.
  • the filtered reference signal s is added to a primary signal n and the sum of the filtered reference signal and the primary signal «, is equal to the desired signal ,.
  • the adaptive filter 104 outputs an adaptively-filtered reference signal yford which is based at least in part on the reference signal X.
  • the adaptively filtered reference signal is subtracted from the desired signal d it yielding an error signal e,.
  • the error signal e is applied to the cost function 106, which outputs a cost function signal J cohesive. Docket: 222106-2960
  • the cost function signal J235 is received by the parameter learning 108, which adjusts the adaptive filter 104 based at least in part on the cost function signal J Coco and/or the error signal e t .
  • the adaptive system 100 is configured to optimize the cost function signal J235 in such a way that the output of the adaptive filter 104 resembles the desired signal d,.
  • the cost function signal Jcetate may be considered optimized when comparison to a threshold meets predefined criterion (e.g., when J Container is at and/or below the threshold).
  • the cost function signal J n is defined to be equal to the correntropy between the desired signal d, and the filter output >,.
  • a normalized Gaussian kernel is used to compute correntropy.
  • the cost function signal J n may be written as:
  • each iteration in EQ. (9) includes an extra scaling factor which is an exponential function of the value of the error signal in that iteration.
  • This scaling factor reflects the outlier rejection property of the correntropy similarity measure.
  • An additional aspect of EQ. (9) is the correntropy kernel width, which is dependent on the input signal.
  • the kernel width defines the amplitude scale within which the similarity of the two random variables is computed.
  • a very large kernel size much larger than the variance of the error will therefore yield a similarity measure close to MSE value.
  • a Docket: 222106-2960 small value of kernel size, of the order of the error e,, variance will be useful for exploiting the properties of a correntropy measure.
  • the Silverman's rule of thumb of density estimation is appropriate to select the kernel size.
  • FIG. 2 shown is a block diagram of an embodiment of an adaptive system 200 configured to identify a system. Identifying a system involves determining the coefficients of an unknown filter by studying its response to a white input signal.
  • An unknown system such as unknown plant 202
  • An unknown system is modeled by the adaptive filter 104 in such a way that for the same input, the difference between the outputs of the unknown plant 202 and the adaptive filter 104 is minimized.
  • the adaptation is complicated by the presence of observation noise, which is added to the output of the unknown plant 202.
  • the difference may be considered minimized when comparison of the cost function signal J235 to a threshold meets predefined criterion (e.g. , when J Website is at and/or below the threshold).
  • the transfer function of the unknown plant 202 changes with time and may be tracked by the adaptive filter 104.
  • the model provided by the adaptive filter 104 may be used in controlling the unknown plant 202.
  • the input (X) to the unknown plant 202 and to the adaptive filter 104 is a white Gaussian signal (white noise) with a zero mean and unit variance.
  • FIGS. 3 and 4 Experimental differences between the adaptive system 200 (FIG. 2) including a MCC cost function with respect to an adaptive system including an MSE cost function are illustrated in FIGS. 3 and 4.
  • FIG. 3 illustrates a non-limiting example of a comparison of the weight signal-to-noise ratio (SNR) of MSE 310 and MCC 320 in impulsive measurement noise, each of which is graphically depicted with respect to the number of iterations.
  • SNR weight signal-to-noise ratio
  • the parameters of the experiment are defined as follows.
  • the coefficients of the unknown plant 202 of the adaptive system 200 are defined to be:
  • impulsive observation noise is simulated using a mixture of Gaussian distributions:
  • the Gaussian component N(0,10) creates strong outliers.
  • the kernel size for the MCC cost function is set to 2 for this case (in general, it should be a fraction of the variance of the expected outliers).
  • a very high kernel size e.g., more than 5 in this example
  • the step sizes for the two update equations (EQS. (9) and (10)) are such that when the observation noise is Gaussian, their performance is similar in terms of the weight SNR (WSNR), as calculated below.
  • FIG. 4 illustrates a non-limiting example of a weight SNR of MSE 410 and a weight SNR of MCC 420 while tracking a time varying system in the presence of impulsive observation noise (e.g., EQ. (12)).
  • EQ. (12) impulsive observation noise
  • a time-varying transfer function of the unknown plant 202 was used in FIG. 4, where the coefficients varied as follows: u (n - ⁇ ) W * EQ. (14) where u(-) is the unit step function.
  • u(-) is the unit step function.
  • the performance of the adaptive system 200 including the MCC cost function is better than performance of an adaptive system including an MSE cost function.
  • FIG. 5 shown is a block diagram of an embodiment of an adaptive system 500 configured to cancel acoustic noise.
  • a desired signal « e.g., speech, music, etc.
  • an acoustic noise signal s e.g., ambient noise in a room
  • an adaptive filter 104 tries to replicate the acoustic noise signal by modeling the acoustic transfer function 502 between the noise source and the noise s, contained in the desired signal ⁇ 3 ⁇ 4.
  • both the noise and the acoustic transfer function 502 may be non-stationary.
  • the difference between the desired signal di and the output signal yi of the adaptive filter 104 is in fact the noise-free signal (cleaned speech) 504.
  • FIG. 6 illustrates a non-limiting example of a speech signal n Vietnamese
  • FIG. 7 illustrates a non-limiting example of a noise signal X, which is the sound of a vacuum cleaner in use and has non-stationary characteristics.
  • the acoustic transfer function 502 may correspond to a typical closed room Docket: 222106-2960 environment for the experiment.
  • FIG. 8 illustrates a non-limiting example of an acoustic transfer function 502 of a small closed room using a 200 tap filter to model the acoustic path.
  • FIG. 9 shows a non-limiting example of the weight tracks of one of the filter weights based upon the conditions illustrated in FIGS. 6-8. Further, the rate of convergence was set to be the same for the adaptive system including the MSE cost function and the adaptive system 500 (FIG. 5) including the MCC cost function. As can be seen by comparing the MSE filter weight track 910 and the MCC filter weight track 920 in FIG. 9, the MSE cost function produced a much higher misadjustment at steady state than the MCC cost function. This is because of the highly impulsive nature of the speech signal «,-. The sudden, high amplitude "bursts" of samples which occur in speech signals can easily disturb the MSE weight track 910.
  • the MCC cost function places exponentially decreasing weights on samples that are distant and impulsive. Therefore, the steps and the overall weight track 920 of the MCC cost function are much less affected by such types of non-stationarities in the desired signal d,-.
  • FIG. 10 shows a non-limiting example of the SNR between the original speech n, signal and the cleaned signal (after averaging over a moving window of 10,000 samples, for visual clarity).
  • Echo Return Loss Enhancement is another measure of performance of an adaptive system 500 (FIG. 5) configured to cancel acoustic noise, and ERLE is more commonly used for echo cancellers. It measures the amount of noise (or echo) power that has been removed from the desired signal by subtracting the output of the adaptive filter 104. It is a ratio of the power of the desired signal and the error, expressed in decibels: Docket: 222106-2960
  • FIG. 1 1 shows a non-limiting example of ERLE values of the MCC and MSE techniques, after smoothing using a running window of 10,000 samples, for visual clarity.
  • the MCC cost function is able to remove more noise power from the desired signal d t than the MSE cost function.
  • the MCC cost function is robust against outliers and impulsive characteristics in an error signal. This robustness is attained by weighting the errors e, with exponential functions, and the kernel size controls how this weighting is done.
  • the kernel size ⁇ effectively discriminates between large valued errors (outliers) and small errors, by weighting them differently. As ⁇ - ⁇ , this discriminability is lost, and the behavior of the MCC cost function approaches a MSE cost function. Therefore, it is reasonable to believe that any value of ⁇ that is not very large is likely to result in a more robust algorithm than an MSE cost function.
  • a very small kernel size can result in very small weighting being given to most of the error samples, in which case the learning rate of the weights would be very slow.
  • FIG. 12 illustrates an embodiment of an adaptive system 1200 including an MCC cost function with an adaptive kernel width 1210 and parameter learning 1212.
  • KL divergence is the amount of information in a function /for discriminating against another function g, and KL divergence is defined as: -2960
  • f (e) is defined to be the estimated density function of a window of N samples of the error, evaluated using Gaussian kernel with kernel size ⁇ :
  • the first term in EQ. (20) is independent of the kernel width. Therefore, minimizing D KL (/II with respect to ⁇ is equivalent to maximizing the second term of EQ. (20). Therefore, the improved MCC cost function including an adapted kernel width is:
  • a log-likelihood based cross validatory cost function can be obtained by minimizing the KL divergence between the true and estimated densities
  • the KL divergence may be considered minimized when compared to a predefined threshold.
  • the kernel width update equation of EQ. (27) can potentially be used to update the MCC cost function of the adaptive system before updating the system parameters. Therefore, before each weight update step, we also update the kernel width with the adapted kernel width 1210. The weight update is performed by the parameter learning 1212, after incorporating the new kernel width, at each iteration.
  • An embodiment of the adaptive system 1200 is illustrated in FIG. 12. Docket: 222106-2960
  • FIG. 13 shows a non-limiting example of the errors 1310 that are produced in the adaptive filter setting as described above. It can be seen that the kernel width 1320 adapts to the values of the error 1310. When the weights of the plant are changed at the 800th iteration, large errors 1310 are produced. The kernel size 1320 also increases, and the filter 104 (FIG. 12) adapts, the errors 1310 become smaller, and the kernel width 1320 also anneals.
  • FIG. 14 shows a non-limiting example of the Weight SNR plots obtained using the MCC cost function with different values of the kernel size.
  • a small kernel size like ⁇ 0.4 effectively creates a very narrow cost function to adapt on. Therefore, as shown by plot 1420, when the system is away from the optimal solution, the adaptation rate is very slow.
  • the adaptive kernel size is able to strike an effective compromise between the two scenarios as illustrated by plot 1430. Docket: 222106-2960
  • the computational complexity of the correntropy cost function is simpler than the other methods and approximately the same as the MSE cost function. Therefore, an adaptive system implementing a correntropy cost function may be a practical way of adapting filters in real life situations, without making assumptions about the nature or statistics of the error signal e,-. Moreover, the correntropy cost function is useful for non-Guassian errors, whereas the MSE cost function is only effective for Guassian errors.
  • a reference signal is obtained in block 1502.
  • the reference signal e.g., white noise
  • a secondary source e.g., noise from an automobile, vacuum cleaner, or another individual.
  • an adaptive filter output is provided based at least in part upon one or more coefficients (or weights) of an adaptive filter and the reference signal.
  • An error signal is determined in block 1506 based at least in part upon a received signal and the adaptive filter output.
  • the received signal may be a desired signal including a filtered reference signal and a primary signal n t .
  • the received signal may include, but is not limited to, the output of the unknown plant and observation noise or speech, music, etc. corrupted by an acoustic noise signal such as, e.g. , noise from a vacuum cleaner that is affected by the acoustic transfer function of a room or road and/or engine noise that is affected by the acoustic transfer function of an automobile.
  • an acoustic noise signal such as, e.g. , noise from a vacuum cleaner that is affected by the acoustic transfer function of a room or road and/or engine noise that is affected by the acoustic transfer function of an automobile.
  • a correntropy cost function signal may be determined in block 1508.
  • the correntropy cost function signal may be used to determine if one or more coefficients (or weights) of an adaptive filter should be adjusted based upon a comparison with a predefined threshold. For example, if correntropy cost function signal is exceeds the threshold, than the coefficient(s) may be incrementally adjusted.
  • One or more of the coefficients of the adaptive filter may be adjusted based at least in part upon the correntropy cost function in block 1510.
  • the coefficient(s) of the adaptive filter may be adjusted based at least in part upon a positive gradient of the correntropy cost function.
  • the error signal of block 1506 may be used to an adjustment for a step size. If it is determined in block 1512 that another iteration should be performed, the adaptive filtering returns to block 1502 where another reference signal is obtained and used to provide an adaptive filter output in block 1504 based at least in part upon the adjusted filter coefficients.
  • the systems and methods provided herein can be implemented in hardware, software, firmware, or a combination thereof.
  • the method can be implemented in software or firmware that is stored in a memory and that is executed by a suitable instruction execution system.
  • the system can be implemented with any or a combination of the following technologies, which are all well known in the art: a discrete logic circuit(s) having logic gates for implementing logic functions upon data signals, an application specific integrated circuit (ASIC) having appropriate combinational logic gates, a programmable gate array(s) (PGA), a field programmable gate array (FPGA), digital signal processor (DSP), etc.
  • ASIC application specific integrated circuit
  • PGA programmable gate array
  • FPGA field programmable gate array
  • DSP digital signal processor
  • the adaptive systems described above may be implemented in a signal processing device 1600 such as the one illustrated in FIG. 16.
  • the signal processing device 1600 includes a receiver 1610, transmitter 1612, processing unit 1614, a bus 1616 and a memory 1618.
  • the memory 1618 stores an application specific software 1620 including modules 1622, which include instructions Docket: 222106-2960 that when executed by the processing unit 1614 perform various operations.
  • the modules 1622 may be, for example, an adaptive filter module 1604, a parameter learning module 1608, a criterion or cost function module 1606, and an adaptive kernel width module 1602.
  • the various modules 1622 may correspond to the blocks shown in the adaptive systems of FIGS. 1 , 2, 5, and/or 12.
  • each block represents a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s).
  • the functions noted in the blocks may occur out of the order noted in FIG. 15.
  • two blocks shown in succession in FIG. 15 may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved, as will be further clarified hereinbelow.
  • the application specific software 1620 can also be stored on a variety of computer-readable media for use by, or in connection with, a variety of computer-related systems or methods.
  • a "computer-readable medium” stores, communicates, propagates, or transports the program for use by or in connection with the instruction execution system, apparatus, or device.
  • the computer readable medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device.
  • the computer-readable medium may include the following: an electrical connection (electronic) having one or more wires, a portable computer diskette (magnetic), a random access memory (RAM) (electronic), a read-only memory (ROM) (electronic), an erasable programmable read-only memory (EPROM, EEPROM, or Flash memory) (electronic), an optical fiber (optical), a portable compact Docket: 222106-2960 disc read-only memory (CD OM) (optical), a digital versatile disc (optical), a high definition digital versatile disc (optical), and a Blu-ray Disc (optical).
  • ratios, concentrations, amounts, and other numerical data may be expressed herein in a range format. It is to be understood that such a range format is used for convenience and brevity, and thus, should be interpreted in a flexible manner to include not only the numerical values explicitly recited as the limits of the range, but also to include all the individual numerical values or sub-ranges encompassed within that range as if each numerical value and sub-range is explicitly recited.
  • a concentration range of "about 0.1 % to about 5%” should be interpreted to include not only the explicitly recited concentration of about 0.1 wt% to about 5 wt%, but also include individual concentrations (e.g., 1 %, 2%, 3%, and 4%) and the sub-ranges (e.g., 0.5%, 1.1 %, 2.2%, 3.3%, and 4.4%) within the indicated range.
  • the term "about” can include traditional rounding according to significant figures of 1 024435

Abstract

Various methods and systems are provided for related to adaptive systems using correntropy. In one embodiment, a signal processing device includes a processing unit and a memory storing an adaptive system executable in the at least one processing unit. The adaptive system includes modules that, when executed by the processing unit, cause the signal processing device to adaptively filter a desired signal using a correntropy cost function. In another embodiment, a method includes adjusting a coefficient of an adaptive filter based at least in part on a correntropy cost function signal, providing an adaptive filter output signal based at least in part on the adjusted coefficient and a reference signal, and determining an error signal based at least in part on a received signal and the adaptive filter output signal.

Description

Docket: 222106-2960
ADAPTIVE SYSTEMS USING CORRENTROPY
CROSS REFERENCE TO RELATED APPLICATIONS
[0001] This application claims priority to copending U.S. provisional application entitled "ROBUST TRAINING OF ADAPTIVE SYSTEMS USING CORRENTROPY" having serial no. 61/303,751 , filed February 12, 2010, the entirety of which is hereby incorporated by reference.
STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT
[0002] This invention was made with government support under agreement ECS-
0601271 awarded by the National Science Foundation. The Government has certain rights in the invention.
BACKGROUND
[0003] Adaptive filters are used in a wide range of signal processing applications. While a least means square (LMS) adaptive algorithm can be used to adapt a filter by minimizing the mean squared value (second order moment) of the error between the output of the adaptive filter and the desired response, the performance of such adaptive filters can be limited when operating under impulsive noise conditions.
BRIEF DESCRIPTION OF THE DRAWINGS
[0004] Many aspects of the present disclosure can be better understood with reference to the following drawings. The components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present disclosure. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views. Docket: 222106-2960
[0005] FIGS. 1 and 2 are block diagrams of examples of adaptive systems in accordance with various embodiments of the present disclosure.
[0006] FIG. 3 is a graphical representation illustrating a non-limiting example of a comparison of the weight signal-to-noise ratio (SNR) of an adaptive system of FIG. 2 using mean squared error (MSE) and maximum correntropy criterion (MCC) in the presence of impulsive measurement noise in accordance with various embodiments of the present disclosure.
[0007] FIG. 4 is a graphical representation illustrating a non-limiting example of a weight SNR of an adaptive system of FIG. 2 using MSE and MCC while tracking a time varying system in the presence of impulsive observation noise in accordance with various embodiments of the present disclosure.
[0008] FIG. 5 is a block diagram of another example of an adaptive system in accordance with various embodiments of the present disclosure.
[0009] FIGS. 6-8 are graphical representations illustrating non-limiting examples of a speech signal, a noise signal, and an acoustic transfer function of Fig. 5 in accordance with various embodiments of the present disclosure.
[0010] FIG. 9 is a graphical representation illustrating a non-limiting example of the weight tracks of one of the filter weights of an adaptive system of FIG. 5 based upon the conditions illustrated in FIGS. 6-8 in accordance with various embodiments of the present disclosure.
[0011] FIG. 10 is a graphical representation illustrating a non-limiting example of the SNR between an original speech signal and a cleaned signal with MSE and MCC in accordance with various embodiments of the present disclosure.
[0012] FIG. 1 1 is a graphical representation illustrating a non-limiting example of ERLE values of the MCC and MSE techniques in accordance with various embodiments of the present disclosure. Docket: 222106-2960
[0013] FIG. 12 is a block diagram of another example of an adaptive system including an adaptive kernel width in accordance with various embodiments of the present disclosure.
[0014] FIG. 13 is a graphical representation illustrating a non-limiting example of the errors and kernel size that are produced in the adaptive system of FIG. 12 in accordance with various embodiments of the present disclosure.
[0015] FIG. 14 is a graphical representation illustrating a non-limiting example of the Weight SNR plots obtained using an MCC cost function with different values of the kernel size in accordance with various embodiments of the present disclosure.
[0016] FIG. 15 is a flow chart illustrating an example of adaptive filtering using correntropy in accordance with various embodiments of the present disclosure.
[0017] FIG. 16 is a graphical representation illustrating an example of a signal processing device in accordance with various embodiments of the present disclosure.
DETAILED DESCRIPTION
[0018] Disclosed herein are various embodiments of systems and methods related to adaptive systems using correntropy. Reference will now be made in detail to the description of the embodiments as illustrated in the drawings, wherein like reference numbers indicate like parts throughout the several views.
[0019] The present disclosure describes various embodiments using a correntropy cost function for training of adaptive systems. Adaptive systems are useful in a variety of signal processing applications such as channel equalization, noise cancellation, system modeling, etc. For example, adaptive systems may be used in communication devices such as, but not limited to, cellular telephones, headphones, and other devices utilizing noise cancellation and/or equalization. An adaptive system that implements a correntropy cost function may be useful in reducing the detrimental effects of outliers and impulsive noise. In particular, because correntropy is a localized measure as Docket: 222 06-2960 opposed to a global measure, a correntropy cost function in an adaptive system is useful for reducing the detrimental effects of non-Gaussian noise.
[0020] Correntropy is a localized similarity measure between two arbitrary scalar random variables Zand 7, and correntropy is a function of higher order moments of the difference between the random variables. Correntropy is defined by:
V (X, Y) = E[K (X - Y)} EQ. (1 ) where K<J( ) is a positive definite kernel. Accordingly, correntropy is a measure of the similarity of two random variables within a small neighborhood determined by the kernel width (or size).
[0021] In practice, only a finite number of samples { ( x„ yt- ) } of the variables Zand
7 are available ( = 1 , N). Accordingly, the sample estimator for the expectation operator is expressed as:
Figure imgf000005_0001
where Κο(-) is assumed to be a Gaussian function (also called the kernel) with a standard deviation σ, which is also called the kernel bandwidth or kernel size. Using a Taylor series expansion of the exponential function in the Gaussian kernel and substituting in EQ. (1), the sample estimator for the expectation operator may also be expressed as:
Figure imgf000005_0002
[0022] As can be seen above, correntropy may be viewed as a correlation function between two random variables, containing higher (even) order moments of the error Docket: 222106-2960
(X- Y) between the two random variables, within a small neighborhood determined by the kernel width (or size).
[0023] In contrast, adaptive systems implementing a mean squared error (MSE) technique (i.e. , minimizing the MSE between a desired signal and the filter output) can provide a global measure. Yet, such second-order statistical techniques are applicable only when the distribution of error is Gaussian. In other practical scenarios, errors may be non-Gaussian such as filtering under impulsive noise conditions, or noise
cancellation in speech signals, which are typically heavy-tailed. A measure based on just second order statistics, like MSE, can easily become biased under such conditions. However, the localization provided by the kernel width in the correntropy calculation may be useful in reducing the detrimental effects of outliers and impulsive noise.
[0024] Turning now to FIG. 1 , shown is a block diagram of an embodiment of an adaptive system 100. The adaptive system 100 is configured to learn the parameters of a filter 102, which are unknown, by exploiting a correntropy measure between a primary input (or the desired signal) and the output of the filter 102. The adaptive system 100 implements a cost function 106 (i.e. , criterion function), a parameter learning 108, and an adaptive filter 104 including a set of filter weights W. As illustrated in FIG. 1 , the filter 102 and the adaptive filter 104 each receive a reference signal X. The filter 102 outputs a filtered reference signal s„ which is based at least in part on the reference signal X.
[0025] The filtered reference signal s, is added to a primary signal n and the sum of the filtered reference signal and the primary signal «, is equal to the desired signal ,. The adaptive filter 104 outputs an adaptively-filtered reference signal y„ which is based at least in part on the reference signal X. The adaptively filtered reference signal is subtracted from the desired signal dit yielding an error signal e,. The error signal e, is applied to the cost function 106, which outputs a cost function signal J„. Docket: 222106-2960
[0026] The cost function signal J„ is received by the parameter learning 108, which adjusts the adaptive filter 104 based at least in part on the cost function signal J„ and/or the error signal et. The adaptive system 100 is configured to optimize the cost function signal J„ in such a way that the output of the adaptive filter 104 resembles the desired signal d,. For example, the cost function signal J„ may be considered optimized when comparison to a threshold meets predefined criterion (e.g., when J„ is at and/or below the threshold).
[0027] The cost function signal Jn is defined to be equal to the correntropy between the desired signal d, and the filter output >,. In this example, a normalized Gaussian kernel is used to compute correntropy.
Figure imgf000007_0001
[0028] If the filter weights at the nth time instant are Wn , the cost function signal Jn may be written as:
Figure imgf000007_0002
[0029] Analytically, it is challenging to determine the set of filter weights W„ , which can maximize the cost function signal J„. Therefore, an iterative gradient ascent approach is used where the next set of filter weights W of the adaptive filter 104 are determined by taking a small step μ along the positive gradient of the cost function Jn in the weight space. The step size μ is chosen according to the application, but the theory of gradient descent states that it has an upper bound which is determined by the largest eigenvalue of the input autocorrelation matrix. Therefore,
W n + 1 = W n + ' uVJ n EQ. (7) Docket: 222106-2960
Substituting the cost function signal Jn and computing the gradient with respect to W„ the next set of filter weights may be expressed as:
Figure imgf000008_0001
where e, = dj - jX, . Approximating the sum by the current value (N= 1) inspired by the stochastic gradient, the maximum correntropy criterion (MCC) update equation is
Figure imgf000008_0002
The weight update equation using the correntropy cost function has therefore been reduced to a simple form, as can be seen above.
[0030] For purposes of comparison with the MCC update equation, the LMS weight update is presented below:
L MS L M S „
W , (n + l ,) = W n + n X EQ. (10)
As can be seen above, each iteration in EQ. (9) includes an extra scaling factor which is an exponential function of the value of the error signal in that iteration. This scaling factor reflects the outlier rejection property of the correntropy similarity measure.
Therefore, the adaptation of weights using a MCC cost function is expected to be more stable if the desired signal dt has strong outliers or impulsive characteristics. This robustness comes at almost no greater algorithmic complexity than a least means square (LMS) algorithm.
[0031] An additional aspect of EQ. (9) is the correntropy kernel width, which is dependent on the input signal. From the definition of correntropy, the kernel width defines the amplitude scale within which the similarity of the two random variables is computed. Using in the correntropy measure a very large kernel size much larger than the variance of the error will therefore yield a similarity measure close to MSE value. A Docket: 222106-2960 small value of kernel size, of the order of the error e,, variance will be useful for exploiting the properties of a correntropy measure. The Silverman's rule of thumb of density estimation is appropriate to select the kernel size.
[0032] Turning now to FIG. 2, shown is a block diagram of an embodiment of an adaptive system 200 configured to identify a system. Identifying a system involves determining the coefficients of an unknown filter by studying its response to a white input signal. An unknown system, such as unknown plant 202, is modeled by the adaptive filter 104 in such a way that for the same input, the difference between the outputs of the unknown plant 202 and the adaptive filter 104 is minimized. The adaptation is complicated by the presence of observation noise, which is added to the output of the unknown plant 202. The difference may be considered minimized when comparison of the cost function signal J„ to a threshold meets predefined criterion (e.g. , when J„ is at and/or below the threshold). In some embodiments, the transfer function of the unknown plant 202 changes with time and may be tracked by the adaptive filter 104. The model provided by the adaptive filter 104 may be used in controlling the unknown plant 202. In the configuration illustrated in FIG. 2, the input (X) to the unknown plant 202 and to the adaptive filter 104 is a white Gaussian signal (white noise) with a zero mean and unit variance.
[0033] Experimental differences between the adaptive system 200 (FIG. 2) including a MCC cost function with respect to an adaptive system including an MSE cost function are illustrated in FIGS. 3 and 4. Specifically, FIG. 3 illustrates a non-limiting example of a comparison of the weight signal-to-noise ratio (SNR) of MSE 310 and MCC 320 in impulsive measurement noise, each of which is graphically depicted with respect to the number of iterations. For FIG. 3, the parameters of the experiment are defined as follows. The coefficients of the unknown plant 202 of the adaptive system 200 are defined to be:
W = [0.1, 0.2, 0.3, 0.4, 0.5, 0.4, 0.3, 0.2, 0. l]' Ea Docket: 222106-2960
Further, the impulsive observation noise is simulated using a mixture of Gaussian distributions:
Figure imgf000010_0001
Note that the Gaussian component N(0,10) creates strong outliers. Further, the kernel size for the MCC cost function is set to 2 for this case (in general, it should be a fraction of the variance of the expected outliers). A very high kernel size (e.g., more than 5 in this example) would make the correntropy cost function behave similar to the MSE cost function, while a very low step size (e.g., less than 0.5 in this example) would slow the adaptation. The step sizes for the two update equations (EQS. (9) and (10)) are such that when the observation noise is Gaussian, their performance is similar in terms of the weight SNR (WSNR), as calculated below.
Figure imgf000010_0002
[0034] As can be seen in FIG. 3, when adaptation is performed in the presence of impulsive noise (e.g., EQ. (12)), there is a significant difference in performance between the MSE cost function and the MCC cost function. Whenever a high amplitude outlier is encountered in the desired signal d, (or in the error signal e,) of FIG. 2, a LMS weight update equation corresponding to the MSE cost function will be forced to make a large increment, which takes the weights away from the true values. The overall effect of several such outliers is that the adaptive filter weights W keep jittering around the optional values, and exhibit noisy weight tracks as illustrated in by the MSE weight SNR 310 of FIG. 3. In contrast, the MCC weight tracks are more robust in such situations. Referring back to the MCC update equation (see EQ. (9) above), the exponential function of the error signal et provides stability when high amplitude outliers are encountered, keeping the overall weight track close to the optimal as illustrated in by the higher MCC weight SNR 320 of FIG. 3. Docket: 222106-2960
[0035] Additionally, FIG. 4 illustrates a non-limiting example of a weight SNR of MSE 410 and a weight SNR of MCC 420 while tracking a time varying system in the presence of impulsive observation noise (e.g., EQ. (12)). Instead of the weights of EQ. (1 1 ), a time-varying transfer function of the unknown plant 202 was used in FIG. 4, where the coefficients varied as follows: u (n - ΙΟΟθ) W* EQ. (14)
Figure imgf000011_0001
where u(-) is the unit step function. As can be seen by comparing the weight SNR of
MSE 410 and the weight SNR of MCC 420 in FIG. 4, the performance of the adaptive system 200 including the MCC cost function is better than performance of an adaptive system including an MSE cost function.
[0036] Moving now to FIG. 5, shown is a block diagram of an embodiment of an adaptive system 500 configured to cancel acoustic noise. Given a desired signal «, (e.g., speech, music, etc.) corrupted by an acoustic noise signal s, (e.g., ambient noise in a room), an adaptive filter 104 tries to replicate the acoustic noise signal by modeling the acoustic transfer function 502 between the noise source and the noise s, contained in the desired signal <¾. In practical scenarios, both the noise and the acoustic transfer function 502 may be non-stationary. The difference between the desired signal di and the output signal yi of the adaptive filter 104 is in fact the noise-free signal (cleaned speech) 504.
[0037] Experimental results corresponding to the adaptive system 500 configured to cancel acoustic noise are discussed below. FIG. 6 illustrates a non-limiting example of a speech signal n„ and FIG. 7 illustrates a non-limiting example of a noise signal X, which is the sound of a vacuum cleaner in use and has non-stationary characteristics. The acoustic transfer function 502 may correspond to a typical closed room Docket: 222106-2960 environment for the experiment. FIG. 8 illustrates a non-limiting example of an acoustic transfer function 502 of a small closed room using a 200 tap filter to model the acoustic path.
[0038] FIG. 9 shows a non-limiting example of the weight tracks of one of the filter weights based upon the conditions illustrated in FIGS. 6-8. Further, the rate of convergence was set to be the same for the adaptive system including the MSE cost function and the adaptive system 500 (FIG. 5) including the MCC cost function. As can be seen by comparing the MSE filter weight track 910 and the MCC filter weight track 920 in FIG. 9, the MSE cost function produced a much higher misadjustment at steady state than the MCC cost function. This is because of the highly impulsive nature of the speech signal «,-. The sudden, high amplitude "bursts" of samples which occur in speech signals can easily disturb the MSE weight track 910. However, since correntropy is a localized comparison measure, the MCC cost function places exponentially decreasing weights on samples that are distant and impulsive. Therefore, the steps and the overall weight track 920 of the MCC cost function are much less affected by such types of non-stationarities in the desired signal d,-.
[0039] FIG. 10 shows a non-limiting example of the SNR between the original speech n, signal and the cleaned signal (after averaging over a moving window of 10,000 samples, for visual clarity). As can be seen by comparing the MSE SNR 1010 and the MCC SNR 1020 in FIG. 10, there is a substantial improvement in the SNR as obtained by MCC cost function and that obtained with the MSE cost function.
[0040] Echo Return Loss Enhancement (ERLE) is another measure of performance of an adaptive system 500 (FIG. 5) configured to cancel acoustic noise, and ERLE is more commonly used for echo cancellers. It measures the amount of noise (or echo) power that has been removed from the desired signal by subtracting the output of the adaptive filter 104. It is a ratio of the power of the desired signal and the error, expressed in decibels: Docket: 222106-2960
ERLE ( db ) = 10 log EQ. (15)
Figure imgf000013_0001
FIG. 1 1 shows a non-limiting example of ERLE values of the MCC and MSE techniques, after smoothing using a running window of 10,000 samples, for visual clarity. As can be seen by comparing the MSE ERLS track 1 110 and the MCC ERLS track 1120 in FIG. 1 1 , the MCC cost function is able to remove more noise power from the desired signal dt than the MSE cost function.
[0041] As demonstrated above, the MCC cost function is robust against outliers and impulsive characteristics in an error signal. This robustness is attained by weighting the errors e, with exponential functions, and the kernel size controls how this weighting is done. The kernel size σ effectively discriminates between large valued errors (outliers) and small errors, by weighting them differently. As σ - ∞, this discriminability is lost, and the behavior of the MCC cost function approaches a MSE cost function. Therefore, it is reasonable to believe that any value of σ that is not very large is likely to result in a more robust algorithm than an MSE cost function. However, a very small kernel size can result in very small weighting being given to most of the error samples, in which case the learning rate of the weights would be very slow.
[0042] It is also possible to adapt the kernel width in a way that improves the performance of the MCC cost function. FIG. 12 illustrates an embodiment of an adaptive system 1200 including an MCC cost function with an adaptive kernel width 1210 and parameter learning 1212. To formulate such an adaptive kernel width, the Kullback-Leibler (KL) divergence between the density of the error samples estimated using the kernel and the true density is computed. KL divergence or "discriminant information" is the amount of information in a function /for discriminating against another function g, and KL divergence is defined as: -2960
Figure imgf000014_0001
[0043] In an adaptive filter configuration, f (e) is defined to be the estimated density function of a window of N samples of the error, evaluated using Gaussian kernel with kernel size σ:
1 L
EQ. (17)
[0044] Substituting the true density function / ye) for the function_ (x) and the
σ
estimated density function .(e) for g(x) in the KL divergence equation and simplifying the equation further,
Figure imgf000014_0002
[0045] The first term in EQ. (20) is independent of the kernel width. Therefore, minimizing DKL (/II with respect to σ is equivalent to maximizing the second term of EQ. (20). Therefore, the improved MCC cost function including an adapted kernel width is:
Figure imgf000014_0003
[0046] This is also the cross entropy of the estimated density function, by using the true density function to compute the expected value. Using the simple estimator for the expectation operator, and using EQ. (17),
(', - *, )] J EQ (22)
Figure imgf000014_0004
Docket: 222106-2960
[0047] Therefore, a log-likelihood based cross validatory cost function can be obtained by minimizing the KL divergence between the true and estimated densities The KL divergence may be considered minimized when compared to a predefined threshold. Taking the derivative of JKL( ) with respect to σ yields,
Figure imgf000015_0001
[0048] Going back to the original adaptive filter configuration, using the above equation we can formulate a simple gradient ascent-based search rule for the adapted kernel width 1210 to update the kernel size, at every weight update step of the adaptive filter. Therefore,
dJKL ( σ ) EQ. (25) δ σ
Figure imgf000015_0002
[0049] The stochastic approximation of the gradient can be used by dropping the expectation operator and evaluating the operand at the current sample of the error. Therefore the final update equation becomes: Docket: 222106-2960
Figure imgf000016_0001
[0050] The computational complexity of this adaptation technique, per iteration, is O(L), where L is the length of the window used for computing the density estimate. L should be selected to appropriately estimate the error dynamics for the application. A special case of the above update rule arises when L = 1 , in which case the equation takes the simple form,
Figure imgf000016_0002
This simplified form has been experimentally shown to be sufficient in many practical applications.
[0051] Although the stochastic gradient approximation of EQ. (27) reduces computational complexity as compared to EQ. (26), it is more easily affected by an outlier in the data. If en is an outlier in the stochastic gradient of EQ. (27), the denominator in the gradient diminishes, causing a "jump" in adaptation. In practice, this effect can be alleviated by having a small regularization constant, e, added to the denominator.
[0052] The kernel width update equation of EQ. (27) can potentially be used to update the MCC cost function of the adaptive system before updating the system parameters. Therefore, before each weight update step, we also update the kernel width with the adapted kernel width 1210. The weight update is performed by the parameter learning 1212, after incorporating the new kernel width, at each iteration. An embodiment of the adaptive system 1200 is illustrated in FIG. 12. Docket: 222106-2960
[0053] Such a configuration was tested on an example of a system identification problem. The input (X) to the plant and the adaptive filter was unit variance white Gaussian signal. For the first 800 samples (or iterations), the coefficients of the plant were chosen to be:
W* = 3 * [0.1, 0.5, 0.3, 0.2]r EQ. (29)
After 800 samples, the coefficients of the plant were changed to be:
W* = 0.5 * [0.1, 0.5, 0.3, 0.2]' EQ- (3°)
[0054] For the experiment, this allows one to see how quickly the system adapts to new weights, for various kernel widths. To create non-Gaussian errors, the following is added as observation noise:
0.95N(0, 10"4 ) + 0.05N (0, 5) EQ. (31)
[0055] FIG. 13 shows a non-limiting example of the errors 1310 that are produced in the adaptive filter setting as described above. It can be seen that the kernel width 1320 adapts to the values of the error 1310. When the weights of the plant are changed at the 800th iteration, large errors 1310 are produced. The kernel size 1320 also increases, and the filter 104 (FIG. 12) adapts, the errors 1310 become smaller, and the kernel width 1320 also anneals.
[0056] FIG. 14 shows a non-limiting example of the Weight SNR plots obtained using the MCC cost function with different values of the kernel size. As indicated by plot 1410, a relatively large kernel size such as σ = 2, for example, helps in faster learning, but the robustness to outliers reduces and therefore the weights get affected by them, resulting in lower WSNR values. A small kernel size like σ = 0.4 effectively creates a very narrow cost function to adapt on. Therefore, as shown by plot 1420, when the system is away from the optimal solution, the adaptation rate is very slow. The adaptive kernel size is able to strike an effective compromise between the two scenarios as illustrated by plot 1430. Docket: 222106-2960
[0057] Other ways have been proposed to address non-Guassian errors. For example, optimization of information theoretic quantities like entropy of the error have also been proposed as training principles, and their superiority over the conventional MSE cost function has been demonstrated. However, the computational complexity of those methods is high.
[0058] In contrast, the computational complexity of the correntropy cost function is simpler than the other methods and approximately the same as the MSE cost function. Therefore, an adaptive system implementing a correntropy cost function may be a practical way of adapting filters in real life situations, without making assumptions about the nature or statistics of the error signal e,-. Moreover, the correntropy cost function is useful for non-Guassian errors, whereas the MSE cost function is only effective for Guassian errors.
[0059] Referring next to FIG. 15, shown is a flow chart 1500 illustrating an example of adaptive filtering using correntropy in accordance with various embodiments of the present disclosure. Initially, a reference signal is obtained in block 1502. For example, the reference signal (e.g., white noise) may be an input signal to an unknown plant or noise from a secondary source (e.g., noise from an automobile, vacuum cleaner, or another individual). In block 1504, an adaptive filter output is provided based at least in part upon one or more coefficients (or weights) of an adaptive filter and the reference signal. An error signal is determined in block 1506 based at least in part upon a received signal and the adaptive filter output. The received signal may be a desired signal including a filtered reference signal and a primary signal nt. For example, the received signal may include, but is not limited to, the output of the unknown plant and observation noise or speech, music, etc. corrupted by an acoustic noise signal such as, e.g. , noise from a vacuum cleaner that is affected by the acoustic transfer function of a room or road and/or engine noise that is affected by the acoustic transfer function of an automobile. Docket: 222106-2960
[0060] A correntropy cost function signal may be determined in block 1508. In some implementations, the correntropy cost function signal may be used to determine if one or more coefficients (or weights) of an adaptive filter should be adjusted based upon a comparison with a predefined threshold. For example, if correntropy cost function signal is exceeds the threshold, than the coefficient(s) may be incrementally adjusted. One or more of the coefficients of the adaptive filter may be adjusted based at least in part upon the correntropy cost function in block 1510. The coefficient(s) of the adaptive filter may be adjusted based at least in part upon a positive gradient of the correntropy cost function. The error signal of block 1506 may be used to an adjustment for a step size. If it is determined in block 1512 that another iteration should be performed, the adaptive filtering returns to block 1502 where another reference signal is obtained and used to provide an adaptive filter output in block 1504 based at least in part upon the adjusted filter coefficients.
[0061] The systems and methods provided herein can be implemented in hardware, software, firmware, or a combination thereof. In one embodiment, the method can be implemented in software or firmware that is stored in a memory and that is executed by a suitable instruction execution system. If implemented in hardware, as in an alternative embodiment, the system can be implemented with any or a combination of the following technologies, which are all well known in the art: a discrete logic circuit(s) having logic gates for implementing logic functions upon data signals, an application specific integrated circuit (ASIC) having appropriate combinational logic gates, a programmable gate array(s) (PGA), a field programmable gate array (FPGA), digital signal processor (DSP), etc.
[0062] In some embodiments, the adaptive systems described above may be implemented in a signal processing device 1600 such as the one illustrated in FIG. 16. The signal processing device 1600 includes a receiver 1610, transmitter 1612, processing unit 1614, a bus 1616 and a memory 1618. The memory 1618 stores an application specific software 1620 including modules 1622, which include instructions Docket: 222106-2960 that when executed by the processing unit 1614 perform various operations. As illustrated in FIG. 16, the modules 1622 may be, for example, an adaptive filter module 1604, a parameter learning module 1608, a criterion or cost function module 1606, and an adaptive kernel width module 1602. In some embodiments, the various modules 1622 may correspond to the blocks shown in the adaptive systems of FIGS. 1 , 2, 5, and/or 12.
[0063] The flow chart of FIG. 15 shows the architecture, functionality, and operation of a possible implementation of the adaptive filtering software 1622. In this regard, each block represents a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of the order noted in FIG. 15. For example, two blocks shown in succession in FIG. 15 may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved, as will be further clarified hereinbelow.
[0064] The application specific software 1620 can also be stored on a variety of computer-readable media for use by, or in connection with, a variety of computer-related systems or methods. In the context of this disclosure, a "computer-readable medium" stores, communicates, propagates, or transports the program for use by or in connection with the instruction execution system, apparatus, or device. The computer readable medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device. More specific examples (a non-exhaustive list) of the computer-readable medium may include the following: an electrical connection (electronic) having one or more wires, a portable computer diskette (magnetic), a random access memory (RAM) (electronic), a read-only memory (ROM) (electronic), an erasable programmable read-only memory (EPROM, EEPROM, or Flash memory) (electronic), an optical fiber (optical), a portable compact Docket: 222106-2960 disc read-only memory (CD OM) (optical), a digital versatile disc (optical), a high definition digital versatile disc (optical), and a Blu-ray Disc (optical).
[0065] Any process descriptions or blocks should be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps in the process, and alternate implementations are included within the scope of the embodiments described in the present disclosure in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those reasonably skilled in the art of the present disclosure.
[0066] It should be emphasized that the above-described embodiments of the present disclosure are merely possible examples of implementations set forth for a clear understanding of the principles of the disclosure. Many variations and modifications may be made to the above-described embodiment(s) without departing substantially from the spirit and principles of the disclosure. All such modifications and variations are intended to be included herein within the scope of this disclosure and protected by the following claims.
[0067] It should be noted that ratios, concentrations, amounts, and other numerical data may be expressed herein in a range format. It is to be understood that such a range format is used for convenience and brevity, and thus, should be interpreted in a flexible manner to include not only the numerical values explicitly recited as the limits of the range, but also to include all the individual numerical values or sub-ranges encompassed within that range as if each numerical value and sub-range is explicitly recited. To illustrate, a concentration range of "about 0.1 % to about 5%" should be interpreted to include not only the explicitly recited concentration of about 0.1 wt% to about 5 wt%, but also include individual concentrations (e.g., 1 %, 2%, 3%, and 4%) and the sub-ranges (e.g., 0.5%, 1.1 %, 2.2%, 3.3%, and 4.4%) within the indicated range. The term "about" can include traditional rounding according to significant figures of 1 024435
Docket: 222106-2960 numerical values. In addition, the phrase "about 'x' to 'y'" includes "about 'χ' to about V".

Claims

Docket: 222106-2960 CLAIMS Therefore, at least the following is claimed:
1. A signal processing device, comprising:
at least one processing unit; and
a memory storing an adaptive system executable in the at least one processing unit, the adaptive system comprising modules that when executed by the at least one processing unit cause the signal processing device to adaptively filter a desired signal using a correntropy cost function.
2. The signal processing device of claim 1 , wherein the adaptive system
comprises:
an adaptive filter module including a set of filter weights, the adaptive filter configured to provide an adaptive filter output signal based at least in part upon the set of filter weights and a reference signal; and
a parameter learning module configured to adjust the set of filter weights based at least in part upon the correntropy cost function and an error signal based at least in part upon the difference between the desired signal and the adaptive filter output signal.
3. The signal processing device of claim 2, wherein the adaptive system adjusts the set of filter weights based upon the error signal associated with the previous filter weight adjustment.
4. The signal processing device of claim 2, wherein the adaptive system iteratively adjusts the set of filter weights based upon a series of previous error signals, each error signal associated with one of a series of previous filter weight adjustments. Docket: 222106-2960
5. The signal processing device of claim 2, wherein the correntropy cost function is scaled by an exponential function of the error signal.
6. The signal processing device of claim 2, further configured to obtain the desired signal and the reference signal.
7. The signal processing device of claim 2, wherein the reference signal is noise from a secondary source and the desired signal comprises speech and acoustic noise from the secondary source.
8. The signal processing device of claim 2, wherein the reference signal is an input signal to an unknown plant and the desired signal comprises observation noise and an output signal from the unknown plant.
9. The signal processing device of claim 1 , wherein the signal processing device is a cellular telephone.
10. A method, comprising:
adjusting, by a signal processing device, a coefficient of an adaptive filter based at least in part on a correntropy cost function signal;
providing, by the signal processing device, an adaptive filter output signal based at least in part on the adjusted coefficient and a reference signal; and determining, by the signal processing device, an error signal based at least in part on a received signal and the adaptive filter output signal.
11. The method of claim 10, wherein the coefficient of the adaptive filter is adjusted based at least in part upon a positive gradient of the correntropy cost function.
12. The method of claim 1 1 , wherein the coefficient of the adaptive filter is
incrementally adjusted based at least in part upon the positive gradient of the cost function and a previous error signal. Docket: 222106-2960
13. The method of claim 1 0, further comprising further adjusting coefficient of the adaptive filter based at least in part upon the determined error signal and a positive gradient of the correntropy cost function .
14. The method of claim 13, wherein the providing, determining, and further
adjusting steps are iteratively repeated.
15. The method of claim 10, wherein the error signal is the difference between the received signal and the adaptive filter output signal.
16. The method of claim 10, wherein the correntropy cost function signal is scaled by an exponential function of the value of the error signal.
17. The method of claim 10, wherein the coefficient of the adaptive filter is adjusted in response to a comparison of the correntropy cost function signal with a predetermined threshold.
18. The method of claim 17, wherein the coefficient of the adaptive filter is adjusted when the correntropy cost function signal exceeds the predetermined threshold.
19. A method of canceling noise, comprising:
modeling, by a signal processing device, a noise signal using a correntropy cost function;
subtracting, by a signal processing device, the modeled noise signal from a received signal to yield a noise-free output signal; and
providing, by a signal processing device, the noise-free output signal.
20. The method of claim 1 9, wherein modeling the noise signal using the correntropy cost function comprises:
iteratively adjusting a set of filter coefficients of an adaptive filter based at least in part upon the correntropy cost function and the noise-free output signal Docket: 222106-2960 at a previous interation; and
determining the noise signal based at least in part upon the set of adjusted filter coefficients and a current input noise signal.
The method of claim 20, wherein the set of filter coefficients is adjusted based at least in part upon a positive gradient of the correntropy cost function.
PCT/US2011/024435 2010-02-12 2011-02-11 Adaptive systems using correntropy WO2011100491A2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/578,358 US9269371B2 (en) 2010-02-12 2011-02-11 Adaptive systems using correntropy

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US30375110P 2010-02-12 2010-02-12
US61/303,751 2010-02-12

Publications (2)

Publication Number Publication Date
WO2011100491A2 true WO2011100491A2 (en) 2011-08-18
WO2011100491A3 WO2011100491A3 (en) 2011-11-24

Family

ID=44368445

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2011/024435 WO2011100491A2 (en) 2010-02-12 2011-02-11 Adaptive systems using correntropy

Country Status (2)

Country Link
US (1) US9269371B2 (en)
WO (1) WO2011100491A2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106093724A (en) * 2016-06-20 2016-11-09 山东大学 A kind of shelf depreciation Wavelet noise-eliminating method based on hybrid particle swarm
CN107342751A (en) * 2017-06-28 2017-11-10 西安交通大学 A kind of variable step SA adaptive filter algorithms based on cross-correlation entropy
CN108768560A (en) * 2018-05-03 2018-11-06 上海电机学院 Adaptive non-integer delay time estimation method under low signal-to-noise ratio impulse noise environment

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2806053A1 (en) 2010-07-22 2012-01-26 University Of Florida Research Foundation, Inc. Classification using correntropy
WO2013003026A2 (en) 2011-06-30 2013-01-03 University Of Florida Research Foundation, Inc. Adaptive background estimation
US10147442B1 (en) * 2015-09-29 2018-12-04 Amazon Technologies, Inc. Robust neural network acoustic model with side task prediction of reference signals
CN110070136B (en) * 2019-04-26 2022-09-09 安徽工程大学 Image representation classification method and electronic equipment thereof
CN111416595B (en) * 2020-04-08 2022-04-08 北京航空航天大学 Big data filtering method based on multi-core fusion
CN114614797B (en) * 2022-05-12 2022-09-30 之江实验室 Adaptive filtering method and system based on generalized maximum asymmetric correlation entropy criterion
CN116741151B (en) * 2023-08-14 2023-11-07 成都筑猎科技有限公司 User call real-time monitoring system based on call center

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5157596A (en) * 1987-07-17 1992-10-20 Hughes Aircraft Company Adaptive noise cancellation in a closed loop control system
US5974434A (en) * 1997-10-07 1999-10-26 Ralph E. Rose Method and apparatus for automatically tuning the parameters of a feedback control system
US20080293372A1 (en) * 2005-10-31 2008-11-27 University Of Florida Research Foundation, Inc. Optimum Nonlinear Correntropy Filted

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5943661A (en) * 1991-07-11 1999-08-24 Texas Instruments Incorporated Hybrid neural network classifier, systems and methods
WO2007027839A2 (en) * 2005-09-01 2007-03-08 University Of Florida Research Foundation, Inc. Device and methods for enhanced matched filtering based on correntropy
US8611839B2 (en) * 2007-04-26 2013-12-17 University Of Florida Research Foundation, Inc. Robust signal detection using correntropy
US8428661B2 (en) * 2007-10-30 2013-04-23 Broadcom Corporation Speech intelligibility in telephones with multiple microphones

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5157596A (en) * 1987-07-17 1992-10-20 Hughes Aircraft Company Adaptive noise cancellation in a closed loop control system
US5974434A (en) * 1997-10-07 1999-10-26 Ralph E. Rose Method and apparatus for automatically tuning the parameters of a feedback control system
US20080293372A1 (en) * 2005-10-31 2008-11-27 University Of Florida Research Foundation, Inc. Optimum Nonlinear Correntropy Filted

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106093724A (en) * 2016-06-20 2016-11-09 山东大学 A kind of shelf depreciation Wavelet noise-eliminating method based on hybrid particle swarm
CN107342751A (en) * 2017-06-28 2017-11-10 西安交通大学 A kind of variable step SA adaptive filter algorithms based on cross-correlation entropy
CN108768560A (en) * 2018-05-03 2018-11-06 上海电机学院 Adaptive non-integer delay time estimation method under low signal-to-noise ratio impulse noise environment

Also Published As

Publication number Publication date
WO2011100491A3 (en) 2011-11-24
US20130211829A1 (en) 2013-08-15
US9269371B2 (en) 2016-02-23

Similar Documents

Publication Publication Date Title
US9269371B2 (en) Adaptive systems using correntropy
CN110100457B (en) Online dereverberation algorithm based on weighted prediction error of noise time-varying environment
US11315587B2 (en) Signal processor for signal enhancement and associated methods
US9805734B2 (en) Signal processing device, signal processing method and signal processing program for noise cancellation
EP3329488B1 (en) Keystroke noise canceling
US10056092B2 (en) Residual interference suppression
Albu et al. New variable step size affine projection algorithms
KR101568937B1 (en) Apparatus and method for supressing non-linear echo talker using volterra filter
Wung et al. Robust multichannel linear prediction for online speech dereverberation using weighted householder least squares lattice adaptive filter
Malik et al. Double-talk robust multichannel acoustic echo cancellation using least-squares MIMO adaptive filtering: transversal, array, and lattice forms
Sunnydayal et al. A survey on statistical based single channel speech enhancement techniques
Kim et al. Variable step-size affine projection algorithm based on global speech absence probability for adaptive feedback cancellation
Chinaev et al. A priori SNR Estimation Using a Generalized Decision Directed Approach.
Ciochină et al. An optimized affine projection algorithm for acoustic echo cancellation
Bachute et al. Performance analysis and comparison of complex LMS, sign LMS and RLS algorithms for speech enhancement application
US11837248B2 (en) Filter adaptation step size control for echo cancellation
Zhao et al. Spline Adaptive Filter
Jungmann et al. Room impulse response reshaping by p-norm optimization based on estimates of room impulse responses
Liu et al. Steady-state performance for the sign normalized algorithm based on hammerstein spline adaptive filtering
Malenovsky et al. Optimal step-size LMS algorithm using exponentially averaged gradient vector
Triki Performance issues in recursive least-squares adaptive GSC for speech enhancement
Bekrani et al. Convergence analysis of clipped input adaptive filters applied to system identification
Gabrea An adaptive Kalman filter for the enhancement of speech signals in colored noise
Ravi et al. Speech Enhancement Using Kernel and Normalized Kernel Affine Projection Algorithm
Bhotto Improved robust adaptive-filtering algorithms

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11742832

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 13578358

Country of ref document: US

122 Ep: pct application non-entry in european phase

Ref document number: 11742832

Country of ref document: EP

Kind code of ref document: A2