Field of the invention and description of prior art
[0001] The present invention relates to a method for estimation of at least one signal of
interest from at least one discrete-time signal. Furthermore, the invention relates
to a device for carrying out a method according to the invention.
[0002] In many applications, signals of interest are corrupted by noise sources and/or other
signals. Therefore, depending on the requirements of a given application, efforts
were taken to reduce the level of noise and/or other signals to a tolerable level.
[0003] In case of the signal of interest (target signal) being a continuous-time signal
to be processed by digital data processing, the target signal is usually measured
and transformed into a quantized discrete-time signal. Provided that the sampling
rate and the quantization levels were chosen properly the quantized discrete-time
signal comprises the desired target signal and, as an undesirable effect, also noise
sources and/or other signals.
[0004] Conventional phase-unaware amplitude estimation methods separate the target signal
(signal of interest) from noise and/or other signals by applying a frequency-dependent
gain function (mask) on observed noisy amplitude spectrum. Examples for such gain
functions are Wiener filter (as softmask) and binary mask. Noise reduction capability
obtained by conventional methods is limited since they only modify the amplitude or
phase individually.
Summary of the invention
[0005] It is an object of the present invention to provide an enhanced method for the estimation
of at least one target signal.
[0006] In a first aspect of the invention, this aim is achieved by means of above-mentioned
method, comprising the following steps:
- a) transforming the at least one discrete-time signal into a frequency domain to obtain
a complex spectrum of the at least one discrete-time signal;
- b) performing an amplitude estimation on the complex spectrum to obtain an estimated
amplitude spectrum of the at least one signal of interest;
- c) performing a phase estimation on the complex spectrum, said phase estimation being
an amplitude-aware phase estimation using an input signal to obtain an estimated phase
spectrum of the at least one signal of interest;
- d) performing an amplitude estimation on the complex spectrum, said amplitude estimation
being a phase-aware amplitude estimation using the result of the phase estimation
of step c) to obtain an enhanced complex spectrum of the at least one signal of interest.
[0007] By virtue of this approach according to the invention it is possible to perform accurate
estimations of at least one target signal comprised in at least one discrete-time
signal even under adverse conditions, i.e. highly correlated noise sources and/or
other signals. For example, the signal of interest can be any target signal included
in the at least one discrete-time signal. This approach according to the invention
pushes the limits of the conventional methods by introducing interaction between amplitude
estimation and phase estimation stages.
[0008] Step c) and d) according to the invention are based upon certain conditions. Amplitude-aware
phase estimation according to step c) requires at least an estimation of the amplitude
spectrum (signal magnitude spectrum) of the at least one signal of interest and preferably
estimation of the amplitude spectrum of the vector sum of all other sources. In particular,
the amplitude-aware phase estimator has been derived from the non-patent literature
"Phase estimation for signal reconstruction in single-channel speech separation" (P.
Mowlaee, R. Saiedi, and R. Martin, in Proceedings of the International Conference
on Spoken Language Processing, 2012, see appendix AP1) and "STFT phase improvement
for single channel speech enhancement" (
M. Krawczyk and T. Gerkmann, in International Workshop on Acoustic Signal Enhancement;
Proceedings of IWAENC, 2012, pp. 1-4). The phase-aware amplitude estimator has been derived from the non-patent literature
"On phase importance in parameter estimation in single-channel speech enhancement"
(
P. Mowlaee and R. Saeidi, in IEEE International Conference on Acoustics, Speech and
Signal Processing, May 2013, pp 7462-7466, see appendix AP2) and "MMSE-optimal spectral amplitude estimation given the STFT-phase"
(
T. Gerkmann and M. Krawczyk, Signal Processing Letters, IEEE, vol. 20, no. 2, pp.
129 - 132, Feb. 2013).
[0009] The at least one discrete-time signal can be of any source or the interaction of
sources, for example a noisy speech signal or the superposition of several speech
and/or noise signals. The at least one discrete-time signal could be obtained by observation,
measurement and/or calculation.
[0010] In a variant of the invention, the amplitude estimation on the complex spectrum in
step b) is performed irrespectively of the phase spectrum of the signal of interest.Such
an amplitude estimation can be a "phase-unaware amplitude estimation" referring to
any conventional amplitude estimation method which is performed irrespectively of
the phase spectrum of at least the at least one signal of interest.
[0011] Preferably, after step b) the result of the amplitude estimation of the preceding
step b) is used as an input signal in step c). In particular, the result of the amplitude
estimation is only used as an input signal in step c) if step c) follows in direct
order to step b).
[0012] According to a development of the invention, the amplitude estimation on the complex
spectrum in step b) is performed by a frequency-dependent time-frequency mask, in
particular by Wiener filtering of the complex spectrum.
[0013] In a further development of the invention, after step d) the steps c) and d) are
repeated iteratively wherein as input signal in step c) the result of the phase-aware
amplitude estimation of the preceding step d) is used. Therefore, a loop is closed
by a feedback from an output of a phase-aware amplitude estimator to an input of an
amplitude-aware phase estimator. Previous iterative speech enhancement methods aimed
at improving the spectral amplitude estimates only within the iterations. In these
methods, neither a phase enhancement stage nor a combined synthesis-analysis stage
was used within the feedback loop for the iterations. Instead, a noisy phase was exploited
in signal reconstruction. No phase information was taken into account to update the
signal parameters of an enhanced target signal. A synergistic effect in this closed
loop according to the invention stems from the fact that better amplitude estimation
assists the phase estimation and a better phase estimation assists the amplitude estimation.
These improvements can be continued by alternating between the two estimators multiple
times until a sufficient quality of the joint amplitude and phase estimates is obtained.
[0014] In yet another development of the invention, the consistency between the phase and
amplitude estimations of the enhanced complex spectrum (as the enhanced complex spectrum
provides an input in step c)) of the at least one signal of interest is monitored
according to the following comparison criterion:
F(
X)
= STFT(
STFT-1(
X))
- X, X being a matrix composed of a complex time-frequency representation of the enhanced
complex spectrum, wherein at least one quality index ε
(i) is established to measure inconsistency of complex time-frequency representations
denoted by
D(i), obtained at each loop-iteration, and defined for the i-th loop iteration as follows:

wherein the i-th ε
(i) quality index is calculated by

and the loop-iterations are stopped at least when a quality indexε
(i) gets lower than a pre-defined threshold ε
th. Establishing a quality index ε
(i) and comparison with a defined threshold ε
th allows to measure the decrease of the amount of inconsistency observed between the
phase and amplitude estimates obtained by each iteration before feedback to the phase
estimation. Therefore, the iterations can be stopped when the quality index ε
(i) gets lower than the pre-defined threshold ε
th, allowing fast and efficient processing of the transformed signal.
[0015] Advantageously, the threshold is ε
th = 0,05, which is especially suited as a comparison criterion for the quality index
ε
(i).
[0016] According to another development of the invention the iterations are stopped at least
after a predefined number of iterations, in particular after five, six or seven iterations.
This allows to limit the number of iterations and therefore to limit the computing
efforts. It is also possible to relate to the above mentioned comparison criterion
and to limit the number of iterations in case ε
(i) does not fall below the threshold ε
th.
[0017] In a variant of the invention the transformation method in step a) is a spectro-temporal
transformation, in particular STFT, Wavelet or sinusoidal signal modelling. For example
by replacing STFT representation with sinusoidal model, it is possible to reduce the
dimensionality of signal feature to a great level, hence less computational effort.
On the other hand, replacing STFT with other time-frequency transformations including
Wavelet or Wigner ville time-frequency representation for amplitude estimation or
Chirplet signal transformation and complex Wavelet transformation for representation
of both amplitude and phase enables to have a non-uniform resolution to analyze different
frequency bands, which is advantageous when applied to audio or speech signals.
[0018] Preferably, the at least one discrete-time signal can be a bio-medical, radar, image
or video signal. In this case, the complex time-frequency representation X can be
either one- or multidimensional. The matrix X is typically composed of frames as rows
and frequency bins as its columns (rows are often larger than the columns). For speech
signals, it is composed of a wide dynamic range of values (80 dB). For bio-medical
signals, the dynamic range is often much lower as the signal is sparse in time-frequency.
[0019] Alternatively, the method according to the invention is especially suited if the
at least one discrete-time signal is an audio signal.
[0020] In a further development of the invention the at least one discrete-time signal comprises
at least one speech signal. The speech signal can be the target signal, which is true
for many everyday life speech-related applications, in particular for automatic speech
recognition (ASR) applications. According to a challenging scenario the at least one
discrete-time signal can comprise two or even more speech signals. The target signal
is represented by one speech signal to be separated from the accompanying signals.
[0021] Furthermore, the at least one discrete-time signal can be derived from a single channel
signal. Single channel signals are common in many applications as they rely on a signal
obtained by a single microphone (cell phones, headsets,...) but usually do provide
less information than multi channel devices. Therefore, the requirements on signal
enhancement are very high, especially in case of single channel speech separation
(SCSS). Since the method according to the invention provides strongly enhanced target
signals it is exceptionally suited to be applied on single channel signals.
[0022] Alternatively, the at least one discrete-time signal can be derived from a multi
channel signal. An additional information provided by at least a second measurement
device can be processed to give an extraordinary accurate estimation of the at least
one target signal.
[0023] Of course, the method according to the invention is also suited to estimate two or
more target signals.
[0024] In a second aspect of the invention the aim to provide an enhanced method for the
estimation of at least one target signal is achieved by means of a device for carrying
out a method according to any of the preceding claims.
Brief description of the drawings
[0025] The specific features and advantages of the present invention will be better understood
through the following description. In the following, the present invention is described
in more detail with reference to exemplary embodiments (which are not to be construed
as limitative) shown in the drawings, which show:
Fig. 1 a schematic block-diagram illustrating the object of the invention,
Fig. 2 an exemplary schematic block-diagram of a state of the art multi-sensor speech
enhancement method,
Fig. 3 exemplary state of the art modifications of Fig. 2,
Fig. 4 an exemplary schematic block-diagram of a variant of the invention,
Fig. 5 an exemplary schematic block-diagram of another variant of the invention,
Fig. 6 a schematic block-diagram of the block "New Enhancement" according to the invention
shown in fig. 4 and 5,
Fig. 7 a detailed schematic block-diagram of the stopping rule block shown in fig.
4,
Fig. 8 a schematic block-diagram of a typical single-channel separation algorithm
based on amplitude estimation on a complex spectrum of a noisy signal described in
appendix AP1 in detail, said amplitude estimation being performed phase-unaware,
Fig. 9 a schematic block-diagram of amplitude-aware phase estimation described in
appendix AP1 in detail,
Fig. 10 two schematic block-diagrams of two different single-channel speech separation
algorithms described in appendix AP2 in detail.
Detailed description of the invention
[0026] Fig. 1 shows a schematic block-diagram illustrating the object of the invention.
Given an exemplary continuous-time signal y(t) which includes for example two different
signals s
1(t) and s
2(t), it is an object of the invention to estimate the clean signal ŝ
1(n) and/or ŝ
2(n). Assuming that the signal s
1(t) is a signal of interest and the signal s
2(t) represents for example interfering noise (the signal s
2(t) could stem from any other source or from a superposition of sources), a typical
approach to estimate the signal of interest s
1(t) consists of transforming the continuous-time signal y(t) into a quantized discrete-time
signal y(n) by applying an analog digital converter 1 on the continuous-time signal
y(t). As a next step, a signal estimation device 2 processes the discrete time signal
y(n) using a priori information to provide an estimate of at least the signal of interest
ŝ
1(n). In the given example an estimate of the signal ŝ
2(n) representing noise is provided as well.
[0027] Fig. 2 shows an exemplary schematic block-diagram of a state of the art multi-sensor
speech enhancement method (which can be applied by a signal estimation device 2 according
to Fig. 1) to be applied on M discrete-time signals exploited from a number of M sensors,
said speech enhancement method composed of three stages, i.e. analysis, modification
and synthesis. The analysis stage might consist in different signal representations
including short-time Fourier transformation (STFT), Sinusoidal modeling, polyphase
filter banks, Mel-frequency Cepstral analysis and/or any other suitable transformation
applicable on at least one discrete-time signal. The discrete-time signals exploited
from the number of M sensors are therefore transformed in a complex format providing
amplitude and phase parts of the signals. Furthermore, the analysis stage is required
to decompose the complex signals into a number of N different frequency channels,
hence a product of N x M samples are provided for the modification stage. The modification
stage known from the state of the art can be in two ways: a) amplitude enhancement,
in which any frequency-dependent gain function as amplitude estimator (e.g. Wiener
filter as a common choice) is employed together with a noise estimator given either
by a reference microphone or a noise tracking method, or b) phase enhancement, in
which the noisy phase is often directly copied to synthesize enhanced output signal.
Finally, the synthesis stage is applied on the resulting N x M samples of the modification
stage to reconstruct enhanced signals, in particular enhanced speech signals.
[0028] Fig. 3 shows exemplary state of the art modifications of the modification stage of
Fig. 2 (if not stated otherwise in the description of the figures, same reference
signs describe same features). The M discrete-time signals exploited from the number
of M sensors are analyzed in block A, providing N x M samples in a complex format
as described in Fig. 2. In block 3 the amplitude part contained in the complex format
of the samples is exploited (block 4 exploits the phase part contained in the complex
format of the samples). The samples are processed through amplitude or phase enhancement
stages, wherein the amplitude enhancement stage is provided with a noise estimate,
and finally synthesized in block S to provide an enhanced signal, in particular an
enhanced speech signal. The modifications of the samples can be categorized in four
different groups.
- A first group provides an estimate for the clean speech spectral amplitude based on
a noise estimate from a noise tracker or a reference sensor and a speech estimate
using a decision-directed method (see US 2009/0163168 A1 and Y. Ephraim and D. Malah, "Speech enhancement using a minimum-mean square error short-time
spectral amplitude estimator", IEEE Trans. Acoust., Speech, Signal Processing, vol.
32. no. 6, pp. 1109-1121, Dec 1984). The noisy phase is directly used unaltered when reconstructing a time-domain enhanced
speech signal at an output. This group can be represented in Fig. 3 by an amplitude
switch ASW being in a position P2 and a phase switch PSW being in a position P3.
- A second group (ASW in position P2 and PSW in position P4) refers to phase enhancement
only methods. For example non-patent literature see appendix AP1 and AP2 suggested
to employ Griffin and Lim iterations to estimate the signal phase for signal reconstruction
given the Wiener filtered amplitude spectrum, using synthesis-analysis in iterations.
- A third group (ASW in position P1 and PSW in position 4) refers to phase enhancement
only methods used with the noisy amplitude. The phase estimation often requires strong
assumptions knowing exact onsets and fundamental frequency of clean signals, in particular
speech signals and previous frame phase values.
- A fourth group (ASW in position P2 and PSW in position P4, but in contrast to the
second group no iterations) refers to a method assuming a clean spectral amplitude
is available and spectral amplitude is estimated in a phase-aware way in an open loop
configuration.
[0029] Fig. 4 shows an exemplary schematic block-diagram of a variant of the invention.
Block A and block S represent analysis and synthesis blocks as described in Fig. 2
and 3, wherein block A is provided with at least one discrete-time signal y(n). Block
A transforms the at least one discrete-time signal y(n) for example by an N-point
Fourier transform (any other time-frequency transformation providing amplitude and
phase spectra suffice for the method according to the invention) into a frequency
domain to obtain a complex spectrum, i.e.

where n = [0 ... N-1] with N the window size and Y(k) and φ
y(
k) as the kth frequency component for the magnitude and phase spectrum of y(n), respectively.
F({y(n)} of the at least one discrete-time signal y(n).
[0030] In contrast to the signal modification and enhancement methods described in Fig.
3, an enhanced signal modification method according to the invention is provided,
which is described in detail in Fig. 6. A block "New Enhancement" is provided with
a noise estimate, N x M samples, and, depending on the switching position of a loop
switch LSW,
- with an amplitude estimate of the amplitude part of the complex spectrum of at least
one corresponding signal of interest s1(n) (preferably from noise and/or other signals s2(n) as well) provided by a conventional enhancement block C (loop switch LSW in position
P1) or
- with an output signal of the new enhancement method, which is looped back as an input
signal sin(n) (loop switch LSW in position P2).
[0031] The conventional enhancement block C represents any phase-unaware amplitude estimator
or phase-unaware amplitude estimation methods (or any amplitude estimation method
performed irrespectively of the phase spectrum of at least the signal of interest
s
1(t)), which separate signal of interest s
1(n) from noise and/or other signals s
2(n) for example by applying a frequency-dependent gain function (mask) on observed
noisy amplitude spectrum. Examples for such gain functions are Wiener filter (as softmask)
and binary mask. Noise reduction capability obtained by conventional methods is limited
since they only modify the amplitude or phase individually. Preferably, the block
C and the block "New Enhancement" is provided with a noise estimate. Block C performs
an amplitude estimation on the complex spectrum

to obtain an estimated amplitude spectrum of the at least one signal of interest
s
1(n) (preferably from noise and/or other signals s
2(n) as well).
[0032] Furthermore, Fig. 4 shows a block "stopping rule", which provides a criterion to
stop the feedback loop. The block "new enhancement" is first provided with the amplitude
estimate of the complex spectrum

of the conventional block C. The output of the block "New Enhancement" can be looped
back as an input signal s
in(n) (the input signal s
in(n) can be in complex format) for the block "New Enhancement" in a following iteration.
The block "New Enhancement" is described in more detail in Fig. 6.
[0033] Fig. 5 shows an exemplary schematic block-diagram of another variant of the invention,
wherein the feedback loop differs from the variant shown in Fig. 4. Herein, the output
of the block "New Enhancement" is synthesized in Block S and analyzed in a following
analysis block A, before being looped back as an input signal s
in(n) to the block "New Enhancement", provided that the loop switch LSW being in position
P2. This allows monitoring the consistency between the phase and amplitude estimations
of the enhanced complex spectrum of the at least one signal of interest according
to the following comparison criterion:
F(X) = STFT(STFT-1(X)) - X, X being a matrix composed of a complex time-frequency representation of the enhanced
complex spectrum, wherein at least one quality index ε(i) is established to measure inconsistency of complex time-frequency representations
denoted by D(i), obtained by each loop-iteration, and defined for the i-th loop iteration as follows:

wherein the i-th ε
(i) quality index is calculated by

and the loop-iterations are stopped at least when a quality indexε
(i) gets lower than a pre-defined threshold ε
th. The threshold ε
th is preferably ε
th = 0,05.
[0034] Fig. 6 shows a schematic block-diagram of the block "New Enhancement" according to
the invention shown in Fig. 4 and 5. Herein, two blocks are shown processing the N
x M samples described in the preceding figures. Generally, the a block "amplitude-aware
phase estimation" performs a phase estimation on the complex spectrum

said phase estimation being an amplitude-aware phase estimation using the input signal
s
in(n) to obtain an estimated phase spectrum of the at least one signal of interest s
1(n), wherein the result of the phase-unaware amplitude estimation of the conventional
enhancement block C (see Fig. 4 and 5) is used as said input signal s
in(n). The block "amplitude-aware phase estimation" provides an enhanced phase estimation
of the at least one signal of interest s
1(t) (preferably, an enhanced phase estimation of the noise or any other signal s
2(t) as well) to a following block "phase-aware amplitude estimator". Within the block
"phase-aware amplitude estimator" an amplitude estimation on the complex spectrum

is performed, said amplitude estimation being a phase-aware amplitude estimation
using the result of the phase estimation of the block "amplitude-aware phase estimation"
to obtain an enhanced complex spectrum

of the at least one signal of interest s
1(n).
[0035] Fig. 7 shows a detailed schematic block-diagram of the block "stopping rule" shown
in Fig. 4. A block "consistency check" is provided with
- the estimated phase spectrum of the at least one signal of interest derived from the
block "amplitude-aware phase estimation" (see Fig. 6) and with
- the estimated amplitude spectrum of the at least one signal of interest derived from
the block "phase-aware Amplitude Estimator" (see Fig. 6).
[0036] The consistency of the enhanced complex spectrum

can either be assumed to converge after a certain number of iterations (for example
five, six or seven iterations) or a inconsistency criterion can be applied (for example
the quality index ε
(i) mentioned above) limiting the number of iterations.
[0037] Fig. 8 shows a schematic block-diagram of a typical single-channel separation algorithm
based on amplitude estimation on a complex spectrum of a noisy signal described in
appendix AP1, said amplitude estimation being performed phase-unaware. Herein, a signal
y comprises two signals s1 and s2 to be separated, wherein amplitude estimates Ŝ
1 and Ŝ
2 a noisy phase signal φ
y is applied to reconstruct the clean signals ŝ
1 and ŝ
2.
[0038] Fig. 9 shows a schematic block-diagram of amplitude-aware phase estimation. In contrast
to Fig. 8, the signal reconstruction is provided with phase information corresponding
to the signals s1 and s2 respectively. An minimum mean square error (MMSE) phase estimation
block is shown, which is provided with the amplitude estimates Ŝ
1 and Ŝ
2 and the signal y, said phase estimation being amplitude-aware and providing phase
signals φ̂
1 and φ̂
2. Detailed description of the algorithm is given in appendix AP1.
[0039] Fig. 10 shows two schematic block-diagrams of two different single-channel speech
separation algorithms. A typical method to estimate a clean speech amplitude
X̂ (corresponding to Ŝ
1 of Fig. 8 and 9) is shown in (a), wherein the amplitude estimation (within the block
"Gain function") is not provided with any phase information. Within the scope of this
specification such a amplitude estimation is referred to as being phase-unaware. In
contrast, Fig. 10 (b) provides an example of a phase-aware amplitude estimation (block
"Gain function"), wherein the amplitude estimation is based at least on the magnitude
spectra Y of the signal y and the phase spectra φ
x of a speech signal x, the phase spectra φ
x being the phase spectra of a speech signal x. Taking into account the phase spectra
φ
x of the speech signal x to calculate the a clean speech amplitude
X̂ makes the amplitude estimation phase-aware. Detailed description of the algorithm
is given in appendix AP2.
[0040] Of course, the terms phase-aware amplitude estimation and amplitude-aware phase estimation
defined herein do not relate to speech signals only. In fact, phase-aware amplitude
estimation and amplitude-aware phase estimation is applicable to a plurality of signals
and the speech signals described in appendix AP2 and AP1 just represent one utilization
of phase-aware amplitude estimation and amplitude-aware phase example, respectively.
Therefore, the invention is not limited to the examples given in this specification
and can be adjusted in any manner known to a person skilled in the art.
1. Method for estimation of at least one signal of interest (s
1(t), s
1(n)) from at least one discrete-time signal (y(n)), said method comprising the steps
of
a) transforming the at least one discrete-time signal (y(n)) into a frequency domain
to obtain a complex spectrum

of the at least one discrete-time signal (y(n));
b) performing an amplitude estimation on the complex spectrum

to obtain an estimated amplitude spectrum of the at least one signal of interest
(s1(t), s1(n));
c) performing a phase estimation on the complex spectrum

said phase estimation being an amplitude-aware phase estimation using an input signal
(sin(n)) to obtain an estimated phase spectrum of the at least one signal of interest
(s1(t), s1(n));
d) performing an amplitude estimation on the complex spectrum

said amplitude estimation being a phase-aware amplitude estimation using the result
of the phase estimation of step c) to obtain an enhanced complex spectrum

of the at least one signal of interest (s1(t), s1(n)).
2. Method of claim 1, wherein in step b) the amplitude estimation on the complex spectrum

is performed irrespectively of the phase spectrum of the signal of interest (s
1(t), s
1(n)).
3. Method of claim 1 or 2, wherein after step b) the result of the amplitude estimation
of the preceding step b) is used as an input signal (sin(n)) in step c).
4. Method of any of the claims 1 to 3, wherein in step b) the amplitude estimation on
the complex spectrum

is performed by a frequency-dependent time-frequency mask, in particular by Wiener
filtering of the complex spectrum
5. Method of any of the claims 1 to 4, wherein after step d) the steps c) and d) are
repeated iteratively wherein as input signal (sin(n)) in step c) the result of the phase-aware amplitude estimation of the preceding
step d) is used.
6. Method of any of the claims 1 to 5, wherein the consistency between the phase and
amplitude estimations of the enhanced complex spectrum

of the at least one signal of interest (s
1(t), s
1(n)) is monitored according to the following comparison criterion:
F(X) = STFT(STFT-1(X)) - X, X being a matrix composed of a complex time-frequency representation of the enhanced
complex spectrum, wherein at least one quality index ε (i) is established to measure inconsistency of complex time-frequency representations
denoted by D(i), obtained by each loop-iteration, and defined for the i-th loop iteration as follows:

wherein the i-th ε (i) quality index is calculated by

and the loop-iterations are stopped at least when a quality index ε (i) gets lower than a pre-defined threshold εth.
7. Method of claim 6, wherein the threshold is εth = 0,05.
8. Method of any of the claims 5 to 7, wherein the iterations are stopped at least after
a predefined number of iterations, in particular after five, six or seven iterations.
9. Method of any of the claims 1 to 8, wherein the transformation method in step a) is
a spectro-temporal transformation, in particular STFT, Wavelet or sinusoidal signal
modeling.
10. Method of any of the claims 1 to 9, wherein the at least one discrete-time signal
(y(n)) is a bio-medical, radar, image or video signal.
11. Method of any of the claims 1 to 9, wherein the at least one discrete-time (y(n))
signal is an audio signal.
12. Method of claim 11, wherein the at least one discrete-time signal (y(n)) comprises
at least one speech signal.
13. Method of claims 11 or 12, wherein the at least one discrete-time signal (y(n)) is
derived from a single channel signal.
14. Method of claims 11 or 12, wherein the at least one discrete-time signal (y(n)) is
derived from a multi channel signal.
15. Device for carrying out a method according to any of the preceding claims.