Field of the Invention
[0001] The present invention relates generally to the field of perceptual audio coding and
more particularly to a method for determining masking thresholds using a psychoacoustic
model.
Background of the Invention
[0002] In present state of the art audio coders, perceptual models based on characteristics
of a human ear are typically employed to reduce the number of bits required to code
a given input audio signal. The perceptual models are based on the fact that a considerable
portion of an acoustic signal provided to the human ear is discarded - masked - due
to the characteristics of the human hearing process. For example, if a loud sound
is presented to the human ear along with a softer sound, the ear will likely hear
only the louder sound. Whether the human ear will hear both, the loud and soft sound,
depends on the frequency and intensity of each of the signals. As a result, audio
coding techniques are able to effectively ignore the softer sound and not assign any
bits to its transmission and reproduction under the assumption that a human listener
is not capable of hearing the softer sound even if it is faithfully transmitted and
reproduced. Therefore, psychoacoustic models for calculating a masking threshold play
an essential role in state of the art audio coding. An audio component whose energy
is less than the masking threshold is not perceptible and is, therefore, removed by
the encoder. For the audible components, the masking threshold determines the acceptable
level of quantization noise during the coding process.
[0003] However, it is a well-known fact that the psychoacoustic models for calculating a
masking threshold in state of the art audio coders are based on simple models of the
human auditory system resulting in unacceptable levels of quantization noise or reduced
compression. Hence, it is desirable to improve the state of the art audio coding by
employing better - more realistic - psychoacoustic models for calculating a masking
threshold.
[0004] Furthermore, the MPEG-1 Layer 2 audio encoder is widely used in Digital Audio Broadcasting
(DAB) and digital receivers based on this standard have been massively manufactured
making it impossible to change the decoder in order to improve sound quality. Therefore,
enhancing the psychoacoustic model is an option for improving sound quality without
requiring a new standard.
Summary of the Invention
[0005] It is, therefore, an object of the present invention to provide a method for encoding
an audio signal employing an improved psychoacoustic model for calculating a masking
threshold.
[0006] It is further an object of the present invention to provide an improved psychoacoustic
model incorporating non-linear perception of natural characteristics of an audio signal
by a human auditory system.
[0007] In accordance with the present invention there is provided, a method for encoding
an audio signal comprising the steps of:
receiving the audio signal;
determining a masking index in dependence upon the received audio signal;
determining a masking threshold in dependence upon the masking index using a psychoacoustic
model; and,
encoding the audio signal in dependence upon the masking threshold;
the improvement characterized in that the masking index is determined in dependence
upon characteristics of the audio signal other than intensity or tonality such that
a human perceptible sound quality of the audio signal is retained.
[0008] In accordance with a first aspect of the present invention there is provided, a method
for encoding an audio signal comprising the steps of:
receiving the audio signal;
providing a model relating to temporal masking of sound provided to a human ear;
determining a temporal masking index in dependence upon the received audio signal
and the model;
determining a masking threshold in dependence upon the temporal masking index using
a psychoacoustic model; and,
encoding the audio signal in dependence upon the masking threshold.
[0009] In accordance with a second aspect of the present invention there is provided, a
method for encoding an audio signal comprising the steps of:
receiving the audio signal;
decomposing the audio signal using a plurality of bandpass auditory filters, each
of the filters producing an output signal;
determining an envelope of each output signal using a Hilbert transform;
determining a pitch value of each envelope using autocorrelation;
determining an average pitch error for each pitch value by comparing the pitch value
with the other pitch values;
calculating a pitch variance of the average pitch errors;
determining an inharmonicity index as a function of the pitch variance;
determining a masking threshold in dependence upon the inharmonicity index using a
psychoacoustic model; and,
encoding the audio signal in dependence upon the masking threshold.
[0010] In accordance with the present invention there is further provided, a method for
encoding an audio signal comprising the steps of:
receiving the audio signal;
determining a non-linear masking index in dependence upon human perception of natural
characteristics of the audio signal;
determining a masking threshold in dependence upon the non-linear masking index using
a psychoacoustic model; and,
encoding the audio signal in dependence upon the masking threshold.
[0011] In accordance with the present invention there is yet further provided, a method
for encoding an audio signal comprising the steps of:
receiving the audio signal;
determining a masking index dependence upon human perception of natural characteristics
of the audio signal by considering at least a wideband frequency spectrum of the audio
signal;
determining a masking threshold in dependence upon the masking index using a psychoacoustic
model; and,
encoding the audio signal in dependence upon the masking threshold.
Brief Description of the Drawings
[0012] Exemplary embodiments of the invention will now be described in conjunction with
the drawings in which:
[0013] Fig. 1 is a simplified flow diagram of a first embodiment of a method for encoding
an audio signal according to the present invention;
[0014] Fig. 2 is a diagram illustrating reduction in SMR due to temporal masking;
[0015] Figs. 3a and 3b are diagrams illustrating an example of a harmonic and an inharmonic
signal, respectively;
[0016] Fig. 4 is a simplified flow diagram illustrating a process for determining inharmonicity
of an audio signal according to the invention;
[0017] Figs. 5a and 5b are diagrams illustrating the outputs of a gammatone filterbank for
a harmonic and an inharmonic signal, respectively;
[0018] Figs. 6a and 6b are diagrams illustrating the envelope autocorrelation for a harmonic
and an inharmonic signal, respectively; and,
[0019] Fig. 7 is a simplified flow diagram of a second embodiment of a method for encoding
an audio signal according to the present invention.
Detailed Description of the Invention
[0020] Most psychoacoustic models are based on the auditory "simultaneous masking" phenomenon
where a louder sound renders a weaker sound occurring at a same time instance inaudible.
Another less prominent masking effect is "temporal masking". Temporal masking occurs
when a masker - louder sound - and a maskee - weaker sound - are presented to the
hearing system at different time instances. Detailed information about the temporal
masking is disclosed in the following references:
B. Moore, "An Introduction to the Psychology of Hearing", Academic Press, 1997;
E. Zwicker, and T. Zwicker, "Audio Engineering and Psychoacoustics, Matching Signals
to the Final Receiver, the Human Auditory System", J. Audio Eng. Soc., Vol. 39, No.
3, pp 115-126, Mar. 1991; and,
E. Zwicker and H. Fastl, "Psychoacoustics Facts and Models", Springer Verlag, Berlin,
1990.
[0021] The temporal masking characteristic of the human hearing system is asymmetric, i.e.
"backward masking" is effective approximately 5 msec before occurrence of a masker,
whereas "forward masking" lasts up to 200 msec after the end of the masker. Different
phenomena contributing to temporal auditory masking effects include temporal overlap
of basilar membrane responses to different stimuli, short term neural fatigue at higher
neural levels and persistence of the neural activity caused by a masker, disclosed
in
B. Moore, "An Introduction to the Psychology of Hearing", Academic Press, 1997; and
A. Harma, "Psychoacoustic Temporal Masking Effects with Artificial and Real Signals",
Hearing Seminar, Espoo, Finland, pp. 665-668, 1999.
[0022] Since psychoacoustic models are used for adaptive bit allocation, the accuracy of
those models greatly affects the quality of encoded audio signals. Since digital receivers
have been massively manufactured and are now readily available, it is not desirable
to change the decoder requirements by introducing a new standard. However, enhancing
the psychoacoustic model employed within the encoders allows for improved sound quality
of an encoded audio signal without modifying the decoder hardware. Incorporating non-linear
masking effects such as temporal masking and inharmonicity into the MPEG-1 psychoacoustic
model 2 significantly reduces the bit rate for transparent coding or equivalently,
improves the sound quality of an encoded audio signal at a same bit rate.
[0023] In a first embodiment of a method for encoding an audio signal according to the invention
a temporal masking index is determined in a non-linear fashion in time domain and
implemented into a psychoacoustic model for calculating a masking threshold. In particular,
a combined masking threshold considering temporal and simultaneous masking is calculated
using the MPEG-1 psychoacoustic model 2. Listening tests have been performed with
MPEG-1 Layer 2 audio encoder using the combined masking threshold. In the following
it will become apparent to those of skill in the art that the method for encoding
an audio signal according to the invention has been implemented into the MPEG-1 psychoacoustic
model 2 in order to use a standard state of the art implementation but is not limited
thereto.
[0024] Since the temporal masking method according to the invention is implemented in the
MPEG-1 Layer 2 encoder, the relation between some of the encoder parameters and the
temporal masking method will be discussed in the following. In the MPEG-1 psychoacoustic
model 32 Signal-to-Mask-Ratios (SMR) corresponding to 32 subbands are calculated for
each block of 1152 input audio samples. Since the time-to-frequency mapping in the
encoder is critically sampled, the filterbank produces a matrix - frame - of 1152
subband samples, i.e. 36 subband samples in each of the 32 subbands. Accordingly,
the temporal masking method according to the invention as implemented in the MPEG-1
psychoacoustic model acquires 72 subband samples - 36 samples belonging to a current
frame and 36 samples belonging to a previous frame - in each subband and provides
32 temporal masking thresholds.
[0025] Referring to Fig. 1 a simplified flow diagram of the first embodiment of a method
for encoding an audio signal is shown. The temporal masking method has been implemented
using the following model suggested by
W. Jesteadt, S. Bacon, and J. Lehman, "Forward masking as a function of frequency,
masker level, and signal delay", J. Acoust. Soc. Am., Vol. 71, No. 4, pp. 950-962,
April 1982:

where
M is the amount of masking in dB,
t is the time distance between the masker and the maskee in msec,
Lm is the masker level in dB, and
a,
b, and
c are parameters found from psychoacoustic data.
[0026] For determining the parameters in the above model the fact that forward temporal
masking lasts for up to 200 msec whereas backward temporal masking decays in less
than 5 msec has been considered. Furthermore, temporal masking at any time index is
taken into account if the masker level is greater than 20 dB. Considering the above
mentioned assumptions and based on listening tests of numerous audio materials the
following forward and backward temporal masking functions have been determined, respectively.
For forward masking

where
j =
i + 1,...,36 is the subband sample index, τ is the time distance between successive
subband samples - in msec, and
Lf(
i) is the forward masker level in dB. For backward masking

where
j = 1,...,
i-1 is the subband sample index, τ is the time distance between successive subband
samples - in msec, and
Lb(
i) is the backward masker level in dB. For the backward temporal masking function the
time axis is reversed.
[0027] The time distance τ between successive subband samples is a function of the sampling
frequency. Since the filterbank in the MPEG audio encoder is critically sampled -
box 10 - one subband sample in each subband is produced for 32 input time samples.
Therefore, the time distance τ between successive subband samples is 32/
fs msec, where
fs is the sampling frequency in kHz.
[0028] The masker level in forward masking at time index
i is given by

where
s(
k) denotes the subband sample at time index
k - box 12. At any time index
i the masker level is calculated as the average energy of the 36 subband samples in
the corresponding subband in the previous frame and the subband samples in the current
frame up to time index
i.
[0029] Similarly, the masker level in backward masking - box 14 - at time index
i is given by

The above equation gives the backward masker level at any time as the average energy
of the current and future subband samples.
[0030] The forward temporal masking level at time index
j is then calculated - box 16 - as follows,

[0031] Similarly, the backward temporal masking level at time index
j is then calculated - box 18 - as,

[0032] The total temporal masking energy at time index
j is the sum of the two components - box 20,

where
Mf and
Mb are the forward and the backward temporal masking level in dB at time index
j, respectively.
[0033] The SMR at each subband sample is then calculated - box 22 - as,

where
s(
j) is the
j -th subband sample.
[0034] Since in the MPEG audio encoder all the subband samples in each frame are quantized
with the same number of bits, the maximum value of the 36 SMRs in each subband is
taken to determine the required precision in the quantization process - box 24,

where SMR
(n) is the required Signal-to-Mask-Ratio in subband
n.
[0035] A combined masking threshold is then calculated considering the effect of both temporal
and simultaneous masking. First the SMRs due to temporal masking are translated into
allowable noise levels within a frequency domain. In order to achieve a same SMR in
each subband in the frequency domain, the noise level in a corresponding subband in
the frequency domain is calculated - box 26 - as,

where

is the allowable noise level due to temporal masking - temporal masking index - in
subband
n in the frequency domain, and

is the energy of the DFT components in subband
n in the frequency domain. Alternatively, Parseval's theorem is used to calculate the
equivalent noise level in the frequency domain.
[0036] In the following step, the noise levels due to temporal and simultaneous masking
are combined - box 28. One possibility is to linearly sum the masking energies. However,
according to psychoacoustic experiments the linear combination results in an under-estimation
of the net masking threshold. Instead, a "power law" method is used for combining
the noise levels,

where
NTM and
NSM are the allowable noise due to temporal and simultaneous masking, respectively, and
Nnet is the net masking energy. For the parameter
p, a value of 0.4 has been found to provide an accurate combined masking threshold.
[0037] The net masking energy is used in the MPEG-1 psychoacoustic model 2 to calculate
the corresponding SMR - masking threshold - in each subband - box 30,

[0038] Finally, the acoustic signal is encoded using the masking threshold determined above
- box 32.
[0039] Figure 2 shows an amount of reduction in SMR due to temporal masking in a frame of
1152 subband samples - 36 samples in each of 32 subbands.
[0040] Numerous audio materials have been encoded and decoded with the MPEG-1 Layer 2 audio
encoder using psychoacoustic model 2 based on simultaneous masking and the method
for encoding an audio signal according to the invention based on the improved psychoacoustic
model including temporal masking. Bit allocation has been varied adaptively to lower
the quantization noise below the masking threshold in each frame. Use of the combined
masking model resulted in a bit-rate reduction of 5-12%.
Table 1
Audio Material |
Average Bit Rate Without TM |
Average Bit Rate With TM |
Susan Vega |
153.8 |
138.1 |
Tracy Chapman |
167.2 |
157.7 |
Sax+Double Bass |
191.2 |
177.4 |
Castanets |
150.2 |
132.0 |
Male Speech |
120.1 |
112.4 |
Electric Bass |
145.6 |
129.9 |
[0041] Table 1 shows the average bit rate for a few test files coded with a MPEG-1 Layer
2 encoder using the standard psychoacoustic model 2 and using the modified psychoacoustic
model. The test files were 2-channel stereo audio signals sampled at 48 kHz with 16-bit
resolution.
[0042] In order to compare the subjective quality of the compressed audio materials semiformal
listening tests involving six subjects have been conducted. The listening tests showed
that using the method for encoding an audio signal according to the invention the
subjective high quality of the decoded compressed sounds has been maintained while
the bit rate was reduced by approximately 10%.
[0043] Since psychoacoustic models are used for adaptive bit allocation, the accuracy of
those models greatly affects the quality of encoded audio signals. For instance, the
MPEG-1 Layer 2 audio encoder is used in Digital Audio Broadcasting (DAB) in Europe
and in Canada. Since digital receivers have been massively manufactured and are now
readily available, it is not possible to change the decoder without introducing a
new standard. However, enhancing the psychoacoustic model allows improving the sound
quality of an encoded audio signal without modifying the decoder. Incorporating temporal
masking into the MPEG-1 psychoacoustic model 2 significantly reduces the bit rate
for transparent coding or equivalently, improves the sound quality of an encoded audio
signal at a same bit rate.
[0044] W.C. Treumiet, and D.R. Boucher have shown in "A masking level difference due to harmonicity",
J. Acoust. Soc. Am., 109(1), pp. 306-320, 2001, that the harmonic structure of a complex - multi-tonal - masker has an impact on
the masking pattern. It has been found that if the partials in a multi-tonal signal
are not harmonically related the resulting masking threshold increases by up to 10
dB. The amount of the increase depends on the frequency of the maskee and the frequency
separation between the partials and the level of masker inharmonicity. For example,
it has been found that for two different multi-tonal maskers having the same power,
the one with a harmonic structure produces a lower masking threshold. This finding
has been incorporated into a second embodiment of an audio encoder comprising a modified
MPEG-1 psychoacoustic model 2.
[0045] A sound is harmonic if its energy is concentrated in equally spaced frequency bins,
i.e. harmonic partials. The distance between successive harmonic partials is known
as the fundamental frequency whose inverse is called pitch. Many natural sounds such
as harpsichord or clarinet consist of partials that are harmonically related. Contrary
to harmonic sounds, inharmonic signals consist of individual sinusoids, which are
not equally separated in the frequency domain.
[0046] A model developed to measure inharmonicity recognizes that an auditory filter output
envelope is modulated when the filter passes two or more sinusoids as shown in Appendix
A. since a harmonic masker has constant frequency differences between its adjacent
partials, most auditory filters will have the same dominant modulation rate. On the
other hand, for an inharmonic masker, the envelope modulation rate varies across auditory
filters because the frequency differences are not constant.
[0047] When the signal is a complex masker comprising a plurality of partials, interaction
of neighboring partials causes local variations of the basilar membrane vibration
pattern. The output signal from an auditory filter centered at the corresponding frequency
has an amplitude modulation corresponding to that location. To a first approximation,
the modulation rate of a given filter is the difference between the adjacent frequencies
processed by that filter. Therefore, the dominant output modulation rate is constant
across filters for a harmonic signal because this frequency difference is constant.
However, for inharmonic maskers, the modulation rate varies across filters. Consequently,
in the case of a harmonic masker the modulation rate for each filter output signal
is the fundamental frequency. When inharmonicity is introduced by perturbing the frequencies
of the partials, a variation of the modulation rate across filters is noticeable.
The variation increases with increasing inharmonicity. In general, the harmonicity
nature of a complex masker is characterized by the variance calculated from the envelope
modulation rates across a plurality of auditory filters.
[0048] Since a harmonic signal is characterized by particular relationships among sharp
peaks in the spectrum, an appropriate starting point for measuring the effect of harmonicity
is a masker having a similar distribution of energy across filters, but with small
perturbations in the relationships among the spectral peaks. Fig. 3a shows an example
of a harmonic signal comprising a fundamental frequency of 88 Hz, and a total of 45
equally spaced partials covering a range from 88 Hz to 3960 Hz. Fig. 3b shows an inharmonic
signal generated by slightly perturbing the frequencies and randomizing the phases
of the harmonic signal partials.
[0049] A process for estimating the harmonicity is illustrated in the flow chart of Fig.
4. The signal is analyzed using a "gammatone" filterbank based on the concept of critical
bands disclosed in
E. Zwicker, and E. Terhardt, "Analytical expressions for critical-band rate and critical
bandwidth as a function of frequency", J. Acoust. Soc. Am., 68(5), pp. 1523-1525,
1980. The output of each filter is processed with a Hilbert transform to extract the envelope.
An autocorrelation is then applied to the envelope to estimate its period. Finally,
the harmonicity measure is related to the variance of the modulation rates, i.e. envelope
periods. This variance is negligible for a harmonic masker. However, for an inharmonic
masker the variance is expected to be very large since the modulation rates vary across
filters. For example, the two signals shown in Figs. 3a and 3b have been analyzed
to verify the process. Figs. 5a, 5b, 6a, and 6b illustrate the output signals of the
gammatone filterbank - channels 7-12 - and the corresponding autocorrelation functions
for the harmonic - Figs. 5a and 6a - and inharmonic inputs- Figs. 5b and 6b. As shown
in Figs. 6a and 6b, there is a notable difference between the autocorrelation functions.
In the case of the harmonic signal all the peaks related to the dominant modulation
rate are coincident. Consequently, the variance of the modulation rates is negligible.
On the other hand, for the inharmonic signal, the peaks are not coincident. Therefore,
the variance is much larger. A harmonicity estimation model based on the variability
of envelope modulation rates differentiates harmonic from inharmonic maskers. The
variance of the modulation rate measures the degree to which an audio signal departs
from harmonicity, i.e. a near zero value implies a harmonic signal while a large value
- a few hundreds - corresponds to a noise-like signal.
[0050] In the MPEG-1 Layer 2 psychoacoustic model 2, in order to achieve transparent coding,
the minimum SMRs are computed for 32 subbands as follows. A block of 1056 input samples
is taken from the input signal. The first 1024 samples are windowed using a Hanning
window and transformed into the frequency domain using a 1024-point FFT. The tonality
of each spectral line is determined by predicting its magnitude and phase from the
two corresponding values in the previous transforms. The difference of each DFT coefficient
and its predicted value is used to calculate the unpredictability measure. The unpredictability
measure is converted to the "tonality" factor using an empirical factor with a larger
value indicating a tonal signal. The required SNR for transparent coding is computed
from the tonality using the following empirical formula

where
tj is the tonality factor,
TMNj and
NMTj are the value for tone-masking-noise and noise-masking-tone in subband
j, respectively.
NMTj is set to 5.5 dB and
TMNj is given in a table provided in the MPEG audio standard. In order to take into account
stereo unmasking effects
SNRj is determined to be larger than the minimum SNR
minvalj given in the standard. The SMR is calculated for each of the 32 subbands from the
corresponding SNR. The above process is repeated for the next block of 1056 time samples
- 480 old and 576 new samples - and another set of 32 SMR values is computed. The
two sets of SMR values are compared and the larger value for each subband is taken
as the required SMR.
[0051] Since the masking threshold due to a tonal and a noise-like signal is different,
a tonality factor is calculated for each spectral line. The tonality factor is based
on the unpredictability of the spectral components, meaning that higher unpredictability
indicates a more noise-like signal. However, this measure does not distinguish between
harmonic and inharmonic input signals as it is possible that they are equally predictable.
In the second embodiment of a method for encoding an audio signal, the MPEG-1 psychoacoustic
model 2 has been modified considering imperfect harmonic structures of complex tonal
sounds. It will become apparent to those skilled in the art that the method considering
imperfect harmonic structures is not limited to the implementation in the MPEG-1 psychoacoustic
model 2 but is also implementable into other psychoacoustic models. The example shown
hereinbelow has been chosen because the MPEG-1 Layer 2 encoding is a widely used state
of the art standard encoding process. The inharmonicity of an audio signal raises
the masking threshold and, therefore, incorporating this effect into the encoding
process of inharmonic input signals substantially reduces the bit rate.
[0052] In the MPEG-1 psychoacoustic model 2 the TMN parameter is given in a table. The values
for the TMNs are based on psychoacoustic experiments in which a pure tone is used
to mask a narrowband noise. In these experiments the masker is periodic, which is
the case with an inharmonic masker. In fact, a noise probe is detected at a lower
level when the masker is harmonic. This is likely caused by a disruption of the pitch
sensation due to the periodic structure of the masker's temporal envelope, as taught
in
W.C. Treumiet, and D.R. Boucher, "A masking level difference due to harmonicity",
J. Acoust. Soc. Am., 109(1), pp. 306-320, 2001. In the second embodiment of a method for encoding an audio signal, the TMN parameter
is modified in dependence upon the input signal inharmonicity, as shown in the flow
diagram of Fig. 7. Since in the MPEG-1 Layer 2 psychoacoustic model 2 a set of 32
SMRs is calculated for each 1152 time samples, the same time samples are analyzed
for measuring the level of input signal inharmonicity. After determining the input
signal inharmonicity, an inharmonicity index is calculated and subtracted from the
TMN values. The inharmonicity index as a function of the periodic structure of the
input signal is calculated as follows. The input block of 1632 time samples is decomposed
using a gammatone filterbank - box 100. The envelope of each bandpass auditory filter
output is detected using the Hilbert transform - box 102. The pitch of each envelope
is calculated based on the autocorrelation of the envelope - box 104. Each pitch value
is then compared with the other pitch values and an average error is determined and
the variance of the average errors is calculated - box 106. According to W.C. Treumiet,
and D.R. Boucher inharmonicity causes an increase of up to 10 dB in the masking threshold.
Therefore, the inharmonicity index
δih as a function of the pitch variance
Vp has been defined by the inventors to cover a range of 10 dB - box 108,

The above equation produces a zero value for a perfect harmonic signal and up to
10 dB for noise-like input signals. The new inharmonicity index is incorporated into
the MPEG-1 psychoacoustic model 2 for calculating the masking threshold as

and the acoustic signal is encoded using the masking threshold determined above -
box 110.
[0053] As shown above, the level of inharmonicity is defined as the variance of the periods
of the envelopes of auditory filters outputs. The period of each envelope is found
using the autocorrelation function. The location of the second peak of the autocorrelation
function - ignoring the largest peak at the origin - determines the period. Since
the autocorrelation function of a periodic signal has a plurality of peaks, the second
largest peak sometimes does not correspond to the correct period. In order to overcome
this problem in calculating the difference between two periods the smaller period
is compared to a submultiple of the larger period if the difference becomes smaller.
A MATLAB script for calculating the pitch variance is presented in Appendix B. Another
problem occurs when there is no peak in the autocorrelation function. This situation
implies an aperiodic envelope. In this case the period is set to an arbitrary or random
value.
[0054] As shown in Appendix A, if at least two harmonics pass through an auditory filter
the envelope of the output signal is periodic. Therefore, in order to correctly analyze
an audio signal the lowest frequency of the gammatone filterbank is chosen such that
the auditory filter centered at this frequency passes at least two harmonics. Therefore,
the corresponding critical bandwidth centered at this frequency is chosen to be greater
than twice the fundamental frequency of the input signal. The fundamental frequency
is determined by analyzing the input signal either in the time domain or the frequency
domain. However, in order to avoid extra computation for determining the fundamental
frequency the median of the calculated pitch values is assumed to be the period of
the input signal. The fundamental frequency of the input signal is then simply the
inverse of the pitch value. Therefore, the lower bound for the analysis frequency
range is set to twice the inverse of the pitch value.
[0055] In order to compare the subjective quality of the compressed audio materials informal
listening tests have been conducted. Several audio files have been encoded and decoded
using the standard MPEG-1 psychoacoustic model 2 and the modified version according
to the invention. The bit allocation has been varied adaptively on a frame by frame
basis. When the inharmonicity model was included the bit rate was reduced without
adverse effects on the sound quality. The informal listening tests have shown that
for multi-tonal audio-material the required bit rate decreases by approximately 10%.
[0056] As disclosed above a single value has been used to adjust the masking threshold for
the entire frequency range of the input signal based on the complete frequency spectrum
of the input signal. Alternatively, the masking threshold is modified based on the
local harmonic structure of the input signal based on a local wideband frequency spectrum
of the input signal.
[0057] Optionally, a combination of both non-linear masking effects indicated by the temporal
masking index and the inharmonicity index are implemented into the MPEG-1 psychoacoustic
model 2.
[0058] Of course, numerous other embodiments of the invention will be apparent to persons
skilled in the art without departing from the spirit and scope of the invention as
defined in the appended claims.

1. A method for determining temporal masking thresholds comprising the steps of:
receiving digital data indicative of samples of an analog audio signal (10);
partitioning the digital data into overlapping blocks, each block comprising a predetermined
number of samples (10);
transforming the overlapping blocks into frequency domain using a filterbank, each
transformed overlapping block comprising digital data indicative of a predetermined
number of frequency subbands (10); and,
for each frequency subband determining a temporal masking threshold in dependence
upon the transformed overlapping blocks (12, 14).
2. A method for determining temporal masking thresholds as defined in claim 1 comprising
providing the temporal masking threshold for each frequency subband for incorporation
into a psychoacoustic model (32).
3. A method for determining temporal masking thresholds as defined in any one of claims
1 and 2, characterized in that the psychoacoustic model is the MPEG-1 psychoacoustic model 2 (32).
4. A method for determining temporal masking thresholds as defined in any one of claims
1 to 3 comprising combining for each frequency subband the temporal masking threshold
with a simultaneous masking threshold (28).
5. A method for determining temporal masking thresholds as defined in any one of claims
1 to 4 comprising determining a forward temporal masker level for each current subband
sample using past subband samples (16).
6. A method for determining temporal masking thresholds as defined in claim 5 characterized in that the forward temporal masker level is determined using a forward temporal masking
function (16).
7. A method for determining temporal masking thresholds as defined in claim 5 or 6 comprising
determining a backward temporal masker level for each current subband sample using
future subband samples (18).
8. A method for determining temporal masking thresholds as defined in claim 7 characterized in that the backward temporal masker level is determined using a backward temporal masking
function (18).
9. A method for determining temporal masking thresholds as defined in claim 8 comprising
determining a total temporal masking level for each subband sample using the forward
temporal masker level and the backward temporal masker level (20).
10. A method for determining temporal masking thresholds as defined in claim 9 comprising
determining a signal-to-mask-ratio for each subband sample using the total temporal
masking level of the corresponding subband sample (22).
11. A method for determining temporal masking thresholds as defined in claim 10 comprising
determining a required signal-to-mask-ratio for each frequency subband, the required
signal-to-mask-ratio being a maximum of the signal-to-mask-ratios of the subband samples
of the corresponding frequency subband (24).
12. A method for determining temporal masking thresholds as defined in claim 11 comprising
determining for each frequency subband an allowable noise level due to temporal masking
using the required signal-to-mask-ratio for the corresponding frequency subband (26).
13. A method for determining temporal masking thresholds as defined in claim 12 comprising
determining for each frequency subband a combined allowable noise level using the
allowable noise level due to temporal masking and an allowable noise level due to
simultaneous masking for the corresponding frequency subband (28).
14. A method for determining temporal masking thresholds as defined in claim 13 comprising
determining for each frequency subband a combined signal-to-mask-ratio using the combined
allowable noise level for the corresponding frequency subband (30).