FIELD OF THE INVENTION
[0001] The invention relates to generating an audio signal and/or an audio data signal,
and in particular, but not exclusively, to generating such signals to support e.g.,
an eXtended Reality application.
BACKGROUND OF THE INVENTION
[0002] The variety and range of experiences based on audiovisual content have increased
substantially in recent years with new services and ways of utilizing and consuming
such content continuously being developed and introduced. In particular, many spatial
and interactive services, applications and experiences are being developed to give
users a more involved and immersive experience.
[0003] Examples of such applications are Virtual Reality (VR), Augmented Reality (AR), and
Mixed Reality (MR) applications (commonly referred to as eXtended Reality XR applications)
which are rapidly becoming mainstream, with a number of solutions being aimed at the
consumer market. A number of standards are also under development by a number of standardization
bodies. Such standardization activities are actively developing standards for the
various aspects of VR/AR/MR/XR systems including e.g., streaming, broadcasting, rendering,
etc.
[0004] VR applications tend to provide user experiences corresponding to the user being
in a different world/ environment/ scene whereas AR (including Mixed Reality MR) applications
tend to provide user experiences corresponding to the user being in the current environment
but with additional information or virtual objects or information being added. Thus,
VR applications tend to provide a fully immersive synthetically generated world/ scene
whereas AR applications tend to provide a partially synthetic world/ scene which is
overlaid the real scene in which the user is physically present. However, the terms
are often used interchangeably and have a high degree of overlap. In the following,
the term Virtual Reality/ VR will be used to denote both Virtual Reality and Augmented
Reality.
[0005] VR applications typically provide a virtual reality experience to a user allowing
the user to (relatively) freely move about in a virtual environment and dynamically
change his position and where he is looking. Typically, such virtual reality applications
are based on a three-dimensional model of the scene with the model being dynamically
evaluated to provide the specific requested view. This approach is well known from
e.g., game applications, such as in the category of first person shooters, for computers
and consoles.
[0006] In addition to the visual rendering, most XR applications further provide a corresponding
audio experience. In many applications, the audio preferably provides a spatial audio
experience where audio sources are perceived to arrive from positions that correspond
to the positions of the corresponding objects in the visual scene (including both
objects that are currently visible and objects that are not currently visible (e.g.,
behind the user)). Thus, the audio and video scenes are preferably perceived to be
consistent and with both providing a full spatial experience.
[0007] For audio, headphone reproduction using binaural audio rendering technology is widely
used. In many scenarios, headphone reproduction enables a highly immersive, personalized
experience to the user. Using headtracking, the rendering can be made responsive to
the user's head movements, which highly increases the sense of immersion.
[0008] However, whereas such applications may provide suitable user experiences in many
embodiments they tend to not provide optimal user experiences in all situations. In
particular, in many situations, a suboptimum audio quality may be provided, and typically
a distortion may result for the rendered audio compared to the original or desired
content side audio. In particular, in many applications, the approach may not accurately
reflect or compensate for variations in signal processing or audio reproduction at
the source or render side.
[0009] Hence, an improved approach for distribution and/or rendering and/or processing of
audio signals, in particular for a Virtual/ Augmented/ Mixed/ eXtended Reality experience/
application, would be advantageous. In particular, an approach that allows improved
operation, increased flexibility, reduced complexity, facilitated implementation,
an improved user experience, improved audio quality, improved adaptation to audio
reproduction functions, facilitated and/or improved adaptation to changes in listener
position/orientation (e.g., a virtual listener position/orientation); an improved
Virtual Reality experience, and/or improved performance and/or operation would be
advantageous.
SUMMARY OF THE INVENTION
[0010] Accordingly, the Invention seeks to preferably mitigate, alleviate or eliminate one
or more of the above mentioned disadvantages singly or in any combination.
[0011] According to an aspect of the invention there is provided an apparatus for generating
an output audio signal, the apparatus comprising: a receiver arranged to receive a
data signal comprising at least a first audio signal and metadata including: a pose
indication for an audio source of the first audio signal, and frequency equalization
data indicative of a reference audio reproduction frequency response for the first
audio signal; a listener pose processor arranged to determine a listening pose; a
binaural transfer function processor arranged to determine a binaural transfer function
dependent on the listening pose and the pose indication for the audio source; a processor
arranged to determine an audio rendering frequency response indicative of an audio
reproduction frequency response of an audio reproduction path for the output audio
signal; an adapter arranged to generate a binaural filter having a frequency response
dependent on a combination of the audio rendering frequency response, the reference
audio reproduction frequency response, and the binaural transfer function; and a renderer
arranged to generate the output audio signal in dependence on the binaural filter.
[0012] The approach may provide an improved output audio signal to be generated. The approach
may in many embodiments and scenarios provide an improved audio quality and may e.g.
reduce frequency distortion or degradations caused by the processing path. The approach
may for example allow a rendering side compensation for distortion caused by the audio
reproduction equipment (e.g. headphones) used by the content side to adapt/ generate
the audio signal at the encoding/creation side. For example, a content creator or
sound engineer may adapt or generate the audio signal with a desired audio quality
and content, and as this is based on the sound heard by the content creator/ sound
engineer the frequency response of the used headphones (or other audio reproduction
means) may impact the generated audio. The approach may allow rendering side compensation
of such effects while at the same time compensating for the local rendering side audio
reproduction means.
[0013] Further, the approach may allow for a very efficient implementation and operation
and may in particular allow low complexity and/or resource usage. The approach may
in particular allow a single filtering of the received audio signal to provide compensation
for both content creator side audio reproduction effects and rendering side audio
reproduction effects, as well as provide effective binaural filtering. Indeed, the
approach may provide such improved effects with substantially no computational resource
increase over a conventional generation of binaural signals. For example, the binaural
filter may only be updated at low rate to reflect user movement (or audio source movement),
and thus the updating may require very little computational resource in comparison
to the resource required for continuous filtering of the audio signal.
[0014] A pose, which also may be referred to as a placement, may be a position and/or orientation.
The listening pose may the pose for which the (spatial) output audio signal is generated.
The output audio signal may be a stereo audio signal.
[0015] The binaural transfer function processor may be arranged to determine a frequency
response of a binaural transfer function dependent on the listening pose and the pose
indication for the audio source. The binaural transfer function processor may be arranged
to determine the binaural transfer function dependent on a difference between the
listening pose and the pose indication for the audio source.
[0016] The audio reproduction path for the output audio signal may comprise or consist in
a function converting the output audio signal from an electrical signal to an audio/sound/acoustic
signal. The audio reproduction path for the output audio signal may comprise or consist
in an audio transducer, such as headphones and/or loudspeakers.
[0017] The audio reproduction frequency response for the output audio signal may comprise
or consist in a frequency response of an audio transducer, such as a pair of headphones
or loudspeaker(s).
[0018] The reference audio reproduction frequency response for the first audio signal may
comprise or consist in a frequency response of an audio transducer, such as a pair
of headphones or loudspeaker(s).
[0019] The adapter may be arranged to generate the binaural filter to have a frequency response
which is the combination (in particular a series coupling) of a filter having a frequency
response matching the audio rendering frequency response, a filter having a frequency
response matching the reference audio reproduction frequency response, and a filter
having a frequency response matching a frequency response of the binaural transfer
function.
[0020] In accordance with an optional feature of the invention, the reference audio reproduction
frequency response is a target frequency response for the first audio signal.
[0021] This may provide improved and/or facilitated operation or performance in many embodiments.
The target frequency response may be indicative of a (desired/target) frequency response
for the entire audio path or e.g., only for the rendering audio path.
[0022] In accordance with an optional feature of the invention, the reference audio reproduction
frequency response is a source frequency response applied in generation of the first
audio signal.
[0023] This may provide improved and/or facilitated operation or performance in many embodiments.
[0024] In accordance with an optional feature of the invention, the frequency equalization
data comprises an indication of an audio reproduction device, and the adapter is arranged
to determine the reference reproduction frequency response as a predetermined frequency
response for the audio reproduction device.
[0025] This may provide improved and/or facilitated operation or performance in many embodiments.
The audio reproduction device may be (or include) a headphone and/or loudspeaker(s).
The predetermined frequency response may for example be a stored frequency response
for the audio reproduction device. The predetermined frequency response may for example
be stored internally or remote from the audio apparatus. For example, the adapter
may be arranged to retrieve the frequency response from a remote server.
[0026] In accordance with an optional feature of the invention, the frequency equalization
data comprises data describing a Finite Impulse Response, FIR, filter having a frequency
response matching the reference audio reproduction frequency response.
[0027] This may provide improved and/or facilitated operation or performance in many embodiments.
It may allow a particularly advantageous representation that may be indicated effectively
and with a relatively low data rate. It may further provide improved audio quality
by allowing an accurate representation.
[0028] In accordance with an optional feature of the invention, the frequency equalization
data comprises data describing a combination of a set of sections, each section being
a First Order Section or a Second Order Section and the combination having a frequency
response matching the reference audio reproduction frequency response.
[0029] This may provide improved and/or facilitated operation or performance in many embodiments.
It may allow a particularly advantageous representation that may be indicated effectively
and with a relatively low data rate. It may further provide improved audio quality
by allowing an accurate representation.
[0030] In accordance with an optional feature of the invention, the frequency equalization
data comprises data describing a set of parallel filters, the set of parallel filters
having a frequency response matching the reference audio reproduction frequency response.
[0031] This may provide improved and/or facilitated operation or performance in many embodiments.
It may allow a particularly advantageous representation that may be indicated effectively
and with a relatively low data rate. It may further provide improved audio quality
by allowing an accurate representation. For example, the approach may facilitate processing
reflecting perceptual significance, such as e.g. with the reference audio reproduction
frequency response reflecting perceptual frequency bands etc.
[0032] In accordance with an optional feature of the invention, the frequency equalization
data comprises frequency response identification data, and the adapter is arranged
to access a remote server to retrieve the reference reproduction frequency response
based on the frequency response identification data.
[0033] This may provide improved and/or facilitated operation or performance in many embodiments.
[0034] In accordance with an optional feature of the invention, the metadata further comprises
a reference playback level; and wherein audio apparatus is arranged to adapt the frequency
response of the binaural filter in dependence on a first play back level for a reproduction
of the output audio signal relative to the reference playback level.
[0035] This may provide improved and/or facilitated operation or performance in many embodiments.
[0036] In accordance with an optional feature of the invention, the metadata includes data
indicating a dependency of the reference audio reproduction frequency response on
a playback level and the adapter is arranged to adapt the reference audio reproduction
frequency response in dependence on the playback level.
[0037] This may provide improved and/or facilitated operation or performance in many embodiments.
[0038] According to an aspect of the invention there is provided an apparatus for generating
an audio data signal, the apparatus comprising: a receiver arranged to receive at
least a first audio signal from a first audio source having a pose in a scene; a reproduction
processor (arranged to determine a reference audio reproduction frequency response
for the first audio signal; a data signal generator arranged to generate the audio
data signal to comprise: audio data for the first audio signal; a pose indication
indicative of the pose of the audio source; and frequency equalization data indicative
of the reference reproduction frequency response.
[0039] According to an aspect of the invention there is provided a method of generating
an output audio signal, the method comprising: receiving a data signal comprising
at least a first audio signal and metadata including: a pose indication for an audio
source of the first audio signal, and frequency equalization data indicative of a
reference audio reproduction frequency response for the first audio signal; determining
a listening pose; determining a binaural transfer function dependent on the listening
pose and a pose indication for the audio source; determining an audio rendering frequency
response indicative of an audio reproduction frequency response of an audio reproduction
path for the output audio signal; generating a binaural filter having a frequency
response dependent on a combination of the audio rendering frequency response, the
reference audio reproduction frequency response, and the binaural transfer function;
and generating the output audio signal in dependence on the binaural filter.
[0040] According to an aspect of the invention there is provided an audio data signal, the
method comprising: receiving at least a first audio signal from a first audio source
having a pose in a scene; determining a reference reproduction frequency response
for the first audio signal; a data signal generator arranged to generate the audio
data signal to comprise: audio data for the first audio signal; a pose indication
indicative of the pose of the audio source; and frequency equalization data indicative
of the reference reproduction frequency response.
[0041] According to an aspect of the invention there is provided an audio data signal comprising
at least a first audio signal and metadata including: a pose indication for an audio
source of the first audio signal; and frequency equalization data indicative of a
reference reproduction frequency response for the first audio signal.
[0042] These and other aspects, features and advantages of the invention will be apparent
from and elucidated with reference to the embodiment(s) described hereinafter.
BRIEF DESCRIPTION OF THE DRAWINGS
[0043] Embodiments of the invention will be described, by way of example only, with reference
to the drawings, in which
FIG. 1 illustrates an example of a client server based Virtual Reality system;
FIG. 2 illustrates an example of elements of an audio rendering apparatus in accordance
with some embodiments of the invention;
FIG. 3 illustrates an example of elements of an audio data signal generating apparatus
in accordance with some embodiments of the invention; and
FIG. 4 illustrates some elements of a possible arrangement of a processor for implementing
elements of an apparatus in accordance with some embodiments of the invention.
DETAILED DESCRIPTION OF SOME EMBODIMENTS OF THE INVENTION
[0044] The following description will focus on eXtended Reality applications where audio
is rendered to following a user position in an audio scene to provide an immersive
user experience. Typically, the audio rendering may be accompanied by a rendering
of images such that a complete audiovisual experience is provided to the user. However,
it will be appreciated that the described approaches may be used in many other applications.
[0045] EXtended Reality (including Virtual Augmented and Mixed Reality) experiences allowing
a user to move around in a virtual or augmented world are becoming increasingly popular
and services are being developed to improve such applications. In many such approaches,
visual and audio data may dynamically be generated to reflect a user's (or viewer's)
current pose.
[0046] In the field, the terms placement and pose are used as a common term for position
and/or orientation / direction. The combination of the position and direction/ orientation
of e.g., an object, a camera, a head, or a view may be referred to as a pose or placement.
Thus, a placement or pose indication may comprise up to six values/ components/ degrees
of freedom with each value/ component typically describing an individual property
of the position/ location or the orientation/ direction of the corresponding object.
Of course, in many situations, a placement or pose may be represented by fewer components,
for example if one or more components is considered fixed or irrelevant (e.g., if
all objects are considered to be at the same height and have a horizontal orientation,
four components may provide a full representation of the pose of an object). In the
following, the term pose is used to refer to a position and/or orientation which may
be represented by one to six values (corresponding to the maximum possible degrees
of freedom).
[0047] Many XR applications are based on a pose having the maximum degrees of freedom, i.e.,
three degrees of freedom of each of the position and the orientation resulting in
a total of six degrees of freedom. A pose may thus be represented by a set or vector
of six values representing the six degrees of freedom and thus a pose vector may provide
a three-dimensional position and/or a three-dimensional direction indication. However,
it will be appreciated that in other embodiments, the pose may be represented by fewer
values.
[0048] A system or entity based on providing the maximum degree of freedom for the viewer
is typically referred to as having 6 Degrees of Freedom (6DoF). Many systems and entities
provide only an orientation or position and these are typically known as having 3
Degrees of Freedom (3DoF).
[0049] Typically, the Virtual Reality application generates a three-dimensional output in
the form of separate view images for the left and the right eyes. These may then be
presented to the user by suitable means, such as typically individual left and right
eye displays of a VR headset. In other embodiments, one or more view images may e.g.,
be presented on an autostereoscopic display, or indeed in some embodiments only a
single two-dimensional image may be generated (e.g., using a conventional two-dimensional
display).
[0050] Similarly, for a given viewer/ user/ listener pose, an audio representation of the
scene may be provided. The audio scene is typically rendered to provide a spatial
experience where audio sources are perceived to originate from desired positions.
As audio sources may be static in the scene, changes in the user pose will result
in a change in the relative position of the audio source with respect to the user's
pose. Accordingly, the spatial perception of the audio source may change to reflect
the new position relative to the user. The audio rendering may accordingly be adapted
depending on the user pose.
[0051] The listener pose input may be determined in different ways in different applications.
In many embodiments, the physical movement of a user may be tracked directly. For
example, a camera surveying a user area may detect and track the user's head (or even
eyes (eye-tracking)). In many embodiments, the user may wear a VR headset which can
be tracked by external and/or internal means. For example, the headset may comprise
accelerometers and gyroscopes providing information on the movement and rotation of
the headset and thus the head. In some examples, the VR headset may transmit signals
or comprise (e.g., visual) identifiers that enable an external sensor to determine
the position and orientation of the VR headset.
[0052] In many systems, the VR/ scene data may be provided from a remote device or server.
For example, a remote server may generate audio data representing an audio scene and
may transmit audio signals corresponding to audio components/ objects/ channels, or
other audio elements corresponding to different audio sources in the audio scene together
with position information indicative of the position of these (which may e.g., dynamically
change for moving objects). The audio signals/elements may include elements associated
with specific positions but may also include elements for more distributed or diffuse
audio sources. For example, audio elements may be provided representing generic (non-localized)
background sound, ambient sound, diffuse reverberation etc.
[0053] The local VR device may then render the audio elements appropriately, and specifically
by applying appropriate binaural processing reflecting the relative position of the
audio sources for the audio components.
[0054] Similarly, a remote device may generate visual/video data representing a visual audio
scene and may transmit visual scene components/ objects/ signals or other visual elements
corresponding to different objects in the visual scene together with position information
indicative of the position of these (which may e.g., dynamically change for moving
objects). The visual items may include elements associated with specific positions
but may also include video items for more distributed sources.
[0055] In some embodiments, the visual items may be provided as individual and separate
items, such as e.g., as descriptions of individual scene objects (e.g., dimensions,
texture, opaqueness, reflectivity etc.). Alternatively or additionally, visual items
may be represented as part of an overall model of the scene e.g. including descriptions
of different objects and their relationship to each other.
[0056] For a VR service, a central server may accordingly in some embodiments generate audiovisual
data representing a three dimensional scene, and may specifically represent the audio
by a number of audio signals representing audio sources in the scene which can then
be rendered by the local client/ device.
[0057] FIG. 1 illustrates an example of a VR/XR system in which a central server 101 liaises
with a number of remote clients 103 e.g., via a network 105, such as e.g., the Internet.
The central server 101 may be arranged to simultaneously support a potentially large
number of remote clients 103.
Such an approach may in many scenarios provide an improved trade-off e.g., between
complexity and resource demands for different devices, communication requirements
etc. For example, the scene data may be transmitted only once or relatively infrequently
with the local rendering device (the remote client 103) receiving a viewer pose and
locally processing the scene data to render audio and/or video to reflect changes
in the viewer pose. This approach may provide for an efficient system and attractive
user experience. It may for example substantially reduce the required communication
bandwidth while providing a low latency real time experience while allowing the scene
data to be centrally stored, generated, and maintained. It may for example be suitable
for applications where a VR experience is provided to a plurality of remote devices.
[0058] FIG. 2 illustrates elements of an apparatus for generating an output audio signal,
henceforth also referred to as an audio rendering apparatus, which may generate an
improved audio signal in many applications and scenarios. In particular, the audio
rendering apparatus may provide improved rendering for many VR applications, and the
audio rendering apparatus may specifically be arranged to perform the audio processing
and rendering for a VR client 103 of FIG. 1. FIG. 3 illustrates an example of an apparatus,
also henceforth referred to as an audio encoding apparatus, for generating an audio
data signal that represents audio for one or more audio sources of a scene. In particular,
the audio encoding apparatus may provide improved representation of an audio source
or scene for many VR applications, and the audio encoding apparatus may specifically
be arranged to perform the audio processing and function for a VR server 101 of FIG.
1
[0059] The audio apparatus of FIG. 2 is arranged to render audio of a three dimensional
scene to provide a three dimensional perception of the scene. The specific description
will focus on audio rendering, but it will be appreciated that in many embodiments
this may be supplemented by a visual rendering of the scene. Specifically view images
may in many embodiments be generated and presented to the user.
[0060] The audio rendering apparatus comprises a first receiver 201 which is arranged to
receive audio items from a local or remote source. In the specific example, the first
receiver 201 receives data describing the audio items from the server 101. The first
receiver 201 may be arranged to receive data describing the virtual scene, and specifically
the audio scene. The data may include data providing a visual description of the scene
and may include data providing an audio description of the scene. Thus, an audio scene
description and a visual scene description may be provided by the received data.
[0061] The audio signals/items may be encoded audio data, such as encoded audio signals.
The audio signals may be different types of audio elements including different types
of elements and components, and indeed in many embodiments the first receiver 201
may receive audio data which defines different types/ formats of audio. For example,
the audio data may include audio represented by audio channel signals, individual
audio objects, scene based audio such as Higher Order Ambisonics (HOA) etc. The audio
may for example be represented as encoded audio for a given audio component that is
to be rendered.
[0062] The received data signal further comprises data indicative of the poses (position
and/or orientation) of the audio sources for which the audio signals are provided.
In many embodiments, at least some of the audio signals are linked to data describing
the position of the audio source in the scene. In some cases, data may alternatively,
or typically additionally, comprise data defining the orientation of the audio sources
(e.g., allowing non-omni directional audio sources).
[0063] The received signal may specifically comprise metadata which includes pose data that
indicates a position and/or orientation of the audio items and specifically the position
and/or orientation of an audio source and/or a visual scene object or element. The
pose data may for example include absolute position and/or orientation data defining
a position of each, or at least some of the audio sources.
[0064] The first receiver 201 is coupled to a renderer 203 which proceeds to render the
audio scene based on the received data describing the audio items. In case of encoded
data, the renderer 203 may also be arranged to decode the audio data (or in some embodiments
decoding may be performed by the first receiver 201).
[0065] The renderer 203 is arranged to render the audio scene by generating audio signals
based on the received audio data for the audio sources. In the example, the renderer
203 is a binaural audio renderer which generates binaural audio signals for a left
and right ear of a user. The binaural audio signals are generated to provide a desired
spatial experience and are typically reproduced by headphones or earphones that specifically
may be part of a headset worn by a user (the headset typically also comprises left
and right eye displays).
[0066] Thus, in many embodiments, the audio rendering by the renderer 203 is a binaural
render process using suitable binaural transfer functions to provide the desired spatial
effect for a user wearing a headphone. For example, the renderer 203 may be arranged
to generate an audio component to be perceived to arrive from a specific position
using binaural processing.
[0067] Binaural processing is known to be used to provide a spatial experience by virtual
positioning of sound sources using individual signals for the listener's ears. With
an appropriate binaural rendering processing, the signals required at the eardrums
in order for the listener to perceive sound from any desired direction can be calculated,
and the signals can be rendered such that they provide the desired effect. These signals
are then recreated at the eardrum using either headphones or a crosstalk cancelation
method (suitable for rendering over closely spaced speakers). Binaural rendering can
be considered to be an approach for generating signals for the ears of a listener
resulting in tricking the human auditory system into perceiving that a sound is coming
from the desired positions.
[0068] The binaural rendering is based on binaural transfer functions which vary from person
to person due to the acoustic properties of the head, ears and reflective surfaces,
such as the shoulders. Binaural transfer functions may therefore be personalized for
an optimal binaural experience. For example, binaural filters can be used to create
a binaural recording simulating multiple sources at various locations. This can be
realized by convolving each sound source with the pair of e.g., Head Related Impulse
Responses (HRIRs) that correspond to the position of the sound source.
[0069] A well-known method to determine binaural transfer functions is binaural recording.
It is a method of recording sound that uses a dedicated microphone arrangement and
is intended for replay using headphones. The recording is made by either placing microphones
in the ear canal of a subject or using a dummy head with built-in microphones, a bust
that includes pinnae (outer ears). The use of such dummy head including pinnae provides
a very similar spatial impression as if the person listening to the recordings was
present during the recording.
[0070] By measuring e.g., the responses from a sound source at a specific location in 2D
or 3D space to microphones placed in or near the human ears, the appropriate binaural
filters can be determined. Based on such measurements, binaural filters reflecting
the acoustic transfer functions to the user's ears can be generated. The binaural
filters can be used to create a binaural recording simulating multiple sources at
various locations. This can be realized e.g., by convolving each sound source with
the pair of measured impulse responses for a desired position of the sound source.
In order to create the illusion that a sound source is moving around the listener,
a large number of binaural filters is typically required with a certain spatial resolution,
e.g., 10 degrees.
[0071] The head related binaural transfer functions may be represented e.g., as Head Related
Impulse Responses (HRIR), or equivalently as Head Related Transfer Functions (HRTFs)
or, Binaural Room Impulse Responses (BRIRs), or Binaural Room Transfer Functions (BRTFs).
The (e.g., estimated or assumed) transfer function from a given position to the listener's
ears (or eardrums) may for example be represented in the frequency domain in which
case it is typically referred to as an HRTF or BRTF, or in the time domain in which
case it is typically referred to as a HRIR or BRIR. In some scenarios, the head related
binaural transfer functions are determined to include aspects or properties of the
acoustic environment and specifically of the room in which the measurements are made,
whereas in other examples only the user characteristics are considered. Examples of
the first type of functions are the BRIRs and BRTFs.
[0072] The renderer 203 may be arranged to individually apply binaural processing to a plurality
of audio signals/ sources and may then combine the results into a single binaural
output audio signal representing the audio scene with a number of audio sources positioned
at appropriate positions in the sound stage.
[0073] The audio rendering apparatus further comprises a listener pose processor 205 which
is arranged to determine a listening pose for which the output audio signal is generated.
The listener pose may accordingly correspond to the position in the scene of the user/listener.
[0074] The listener pose may specifically be determined in response to sensor input, e.g.,
from suitable sensors being part of a headset. It will be appreciated that many suitable
algorithms will be known to the skilled person and for brevity this will not be described
in more detail.
[0075] The renderer 203 of the apparatus of FIG.2 is arranged to generate the output audio
signal by a binaural rendering process such that spatial cues are provided from e.g.,
rendering using headphones. The renderer may filter the audio signal by suitable binaural
filters corresponding to respectively the left and right ear in order to generate
a binaural output signal that when presented to the ears of a user will provide a
spatial audio experience.
[0076] The listener pose processor 205 is coupled to a binaural transfer function processor
207 which receives the listener pose from the listener pose processor 205 and which
is arranged to determine a binaural transfer function for this listener pose and for
a given audio source pose. The binaural transfer function processor 207 is arranged
to determine the appropriate binaural transfer function for a given audio signal and
audio source dependent on the relative position/pose of the audio source relative
to the listener pose.
[0077] The binaural transfer function processor 207 may comprise a store/ service/ database
with binaural transfer functions for a, typically high, number of different positions,
and sometimes orientations, with each binaural transfer function providing information
on how an audio signal should be processed/ filtered in order to be perceived to originate
from that position.
[0078] The binaural transfer function processor 207 may for a given audio signal that is
to be perceived to originate from a given position relative to the user's head, select
and retrieve the stored binaural transfer function that most closely matches the desired
relative position, i.e., that most closely matches the audio source position relative
to the listener position. In some embodiments, the binaural transfer function processor
207 may be arranged to generate the binaural transfer function by interpolating between
a plurality of close stored binaural transfer functions. It may then apply the selected
binaural transfer function to the audio signal of the audio element thereby generating
an audio signal for the left ear and an audio signal for the right ear.
[0079] However, in the apparatus of FIG. 2, the extracted binaural transfer function is
not used directly by the renderer 203 but rather it is adapted prior to the rendering.
The binaural transfer function processor 207 is coupled to an adapter 209 which is
arranged to adapt the binaural transfer function and specifically it is arranged to
generate a binaural filter that has a frequency response which corresponds to the
frequency response of the binaural transfer function but with this being adapted/modified.
The binaural filter is then fed to the renderer 203 where it is used to perform the
binaural rendering. Typically, the binaural filter is a stereo filter comprising both
a transfer function/ impulse response/ frequency response for generating a left ear
signal and for generating a right ear signal (equivalently the adapter 207 may generate
two filters that are used by the renderer 203 to generate respectively the left ear
signal and the right ear signal. The described process may in such a case be applied
for both binaural filters).
[0080] The renderer may specifically generate an output stereo signal by filtering the audio
signal by the binaural filter(s). The generated output stereo signal in the form of
the left and right ear signal is then suitable for headphone rendering and may be
amplified to generate drive signals that are fed to the headset of a user. The user
will then perceive the audio signal/source to originate from the desired position.
[0081] It will be appreciated that the audio signals may in some embodiments also be processed
to e.g., add acoustic environment effects. For example, the audio signal may be processed
to add reverberation or e.g., decorrelation/ diffuseness. In many embodiments, this
processing may be performed on the generated binaural signal rather than directly
on the audio element signal.
[0082] Thus, the renderer 203 may be arranged to generate the output audio (stereo) signal
such that a given audio source is rendered with a user wearing headphones perceiving
the audio to be received from a desired position. The approach is typically applied
to a plurality of audio sources and signals which are then combined by the renderer
203 to generate the output signal.
[0083] It will be appreciated that many algorithms and approaches for rendering of spatial
audio, e.g., using headphones, and specifically for binaural rendering, will be known
to the skilled person and that any suitable approach may be used without detracting
from the invention.
[0084] The adapter 209 may specifically be arranged to compensate/adapt the binaural transfer
functions for properties of audio reproduction at both the rendering and source side.
[0085] The apparatus of FIG. 2 specifically comprises a reproduction processor 211 which
is arranged to determine an audio rendering frequency response which reflects the
audio reproduction frequency response of the audio reproduction path for the output
audio signal. The audio rendering frequency response may specifically reflect a frequency
response of a pair of headphones that are used to render the generated output audio
signal to the user. More generally, the audio rendering frequency response typically
is or includes a frequency response of an audio transducer arranged to convert an
electrical signal to an acoustic signal. The audio transducer may e.g., be a headphone
and/or loudspeaker.
[0086] In the approach, the received data signal further comprises frequency equalization
data that is indicative of a reference audio reproduction frequency response for the
generated audio. The received data may comprise frequency equalization data that for
example indicate a desired frequency spectrum compensation, a target frequency response,
and/or a frequency response of an audio reproduction device that was used at the content
side, e.g., in the generation or post-processing of the audio of the audio source.
[0087] The reproduction processor 211 is coupled to the adapter 209 which is fed the audio
rendering frequency response, and which further is fed the reference audio reproduction
frequency response from the receiver 201. The adapter 209 is arranged to adapt the
binaural transfer function in dependence on both the audio rendering frequency response
and the reference audio reproduction frequency response to generate the binaural filter
which is fed to the renderer 203 and applied to the audio signal to generate the binaural
audio signal.
[0088] The exact adaptation will depend on the specific preferences and requirements of
the individual embodiment, and will depend on e.g., the format and specific properties
of the provided audio rendering frequency response and reference audio reproduction
frequency response.
[0089] In many embodiments, the reference audio reproduction frequency response may be indicative
of a frequency response of an audio reproduction device, such as a headphone, that
was used by a content creator to generate (e.g., equalize) the audio signal. Similarly,
the reference audio reproduction frequency response may characterize the frequency
response of a set of headphones or a loudspeaker that are used to render the output
stereo signal to the user. In such a case, the adapter 209 may be arranged to compensate
the frequency response of the extracted binaural transfer function to compensate for
the frequency responses of the headphones/ loudspeaker. For example, the gain of the
binaural transfer function for a first frequency may be multiplied by a (normalized)
gain of the reference audio reproduction frequency response and divided by a (normalized)
gain of the audio rendering frequency response. This may be repeated for all frequency
values resulting in a frequency response of the binaural filter. This frequency response
thus represents an audio reproduction compensated binaural transfer function/ filter
that, when applied to the audio signal may provide improved audio quality.
[0090] In the approach, the rendering side may accordingly adapt/ compensate for frequency
distortion at both the encoder and decoder side. The compensation and audio quality
may be aided or controlled by the content side thereby allowing a content creator
to affect the rendering properties. Further, the adaptation of both the content creator
side and the rendering side properties are synergistically combined with the binaural
processing. Indeed, the binaural filter pre-processing (rather than e.g., applying
a filter as postprocessing of the output stereo signal), provides a highly efficient
processing and operation which may provide improved audio quality for a low implementation
complexity and resource usage. In particular, in the approach only a single filtering
of the audio signal is required to achieve the beneficial effects.
[0091] The encoder of FIG. 3 comprises an encoder receiver 301 arranged to receive at least
a first audio signal from a first audio source having a given pose in a(n audio) scene.
Typically, the receiver will receive a plurality of audio sources such as audio sources
at different positions and/or with different orientations, and/or audio sources with
no specific position and/or orientation, such as e.g., a source representing the diffuse
reverberation audio in the audio scene.
[0092] The encoder further comprises an encoder reproduction processor 303 which is arranged
to determine a reference audio reproduction frequency response for the first audio
signal. The reference audio reproduction frequency response may be determined in dependence
on a frequency response of an audio reproduction path or device that has been used
at the encoder/content side. For example, the reference audio reproduction frequency
response may be determined on the basis of the frequency response of a pair of headphones
or a loudspeaker that was used by a content creator to generate the audio signal.
In many embodiments, the reference audio reproduction frequency response may be or
include a frequency response of an audio transducer arranged to convert an electrical
signal to an acoustic signal. The audio transducer may e.g., be a headphone and/or
loudspeaker.
[0093] The encoder reproduction processor 303 may for example include a user interface allowing
a user or content creator to manually input data describing a frequency response,
e.g., of a headphone or speaker. As another example, in some embodiments, the encoder
reproduction processor 303 may be arranged to control a test system to measure a frequency
response of a loudspeaker. For example, the encoder reproduction processor 303 may
comprise a variable frequency tone generator that may be coupled to an external audio
reproduction device, e.g., headphone or speaker. A frequency sweep may be performed,
and a microphone positioned appropriately for a given audio reproduction device may
generate a microphone signal capturing the generated audio. The level of the captured
audio for different frequencies may then be used to determine the frequency response
of the reproduction device/headphone. As another example, the reproduction device
may have functionality for identifying e.g., a brand, model, type etc. and the encoder
reproduction processor 303 may be arranged to retrieve a stored nominal frequency
response for that device, e.g., from a remote database. For example, a Bluetooth
™ headphone may provide identification data that allows the encoder reproduction processor
303 to access e.g., a remote database to retrieve the frequency response for the headphone.
[0094] The encoder reproduction processor 303 and the encoder receiver 301 are coupled to
a data signal generator 305 which is arranged to generate an output audio data signal
which may be transmitted or distributed to the decoder of FIG. 1 (as well as possibly
to other decoder devices). The data signal generator 305 may encode the audio signal
to generate encoded audio data that is included in the audio data signal. It may typically
encode other audio sources to provide a combined audio representation of the audio
scene. The data signal generator 305 may further receive pose data for the audio source
of the first audio signal, and typically for other audio sources, and may be arranged
to include pose data for the audio sources in the output audio signal.
[0095] The data signal generator 305 may further receive the frequency equalization data
indicative of the reference reproduction frequency response, such as specifically
the frequency response of a reproduction device used by the content creation side.
The data signal generator 305 may then include data in the output audio data signal
that indicates and describes the reference audio reproduction frequency response.
[0096] In some embodiments, the reference audio reproduction frequency response may be a
target frequency response for the reproduction of the first audio signal at the decoder
side. For example, the reference audio reproduction frequency response may indicate
a desired frequency response for the overall audio path, including the frequency response
of the reproduction path at the rendering device.
[0097] For example, a content creator may manually process and adapt an audio signal to
sound as desired and such a process typically include frequency equalization and filtering,
i.e., the content creator may adapt the frequency spectrum of the audio signal to
result in a desired sound. It may further be desired that the overall frequency response
of the entire signal processing is a flat frequency response such that the rendered
sound corresponds more directly to the audio manually generated by the content creator.
However, the content creator will inherently use an audio reproduction device, such
as headphones or speakers, and this will inherently have a frequency response that
will tend to not be flat. Accordingly, the audio as heard by the content creator may
be colored by the reproduction. In the described approach, the reference audio reproduction
frequency response may be used to compensate for such distortion by indicating a target
reproduction frequency response that will compensate for the frequency response of
the content side reproduction path.
[0098] Specifically, the target reference audio reproduction frequency response may be indicated
as frequency response that compensates, and e.g., has the inverse frequency response
of the determined content side reproduction device frequency response. For example,
when the content side reproduction frequency response has increased attenuation, the
target reference audio reproduction frequency response may be determined to have the
corresponding gain. In particular, the reference audio reproduction frequency response
may be determined as the inverse/ reciprocal of the determined reproduction frequency
response for the content side reproduction device. When receiving such a target reference
audio reproduction frequency response, the rendering apparatus of FIG. 2 may proceed
to perform a frequency equalization that includes applying the corresponding reference
audio reproduction frequency response. The rendering apparatus may further include
a compensation for the audio rendering frequency response. Thus, specifically, the
binaural transfer function/HRTF may be modified to provide a desired target response
while compensating for the local audio reproduction. As a specific example, the frequency
response of a binaural filter used by the renderer 203 may be determined as:

where H
T(f) indicates the received target reference audio reproduction frequency response,
H
R(f) indicates the local audio rendering frequency response, and H
B(f) indicates the determined binaural transfer function (specifically HRTF) for the
position of the audio source relative to the listening position.
[0099] In some embodiments, the reference audio reproduction frequency response is a source
frequency response applied in generation of the first audio signal. Specifically,
the reference audio reproduction frequency response may provide an indication of the
frequency response of e.g., a headphone or a loudspeaker that has been used by e.g.,
a sound engineer or content creator on the content side. In such a case, the adapter
209 may be arranged to adapt the frequency response of the binaural transfer function
to compensate for the filter response. For example, the frequency response of a binaural
filter used by the renderer 203 may be determined using the same approach as described
above (bearing in mind that the impact of the frequency response of the headphone/loudspeaker
on the transmitted signal is the inverse of the frequency response, and thus that
the frequency response of the headphone is also the desired target frequency response
to compensate for the headphone).
[0100] The frequency equalization data may use different approaches for indicating the reference
audio reproduction frequency response.
[0101] In some embodiments, the frequency equalization data may comprise data that describes
a Finite Impulse Response, FIR, filter having a frequency response matching the reference
audio reproduction frequency response. The reference audio reproduction frequency
response may specifically be described as a FIR filter that has a given frequency
response. The FIR filter may be described in the frequency domain and/or in the time
domain. For example, the frequency equalization data may include a set of coefficients
for a FIR filter, and specifically the coefficients may correspond to suitable samples
of an impulse response for a FIR filter having a frequency response that represents/matches
the reference audio reproduction frequency response.
[0102] Such an approach may provide a highly efficient representation that is easy to process
at the renderer side. It may also be generated with relatively low complexity at the
content side. It may further be advantageous in that a FIR filter may effectively
provide equalization of both amplitude and phase, and indeed even with these being
done independently from each other.
[0103] In some embodiments, the frequency equalization data may include data that describes
the reference audio reproduction frequency response as an Infinite Impulse Response,
IIR, filter. This may in some scenarios provide for a more compressed representation
and/or lower complexity implementation.
[0104] In some embodiments, the frequency equalization data may comprise data that describes
a combination of a set of sections with the combination having a frequency response
matching the reference audio reproduction frequency response. Each section may be
a First Order Section, FOS, or a Second Order Section, SOS. The combination may specifically
be a series of the sections of FOS and SOSs. Each section may thus correspond to a
first order or second order filter, and the combination may correspond to a filter
formed by the series of these first and second order filters. The combined frequency
response of the series of filters may be considered by the renderer 203 as the reference
audio reproduction frequency response.
[0105] Such an approach may be particularly advantageous in many embodiments and scenarios
and may in particular allow computationally efficient implementations and functions.
[0106] In some embodiments, the frequency equalization data may include data describing
a set of parallel filters and the reference audio reproduction frequency response
may be given as the combined frequency response of these parallel filters. For example,
a set of parallel band pass filters may be provided with each bandpass filter providing
the frequency response for a given frequency band and with other the pass bands adding
up to the overall frequency response. As a specific example, parallel filters may
for example be provided an octave apart. In some embodiments, parallel filters may
be provided with different bandwidths and an advantage of such an approach is that
the filter set may e.g., be designed on a perceptually motivated logarithmic frequency
scale, which is an advantage over FIR filters that are designed on a linear frequency
scale. For example, the filters and filter bandwidths may be determined to correspond
to e.g., a logarithmic, octave, 1/3rd octave, equivalent rectangular bandwidth (ERB),
Bark scale, etc.
[0107] As a specific example, the audio data signal may comprise frequency equalization
data in the form of an EQ ((frequency EQualization) field specifier that describes
e.g., a type and method of frequency equalization. Such an EQ field may indicate a
method of frequency equalization and the equalization filters that can or may be used.
The frequency equalization data may for example include data indications for one or
more of the following examples of frequency equalization indications:
The reference audio reproduction frequency response represented as the coefficients
of an FIR filter specified at certain sample rate.
[0108] The reference audio reproduction frequency response represented as a series of one
or more of First and Second Order Sections (FOS & SOS) specified either by their coefficients
and sample rate or by their descriptive parameters. For example, there may be provided
a second order peak-EQ filter with a center frequency, a gain and a quality factor,
followed by first order shelving filter with a given gain and center frequency, etc.
[0109] The reference audio reproduction frequency response represented as a sum of parallel
filters.
[0110] It will be appreciated that other approaches may be used depending on the requirements
and preferences of the individual embodiment.
[0111] In some embodiments, the audio rendering apparatus may receive frequency equalization
data that comprises frequency response identification data. This identification data
may not fully describe a frequency response but may provide an identification of a
frequency response, such as specifically a specific name and/or number.
[0112] In some embodiments, the audio rendering apparatus may be arranged to extract a corresponding
reference audio reproduction frequency response from a set of locally stored reference
audio reproduction frequency responses. For example, during manufacturing, a plurality
of reference audio reproduction frequency responses may be stored in local memory
together with an associated identity for each reference audio reproduction frequency
response. During operation, the rendering apparatus may receive an audio data signal
with an identification of a reference audio reproduction frequency response, and it
may proceed to access the memory to attract the reference audio reproduction frequency
response that is stored for the received identification. It may then proceed to use
this reference audio reproduction frequency response to modify the binaural transfer
function.
[0113] In some embodiments, the adapter 209 may be arranged to access a remote server to
retrieve a reference reproduction frequency response that matches the frequency response
identification received in the data audio signal. For example, the adapter 209 may
comprise a network interface that e.g., may allow it to connect to the Internet to
access a remote server with a request that includes the frequency response identity.
The server may extract the corresponding reference audio reproduction frequency response
and transmit it back to the rendering apparatus.
[0114] Such an approach may be highly advantageous in many embodiments as it may allow a
single (or a few) central servers to serve a large number of rendering devices. This
may substantially facilitate the distribution and updating of the reference audio
reproduction frequency responses.
[0115] In many embodiments, the identification may specifically be an identification of
an audio reproduction device, and specifically of an audio reproduction device that
may be used at the content side operation. For example, the identification may identify
headphones, speakers, amplifiers etc. that may be used as part of the content side
setup. The identification may in many examples simply be data describing the make
and model of the device. The adapter 209 may then access a remote database/ server
to retrieve a frequency response for that device and the binaural transfer function
may be modified to compensate for this response.
[0116] As a specific example, if a headphone brand and type is indicated, the adapter 209
may access a (third-party) database containing the preferred equalization filter frequency
response or the headphone frequency response and this may then be used to provide
a desired target frequency response, such as specifically a flat frequency response.
[0117] In some embodiments, the encoder may further be arranged to generate the audio data
signal to include metadata that provides an indication of a reference playback level.
The reference playback level may be a nominal value or may e.g., be dynamically determined
to directly indicate the level of the playback that has been applied, e.g., by a content
creator or sound engineer manually adjusting the frequency spectrum of the first audio
signal. For example, a sound engineer may perform a manual frequency equalization
while listening to the stereo audio signal. The level of the reproduction of the stereo
signal may be measured or determined (e.g., from a volume setting by the sound engineer)
and data may be included in the audio data signal indicating this playback level.
[0118] The playback level may for example be provided as a relative or absolute value. For
example, in some embodiments, a general audio level in a predetermined range (e.g.,
from 0 to 11) may be provided. In other embodiments, the reference playback level
may be provided as an indication of a Sound Pressure Level, SPL.
[0119] The renderer may receive such an audio data signal and extract the reference playback
level from the metadata of the audio data signal and the adapter 209 may be arranged
to adapt the binaural filter in dependence on the reference playback level and a playback
level for the generated output audio signal.
[0120] For example, the current level setting (e.g., a volume setting) for the rendering/
reproduction of the output stereo signal may be provided to the adapter 209 and depending
on this and the reference playback level, the adapter 209 may for example modify the
frequency response adaptation applied to the binaural transfer function. For example,
if the reference playback level is indicative of a low level and the current volume
setting is set to a low reproduction sound level, the adapter 209 not provide any
additional sound level frequency adaptation. However, if the volume setting is currently
set to correspond to a high reproduction sound level, the adapter 209 may proceed
to perform an additional frequency response adjustment to attenuate low and high frequencies
(e.g., to prevent hearing damage). If instead, the reference playback level indicates
a high sound level, the adapter 209 may be arranged to not include any additional
frequency response adaptation for a high volume setting, but may increase the gain
for high and low frequencies for a low volume setting (e.g., to provide a loudness
effect).
[0121] Such an approach may provide an improved listening experience in many embodiments.
[0122] In some embodiments, the metadata may include data indicative of a dependency of
the reference audio reproduction frequency response on a playback level. Specifically,
the metadata may not only provide data describing a reference audio reproduction frequency
response but also include data describing how this may vary for different reproduction
sound levels.
[0123] As a low complexity example, the metadata may include different reference audio reproduction
frequency responses for different sound reproduction levels. For example, the metadata
may include a first reference audio reproduction frequency response for a first (e.g.,
low) reference playback level and a second reference audio reproduction frequency
response for a second (e.g., high) reference playback level. As another example, the
metadata may include a reference audio reproduction frequency response with a gain
for one or more frequencies being provided as a function of a sound/playback level.
[0124] The adapter 209 may then adapt the reference audio reproduction frequency response
that is used for modifying the binaural transfer function in dependence on a current
playback level for the output audio signal. For example, dependent on the current
volume setting (e.g., whether it is a "low range" or a "high range" - e.g., whether
it is below or above a setting of 5), the adapter 209 may select between different
reference audio reproduction frequency responses provided in the metadata. As another
example, in some embodiments, the actual current sound reproduction level may be measured
(e.g., by a microphone) and the adapter 209 may proceed to determine various gains
of the reference audio reproduction frequency response by evaluating the functions
provided by the metadata for the given measured sound reproduction level.
[0125] Such an approach may provide a highly advantageous user experience in many embodiments
and may in particular allow content side control or assistance in adapting render
side systems to provide a desired audio reproduction.
[0126] FIG. 4 is a block diagram illustrating an example processor 400 according to embodiments
of the disclosure. Processor 400 may be used to implement one or more processors implementing
an apparatus as previously described or elements thereof. Processor 400 may be any
suitable processor type including, but not limited to, a microprocessor, a microcontroller,
a Digital Signal Processor (DSP), a Field ProGrammable Array (FPGA) where the FPGA
has been programmed to form a processor, a Graphical Processing Unit (GPU), an Application
Specific Integrated Circuit (ASIC) where the ASIC has been designed to form a processor,
or a combination thereof.
[0127] The processor 400 may include one or more cores 402. The core 402 may include one
or more Arithmetic Logic Units (ALU) 404. In some embodiments, the core 402 may include
a Floating Point Logic Unit (FPLU) 406 and/or a Digital Signal Processing Unit (DSPU)
408 in addition to or instead of the ALU 404.
[0128] The processor 400 may include one or more registers 312 communicatively coupled to
the core 402. The registers 412 may be implemented using dedicated logic gate circuits
(e.g., flip-flops) and/or any memory technology. In some embodiments the registers
412 may be implemented using static memory. The register may provide data, instructions
and addresses to the core 402.
[0129] In some embodiments, processor 400 may include one or more levels of cache memory
410 communicatively coupled to the core 402. The cache memory 410 may provide computer-readable
instructions to the core 402 for execution. The cache memory 410 may provide data
for processing by the core 402. In some embodiments, the computer-readable instructions
may have been provided to the cache memory 410 by a local memory, for example, local
memory attached to the external bus 416. The cache memory 410 may be implemented with
any suitable cache memory type, for example, Metal-Oxide Semiconductor (MOS) memory
such as Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), and/or
any other suitable memory technology.
[0130] The processor 400 may include a controller 414, which may control input to the processor
400 from other processors and/or components included in a system and/or outputs from
the processor 400 to other processors and/or components included in the system. Controller
414 may control the data paths in the ALU 404, FPLU 406 and/or DSPU 408. Controller
414 may be implemented as one or more state machines, data paths and/or dedicated
control logic. The gates of controller 414 may be implemented as standalone gates,
FPGA, ASIC or any other suitable technology.
[0131] The registers 412 and the cache 410 may communicate with controller 414 and core
402 via internal connections 420A, 420B, 420C and 420D. Internal connections may be
implemented as a bus, multiplexer, crossbar switch, and/or any other suitable connection
technology.
[0132] Inputs and outputs for the processor 400 may be provided via a bus 416, which may
include one or more conductive lines. The bus 416 may be communicatively coupled to
one or more components of processor 400, for example the controller 414, cache 410,
and/or register 412. The bus 416 may be coupled to one or more components of the system.
[0133] The bus 416 may be coupled to one or more external memories. The external memories
may include Read Only Memory (ROM) 432. ROM 432 may be a masked ROM, Electronically
Programmable Read Only Memory (EPROM) or any other suitable technology. The external
memory may include Random Access Memory (RAM) 433. RAM 433 may be a static RAM, battery
backed up static RAM, Dynamic RAM (DRAM) or any other suitable technology. The external
memory may include Electrically Erasable Programmable Read Only Memory (EEPROM) 435.
The external memory may include Flash memory 434. The External memory may include
a magnetic storage device such as disc 436. In some embodiments, the external memories
may be included in a system.
[0134] It will be appreciated that the above description for clarity has described embodiments
of the invention with reference to different functional circuits, units and processors.
However, it will be apparent that any suitable distribution of functionality between
different functional circuits, units or processors may be used without detracting
from the invention. For example, functionality illustrated to be performed by separate
processors or controllers may be performed by the same processor or controllers. Hence,
references to specific functional units or circuits are only to be seen as references
to suitable means for providing the described functionality rather than indicative
of a strict logical or physical structure or organization.
[0135] The invention can be implemented in any suitable form including hardware, software,
firmware or any combination of these. The invention may optionally be implemented
at least partly as computer software running on one or more data processors and/or
digital signal processors. The elements and components of an embodiment of the invention
may be physically, functionally and logically implemented in any suitable way. Indeed
the functionality may be implemented in a single unit, in a plurality of units or
as part of other functional units. As such, the invention may be implemented in a
single unit or may be physically and functionally distributed between different units,
circuits and processors.
[0136] Although the present invention has been described in connection with some embodiments,
it is not intended to be limited to the specific form set forth herein. Rather, the
scope of the present invention is limited only by the accompanying claims. Additionally,
although a feature may appear to be described in connection with particular embodiments,
one skilled in the art would recognize that various features of the described embodiments
may be combined in accordance with the invention. In the claims, the term comprising
does not exclude the presence of other elements or steps.
[0137] Furthermore, although individually listed, a plurality of means, elements, circuits
or method steps may be implemented by e.g., a single circuit, unit or processor. Additionally,
although individual features may be included in different claims, these may possibly
be advantageously combined, and the inclusion in different claims does not imply that
a combination of features is not feasible and/or advantageous. Also the inclusion
of a feature in one category of claims does not imply a limitation to this category
but rather indicates that the feature is equally applicable to other claim categories
as appropriate. Furthermore, the order of features in the claims do not imply any
specific order in which the features must be worked and in particular the order of
individual steps in a method claim does not imply that the steps must be performed
in this order. Rather, the steps may be performed in any suitable order. In addition,
singular references do not exclude a plurality. Thus, references to "a", "an", "first",
"second" etc. do not preclude a plurality. Reference signs in the claims are provided
merely as a clarifying example shall not be construed as limiting the scope of the
claims in any way.