FIELD
[0001] The present disclosure is directed to personalized sound delivery and hearing aid
               devices and systems.
 
            BACKGROUND
[0002] Hearing systems, including earphones, headphones, head sets, hearing aids and other
               such devices, deliver sound to the ears of the user. These hearing systems may include
               certain filtering, amplifying, and other processing techniques applied to detected
               or received audio data for delivering through speakers as sound to the user's ears
               to enhance sound quality and filter unwanted background sound noise. Digital Signal
               Processing "DSP" is a known field including techniques for filtering and processing
               recorded signals and generating improved signals. Hearing systems are generally not
               personalized to user preferences in specific sound environments. Accordingly, the
               settings on hearing devices are generally optimized for the average listener based
               on hearing thresholds, and average sound environment and are not individually tailored
               or optimized to certain noise environments. Additionally, users cannot predict the
               sound levels in restaurants or bars they plan on frequenting. Also, clinicians generally
               only personalize the settings of a hearing device in their offices, requiring a user
               to make an appointment and travel to the office whenever a change in settings is desired.
 
            [0003] One known hearing aid personalization system is disclosed in 
U.S. Pat. No. 8,379,871 (Michael et al), as: "A personalized hearing profile is generated for an ear-level device comprising
               a memory, microphone, speaker and processor. Communication is established between
               the ear-level device and a companion device, having a user interface. A frame of reference
               in the user interface is provided, where positions in the frame of reference are associated
               with sound profile data. A position on the frame of reference is determined in response
               to user interaction with the user interface, and certain sound profile data associated
               with the position. Certain data is transmitted to the ear level device. Sound can
               be generated through the speaker based upon the audio stream data to provide real-time
               feedback to the user. The determining and transmitting steps are repeated until detection
               of an end event." (Abstract).
 
            [0004] What is needed is methods and systems that are personalized to user preferences in
               specific sound environments. More specifically, the systems and methods need to be
               individually tailored or optimized to certain noise environments and need to be capable
               of predicting the sound levels in restaurants or bars that they frequent or plan on
               frequenting. Further, a system is needed that clinicians can personalize more efficiently
               without the users needing to make an appointment and travel to the office whenever
               a change in a setting is desired.
 
            SUMMARY
[0005] Various systems and methods are disclosed herein to increase the quality of the sound
               delivered to a user when using a hearing system. For example, users of hearing systems
               generally cannot determine the noise level and quality at an establishment prior to
               frequenting it, and therefore cannot adjust their hearing devices in advance or make
               an informed decision to avoid certain establishments (e.g. restaurants, bars, etc.)
               that are too noisy. For instance, there is no accurate database of sound levels for
               establishments beyond subjective assessments and personal opinions posted on rating
               sites such as Yelp.
 
            [0006] Furthermore, regardless of whether users are aware of the sound level in advance,
               hearing devices are rarely optimized for any specific sound environment (e.g., restaurant).
               Because sound delivery devices are generally not customizable, users cannot adjust
               the sound settings to optimize them when they enter an environment for which their
               hearing device is not optimized. Accordingly, hearing systems, including hearing aids,
               are set to be the most useful for the average noise environment, and therefore may
               be poorly suited to specific environments that have noise profiles that deviate considerably
               from the average noise environment (e.g., divergent frequencies and amplitudes of
               sound from the average).
 
            [0007] In fact, clinicians can only customize the audio settings of most hearing devices
               in the office after setting up an appointment. Therefore, adjusting the device settings
               for particular environments such as, but not limited to restaurants, concerts and
               group settings, while in a clinician's office is not practical or effective in addressing
               situation-specific listening complaints. Hearing devices generally are not optimized
               for any particular environment beyond that of a clinician's office. Accordingly, many
               users suffer situational hearing problems due to poor environmental acoustics and
               excessive background noise. Moreover, users may enter environments that have dangerous
               levels of noise. In fact, exposure to dangerous noise is a leading cause of preventable
               hearing loss and is even more profound in individuals hearing aid devices who already
               suffer hearing impairment.
 
            [0008] Systems and methods have been developed to more effectively rate the noise level
               and quality of various locations. For instance, the hearing system may include a hearing
               device (e.g. hearing aid) and an integrated control, for instance a mobile phone or
               computing device that is wirelessly linked to the sound delivery device. For example,
               the hearing aid and associated device, e.g., mobile phone or PDA, may be wirelessly
               linked using radio technology, such as Bluetooth® compliant synchronous connection-oriented
               "SCO" channel protocol (e.g., Core Version 5.0 and predecessor versions) as regularly
               updated and enhanced by Bluetooth SIG Working Groups to meet evolving technology and
               market needs. The hearing aid and/or wireless device is enabled with Bluetooth compliant
               serial port profile "SPP", control channel, SCO channel and other necessary protocols
               and functions to insure proper operation and communication of signals and data. Other
               compatible platforms and specifications may be used to establish wireless links between
               devices. Microphone is a type of transducer that transfers or translates mechanical
               energy-such as sound vibration and converts that energy into an electrical signal,
               which may be amplified and/or further modified such as by an analog-to-digital converter
               to generate a digital signal for use in data/signal processing systems. The microphones,
               such as omnidirectional microphones, on the sound delivery device and/or the associated
               mobile device may then detect sound levels at different locations at different times.
               This will allow a significant amount of data from different users to be collected,
               aggregated and uploaded to a server for analysis to make the data available for these
               and other users to determine the current or average sound levels at particular establishments.
               Accordingly, a database could be created that includes sound level and characteristic
               information for different cities, restaurants, sporting venues, public transportation,
               and others places. These sound level ratings may then be aggregated by a server in
               a database and accessed by users through an application on their mobile phone, a website,
               or through integration into websites like Yelp.com. To do this, GPS data may be tagged
               to the sound level data recorded through the microphones to identify the location.
               Additionally, signal-to-noise ratio, and time-stamp data maybe tagged to the sound
               level. Alternatively, users may tag or indicate which establishment they are attending
               after receiving notifications on their mobile device or any combination of the tagging
               methods for confirmation.
 
            [0009] Accordingly, once a hearing system has recorded sound data from the user's current
               environment, that data may be utilized to adjust the settings on the hearing device
               manually or automatically. For example, in some embodiments, the user may modify the
               sound settings manually while gaging the effect in a new environment. This provides
               a user real time feedback to optimize the sound based on their preferences. For example,
               the user may set the gain, compression, frequency shaping, output volume, filters
               or amplifiers (e.g., filtering non-human voice frequencies and amplifying human voice
               frequencies), and microphone directionality. This may be performed via different methods
               through a graphical user interface, such as presented on a mobile device in communication
               with a hearing aid, e.g., via Bluetooth connection. In this manner, the invention
               enables a user to intuitively and/or graphically optimize the settings while listening
               in a new or changing environment.
 
            [0010] In other embodiments, the sound settings, including Digital Signal Processing (DSP)
               and filtering algorithms may be automatically and adaptively modified based on the
               acoustic data that is recorded through the microphones. For example, the audio data
               detected in certain environments may be monitored for certain amplitudes and frequencies
               and filtered if they cross a certain threshold. In other embodiments, the device may
               adaptively adjust the settings to amplify or filter certain frequencies more or less
               based on the magnitude and presence of certain frequencies. For example, as a voice
               is getting louder in a certain frequency range and outside of the algorithm target
               amplitude range, the system may adaptively dampen the sound in that frequency range,
               and vice versa. As another example, if music is newly detected that has a high amplitude,
               certain algorithms may be initiated or triggered to filter those additional frequencies.
 
            [0011] These optimized sound settings for particular locations may be tagged with GPS or
               other location data and time stamped to allow a user to easily recall those settings
               when reentering the same or similar acoustic environment. For example, if a user travels
               to the same coffee shop each day, the device or mobile phone associated with it may
               collect GPS data that indicates the user has entered the same geographic location
               as the coffee shop and retrieve the sound settings for that particular shop. Additionally,
               if the user attends the shop later in the day when the patrons are more lively and
               talkative, the noise reduction algorithms may be automatically increased based on
               the stored settings for that location and time of day. Additionally, these saved settings
               may be uploaded to the Internet and aggregated along with the noise data for those
               locations to allow other users to access the sound settings for the location and place
               uploaded by different users. Alternatively, the mobile device may associate the location
               with a coffee shop in general and apply a profile or set of settings modeled after
               typical same-store conditions.
 
            [0012] In some embodiments, the database and server may also analyze the data to output
               settings that are optimal for the particular location or provide ranges of optimal
               settings and allow a user to choose along a narrowed continuum, which would make the
               personalization process more efficient. In other embodiments, the average settings
               utilized by users (as indicative of the choices by users after real time feedback)
               may be available to other users entering that environment or other environments that
               are acoustically similar. Accordingly, the user may be presented with options to select
               particular sound settings upon arriving at a known location based on their own history
               or other users' history. These settings may also be updated, dynamically, in real
               time as the microphones of the user's sound delivery device and/or other users' microphones
               detect changes or updates in the acoustic environment.
 
            [0013] In other embodiments, users may have a personal profile associated with their settings
               that contain certain physiological hearing characteristics entered in by professionals
               and/or the user. Accordingly, users could be offered sound settings based on the choices
               of similar users and not based on the entire average of all users that attended a
               particular location.
 
            [0014] In other embodiments, instead of having to travel to the clinician's office to change
               settings in the device, and particularly out of the box, e.g., when initially setting
               up the hearing assistance device, the user may first adjust the hearing settings using
               a pre-recorded audio sample that is delivered to the user's ears through the speaker
               of the hearing device (e.g. hearing aid or an associated phone/computer's speakers).
               Accordingly, the audio sample may represent the average or most frequent speech experienced
               by the average user, or include samplings of the most often experienced frequencies
               and amplitudes of noises in certain environments. The user may then automatically
               adjust the settings to an optimal level based on their preferences after hearing the
               pre-recorded sample. Otherwise, users often will first adjust a sound delivery device
               out of the box (e.g., a hearing aid) based on the first sound environment they are
               in (e.g., home) which may not be optimal for the most problematic places, or even
               work. Accordingly, a pre-recorded sample with a wide range of frequencies and noise
               will allow the user to adjust the sound settings to an average that is more likely
               appropriate early in the adjustment process.
 
            [0015] Furthermore, systems and methods have been developed to allow a clinician to adjust
               the sound settings of a sound delivery system remotely. The remotely located clinician
               will be able to access an individual's hearing device via a network connection to
               the clinician's computing device (e.g., computer and/or mobile device) and be able
               to conduct a hearing test, sample the acoustic environment including decibel sound
               pressure level, spectral content, types of sounds and location of sounds and then
               tune the device while the person is wearing it in situ. The professional may do this
               by using an application on their computer or mobile device that could then connect
               via the Internet to the user's hearing device, mobile device (phone or tablet) and/or
               computer. The system would allow the professional to manipulate the settings of the
               hearing device remotely through a data connection using their expertise while receiving
               direct feedback from a user based on the noise of that particular environment. Then,
               these settings could be saved by a user and/or the system in a database linked to
               the systems servers, including with associated GPS and time data. These settings could
               then be easily accessed and restored, or analyzed for progressive changes in a user's
               hearing over time.
 
            [0016] These changes could also be utilized for proactive identification of problematic
               hearing changes that may indicate certain hearing pathologies experienced by a user.
               Certain changes within certain thresholds may cause a message to be sent to a clinician
               and/or the user to immediately make an appointment to evaluate and diagnose the hearing
               of the user.
 
            [0017] In some embodiments, while the microphone on the hearing device is recording sound
               it may also save a finite number of seconds or minutes of the audio file in a memory
               or buffer for instant replay for the user. Therefore, when a user presses a button
               on the hearing device or their mobile phone, the hearing device or mobile phone may
               play back the stored audio of the last X seconds. In some embodiments, pressing the
               button for replay could automatically send the buffered and/or stored audio data to
               the mobile phone and/or server to be filtered for voice signals or as indicated by
               a user and saved for continual replay. For example, if a name is repeated or directions
               are given a user may want it replayed and/or stored. In some embodiments a button
               on the hearing device or mobile phone may automatically send the previous certain
               number of seconds of sound recorded from the device microphones to the mobile phone
               without replaying at the time, but for easy replaying later.
 
            BRIEF DESCRIPTION OF THE DRAWINGS
[0018] The accompanying drawings, which are incorporated in and constitute a part of this
               specification, exemplify the embodiments of the present invention and, together with
               the description, serve to explain and illustrate principles of the invention. The
               drawings are intended to illustrate major features of the exemplary embodiments in
               a diagrammatic manner. The drawings are not intended to depict every feature of actual
               embodiments nor relative dimensions of the depicted elements, and are not drawn to
               scale.
               
               
Figure 1 is an overview of an example hearing system.
               Figure 2 is a diagram of an example of a hearing device.
               Figure 3 is a flow chart illustrating a process for aggregating sound level ratings
                  for specific locations.
               Figure 4 is a flow chart illustrating a process for suggesting sound settings for
                  specific locations.
               Figure 5 is a flow chart illustrating a process for outputting suggested sound settings
                  when a user enters a specific location.
               Figure 6 is a flow chart illustrating a process by which a remotely located clinician
                  may tune a hearing device.
               Figure 7 is a flow chart illustrating a process for tuning a hearing device using
                  a pre-recorded audio sample.
               Figure 8 is a flow chart illustrating a process for implementing an instant replay
                  feature for a hearing device.
 
            [0019] In the drawings, the same reference numbers and any acronyms identify elements or
               acts with the same or similar structure or functionality for ease of understanding
               and convenience. To easily identify the discussion of any particular element or act,
               the most significant digit or digits in a reference number refer to the Figure number
               in which that element is first introduced.
 
            DETAILED DESCRIPTION
[0020] Various examples of the invention will now be described. The following description
               provides specific details for a thorough understanding and enabling description of
               these examples. One skilled in the relevant art will understand, however, that the
               invention may be practiced without many of these details. Likewise, one skilled in
               the relevant art will also understand that the invention can include many other obvious
               features not described in detail herein. Additionally, some well-known structures
               or functions may not be shown or described in detail below, so as to avoid unnecessarily
               obscuring the relevant description.
 
            [0021] The terminology used below is to be interpreted in its broadest reasonable manner,
               even though it is being used in conjunction with a detailed description of certain
               specific examples of the invention. Indeed, certain terms may even be emphasized below;
               however, any terminology intended to be interpreted in any restricted manner will
               be overtly and specifically defined as such in this Detailed Description section.
 
            Hearing System Overview
[0022] Figure 1 illustrates an overview of an example hearing system 150 according to the
               present disclosure. The system 150 may include a hearing device 100, associated or
               used by, a user 105, a mobile device 110 linked with the hearing device using antennas
               115, and a personal profile for the user 105 that may be stored optionally on the
               mobile device 1 10 or elsewhere (e.g., server). In some embodiments, the hearing device
               100 may include a charging case that can store the audio data, and the audio data
               may be uploaded to a computer (for instance for users without a mobile device) which
               could then upload data over the network 120, such as, to a server 130. Additionally,
               a network 120 may also link the mobile device 110 and/or hearing device 100 to a server
               130 and database 160 that stores personal profiles, including software for analysis
               of sound data and performing other functions as disclosed herein. Furthermore, other
               users 107 with operating hearing devices 100 may also be linked to the network 120
               and sever 130 and sound/hearing data from the other users 107 hearing devices may
               be aggregated and stored in the database 160. In addition, a clinician 140 operating
               a computing device 135 may be connected to the network 120 via a communications component
               of a computing device 135 to allow the user to diagnose and make changes to the settings
               of the hearing devices 100. The changes made by clinicians 140 may also be stored
               in the database 160 for separate or combined reference. This will allow the clinician
               to remotely diagnose the users 105 and change the settings on their hearing devices
               100 while the user is in a noisy environment.
 
            [0023] The hearing system 150 efficiently optimizes the hearing device 100 in certain environments
               based on an accumulation of data from both the user 105 and other users 107, and in
               some cases the clinician 140. This accumulated data can be utilized to present the
               user options or automatically set the audio settings on a user's hearing device 100.
               Furthermore, this data may be utilized by clinicians to evaluate certain settings
               and improve their recommend settings for a given user 105 and noise environment.
 
            Personal Profiles
[0024] The hearing system 150 may include personal profiles 125 of users 105 that include
               their demographic information and individual hearing characteristics of a particular
               user 105. For example, the hearing system 150 may, upon initialization or the first
               time connecting a hearing device 105 to a computer or mobile device 1 10, prompt the
               user with a series of questions designed to obtain information for the personal profile
               to optimize the settings for a particular user 105.
 
            [0025] For example, the hearing system 150 may ask the user 105 its age, gender, hearing
               history, medication, and certain information regarding their hearing related medical
               history. For example, the system 150 may ask the user: whether they have tinnitus,
               whether they are deaf in any ears, and other questions related to hearing and hearing
               health. Additionally, clinical data relating to the user's 105 hearing history may
               be downloaded or manually entered to a user's profile 125 that includes information
               regarding a user's 105 sensitivity to certain frequency ranges and amplitudes or combinations
               of frequencies ranges and amplitudes.
 
            [0026] Additionally, the hearing device 100 may perform a hearing test automatically upon
               initialization that may then be stored with the personal profile 125. The information
               linked to a user's 105 personal profile 125 may be utilized to make suggestions for
               certain noise environments by, for example, recommending settings from other users
               107 that have similar profiles 125 or have similar characteristics in their profiles
               125. Furthermore, the personal profile 125 data could then be associated with the
               user's 105 audio data that is uploaded to the server 130 and stored in databases 160.
               This will allow the user's 105 data to be pooled with data associated with other users
               107 and data from the clinician 140, including the associated personal profile 125
               data for comparisons, and aggregation to discern trends and patterns to improve the
               recommend settings for users 105. Also, the history of settings associated with a
               particular user 105 may be used to notify a clinician if the system 150 detects hearing
               changes (based on the settings chosen) that indicate a potentially problematic hearing
               issue that is developing. Accordingly, prevention and treatment may be addressed earlier
               than previously possible and before a hearing condition has deteriorated further or
               to a point where it is not treatable.
 
            Hearing Device
[0027] Figure 2 is a diagram of an example hearing device 100 that includes a control system
               220 and memory 235 that may be flash memory, solid state memory, a combination, or
               any other types of computer memory. The control system 220 may be any type of controller
               or combination of controllers, including a microprocessor, microcontroller, or other
               processors and associated components. The hearing device 100 may also include an antenna
               1 15 for transmitting and/or receiving wireless signals. The hearing device 100 may
               employ any well-known radio and transceiver components including components using
               cellular, wideband, Bluetooth, Wi-Fi, NFC or other frequencies to transmit/receive
               data to/from an associated mobile device 110. Additionally, the hearing device 100
               may include a single or multiple microphone(s) 225 for detecting sound in the ambient
               environment and a speaker 215 for delivering sound to the user's 205 ear. In some
               embodiments, the hearing device 100 may include a hearing device that fits into a
               user's ear, along with a remote microphone that can be located next to the source
               of the desired sound that would connect to the ear portion of the hearing device with
               Bluetooth. The hearing device 100 may also include any combination of data and signal
               processing components 230 including analog-to-digital converters, filters, noise reduction
               circuitry, and other signal processing components. These components may be in any
               combination in order to condition the audio data for transmitting over the speaker
               or for filtering from the microphone(s).
 
            [0028] In some embodiments, the hearing device 100 may contain one or more processors, one
               or more analog-to-digital converters (ADC) and/or digital-to-analog converters (DAC),
               power supply including, for example, a battery, a charging function, user interface
               and indication equipment, such as buttons, and lights/LEDs, and one or more memory
               235. The various components may communicate via one or more control/signal buses.
               The hearing device 100 may buffer and/or store in memory 235 a certain, preselected
               or configurable amount of audio data from the speaker 215 continuously. In addition,
               memory 235 may have stored therein one or more of: audio data associated with functions
               performed on the mobile device 200; program and configuration data, user profile data,
               sound profile data; radio protocol/configuration/parameter program and data; and control
               parameters and other data for establishing communication links. In some embodiments,
               memory 235 may be in the form of flash memory or other nonvolatile memory and may
               include different types of memory 235. In some embodiments, the amount of audio data
               stored or recorded may be the last "X" seconds of audio data detected. This audio
               data may also be filtered accordingly, or stored in compressed or raw audio format.
               In some cases, the audio data will be stored in the buffer in distinct time blocks
               (e.g. 1 , 2, 3, 4 or 5 second time blocks) and may store 5, 10, 15 seconds, 1 minute,
               or other amounts of audio data in a buffer or other memory for purposes of: (1) sending
               data via the antenna 1 15 to the mobile device and/or over the network 120 to the
               servers 130 for storage on the database 160. In some embodiments, the database 160
               may be on the mobile device or other computing device itself or may be in the "cloud"
               connected to a remote server 130; (2) accessing via instant replay function directly
               by replaying over the speaker; and/or (3) other uses. For instance, if you are in
               a location (e.g., plane, basement) with no Internet access, the audio data may be
               saved on local memory 235 or on memory 235 of the mobile device 1 10 until an Internet
               connection is reached to upload to the remote server 130 and database 160. In some
               embodiments, the buffer or memory will add a new block of audio data continuously
               as a new block is recorded, and discard the last block of data.
 
            Sound Ratings by Location
[0029] Figure 3 illustrates a method for providing access to the current and/ or historical
               sound levels at certain establishments based on data collected from users' (105 or
               107) hearing devices 100 and mobile devices 110. The system 150 may also be utilized
               to record sound data 302 at various locations and establishments, and upload 310 for
               aggregation of that data 315 in order to output or send 320 it to users 105 or other
               people to provide them with information about the sound level and type. The audio
               data may be saved on local memory 235 or on memory 235 of the mobile device 1 10 until
               an Internet connection is reached to upload to the remote server 130 and database
               160.
 
            [0030] For example, the microphone(s) 225 on the hearing device 100, mobile device 110 or
               other associated or connected microphone(s) 225 may record ambient sound data and
               store it in the memory 235 and/or upload the data over the network for storage, and/or
               analysis by the server 130 on the database 160. The audio data may be saved on local
               memory 235 or on memory 235 of the mobile device 110 until an Internet connection
               is reached to upload to the remote server 130 and database 160. As each user 105 and
               all other users 107 each have microphones 225 and would attend different establishments,
               the pool of users 105 may create data that can be processed to provide detailed information
               about the sound level and quality at various establishments.
 
            [0031] Additionally, a GPS sensor in the mobile device 110 and/or hearing device 100 may
               tag or associate GPS coordinates to the sound data 305 so that the sound data can
               be referenced to a particular establishment with a time stamp. Accordingly, all of
               the sound data for a particular establishment can be aggregated 315 and analyzed to
               provide feedback to users on the current and typical noise levels 320. Additionally,
               the GPS data may also be date and time stamped to include additional metrics and averages
               for certain times and days that may have more noise (e.g., nights and weekends, or
               special nights).
 
            [0032] In order to associate the sound data with a particular establishment or location
               305, in some embodiments, the user 105 will have to choose the location 305 from a
               list of nearby locations. Accordingly, systems and methods are disclosed herein for
               the system 150 to associate the GPS location data (and in some embodiments time stamp
               data) with a list of establishments within a certain distance of the current GPS coordinates.
               This may be useful as some mobile phones 110 may experience interference or have GPS
               locations that are inaccurate enough that automatic association with a map labeled
               establishment may not function properly, and be assigned to an incorrect establishment
               (e.g. restaurant, bar).
 
            [0033] Additionally, different establishments may be on different levels of the same structure,
               and therefore either altitude data may be combined with GPS to determine the location
               or the user 105 may be provided a list of potential locations that are within a certain
               distance of the detected GPS coordinates and the user 105. Once the system 150 identifies
               the closest restaurants or establishments the user 105 could then select their current
               location on a display or other graphical interface (e.g. on the mobile device). The
               sound data could then be associated with the location 305 until the user moves outside
               of a certain range of the location and has left the establishment.
 
            [0034] Next, the aggregate of all of the sound data for a particular location must be aggregated/stored
               and analyzed 215 by the system 150. The data may first be uploaded 310 to the server
               130 over the network 120, for analysis and storage in the database 160. The algorithms
               utilized for analyzing all of the sound data may employee averages, medians, time
               averages, day averages, current noise levels (e.g., noise recorded within the last
               hour, 30 minutes, 10 minutes, etc.), pitch of the highest bands of noise, comparison
               of human voice frequencies with background noises, decibel levels in certain ranges,
               overall decibel levels and other data processing techniques.
 
            [0035] Furthermore, this information must then be accessible or output/sent 320 to users
               105 to determine the current or history of noise levels at certain establishments.
               For example, a user 105 may want to try a new restaurant on a Saturday night in downtown
               San Francisco but wants to check on the noise levels first. Thus, the user 105 may
               then send a request to the database to access the current noise levels, average noise
               levels for a Saturday night (with specific time - also, they can dynamically see what
               the sound level is in real-time or near real-time, based on current data being continuously
               uploaded), and any other metrics. Additionally, the user 105 may be more sensitive
               to noises in certain frequency ranges and the noise level data may be displayed in
               bands of frequencies of interest. For instance, the output 320 data or indication
               may be the decibel level in certain frequencies, including human voices, higher pitched
               noises (like coffee shop espresso machines) or lower frequency music at dance clubs
               or restaurants. Thus, the user 105 may get an idea of both the level and characteristics
               of noise at a particular restaurant prior to travelling there. This data could also
               be integrated with review sites such as Yelp in order to allow a user 105 to easily
               view noise levels, and/or filter by certain noise levels and types.
 
            [0036] The user could use this data 320 to adjust the sound settings on the hearing device
               100 based on the indication of noise levels received 320. Accordingly, the data output
               320 may include particular frequencies that should be filtered out or that could be
               automatically removed, including based on user preferences or sensitivities. Additionally,
               the data output 320 could indicate the user 105 must lower the sound level, or the
               hearing device 100 could automatically lower the volume of the hearing device 100
               based on a relative comparison of the current sound level to a baseline sound level.
 
            Crowd Sourced Sound Settings by Location
[0037] Figure 4 illustrates a method of personalizing sound settings for a particular location
               that is based on a user's 105 and or other user's 107 preferences and the particular
               characteristics of that sound environment or location. As described above with respect
               to the sound ratings, the hearing device 100 or mobile phone 110 may first record
               sound data 302 in an establishment. Additionally, the location may be tagged 305 (as
               described herein) to the data, along with the date and time. Then, the sound settings
               for the hearing device 100 may be adjusted 405 manually 409 or adaptively using an
               automatic algorithm 407. Once the settings have been adjusted, these settings may
               be uploaded 410 to a server 130 for aggregation 315 and analysis, which may then be
               utilized to output suggested sound settings 420 to the same user 105 or other users
               107 as individual users return to locations where sound data has been aggregated 315.
               In some embodiments, the uploading and/or saving the mobile device 110 of sound settings
               will allow the same user 105 to easily recall the settings when arriving at a new
               location.
 
            [0038] Once a user's 105 location is identified and/or confirmed via manual selection by
               the user 105, the sound settings may be adjusted or personalized for the location.
               In some embodiments, the user 105 will manually personalize the sound settings for
               the particular location. An example of an interface and process for personalization
               of a hearing device 100 for a particular sound environment is disclosed in 
U.S. Patent No. 8,379,871, issued February 19, 2013 titled Personalized Hearing Profile Generation with Real-Time Feedback. The sound
               processing modes that may be changed or switched on and off include volume controls,
               allow certain bandwidths of frequency to be amplified or filtered (e.g., filtering
               non-human voice frequencies and amplifying human voice frequencies), compression,
               microphone directionality and general environmental noise reduction algorithms.
 
            [0039] In some embodiments, instead of or in addition to manual personalization, the hearing
               device 100 and or mobile device 110 may automatically evaluate recently detected sound
               data and determine the optimal or likely optimal settings for the hearing system 150
               and/or device 100. For example, the system 150 may adjust the volume level based on
               the detected decibel level of the environment. Additionally, various filtering algorithms
               may be applied to certain bandwidths and techniques such as Adaptive Constructive
               Interference, Destructive Interference, Active Noise Cancellation may be used. For
               instance, bandwidths that do not generally include frequencies of a human voice for
               a normal conversational tone (as opposed to singing with music in the background)
               may be amplified or preserved, while other frequencies associated with ambient noise
               (i.e. car driving noise, coffee machines, music, etc.) may be reduced or eliminated
               from the sound that is eventually produced through the speakers 215 of the hearing
               device 100.
 
            [0040] These settings may then be tagged to the location and also the date and time of the
               setting. These settings and tags may then be uploaded 410 to the server 130 to be
               aggregated 315 and analyzed by the server 130, and the data stored in the database
               160. The settings may also be saved on local memory 235 or on memory 235 of the mobile
               device 110 until an Internet connection is reached to upload to the remote server
               130 and database 160. In some embodiments, the system could pre-populate the settings
               based on the user's home settings, and then provide an option to the user to update
               the settings with locally stored new settings when connecting the device to the network,
               computer or other computing device.
 
            [0041] In other embodiments, the data may be sent to over the antenna 1 15 to the mobile
               device 1 10 or the sound data may be sent to the mobile device 110 the GPS data from
               the mobile device 110 would be tagged to the sound/settings data. In some embodiments,
               the mobile device 1 10 and/or the server 130 and database 160 may store the settings
               to be recalled by the system 150. In some embodiments these settings may be aggregated
               315 from all users 107 and analyzed by the server 130 to provide averages or associate
               certain types of sound settings with various personal profiles 125 or certain characteristics
               in personal profiles 125. Accordingly, this information may be utilized to output
               suggested sound settings 420 based on similar users 105 or the same user 105 in similar
               circumstances.
 
            [0042] There are several ways by which sound settings may be output or suggested 420 to
               a user 105. For instance, if a user 105 has attended a certain location (and even
               at a certain time) the user 105 may have set the hearing device's 100 settings for
               that location as disclosed herein. Then the system 150 would have saved those settings
               and associated with that location (and potentially date and time). Then, once the
               user 105 reenters that location the system 150 could automatically recall those settings,
               or pop up a message on the display or user interface of an associated mobile device
               1 10 or an audio message through hearing device's 100 speaker 215 asking whether the
               user 105 has reentered a suggested location.
 
            [0043] If the user 105 indicates it is the correct location, the user's 105 own saved settings
               may be recalled to allow the user 105 to easily adjust its hearing to that environment.
               In other embodiments, if the user 105 is new to a location, once the system 150 determines
               (and/or the user 105) confirms the actual location, setting and/or sound data from
               other users 107 may be suggested to the user 105, or automatically switched on if
               the user does not have any saved settings.
 
            [0044] For example, in some embodiments, the user's 105 profile 125 may be compared with
               other user's 107 profiles 125 (and/or data from clinician 140) that have attended
               the same location, to identify another user's profile 125 with similar characteristics
               (that are relevant to sound perception and hearing) and recommend settings or ranges
               of settings to the user 105 based on the similar user 107. In some embodiments, after
               those sound settings are output as suggested 510 (as shown in Figure 5), the user
               105 may then further manually adjust the settings 407 or the system may automatically
               adjust the settings 409 to further refine the settings based on the user's profile
               125, the user's history of sound settings, the currently detected ambient sound, or
               through manual adjustment 409. In some embodiments, the settings may be selected by
               finding settings from other users 107 that were based on sound data that was most
               similar to the current sound levels, for example, within certain frequencies. Various
               matching, machine learning, statistical analysis, or other algorithms may be utilized
               to match the settings to one another from different user's 107 and profiles 125.
 
            [0045] Then, after the user 105 further refines the settings that were suggested, the settings
               may be uploaded 410 along with the sound data and further aggregated and saved for
               later use or analysis. Then, as above, that individual user's 105 sound settings may
               be recalled when the user 105 re-enters the same location 520, with possible further
               refinements as described above with respect to Figure 5.
 
            Tuning by Remote Clinician
[0046] In addition to being set automatically or manually by a user 105, a remotely located
               clinician may also be able to tune the audio settings of a hearing device 100 as disclosed
               herein using the present system 150. This will allow a professional with expert training
               to remotely adjust the settings, based on the sound data detected by the hearing device
               100 and real-time feedback from the user 105 while the user 105 is in the actual environment
               of interest. This may increase the speed and accuracy with which the device 100 is
               set, and also the frequency at which the device 100 may be expertly personalized for
               particular locations.
 
            [0047] Figure 6 illustrates a flow chart of a process 600 by which a remotely located clinician
               may adjust the hearing device 100. For instance, the hearing device 100 or mobile
               device 110 may first detect or record sound data 302. Then the user 105 may also provide
               verbal or electronic feedback 505 through the mobile device or hearing device 100
               based on the current settings. Then, this data and feedback may be uploaded 610 and
               sent over the network 120 the server 130 and/or a remote clinician's computing device
               135 for analysis and evaluation by the clinician. Then the clinician may then remote
               send commands to modify the settings of the hearing device 100 over the network 120.
               The hearing device's 100 processor or control system 220 may then receive the commands
               and implement the settings. Then, the process may begin again with new user feedback
               605. If the feedback is negative, then the clinician may further modify the settings.
 
            [0048] Finally, once the settings are finalized, they may be stored and associated with
               the user's profile 635 at some point in the system 150 for recalling. For instance,
               they may be stored on the mobile device 1 10 or the database 160 or other points in
               the system 150. Then those settings may be recalled and output to suggest settings
               420 once the user 105 returns to the specific location, an acoustically similar situation,
               or generally attempts to recall its favorite settings. This process will allow a clinician
               to evaluate user feedback of a real sound environment while also analyzing the acoustic
               properties of the sound. Also, it will immensely increase the practicality of personalizing
               the hearing system 150 for specific environments.
 
            Tuning using Pre-Recorded Audio Sample
[0049] In some embodiments, part of the setting process for a hearing system 150 as disclosed
               herein may be to utilize pre-recorded audio samples as described in the pre-recorded
               setting process 700 illustrated in Figure 7. First, pre-recorded audio samples may
               be played 702 through the speaker 215 of the hearing device 100 or mobile device 110
               (or computer or other Internet connected computing device with speaker(s)). Then the
               user 105 may adjust the sound processing settings 405 accordingly based on their own
               feedback after hearing the prerecorded audio sample played through one of the speakers.
               In some embodiments, the pre-recorded sound sample may be played several times, with
               the user 105 adjusting the sound settings 705 in-between playbacks of the pre-recorded
               audio sample 702. Eventually, once the user has arrived at a sound setting that they
               feel is adequate based on the pre-recorded sample, the settings may be saved as disclosed
               herein, and uploaded to be associated with the user's 105 profile 710. As disclosed
               herein, the settings may then be aggregated 315 and analyzed. Then this analyzed data
               may be utilized to output suggested sound settings 420 or the system 150 may simply
               recall a user's 105 previously saved sound settings.
 
            [0050] The pre-recorded audio sample may be played over the hearing device's 100 speaker
               125, the mobile device's 110 speaker, or another speaker associated with the system
               150 that is within hearing range of the user 105. When the sound is played over the
               mobile device it may have different characteristics than when played over the hearing
               device speaker 110 due to the difference in range and acoustics. Therefore, a different
               pre-recorded audio sample may be played 702 based on which speaker 125 plays the sound.
 
            [0051] The pre-recorded audio sample may be any mixture of frequencies, amplitudes and temporal
               information. For instance, in some embodiments, the pre-recorded sample may be in
               the human voice frequency range or consist of a human voice speaking. In other embodiments,
               it may just be non-speaking sounds in the human frequency range. Additionally, other
               frequencies may be included in the pre-recorded sample, including frequencies that
               are typical for environment noise that the average user 105 experiences. The noise
               samples could be of varying length, from 5 seconds to a minute to 30 seconds to 20
               seconds or 2 minutes or other lengths. Additionally, the sample may contain any combination
               of frequencies and amplitudes associated with each of the frequencies.
 
            [0052] The pre-recorded audio sample may be obtained from a variety of sources including
               from recording noise and human voices in a sound study or from other sources. For
               instance, the pre-recorded audio sample could be obtained from the aggregated data
               315 from all other users 107 and analyzed for the noise frequencies with highest amplitudes
               on average. In some cases, the system could prompt the user to ask what language they
               speak or will be listening to, and play pre-recorded audio samples within the frequency
               and amplitude range for specific languages and dialects selected.
 
            [0053] In other embodiments, perhaps a particular noise type is frequently experienced by
               all users, or all users within the same geographic region (and thus likely to experience
               similar noise) that competes with more desirable sounds such as speech. As one example,
               everyone in New York City may experience constant car horns or everyone in San Francisco
               may experience loud noise while riding the Bay Area Rapid Transit (BART) system. The
               system could analyze the data and pull out the frequency ranges that are most likely
               experienced and add those frequencies to the pre-recorded sample or similar frequencies
               that are averages within the same band so that the listener could optimize settings
               to minimize the impact of these sounds on their ability to detect and make use of
               more desirable sound signals.
 
            [0054] In some embodiments, the pre-recorded sample may contain multiple steps for tuning
               the hearing device 100 and therefore may contain multiple segments of pre-recorded
               audio sample for the user to separately provide feedback. For instance, if all of
               the relative noise is played at once it may be harder for a user 105 to evaluate the
               effects of programming changes as noise is general experienced in more isolated incidents.
               Therefore, each audio sample might have a different frequency of common noise (e.g..
               the car horn/traffic noise) and an underlying human voice for the user 105 to provide
               feedback. In some embodiments, this will allow the user 105 to optimize certain noise
               filters for each bandwidth of relevant noise. Also, the user 105 may be presented
               with varying pitches of voices and amplitudes for further modification to ensure a
               robust tuning.
 
            [0055] In some embodiments, prior to tuning with a pre-recorded audio sample, the device
               may record sound data 302 for a set period of time (e.g. 1 week, 2 weeks) while a
               user 105 experiences a normal work week, commuting to work, going to normal restaurants
               and shops, and other normal activities. This will allow the device to average or detect
               the most relevant sound signatures (i.e., non-human voice) a specific user 105 experiences
               on average. Then, the system 150 may be able to make a custom noise sample that could
               be overlaid on top of a human voice for the pre-recorded sample tuning method 700.
               Accordingly, in this embodiment, the pre-recorded sample will be customized and specialized
               for the particular user's 105 normal activities.
 
            Instant Replay for a Hearing Device
[0056] In some embodiments, and with reference to Figure 8, the hearing system 150 may include
               an instant replay process 800 to allow the user 105 to replay the last few seconds
               or minutes of audio data recorded 302 by the hearing system 150. This may be useful
               in noisy situations, when a user 105 is distracted, or when a user 105 is trying to
               remember a phone number or directions. For instance, the microphone 225 on either
               the hearing device 100, the user's 105 mobile device, or other associated microphone,
               may record sound data 302 continuously and keep the last few seconds of audio data
               recorded 805, and discard or save older sound data. Then, when the user 105 requests
               a replay 810 by interacting through its mobile phone 1 10 or hearing device 100, the
               device may play back the last few seconds of data either through the speaker 215 on
               the hearing device 100 or mobile phone 110.
 
            [0057] In some embodiments, the user 105 may have a choice of how many seconds of data are
               replayed. For example, the system 150 may store a minute or longer of audio data continuously,
               and the user might have an option on their phone as to how many seconds of replay
               the user 105 would like to request or certain windows of time. In some embodiments,
               the hearing device 100 may contain an instant replay button that either immediately
               replays the sound, sends it and saves it to other components of the system such as
               the mobile device 110 or pops up a notification on the mobile device requesting the
               amount of time for which playback is requested. The system 150 may include the ability
               to record locally with finite memory 235, local memory 235 and/or on memory 235 of
               the mobile device 110, and upload recordings to the remote server 130 and database
               160 and synchronize recordings on all devices.
 
            [0058] This sound data may be stored in the memory 235 of the hearing device 100 or the
               mobile device 110, or on other places in the system 150. In some embodiments, this
               may include storage by the server 130 in the database 160. The audio data loop may
               be saved in a buffer, flash memory, or other memory. In some embodiments, the audio
               data will be stored every few seconds from the buffer into other memory, so that the
               audio data is recorded in a few second, one second, five second, or other blocks of
               time. These blocks of time are then continuously saved 805 and time stamped, so that,
               in some embodiments, as new blocks of time are added, the oldest time block is discarded.
               This process may continue until a user 105 initiates instant replay 810 by pressing
               a button on the hearing device 100, by pressing a button or interacting with their
               mobile phone 1 10 or by other initiation methods.
 
            [0059] Then, once instant replay command is received 810, the stored audio data will be
               placed back 820 to the user 105 over one of the speakers 215. Alternately, the system
               150 may have pre-recorded sound played back on phone/tablet 1 10, computer, or server.
               In some embodiments, pushing instant replay will also save the audio file to memory
               235 in the hearing device 100, and/or send the data to the mobile device 1 10 or a
               remote device 815 (e.g. server 130 and database 160) so that it may be further replayed
               by accessing through the system 150.
 
            [0060] In some embodiments, the instant replay function may first filter the blocks of data
               using a different filter than currently being utilized or set by the hearing system
               150. For example, in some embodiments, the instant replay function may have its own
               filter, and in particular, may filter out all other frequencies except human voice
               related frequencies. In other embodiments, the noise filter for voice may include
               an adaptive filter that filters out all other noise besides a specific voice, so that
               playback for instant replay only includes primarily a human voice. This is because
               instant replay will likely be used to repeat conversations, and would generally not
               be used to repeat other bands of noise that are not related to human voices. Additionally,
               directionality filters or other techniques may be utilized to filter out everything
               but the noise from a human voice with which a user 105 is having a conversation.
 
            [0061] Particular implementations of the subject matter have been described. Other implementations
               are within the scope of the following claims. In some cases, the actions recited in
               the claims can be performed in a different order and still achieve desirable results.
               In addition, the processes depicted in the accompanying figures do not necessarily
               require the particular order shown, or sequential order, to achieve desirable results.
 
            [0062] While this specification contains many specific implementation details, these should
               not be construed as limitations on the scope of any inventions or of what may be claimed,
               but rather as descriptions of features specific to particular implementations of particular
               inventions. Certain features that are described in this specification in the context
               of separate implementations can also be implemented in combination in a single implementation.
               Conversely, various features that are described in the context of a single implementation
               can also be implemented in multiple implementations separately or in any suitable
               sub combination. Moreover, although features may be described above as acting in certain
               combinations and even initially claimed as such, one or more features from a claimed
               combination can in some cases be excised from the combination, and the claimed combination
               may be directed to a sub combination or variation of a sub combination.
 
            [0063] Similarly while operations may be depicted in the drawings in a particular order,
               this should not be understood as requiring that such operations be performed in the
               particular order shown or in sequential order, or that all illustrated operations
               be performed, to achieve desirable results. In certain circumstances, multitasking
               and parallel processing may be advantageous. Moreover, the separation of various system
               components in the implementations described above should not be understood as requiring
               such separation in all implementations, and it should be understood that the described
               program components and systems can generally be integrated together in a single software
               product or packaged into multiple software products.
 
            [0064] It should initially be understood that the disclosure herein may be implemented with
               any type of hardware and/or software, and may be a pre-programmed general purpose
               computing device. For example, the system may be implemented using a server, a personal
               computer, a portable computer, a thin client, or any suitable device or devices. The
               disclosure and/or components thereof may be a single device at a single location,
               or multiple devices at a single, or multiple, locations that are connected together
               using any appropriate communication protocols over any communication medium such as
               electric cable, fiber optic cable, or in a wireless manner.
 
            [0065] It should also be noted that the disclosure is illustrated and discussed herein as
               having a plurality of modules which perform particular functions. It should be understood
               that these modules are merely schematically illustrated based on their function for
               clarity purposes only, and do not necessary represent specific hardware or software.
               In this regard, these modules may be hardware and/or software implemented to substantially
               perform the particular functions discussed. Moreover, the modules may be combined
               together within the disclosure, or divided into additional modules based on the particular
               function desired. Thus, the disclosure should not be construed to limit the present
               invention, but merely be understood to illustrate one example implementation thereof.
 
            [0066] The computing system can include clients and servers. A client and server are generally
               remote from each other and typically interact through a communication network. The
               relationship of client and server arises by virtue of computer programs running on
               the respective computers and having a client-server relationship to each other. In
               some implementations, a server transmits data (e.g., an HTML page) to a client device
               (e.g., for purposes of displaying data to and receiving user input from a user interacting
               with the client device). Data generated at the client device (e.g., a result of the
               user interaction) can be received from the client device at the server.
 
            [0067] Implementations of the subject matter described in this specification can be implemented
               in a computing system that includes a back-end component, e.g., as a data server,
               or that includes a middleware component, e.g., an application server, or that includes
               a front-end component, e.g., a client computer having a graphical user interface or
               a Web browser through which a user can interact with an implementation of the subject
               matter described in this specification, or any combination of one or more such back-end,
               middleware, or front-end components. The components of the system can be interconnected
               by any form or medium of digital data communication, e.g., a communication network.
               Examples of communication networks include a local area network ("LAN") and a wide
               area network ("WAN"), an internetwork (e.g., the Internet), and peer-to-peer networks
               (e.g., ad hoc peer-to-peer networks).
 
            [0068] Implementations of the subject matter and the operations described in this specification
               can be implemented in digital electronic circuitry, or in computer software, firmware,
               or hardware, including the structures disclosed in this specification and their structural
               equivalents, or in combinations of one or more of them. Implementations of the subject
               matter described in this specification can be implemented as one or more computer
               programs, i.e., one or more modules of computer program instructions, encoded on computer
               storage medium for execution by, or to control the operation of, data processing apparatus.
               Alternatively or in addition, the program instructions can be encoded on an artificially-generated
               propagated signal, e.g., a machine-generated electrical, optical, or electromagnetic
               signal that is generated to encode information for transmission to suitable receiver
               apparatus for execution by a data processing apparatus. A computer storage medium
               can be, or be included in, a computer-readable storage device, a computer-readable
               storage substrate, a random or serial access memory array or device, or a combination
               of one or more of them. Moreover, while a computer storage medium is not a propagated
               signal, a computer storage medium can be a source or destination of computer program
               instructions encoded in an artificially-generated propagated signal. The computer
               storage medium can also be, or be included in, one or more separate physical components
               or media (e.g., multiple CDs, disks, or other storage devices).
 
            [0069] The operations described in this specification can be implemented as operations performed
               by a "data processing apparatus" on data stored on one or more computer-readable storage
               devices or received from other sources.
 
            [0070] The term "data processing apparatus" encompasses all kinds of apparatus, devices,
               and machines for processing data, including by way of example a programmable processor,
               a computer, a system on a chip, or multiple ones, or combinations, of the foregoing
               The apparatus can include special purpose logic circuitry, e.g., an FPGA (field programmable
               gate array) or an ASIC (application-specific integrated circuit). The apparatus can
               also include, in addition to hardware, code that creates an execution environment
               for the computer program in question, e.g., code that constitutes processor firmware,
               a protocol stack, a database management system, an operating system, a cross-platform
               runtime environment, a virtual machine, or a combination of one or more of them. The
               apparatus and execution environment can realize various different computing model
               infrastructures, such as web services, distributed computing and grid computing infrastructures.
 
            [0071] A computer program (also known as a program, software, software application, script,
               or code) can be written in any form of programming language, including compiled or
               interpreted languages, declarative or procedural languages, and it can be deployed
               in any form, including as a stand-alone program or as a module, component, subroutine,
               object, or other unit suitable for use in a computing environment. A computer program
               may, but need not, correspond to a file in a file system. A program can be stored
               in a portion of a file that holds other programs or data (e.g., one or more scripts
               stored in a markup language document), in a single file dedicated to the program in
               question, or in multiple coordinated files (e.g., files that store one or more modules,
               sub-programs, or portions of code). A computer program can be deployed to be executed
               on one computer or on multiple computers that are located at one site or distributed
               across multiple sites and interconnected by a communication network.
 
            [0072] The processes and logic flows described in this specification can be performed by
               one or more programmable processors executing one or more computer programs to perform
               actions by operating on input data and generating output. The processes and logic
               flows can also be performed by, and apparatus can also be implemented as, special
               purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC
               (application-specific integrated circuit).
 
            [0073] Processors suitable for the execution of a computer program include, by way of example,
               both general and special purpose microprocessors, and any one or more processors of
               any kind of digital computer. Generally, a processor will receive instructions and
               data from a read-only memory or a random access memory or both. The essential elements
               of a computer are a processor for performing actions in accordance with instructions
               and one or more memory devices for storing instructions and data. Generally, a computer
               will also include, or be operatively coupled to receive data from or transfer data
               to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical
               disks, or optical disks. However, a computer need not have such devices. Moreover,
               a computer can be embedded in another device, e.g., a mobile telephone, a personal
               digital assistant (PDA), a mobile audio or video player, a game console, a Global
               Positioning System (GPS) receiver, or a portable storage device (e.g., a universal
               serial bus (USB) flash drive), to name just a few. Devices suitable for storing computer
               program instructions and data include all forms of non-volatile memory, media and
               memory devices, including by way of example semiconductor memory devices, e.g., EPROM,
               EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable
               disks; magneto-optical disks; and CD-ROM and DVD-ROM disks. The processor and the
               memory can be supplemented by, or incorporated in, special purpose logic circuitry.
 
            [0074] Various example configurations of the present invention are set out in the following
               numbered clauses:
               
               
                  - 1. A mobile device for calculating and generating sound level rating for various establishments
                     comprising:
                     
                     an antenna that receives audio data from a hearing device; a GPS receiver configured to output location data indicative of a location of the
                        mobile device; and a memory containing machine readable medium comprising machine executable code having
                        stored thereon instructions; a control system having at least one processor coupled to the memory, the processor
                        configured to execute the machine executable code to cause the control system to:
                        
                         associate the location data with the audio data; and transmit the associated location data and audio data over a wireless network to a
                           remote server. 
- 2. The mobile device of clause 1, wherein location data is associated with audio data
                     that was created on the same date and within a specified time period as the location
                     data.
- 3. The mobile device of clause 2, wherein the specified time period is 30 seconds.
- 4. The mobile device of clause 1 , wherein the remote server determines a set of establishments
                     within a certain distance of the location represented by the location data.
- 5. The mobile device of clause 1, wherein the control system is configured to determine
                     a set of establishments within a certain distance of the location represented by the
                     location data.
- 6. The mobile device of clause 5, wherein the control system is further configured
                     to present the set of establishments on a user interface screen to request a user
                     to select one of the possible establishments.
- 7. The mobile device of clause 4, wherein the control system is further configured
                     to receive user selection of one of the possible establishments and associate the
                     establishment with the audio data.
- 8. A system for processing of sound data relating to specific establishments comprising:
                     a database;
 a memory containing non-transitory machine readable medium and having stored thereon
                     machine executable code adapted to process sound data;
 a control system including at least one processor coupled to the memory, the processor
                     configured to execute the machine executable code to cause the control system to:receive, at a server, a first set of audio data recorded at a location from a first
                        mobile device and a second set of audio data recorded at the location from a second
                        mobile device; and process, by the server, the first and second sets of audio data to determine an indication
                        of a noise level at the location and store the indication in the database. 
- 9. The system of clause 8, wherein the location is a restaurant, bar, or concert venue.
- 10. The system of clause 8, wherein the first set of audio data is referenced to a
                     first set of location data detected at the location and the second set of audio data
                     is referenced to a second set of location data detected at the location.
- 11. The system of clause 10, wherein data that is referenced comprises data that is
                     received by the server in the same batch or within a threshold time window from the
                     same mobile device.
- 12. The system of clause 8, wherein the first and second sets of audio data are referenced
                     to a date and time recording.
- 13. The system of clause 8, wherein the control system is further configured to output
                     the indication of the noise level at the location.
- 14. The system of clause 8, wherein the stored noise level indication is further characterized
                     by one or more from the group consisting of: a current noise level; a historical noise
                     level; an average noise level; a category of noise level; an output in decibels; an
                     output in decibels by frequency band; a color-coded system; a numeric decibel value;
                     and a qualitative assessment.
- 15. A hearing device and associated application for calculating and generating sound
                     level ratings for various establishments comprising:
                     
                     a microphone adapted to transduce detected sound energy into an electrical signal;
                        an analog-to-digital converter adapted to convert the electrical signal into a digital
                        signal; a control system having a processor coupled with a memory, the memory having stored
                        thereon executable instructions and being adapted to store audio data representing
                        the digital signal; and a radio transmitter adapted to transmit the audio data to a mobile device having an
                        antenna within range of the radio transmitter, the mobile device adapted to receive
                        the audio data and having a mobile device memory adapted to store the audio data and
                        an application comprising instructions for processing the audio data, the mobile device
                        having a GPS module adapted to generate location data and having a processor adapted
                        to execute the instructions stored in the mobile device memory and associate GPS location
                        data with stored audio data. 
- 16. The hearing device and associated application of clause 15, wherein the application,
                     when executed by the mobile device processor, is adapted to cause the mobile device
                     to:
                     
                     associate the GPS location data with the audio data; and transmit the associated GPS location data and audio data over a wireless network to
                        a remote server. 
- 17. The hearing device and associated application of clause 15, wherein the application,
                     when executed by the mobile device processor, is adapted to cause the mobile device
                     to filter the audio data to remove frequencies outside the human speech range.
- 18. The hearing device and associated application of clause 15, wherein the application,
                     when executed by the mobile device processor, is adapted to cause the mobile device
                     to filter and process the audio data to provide an indication of a level or content
                     of specific frequency bands in order to send only the indication of the level of the
                     specific frequency bands to the remote server.
- 19. A hearing device capable of being remotely tuned by clinician comprising:
                     
                     a speaker configured to direct sound into a patient's ear; a microphone configured to output audio data representing ambient sound detected by
                        the microphone; a memory containing machine readable medium comprising machine executable code having
                        stored thereon instructions; a control system coupled to the memory, the control system configured to execute the
                        machine executable code to cause the control system to:
                        
                         transmit the audio data to a computing device via a communications link established
                           with the hearing device; receive control data from the computing device that includes instructions for modifying
                           audio settings that control the output of the speaker; and execute the control data to modify the audio settings. 
- 20. The hearing device of clause 19, further comprising a signal processor adapted
                     to process the audio data into processed audio data and wherein the audio data transmitted
                     via the communications link comprises processed audio data.
- 21. The hearing device of clause 20, wherein the instructions for modifying the audio
                     settings include instructions for modifying the processing of the audio data.
- 22. The hearing device of clause 20, where the signal processor includes one or both
                     of noise filters and amplifiers.
- 23. The hearing device of clause 19, wherein the instructions for modifying the audio
                     settings include instructions for modifying the volume controls and/or instructions
                     for modifying a directionality of the microphone.
- 24. The hearing device of clause 20, wherein the processed audio data comprises amplitudes
                     and content of certain frequency bands.
- 25. The hearing device of clause 19, wherein the computing device is one from a group
                     consisting of a mobile phone; a remote computer; a Personal Digital Assistant; an
                     electronic tablet; and a remote server.
- 26. A hearing device capable of replaying recent sound data recorded by a hearing
                     device comprising:
                     
                     a speaker configured to direct sound into a patient's ear; a microphone configured to detect ambient sound and output audio data representing
                        the ambient sound; a memory containing machine readable medium comprising machine executable code having
                        stored thereon instructions; and a control system coupled to the memory, the control system configured to execute the
                        machine executable code to cause the control system to:
                        
                         store a predetermined length of the audio data that represents recently recorded audio
                           data; and replay the stored predetermined length of audio data upon receiving a command to replay
                           the audio data. 
- 27. The hearing device of clause 26, wherein the predetermined length of audio data
                     that represents the recently recorded audio data is transmitted to a mobile device
                     for storage.
- 28. The hearing device of clause 26, wherein the replaying is performed through the
                     speaker.
- 29. The hearing device of clause 27, wherein the replaying is performed through a
                     second speaker on a mobile device.
- 30. The hearing device of clause 26, wherein the predetermined length is adjustable.
- 31. The hearing device of clause 26, wherein the memory includes a buffer.
- 32. The hearing device of clause 26, wherein the replay command is received from mobile
                     device.
- 33. The hearing device of clause 26, wherein the storing comprises storing the audio
                     data in a replay set of audio blocks, wherein each audio block is shorter in length
                     that the predetermined length.
- 34. The hearing device of clause 33, wherein the storing further comprises continuously
                     referencing the most recently recorded audio block as the last to be played in the
                     replay set of audio blocks and de-referencing the earliest recorded time block in
                     the replay set of audio blocks so as to exclude it from the replay set of audio blocks.
- 35. The hearing device of clause 34, wherein the de-referencing comprises deleting.
- 36. A mobile device for storing sound settings for an associated hearing device, the
                     mobile device comprising:
                     
                     an antenna that receives audio data from a hearing device; a GPS receiver configured to output location data indicative of a location of the
                        mobile device; a memory containing machine readable medium comprising machine executable code having
                        stored thereon instructions; a control system coupled to the memory, the control system configured to execute the
                        machine executable code to cause the control system to:
                        
                         transmit the location data to a server; and receive control data from the server with instructions for setting audio controls
                           on the hearing device from a first setting to a second setting. 
- 37. The mobile device of clause 36, wherein the audio controls comprise one or more
                     of noise filters, volume controls, and amplifiers.
- 38. The mobile device of clause 36, wherein the control system is configured to receive
                     data representing a set of establishments within a certain distance of the location
                     represented by the location data.
- 39. The mobile device of clause 38, further comprising a display for presenting a
                     user interface screen, and wherein the control system is further configured to present
                     the set of establishments on a user interface screen comprising a set of user selectable
                     elements representing the set of establishments.
- 40. The mobile device of clause 39, wherein the control system is further configured
                     to receive user selection of one of the set of establishments via the set of user
                     selectable elements and transmit the user selection to the server.
- 41. The mobile device of clause 36, wherein the control system is further configured
                     to transmit instructions to the hearing device to change the audio settings on the
                     hearing device to a different setting if the controller determines based on the data
                     output by the GPS receiver that the mobile device has changed location.
- 42. A system for aggregation and processing of sound data relating to specific establishments
                     comprising:
                     
                     a database a memory containing non-transitory machine readable medium and having stored thereon
                        machine executable code adapted to process sound data; a control system coupled to the memory, the control system configured to execute the
                        machine executable code to cause the control system to:
                        
                         receive, at a server, a first set of audio settings recorded at a location from a
                           first mobile device and a second set of audio settings recorded at the location from
                           a second mobile device; and process, by the server, the first and second sets of audio settings in order to determine
                           suggested audio settings associated with the location and store the suggested settings
                           in the database. 
- 43. The system of clause 42 wherein the first set of audio settings are referenced
                     to a first set of location data detected at the location and the second set of audio
                     settings are referenced to a second set of location data detected at the location.
- 44. The system of clause 43 wherein data that is referenced to other data comprises
                     data that is received in the same batch by the server or is received within a threshold
                     time window by the server.
- 45. The system of clause 42 wherein the first and second sets of audio settings are
                     date and time stamped.
- 46. The system of clause 42 wherein the control system is further configured to send,
                     to a third mobile device, suggested audio settings after receiving an indication from
                     the third mobile device that the third mobile device is co-located.
- 47. The system of clause 42 wherein the audio settings comprise one or more of a noise
                     filter, amplifiers, and volume controls.
- 48. A hearing device capable of being tuned by a pre-recorded audio sample comprising:
                     
                     a speaker configured to direct sound into a patient's ear; a microphone configured to output audio data representing ambient sound; a memory containing machine readable medium comprising machine executable code having
                        stored thereon instructions; a control system coupled to the memory, the control system configured to execute the
                        machine executable code to cause the control system to:
                        
                         play, by the speaker, a pre-recorded audio sample in response to receiving instructions
                           to initiate a tuning process; receive control data that includes instructions for modifying audio settings that
                           control the output of the speaker after playing the pre-recorded audio sample; and
                           execute the control data to modify the audio settings. 
- 49. The hearing device of clause 48, further comprising signal processing equipment
                     for processing the audio data into processed audio data.
- 50. The hearing device of clause 49, wherein the modifying the audio settings include
                     modifying the processing of the audio data.
- 51. The hearing device of clause 49, where the processing includes one or both of
                     noise filters and amplifiers.
- 52. The hearing device of clause 48, wherein the instructions for modifying the audio
                     settings include instructions for modifying the volume controls and/or instructions
                     for modifying a directionality of the microphone.
- 53. The hearing device of clause 48, wherein the control data is received from a computing
                     device or a mobile device.
- 54. The hearing device of clause 48, wherein the control data is received from manual
                     controls on the hearing device.
- 55. The hearing device of clause 48, wherein the control system is further configured
                     to play, by the speaker, multiple pre-recorded audio samples and receive control data
                     after playing each pre-recorded audio sample.
- 56. A method for tuning a hearing device comprising:
                     
                     playing a first pre-recorded audio sample from a speaker of a hearing device; receiving a first set of feedback data after playing the pre-recorded audio sample;
                        after receiving the first set of feedback data, playing a second pre-recorded audio
                        sample from the speaker; receiving a second set of feedback data after playing the second pre-recorded audio
                        sample; and configuring the audio settings of the hearing device based on at least the first and
                        second sets of feedback data. 
- 57. The method of clause 56, wherein the configuring the audio settings comprises
                     modifying one or more of noise filters, amplifiers, and volume.
- 58. The method of clause 56, wherein the first pre-recorded audio sample includes
                     sound in the human voice frequency range.
 
          
         
            
            1. A system for aggregation and processing of sound data relating to specific establishments
               comprising:
               
               
a database;
               
               a memory containing non-transitory machine readable medium and having stored thereon
                  machine executable code adapted to process sound data;
               
               a control system coupled to the memory, the control system configured to execute the
                  machine executable code to cause the control system to:
                  
                  
store in the database, in respect of a first user, hearing device audio settings with
                     an associated personal profile;
                  
                  receive at a server, from a first mobile device, a first set of audio data recorded
                     at a location by the first user;
                  
                  process by the server the first set of audio data in order to determine suggested
                     hearing device audio settings associated with the location based on comparison with
                     users that have similar characteristics in their profiles; and
                  
                  store the suggested hearing device audio settings in the database.
                 
            2. The system according to claim 1, wherein the control system is configured to recommend
               hearing device audio settings or ranges of hearing device audio settings to the first
               user based on the similar user.
 
            3. The system according to claim 1, wherein the control system is configured to receive
               at the server, from a second mobile device, a second set of audio data recorded at
               said location.
 
            4. The system according to claim 1, wherein the control system is configured to provide
               to the user options to select particular hearing device audio settings upon arriving
               at a known location based on the history of the user or of other users.
 
            5. The system according to claim 4, wherein the control system is configured to update
               dynamically the hearing device audio settings as the microphones of the user's hearing
               device or another user's microphones detect changes or updates in the acoustic environment.
 
            6. The system according to claim 1, wherein the control system is configured to provide
               to the user options to update and save his or her hearing device audio settings upon
               arriving at a known location, and to analyze these settings for progressive changes
               in a user's hearing over time.
 
            7. The system according to claim 1, wherein the user's personal profile includes individual
               hearing characteristics, such as a user's sensitivity to certain frequency ranges
               and amplitudes.
 
            8. The system according to claim 1, wherein the control system is configured to select
               the hearing device audio settings by utilizing matching, machine learning, statistical
               analysis, or other algorithms to match the hearing device audio settings from different
               users and profiles to one another.
 
            9. A method for aggregation and processing of sound data relating to specific establishments
               comprising:
               
               
storing in a database, in respect of a first user, hearing device audio settings with
                  an associated personal profile;
               
               receiving at a server, from a first mobile device, a first set of audio data recorded
                  at a location by the first user;
               
               processing by the server the first set of audio data in order to determine suggested
                  hearing device audio settings associated with the location based on comparison with
                  users that have similar characteristics in their profiles; and
               
               storing the suggested hearing device audio settings in the database.
  
            10. The method according to claim 9, comprising recommending hearing device audio settings
               or ranges of hearing device audio settings to the first user based on the similar
               user.
 
            11. The method according to claim 9, comprising receiving at the server, from a second
               mobile device, a second set of audio data recorded at said location.
 
            12. The method according to claim 9, comprising providing to the user options to select
               particular hearing device audio settings upon arriving at a known location based on
               the history of the user or of other users.
 
            13. The method according to claim 12, comprising dynamically updating the hearing device
               audio settings as the microphones of the user's hearing device or another user's microphones
               detect changes or updates in the acoustic environment.
 
            14. The method according to claim 9, comprising providing to the user options to update
               and save his or her hearing device audio settings upon arriving at a known location,
               and analyzing these settings for progressive changes in a user's hearing over time.
 
            15. The method according to claim 9, comprising selecting the hearing device audio settings
               by utilizing matching, machine learning, statistical analysis, or other algorithms
               to match the hearing device audio settings from different users and profiles to one
               another.