<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE ep-patent-document PUBLIC "-//EPO//EP PATENT DOCUMENT 1.7//EN" "ep-patent-document-v1-7.dtd">
<!-- This XML data has been generated under the supervision of the European Patent Office -->
<ep-patent-document id="EP23461601A1" file="EP23461601NWA1.xml" lang="en" country="EP" doc-number="4478362" kind="A1" date-publ="20241218" status="n" dtd-version="ep-patent-document-v1-7">
<SDOBI lang="en"><B000><eptags><B001EP>ATBECHDEDKESFRGBGRITLILUNLSEMCPTIESILTLVFIROMKCYALTRBGCZEEHUPLSKBAHRIS..MTNORSMESMMAKHTNMD..........</B001EP><B005EP>J</B005EP><B007EP>0009012-RPUB02</B007EP></eptags></B000><B100><B110>4478362</B110><B120><B121>EUROPEAN PATENT APPLICATION</B121></B120><B130>A1</B130><B140><date>20241218</date></B140><B190>EP</B190></B100><B200><B210>23461601.9</B210><B220><date>20230612</date></B220><B250>pl</B250><B251EP>en</B251EP><B260>en</B260></B200><B400><B405><date>20241218</date><bnum>202451</bnum></B405><B430><date>20241218</date><bnum>202451</bnum></B430></B400><B500><B510EP><classification-ipcr sequence="1"><text>G10L  25/03        20130101AFI20231103BHEP        </text></classification-ipcr><classification-ipcr sequence="2"><text>G10L  25/66        20130101ALI20231103BHEP        </text></classification-ipcr><classification-ipcr sequence="3"><text>G10L  25/75        20130101ALI20231103BHEP        </text></classification-ipcr></B510EP><B520EP><classifications-cpc><classification-cpc sequence="1"><text>G10L  25/03        20130101 FI20231026BHEP        </text></classification-cpc><classification-cpc sequence="2"><text>G10L  25/66        20130101 LI20231026BHEP        </text></classification-cpc><classification-cpc sequence="3"><text>G10L  25/75        20130101 LA20231026BHEP        </text></classification-cpc></classifications-cpc></B520EP><B540><B541>de</B541><B542>VERFAHREN ZUR ERKENNUNG DER CHARAKTERISTISCHEN MERKMALE DES KLANGFARBE AUF DER BASIS VON KLANGOBJEKTEN UND SYSTEM, COMPUTERPROGRAMM UND COMPUTERPROGRAMMPRODUKT DAFÜR</B542><B541>en</B541><B542>METHOD OF RECOGNITION OF THE CHARACTERISTIC FEATURES OF THE SOUND TIMBRE BASED ON SOUND OBJECTS AND SYSTEM, COMPUTER PROGRAM AND COMPUTER PROGRAM PRODUCT THEREFOR</B542><B541>fr</B541><B542>PROCÉDÉ DE RECONNAISSANCE DES CARACTÉRISTIQUES DU TIMBRE SONORE BASÉ SUR DES OBJETS SONORES ET SYSTÈME, PROGRAMME INFORMATIQUE ET PRODUIT DE PROGRAMME INFORMATIQUE ASSOCIÉS</B542></B540><B590><B598>1</B598></B590></B500><B700><B710><B711><snm>Vivid Mind PSA</snm><iid>101999212</iid><irf>P32023EP00/ABL</irf><adr><str>ul. Ksawerów 3</str><city>02-656 Warszawa</city><ctry>PL</ctry></adr></B711></B710><B720><B721><snm>Pluta, Adam</snm><adr><city>02-956 Warszawa</city><ctry>PL</ctry></adr></B721></B720><B740><B741><snm>Patpol Kancelaria Patentowa Sp. z o.o.</snm><iid>101432579</iid><adr><str>Nowoursynowska 162J</str><city>02-776 Warszawa</city><ctry>PL</ctry></adr></B741></B740></B700><B800><B840><ctry>AL</ctry><ctry>AT</ctry><ctry>BE</ctry><ctry>BG</ctry><ctry>CH</ctry><ctry>CY</ctry><ctry>CZ</ctry><ctry>DE</ctry><ctry>DK</ctry><ctry>EE</ctry><ctry>ES</ctry><ctry>FI</ctry><ctry>FR</ctry><ctry>GB</ctry><ctry>GR</ctry><ctry>HR</ctry><ctry>HU</ctry><ctry>IE</ctry><ctry>IS</ctry><ctry>IT</ctry><ctry>LI</ctry><ctry>LT</ctry><ctry>LU</ctry><ctry>LV</ctry><ctry>MC</ctry><ctry>ME</ctry><ctry>MK</ctry><ctry>MT</ctry><ctry>NL</ctry><ctry>NO</ctry><ctry>PL</ctry><ctry>PT</ctry><ctry>RO</ctry><ctry>RS</ctry><ctry>SE</ctry><ctry>SI</ctry><ctry>SK</ctry><ctry>SM</ctry><ctry>TR</ctry></B840><B844EP><B845EP><ctry>BA</ctry></B845EP></B844EP><B848EP><B849EP><ctry>KH</ctry></B849EP><B849EP><ctry>MA</ctry></B849EP><B849EP><ctry>MD</ctry></B849EP><B849EP><ctry>TN</ctry></B849EP></B848EP></B800></SDOBI>
<abstract id="abst" lang="en">
<p id="pa01" num="0001">The invention pertains to a method for recognizing the characteristics of the sound timbre based on the sound objects, in which statistical parameters are determined, after which the stages of separating sound objects into categories and determining at least one of: phase instability; the share of the energy of each category of objects in the total energy of the sound signal; harmonic frequency energy to noise ratio; amplitude slope coefficients; the slope of the graph of fundamental frequency energy and strong harmonics; local and global phase shift of each harmonic frequency; phase Fk average value, phase Fk drift; the number of sound objects and the percentage number of objects contained in the fundamental frequency and in all strong harmonic frequencies, follow.</p>
<p id="pa02" num="0002">The invention also pertains to a computer program and a computer program product comprising instructions that, when executed, perform the steps of the method for recognizing timbre characteristics based on the sound objects.
<img id="iaf01" file="imgaf001.tif" wi="78" he="100" img-content="drawing" img-format="tif"/></p>
</abstract>
<description id="desc" lang="en"><!-- EPO <DP n="1"> -->
<heading id="h0001"><b>Field of invention</b></heading>
<p id="p0001" num="0001">The invention relates to a method, system, computer program and computer program product for recognizing characteristic features of sound timbre (audio signal), for example sound abnormalities, based on sound objects. The invention is applicable in the field of analysis and synthesis of acoustic signals (also referred to as sound signals or audio signals), for example, for measuring characteristic parameters of sound timbre of a human voice or characteristic sound parameters of a device, for subsequent analysis of these parameters. Parameterization of the physical phenomenon of sound emission is important in such technical fields as acoustics, mechanics, construction, aviation, astronautics, geological measurements, and many others. It also plays a role in medicine, forensics, speaker identification and speech recognition.</p>
<heading id="h0002"><b>State of the art</b></heading>
<p id="p0002" num="0002">In recent years, intensive development of speech analysis systems has been observed. Many programs have been created for recording speech in the form of text (the so-called "Speech-to-Text Software") or for translating text from one language to another. Researchers admit that speech and voice are also very good markers of the speaker's health. Tools are being developed to measure speech parameters and, based on a comparison with stored patterns, to diagnose various disease states.</p>
<p id="p0003" num="0003">One of the most commonly used solutions is a direct analysis of an acoustic signal. An example of such an approach is patent application <patcit id="pcit0001" dnum="US15686057A"><text>US15686057A, published on March 1, 2018</text></patcit>, "Method for evaluating a quality of voice onset of a speaker", in which the assessment of health was made by measuring the parameters of the rate of change of energy in the fundamental component and harmonic components. This study is particularly suitable for the treatment of stuttering disorders.</p>
<p id="p0004" num="0004">In patent application <patcit id="pcit0002" dnum="CN202011522343A"><text>CN202011522343A, published on April 20, 2021</text></patcit>, "Voice state classification method, device, electronic device and storage medium", for measuring voice parameters such as fundamental frequency jitter, amplitude shimmer, noise-to-signal ratio, a<!-- EPO <DP n="2"> --> multidimensional voice processing program was applied. The advantage of the system over the patient's examination by a doctor is the objective assessment of voice parameters.</p>
<p id="p0005" num="0005">Some solutions look for voice biomarkers in spectral features and prosodic features. Melody, volume, tempo, accent, dynamics, and rhythmicity of the voice are assessed. In order to calculate these parameters, the spectrum of the acoustic signal should be determined using the Fourier Transform, or the parameters of the signal cepstrum presented in the mel scale (MFCC Mel-Frequency Cepstral Coefficients)).</p>
<p id="p0006" num="0006">An example of the use of this technology is patent application <patcit id="pcit0003" dnum="CN202110775311A"><text>CN202110775311A, published on November 12, 2021</text></patcit>, "A method, system and device for identifying depression based on voice analysis". This document reveals how a Digital Signal Processor with the help of an FPGA integrated circuit calculates the MFCC coefficients 1-12 and performs the classification using a decision tree algorithm.</p>
<p id="p0007" num="0007">The system presented in patent application <patcit id="pcit0004" dnum="CN202111677396A"><text>CN202111677396A, published on April 8, 2022</text></patcit>, "Voice analysis method and system of Parkinson's disease freezing gait symptom key characteristic parameter based on AdaBoost algorithm" works on a similar principle. In this method, the acoustic signal is analyzed using MFCC, and a CART algorithm was used to select the characteristic features. The program searches for the characteristics of Parkinson's disease in voice, which include slow speech, hoarseness, low volume, and vocal tremors.</p>
<p id="p0008" num="0008">Another solution to the problem of decomposing a signal into components, allowing to easily determine the parameters describing the acoustic signal, is presented in patent application <patcit id="pcit0005" dnum="EP16741938A"><text>EP16741938.1A, published on April 11, 2018</text></patcit>, "A method and a system for decomposition of acoustic signal into sound objects, a sound object and its use". This application presents the structure of the acoustic signal spectrum, wherein the structure is illustrated using a bank of zero-phase filters with a logarithmic frequency distribution and a resolution of 48 filters per octave. From the spectrum created in this way, sound objects were created with the use of linear prediction and phase continuity tracking algorithm. The resulting sound objects very precisely reflect frequency components present in the acoustic signal, without distortion of the information contained in them about amplitude, frequency and phase.</p>
<p id="p0009" num="0009">The above-mentioned solutions known in the art offer methods for analyzing the characteristics of audio signals, allowing, among other things, to identify abnormalities in audio signals. However, none of the solutions known from the state of art use in its analysis a sound<!-- EPO <DP n="3"> --> signal, recorded in the form of sound objects. Standard solutions, that do not use sound objects for analysis, require collection of large amounts of data from many samples to provide an accurate result. Analysis with the use of sound objects allows to reduce the number of samples necessary to be analyzed to obtain the appropriate parameters, allows the analysis of the sound timbre, shimmer and of the sound jitter. In addition, sound recording in the form of sound objects provides high frequency resolution of the recording, thanks to which it is possible to precisely measure the component signals of the analyzed sound. The solutions known in the art do not provide such precise information about the sound components , in particular about the harmonic structures of the signal and their mutual relations, especially phase relations.</p>
<heading id="h0003"><b>Summary</b></heading>
<p id="p0010" num="0010">A method of recognizing the characteristic features of the sound timbre based on sound objects, where a sound object is analyzed, wherein a sound object is a single acoustic signal with a slowly varying amplitude and a single slowly varying frequency and continuous phase, for which sound object statistical parameters are determined:
<ul id="ul0001" list-style="dash" compact="compact">
<li>for each time interval, local parameters are calculated for at least one sound object:<br/>
parameters of amplitude, frequency, number of measurement points;</li>
<li>global parameters are calculated from local parameters,</li>
</ul>
characterized in that, after calculating the statistical parameters of sound objects, the method comprises the steps of:
<ul id="ul0002" list-style="dash" compact="compact">
<li>separating the sound objects into categories, wherein the categories are formed depending on frequency, wherein at least one of the categories is dependent on fundamental frequency and its associated harmonic frequencies;</li>
<li>calculating categorized local and global parameters for at least one category of sound objects; and</li>
<li>determining at least one of:
<ul id="ul0003" list-style="none" compact="compact">
<li>∘ phase instability as the value of standard deviation of at least one instantaneous phase of the harmonic frequency of category of objects from at least one category from the phase of fundamental frequency,</li>
<li>∘ the share of energy of each category of objects in the total energy of the sound signal;</li>
<li>∘ the ratio of harmonic frequency energy to noise;<!-- EPO <DP n="4"> --></li>
<li>∘ local and global amplitude slope coefficients;</li>
<li>∘ directional slope coefficient of the energy graph for the fundamental frequency and strong harmonic ;</li>
<li>∘ a local and global phase shift of each harmonic frequency relative to the fundamental frequency at the moment, when the fundamental frequency phase is 0;</li>
<li>∘ the average value of phase Fk at the moment when the phase F1 of the fundamental frequency component has a given value; where k is the order number of the harmonic frequency;</li>
<li>∘ the phase Fk drift, at the moment when the phase F1 of the fundamental frequency component has a given value; where k is the order number of the harmonic frequency;</li>
<li>∘ the number of sound objects contained in the fundamental frequency and in all the strong harmonic frequencies;</li>
<li>∘ the percentage number of objects contained in the fundamental frequency and in all the strong harmonic frequencies.</li>
</ul></li>
</ul></p>
<p id="p0011" num="0011">Preferably, the value of the standard deviation of at least one instantaneous phase of harmonic frequency of the category of objects is calculated by the formula <maths id="math0001" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseStandDev</mi></mstyle><mo>=</mo><msqrt><msup><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>q</mi><mo>=</mo><mn>1</mn></mrow><mi>Q</mi></msubsup><mrow><msup><mfenced separators=""><mi mathvariant="italic">TabPhase</mi><mfenced open="[" close="]"><mi>q</mi></mfenced></mfenced><mn>2</mn></msup><mo>/</mo><mi>Q</mi><mo>−</mo><mfenced separators=""><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>q</mi><mo>=</mo><mn>1</mn></mrow><mi>Q</mi></msubsup><mrow><mi mathvariant="italic">TabPhase</mi><mfenced open="[" close="]"><mi>q</mi></mfenced></mrow></mstyle></mfenced><mo>/</mo><mi>Q</mi></mfenced></mrow></mstyle><mn>2</mn></msup></msqrt></math><img id="ib0001" file="imgb0001.tif" wi="159" he="15" img-content="math" img-format="tif"/></maths> where TabPhase[q] is the data for the q-th measurement stored in the table, and Q is the number of measurements.</p>
<p id="p0012" num="0012">Preferably, the global parameters are calculated as a weighted average of the local parameters.</p>
<p id="p0013" num="0013">Preferably, the energy of each sound object is calculated by the formula <maths id="math0002" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoEnergy</mi></mstyle><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">n</mi><mo>=</mo><mn>1</mn></mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoNumOfPoint</mi></mstyle></msubsup><msup><mfenced separators=""><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi mathvariant="bold-italic">a</mi></mfenced><mn>2</mn></msup></mstyle></math><img id="ib0002" file="imgb0002.tif" wi="71" he="11" img-content="math" img-format="tif"/></maths> where n is the ordinal number of the sound object point, and Sp<sub>n</sub>a is the amplitude of the sound object point for the n-th point.</p>
<p id="p0014" num="0014">Preferably, the amplitude slope coefficient is calculated by the formula <maths id="math0003" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmpLean</mi></mstyle><mo>=</mo><mfenced separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmpPos</mi></mstyle><mo>∗</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">Num</mi></mstyle><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmp</mi></mstyle><mo>∗</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumPos</mi></mstyle></mfenced><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">Delta</mi></mstyle><mo>;</mo></math><img id="ib0003" file="imgb0003.tif" wi="136" he="5" img-content="math" img-format="tif"/></maths> wherein <maths id="math0004" num=""><math display="block"><mi mathvariant="italic">Num</mi><mo>=</mo><mi mathvariant="italic">SoNumOfPoint</mi><mo>;</mo></math><img id="ib0004" file="imgb0004.tif" wi="47" he="5" img-content="math" img-format="tif"/></maths><!-- EPO <DP n="5"> --> <maths id="math0005" num=""><math display="block"><mi mathvariant="italic">SumPos</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mrow></mstyle><mo>;</mo></math><img id="ib0005" file="imgb0005.tif" wi="44" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0006" num=""><math display="block"><mi mathvariant="italic">SumSquarePos</mi><mo>=</mo><msup><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mfenced separators=""><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mfenced></mstyle><mn>2</mn></msup><mo>;</mo></math><img id="ib0006" file="imgb0006.tif" wi="61" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0007" num=""><math display="block"><mi mathvariant="italic">Delta</mi><mo>=</mo><mi mathvariant="italic">SumSquarePos</mi><mo>∗</mo><mi mathvariant="italic">Num</mi><mo>−</mo><mi mathvariant="italic">SumPos</mi><mo>∗</mo><mi mathvariant="italic">SumPos</mi><mo>;</mo></math><img id="ib0007" file="imgb0007.tif" wi="98" he="5" img-content="math" img-format="tif"/></maths> <maths id="math0008" num=""><math display="block"><mi mathvariant="italic">SumAmp</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>a</mi></mrow></mstyle><mo>;</mo></math><img id="ib0008" file="imgb0008.tif" wi="45" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0009" num=""><math display="block"><mi mathvariant="italic">SumAmpPos</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>a</mi><mo>∗</mo><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mrow></mstyle></math><img id="ib0009" file="imgb0009.tif" wi="65" he="11" img-content="math" img-format="tif"/></maths> where n is the ordinal number of the sound object point, SoNumOfPoint is the number of points, Sp<sub>n</sub>a is the amplitude of the sound object point, and Sp<sub>n</sub>p is the position of the sound object point.</p>
<p id="p0015" num="0015">Preferably, the average phase Fk value is calculated from the formula: <maths id="math0010" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseAvarage</mi></mstyle><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced></mrow></mstyle></mfenced><mo>/</mo><mi mathvariant="bold-italic">Q</mi><mo>;</mo></math><img id="ib0010" file="imgb0010.tif" wi="99" he="12" img-content="math" img-format="tif"/></maths> where TabPhase[q] is the data for the q-th measurement stored in the table.</p>
<p id="p0016" num="0016">Preferably, the phase drift is calculated by the formula: <maths id="math0011" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseDrift</mi></mstyle><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>2</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mfenced open="|" close="|" separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]" separators=""><mi mathvariant="bold-italic">q</mi><mo>−</mo><mn>1</mn></mfenced></mfenced></mstyle></mfenced><mo>/</mo><mfenced separators=""><mi mathvariant="bold-italic">Q</mi><mo>−</mo><mn>1</mn></mfenced><mo>,</mo></math><img id="ib0011" file="imgb0011.tif" wi="145" he="12" img-content="math" img-format="tif"/></maths> where TabPhase[q] is the data for the q-th measurement stored in the table.</p>
<p id="p0017" num="0017">Preferably, the sound objects are divided into at least the following categories:
<ul id="ul0004" list-style="dash" compact="compact">
<li>fundamental frequency category;</li>
<li>the category of strong harmonic frequencies, in which the share of these harmonics in the signal energy is more than 5%;</li>
<li>a category of weak harmonic frequencies, in which the share of these harmonics in the signal energy is below 5%;</li>
<li>category of non-harmonic signals of single frequency and amplitude;</li>
<li>low frequency noise category, below 200 Hz;</li>
<li>medium frequency noise category, i.e., in the range from 200 Hz to 2000 Hz;</li>
<li>high-frequency noise category, i.e., in the range from 2,000 Hz to 10,000 Hz.</li>
</ul></p>
<p id="p0018" num="0018">Preferably, after the parameter determination step, all parameters are stored and set of training input data is created for the artificial intelligence neural network.</p>
<p id="p0019" num="0019">Preferably, the time interval is 200 ms.<!-- EPO <DP n="6"> --></p>
<p id="p0020" num="0020">Preferably, the phase of the fundamental frequency component has a predetermined value.</p>
<p id="p0021" num="0021">Preferably, the phase of the fundamental frequency component is equal to 0.</p>
<p id="p0022" num="0022">A system for recognizing characteristic features of sound timbre based on sound objects, where a sound object is analyzed, wherein the sound object is a single acoustic signal with a slowly varying amplitude and a single slowly varying frequency and a continuous phase. The system is characterized by comprising:
<ul id="ul0005" list-style="dash" compact="compact">
<li>recording module adapted to record sounds;</li>
<li>a processing module adapted to process the recorded sound so that the sound object is extracted;</li>
<li>a computing module adapted to perform the above method of recognizing characteristic features of sound timbre based on the sound objects;</li>
<li>a classifying module adapted to analyze the sound object based on the results obtained from the computing module.</li>
</ul></p>
<p id="p0023" num="0023">A computer program for recognizing characteristic features of sound timbre based on sound objects, comprising instructions for performing the above method of identifying characteristic features of sound timbre based on sound objects.</p>
<p id="p0024" num="0024">A computer program product for recognizing characteristic features of sound timbre based on sound objects comprising a computer-readable code performing the steps of the above method for recognizing characteristic features of sound timbre based on sound objects.</p>
<heading id="h0004"><b>Brief Description of the Drawings</b></heading>
<p id="p0025" num="0025">The invention has been shown with reference to the figures on which:
<ul id="ul0006" list-style="none" compact="compact">
<li><figref idref="f0001">Fig. 1</figref> shows a block diagram of the analyzed person's speech processing system;</li>
<li><figref idref="f0002">Fig. 2</figref> is a graph of the audio signal contained in the sound object;</li>
<li><figref idref="f0003">Fig. 3</figref> - examples of presentation of acoustic signals using sound objects;</li>
<li><figref idref="f0004">Fig. 4</figref> - amplitude parameters for sound object;</li>
<li><figref idref="f0005">Fig. 5</figref> - frequency parameters for sound object;</li>
<li><figref idref="f0006">Fig. 6</figref> - phase parameters for sound object;</li>
<li><figref idref="f0007">Fig. 7</figref> - parameters of harmonic relations.</li>
</ul><!-- EPO <DP n="7"> --></p>
<heading id="h0005"><b>Description of embodiments</b></heading>
<p id="p0026" num="0026">An exemplary application of the method according to the invention is presented below.</p>
<p id="p0027" num="0027"><figref idref="f0001">Fig. 1</figref> shows a block diagram presenting the operation of the method according to the invention as well as the sequence of processes taking place in the modules of the system performing the method according to the invention.</p>
<p id="p0028" num="0028">According to the present method of recognizing the characteristic features of sound timbre, as shown in <figref idref="f0001">Fig. 1</figref>, using the appropriate element of the recording module (A), for example, the Mobile App (10), which can be installed on a smartphone, portable computer, notebook or other appropriate electronic device of the user, the sound is recorded in the form of a digital acoustic signal and then sent as a file without compression to the processing module (B), adapted to process the recorded sound.</p>
<p id="p0029" num="0029">The digital audio signal is stored in the .wav file in the form of a series of electrical values of the microphone signal at fixed time intervals, i.e., the so-called sampling frequency, wherein for human speech, it is sufficient to use a value of 22 050 samples per second.</p>
<p id="p0030" num="0030">In the processing module (B), the sent file goes to the Internet Network Communication Module (11), where the licensing rights of the client ordering acoustic signal processing and the data concerning the file containing information regarding what the recording is about and what file processing is to be performed are verified, all this information together can be referred to as an "order". If the client is authorized to send the file and the system is prepared to realize such processing, the authorization verification is completed correctly, the file is transferred to the audio file archive (12) and saved there.</p>
<p id="p0031" num="0031">In parallel, a new order entry is created in the database (13), for example PostgreSQL, in which order, the processing parameters are saved, including the date, time, ordering party's identifier and the type of ordered processing and, as the order is processed, the following will be added: the location of the audio file, the location of the file containing the designated sound objects (single acoustic signals with a slowly varying amplitude and a single slowly varying frequency and continuous phase), global and local parameters describing the acoustic signal, classification assessment, time of processing completion and the content of the final report sent to the client.</p>
<p id="p0032" num="0032">The audio file archive (12) and the database (13) are asynchronously connected to the process control module (14), which, after the steps of saving the file in the audio file archive (12) and creating an order record in the database (13), successively activates the computing module (C),<!-- EPO <DP n="8"> --> namely sound file conversion module (15), in which at least one sound object is separated, and sound object analysis module (16) and finally the classifying module (D), also called the measurement result classifier module (17).</p>
<p id="p0033" num="0033">The sound file conversion module (15) decomposes the file, which is recorded in .wav format, from the sound file archive (12) according to the method disclosed in the application <patcit id="pcit0006" dnum="EP3304549A1"><text>EP 3304549 A1</text></patcit> to create at least one sound object, and then saves this at least one sound object in the appropriate .uho format in the sound file archive (12). Information about the sampling frequency with which the signal was recorded is also sent to the .uho file. Typically, this information is in the header of the .wav file and is necessary for the correct determination of component frequencies.</p>
<p id="p0034" num="0034">The computing module (C) of <figref idref="f0001">Fig. 1</figref> recognizes and is able to perform operations on various .uho recording formats, including UH00 as shown in the aforementioned patent document <patcit id="pcit0007" dnum="EP3304549A1"><text>EP 3304549 A1</text></patcit>.</p>
<p id="p0035" num="0035">In the next stage of the method of recognizing the characteristic features of sound timbre, the sound object analysis module (16) reads the indicated .uho file from the sound file archive (12), and then determines, for at least one sound object stored in the archive and the harmonic structures of this at least one sound object, the statistical parameters of this object, specifically local parameters: amplitude parameters, frequencies, the number of measurement points in the time interval and the length of the time interval, as well as global parameters, wherein the local parameters are calculated for time intervals and then grouped into categories according to the method of the invention, as will be described in more detail below, wherein time intervals are sections into which the sound signal recorded by the module (10) is divided. The sections may be, for example, 200 ms long (4410 samples of the acoustic signal) and local parameters are calculated for each of these sections. For example, for a 10-second recording, there are 50 time intervals, and the number of sound objects can range from 5,000 objects in the case of a low-quality recording to over 50,000 when the recording is of good quality.</p>
<p id="p0036" num="0036">The determined parameters and categories are stored in a table, which is then stored in the audio file archive (12) and, in addition, a pointer is created to the table stored in the archive (12), which is then stored in the database (13).</p>
<p id="p0037" num="0037">After module (16) conducts the analysis and creates tables stored in the archive (12), the classification module (D) reads the parameters stored in the table and compares them using<!-- EPO <DP n="9"> --> artificial intelligence algorithms with data obtained from other authorized measurements, issuing a classification assessment, which is described below. The assessment will be saved in the file archive (12), after which the result of this comparison, together with the report on the execution of the order, will be sent to the user's electronic device.</p>
<p id="p0038" num="0038">Turning now to a detailed description of the method for recognizing characteristic features of sound timbre based on sound objects, the individual steps of the method are presented below.</p>
<p id="p0039" num="0039">To facilitate understanding of the equations used in the process steps, key terms of the equations that are used therein are shown below:
<ul id="ul0007" list-style="none" compact="compact">
<li>FileHeader -file marker;</li>
<li>NumOfTrack - number of tracks;</li>
<li>SampleFreq - sampling frequency;</li>
<li>TrackStat - track status;</li>
<li>TrackNum - track number;</li>
<li>TrackNumOfSo - number of sound objects;</li>
<li>TrackPos - track position;</li>
<li>SoNumOfPoint - number of points in the sound object;</li>
<li>SoStartPhase - initial phase in the sound object;</li>
<li>SoStartPos - start position of the sound object in relation to the TrackPos position;</li>
<li>SpAmp - point amplitude;</li>
<li>SpTon - point tone;</li>
<li>SpDeltaPos - distance from the previous point;</li>
<li>k - harmonic frequency number</li>
<li>n - ordinal number of the sound object point;</li>
<li>m - ordinal number of the acoustic signal sample;</li>
<li>Sp<sub>n</sub> - Point number n of the analyzed sound object;</li>
<li>Sp<sub>n</sub>p - Position of the sound object point;</li>
<li>Sp<sub>n</sub>f - Frequency of the sound object point;</li>
<li>Sp<sub>n</sub>ω - Pulsation of the sound object point;</li>
<li>Sp<sub>n</sub>ϕ- Phase of the sound object point;</li>
<li>Amp<sub>m</sub> -Amplitude of the sample m of the audio signal;</li>
<li>Φ<sub>m</sub> - Phase of the sample m of the audio signal.</li>
</ul><!-- EPO <DP n="10"> --></p>
<p id="p0040" num="0040">In the method, after the sound is recorded, saved as a digital audio signal (as a file), the file is decomposed according to the method disclosed in <patcit id="pcit0008" dnum="EP3304549A1"><text>application EP 3304549A1</text></patcit> and at least one sound object is thus obtained, further, local and global statistical parameters are determined for such a sound object.</p>
<p id="p0041" num="0041">An exemplary sound object, for example extracted by the sound file conversion module (15), is shown in <figref idref="f0002">Fig. 2</figref>.</p>
<p id="p0042" num="0042"><figref idref="f0002">Fig. 2</figref>, illustrates local parameters, determined in present method, calculated for time intervals, for each nth (n { 1... 5}) point of the analyzed sound object, in accordance with the following formulas:
<tables id="tabl0001" num="0001">
<table frame="none">
<tgroup cols="2" colsep="0" rowsep="0">
<colspec colnum="1" colname="col1" colwidth="60mm"/>
<colspec colnum="2" colname="col2" colwidth="50mm"/>
<tbody>
<row>
<entry>Sp<sub>n</sub>p = Sp<sub>n-1</sub>p + Sp<sub>n</sub>DeltaPos</entry>
<entry>Point Position [Sample number]</entry></row>
<row>
<entry>Sp<sub>n</sub>a = Sp<sub>n</sub>Amp</entry>
<entry>Point Amplitude</entry></row>
<row>
<entry>Sp<sub>n</sub>f = 16.3516 <sup>∗</sup> 2 ^ Sp<sub>n</sub>Ton/(32<sup>∗</sup>48)</entry>
<entry>Point Frequency [Hz]</entry></row>
<row>
<entry>Sp<sub>n</sub>ω = 2 <sup>∗</sup> π <sup>∗</sup> Sp<sub>n</sub>f / SampleFreq</entry>
<entry>Point Pulsation [rad/sample]</entry></row>
<row>
<entry>Sp<sub>n</sub>ϕ = Sp<sub>n-1</sub>ϕ + Sp<sub>n</sub>DeltaPoz <sup>∗</sup> Sp<sub>n</sub>ω</entry>
<entry>Point Phase [rad]</entry></row></tbody></tgroup>
</table>
</tables></p>
<p id="p0043" num="0043">More specifically, <figref idref="f0002">Fig. 2</figref> shows an audio signal contained in sound object, where a graph of the amplitude, frequency and phase parameters of the determined measurement points is marked point by point and a line section is used to show interpolated parameters for the signal samples between the points.<br/>
The audio signal graph is defined by the following formulas: <maths id="math0012" num=""><math display="block"><msub><mi mathvariant="italic">Audio</mi><mi>m</mi></msub><mo>=</mo><msub><mi mathvariant="italic">Amp</mi><mi>m</mi></msub><mo>∗</mo><mi>cos</mi><mfenced><msub><mi>φ</mi><mi>m</mi></msub></mfenced></math><img id="ib0012" file="imgb0012.tif" wi="51" he="5" img-content="math" img-format="tif"/></maths></p>
<p id="p0044" num="0044">For every n-th point n { 1 ... 4 }<br/>
and for each m-th sample m { Sp<sub>n</sub>p ... Sp<sub>n+1</sub>p }, where: <maths id="math0013" num=""><math display="block"><msub><mi mathvariant="italic">Amp</mi><mi>m</mi></msub><mo>=</mo><mfenced separators=""><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>a</mi><mo>∗</mo><mfenced separators=""><msub><mi mathvariant="italic">Sp</mi><mrow><mi>n</mi><mo>+</mo><mn>1</mn></mrow></msub><mi>p</mi><mo>−</mo><mi>m</mi></mfenced><mo>+</mo><msub><mi mathvariant="italic">Sp</mi><mrow><mi>n</mi><mo>+</mo><mn>1</mn></mrow></msub><mi>a</mi><mo>∗</mo><mfenced separators=""><mi>m</mi><mo>−</mo><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mfenced></mfenced><mo>/</mo><msub><mi mathvariant="italic">Sp</mi><mrow><mi>n</mi><mo>+</mo><mn>1</mn></mrow></msub><mi mathvariant="italic">DeltaPoz</mi></math><img id="ib0013" file="imgb0013.tif" wi="132" he="5" img-content="math" img-format="tif"/></maths> <maths id="math0014" num=""><math display="block"><msub><mi>φ</mi><mi>m</mi></msub><mo>=</mo><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi mathvariant="italic">aφ</mi><mo>+</mo><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>ω</mi><mo>∗</mo><mfenced separators=""><mi>m</mi><mo>−</mo><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mfenced></math><img id="ib0014" file="imgb0014.tif" wi="65" he="5" img-content="math" img-format="tif"/></maths></p>
<p id="p0045" num="0045">The accuracy of mapping of the analyzed waveform of recorded audio signal with the above-described audio objects (calculated sample to sample) reaches 99.5%.</p>
<p id="p0046" num="0046">Described above acoustic (sound) signals, are recorded with the use of sound objects after the decomposition, as shown in <figref idref="f0003">Fig. 3</figref>, where three exemplary acoustic signals are illustrated:
<ul id="ul0008" list-style="dash" compact="compact">
<li>recording number 36 - a signal with weak vibrations, which is a sound signal coming from a person without disease symptoms;<!-- EPO <DP n="11"> --></li>
<li>recording number 65 - signal with medium vibrations, which is a sound signal coming from a person suspected of having moderate dementia aliments;</li>
<li>recording number 08 - a signal with strong vibrations, which is a sound signal coming from a person diagnosed with large dementia lesions.</li>
</ul></p>
<p id="p0047" num="0047">In <figref idref="f0003">Fig. 3</figref>, each of the signals is presented as a graph of the audio signal over time (upper graph of each recording) and as a representation of this audio signal in the form of sound objects representing harmonic structures (the graph on the left for each recording) and non-harmonic structures and noise respectively (the graph on the right for each recording).</p>
<p id="p0048" num="0048">In an embodiment, after calculation of local parameters, it is possible to proceed to obtaining global parameters. These global parameters are calculated for the entire length of the file (for all time intervals) as weighted averages of the corresponding local parameters for all objects.</p>
<p id="p0049" num="0049">Next, after calculating the statistical parameters (local and global) of the sound objects, the sound objects are categorized according to frequency, using the parameters calculated during the previous step.</p>
<p id="p0050" num="0050">The categories into which objects are grouped are:
<ul id="ul0009" list-style="dash" compact="compact">
<li>the category of fundamental frequency - containing objects of fundamental frequency;</li>
<li>the category of strong harmonic frequencies (low harmonics) - harmonic frequencies whose energy is greater than 5% of the energy of the entire signal;</li>
<li>the category of weak harmonic frequencies (high harmonics) - containing objects with other weaker harmonic frequencies, including up to 24 frequencies;</li>
<li>the category of non-harmonic signals - containing objects that are not harmonic in relation to the fundamental frequency, the energy of which exceeds 1% of the energy of the entire signal;</li>
<li>the category of Low Noise (category of low-frequency noise) - containing objects with non-harmonic frequencies, the average frequency of which is lower than 200 Hz;</li>
<li>the category of Medium Noise (category of medium-frequency noise) - containing objects with non-harmonic frequencies, the average frequency of which is between 200 Hz and 2 000 Hz;</li>
<li>the category of High Noise (category of high-frequency noise) - containing objects with non-harmonic frequencies, in the range from 2,000 Hz to 10,000 Hz.</li>
</ul><!-- EPO <DP n="12"> --></p>
<p id="p0051" num="0051">To reduce the number of calculations, for the first stage of the analysis one may, for example, select only objects whose frequency is less than 2,000 Hz and whose energy is not less than 1% of the energy of the entire recording, or whose length exceeds 4410 samples (200 ms).</p>
<p id="p0052" num="0052">For these several dozen sound objects, the frequencies of all pairs of objects are compared with each other (each object is successively paired with other objects), and the group of objects, that have a common divisor and the sum of their energies is the greatest, is selected. This group of objects creates a harmonic structure. The smallest divisor defines the fundamental frequency. It may happen that no object belongs to the fundamental frequency. Objects that have not been qualified to the harmonic structure will remain so-called non-harmonic objects. The objects that make up the harmonic structure will be assigned to one of the 24 harmonics. Since the distances between harmonics higher than the 24th harmonic are too small to be analyzed individually, objects with higher frequencies will be treated as noise. After separating strong objects among harmonics, weak objects will be assigned to harmonics, whose frequency coincides with the objects' harmonic frequency. Noise objects not assigned to harmonics can be divided into Low Noise, Medium Noise and High Noise.</p>
<p id="p0053" num="0053">The categorization process described above yields 24 harmonic groups, groups of non-harmonic signals with energy above 1 % of the total signal energy, and 3 types of noise based on statistical parameters present in the group of sound objects, for categories containing categorized sound objects, categorized global and local parameters are then determined. From the 24 harmonic groups, the Strong Harmonics Group is determined, which includes the fundamental frequency and those harmonics whose total energy is not less than 5% of the total energy of the recording, as well as the Weak Harmonics Group, which includes the remaining harmonics.</p>
<p id="p0054" num="0054">As in the case of statistical parameters, categorized local parameters are determined for each of the time intervals (sections) and global parameters are determined for the entire recording. The calculation of the categorized parameters for the respective categories is shown below.</p>
<p id="p0055" num="0055">In one embodiment, additionally in the method for recognizing characteristic features of the sound timbre, for the category of Low Harmonics it is possible to determine following, illustrated on the amplitude graphs of sound objects of <figref idref="f0004">Fig. 4</figref>, categorized local paramethers according to the presented formulas:<!-- EPO <DP n="13"> -->
<ol id="ol0001" compact="compact" ol-style="">
<li>a) Average value of the object amplitude - equal to the value of the sum of the amplitudes of all points of the sound object divided by the number of points. <maths id="math0015" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpAverage</mi></mstyle><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">n</mi><mo>=</mo><mn>1</mn></mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoNumOfPoint</mi></mstyle></msubsup><mrow><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi mathvariant="bold-italic">a</mi></mrow></mstyle></mfenced><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoNumOfPoint</mi></mstyle></math><img id="ib0015" file="imgb0015.tif" wi="113" he="11" img-content="math" img-format="tif"/></maths></li>
<li>b) Energy of the object - equal to the sum of the squares of the amplitudes of all points of the objects. <maths id="math0016" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoEnergy</mi></mstyle><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">n</mi><mo>=</mo><mn>1</mn></mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoNumOfPoint</mi></mstyle></msubsup><msup><mfenced separators=""><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi mathvariant="bold-italic">a</mi></mfenced><mn>2</mn></msup></mstyle></math><img id="ib0016" file="imgb0016.tif" wi="71" he="11" img-content="math" img-format="tif"/></maths></li>
<li>c) Percent value of the standard deviation of the amplitude - equal to the quotient of the square root of the difference between the object energy divided by the number of points and the square of the average amplitude value and the average object amplitude value. <maths id="math0017" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpSDProc</mi></mstyle><mo>=</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpStandDev</mi></mstyle><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpAvarage</mi></mstyle><mo>;</mo></math><img id="ib0017" file="imgb0017.tif" wi="105" he="5" img-content="math" img-format="tif"/></maths> wherein <maths id="math0018" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpStandDev</mi></mstyle><mo>=</mo><msqrt><mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoEnergy</mi></mstyle><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoNumOfPoint</mi></mstyle><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpAvarag</mi></mstyle><msup><mi mathvariant="bold-italic">e</mi><mn>2</mn></msup></mrow></msqrt></math><img id="ib0018" file="imgb0018.tif" wi="137" he="7" img-content="math" img-format="tif"/></maths></li>
<li>d) Percentage value of the instantaneous change in amplitude, i.e., shimmer - equal to the quotient of the quotient of the sum of the modulus of the difference in the amplitude of the adjacent points and the number of intervals and the value of the average amplitude of the object. <maths id="math0019" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpSimmerProc</mi></mstyle><mo>=</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpSimmer</mi></mstyle><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoAmpAverage</mi></mstyle><mo>;</mo></math><img id="ib0019" file="imgb0019.tif" wi="111" he="5" img-content="math" img-format="tif"/></maths> wherein <maths id="math0020" num=""><math display="block"><mi mathvariant="italic">SoAmpSimmer</mi><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>2</mn></mrow><mi mathvariant="italic">SoNumOfPoint</mi></msubsup><mfenced open="|" close="|" separators=""><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>a</mi><mo>−</mo><msub><mi mathvariant="italic">Sp</mi><mrow><mi>n</mi><mo>−</mo><mn>1</mn></mrow></msub><mi>a</mi></mfenced></mstyle></mfenced><mo>/</mo><mfenced separators=""><mi mathvariant="italic">SoNumOfPoint</mi><mo>−</mo><mn>1</mn></mfenced></math><img id="ib0020" file="imgb0020.tif" wi="138" he="11" img-content="math" img-format="tif"/></maths></li>
<li>e) Coefficient of the slope of the amplitude representing the minimum root mean square error of the amplitude. <maths id="math0021" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmpLean</mi></mstyle><mo>=</mo><mfenced separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmpPos</mi></mstyle><mo>∗</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">Num</mi></mstyle><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmp</mi></mstyle><mo>∗</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumPos</mi></mstyle></mfenced><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">Delta</mi></mstyle><mo>;</mo></math><img id="ib0021" file="imgb0021.tif" wi="136" he="5" img-content="math" img-format="tif"/></maths> wherein <maths id="math0022" num=""><math display="block"><mi mathvariant="italic">Num</mi><mo>=</mo><mi mathvariant="italic">SoNumOfPoint</mi><mo>;</mo></math><img id="ib0022" file="imgb0022.tif" wi="47" he="5" img-content="math" img-format="tif"/></maths><!-- EPO <DP n="14"> --> <maths id="math0023" num=""><math display="block"><mi mathvariant="italic">SumPos</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi><mo>;</mo></mrow></mstyle></math><img id="ib0023" file="imgb0023.tif" wi="44" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0024" num=""><math display="block"><mi mathvariant="italic">SumSquarePos</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msup><mfenced separators=""><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mfenced><mn>2</mn></msup><mo>;</mo></mrow></mstyle></math><img id="ib0024" file="imgb0024.tif" wi="62" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0025" num=""><math display="block"><mi mathvariant="italic">Delta</mi><mo>=</mo><mi mathvariant="italic">SumSquarePos</mi><mo>∗</mo><mi mathvariant="italic">Num</mi><mo>−</mo><mi mathvariant="italic">SumPos</mi><mo>∗</mo><mi mathvariant="italic">SumPos</mi><mo>;</mo></math><img id="ib0025" file="imgb0025.tif" wi="98" he="5" img-content="math" img-format="tif"/></maths> <maths id="math0026" num=""><math display="block"><mi mathvariant="italic">SumAmp</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>a</mi><mo>;</mo></mrow></mstyle></math><img id="ib0026" file="imgb0026.tif" wi="45" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0027" num=""><math display="block"><mi mathvariant="italic">SumAmpPos</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>a</mi><mo>∗</mo><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi><mo>.</mo></mrow></mstyle></math><img id="ib0027" file="imgb0027.tif" wi="67" he="11" img-content="math" img-format="tif"/></maths></li>
</ol></p>
<p id="p0056" num="0056">In another embodiment, additionally in the method for recognizing characteristic features of sound timbre, for the category of Low Harmonics it is possible to determine following, illustrated on the frequency graphs of sound objects of <figref idref="f0005">Fig. 5</figref>, categorized global parameters according to the presented formulas:
<ol id="ol0002" compact="compact" ol-style="">
<li>a) The average value of the frequency of the object - the sum of the frequency values of all points of the sound object divided by the number of points. <maths id="math0028" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoFreqAverage</mi></mstyle><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">n</mi><mo>=</mo><mn>1</mn></mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoNumOfPoint</mi></mstyle></msubsup><mrow><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi mathvariant="bold-italic">f</mi></mrow></mstyle></mfenced><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoNumOfPoint</mi></mstyle></math><img id="ib0028" file="imgb0028.tif" wi="112" he="11" img-content="math" img-format="tif"/></maths></li>
<li>b) Percent value of the frequency standard deviation - equal to the quotient of the square root of the difference between the sum of the squares of the frequency of points divided by the number of points and the square of the average frequency value and the average frequency value of the object. <maths id="math0029" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoFreqSDProc</mi></mstyle><mo>=</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoFreqStandDe</mi></mstyle><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoFreqAvarage</mi></mstyle></math><img id="ib0029" file="imgb0029.tif" wi="102" he="5" img-content="math" img-format="tif"/></maths> wherein <maths id="math0030" num=""><math display="block"><mi mathvariant="italic">SoFreqSquared</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">SoNumOfPoint</mi></msubsup><msup><mfenced separators=""><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>f</mi></mfenced><mn>2</mn></msup></mstyle></math><img id="ib0030" file="imgb0030.tif" wi="79" he="11" img-content="math" img-format="tif"/></maths> <maths id="math0031" num=""><math display="block"><mi mathvariant="italic">SoFreqStandDe</mi><mo>=</mo><msqrt><mrow><mi mathvariant="italic">SoFreqSquared</mi><mo>/</mo><mi mathvariant="italic">SoNumOfPoint</mi><mo>−</mo><msup><mi mathvariant="italic">SoFreqAvarage</mi><mn>2</mn></msup></mrow></msqrt></math><img id="ib0031" file="imgb0031.tif" wi="137" he="7" img-content="math" img-format="tif"/></maths><!-- EPO <DP n="15"> --></li>
<li>c) The percentage value of the instantaneous frequency change, i.e., jitter - equal to the quotient of the quotient of the sum of the modules of the difference in the frequency values of the adjacent points and the number of intervals, and the value of the average frequency of the object. <maths id="math0032" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoFreqJitterProc</mi></mstyle><mo>=</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoFreqJitter</mi></mstyle><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoFreqAvarage</mi></mstyle></math><img id="ib0032" file="imgb0032.tif" wi="102" he="5" img-content="math" img-format="tif"/></maths> <i>wherein</i> <maths id="math0033" num=""><math display="block"><mi mathvariant="italic">SoFreqJitter</mi><mo>=</mo><mfrac bevelled="true"><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>2</mn></mrow><mi mathvariant="italic">SoNumOfPoint</mi></msubsup><mfenced open="|" close="|" separators=""><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>f</mi><mo>−</mo><msub><mi mathvariant="italic">Sp</mi><mrow><mi>n</mi><mo>−</mo><mn>1</mn></mrow></msub><mi>f</mi></mfenced></mstyle></mfenced><mfenced separators=""><mi mathvariant="italic">SoNumOfPoint</mi><mo>−</mo><mn>1</mn></mfenced></mfrac></math><img id="ib0033" file="imgb0033.tif" wi="132" he="11" img-content="math" img-format="tif"/></maths></li>
<li>d) Coefficient of the slope of the frequency representing the minimum root mean square error of the frequency. <maths id="math0034" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumFreqLean</mi></mstyle><mo>=</mo><mfenced separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumFreqPos</mi></mstyle><mo>∗</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">Num</mi></mstyle><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumFreq</mi></mstyle><mo>∗</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumPos</mi></mstyle></mfenced><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">Delta</mi></mstyle></math><img id="ib0034" file="imgb0034.tif" wi="136" he="5" img-content="math" img-format="tif"/></maths> <i>wherein</i> <maths id="math0035" num=""><math display="block"><mi mathvariant="italic">SumFreq</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub></mstyle><mi>f</mi><mo>;</mo></math><img id="ib0035" file="imgb0035.tif" wi="46" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0036" num=""><math display="block"><mi mathvariant="italic">SumFreqPos</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>f</mi><mo>∗</mo><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mrow></mstyle><mo>.</mo></math><img id="ib0036" file="imgb0036.tif" wi="66" he="6" img-content="math" img-format="tif"/></maths></li>
</ol></p>
<p id="p0057" num="0057">The characteristic features of sound timbre recorded in the form of parameters can be used to determine the relationships between the various parameters of the signal, wherein the phase-related parameters are the preferred and most informative. Phase measurement is a very precise tool that allows, for example, to assess the quality of the speaker's voice path control. It can only be performed after properly extracting the actual harmonic frequencies present in the voice. The use of the transform for voice analysis does not give this possibility, because the harmonic frequencies of the transform do not correspond to the real harmonic frequencies. Thanks to the use of sound objects created by the Bank of Zero-Phase Filters using the phase continuity algorithm described in <patcit id="pcit0009" dnum="EP3304549A1"><text>EP 3304549 A1</text></patcit>, the determined harmonics correctly measure the phase of the harmonic components of the acoustic signal.</p>
<p id="p0058" num="0058">Characteristic features with respect to the parameters related to the phases are shown in <figref idref="f0006">Fig. 6</figref>. To determine the phase relations, it is necessary to calculate the categorized parameters analogous to those calculated above (embodiments with reference to <figref idref="f0004">Figs. 4</figref> and <figref idref="f0005">5</figref>) and determine the phase instability. To determine this instability, at least one instantaneous phase of the harmonic<!-- EPO <DP n="16"> --> frequency from at least one category is compared to the phase of the fundamental frequency and the value of the standard deviation between these phases is determined.</p>
<p id="p0059" num="0059">If only a virtual fundamental frequency is determined in the signal, no phase measurement will be performed.</p>
<p id="p0060" num="0060">Preferably, the phase of the fundamental frequency component has a predetermined value and may be equal to 0.</p>
<p id="p0061" num="0061">In an embodiment, the measurement is performed in 200 ms sections when the phase of the fundamental frequency is 0 radians. Points in the sound object for each sound object are determined every two frequency periods of the sound object. Between two successive fundamental frequency points, there will be at least one point where the phase of the frequency will be zero. The positions of this point can be calculated from the following formula: <maths id="math0037" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">FundZeroPos</mi></mstyle><mo>=</mo><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi mathvariant="bold-italic">p</mi><mo>−</mo><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi>φ</mi><mo>/</mo><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi>ω</mi></math><img id="ib0037" file="imgb0037.tif" wi="73" he="5" img-content="math" img-format="tif"/></maths></p>
<p id="p0062" num="0062">Method for calculating the phase instability for the sound object category of strong harmonic frequencies is presented below.</p>
<p id="p0063" num="0063">For each Kth harmonic frequency in the category of strong harmonic frequencies, the phase at the point zero fundamental frequency FundZeroPos can be determined, so that from all the sound objects included in this harmonic category, the sound object whose n point is closest to the designated point is selected. For this point: <maths id="math0038" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseAtPointZero</mi></mstyle><mo>=</mo><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi>φ</mi><mo>+</mo><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi>ω</mi><mo>∗</mo><mfenced separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">FundZeroPos</mi></mstyle><mo>−</mo><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi mathvariant="bold-italic">p</mi></mfenced></math><img id="ib0038" file="imgb0038.tif" wi="135" he="5" img-content="math" img-format="tif"/></maths> The fundamental frequency of the human voice is typically between 80 Hz and 250 Hz, so more than 20 points with zero value can occur in a 200 ms time interval. To determine the phase parameters in each section for each K-th harmonic frequency in the category of Strong Harmonics, the TabPhase[20] parameter was used, in which the phase values were entered in 20 consecutive points, q, zero of the fundamental frequency, where TabPhase[Q] is a table containing the determined parameters and TabPhase[20] is the data stored in the table for the twentieth parameter.</p>
<p id="p0064" num="0064">In the phase graphs of the sound objects shown in <figref idref="f0006">Fig. 6</figref>, phase categorized local parameters are shown, according to the following formulas, where the phase F1 of the fundamental frequency component is assumed to be 0, but in other embodiments the phase of the fundamental frequency component may be assigned any value, which must then be taken into account when calculating other parameters:<!-- EPO <DP n="17"> -->
<ol id="ol0003" compact="compact" ol-style="">
<li>a) The average values of the phase of the K-th harmonic frequency when the phase of the fundamental frequency component, F1 = 0, in the category of Strong Harmonics: <maths id="math0039" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseAvarage</mi></mstyle><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced></mrow></mstyle></mfenced><mo>/</mo><mi mathvariant="bold-italic">Q</mi><mo>,</mo></math><img id="ib0039" file="imgb0039.tif" wi="96" he="8" img-content="math" img-format="tif"/></maths> where TabPhase[q] is the data for the q-th measurement stored in the table, and Q is the number of measurements, where Q can be, for example, 20.<br/>
The above parameter is very important because the average phase value can be a parameter identifying the speaker.</li>
<li>b) The value of the phase standard deviation of the Kth harmonic frequency in the Strong Harmonics category <maths id="math0040" num=""><math display="block"><mtable columnalign="left"><mtr><mtd><mstyle mathvariant="bold-italic"><mi mathvariant="italic">HarmPhaseStandDev</mi></mstyle></mtd></mtr><mtr><mtd><mo>=</mo><msqrt><mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseSquared</mi></mstyle><mo>/</mo><mi mathvariant="bold-italic">Q</mi><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseAvarag</mi></mstyle><msup><mi mathvariant="bold-italic">e</mi><mn>2</mn></msup></mrow></msqrt></mtd></mtr></mtable></math><img id="ib0040" file="imgb0040.tif" wi="138" he="14" img-content="math" img-format="tif"/></maths> wherein: <maths id="math0041" num=""><math display="block"><mi mathvariant="italic">KHarmPhaseSquared</mi><mo>=</mo><msup><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>q</mi><mo>=</mo><mn>1</mn></mrow><mi>Q</mi></msubsup><mfenced separators=""><mi mathvariant="italic">TabPhase</mi><mfenced open="[" close="]"><mi>q</mi></mfenced></mfenced></mstyle><mn>2</mn></msup><mo>;</mo></math><img id="ib0041" file="imgb0041.tif" wi="90" he="12" img-content="math" img-format="tif"/></maths></li>
<li>c) The value of the instantaneous phase change of the K-th harmonic frequency in the category of Strong Harmonics, when the phase of the fundamental frequency component, F1 = 0, the so-called drift: <maths id="math0042" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseDrift</mi></mstyle><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>2</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mfenced open="|" close="|" separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]" separators=""><mi mathvariant="bold-italic">q</mi><mo>−</mo><mn>1</mn></mfenced></mfenced></mstyle></mfenced><mo>/</mo><mfenced separators=""><mi mathvariant="bold-italic">Q</mi><mo>−</mo><mn>1</mn></mfenced></math><img id="ib0042" file="imgb0042.tif" wi="144" he="15" img-content="math" img-format="tif"/></maths></li>
</ol></p>
<p id="p0065" num="0065">Based on the above categorized local phase parameters, the global parameters are determined; however, more information about the signal can be obtained from the locally computed parameters in 200 ms sections.</p>
<p id="p0066" num="0066">By specifying the value of the standard deviation of the phase, and thus enabling the determination of the aforementioned phase inconsistency, for any number of zero values, and thus for any number of measurements Q, the value of the standard deviation of at least one instantaneous phase of the harmonic frequency for any category of sound objects can be presented as: <maths id="math0043" num=""><math display="block"><mtable columnalign="left"><mtr><mtd><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseStandDev</mi></mstyle></mtd></mtr><mtr><mtd><mo>=</mo><msqrt><msup><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mrow><msup><mfenced separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced></mfenced><mn>2</mn></msup><mo>/</mo><mi mathvariant="bold-italic">Q</mi><mo>−</mo><mfenced separators=""><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced></mrow></mstyle></mfenced><mo>/</mo><mi mathvariant="bold-italic">Q</mi></mfenced></mrow></mstyle><mn>2</mn></msup></msqrt></mtd></mtr></mtable></math><img id="ib0043" file="imgb0043.tif" wi="139" he="22" img-content="math" img-format="tif"/></maths><!-- EPO <DP n="18"> --> where TabPhase[q] is the data for the q-th measurement stored in the table, and Q is the number of measurements.</p>
<p id="p0067" num="0067"><figref idref="f0006">Fig. 6</figref> shows, for three different recordings at four different times, the local harmonic components parameter's graphs, related to the fundamental frequency component. Since the graph of the acoustic signal of successive harmonics corresponds to the cosine function, the graph has a maximum value when the harmonic is in phase 0. In the subsequent parts of the chart a vertical line crossing all harmonic components marks the moment when the phase F1 of the fundamental component has a value of 0. Next to this line there are vertical lines symbolizing the zero points of the phase of successive harmonics. By measuring the distance of the vertical line and the local line sections, we can determine the phase shift of selected harmonics from the phase of the fundamental frequency. Changing phase shift of any harmonic component changes the graph of the total acoustic signal presented in the upper line of the graph. Various measured values for the phases of the harmonic components are indicated below the graphs, including phase average values, phase drifts and phase standard deviations calculated in accordance with the formulas presented in this description. A stable value of the phase of the components guarantees the repeatability of the graph and yields information about the correct control of the speech path.</p>
<p id="p0068" num="0068">The irregular phase shift information allows the determination of the stability of phase values of the frequency components. The stable phase value guarantees the repeatability of the function, which corresponds to a sound signal graph saved in the file, and thus provides information about the correct signal path.</p>
<p id="p0069" num="0069">Determination of the non-constancy of component phases of frequency is important because changes in the relations between the phases of harmonic frequencies, especially between the phases of harmonic frequencies and the phase of the fundamental frequency, indicate whether the sound signal is controlled - a phase shift of the frequency of the fundamental component repeated in subsequent periods for the selected phase of the harmonic frequency indicates a regular and controllable sound signal, and in particular the regular and controlled operation of the sound path. A change in this shift, and more precisely in the relationship (dependence) between the phases, may indicate damage to this path or damage to the components affecting the sound generating system. Identification of the phase of a harmonic frequency that is characterized by an irregular shift can be used, for example, to identify the location of a fault.<!-- EPO <DP n="19"> --></p>
<p id="p0070" num="0070">The average value of phase shift of the selected harmonic frequency carries information about characteristic parameters of the selected fragment of the sound path and is a reference point for measuring the stability of the sound generating system. A large value of the standard deviation of phase shift of the selected harmonic frequency may be a signal that the sound path is not stable. Short-term changes (phase drift) calculated as the modulus of the phase difference at successive measurement points indicate a rapid increase or decrease in the vibration frequency of the tested harmonic and signify a deterioration of control over the analyzed fragment of the sound track.</p>
<p id="p0071" num="0071">The merging of sound objects into categories described above allows one to quickly see which categories contain how many sound objects - it is therefore possible to determine the number of sound objects contained in the fundamental frequency and in all strong harmonic frequencies, and to determine the percentage of the number of objects contained in the fundamental frequency and in all strong harmonic frequencies. Since audio objects are created when there is a step change in phase, the number of objects grouped in one category will suggest the quality and fidelity of the audio signal. The smaller the number of objects, the fewer breaks in the signal continuity occurred in the signal file, and thus the fewer breaks took place in the recorded sound.</p>
<p id="p0072" num="0072">Using methods known from the state of art, which employ signal compression or transformation, it is not possible to measure the phase parameters of the harmonic components of the sound, which carry so much information about the operation of the sound system, because the phase of components of the sound processed in this way is irretrievably distorted.</p>
<p id="p0073" num="0073">After the calculation steps and the determination of parameters, all obtained local and global parameters are placed in a table and then stored in an archive (12).</p>
<p id="p0074" num="0074">In the database (13) a pointer is created to the table stored in the archive (12), so that the database (13) serves as a register of handled orders for processed recordings.</p>
<p id="p0075" num="0075">In one embodiment, relations between the energies of the categorized sound objects are determined; for example, it is possible to analyze the energy distribution:
<ul id="ul0010" list-style="dash" compact="compact">
<li>between all categories - for example, the distribution can be presented in the form of a pie chart which can show, for example, the share of energy of each category of objects in the total energy of the sound signal;</li>
<li>between harmonic frequencies and noise - for example, the distribution can be presented as a Harmonic to Noise Ratio (HNR);<!-- EPO <DP n="20"> --></li>
<li>between strong harmonic frequencies - for example, the distribution can be presented in the form of the slope of the energy distribution line between successive harmonic frequencies;</li>
<li>distribution of energy in time for selected harmonic frequencies - presented, for example, as a directional coefficient of energy slope in successive time sections.</li>
</ul></p>
<p id="p0076" num="0076"><figref idref="f0007">Fig. 7</figref> shows the distribution of energy between the harmonic and noise components of the analyzed audio signal. The pie chart on the left shows the percentage of the energy of the fundamental, low frequency strong harmonics, high frequency harmonics and other objects. The bar graph to the left of the pie chart shows the breakdown of energy for the remaining objects into low noise, medium noise, high noise and strong non-harmonic objects. Based on this graph, it is possible to determine the HNR coefficient.</p>
<p id="p0077" num="0077">The bar graph on the right in <figref idref="f0007">Fig. 7</figref> graphically shows the distribution of energy between strong harmonics along with how many sound objects are contained in each harmonic. Since the sound object is interrupted when there is a step change of phase, the number of objects grouped in one harmonic indicates the quality of voice path control. The smaller the number of objects, the fewer breaks in the voice continuity. The average number of objects per harmonic is also important. As a rule, the lower the harmonic, the fewer broken objects there are. As can be seen from the examples in <figref idref="f0007">Fig. 7</figref>, the speech of people generating weaker vibrations is distinguished by a greater number of harmonics. In a person with severe speech disorders, the fundamental component concentrates more than 1/3 of the energy of the entire signal, and the number of strong harmonics is small. In this case, the harmonic slope declines very strongly.</p>
<p id="p0078" num="0078">The above-mentioned information obtained by using present method of recognizing characteristic features of sound timbre on the basis of sound objects allows to assess the fidelity of the signal, for example by comparing it with other signals stored in the File Archive (12).</p>
<p id="p0079" num="0079">The local and global statistical parameters and category parameters stored in the File Archive (12) and calculated in accordance with the above-described method can be used by the classification module (D) according to the functionalities assigned to it.</p>
<p id="p0080" num="0080">The classification module (D) may be used to recognize the regularity of the signals by computer counting the objects assigned to a category or by comparing parameters of the objects to reference parameters.</p>
<p id="p0081" num="0081">Advantageously, the reference parameters can be transferred and stored in the memory of the classification module (D).<!-- EPO <DP n="21"> --></p>
<p id="p0082" num="0082">In a variant in which classification module (D) compares the parameters, the classification module (D) is informed which of the transmitted parameters are reference parameters, i.e. parameters which have been derived in accordance with the method presented herein, for:
<ul id="ul0011" list-style="dash" compact="compact">
<li>a sound signal known to have the correct sound characteristics, i.e. undisturbed or representing a healthy person or a properly functioning machine;</li>
<li>a sound signal known to belong to an ill person or a malfunctioning machine, where it is known what the disease or damage to the machine is.</li>
</ul></p>
<p id="p0083" num="0083">These reference parameters are provided with information, for example in the form of a flag containing an appropriate description (e.g. a healthy patient/device working properly or an ill patient/device with type failure). This stage can be defined as building an internal classifier and assigning a classification rating.</p>
<p id="p0084" num="0084">Then, any new parameters, derived from recordings of audio signals for which there is no information about the patient/device condition, can be compared in the classification module (D) with the existing reference parameters to determine whether the source of the recording (patient/device) is healthy or works fine. Thanks to this, the new parameters derived are assigned a classification rating.</p>
<p id="p0085" num="0085">In one variant, the classification module (D) may change its reference parameters, for example using artificial intelligence algorithms or neural networks, based on the provided parameters and the provided classification scores.</p>
<p id="p0086" num="0086">The method according to the invention, using sound objects for signal analysis, allows obtaining information about the signal characteristics with a relatively small number of samples, which facilitates building classifiers and further recognition of signals. In addition, the very structure of saving data from these sound objects to tables has a positive effect on the size of files recorded and saved in the database, because the connections between items in the table allow for flexible manipulation of objects and facilitate access to data related to these objects, while the size of these saved objects is small.</p>
<p id="p0087" num="0087">With the method disclosed in the present application, it is also possible to recognize the correctness of a signal, so the invention finds application in fields such as:
<ol id="ol0004" ol-style="">
<li>1) diagnosis of disease states affecting speech mode;</li>
<li>2) monitoring the health condition of the examined person;<!-- EPO <DP n="22"> --></li>
<li>3) voice identification of persons;</li>
<li>4) diagnosis of the state of devices by the sound emitted by the device or its component.</li>
</ol></p>
<p id="p0088" num="0088">All embodiments of the method for recognizing timbre characteristics from sound objects are also applicable to the system, the computer program, and the computer program product.</p>
<p id="p0089" num="0089">In one embodiment, the system of the invention is implemented on a processor in any computing system such as server or PC computing system.</p>
<p id="p0090" num="0090">In another form, the invention is a computer program product comprising computer-readable instructions that cause the processor to perform the method of recognizing the characteristic features of sound timbre according to the invention.</p>
<p id="p0091" num="0091">Each of the schematic blocks of the system for recognizing characteristic features of sound timbre based on sound objects, as well as each of the steps of the method for recognizing characteristic features of sound timbre based on sound objects, may be implemented by computer program instructions. These instructions may be provided to a general purpose computer processor, a special purpose computer or other programmable data processing device such that the instructions that are executed by the computer processor or other programmable data processing device enable implementation of the functions defined in the system diagram and method steps.</p>
<p id="p0092" num="0092">Aspects of the present invention may be implemented by a computer or devices such as a CPU (central processing unit), MPU (microprocessor unit) or MCU (microcontroller unit) that read and execute a computer program product stored in the device memory to perform the functions of the embodiments described above. Aspects of the present invention may also be implemented by a method the steps of which are performed by a computer of the system or device by, for example, reading and executing a program stored on a memory device to perform the functions of the embodiments described above. In this regard, the computer program product is provided to the computer, for example, over a network or other type of recording medium serving as a storage device. The computer program product of the invention also includes a durable machine-readable medium.</p>
<p id="p0093" num="0093">Embodiments are provided herein as non-limiting indications of the invention only and are not intended to limit in any way the scope of protection that is defined by the claims, and terms such as sound timbre refer to sound characteristics and may also be referred to as components, intensity or harmonics, without affecting the scope of the invention. It should be understood that<!-- EPO <DP n="23"> --> each technical solution used in the invention can be implemented using equivalent technologies without going beyond the scope of protection.</p>
</description>
<claims id="claims01" lang="en"><!-- EPO <DP n="24"> -->
<claim id="c-en-0001" num="0001">
<claim-text>A method of recognizing the characteristic features of sound timbre based on sound object, where a sound object is analyzed, wherein the sound object is a single acoustic signal with a slowly varying amplitude and a single slowly varying frequency and continuous phase, for which sound object statistical parameters are determined:
<claim-text>- for each time interval, local parameters are calculated for at least one sound object:<br/>
parameter of amplitude, frequency, number of measurement points;</claim-text>
<claim-text>- global parameters are calculated from local parameters;</claim-text>
<b>characterized in that</b> after calculating the statistical parameters of the sound objects, the method comprises the steps of:
<claim-text>- separating the sound objects into categories, wherein the categories are formed depending on the frequency, wherein at least one of the categories is dependent on the fundamental frequency and its associated harmonic frequencies;</claim-text>
<claim-text>- calculating categorized local and global parameters for at least one category of sound objects; and</claim-text>
<claim-text>- determining at least one of:
<claim-text>o phase instability as the value of the standard deviation of at least one instantaneous phase of the harmonic frequency of category of objects of at least one category from fundamental frequency phase,</claim-text>
<claim-text>o the share of the energy of each category of objects in the total energy of the sound signal;</claim-text>
<claim-text>o the ratio of harmonic frequency energy to noise;</claim-text>
<claim-text>o local and global amplitude slope coefficients;</claim-text>
<claim-text>o directional slope coefficient of the energy graph of the fundamental frequency and strong harmonic;</claim-text>
<claim-text>o a local and global phase shift of each harmonic frequency relative to the fundamental frequency at the moment when the fundamental frequency phase is 0;</claim-text>
<claim-text>o the average value of phase Fk at the moment when the phase F1 of the fundamental frequency component has a given value; where k is the order number of the harmonic frequency<!-- EPO <DP n="25"> --></claim-text>
<claim-text>o the phase Fk drift at the moment when the phase F1 of the fundamental frequency component has a given value; where k is the order number of the harmonic frequency;</claim-text>
<claim-text>o the number of sound objects contained in the fundamental frequency and in all strong harmonic frequencies;</claim-text>
<claim-text>o the percentage number of objects contained in the fundamental frequency and in all strong harmonic frequencies.</claim-text></claim-text></claim-text></claim>
<claim id="c-en-0002" num="0002">
<claim-text>The method according to claim 1, <b>characterized in that</b> the value of the standard deviation of at least one instantaneous phase of the harmonic frequency of category of objects is calculated by the formula <maths id="math0044" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseStandDev</mi></mstyle><mo>=</mo><msqrt><msup><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mrow><msup><mfenced separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced></mfenced><mn>2</mn></msup><mo>/</mo><mi mathvariant="bold-italic">Q</mi><mo>−</mo><mfenced separators=""><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced></mrow></mstyle></mfenced><mo>/</mo><mi mathvariant="bold-italic">Q</mi></mfenced></mrow></mstyle><mn>2</mn></msup></msqrt></math><img id="ib0044" file="imgb0044.tif" wi="159" he="15" img-content="math" img-format="tif"/></maths> where TabPhase[q] is the data for the q-th measurement stored in the table, and Q is the number of measurements.</claim-text></claim>
<claim id="c-en-0003" num="0003">
<claim-text>The method according to any one of claim 1 or claim 2, <b>characterized in that</b> the global parameters are calculated as a weighted average of the local parameters.</claim-text></claim>
<claim id="c-en-0004" num="0004">
<claim-text>The method according to claim 1, <b>characterized in that</b> the energy of each sound object is calculated by a formula: <maths id="math0045" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoEnergy</mi></mstyle><mo>=</mo><msup><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">n</mi><mo>=</mo><mn>1</mn></mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SoNumOfPoint</mi></mstyle></msubsup><mfenced separators=""><mi mathvariant="bold-italic">S</mi><msub><mi mathvariant="bold-italic">p</mi><mi mathvariant="bold-italic">n</mi></msub><mi mathvariant="bold-italic">a</mi></mfenced></mstyle><mn>2</mn></msup></math><img id="ib0045" file="imgb0045.tif" wi="71" he="11" img-content="math" img-format="tif"/></maths> where n is the ordinal number of the sound object point, and Sp<sub>n</sub>a is the amplitude of the sound object point for the n-th point.</claim-text></claim>
<claim id="c-en-0005" num="0005">
<claim-text>The method according to claim 1, <b>characterized in that</b> the amplitude slope coefficient is calculated by the formula <maths id="math0046" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmpLean</mi></mstyle><mo>=</mo><mfenced separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmpPos</mi></mstyle><mo>∗</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">Num</mi></mstyle><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumAmp</mi></mstyle><mo>∗</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">SumPos</mi></mstyle></mfenced><mo>/</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">Delta</mi></mstyle><mo>;</mo></math><img id="ib0046" file="imgb0046.tif" wi="136" he="5" img-content="math" img-format="tif"/></maths> wherein <maths id="math0047" num=""><math display="block"><mi mathvariant="italic">Num</mi><mo>=</mo><mi mathvariant="italic">SoNumOfPoint</mi><mo>;</mo></math><img id="ib0047" file="imgb0047.tif" wi="47" he="5" img-content="math" img-format="tif"/></maths> <maths id="math0048" num=""><math display="block"><mi mathvariant="italic">SumPos</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi><mo>;</mo></mrow></mstyle></math><img id="ib0048" file="imgb0048.tif" wi="44" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0049" num=""><math display="block"><mi mathvariant="italic">SumSquarePos</mi><mo>=</mo><msup><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mfenced separators=""><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mfenced></mstyle><mn>2</mn></msup><mo>;</mo></math><img id="ib0049" file="imgb0049.tif" wi="61" he="6" img-content="math" img-format="tif"/></maths> <maths id="math0050" num=""><math display="block"><mi mathvariant="italic">Delta</mi><mo>=</mo><mi mathvariant="italic">SumSquarePos</mi><mo>∗</mo><mi mathvariant="italic">Num</mi><mo>−</mo><mi mathvariant="italic">SumPos</mi><mo>∗</mo><mi mathvariant="italic">SumPos</mi><mo>;</mo></math><img id="ib0050" file="imgb0050.tif" wi="98" he="5" img-content="math" img-format="tif"/></maths> <maths id="math0051" num=""><math display="block"><mi mathvariant="italic">SumAmp</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>a</mi></mrow></mstyle><mo>;</mo></math><img id="ib0051" file="imgb0051.tif" wi="45" he="6" img-content="math" img-format="tif"/></maths><!-- EPO <DP n="26"> --> <maths id="math0052" num=""><math display="block"><mi mathvariant="italic">SumAmpPos</mi><mo>=</mo><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi>n</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="italic">Num</mi></msubsup><mrow><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>a</mi><mo>∗</mo><msub><mi mathvariant="italic">Sp</mi><mi>n</mi></msub><mi>p</mi></mrow></mstyle></math><img id="ib0052" file="imgb0052.tif" wi="65" he="11" img-content="math" img-format="tif"/></maths> where n is the ordinal number of the sound object point, SoNumOfPoint is the number of points, Sp<sub>n</sub>a is the amplitude of the sound object point, and Sp<sub>n</sub>p is the position of the sound object point.</claim-text></claim>
<claim id="c-en-0006" num="0006">
<claim-text>The method according to claim 1, <b>characterized in that</b> the average phase Fk value is calculated from the formula: <maths id="math0053" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseAvarage</mi></mstyle><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>1</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mrow><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced></mrow></mstyle></mfenced><mo>/</mo><mi mathvariant="bold-italic">Q</mi><mo>;</mo></math><img id="ib0053" file="imgb0053.tif" wi="99" he="12" img-content="math" img-format="tif"/></maths> where TabPhase[q] is the data for the q-th measurement stored in the table.</claim-text></claim>
<claim id="c-en-0007" num="0007">
<claim-text>The method according to claim 1, <b>characterized in that</b> the phase drift is calculated by the formula: <maths id="math0054" num=""><math display="block"><mstyle mathvariant="bold-italic"><mi mathvariant="italic">KHarmPhaseDrift</mi></mstyle><mo>=</mo><mfenced><mstyle displaystyle="true"><msubsup><mo>∑</mo><mrow><mi mathvariant="bold-italic">q</mi><mo>=</mo><mn>2</mn></mrow><mi mathvariant="bold-italic">Q</mi></msubsup><mfenced open="|" close="|" separators=""><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]"><mi mathvariant="bold-italic">q</mi></mfenced><mo>−</mo><mstyle mathvariant="bold-italic"><mi mathvariant="italic">TabPhase</mi></mstyle><mfenced open="[" close="]" separators=""><mi mathvariant="bold-italic">q</mi><mo>−</mo><mn>1</mn></mfenced></mfenced></mstyle></mfenced><mo>/</mo><mfenced separators=""><mi mathvariant="bold-italic">Q</mi><mo>−</mo><mn>1</mn></mfenced><mo>,</mo></math><img id="ib0054" file="imgb0054.tif" wi="145" he="12" img-content="math" img-format="tif"/></maths> where TabPhase[q] is the data for the q-th measurement stored in the table.</claim-text></claim>
<claim id="c-en-0008" num="0008">
<claim-text>The method according to any one of claims 1-7, <b>characterized in that</b> the sound objects are divided into at least the following categories:
<claim-text>- fundamental frequency category;</claim-text>
<claim-text>- the category of strong harmonic frequencies, in which the share of these harmonics in the signal energy is more than 5%;</claim-text>
<claim-text>- a category of weak harmonic frequencies, in which the share of these harmonics in the signal energy is below 5%;</claim-text>
<claim-text>- category of non-harmonic signals of single frequency and amplitude,</claim-text>
<claim-text>- low frequency noise category, below 200 Hz;</claim-text>
<claim-text>- medium frequency noise category, i.e. in the range from 200 Hz to 2000 Hz;</claim-text>
<claim-text>- high-frequency noise category, i.e. in the range from 2,000 Hz to 10,000 Hz.</claim-text></claim-text></claim>
<claim id="c-en-0009" num="0009">
<claim-text>The method according to any one of claims 1-8, <b>characterized in that</b> after the determination step, all parameters are stored and a set of training input data is created for the artificial intelligence neural network.</claim-text></claim>
<claim id="c-en-0010" num="0010">
<claim-text>The method according to any one of claims 1-9, <b>characterized in that</b> the time interval is 200 ms.<!-- EPO <DP n="27"> --></claim-text></claim>
<claim id="c-en-0011" num="0011">
<claim-text>The method according to any one of claims 1-10, <b>characterized in that</b> the phase of the fundamental frequency component has a predetermined value.</claim-text></claim>
<claim id="c-en-0012" num="0012">
<claim-text>The method according to claims 11, <b>characterized in that the</b> phase of the fundamental frequency component is equal to 0.</claim-text></claim>
<claim id="c-en-0013" num="0013">
<claim-text>A system for recognizing characteristic features of sound timbre based on sound objects, where a sound object is analyzed, wherein the sound object is a single acoustic signal with a slowly varying amplitude and a single slowly varying frequency and continuous phase, <b>characterized by</b> comprising:
<claim-text>- recording module adapted to record sounds;</claim-text>
<claim-text>- a processing module adapted to process the recorded sound so that the sound object is extracted;</claim-text>
<claim-text>- a computing module adapted to perform the method of claims 1 to 12;</claim-text>
<claim-text>- a classifying module adapted to analyze the sound object based on the results obtained from the calculation module.</claim-text></claim-text></claim>
<claim id="c-en-0014" num="0014">
<claim-text>A computer program for recognizing characteristic features of sound timbre based on sound objects, comprising instructions for performing the method of any one of claims 1 to 12.</claim-text></claim>
<claim id="c-en-0015" num="0015">
<claim-text>A computer program product for recognizing characteristic features of sound timbre based on sound objects comprising a computer-readable code performing the steps of the method of any of claims 1 to 12.</claim-text></claim>
</claims>
<drawings id="draw" lang="en"><!-- EPO <DP n="28"> -->
<figure id="f0001" num="1"><img id="if0001" file="imgf0001.tif" wi="155" he="200" img-content="drawing" img-format="tif"/></figure><!-- EPO <DP n="29"> -->
<figure id="f0002" num="2"><img id="if0002" file="imgf0002.tif" wi="146" he="218" img-content="drawing" img-format="tif"/></figure><!-- EPO <DP n="30"> -->
<figure id="f0003" num="3"><img id="if0003" file="imgf0003.tif" wi="146" he="218" img-content="drawing" img-format="tif"/></figure><!-- EPO <DP n="31"> -->
<figure id="f0004" num="4"><img id="if0004" file="imgf0004.tif" wi="144" he="218" img-content="drawing" img-format="tif"/></figure><!-- EPO <DP n="32"> -->
<figure id="f0005" num="5"><img id="if0005" file="imgf0005.tif" wi="149" he="216" img-content="drawing" img-format="tif"/></figure><!-- EPO <DP n="33"> -->
<figure id="f0006" num="6"><img id="if0006" file="imgf0006.tif" wi="145" he="218" img-content="drawing" img-format="tif"/></figure><!-- EPO <DP n="34"> -->
<figure id="f0007" num="7"><img id="if0007" file="imgf0007.tif" wi="149" he="214" img-content="drawing" img-format="tif"/></figure>
</drawings>
<search-report-data id="srep" lang="en" srep-office="EP" date-produced=""><doc-page id="srep0001" file="srep0001.tif" wi="161" he="240" type="tif"/><doc-page id="srep0002" file="srep0002.tif" wi="161" he="240" type="tif"/></search-report-data><search-report-data date-produced="20231031" id="srepxml" lang="en" srep-office="EP" srep-type="ep-sr" status="n"><!--
 The search report data in XML is provided for the users' convenience only. It might differ from the search report of the PDF document, which contains the officially published data. The EPO disclaims any liability for incorrect or incomplete data in the XML for search reports.
 -->

<srep-info><file-reference-id>P32023EP00/ABL</file-reference-id><application-reference><document-id><country>EP</country><doc-number>23461601.9</doc-number></document-id></application-reference><applicant-name><name>Vivid Mind PSA</name></applicant-name><srep-established srep-established="yes"/><srep-invention-title title-approval="yes"/><srep-abstract abs-approval="yes"/><srep-figure-to-publish figinfo="by-applicant"><figure-to-publish><fig-number>1</fig-number></figure-to-publish></srep-figure-to-publish><srep-info-admin><srep-office><addressbook><text>MN</text></addressbook></srep-office><date-search-report-mailed><date>20231109</date></date-search-report-mailed></srep-info-admin></srep-info><srep-for-pub><srep-fields-searched><minimum-documentation><classifications-ipcr><classification-ipcr><text>G10L</text></classification-ipcr></classifications-ipcr></minimum-documentation></srep-fields-searched><srep-citations><citation id="sr-cit0001"><nplcit id="sr-ncit0001" npl-type="s"><article><author><name>PEETERS GEOFFROY ET AL</name></author><atl>The Timbre Toolbox: Extracting audio descriptors from musical signals</atl><serial><sertitle>THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, AMERICAN INSTITUTE OF PHYSICS, 2 HUNTINGTON QUADRANGLE, MELVILLE, NY 11747</sertitle><pubdate>20111101</pubdate><vid>130</vid><ino>5</ino><doi>10.1121/1.3642604</doi><issn>0001-4966</issn></serial><location><pp><ppf>2902</ppf><ppl>2916</ppl></pp></location><refno>XP012152867</refno></article></nplcit><category>X</category><rel-claims>1-15</rel-claims><rel-passage><passage>* page 2904, right-hand column, line 43 - page 2915, left-hand column, line 12; figures 3,4; table I *</passage></rel-passage></citation><citation id="sr-cit0002"><nplcit id="sr-ncit0002" npl-type="b"><article><book><author><name>XIN ZHANG ET AL</name></author><book-title>Discriminant Feature Analysis for Music Timbre Recognition and Automatic Indexing</book-title><imprint><name>MINING COMPLEX DATA; [LECTURE NOTES IN COMPUTER SCIENCE], SPRINGER BERLIN HEIDELBERG, BERLIN, HEIDELBERG, PAGE(S) 104 - 115</name><pubdate>20070917</pubdate></imprint><isbn>978-3-540-68415-2</isbn><refno>XP019075844</refno></book></article></nplcit><category>X</category><rel-claims>1-15</rel-claims><rel-passage><passage>* abstract *</passage><passage>* page 106, line 11 - page 110, line 23 *</passage></rel-passage></citation><citation id="sr-cit0003"><nplcit id="sr-ncit0003" npl-type="s"><article><author><name>XIN ZHANG ET AL</name></author><atl>Analysis of Sound Features for Music Timbre Recognition</atl><serial><sertitle>2007 INTERNATIONAL CONFERENCE ON MULTIMEDIA AND UBIQUITOUS ENGINEERING IEEE PISCATAWAY, NJ, USA, IEEE, PISCATAWAY, NJ, USA</sertitle><pubdate>20070401</pubdate><isbn>978-0-7695-2777-2</isbn></serial><location><pp><ppf>3</ppf><ppl>8</ppl></pp></location><refno>XP031086496</refno></article></nplcit><category>X</category><rel-claims>1-15</rel-claims><rel-passage><passage>* page 3, left-hand column, line 18 - page 7, right-hand column, line 16 *</passage></rel-passage></citation><citation id="sr-cit0004"><nplcit id="sr-ncit0004" npl-type="s"><article><author><name>ALICJA A WIECZORKOWSKA ET AL</name></author><atl>Identification of a dominating instrument in polytimbral same-pitch mixes using SVM classifiers with non-linear kernel</atl><serial><sertitle>JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, KLUWER ACADEMIC PUBLISHERS, BO</sertitle><pubdate>20090826</pubdate><vid>34</vid><ino>3</ino><issn>1573-7675</issn></serial><location><pp><ppf>275</ppf><ppl>303</ppl></pp></location><refno>XP019790648</refno></article></nplcit><category>A</category><rel-claims>1-15</rel-claims><rel-passage><passage>* page 275, line 1 - page 287, line 14; figure 1 *</passage></rel-passage></citation></srep-citations><srep-admin><examiners><primary-examiner><name>Dobler, Ervin</name></primary-examiner></examiners><srep-office><addressbook><text>Munich</text></addressbook></srep-office><date-search-completed><date>20231031</date></date-search-completed></srep-admin></srep-for-pub></search-report-data>
<ep-reference-list id="ref-list">
<heading id="ref-h0001"><b>REFERENCES CITED IN THE DESCRIPTION</b></heading>
<p id="ref-p0001" num=""><i>This list of references cited by the applicant is for the reader's convenience only. It does not form part of the European patent document. Even though great care has been taken in compiling the references, errors or omissions cannot be excluded and the EPO disclaims all liability in this regard.</i></p>
<heading id="ref-h0002"><b>Patent documents cited in the description</b></heading>
<p id="ref-p0002" num="">
<ul id="ref-ul0001" list-style="bullet">
<li><patcit id="ref-pcit0001" dnum="US15686057A"><document-id><country>US</country><doc-number>15686057</doc-number><kind>A</kind><date>20180301</date></document-id></patcit><crossref idref="pcit0001">[0003]</crossref></li>
<li><patcit id="ref-pcit0002" dnum="CN202011522343A"><document-id><country>CN</country><doc-number>202011522343</doc-number><kind>A</kind><date>20210420</date></document-id></patcit><crossref idref="pcit0002">[0004]</crossref></li>
<li><patcit id="ref-pcit0003" dnum="CN202110775311A"><document-id><country>CN</country><doc-number>202110775311</doc-number><kind>A</kind><date>20211112</date></document-id></patcit><crossref idref="pcit0003">[0006]</crossref></li>
<li><patcit id="ref-pcit0004" dnum="CN202111677396A"><document-id><country>CN</country><doc-number>202111677396</doc-number><kind>A</kind><date>20220408</date></document-id></patcit><crossref idref="pcit0004">[0007]</crossref></li>
<li><patcit id="ref-pcit0005" dnum="EP16741938A" dnum-type="L"><document-id><country>EP</country><doc-number>16741938</doc-number><kind>A</kind><date>20180411</date></document-id></patcit><crossref idref="pcit0005">[0008]</crossref></li>
<li><patcit id="ref-pcit0006" dnum="EP3304549A1"><document-id><country>EP</country><doc-number>3304549</doc-number><kind>A1</kind></document-id></patcit><crossref idref="pcit0006">[0033]</crossref><crossref idref="pcit0007">[0034]</crossref><crossref idref="pcit0008">[0040]</crossref><crossref idref="pcit0009">[0057]</crossref></li>
</ul></p>
</ep-reference-list>
</ep-patent-document>
