WO2012070212A1 - うるささ判定システム、装置、方法およびプログラム - Google Patents
うるささ判定システム、装置、方法およびプログラム Download PDFInfo
- Publication number
- WO2012070212A1 WO2012070212A1 PCT/JP2011/006435 JP2011006435W WO2012070212A1 WO 2012070212 A1 WO2012070212 A1 WO 2012070212A1 JP 2011006435 W JP2011006435 W JP 2011006435W WO 2012070212 A1 WO2012070212 A1 WO 2012070212A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- speech
- sound
- latency
- annoyance
- unit
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/12—Audiometering
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
- A61B5/377—Electroencephalography [EEG] using evoked responses
- A61B5/38—Acoustic or auditory stimuli
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2225/00—Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
- H04R2225/43—Signal processing in hearing aids to enhance the speech intelligibility
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/70—Adaptation of deaf aid to hearing loss, e.g. initial electronic fitting
Definitions
- the present invention relates to a technique for evaluating (determining) whether or not a user has been able to listen to speech comfortably.
- the hearing aid compensates for the decreased hearing ability of the user by amplifying the amplitude of a signal of a specific frequency among various frequencies constituting a sound that is difficult for the user to hear.
- Hearing aids are required to adjust the amount of sound amplification in accordance with the degree of hearing loss for each user. Therefore, before starting to use the hearing aid, “fitting” that adjusts the amount of sound amplification in accordance with the hearing ability of each user is essential.
- Fitting means that the sound pressure to be output (atmospheric pressure fluctuation, which can be perceived as sound) is set to MCL (most comfortable level: sound pressure that the user feels comfortable) for each sound frequency.
- MCL most comfortable level: sound pressure that the user feels comfortable
- fitting is not appropriate in the case of either (1) insufficient amplification amount or (2) excessive amplification amount.
- the amount of amplification is insufficient, the user cannot hear the voice and the purpose of wearing the hearing aid cannot be achieved.
- the amount of amplification is excessive, the user can distinguish the voice.
- the loudspeaker presents a louder volume than necessary, which may damage the user's ear.
- the first procedure of fitting is audiogram measurement.
- “Audiogram” means measuring the minimum sound pressure of an audible pure tone. For example, it is the figure which plotted the minimum sound pressure (decibel value) which the user can hear about each of the sound of a some frequency according to the frequency (for example, 250Hz, 500Hz, 1000Hz, 2000Hz, 4000Hz).
- the amplification amount for each frequency is determined based on the fitting theory that is a function for estimating the amplification amount for each frequency from the audiogram result.
- the audiogram does not have a one-to-one correspondence with the ability to distinguish between conversations, and the difficulty of adjustment because the hearing-impaired person has a narrow sound pressure range that feels appropriate.
- a hearing aid conformity inspection is carried out using the hearing aid determined and adjusted by the above method (for example, see Non-Patent Document 1).
- a single syllable speech sound is presented with sound pressures of 55 dB SPL (Sound pressure level), 65 dB SPL, 75 dBSPL, and 85 dB SPL when hearing aid is worn and when hearing aid is not worn (bare ear), Plot the speech intelligibility for each sound pressure and compare. Then, if the clarity is improved at the time of wearing as compared with the case of non-wearing, it is determined as conforming.
- Speech intelligibility is an indicator of whether or not a single syllable speech sound has been heard. Speech intelligibility reflects the degree of discrimination during conversation.
- the “single syllable word sound” indicates one vowel or a combination of a consonant and a vowel (for example, “a” / “da” / “shi”).
- Speech intelligibility is evaluated by the following procedure (for example, Non-Patent Document 2).
- the 67S-type word table (20-word sound) established by the Japan Audiological Society is played one by one and sent to the user.
- the user is made to answer by using a method such as utterance or writing, as to which speech has been heard.
- the speech sound presented by the evaluator is compared with the answer, and the correct answer rate, which is the proportion of the speech sound that is correctly heard out of all 20 speech sounds, is calculated.
- the correct answer rate is speech intelligibility.
- Patent Document 1 discloses a speech intelligibility evaluation method that automatically performs correct / incorrect determination using a personal computer (PC).
- PC personal computer
- a single syllable voice is presented to a user using a PC, the user is made to answer with a mouse or a pen touch (touch the pen to the display), an answer is accepted as an input of the PC,
- By receiving an answer input with a mouse or pen touch it is not necessary for the evaluator to identify / decipher the user's answer (speech or writing), and the effort of the evaluator is reduced.
- Patent Document 2 discloses a speech intelligibility evaluation method that presents a selection candidate of speech corresponding to a speech after speech presentation.
- the number of selection candidates is narrowed down to several, and the user's trouble of searching for a character is reduced by selecting a corresponding word sound from among several characters.
- Non-patent Document 2 In the measurement of the permissible level of environmental noise, it is evaluated whether or not the environmental noise is acceptable when the reading sound and the environmental noise are simultaneously presented and the reading sound is heard (Non-patent Document 2). Specifically, the reading sound is presented at 65 dB SPL and the environmental noise is presented at 55 dB SPL, and a subjective impression regarding whether the environmental noise is acceptable is reported. As a subjective impression, it can be reported that either hearing aids can be used when listening to reading sounds under noise, or wearing a hearing aid under noise is difficult. Then, the former case is determined as conforming, and the latter case is determined as nonconforming.
- the conformity state is determined based only on the speech intelligibility, and it was not considered whether the user felt noisy when listening to the speech . Therefore, even when it was felt noisy when listening to the speech sound, it was evaluated as a suitable hearing aid process if the clarity was higher when the hearing aid was worn than when it was not worn. Further, in the measurement of the environmental noise tolerance level, whether or not the environmental noise is acceptable was evaluated, and the annoyance with respect to listening to the speech was not evaluated. In these evaluations, it may be determined that the hearing aid processing that the user feels noisy at the time of listening to the speech sound and is tired by listening is suitable. Annoyance for listening to speech sounds is a burden for users who wear hearing aids on a daily basis.
- An object of the present invention is to provide a mechanism for evaluating a user state relating to annoyance with respect to speech listening.
- An annoyance determination system includes a biological signal measurement unit that measures a user's electroencephalogram signal, and a speech database that stores speech of a plurality of single syllables, and for each speech, corresponds to the speech and the speech
- the speech sound database in which the reference latency of the negative component of the electroencephalogram signal to be associated is associated the speech sound determination unit for determining the speech sound of a single syllable to be presented with reference to the speech sound database, and the user
- An output unit that presents the voice, a peak latency of the negative component of the electroencephalogram signal included in a range of 50 ms to 350 ms starting from the time when the speech was presented, and a determination stored in the speech database
- An annoyance determining unit that compares the reference latency corresponding to the word sound and determines the noisyness of the output word sound.
- the speech sound database may associate a reference latency of a negative component of an electroencephalogram signal corresponding to the speech sound with the length of the consonant included in the speech sound or the intensity of the consonant.
- the annoyance determining unit determines that the audio signal is noisy for the user when the peak latency of the negative component is smaller than the reference latency, and the peak latency of the negative component is greater than the reference latency. If it is large, it may be determined that the audio signal is not noisy for the user.
- the annoyance determination system further includes an event-related potential processing unit that adds event-related potentials of the electroencephalogram signal according to a predetermined standard, the presented word sound determination unit determines a plurality of word sounds, and the output unit includes: The plurality of determined speech sounds are sequentially presented, and the event-related potential processing unit is configured to indicate, for the plurality of speech sounds, a time when the speech sound having the same sound pressure when the speech sound type or the speech sound is presented is presented. The event-related potentials of the electroencephalogram signal as a starting point may be added, and the added result may be output to the annoyance determination unit.
- the annoyance determination unit takes the minimum potential of the negative component of the electroencephalogram signal included in the range of 50 ms or more and 350 ms or less from the time when the determined speech is presented as the peak latency, or A peak latency associated with the template when the degree of coincidence between the brain wave signal and the template of the waveform of the N1 component prepared in advance is the highest may be adopted.
- the event-related potential processing unit may add an event-related potential for each consonant or for each word sound group in which the difference in reference latency is smaller than a predetermined value.
- the annoyance determination system further includes a result storage database that stores information indicating a determination result related to the speech sound annoyance.
- the result storage database includes a predetermined value for a difference between each word sound, each consonant, or a reference latency. For each smaller speech sound group, information indicating a determination result related to the loudness of the speech sound may be accumulated.
- the noisy determination system further includes a hearing aid processing unit that selects a type of hearing aid processing for the presented speech and processes the speech data held in the speech database based on the selected hearing aid processing. Also good.
- the annoyance determination system further includes a switching unit that switches between a calibration mode for obtaining a latency of a negative component for each user and an evaluation mode for evaluating the annoyance, and in the calibration mode, the switching unit includes the presented word sound determination unit. After selecting the vowel and calculating the reference latency for each speech based on the latency of the negative component with respect to the vowel, and after switching to the evaluation mode, the switching unit calculates the peak latency of the negative component The reference latency may be compared.
- the evaluation switching unit sets the latency of the N1 component with respect to the vowel as the reference latency of the vowel, and sets the length of the consonant part during the reference latency of the vowel.
- the reference latency for each consonant may be calculated by adding a positive value determined according to the length or intensity.
- a hearing aid processing system is a speech sound database that stores speech sounds of a plurality of single syllables, and for each speech sound, there is a speech sound and a reference latency of a negative component of an electroencephalogram signal corresponding to the speech sound.
- the speech sound determination unit that determines the speech sound of a single syllable to be presented with reference to the speech sound database associated with the speech sound database, the speech sound Corresponds to the peak latency of the negative component of the electroencephalogram signal included in the range of 50 ms to 350 ms starting from the time when the message is presented to the user by the output unit, and the determined word sound held in the word sound database
- An annoyance determination unit that outputs a result of comparing the reference latency to be output, and a negative component peak of the electroencephalogram signal output by the annoyance determination unit Based on the difference between the reference time latency corresponding to a time to the word sounds, and a hearing aid processor for adjusting said speech.
- the annoyance determination method relates to a step of measuring a user's brain wave signal, a word sound of a plurality of single syllables, and a reference latency of a negative component of an electroencephalogram signal corresponding to the word sound.
- determining a speech of a single syllable to be presented presenting the determined speech to the user, and starting from the time when the speech is presented 50 ms to 350 ms
- the peak latency of the negative component of the electroencephalogram signal included in the following range is compared with the reference latency corresponding to the determined speech stored in the speech database, and the output of the speech Determining noisy.
- a computer program is a computer program executed by a computer provided in a speech sound annoyance determination system, wherein the computer program is implemented in the annoyance determination system.
- a speech sound database in which a step of receiving a user's brain wave signal, a plurality of single syllable speech sounds, and a reference latency of a negative component of the electroencephalogram signal corresponding to the speech sound is stored in association with each other, A step of determining a syllable speech, a step of presenting the determined speech to the user, and a peak of a negative component of the electroencephalogram signal included in a range of 50 ms to 350 ms starting from the time when the speech was presented Latency and the determined and stored in the speech database By comparing the time reference latency corresponding to the sound, to perform the step of determining loudness of the output the word sound.
- An annoyance determination system includes a biological signal measurement unit that measures a user's brain wave signal, a voice input unit that inputs a voice signal uttered by a specific speaker, and the voice signal is input.
- a voice analysis unit that detects a timing and outputs a trigger, and further analyzes a feature relating to a length and intensity of a consonant portion of the speech, and based on the feature analyzed by the speech analysis unit,
- a reference latency estimation unit for estimating a reference latency, a peak latency of a negative component of the electroencephalogram signal included in a range of 50 ms to 350 ms starting from the trigger, and a reference estimated by the reference latency estimation unit
- an annoyance determining unit that determines annoyance by comparing with the latency.
- the annoyance determination system further includes a character output unit that outputs character information indicating a sound to be generated by the specific speaker, and the specific speaker is generated based on the character information output by the character output unit.
- An audio signal may be input to the audio input unit.
- the character output unit further outputs information related to sound pressure indicating the volume when the specific speaker utters the speech of the single syllable, and the character information and information related to sound pressure output by the character output unit Based on the voice signal generated by the specific speaker, the voice input unit may input the voice signal.
- the annoyance determination system further includes a presentation speech sound determination unit that determines a speech sound to be uttered by the specific speaker with reference to a speech sound list prepared in advance, and the character output unit is determined by the presentation speech sound determination unit.
- character information indicating the speech may be output.
- the reference latency estimation unit estimates a reference latency of the negative component based on the characteristics analyzed by the speech analysis unit and the speech to be uttered by the specific speaker determined by the presentation speech determination unit. May be.
- the reference latency estimation unit estimates a reference latency of the negative component by adding a predetermined positive value determined based on a consonant length or a consonant intensity of the voice to a base latency prepared in advance. May be.
- an annoyance determination method includes a step of measuring a user's brain wave signal, a step of inputting a voice signal uttered by a specific speaker, and a timing at which the voice signal is input. Outputting a trigger and further analyzing characteristics related to the length and intensity of the consonant portion of the speech; estimating a reference latency of a negative component based on the characteristics analyzed in the analyzing step; And comparing the peak latency of the negative component of the electroencephalogram signal included in the range of 50 ms or more and 350 ms or less starting from the trigger with the reference latency estimated by the estimating step to determine annoyance. To do.
- FIG. 1 It is a figure which shows the outline
- (A) is a figure which shows the sound pressure level for every condition measured with the sound level meter
- (b) is a figure which shows the gain adjustment amount for every frequency in each of six conditions.
- (A) is the figure which showed the electrode position of the international 10-20 method (10-20 square system).
- (B) is an electrode arrangement equipped with an electroencephalograph. It is a figure which shows the result of having made a subjective report about the annoyance on a visual analog scale.
- FIG. 6 is a flowchart showing a processing procedure of a speech sound annoyance determination system 200 according to Embodiment 2. It is a figure which shows the structure of the functional block of the noisy judgment system 300 of the speech sound listening by Embodiment 3.
- FIG. 10 is a flowchart illustrating a processing procedure of a speech sound annoyance determination system 300 according to Embodiment 3. It is a figure which shows the structure of the functional block of the annoyance determination system 400 of speech sound listening by Embodiment 4.
- FIG. It is a figure which shows the hardware constitutions of the noisy judgment apparatus 4 of the speech sound listening by Embodiment 4.
- 10 is a flowchart illustrating a processing procedure of a speech sound annoyance determination system 400 according to a fourth embodiment.
- the speech sound annoyance determination system is used to evaluate whether the user feels noisy when listening to the speech using the brain wave as a user state at the time of listening to the speech. More specifically, the present system presents a single syllable word sound as a voice, and evaluates the annoyance of the word sound using an event-related potential of a user brain wave measured from the voice presentation as an index.
- An “event-related potential (ERP)” is a part of an electroencephalogram (EEG), and is a transient potential generated in time in relation to an external or internal event. Refers to fluctuations.
- “presenting a voice” means outputting an auditory stimulus (also referred to as a “voice stimulus”). For example, sound is output from a speaker.
- the type of the speaker is arbitrary, and a speaker installed on a floor or a stand may be used, or a headphone type speaker may be used. However, the speaker needs to be able to output accurately at the sound pressure specified for correct evaluation. “Evaluation” is also used as a meaning of “determination”.
- the speech intelligibility evaluation is a short-time evaluation performed at a hearing aid store. Whether or not the user feels noisy is not an object to be evaluated. Therefore, even if the user feels a little noisy if he / she cannot tolerate, he / she will endure and perform the evaluation task.
- the inventors of the present application believe that the user status when listening to speech should be evaluated separately when “tolerance against annoyance” is not necessary and when “tolerance against annoyance” is necessary. It was. Since annoyance is a process in the brain at the time of listening to speech, there is a possibility that it can be evaluated by measuring an electroencephalogram.
- the inventors of the present invention have found that in the event-related potential starting from the voice presentation, the latency of the negative component (N1 component) having a latency of about 200 ms is shortened with an increase in annoyance to the voice. Furthermore, it has been found that the latency of the N1 component varies depending on the difference in characteristics such as consonant length or consonant intensity for each word sound.
- the “latency” refers to the time until the peak of the positive component or negative component appears from the time when the voice stimulus is presented.
- the present inventors have found that it is possible to determine the annoyance of speech listening from the latency of the negative component (N1 component) having a latency of about 200 ms. With this method, it is possible to objectively and quantitatively evaluate for each speech sound whether it was felt noisy as the user state of speech sound listening.
- EEG measurement experiment we investigated the relationship between the subjective report about annoyance obtained after voice presentation and the event-related potential from the voice.
- FIG. 1 to FIG. 6 experimental settings and experimental results of an electroencephalogram measurement experiment will be described.
- the experiment participants were 13 university / graduate students with normal hearing.
- Fig. 1 shows an outline of the experimental procedure for an electroencephalogram measurement experiment.
- procedure A a single syllable voice was presented. Details of the presented voice will be described later.
- procedure B the participant was made to hear the voice and the hiragana corresponding to the heard voice was written down. Only the type of speech was changed without changing the conditions of the presented speech.
- Procedures A and B were repeated 5 times.
- Procedure C the participants were subjected to subjective evaluation regarding the annoyance and the like with respect to the voice presented in Procedure A. Subjective evaluation was performed on a visual analog scale (100-level evaluation) using a touch panel.
- the above procedure A to procedure C were repeated as 12 blocks and repeated 12 blocks (total 60 trials). The sound pressure and distortion conditions of the voice presented for each block were changed in a random order.
- FIG. 2 is a flowchart showing the procedure for one block.
- step S11 a single syllable voice is presented to the experiment participant.
- step S12 the participant hears a single syllable and thinks of the corresponding character.
- step S13 the character corresponding to the voice heard by the participant is written down.
- step S14 the number of voice presentations is counted.
- the process returns to step S11 until the number of presentations is four.
- the process proceeds to step S15 and the number of presentations is reset.
- step S15 the participant answers the subjectivity to the voice heard in step S11.
- the speech sounds presented as stimuli are 8 voices (Ki, Ku, Si, Su, Ta, Te, T C).
- Frequency gain means gain (circuit gain, amplification factor) for each of a plurality of frequency bands.
- LF Large Flat
- LD Large Distorted
- MF Middle Flat
- MD Middle Distorted
- FIG. 3 (a) shows the classification of 6 conditions of speech and distortion.
- FIG. 3B shows a gain adjustment amount for each frequency. The reason why the frequency gain in the high frequency band is reduced is to simulate a high-tone gradual type that is a typical pattern of elderly hearing loss. The voice stimulus was presented from a speaker with a flat frequency characteristic.
- FIG. 4 (a) shows the electrode positions of the international 10-20 method (10-20 system), and FIG. 4 (b) shows the electrode arrangement in which the electrodes were mounted in this experiment.
- the sampling frequency was 200 Hz and the time constant was 1 second.
- a 0.05-20 Hz digital bandpass filter was applied off-line.
- a waveform of ⁇ 200 ms to 1000 ms was cut out from the time when the voice was presented.
- ⁇ 200 ms refers to a time point that is 200 milliseconds before the time when the voice is presented.
- the results of subjective evaluation are shown. Based on the subjective evaluation results, “noisy” / “not noisy” labeling was performed based on the threshold for each participant determined by the method described later.
- the label of the subjective evaluation is a user state at the time of listening to the speech.
- Fig. 5 shows the result of subjective evaluation for each participant regarding annoyance.
- the horizontal axis is the subjective evaluation value (1 to 100 on the visual analog scale), and the vertical axis is the ratio (0 to 1) obtained by dividing the frequency distribution by the total number of trials.
- FIG. 5 shows the percentage of all trials.
- the solid line in FIG. 5 is the distribution of the subjective evaluation results, and the broken line indicates the threshold value obtained by dividing the subjective evaluation (“noisy” / “not noisy”).
- the inventors of the present application determined the threshold based on the ranking of evaluation results (1 to 100 on the visual analog scale) for each individual. Specifically, the inventors of the present application set the threshold value by setting the evaluation value rank within an individual to be “noisy” for one third from the highest evaluation value, and “not noisy” for others. However, the same evaluation results were treated as the same subjective evaluation.
- FIG. 6 shows event-related potentials starting from the time when voice stimulation was presented at the top of the head (Pz). Specifically, FIG. 6 shows a waveform obtained by performing the total addition averaging by dividing according to the criteria of “noisy” / “not noisy” labeled by the above-described method. The addition average was performed based on the subjective evaluation regarding the annoyance for every block in all the six conditions of the said measurement experiment.
- the horizontal axis in FIG. 6 is time and the unit is ms, and the vertical axis is potential and the unit is ⁇ V. As is clear from the scale shown in FIG. 6, the lower direction of the graph corresponds to positive (positive) and the upper direction corresponds to negative (negative).
- the solid line shown in FIG. 6 is the total addition average waveform in the case of “noisy”, and the broken line is the total addition average waveform in the case of “not noisy”.
- FIG. 6 shows that the latency of the negative component (N1 component) caused by the latency of about 200 ms is shorter in the solid line in the “noisy” case than in the broken line in the “noisy” case.
- the latency of the N1 component for each participant was 195 ms for “noisy” and 240 ms for “noisy”.
- p ⁇ 0.05 a significant difference was observed (p ⁇ 0.05). Therefore, it is possible to conclude that the latency was shorter in the “noisy” case than in the “not noisy” case, and the N1 component starting from the time when the presentation of the voice stimulus (hereinafter referred to as “voice presentation”) was performed.
- voice presentation the N1 component starting from the time when the presentation of the voice stimulus
- FIG. 7 shows an example of the speech waveform of the MF condition (8 words: Ki, K, Shi, Su, Ta, Te, G, and C) used in this experiment.
- the consonant length is the time length until the waveform rises relatively large after the generation start time (0 ms). More specifically, the consonant length is a time length until the vowel rises.
- the consonant length was about 80 ms for the Ka line, about 170 ms for the Sa line, about 40 ms for the Ta line, and 130 ms for the Ha line.
- the intensity of the consonant part was stronger in the Sa line than in the Ka Ta Ha line. While the voice waveforms for each row are greatly different, it can be seen that the characteristics of the entire waveform are similar for the same row.
- the K line indicates the sound that begins with k. Specifically, ka, ki, ku, ke, and ko are shown.
- the word “S” indicates a sound that starts with “s”. Specifically, sa, si, su, se, and so are shown.
- a line indicates a sound that begins with t. Specifically, ta, ti, tu, te, and to are shown.
- the latency and amplitude of the N1 component with respect to the word sound may be different for each word sound.
- the feature of the speech sound presented is the cause of the fluctuation of the latency of the N1 component
- the latency of the N1 component varies regardless of “noisiness”. For this reason, there is a risk of mis-evaluation such as annoying voices that are noisy or noisy voices.
- the inventors of the present application performed an averaging operation for each same line having similar voice waveform characteristics. A certain average number of times required for analysis of the latency of the N1 component was ensured by the addition average for each row. Then, from the result of the addition average waveform for each row, the influence of the difference in the characteristics of the voice stimulus on the latency of the N1 component was examined. In addition, in order to clarify the relationship between the sound pressure and the N1 component latency, addition averaging was performed for each sound pressure regardless of the presence or absence of distortion.
- FIGS. 8A to 8C sequentially show the results of the Ta row (Ta Te To), the Ha row (C), and the Sa row (Shi Su).
- the horizontal axis in FIG. 8 is time and the unit is ms, and the vertical axis is potential and the unit is ⁇ V.
- the thick solid line shown in FIG. 8 is the total addition average waveform in the case of the Large condition (85 dB), the broken line is the total addition average waveform in the case of the Middle condition (65 dB), and the thin solid line is in the case of the Small condition (45 dB). It is a total addition average waveform.
- the circles in FIG. 8 indicate the N1 component.
- the N1 component takes a minimum value in a section from 0 ms to 500 ms, for example. Also, in FIGS.
- FIG. 9 shows an example of a change due to sound pressure in the latency of the N1 component obtained by this experiment.
- FIG. 9 also shows the change due to the sound pressure of a pure tone according to conventional research.
- the following features can be found for (1) sound intensity and (2) consonant length.
- (1) sound intensity it can be seen that the N1 latency is shortened as the sound pressure increases in any row.
- (2) consonant length it can be seen that the latency of the N1 component differs depending on the consonant length, by comparing the result of the TA row and the result of the HA row.
- the consonant lengths of the TA row and the HA row are about 40 ms and about 130 ms, respectively, and there is a difference of about 90 ms.
- the latency of the Ta row and the C row under the Small condition is 236 ms and 322 ms, and a difference of about 90 ms is stored. Therefore, it can be considered that this N1 component is caused by the rise of the vowel.
- the influence of the consonant length was different in the Sa line where the consonant intensity was stronger than the other lines.
- the consonant length of the Sa line is about 170 ms, which is longer than the consonant length of the Ha line (about 130 ms), but the latency of the N1 component was shorter in the Sa line than the Ha line in all sound pressure conditions. Further, the latency of the N1 component in the large condition of the Sa line was 156 ms, which was shorter than the consonant length of the Sa line.
- FIG. 10 shows, as an example of the consonant intensity of the speech sound presented as a stimulus, the sum of the mean square amplitude of the consonant part in the MF condition up to that time. From FIG. 10, it can be seen that in the Sa line, the intensity of the consonant part is higher than that of other word sounds.
- “Consonant intensity” means the amplitude of the consonant part of the speech waveform in the time zone. The consonant intensity may be obtained by, for example, a mean square value of the amplitude of the consonant part.
- FIG. 11 shows examples of speech waveforms of a consonant with a weak consonant intensity and a strong consonant. It can be seen that the consonant intensity of ma line “mo” and na line “ne” is stronger than the line “di” and ra line “li”. In addition, it can be said that the consonant intensity is strong, for example, for the word sounds of the ma line, na line, ya line, wa line and ga line.
- an electroencephalogram measurement experiment has revealed the existence of an electroencephalogram component that reflects a subjective evaluation of the user's annoyance with respect to speech listening. Specifically, it was discovered that annoyance is reflected in a negative potential having a peak at a latency of about 200 ms. It was also found that the latency of the N1 component is different for each speech sound due to the influence of the difference in the speech waveform.
- the latency of the N1 component (FIG. 6) with a latency of about 200 ms for each subjective evaluation regarding annoyance at the top of the head (Pz) described above may be, for example, the time when the minimum potential is taken in the corresponding section.
- a template may be created from a typical N1 component waveform, and the peak time of the N1 component of the template when the degree of coincidence between the template and the measured waveform is the highest may be used as the latency.
- the latency of the N1 component can be identified by a method of comparing with a predetermined threshold related to latency. Note that the threshold value / template may be a typical user's previously stored or may be created for each individual.
- negative components can be identified by non-addition or a few additions of several times by devising a feature amount extraction method (for example, wavelet transform of a waveform) or an identification method (for example, support vector machine learning).
- a feature amount extraction method for example, wavelet transform of a waveform
- an identification method for example, support vector machine learning
- a time after a predetermined time elapsed from a certain time point in order to define a component of an event-related potential is expressed as, for example, “latency about 200 ms”. This means that a range centered on a specific time of 200 ms can be included.
- EMP Event-Related Potential
- the terms “about Xms” and “near Xms” mean that a width of 30 to 50 ms can exist around the Xms (for example, 300 ms ⁇ 30 ms, 750 ms ⁇ 50 ms).
- the latency of the N1 component changes according to the characteristics of the speech sound. Therefore, it is preferable to handle it as having a wider width, for example, about 150 ms before and after, including at least the consonant length variation (0 ms: vowel to about 200 ms: consonant). Therefore, in this embodiment, “latency about 200 ms” means a latency included in the range of 50 ms to 350 ms.
- the electroencephalogram measurement experiment conducted by the inventors of the present application revealed that the negative component (N1 component) having a latency of about 200 ms reflects annoyance in the event-related potential starting from the time when the voice was presented. It was. Therefore, the subjective evaluation regarding the annoyance at the time of listening to the speech can be realized by using the event-related potential for voice presentation (voice stimulation) as an index.
- FIG. 12 shows the correspondence between the latency of the N1 component and the noisy determination summarized by the inventors of the present application.
- the latency of the N1 component is shorter than a predetermined threshold, it is determined as “noisy”.
- the latency of the N1 component is longer than a predetermined threshold, it is determined that it is “noisy”.
- negative component means a potential smaller than 0 ⁇ V.
- negative component in order to identify whether or not it is felt “noisy”, the case where the latency is shorter than a predetermined threshold is “negative component”, and the case where the latency is longer than the predetermined threshold is “negative” It is defined as “no ingredients”. A specific example of the threshold will be described later.
- the speech sound annoyance determination system presents a single syllable word sound sequentially as speech, and based on the latency of the negative component of the event-related potential latency of about 200 ms starting from the time when the speech was presented, the speech sound annoyance Make a decision. This is realized for the first time based on the discovery of the present inventors.
- the speech sound annoyance determination system sequentially presents voices and measures event-related potentials starting from each voice presentation time. Then, a negative component having a latency of about 200 ms is detected, and the annoyance of listening to the speech is determined.
- the exploration electrode is provided on the top of the head (Pz)
- the reference electrode is provided on either the left or right mastoid
- an electroencephalogram that is a potential difference between the exploration electrode and the reference electrode is measured.
- the level and polarity of the characteristic component of the event-related potential may vary depending on the part to which the electroencephalogram measurement electrode is attached and how to set the reference electrode and the exploration electrode.
- a person skilled in the art can make an appropriate modification according to the reference electrode and the exploration electrode at that time to detect the characteristic component of the event-related potential and evaluate the speech intelligibility It is. Such modifications are within the scope of the present invention.
- FIG. 13 shows the configuration and usage environment of the speech sound annoyance determination system 100 according to this embodiment.
- the speech sound annoyance determination system 100 is illustrated in correspondence with the system configuration of the first embodiment to be described later.
- the speech sound annoyance determination system 100 includes a speech sound annoyance determination device 1, a sound output unit 11, and a biological signal measurement unit 50.
- the biological signal measuring unit 50 is connected to at least two electrodes A and B.
- the electrode A is affixed to the mastoid of the user 5, and the electrode B is affixed to the top of the scalp of the user 5 (so-called Pz).
- the noisy judgment system 100 for listening to a speech presents a single syllable speech to the user 5 with a sound pressure of a certain sound pressure, and N1 having a latency of about 200 ms in the brain wave (event-related potential) of the user 5 measured from the speech presentation time. Whether the component latency is shorter than a predetermined threshold is identified. And based on the identification result regarding the latency of the present voice and the N1 component, it is determined whether the user feels noisy when listening to the speech.
- the brain wave of the user 5 is acquired by the biological signal measuring unit 50 based on the potential difference between the electrode A and the electrode B.
- the biological signal measurement unit 50 transmits information corresponding to the potential difference (electroencephalogram signal) to the annoyance determination device 1 for listening to the speech sound wirelessly or by wire.
- FIG. 13 illustrates an example in which the biological signal measurement unit 50 transmits the information to the annoyance determination device 1 for listening to speech wirelessly.
- the speech sound annoyance determination device 1 controls the sound pressure of the sound or the sound presentation timing for the speech sound annoyance determination, and presents the sound to the user 5 via the sound output unit 11 (for example, a speaker).
- the sound output unit 11 for example, a speaker
- FIG. 14 shows the hardware configuration of the speech sound annoyance judging device 1 according to the present embodiment.
- the speech sound annoyance determination device 1 includes a CPU 30, a memory 31, and an audio controller 32. These are connected to each other via a bus 34 and can exchange data with each other.
- the CPU 30 executes a computer program 35 stored in the memory 31.
- the computer program 35 describes a processing procedure shown in a flowchart described later.
- the speech listening annoyance determination device 1 uses the speech database (DB) 71 stored in the same memory 31 to perform processing for controlling the speech listening annoyance determination system 100 as a whole. . This process will be described in detail later.
- the audio controller 32 generates a sound to be presented in accordance with a command from the CPU 30, and outputs the generated sound signal to the sound output unit 11 with a designated sound pressure.
- the annoyance determination device 1 for listening to speech may be realized as hardware such as a DSP in which a computer program is incorporated in one semiconductor circuit.
- a DSP can realize all the functions of the CPU 30, the memory 31, and the audio controller 32 described above with a single integrated circuit.
- the computer program 35 described above can be recorded on a recording medium such as a CD-ROM and distributed as a product to the market, or can be transmitted through an electric communication line such as the Internet.
- a device for example, a PC
- the speech sound DB 71 may not be held in the memory 31 and may be stored in, for example, a hard disk (not shown) connected to the bus 34.
- FIG. 15 shows a functional block configuration of the speech sound annoyance determination system 100 according to the present embodiment.
- the speech sound listening annoyance determination system 100 includes an audio output unit 11, a biological signal measurement unit 50, and a speech sound listening annoyance determination device 1.
- FIG. 15 also shows detailed functional blocks of the annoyance determination device 1 for listening to speech. That is, the speech sound annoyance determination device 1 includes an event-related potential processing unit 55, an annoyance determination unit 65, a presented speech sound determination unit 70, a speech sound DB 71, and a result storage DB 80. Note that the block of the user 5 is shown for convenience of explanation.
- Each function block (except for the speech DB 71) of the speech listening annoyance determination device 1 is sometimes changed as a whole by the CPU 30, the memory 31, and the audio controller 32 by executing the program described with reference to FIG. It corresponds to the function realized in.
- the speech sound DB 71 is a speech sound database for determining the annoyance of speech sound listening.
- FIG. 16 shows an example of the speech sound DB 71 when, for example, the 20 word sounds of the 67S word table are used as the inspection sound sounds.
- the reference latency of the N1 component for each speech sound is stored in association with each other.
- the reference latency of the N1 component for each word sound is preferably a length corresponding to the length of the consonant included in the word sound or the intensity of the consonant.
- the sound file may be a standard test sound for 20 words in the 67S word table, for example, or may be a recorded sound in which a voice of the other party who mainly talks with a hearing aid is recorded. It is assumed that the stored sound has been subjected to gain adjustment (hearing aid processing) for each frequency based on a certain fitting theory from the audiogram of the hearing impaired measured in advance.
- 50 sounds in the 57S word table may be used in addition to the 20 sound in the 67S word table.
- the consonant label is used when the user 5 evaluates in which consonant the loudness is high.
- the reference latency of the N1 component is a threshold (unit: ms) of the latency of the N1 component for each speech, taking into account the influence of differences in the features of speech sounds.
- the latency of the N1 component measured for each word sound may be used at a sound pressure at which a general user feels noisy.
- the standard test voice for example, when using a recorded voice in which the voice of the other party who mainly talks wearing a hearing aid is used, it is calculated based on the consonant length and consonant intensity of the presented word sound, for example. A value may be set.
- the annoyance is determined by comparing the reference latency of the N1 component with the measured latency of the N1 component. A method for determining annoyance will be described later.
- the presentation speech sound determination unit 70 refers to the speech sound DB 71 and determines which speech sound is to be presented at which sound pressure.
- the presentation word sounds may be selected and determined in a random order, for example.
- the sound pressure of the presented speech is the sound pressure after the hearing aid processing is performed on the speech of 55 dB SPL, 65 dB SPL, 75 dB SPL, 85 dB SPL of the speech intelligibility curve measurement.
- the sound pressure may be changed in order from a small sound pressure to a large sound pressure, or vice versa. Further, the sound pressures may be selected in a random order.
- the presented speech sound determination unit 70 outputs a trigger to the biological signal measurement unit 50 in accordance with the voice presentation time, and transmits the content of the presented voice to the event-related potential processing unit 55.
- the voice output unit 11 presents the user 5 with a single syllable voice determined by the presentation word sound determination unit 70.
- the biological signal measuring unit 50 is an electroencephalograph that measures a biological signal of the user 5 and measures an electroencephalogram as a biological signal. Then, the EEG data is subjected to frequency filtering of a cutoff frequency suitable for extraction of the N1 component, and an event related to an electroencephalogram in a predetermined section (for example, a section from ⁇ 200 ms to 500 ms) is triggered by the trigger received from the presented word sound determination unit 70. The potential is cut out and the waveform data (electroencephalogram data) is sent to the event-related potential processing unit 55.
- the frequency of the N1 component is about 10 Hz.
- a bandpass filter used as the frequency filter, for example, it may be set so that components of brain waves from 5 Hz to 15 Hz are passed. It is assumed that the user 5 is wearing an electroencephalograph in advance.
- the electroencephalogram measurement electrode is attached to, for example, Pz at the top of the head.
- the event-related potential processing unit 55 performs an addition operation of the event-related potential received from the biological signal measuring unit 50 in accordance with the content of the presented voice received from the presented word sound determining unit 70.
- the event-related potential processing unit 55 selects only the event-related potential for voice presentation of the same word sound, for example, and performs an event-related potential addition operation for each type of word sound.
- event-related potentials are added only with the same word sound, annoyance can be determined for each word sound. Since the features of speech sounds are similar in speech sounds having the same consonant, addition may be performed by selecting event-related potentials of speech sounds having the same consonant. Further, word sounds whose difference in reference latency of the N1 component shown in FIG.
- the 16 is as small as 10 ms or less may be grouped and added.
- speech sounds having the same consonant it is possible to evaluate the annoyance of listening to the speech for each consonant type.
- addition when addition is performed for each word sound having a small difference in the reference latency of the N1 component, it is possible to determine annoyance for each group.
- An addition waveform in which the number of additions is ensured to some extent is obtained for each consonant and for each group having a small difference in the reference latency of the N1 component.
- the event-related potential processing unit 55 may obtain S (signal) / N (noise) using the N1 component as a signal. In the above experiment, the event-related potentials are added and averaged. However, if attention is paid only to the latency of the N1 component, the averaging process is unnecessary.
- the event-related potential processing unit 55 sends the electroencephalogram data obtained by executing a predetermined number of addition operations for each word sound to the annoyance determination unit 65.
- the noisy determination unit 65 receives the electroencephalogram data from the event-related potential processing unit 55 and performs analysis processing described later.
- the annoyance determination unit 65 determines whether or not the user feels noisy based on the latency of the N1 component at the latency of about 200 ms of the electroencephalogram data received from the event-related potential processing unit 55. For example, the annoyance determination unit 65 sets the time (hereinafter also referred to as “peak latency”) at which a negative potential peak is applied between 50 ms and 350 ms starting from the trigger received from the presented word sound determination unit 70 as the N1 component. The latency is compared with a predetermined reference latency (threshold value) stored in the speech sound DB 71.
- the annoyance determining unit 65 may determine not only the binary determination of “noisy” and “not noisy” but also the difference between the peak latency of the N1 component and the reference latency.
- the reference latency for each speech is set as a predetermined threshold.
- the reference latency for each row is set as a predetermined threshold, and when determining the annoyance for each group having a small difference in the reference latency, The reference latency may be set as a predetermined threshold value.
- the result accumulation DB 80 receives the voice information presented from the presented word sound determination unit 70. Further, the result accumulation DB 80 receives information on the result of annoyance determination for each speech from the annoyance determination unit 65. Then, for example, the information of the received annoyance determination result is accumulated for each speech and sound pressure of the presented voice.
- FIG. 17 is an example of data accumulation in the result accumulation DB 80.
- FIG. 17 illustrates an example in which information on annoyance is accumulated for each speech sound and for each sound pressure.
- “1” in FIG. 17 indicates a case where the noisy determination unit 65 determines that the latency of the N1 component is shorter than the reference latency and is “noisy”, and “0” indicates that the latency of the N1 component is The case where it is longer than the reference latency and is determined to be “noisy” is shown.
- FIG. 18A shows an example in which only the speech intelligibility is evaluated for each sound pressure of the presented voice during wearing.
- This example is an evaluation result obtained by a conventional evaluation method.
- the intelligibility is evaluated as 80% or more at a sound pressure of 65 dB SPL or more. Therefore, when the speech intelligibility is improved as compared with the non-wearing state (not shown in FIG. 18), it is determined to be suitable in the hearing aid compatibility test.
- 18 (b) and 18 (c) are examples of evaluation results obtained by adding annoyance determination according to the present embodiment in addition to the measurement results of the speech intelligibility curve at the time of wearing shown in FIG. 18 (a). It is assumed that the speech intelligibility curve is separately measured by a conventional method of performing speech or key input.
- the evaluation of annoyance is generally low. Therefore, it can be evaluated that this is a hearing aid process in which the user feels noisy.
- the evaluation of annoyance is generally high, and particularly high at a large sound pressure of 65 dB SPL or higher. Therefore, it can be evaluated as hearing aid processing that feels noisy at 65 dB SPL, which is the sound pressure of everyday conversation.
- the gain adjustment amount is increased overall. In the case of FIG. It is possible to propose a specific fitting procedure for reducing the gain adjustment amount and further increasing the compression in nonlinear amplification.
- 18 (b) and 18 (c) show the evaluation of annoyance only during wearing, but the annoyance is also evaluated during non-wearing (bare ears), and the annoyance is compared between non-wearing and wearing. May be.
- FIG. 19 is a flowchart illustrating a procedure of processing performed in the annoyance determination system 100 for listening to speech.
- step S101 the presented speech sound determination unit 70 determines the speech and sound pressure of a single syllable to be presented with reference to the speech sound DB 71.
- the voice output unit 11 presents the speech to the user 5 with the determined sound pressure.
- the presented word sound determination unit 70 transmits a trigger to the biological signal measurement unit 50, and transmits sound information related to the presented word sound to the event-related potential processing unit 55.
- the speech sounds to be presented may be selected randomly from the DB 71, or the speech sounds of specific consonants may be selected intensively.
- the sound pressure of the presented speech sound is, for example, the sound pressure after the hearing aid processing is performed on the speech of the speech intelligibility curve measurement of 55 dB SPL, 65 dB SPL, 75 dB SPL, and 85 dB SPL.
- the order in which the sound pressures are presented may be changed in order from a small sound pressure to a large sound pressure, or vice versa. Further, the sound pressures may be selected in a random order.
- step S102 the biological signal measuring unit 50 receives a trigger from the presented word sound determining unit 70, and cuts out an event-related potential from ⁇ 200 ms to 500 ms, for example, starting from the trigger in the measured electroencephalogram. Then, for example, an average potential of ⁇ 200 ms to 0 ms is obtained, and the obtained event-related potential is baseline-corrected so that the average potential becomes 0 ⁇ V.
- the biological signal measurement unit 50 always measures an electroencephalogram during the evaluation and applies a frequency filter suitable for extracting the N1 component to the electroencephalogram data.
- a suitable frequency filter is, for example, a bandpass filter that passes 5 Hz to 15 Hz around the center frequency 10 Hz of the N1 component. Note that when a high-pass filter of, for example, 5 Hz or higher is applied to the electroencephalogram data, baseline correction is not essential because it is hardly affected by a low-frequency baseline change.
- step S103 the event-related potential processing unit 55 adds the event-related potential cut out in step S102 for each word sound and sound pressure based on the information of the presented word sound received from the presented word sound determining unit 70.
- the addition calculation is performed for each word sound / sound pressure, but the target of the addition calculation is not limited to each word sound. For example, it may be performed for each consonant, for each group of word sounds having a small difference in reference latency, or for each sound pressure presented, depending on the type of word sound.
- the event-related potential processing unit 55 presents the sounds belonging to the same classification. What is necessary is just to add the event related electric potential of the electroencephalogram signal obtained when it is done.
- step S104 the event-related potential processing unit 55 determines whether or not the number of event-related potentials added to the word sound presented in step S101 has reached a predetermined number. If the number of additions is less than or equal to the predetermined number, the process returns to step S101 and repeats voice presentation. If the number of additions is equal to or greater than the predetermined number, the process proceeds to step S105.
- the predetermined number is, for example, 20 times. Note that “20 times” is the number of additions frequently used in the field of measuring event-related potentials, but this is only an example.
- the event-related potential processing unit 55 may obtain S (signal) / N (noise) using the N1 component as a signal, and the number of additions at which the S / N becomes a certain value or more may be a predetermined number.
- step S105 the event-related potential processing unit 55 sends the electroencephalogram data that has undergone a predetermined number of addition operations to the annoyance determination unit 65.
- the annoyance determination unit 65 obtains the latency of the N1 component of the electroencephalogram data received from the event-related potential processing unit 55, and compares it with the reference latency of the N1 component received from the speech sound DB 71.
- the latency of the N1 component in the electroencephalogram data is, for example, a time at which the potential is minimum in a section from 0 ms to 500 ms.
- the reference latency for each speech is compared with the latency of the N1 component in the electroencephalogram data.
- step S107 the noisy determination unit 65 determines that the user 5 feels noisy when the latency of the N1 component in the electroencephalogram data is shorter than the reference latency. On the other hand, when the latency of the N1 component in the electroencephalogram data is longer than the reference latency, it is determined that the user 5 feels noisy.
- step S108 the result accumulation DB 80 accumulates information of the annoyance determination result received from the annoyance determination unit 65 for each word sound and sound pressure presented in step S101.
- step S109 the presented speech sound determination unit 70 determines whether or not the stimulus presentation has been completed for all the speech sounds and sound pressures that are to be evaluated for the annoyance of speech listening. If it has not been completed, the process returns to step S101. If it has been completed, the noisy determination for listening to the speech is terminated.
- the annoyance of speech listening was evaluated by processing using the negative component of the latency of about 200 ms of the event-related potential starting from the time when the speech was presented with the setting of presenting the speech of a single syllable as speech. .
- the suitability of hearing aid processing can be evaluated with the annoyance of listening to speech different from speech intelligibility as an axis. Since it is possible to evaluate the hearing aid process with the annoyance as the axis, it is possible to realize a hearing aid process in which the user does not feel annoyed while listening to the speech and is not tired when listening.
- the speech sounding annoyance determination device 1 in the present embodiment is realized by a configuration that can be miniaturized using general hardware as shown in FIG.
- the user can evaluate the comfort of listening to speech in a sound environment in which the user uses a hearing aid by configuring the loudness determination device 1 in a size and weight that can be carried and being carried by the user.
- the audio output unit 11 is a speaker, but the audio output unit 11 may be a headphone. By using headphones, it is easy to carry and the speech intelligibility can be evaluated in the environment used by the user.
- the description has been made assuming Japanese evaluation.
- it may be in English or Chinese as long as it is a single syllable.
- a single syllable word may be presented and the evaluation for each word may be performed.
- Single syllable English words are short-time voices and are composed of consonants and vowels. Therefore, the reference latency can be determined for each word based on the consonant length and the consonant intensity in substantially the same manner as the above-described single syllable word sound in Japanese.
- FIG. 20 shows an example of the result of evaluating the annoyance for each single syllable word. “1” in FIG. 20 indicates that the user feels noisy, and “0” indicates that the user does not feel noisy.
- the speech sound annoyance determination system 100 it is possible to determine how loud (noisy) a user feels when listening to a sound just by listening to the voice. As a result, the “noisiness” that the user feels when listening to the speech is quantified, and the hearing aid process can be evaluated on the axis of noisy, and a fitting that does not feel noisy and is hard to listen to can be realized.
- the biological signal measurement unit 50 extracts an event-related potential in a predetermined range starting from the trigger from the presentation word sound determination unit 70, performs baseline correction, and converts the potential waveform data into the event. It is assumed that it is transmitted to the related potential processing unit 55. However, this process is an example. As another process, for example, the biological signal measurement unit 50 may continuously measure the brain waves, and the event-related potential processing unit 55 may perform necessary event-related potential extraction and baseline correction. If it is the said structure, the presentation sound determination part 70 does not need to transmit a trigger to the biological signal measurement part 50, and should just transmit a trigger to the event related electric potential process part 55. FIG.
- the result of the noisy determination is stored in the result storage DB 80, but it may not be stored.
- each determination result of the annoyance determination unit 65 may be simply output.
- Each determination result can be used as information regarding the annoyance of listening to the speech.
- the annoyance determination system shown in the present embodiment shows how annoying it feels for speech listening according to the latency of the negative component of the user brain wave latency of about 200 ms (more specifically, 50 ms to 350 ms) after voice presentation. Determine (noisy). At that time, by setting different reference latencies for each word sound, it is possible to increase the annoyance with high accuracy regardless of characteristics such as the length of the consonant part (consonant length) and the intensity of the consonant part (consonant intensity). evaluate. By determining the annoyance for listening to the speech sound, it is possible to select a hearing aid process in which the user does not feel annoyance and is less tired even if the hearing aid is worn for a long time.
- a description will be given of a speech sound annoyance determination system that includes a hearing aid processing unit that processes a speech sound to be presented into a sound output from a hearing aid and evaluates the annoyance of each hearing aid process.
- FIG. 21 shows a functional block configuration of the speech sound annoyance determination system 200 according to the present embodiment.
- the speech sound annoyance determination system 200 includes an audio output unit 11, a biological signal measurement unit 50, and a speech sound annoyance determination device 2.
- the same blocks as those in FIG. 15 are denoted by the same reference numerals, and the description thereof is omitted.
- the hardware configuration of the speech sound annoyance judging device 2 is as shown in FIG.
- the annoyance determination system 200 is also referred to as a hearing aid processing system.
- the point that the speech sound listening annoyance determination device 2 according to the present embodiment is greatly different from the speech sound listening annoyance determination device 1 according to the first embodiment is that a hearing aid processing unit 90 is newly provided.
- a hearing aid processing unit 90 is newly provided.
- the same names as those of the first embodiment are used for the component names of the annoyance determination device 2, but different reference numerals are used when they have different operations and / or functions.
- the event-related potential processing unit 55 in order to perform annoyance determination for each of a plurality of hearing aid processes, which is not performed in the first embodiment, the event-related potential processing unit 55, the presented word sound determination unit 70, the word sound DB 71, and the results according to the first embodiment
- an event-related potential processing unit 56 in this embodiment, a presentation word sound determination unit 75, a word sound DB 72, and a result accumulation DB 85 are provided.
- the speech sound DB 72 the presented speech sound determination unit 75, the hearing aid processing unit 90, the event-related potential processing unit 56, the result accumulation DB 85, and the hearing aid processing unit 90 will be described.
- the speech sound DB 72 is a speech sound database for determining the annoyance of listening to a speech sound such as the 20 speech sounds of the 67S-type word table shown in FIG. Similarly to the speech sound DB 71, information on the reference latency of the N1 component is also held for each speech sound.
- the difference between the speech sound DB 72 and the speech sound DB 71 is that the speech sound DB 72 has speech sound data before hearing aid processing.
- the presentation speech sound determination unit 75 refers to the speech sound DB similarly to the presentation speech sound control unit 70 according to the first embodiment, and determines the type and sound pressure of the speech sound.
- the difference between the presenting word sound determining unit 75 and the presenting word sound control unit 70 is that the presenting sound determining unit 75 selects which hearing processing to process the sound and sends the sound data of the presenting word sound to the hearing aid processing unit 90 together. It is.
- the hearing aid processing unit 90 processes the sound data by the instructed hearing aid processing method based on the instruction regarding the hearing aid processing selection received from the presentation word sound determination unit 75 and the sound data.
- Hearing aid processing includes, for example, consonant enhancement, directivity, noise reduction, and the like. If the consonant-enhanced hearing aid process is selected, for example, a process of increasing the gain amplification amount of the consonant frequency is performed to process the audio data.
- the hearing aid processing unit 90 may adjust the gain amplification amount of the sound based on the determination result of the annoyance determination unit 65. For example, the predetermined gain amplification amount is reduced with respect to the voice data of the speech that the noisy determination unit 65 determines to be noisy.
- the gain amplification amount is not adjusted with respect to the voice data of the speech that is determined to be noisy by the noisy determination unit 65.
- the gain amplification amount may be determined based on the difference between the peak latency of the N1 component and the reference latency determined by the noisy determination unit 65. For example, if the magnitude of the difference between the peak latency of the N1 component and the reference latency is within a predetermined range, the hearing aid processing unit 90 does not adjust the gain amplification amount. The greater the difference between the peak latency of the N1 component and the reference latency and the upper limit value or lower limit value of the predetermined range, the smaller the hearing aid processor 90 decreases the gain amplification amount.
- the event-related potential processing unit 56 receives the event-related potential received from the biological signal measurement unit 50 in accordance with the content of the presented voice received from the presentation word sound determination unit 75. Addition operation is performed.
- the difference between the event-related potential processing unit 56 and the event-related potential processing unit 55 is that the event-related potential processing unit 56 receives information on the hearing aid processing from the presented word sound determination unit 75 and receives each sound, sound pressure, and hearing aid processing. This is the point where the addition operation is performed.
- the result storage DB 85 stores information on the annoyance determination result based on the latency of the N1 component received from the annoyance determination unit 65, for example, for each speech and sound pressure.
- the difference between the result accumulation DB 85 and the result accumulation DB 80 is that the result accumulation DB 85 receives information on the type of hearing aid processing from the presentation word sound determination unit 75 in addition to the information on the sound and sound pressure of the presentation stimulus, and the type of hearing aid processing. It is a point to accumulate data every time.
- FIG. 22 (a) and 22 (b) are examples of data accumulation in the result accumulation DB 85.
- FIG. FIG. 22 illustrates a case where the determination results of annoyance are accumulated for each speech sound, each sound pressure, and each hearing aid process.
- FIG. 22A shows a pattern of hearing aid processing A
- FIG. 22B shows a pattern of hearing aid processing B.
- the evaluation results of the annoyance in the case of presenting each hearing aid processed speech are shown.
- “1” indicates that the annoyance determination unit 65 determines that the user 5 felt that the latency of the N1 component was shorter than the reference latency
- “0” represents the latency of the N1 component. Shows a case where it is determined that the user 5 feels noisy longer than the reference latency. Comparing (a) and (b) in FIG. 22, it can be said that the number of “1” is smaller in (b) subjected to hearing aid processing with the pattern of hearing aid processing B, and the user does not feel noisy.
- FIG. 23 shows a processing procedure of the speech intelligibility system 200 according to the present embodiment.
- steps that perform the same process as the process (FIG. 19) of the speech sound annoyance determination system 100 are denoted by the same reference numerals, and description thereof is omitted.
- the processing of the speech listening annoyance determination system 200 according to the present embodiment is different from the processing of the speech listening annoyance determination system 100 according to Embodiment 1 in steps S201, S202, and S203.
- step S201 a single syllable speech processed by the designated hearing aid process is presented.
- step S202 an addition operation is performed for each speech sound / sound pressure / hearing aid process.
- step S203 the results are accumulated for each word sound, each voice, and each hearing aid process. Other steps have already been described with reference to FIG.
- the presented speech sound determination unit 75 refers to the speech sound DB 72 to determine the type and sound pressure of the speech to be presented, and acquires the data. Further, the presented word sound determination unit 75 determines the hearing aid process, and sends information related to the type of the hearing aid process and audio data to the hearing aid processor 90.
- the hearing aid processing unit 90 receives information on the type of hearing aid processing determined by the presentation word sound determination unit 75 and the sound data, and processes the sound data based on the designated hearing aid processing method.
- the audio output unit 11 presents the processed audio data to the user 5.
- step S ⁇ b> 202 the event-related potential processing unit 56 receives the information on the type of presentation word sound, sound pressure, and hearing aid processing received from the presentation word sound determination unit 75, and determines the event-related potential of the electroencephalogram measured by the biological signal measurement unit 50. For example, it is added for each word sound, each sound pressure, and each hearing aid process.
- step S203 the result accumulation DB is annoyance based on the latency of the N1 component determined by the annoyance determination unit 65 for each piece of information (sound type, sound pressure, and hearing aid processing) related to the presented word sound received from the presentation word sound determination unit 75. Are accumulated.
- An example of the result accumulation is as shown in FIG.
- comfort for listening to speech can be evaluated for each hearing aid processing such as consonant enhancement, directivity, and noise reduction.
- hearing aid processing it is assumed that a plurality of types of sounds that have been subjected to hearing aid processing are mixed and presented in a random order. For example, after hearing aid processing A is performed, hearing aid processing B is performed. Furthermore, you may perform annoyance determination in order for every kind of hearing aid process. When implemented for each type of hearing aid processing, there is a feature that the parameter of the next hearing aid processing can be changed according to the annoyance determination result.
- the annoyance for each hearing aid process can be evaluated. As a result, it is possible to realize selection of hearing aid processing in accordance with the purpose of wearing the hearing aid and the usage environment.
- the noisy judgment unit 65 compares the reference latency of the N1 component for each speech sound of a general user with the latency of the N1 component of the measured electroencephalogram data. Annoyance determination was performed.
- the N1 component is an initial component of an event-related potential called an evoked potential, and the individual difference regarding latency / amplitude is considered to be relatively small.
- the N1 component is not completely free of individual differences in latency / amplitude. Therefore, in the identification based on the reference latency obtained from the latency of the N1 component for each general user's speech, there is a limit in determining annoyance with higher accuracy.
- calibration is performed to measure the reference latency of the N1 component for each user before determining the annoyance of listening to the speech, and the annoyance is evaluated based on the characteristics of the N1 component for each individual.
- FIG. 24 shows a functional block configuration of a speech sound annoyance determination system 300 according to this embodiment.
- the speech sound annoyance determination system 300 includes an audio output unit 11, a biological signal measurement unit 50, and a speech sound annoyance determination device 3.
- the same blocks as those in FIG. 15 are denoted by the same reference numerals, and the description thereof is omitted.
- the hardware configuration of the speech sound listening annoyance determination device 3 is as shown in FIG.
- the speech sound annoyance determination device 3 By executing a program that defines processing different from the program 35 (FIG. 14) described in the first embodiment, the speech sound annoyance determination device 3 according to the present embodiment shown in FIG. 24 is realized.
- the speech sound annoyance judging device 3 is greatly different from the speech sound annoyance judging device 1 according to the first embodiment in that a calibration / evaluation switching unit 95 and a reference latency calculating unit 96 are newly provided. It is.
- the presented word sound determination unit 70, the event-related potential processing unit 55, the word sound DB 71 of the first embodiment instead of (FIG. 15), a presenting speech sound determination unit 77, an event related potential processing unit 57, and a speech sound DB 73 are provided.
- the proofreading / evaluation switching unit 95 the reference latency calculation unit 96, the presented speech sound control unit 77, the event-related potential processing unit 57, and the speech sound DB 73 will be described.
- the proofreading / evaluation switching unit 95 is a calibration mode for specifying the reference latency of the N1 component with respect to each word sound for each user, and for determining annoyance based on the specified reference latency and the measured latency of the N1 component. Switch between evaluation modes. Then, information indicating the current mode is sent to the presentation word sound determination unit 77.
- the mode switching may be performed at the time when the reference latency for each speech is written in the speech DB 73, or a predetermined number of times necessary to specify the reference latency of the N1 component for each speech of the user brain wave. You may implement when the speech presentation is completed.
- the presented speech sound determination unit 77 refers to the speech sound DB 73 to determine the type of the speech sound and the sound pressure of the presented speech, and outputs the information to the user 5 via the speech output unit 11 and at the same time provides trigger information to the biological signal measurement unit 50. Send. Further, the information on the calibration mode / evaluation mode is received from the calibration / evaluation switching unit 95, and the audio information and the information on the configuration / evaluation mode are sent to the event-related potential processing unit 57. The presented word sound determination unit 77 switches the operation according to the mode received from the proofreading / evaluation switching unit 95. In the calibration mode, for example, a voice of a vowel (single syllable with no consonant part) is presented at a predetermined sound pressure.
- the predetermined sound pressure is a sound pressure equal to or higher than a threshold at which the user can hear the voice. For example, it may be a sound pressure that the user feels “noisy”.
- the sound pressure that can be heard by the user / the sound pressure that the user feels noisy may be determined with reference to the user's audiogram, for example, or may be measured in advance by subjective evaluation.
- the speech is presented at a predetermined sound pressure, similar to the presented speech control unit 70.
- the event-related potential processing unit 57 performs an addition operation of the event-related potential received from the biological signal measurement unit 50 according to the content of the presented voice received from the presentation word sound determination unit 77. .
- the event-related potential processing unit 57 receives mode information from the presentation word sound determination unit 77 and switches the operation. In the calibration mode, for example, an addition operation is performed for each vowel, and when a predetermined number of additions are completed, an addition waveform for each vowel is sent to the reference latency calculation unit 96.
- the evaluation mode as with the event-related potential processing unit 55, an added waveform for each speech and sound pressure is sent to the annoyance determination unit 65.
- the reference latency calculation unit 96 receives the added waveform data for each vowel from the event-related potential processing unit 57, and obtains the latency of the N1 component at a latency of about 200 ms.
- the latency of the N1 component is, for example, a time at which the potential becomes minimum at a latency of 50 ms to 350 ms.
- the latency of the N1 component for each obtained vowel is set as the reference latency of each vowel.
- the reference latency of the N1 component can be obtained for each vowel according to the difference in sound characteristics for each vowel. Further, the reference latency of the consonant including the consonant part is obtained by adding a predetermined positive value according to the characteristics of the consonant part in the reference latency of each vowel.
- the predetermined positive value is determined for each consonant. For example, in the case of a consonant with a weak consonant part, the consonant length may be set to a predetermined positive value. For example, in the case of a consonant having a strong consonant part, the time until the consonant part becomes equal to or greater than a predetermined value may be a predetermined positive value. Then, the calculated result is written in the speech sound DB 73.
- vowels have a smaller difference in speech characteristics between word sounds than vowels, so the latencies of the N1 component for all vowels may be averaged to serve as the reference latencies for the N1 component of the vowels.
- the latency of vowels for which the N1 component can be stably measured may be averaged to be the reference latency of the N1 component of the vowel.
- the speech sound DB 73 is a speech sound database for determining the annoyance of speech sound listening, similar to the speech sound DB 71 shown as an example in FIG.
- the difference between the speech sound DB 73 and the speech sound DB 71 is that the reference latency of the N1 component can be rewritten. Until the reference latency of the N1 component is rewritten by the reference latency calculation unit 96, 0 indicating that the reference latency is not set for each speech may be held.
- FIG. 25 shows a processing procedure of the speech intelligibility system 300 according to the present embodiment.
- steps that perform the same processing as the processing (FIG. 19) of the speech sound annoyance determination system 100 are denoted by the same reference numerals, and description thereof is omitted.
- the processing of the speech listening annoyance determination system 300 according to the present embodiment is different from the processing of the speech listening annoyance determination system 100 according to Embodiment 1 in steps S301 to S306. Since the other steps have already been described in relation to FIG.
- step S301 the proofreading / evaluation switching unit 95 sets the current mode to the proofreading mode, and sends the information on the proofreading mode to the presentation word sound determining unit 77.
- the proofreading / evaluation switching unit 95 may select the configuration mode when the reference latency value of the N1 component is 0 with reference to the speech sound DB 73. Further, the calibration mode may be set until a predetermined number of voice presentations are completed. The calibration / evaluation mode switching may be performed by selection of a hearing aid fitting expert or the user 5.
- the presented speech sound determination unit 77 refers to the speech sound DB 73, selects, for example, a vowel, and outputs it to the user 5 via the speech output unit 11 with a predetermined sound pressure.
- the predetermined sound pressure is a sound pressure that is equal to or higher than a threshold at which the user can hear the voice. For example, it may be a sound pressure that the user feels “noisy”. The sound pressure that can be heard by the user / the sound pressure that the user feels noisy may be determined with reference to the user's audiogram, for example.
- step S303 the event-related potential processing unit 57 adds the event-related potential measured by the biological signal measuring unit 50 for each vowel.
- the reference latency calculation unit 96 receives the added waveform data for each vowel from the event-related potential processing unit 57, and obtains the latency of the N1 component at the latency of about 200 ms.
- the latency of the N1 component is, for example, a time at which the potential becomes minimum at a latency of 50 ms to 350 ms.
- the presented word sound determination unit 77 sets the sound pressure to a sound pressure at which the user feels “noisy”
- the latency of the N1 component for each obtained vowel is set as the reference latency of the vowel.
- the reference latency of the consonant including the consonant part is obtained by adding a predetermined positive value corresponding to the characteristic of the consonant part in the reference latency of the vowel.
- the predetermined positive value is determined for each consonant.
- the consonant length may be set to a predetermined positive value.
- the time until the consonant part becomes equal to or greater than a predetermined value may be a predetermined positive value.
- step S305 the reference latency calculation unit 96 writes the reference latency of the N1 component for each word sound calculated in step S304 in the word sound DB 73.
- step S306 the proofreading / evaluation switching unit 95 detects that the reference latency of the N1 component is written in the speech sound DB 73, switches from the proofreading mode to the evaluation mode, and presents information on the evaluation mode. It is sent to the speech sound determination unit 77.
- the mode switching may be performed when a predetermined number of voice presentations have been completed, or may be performed by an operation input by a hearing aid fitting expert or the user 5.
- Such processing makes it possible to compare the reference latency of the N1 component for each word sound for each user with the latency of the N1 component of the measured electroencephalogram data. become able to.
- the vowel is selected by the presented word sound determination unit 77 in the calibration mode, the reference latency of the vowel is determined from the latency of the N1 component with respect to the vowel, and the reference latency of the consonant is calculated.
- the reference latency is calculated.
- all speech sounds subject to annoyance determination may be presented in the calibration mode, and the reference latency may be obtained for all speech sounds.
- a pure tone that is generally measured by an audiogram such as 250 Hz, 500 Hz, 1 kHz, 2 kHz, 4 kHz, etc. is held in the presentation word sound determination unit 77, the pure tone is presented in the calibration mode, and the latency of the N1 component with respect to the pure tone From the above, the reference latency for each speech may be calculated.
- the annoyance can be evaluated with high accuracy according to the brain wave characteristics of each user. As a result, it is possible to realize hearing aid processing in which the user does not feel noisy and is not tired when listening.
- the speech sound DB is described as being provided in the speech sound listening annoyance determination device, but this is not essential.
- the speech sound DB may be provided, for example, in a database server (not shown) or an external storage device connected to the speech sound annoyance determination device via a network.
- the speech sound annoyance determination system of each embodiment includes the database server and the external storage device.
- the N1 component of the measured electroencephalogram data is stored by storing the speech already recorded in the speech DB 71 and the reference latency of the N1 component for each speech according to the features of the speech sound.
- the noisy judgment was carried out by comparing with the latency.
- the recorded voice includes, for example, a standard test voice and a voice of a partner (hereinafter referred to as “speaker A”) who mainly talks while wearing a hearing aid.
- the loudness determination for the voice uttered by the speaker A is an evaluation of the voice that the speaker A talks to the user in daily life, and is important.
- the speech sound uttered by the speaker A is analyzed in real time, the reference latency of the N1 component of the user 5 is estimated according to the characteristics of the speech sound of the speaker A, and the estimated reference latency is estimated. And annoyance is evaluated by comparing the measured N1 component of the electroencephalogram with the reference latency.
- FIG. 26 shows a functional block configuration of the speech sound annoyance determination system 400 according to this embodiment.
- the speech sound annoyance determination system 400 includes a character output unit 12, a voice input unit 41, a biological signal measurement unit 50, and a speech sound annoyance determination device 4.
- the same blocks as those in FIG. 15 are denoted by the same reference numerals, and the description thereof is omitted.
- the speech sound listening annoyance determination system 400 according to the present embodiment is different from the speech sound listening annoyance determination system 100 according to the first embodiment in that a speech input unit 41 is newly provided and in place of the speech output unit 11.
- the character output unit 12 is provided.
- the speech listening annoyance determination device 4 has a function different from that of the speech listening annoyance determination device 1 with the addition of these components.
- the character output unit 12 is a display device that outputs character information of speech to the speaker A, for example, a liquid crystal display.
- character information the speech of a single syllable uttered by speaker A (for example, “a”, “da”, “shi”) is presented.
- information regarding the sound pressure of how loud the speaker A speaks may be indicated.
- the information on the sound pressure is, for example, “in a size that is usually spoken”, “in a loud voice”, or “in a small voice”.
- the voice input unit 41 is a microphone that collects the voice uttered by the speaker A.
- the annoyance judging device 4 for listening to speech will be described later.
- FIG. 27 shows a hardware configuration of the speech sound annoyance judging device 4.
- the same symbol is attached
- the difference from FIG. 14 is that a sound controller 46 is newly provided, a graphic controller 45 is provided instead of the audio controller 32, and a speech sound DB 71 is removed from the memory 31.
- the sound controller 46 utters the voice of the speaker A, A / D converts the voice waveform input from the voice input unit 41, and sends the obtained voice digital data to the CPU 30.
- the graphic controller 45 outputs the character information of the speech to be presented to the character output unit 12 according to the instruction of the CPU 30.
- the processing of the speech listening annoyance determination device 4 according to the present embodiment is realized by executing a program that defines processing different from the program 35 (FIG. 14) described in the first embodiment.
- the speech sound annoyance determination device 4 is greatly different from the speech sound annoyance determination device 1 according to the first embodiment in that a speech analysis unit 42 and a reference latency estimation unit 43 are newly provided. It is.
- the presented speech sound determination unit 78 and the noisy determination unit 66 determine the type of the speech sound uttered by the speaker A, and obtain the reference latency of the N1 component for the speech sound uttered by the speaker A. The noisy determination is performed based on the reference latency.
- the presentation word sound determination unit 78 the voice analysis unit 42, the reference latency estimation unit 43, and the annoyance determination unit 66 will be described.
- the presented speech sound determination unit 78 is prepared in advance, refers to the speech sound list held by itself, randomly determines the speech sound uttered by the speaker A, and outputs it to the speaker A via the character output unit 12.
- the determined speech information is sent to the event-related potential processing unit 55 and the reference latency estimation unit 43.
- the word sound list may be, for example, 20 sounds in the 67S word table or 50 sounds in the 57S word table.
- the voice analysis unit 42 detects the timing at which the speaker A uttered the voice from the sound information input to the voice input unit 41, and analyzes characteristics related to the consonant length, consonant intensity, vowel intensity, and the like of the voice. Then, a trigger is sent to the biological signal measuring unit 50 at the timing when the utterance of the speaker A is detected. Further, information about the voice feature is sent to the reference latency estimation unit 42.
- the reference latency estimation unit 43 estimates the reference latency of the N1 component for the word sound based on the information regarding the voice feature received from the voice analysis unit 42 and the word sound information received from the presented word sound determination unit 78.
- the annoyance determining unit 66 like the annoyance determining unit 65 of the first embodiment, has the user felt noisy based on the latency of the N1 component at the latency of about 200 ms of the electroencephalogram data received from the event-related potential processing unit 55? Determine whether or not. For example, the annoyance determination unit 66 compares the latency of the negative potential peak from the latency of 50 ms to 350 ms with the reference latency (threshold) estimated by the reference latency estimation unit 43. When the peak latency of the N1 component is shorter than the threshold, it is determined as “noisy”, and when the peak latency is smaller than the predetermined threshold, it is determined as “not too loud”.
- FIG. 28 shows a processing procedure of the speech intelligibility system 400 according to the present embodiment.
- steps that perform the same processing as the processing of the speech sound listening annoyance determination system 100 are denoted by the same reference numerals, and description thereof is omitted.
- the processing of the speech listening annoyance determination system 400 according to the present embodiment is different from the processing of the speech listening annoyance determination system 100 according to Embodiment 1 in steps S401 to S407. Since the other steps have already been described in relation to FIG.
- step S401 the presented speech sound determination unit 78 refers to the speech sound list held by the presented speech sound determination unit 78, randomly determines the type of the speech sound generated by the speaker A, and uses the determined speech sound as the character output unit. 12 to the speaker A.
- step S402 the voice analysis unit 42 detects the timing at which the speaker A utters the voice from the sound information input to the voice input unit 41, and sends a trigger to the biological signal measurement unit 50 at the detected timing.
- step S403 the voice analysis unit 42 analyzes characteristics related to the consonant length, consonant intensity, vowel intensity, and the like of the voice detected from the sound information input to the voice input unit 41. Then, the analysis result is sent to the reference latency estimation unit 43.
- the reference latency estimation unit 43 estimates the reference latency of the N1 component with respect to the speech based on the information regarding the speech features received from the speech analysis unit 42 and the speech information received from the presented speech sound determination unit 78. To do.
- the reference latency is estimated by adding a predetermined positive value based on the consonant length or consonant intensity of the speech to a predetermined base latency.
- the predetermined base latency may be, for example, an average N1 component latency of a general user when a vowel is heard at 90 dB SPL. Specifically, it may be 100 ms.
- the predetermined positive value is determined for each consonant.
- the consonant length analyzed by the speech analysis unit 42 may be set to a predetermined positive value.
- the time until the intensity of the consonant part analyzed by the speech analysis unit 42 or the intensity of the specific frequency in the consonant part exceeds a predetermined value is set to a predetermined positive value. It is good also as the value of.
- the strength of the consonant intensity may be determined based on the speech information received from the presented speech sound determination unit 78. Then, the estimated reference latency is sent to the annoyance determination unit 66.
- step S405 the event-related potential processing unit 55 sends the content of the presented voice received from the presented word sound determining unit 70 and the event-related potential received from the biological signal measuring unit 50 to the annoyance determining unit 66.
- the annoyance determining unit 66 determines whether or not the user feels noisy based on the latency of the N1 component at the latency of about 200 ms of the electroencephalogram data received from the event-related potential processing unit 55. For example, the annoyance determination unit 65 compares the latency of the negative potential peak from the latency of 50 ms to 350 ms with the reference latency received from the reference latency estimation unit 43. When the peak latency of the N1 component is shorter than the reference latency, it is determined as “noisy”, and when the peak latency is smaller than the reference latency, it is determined as “not too loud”.
- step S407 the result storage DB 80 receives the type of the speech sound from the presented speech sound determination unit 77, and receives information on the result of the noisy determination from the noisy determination unit 66. Then, for example, information on the determination result of annoyance is accumulated for each speech sound.
- the feature of the voice uttered by the speaker A is analyzed, the reference latency of the N1 component is estimated for each voice feature, and the estimated reference latency of the N1 component and the N1 component of the measured electroencephalogram data Comparison with the latency of Note that when the speaker A freely utters a single syllable speech and causes the speech analysis unit 42 to perform speech recognition, the presented speech sound determination unit 78 and the character output unit 12 may be omitted.
- the speech sound listening annoyance determination system 400 of the present embodiment it is possible to determine speech sound annoyance in real time using the voice uttered by the speaker A. As a result, it is possible to realize a hearing aid process in which the speaker A and the hearing aid fitting specialist are less troublesome and the user does not feel noisy and is not tired of listening.
- the voice analysis unit 42 may transmit a trigger to the event-related potential processing unit 55 instead of transmitting the trigger to the biological signal measurement unit 50.
- the biological signal measurement unit 50 may continuously measure the brain wave, and the event-related potential processing unit 55 may perform necessary event-related potential extraction and baseline correction.
- the annoyance determining unit performs the operation of “determining”
- the presentation word sound determining unit performs the operation of “determining”.
- these operations are expressions for convenience of human understanding, and are not intended to output the fact that the apparatus has positively “determined” and “determined” to the outside.
- the “noisiness determination unit” and the “presentation speech sound determination unit” as one component of the device may perform a predetermined process when a predetermined condition is satisfied.
- the annoyance determination unit may accumulate the result according to the classification that the user felt noisy, and the N1 component in the electroencephalogram data If the latency is longer than the reference latency, the results may be accumulated according to the classification that the user felt noisy.
- the presented speech sound determination unit may refer to the speech sound DB, select the speech sounds in a random order, select the sound pressures in a random order, and output them to the speech output unit. It should be noted that these processes are only described using expressions such as “determination” and “determination”.
- the speech sound annoyance determination device and the speech sound annoyance determination system in which the speech sound annoyance determination device of the present invention is incorporated in addition to the intelligibility of whether or not the sound can be discriminated, the electroencephalogram at the time of listening to the sound Based on the above, it is possible to quantitatively determine the annoyance of listening to speech. As a result, the user can select a hearing aid process that does not feel annoying and does not get tired while listening, so that it can be used for fitting all hearing aid users.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Heart & Thoracic Surgery (AREA)
- Surgery (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Veterinary Medicine (AREA)
- Biomedical Technology (AREA)
- Acoustics & Sound (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Public Health (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Multimedia (AREA)
- Psychology (AREA)
- Psychiatry (AREA)
- Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
Description
本願発明者らは、語音聴取のうるささ判定の実現を目指し、うるささを反映した脳波特徴成分を特定するために以下の実験を実施した。
脳波計測実験では、音声呈示後に取得したうるささに関する主観報告と、音声を起点とした事象関連電位との関係を調べた。以下、図1から図6を参照しながら、脳波計測実験の実験設定および実験結果を説明する。
以下では、まず、語音聴取のうるささ判定システムの概要を説明する。その後、語音聴取のうるささ判定装置を含む語音聴取のうるささ判定システムの構成および動作を説明する。
実施形態1による語音聴取のうるささ判定システム100では、語音DB71に保存されたある1種類の補聴処理に基づいてあらかじめ調整された所定の音声に対して、潜時約200msの陰性成分の有無に基づき、語音聴取時のうるささを判定した。
実施形態1による語音聴取のうるささ判定システム100では、うるささ判定部65において、一般的なユーザにおける語音ごとのN1成分の基準潜時と、計測した脳波データのN1成分の潜時との比較により、うるささ判定を実施した。
実施形態1による語音聴取のうるささ判定システム100では、語音DB71に収録済みの音声と、その語音音声の特徴に合わせた語音ごとのN1成分の基準潜時を保持し、計測した脳波データのN1成分の潜時との比較によりうるささ判定を実施した。収録済みの音声には、たとえば標準的なテスト音声、補聴器を装用して主に会話する相手(以下では「話者A」と呼ぶ。)の音声がある。話者Aが発声した音声に対するうるささ判定は、日常生活において話者Aがユーザにしゃべりかける音声の評価であり、重要である。
5 ユーザ
11 音声出力部
12 文字出力部
41 音声入力部
42 音声分析部
43 基準潜時推定部
50 生体信号計測部
55、56、57 事象関連電位処理部
65、66 うるささ判定部
70、75、77、78 呈示語音決定部
71、72、73 語音DB
80、85 結果蓄積DB
90 補聴処理部
95 評価切替部
96 基準潜時算出部
100、200、300、400 語音聴取のうるささ判定システム
Claims (20)
- ユーザの脳波信号を計測する生体信号計測部と、
複数の単音節の語音を保持する語音データベースであって、各語音について、語音と前記語音に対応する脳波信号の陰性成分の基準潜時とが対応付けられている語音データベースと、
前記語音データベースを参照して、呈示する単音節の語音を決定する呈示語音決定部と、
前記ユーザに、決定された語音を呈示する出力部と、
前記語音が呈示された時刻を起点として50ms以上350ms以下の範囲に含まれる前記脳波信号の陰性成分のピーク潜時と、前記語音データベースに保持されている、決定された前記語音に対応する基準潜時とを比較して、出力された前記語音のうるささを判定するうるささ判定部と
を備えたうるささ判定システム。 - 語音データベースは、前記語音に含まれる子音の長さ又は子音の強度に応じて、前記語音と前記語音に対応する脳波信号の陰性成分の基準潜時を対応付けている、請求項1に記載のうるささ判定システム。
- 前記うるささ判定部は、前記陰性成分のピーク潜時が前記基準潜時よりも小さい場合に、前記音声信号が前記ユーザにとってうるさいと判定し、前記陰性成分のピーク潜時が前記基準潜時よりも大きい場合に、前記音声信号が前記ユーザにとってうるさくないと判定する、請求項1に記載のうるささ判定システム。
- 前記脳波信号の事象関連電位を、所定の基準にしたがって加算する事象関連電位処理部をさらに備え、
前記呈示語音決定部は、複数の語音を決定し、
前記出力部は、決定された前記複数の語音を順次呈示し、
前記事象関連電位処理部は、前記複数の語音について、前記語音の種類又は前記語音呈示されたときの音圧が同じ語音が呈示された時刻を起点とした前記脳波信号の事象関連電位を加算して、前記加算した結果を前記うるささ判定部に出力する、請求項1に記載のうるささ判定システム。 - 前記うるささ判定部は、前記ピーク潜時として、前記決定された前記語音が呈示された時刻を起点として50ms以上350ms以下の範囲に含まれる前記脳波信号の陰性成分の最小の電位を取る時刻、または、前記脳波信号と予め用意されたN1成分の波形のテンプレートとの一致度が最も高くなる場合の、前記テンプレートに対応付けられたピーク潜時を採用する、請求項1に記載のうるささ判定システム。
- 前記事象関連電位処理部は、子音ごとまたは基準潜時の差が所定の値よりも小さい語音グループごとに事象関連電位を加算する、請求項5に記載のうるささ判定システム。
- 前記語音のうるささに関する判定結果を示す情報を蓄積する結果蓄積データベースをさらに備え、
前記結果蓄積データベースには、語音ごと、子音ごと、または、基準潜時の差が所定の値よりも小さい語音グループごとに、前記語音のうるささに関する判定結果を示す情報が蓄積されている、請求項1に記載のうるささ判定システム。 - 前記呈示された前記語音に関して補聴処理の種類を選択し、選択された補聴処理に基づいて前記語音データベースに保持されている語音データを加工する補聴処理部をさらに備えた、請求項1に記載のうるささ判定システム。
- ユーザごとの陰性成分の潜時を求める校正モードと、うるささを評価する評価モードとを切り替える切替部をさらに備え、
前記校正モードにおいて、前記切替部は、前記呈示語音決定部において母音を選択し、母音に対する前記陰性成分の潜時に基づいて語音ごとの基準潜時を算出し、
前記評価モードに切り替えられた後において、前記切替部は、前記陰性成分のピーク潜時と算出した前記基準潜時とを比較する、請求項1に記載のうるささ判定システム。 - 前記校正モードにおいて、前記評価切替部は、前記呈示語音決定部において母音が選択された場合に、母音に対するN1成分の潜時を母音の基準潜時に設定し、母音の基準潜時に子音部分の長さまたは強度に応じて決定された正の値を加えて、子音ごとの基準潜時を算出する、請求項9に記載のうるささ判定システム。
- 複数の単音節の語音を保持する語音データベースであって、各語音について、語音と前記語音に対応する脳波信号の陰性成分の基準潜時とが対応付けられている語音データベースと、
前記語音データベースを参照して、呈示する単音節の語音を決定する呈示語音決定部と、
生体信号計測部によって計測された前記ユーザの脳波信号のうち、前記語音が出力部によってユーザに呈示された時刻を起点として、50ms以上350ms以下の範囲に含まれる脳波信号の陰性成分のピーク潜時と、前記語音データベースに保持されている、決定された前記語音に対応する基準潜時とを比較した結果を出力するうるささ判定部と、
前記うるささ判定部が出力した、前記脳波信号の陰性成分のピーク潜時と前記語音に対応する基準潜時との差に基づいて、前記語音を調整する補聴処理部と
を備えた補聴処理システム。 - ユーザの脳波信号を計測するステップと、
複数の単音節の語音と、前記語音に対応する脳波信号の陰性成分の基準潜時とが対応付けて保持されている語音データベースを参照して、呈示する単音節の語音を決定するステップと、
前記ユーザに、決定された語音を呈示するステップと、
前記語音が呈示された時刻を起点として50ms以上350ms以下の範囲に含まれる前記脳波信号の陰性成分のピーク潜時と、前記語音データベースに保持されている、決定された前記語音に対応する基準潜時とを比較して、出力された前記語音のうるささを判定するステップと
を包含する、うるささ判定方法。 - 語音聴取のうるささ判定システムに設けられたコンピュータによって実行されるコンピュータプログラムであって、
前記コンピュータプログラムは、前記うるささ判定システムに実装されるコンピュータに対し、
ユーザの脳波信号を受け取るステップと、
複数の単音節の語音と、前記語音に対応する脳波信号の陰性成分の基準潜時とが対応付けて保持されている語音データベースを参照して、呈示する単音節の語音を決定するステップと、
前記ユーザに、決定された語音を呈示するステップと、
前記語音が呈示された時刻を起点として50ms以上350ms以下の範囲に含まれる前記脳波信号の陰性成分のピーク潜時と、前記語音データベースに保持されている、決定された前記語音に対応する基準潜時とを比較して、出力された前記語音のうるささを判定するステップと
を実行させる、コンピュータプログラム。 - ユーザの脳波信号を計測する生体信号計測部と、
特定話者が発声した音声信号を入力するための音声入力部と、
前記音声信号が入力されたタイミングを検出してトリガを出力し、さらに前記音声の子音部分の長さおよび強度に関する特徴を分析する音声分析部と、
前記音声分析部で分析された前記特徴に基づいて、陰性成分の基準潜時を推定する基準潜時推定部と、
前記トリガを起点として50ms以上350ms以下の範囲に含まれる前記脳波信号の陰性成分のピーク潜時と、前記基準潜時推定部によって推定された基準潜時とを比較してうるささを判定するうるささ判定部と
を備えたうるささ判定システム。 - 前記特定話者に発生させるための語音を示す文字情報を出力する文字出力部をさらに備え、
前記文字出力部によって出力された文字情報に基づいて前記特定話者が発生した音声信号が前記音声入力部に入力される、請求項14に記載のうるささ判定システム。 - 前記文字出力部は、前記単音節の語音を前記特定話者が発声する際の大きさを示す音圧に関する情報をさらに出力し、
前記文字出力部によって出力された文字情報および音圧に関する情報に基づいて前記特定話者が発生した音声信号が前記音声入力部に入力される、請求項15に記載のうるささ判定システム。 - 予め用意された語音リストを参照して、前記特定話者に発声させる語音を決定する呈示語音決定部をさらに備え、
前記文字出力部は、前記呈示語音決定部によって決定された前記語音を示す文字情報を出力する、請求項15に記載のうるささ判定システム。 - 前記基準潜時推定部は、前記音声分析部で分析された前記特徴、および、呈示語音決定部が決定した前記特定話者に発声させる語音に基づいて、前記陰性成分の基準潜時を推定する、請求項17に記載のうるささ判定システム。
- 前記基準潜時推定部は、予め用意されたベース潜時に、前記音声の子音長あるいは子音強度に基づいて決定される所定の正の値を加えることにより、前記陰性成分の基準潜時を推定する、請求項17に記載のうるささ判定システム。
- ユーザの脳波信号を計測するステップと、
特定話者が発声した音声信号を入力するステップと、
前記音声信号が入力されたタイミングを検出してトリガを出力し、さらに前記音声の子音部分の長さおよび強度に関する特徴を分析するステップと、
分析する前記ステップで分析された前記特徴に基づいて、陰性成分の基準潜時を推定するステップと、
前記トリガを起点として50ms以上350ms以下の範囲に含まれる前記脳波信号の陰性成分のピーク潜時と、推定する前記ステップによって推定された基準潜時とを比較してうるささを判定するステップと
を包含する、うるささ判定方法。
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2012512101A JP5144835B2 (ja) | 2010-11-24 | 2011-11-18 | うるささ判定システム、装置、方法およびプログラム |
| CN2011800378518A CN103053179A (zh) | 2010-11-24 | 2011-11-18 | 嘈杂度判定系统、装置、方法以及程序 |
| US13/630,113 US9149214B2 (en) | 2010-11-24 | 2012-09-28 | Annoyance judgment system, apparatus, method, and program |
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2010-261372 | 2010-11-24 | ||
| JP2010261373 | 2010-11-24 | ||
| JP2010-261373 | 2010-11-24 | ||
| JP2010261372 | 2010-11-24 |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/630,113 Continuation US9149214B2 (en) | 2010-11-24 | 2012-09-28 | Annoyance judgment system, apparatus, method, and program |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2012070212A1 true WO2012070212A1 (ja) | 2012-05-31 |
Family
ID=46145583
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2011/006435 Ceased WO2012070212A1 (ja) | 2010-11-24 | 2011-11-18 | うるささ判定システム、装置、方法およびプログラム |
Country Status (4)
| Country | Link |
|---|---|
| US (1) | US9149214B2 (ja) |
| JP (1) | JP5144835B2 (ja) |
| CN (1) | CN103053179A (ja) |
| WO (1) | WO2012070212A1 (ja) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2017189488A (ja) * | 2016-04-15 | 2017-10-19 | 日本電信電話株式会社 | コンテンツ評価装置、コンテンツ評価方法、プログラム |
| JPWO2019022242A1 (ja) * | 2017-07-28 | 2019-07-25 | 国立大学法人大阪大学 | 快不快の判別 |
| CN110068466A (zh) * | 2019-04-30 | 2019-07-30 | 山东理工大学 | 基于脑电波的车辆声品质评价方法 |
Families Citing this family (20)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10049336B2 (en) | 2013-02-14 | 2018-08-14 | Sociometric Solutions, Inc. | Social sensing and behavioral analysis system |
| US9443521B1 (en) * | 2013-02-14 | 2016-09-13 | Sociometric Solutions, Inc. | Methods for automatically analyzing conversational turn-taking patterns |
| JP2015040903A (ja) * | 2013-08-20 | 2015-03-02 | ソニー株式会社 | 音声処理装置、音声処理方法、及び、プログラム |
| CN105193430A (zh) * | 2015-08-31 | 2015-12-30 | 京东方科技集团股份有限公司 | 一种保护装置、可穿戴设备及保护方法和显示系统 |
| WO2018166625A1 (en) * | 2017-03-17 | 2018-09-20 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and appartus for adaptive audio signal alteration |
| EP3684463B1 (en) | 2017-09-19 | 2025-05-14 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement |
| US10580427B2 (en) * | 2017-10-30 | 2020-03-03 | Starkey Laboratories, Inc. | Ear-worn electronic device incorporating annoyance model driven selective active noise control |
| JP6337362B1 (ja) * | 2017-11-02 | 2018-06-06 | パナソニックIpマネジメント株式会社 | 認知機能評価装置、及び、認知機能評価システム |
| US11717686B2 (en) | 2017-12-04 | 2023-08-08 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to facilitate learning and performance |
| US12280219B2 (en) | 2017-12-31 | 2025-04-22 | NeuroLight, Inc. | Method and apparatus for neuroenhancement to enhance emotional response |
| US11273283B2 (en) | 2017-12-31 | 2022-03-15 | Neuroenhancement Lab, LLC | Method and apparatus for neuroenhancement to enhance emotional response |
| US11364361B2 (en) | 2018-04-20 | 2022-06-21 | Neuroenhancement Lab, LLC | System and method for inducing sleep by transplanting mental states |
| CA3112564A1 (en) | 2018-09-14 | 2020-03-19 | Neuroenhancement Lab, LLC | System and method of improving sleep |
| US10922044B2 (en) * | 2018-11-29 | 2021-02-16 | Bose Corporation | Wearable audio device capability demonstration |
| US10817251B2 (en) | 2018-11-29 | 2020-10-27 | Bose Corporation | Dynamic capability demonstration in wearable audio device |
| US10923098B2 (en) | 2019-02-13 | 2021-02-16 | Bose Corporation | Binaural recording-based demonstration of wearable audio device functions |
| US11786694B2 (en) | 2019-05-24 | 2023-10-17 | NeuroLight, Inc. | Device, method, and app for facilitating sleep |
| KR102774638B1 (ko) * | 2019-07-11 | 2025-02-27 | 현대자동차주식회사 | 오류 모니터링을 이용한 교통 감시 시스템 |
| EP3961624B1 (de) * | 2020-08-28 | 2024-09-25 | Sivantos Pte. Ltd. | Verfahren zum betrieb einer hörvorrichtung in abhängigkeit eines sprachsignals |
| US20230222359A1 (en) * | 2022-01-11 | 2023-07-13 | Dell Products L.P. | Conversational artificial intelligence system with live agent engagement based on automated frustration level monitoring |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS60219000A (ja) * | 1984-04-14 | 1985-11-01 | Rion Co Ltd | 補聴器 |
| JPH08275297A (ja) * | 1995-03-31 | 1996-10-18 | Rion Co Ltd | リモートコントローラ |
| WO2010073614A1 (ja) * | 2008-12-22 | 2010-07-01 | パナソニック株式会社 | 語音明瞭度評価システム、その方法およびそのコンピュータプログラム |
Family Cites Families (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH06114038A (ja) | 1992-10-05 | 1994-04-26 | Mitsui Petrochem Ind Ltd | 聴覚検査・訓練装置 |
| JPH0938069A (ja) | 1995-08-02 | 1997-02-10 | Nippon Telegr & Teleph Corp <Ntt> | 語音聴力検査方法およびこの方法を実施する装置 |
| ATE407622T1 (de) * | 2000-05-19 | 2008-09-15 | Baycrest Ct For Geriatric Care | Vorrichtung zur objektiven hörbewertung bei anwendung von auditiven stationären evozierten potentialen |
| US8311228B2 (en) * | 2006-06-01 | 2012-11-13 | Personics Holdings Inc. | Ear input sound pressure level monitoring system |
| US8165687B2 (en) * | 2008-02-26 | 2012-04-24 | Universidad Autonoma Metropolitana, Unidad Iztapalapa | Systems and methods for detecting and using an electrical cochlear response (“ECR”) in analyzing operation of a cochlear stimulation system |
| EP2164066B1 (en) | 2008-09-15 | 2016-03-09 | Oticon A/S | Noise spectrum tracking in noisy acoustical signals |
| JP5144836B2 (ja) * | 2010-06-11 | 2013-02-13 | パナソニック株式会社 | 語音聴取の評価システム、その方法およびそのプログラム |
-
2011
- 2011-11-18 JP JP2012512101A patent/JP5144835B2/ja active Active
- 2011-11-18 CN CN2011800378518A patent/CN103053179A/zh active Pending
- 2011-11-18 WO PCT/JP2011/006435 patent/WO2012070212A1/ja not_active Ceased
-
2012
- 2012-09-28 US US13/630,113 patent/US9149214B2/en active Active
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS60219000A (ja) * | 1984-04-14 | 1985-11-01 | Rion Co Ltd | 補聴器 |
| JPH08275297A (ja) * | 1995-03-31 | 1996-10-18 | Rion Co Ltd | リモートコントローラ |
| WO2010073614A1 (ja) * | 2008-12-22 | 2010-07-01 | パナソニック株式会社 | 語音明瞭度評価システム、その方法およびそのコンピュータプログラム |
Non-Patent Citations (2)
| Title |
|---|
| SHIN'YA KURISHIRO: "Measurements of auditory evoked neuromagnetic field using a multichannel SQUID magnetometer", JOURNAL OF THE ACOUSTICAL SOCIETY OF JAPAN, vol. 48, no. 5, 1 May 1992 (1992-05-01), pages 320 - 327 * |
| TAKAHIRO TAMESUE ET AL.: "Kioku Seishin Sagyoji no Soon ni Taisuru Shinri-Seiriteki Teiryo Hyoka ni Kansuru Kisoteki Kosatsu", REPORT OF THE 2009 SPRING MEETING, THE ACOUSTICAL SOCIETY OF JAPAN KOEN YOSHI-KOEN RONBUN CD-ROM, 10 March 2009 (2009-03-10), pages 1031 - 1032 * |
Cited By (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2017189488A (ja) * | 2016-04-15 | 2017-10-19 | 日本電信電話株式会社 | コンテンツ評価装置、コンテンツ評価方法、プログラム |
| JPWO2019022242A1 (ja) * | 2017-07-28 | 2019-07-25 | 国立大学法人大阪大学 | 快不快の判別 |
| JP2020073084A (ja) * | 2017-07-28 | 2020-05-14 | 国立大学法人大阪大学 | 快不快の判別 |
| JP7203388B2 (ja) | 2017-07-28 | 2023-01-13 | 国立大学法人大阪大学 | 快不快の判別 |
| US11690547B2 (en) | 2017-07-28 | 2023-07-04 | Osaka University | Discernment of comfort/discomfort |
| CN110068466A (zh) * | 2019-04-30 | 2019-07-30 | 山东理工大学 | 基于脑电波的车辆声品质评价方法 |
Also Published As
| Publication number | Publication date |
|---|---|
| US20130039498A1 (en) | 2013-02-14 |
| CN103053179A (zh) | 2013-04-17 |
| JP5144835B2 (ja) | 2013-02-13 |
| US9149214B2 (en) | 2015-10-06 |
| JPWO2012070212A1 (ja) | 2014-05-19 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| JP5144835B2 (ja) | うるささ判定システム、装置、方法およびプログラム | |
| JP5144836B2 (ja) | 語音聴取の評価システム、その方法およびそのプログラム | |
| JP4690507B2 (ja) | 語音明瞭度評価システム、その方法およびそのプログラム | |
| JP5042398B1 (ja) | 脳波記録装置、補聴器、脳波記録方法およびそのプログラム | |
| JP4769336B2 (ja) | 補聴器の調整装置、方法およびプログラム | |
| JP5002739B2 (ja) | 聴力判定システム、その方法およびそのプログラム | |
| JP5215508B1 (ja) | 不快閾値推定システム、方法およびそのプログラム、補聴器調整システムおよび不快閾値推定処理回路 | |
| JP4638558B2 (ja) | 語音明瞭度評価システム、その方法およびそのコンピュータプログラム | |
| US8849391B2 (en) | Speech sound intelligibility assessment system, and method and program therefor | |
| WO2012063423A1 (ja) | 音圧評価システム、その方法およびそのプログラム | |
| JPWO2013057929A1 (ja) | 不快音圧推定システム、不快音圧推定装置、不快音圧推定システムの作動方法およびそのコンピュータプログラム |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| WWE | Wipo information: entry into national phase |
Ref document number: 201180037851.8 Country of ref document: CN |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2012512101 Country of ref document: JP |
|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11842780 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 11842780 Country of ref document: EP Kind code of ref document: A1 |