Compensation for self-voice block

文档序号:197551 发布日期:2021-11-02 浏览:71次 中文

阅读说明:本技术 自身话音闭塞的补偿 (Compensation for self-voice block ) 是由 Z·陈 B·斯蒂尔 T·I·哈维 于 2020-03-12 设计创作,主要内容包括:一种均衡耳机中的声音的方法,该耳机包括被配置成生成第一音频信号的内部麦克风、被配置成生成第二音频信号的外部麦克风、扬声器以及耦合在该扬声器、该外部麦克风和该内部麦克风之间的一个或多个处理器,该方法包括:当该耳机由用户佩戴时:确定在该外部麦克风处存在声音的情况下该第一音频信号与该第二音频信号之间的第一音频传递函数;以及在扬声器由扬声器输入信号驱动的情况下确定扬声器输入信号与第一音频信号之间的第二音频传递函数;确定该一个或多个处理器的电传递函数;基于该第一音频传递函数、该第二音频传递函数和该电气传递函数来确定闭耳传递函数;以及基于该闭耳传递函数与开耳传递函数之间的比较来均衡该第一音频信号以生成均衡的第一音频信号。(A method of equalizing sound in a headset, the headset comprising an internal microphone configured to generate a first audio signal, an external microphone configured to generate a second audio signal, a speaker, and one or more processors coupled between the speaker, the external microphone, and the internal microphone, the method comprising: when the headset is worn by a user: determining a first audio transfer function between the first audio signal and the second audio signal in the presence of sound at the external microphone; and determining a second audio transfer function between the loudspeaker input signal and the first audio signal if the loudspeaker is driven by the loudspeaker input signal; determining an electrical transfer function of the one or more processors; determining a closed-ear transfer function based on the first audio transfer function, the second audio transfer function, and the electrical transfer function; and equalizing the first audio signal based on a comparison between the closed-ear transfer function and the open-ear transfer function to generate an equalized first audio signal.)

1. A method of equalizing sound in a headset, the headset comprising an internal microphone configured to generate a first audio signal, an external microphone configured to generate a second audio signal, a speaker, and one or more processors coupled between the speaker, the external microphone, and the internal microphone, the method comprising:

when the user wears the headset:

determining a first audio transfer function between the first audio signal and the second audio signal in the presence of sound at the external microphone; and

determining a second audio transfer function between a speaker input signal and the first audio signal, the speaker being driven by the speaker input signal;

determining an electrical transfer function of the one or more processors;

determining a closed-ear transfer function based on the first audio transfer function, the second audio transfer function, and the electrical transfer function; and

equalizing the first audio signal based on a comparison between the closed-ear transfer function and an open-ear transfer function to generate an equalized first audio signal.

2. The method of claim 1, wherein the comparison is a frequency domain ratio between the closed-ear transfer function and the open-ear transfer function, or wherein the comparison is a time domain difference between the closed-ear transfer function and the open-ear transfer function.

3. The method of claim 1, wherein:

a) the open-ear transfer function is an open-ear transfer function measured between the ear entrances or tympanic membranes of the user; or

b) The open-ear transfer function is an open-ear transfer function measured between an ear entrance of the head simulator and an eardrum; or

c) The open-ear transfer function is an average open-ear transfer function of a portion of the general population.

4. The method of claim 1, further comprising:

a) measuring an open-ear transfer function between the ear entrance or tympanic membrane of the user; or

b) Measuring the open-ear transfer function between an ear entrance of a head simulator and an eardrum; or

c) Determining the open-ear transfer function based on an average open-ear transfer function of a portion of the general population.

5. The method according to any of the preceding claims, wherein the step of determining the first audio transfer function is performed with the loudspeaker muted.

6. The method according to any of the preceding claims, wherein the step of determining the second audio transfer function is performed in the presence of little or no sound outside the headphone.

7. The method of any of the preceding claims, wherein determining the electrical path transfer function comprises determining a frequency response of a feedforward ANC filter implemented by the one or more processors and/or a frequency response of a feedback ANC filter implemented by the one or more processors.

8. The method of any of the preceding claims, wherein determining the electrical path transfer function comprises determining a gain associated with the one or more processors.

9. The method of any one of the preceding claims, wherein determining the open-ear transfer function between the user's ear entrance and tympanic membrane comprises approximating the open-ear transfer function of the user.

10. The method of any of the preceding claims, further comprising:

outputting the equalized first audio signal to the speaker.

11. The method of any of the preceding claims, further comprising:

determining a third audio transfer function between the first audio signal and the second audio signal when the user is wearing headphones and the user is speaking; and

further equalizing the equalized first audio signal based on the third transfer function.

12. The method of claim 11, further comprising:

upon determining that the user is speaking, outputting a voice equalized first audio signal to the speaker.

13. The method of claim 11 or 12, further comprising:

determining that the one or more processors are implementing Active Noise Cancellation (ANC); and

adjusting the further equalization to account for the one or more processors implementing ANC.

14. The method according to any one of claims 11-13, further comprising:

requesting the user to speak a phoneme balanced sentence or phrase;

wherein the third audio transfer function is determined when the user speaks the phoneme balance sentence.

15. An apparatus, comprising:

a headset, comprising:

an internal microphone configured to generate a first audio signal;

an external microphone configured to generate a second audio signal; a speaker; and

one or more processors configured to:

when the user wears the headset:

determining a first audio transfer function between the first audio signal and the second audio signal in the presence of sound at the external microphone;

determining a second audio transfer function between a speaker input signal and the first audio signal, the speaker being driven by the speaker input signal;

determining an electrical transfer function of the one or more processors;

determining a closed-ear transfer function based on the first audio transfer function, the second audio transfer function, and the electrical transfer function; and

equalizing the first audio signal based on a comparison between the closed-ear transfer function and an open-ear transfer function to generate an equalized first audio signal.

16. The apparatus of claim 15, wherein the comparison is a frequency domain ratio between the closed-ear transfer function and the open-ear transfer function, or wherein the comparison is a time domain difference between the closed-ear transfer function and the open-ear transfer function.

17. The apparatus of claim 15 or 16, wherein:

a) the open-ear transfer function is an open-ear transfer function measured between the ear entrances or tympanic membranes of the user; or

b) The open-ear transfer function is an open-ear transfer function measured between an ear entrance of the head simulator and an eardrum; or

c) The open-ear transfer function is an average open-ear transfer function of a portion of the general population.

18. The device of claim 15 or 16, wherein the one or more processors are further configured to:

a) measuring an open-ear transfer function between the ear entrance or tympanic membrane of the user; or

b) Measuring the open-ear transfer function between an ear entrance of a head simulator and an eardrum; or

c) Determining the open-ear transfer function based on an average open-ear transfer function of a portion of the general population.

19. The apparatus of any of claims 15-18, wherein the step of determining the first audio transfer function is performed with the speaker muted.

20. The apparatus of any of claims 15-19, wherein the step of determining the second audio transfer function is performed in the presence of little or no sound external to the headset.

21. The device of any one of claims 15-20, wherein determining the electrical path transfer function comprises determining a gain associated with the one or more processors.

22. The device of any one of claims 15-21, wherein determining the electrical path transfer function comprises determining a gain associated with the one or more processors.

23. The device of any one of claims 15-22, wherein determining the electrical path transfer function comprises determining a gain associated with the one or more processors.

24. The apparatus of any of claims 15-23, wherein determining the open-ear transfer function between the user's ear entrance and tympanic membrane comprises approaching the open-ear transfer function of the user.

25. The device of any of claims 14-24, wherein the one or more processors are further configured to:

outputting the equalized first audio signal to the speaker upon determining that the user is not speaking.

26. The device of any of claims 14-25, wherein the one or more processors are further configured to:

determining a third audio transfer function between the first audio signal and the second audio signal when the user is wearing headphones and the user is speaking; and

equalizing the equalized first audio signal further based on a difference between the open-ear transfer function and the closed-ear transfer function to generate a voice-equalized first audio signal.

27. The device of claim 26, wherein the one or more processors are further configured to: upon determining that the user is speaking, outputting the voice equalized first audio signal to the speaker.

28. The device of claim 26 or 27, wherein the one or more processors are further configured to:

determining that the one or more processors are implementing Active Noise Cancellation (ANC); and

adjusting the further equalization to account for the one or more processors implementing ANC.

29. The device of any one of claims 26-28, wherein the one or more processors are further configured to:

outputting a request to the user to speak a phoneme balanced sentence or phrase;

wherein the third audio transfer function is determined when the user speaks the phoneme balance sentence.

30. A method of equalizing sound in a headset, the headset comprising an internal microphone configured to generate a first audio signal, an external microphone configured to generate a second audio signal, a speaker, and one or more processors coupled between the speaker, the external microphone, and the internal microphone, the method comprising: determining a first audio transfer function between the first audio signal and the second audio signal when the user is wearing headphones and the user is speaking; and equalizing the first audio signal based on the first audio transfer function.

31. The method of claim 30, further comprising:

upon determining that the user is speaking, outputting a voice equalized first audio signal to the speaker.

32. The method of claim 30 or 31, further comprising:

determining that the one or more processors are implementing Active Noise Cancellation (ANC); and

adjusting the equalization to account for the ANC.

33. The method of any of claims 30 to 33, further comprising:

requesting the user to speak a phoneme balanced sentence or phrase;

wherein the first audio transfer function is determined when the user speaks a phoneme balance sentence.

34. An apparatus, comprising:

a headset, comprising:

an internal microphone configured to generate a first audio signal;

an external microphone configured to generate a second audio signal; non-viable cells

A speaker; and

one or more processors configured to:

determining a first audio transfer function between the first audio signal and the second audio signal when the user is wearing headphones and the user is speaking; and

equalizing the first audio signal based on a difference between the open-ear transfer function and the closed-ear transfer function to generate an equalized first audio signal.

35. The device of claim 34, wherein the one or more processors are configured to:

outputting an equalized first audio signal to the speaker upon determining that the user is speaking.

36. The device of claim 34 or 35, wherein the one or more processors are configured to:

determining that the one or more processors are implementing Active Noise Cancellation (ANC); and

adjusting the equalization to account for the ANC.

37. The device of any one of claims 34-36, wherein the one or more processors are configured to:

requesting the user to speak a phoneme balanced sentence or phrase;

wherein the first audio transfer function is determined when the user speaks a phoneme balance sentence.

38. The device of any of claims 1-14 and 34-37, wherein the headset comprises one or more of the one or more processors.

39. An electronic device comprising the apparatus of any of claims 1-14 and 34-37.

40. A non-transitory computer-readable storage medium storing instructions that, when executed by a computer, cause the computer to perform the method of any one of claims 1-14 and 30-33.

Technical Field

The present disclosure relates to methods and devices for compensating for ear occlusion.

Background

Many hearing devices, such as headphones, hearing aids, and hearing protectors, have tightly sealed ear plugs or ear muffs that occlude the ear and isolate the user from ambient noise. This isolation has two side effects when a user wants to listen to their Own Voice (OV), such as when making a phone call or talking to a nearby person without taking the device off their ear. One of the side effects is Passive Loss (PL) at high frequencies, which makes the user inaudible to his voice. Another effect is to amplify the user's own low frequency voice, which makes their voice sound buzzing. Amplification of the user's own voice at low frequencies is commonly referred to as an Occlusion Effect (OE).

OE occurs primarily below 1kHz and depends on the ear canal structure of the user, the fitting tightness of the hearing device and the phoneme of the user's pronunciation. For example, OE is typically only a few decibels (dB) for an open-front vowel, such as [ i: ], while OE can exceed 30dB for a closed-back vowel, such as [ i: ].

Feedback Active Noise Cancellation (ANC) is a common method for canceling headphone noise to compensate for OE. Feedback ANC uses an internal microphone and a headset speaker located near the eardrum to form a feedback loop to cancel sound near the eardrum. The use of feedback ANC to counteract OE is described in U.S. patent No. 4,985,925 and U.S. patent No. 5,267,321, the contents of each of which are incorporated herein by reference in their entirety. The methods described in these patents require that all parameters of the feedback ANC be preset based on the average OE of the user. Us patent No. 9,020,160 (the contents of which are incorporated herein by reference in their entirety) describes updating feedback loop variables of a feedback ANC filter to account for changes in the phenomenon expressed by the user.

The discussion of any document, act, material, device, article, etc. contained in this specification is not to be taken as an admission that any or all of these matters form part of the prior art base or were common general knowledge in the field relevant to the present disclosure as it existed before the priority date of each of the appended claims.

Disclosure of Invention

The present disclosure provides a method for restoring the naturalness of a user's own voice using novel signal analysis and processing.

According to an aspect of the present disclosure, there is provided a method of equalizing sound in a headset, the headset including an internal microphone configured to generate a first audio signal, an external microphone configured to generate a second audio signal, a speaker, and one or more processors coupled between the speaker, the external microphone, and the internal microphone, the method comprising: when the headset is worn by a user: determining a first audio transfer function between the first audio signal and the second audio signal in the presence of sound at the external microphone; and determining a second audio transfer function between the loudspeaker input signal and the first audio signal if the loudspeaker is driven by the loudspeaker input signal; determining an electrical transfer function of the one or more processors; determining a closed-ear transfer function based on the first audio transfer function, the second audio transfer function, and the electrical transfer function; and equalizing the first audio signal based on a comparison between the closed-ear transfer function and the open-ear transfer function to generate an equalized first audio signal.

The comparison may be a frequency domain ratio between the closed-ear transfer function and the open-ear transfer function. The comparison may be a time domain difference between the closed-ear transfer function and the open-ear transfer function.

The open-ear transfer function may be a measured open-ear transfer function between the ear entrances or tympanic membranes of the user. Alternatively, the open-ear transfer function may be a measured open-ear transfer function between the ear entrance of the head simulator and the auricular periosteum. Alternatively, the open-ear transfer function may be an average open-ear transfer function of a portion of the general population.

The method may further include a) measuring an open-ear transfer function between an entrance to the ear or the tympanic membrane of the user; or b) measuring an open-ear transfer function between an ear entrance of the head simulator and the auricular periosteum; or c) determining the open-ear transfer function based on an average open-ear transfer function of a portion of the general population.

The step of determining the first audio transfer function may be performed in case the loudspeaker is muted.

The step of determining the second audio transfer function may be performed in the presence of little or no sound external to the headset.

Determining the electrical path transfer function may include determining a frequency response of a feed-forward ANC filter implemented by the one or more processors and/or a frequency response of a feedback ANC filter implemented by the one or more processors.

Determining the frequency response may include determining a gain associated with the one or more processors.

The method may further include determining that an open-ear transfer function between the ear entrance of the user and the tympanic membrane comprises an open-ear transfer function proximate to the user.

The method may further include outputting the equalized first audio signal to the speaker.

The method may further comprise: determining a third audio transfer function between the first audio signal and the second audio signal when the headset is worn by the user and the user is speaking; and further equalizing the first audio signal based on the third transfer function.

The method may further include, upon determining that the user is speaking, outputting the voice equalized first audio signal to a speaker.

The method may further include determining that the one or more processors are implementing Active Noise Cancellation (ANC); and adjusting the further equalization to account for the one or more processors implementing ANC.

The method may also include requesting the user to speak a phoneme balanced sentence or phrase. The third audio transfer function may be determined when the user speaks the phoneme balance sentence.

According to another aspect of the present disclosure, there is provided an apparatus comprising: a headset, comprising: an internal microphone configured to generate a first audio signal; an external microphone configured to generate a second audio signal; a speaker; and one or more processors configured to: when the headset is worn by a user: determining a first audio transfer function between the first audio signal and the second audio signal in the presence of sound at the external microphone; and determining a second audio transfer function between the loudspeaker input signal and the first audio signal if the loudspeaker is driven by the loudspeaker input signal; determining an electrical transfer function of the one or more processors; determining a closed-ear transfer function based on the first audio transfer function, the second audio transfer function, and the electrical transfer function; and equalizing the first audio signal based on a comparison between the closed-ear transfer function and the open-ear transfer function to generate an equalized first audio signal.

The comparison may be a frequency domain ratio between the closed-ear transfer function and the open-ear transfer function. The comparison may be a time domain difference between the closed-ear transfer function and the open-ear transfer function.

The open-ear transfer function may be a measured open-ear transfer function between the ear entrances or tympanic membranes of the user. Alternatively, the open-ear transfer function may be a measured open-ear transfer function between the ear entrance of the head simulator and the auricular periosteum. Alternatively, the open-ear transfer function may be an average open-ear transfer function of a portion of the general population.

The one or more processors may be further configured to: a) measuring an open-ear transfer function between an ear entrance or tympanic membrane of a user; or b) measuring an open-ear transfer function between an ear entrance of the head simulator and the auricular periosteum; or c) determining the open-ear transfer function based on an average open-ear transfer function of a portion of the general population.

The step of determining the first audio transfer function may be performed in case the loudspeaker is muted.

The step of determining the second audio transfer function may be performed in the presence of little or no sound external to the headset.

Determining the electrical path transfer function may include determining a frequency response of a feed-forward ANC filter implemented by the one or more processors and/or a frequency response of a feedback ANC filter implemented by the one or more processors.

Determining the electrical path transfer function may include determining a gain associated with the one or more processors.

Determining an open-ear transfer function between the ear entrance of the user and the tympanic membrane includes approximating the open-ear transfer function.

The one or more processors may be further configured to output the equalized first audio signal to the speaker when it is determined that the user is not speaking.

The one or more processors may be further configured to determine a third audio transfer function between the first audio signal and the second audio signal when the headset is worn by the user and the user is speaking; and further equalizing the equalized first audio signal based on a difference between the open-ear transfer function and the closed-ear transfer function to generate a voice-equalized first audio signal.

The one or more processors may be further configured to output the voice equalized first audio signal to the speaker when it is determined that the user is speaking.

The one or more processors may also be configured to determine that the one or more processors are implementing Active Noise Cancellation (ANC); and adjusting the further equalization to account for the one or more processors implementing ANC.

The one or more processors may be further configured to output a request to the user to speak a phoneme balanced sentence or phrase, wherein the third audio transfer function is determined when the user speaks the phoneme balanced sentence.

According to another aspect of the present disclosure, there is provided a method of equalizing sound in a headset, the headset including an internal microphone configured to generate a first audio signal, an external microphone configured to generate a second audio signal, a speaker, and one or more processors coupled between the speaker, the external microphone, and the internal microphone, the method comprising: determining a first audio transfer function between the first audio signal and the second audio signal while the headset is worn by the user and the user is speaking; and equalizing the first audio signal based on the first audio transfer function.

The method may further include, upon determining that the user is speaking, outputting the voice equalized first audio signal to a speaker.

The method may further include determining that the one or more processors are implementing Active Noise Cancellation (ANC); and adjusting the equalization to account for the ANC.

The method may also include requesting the user to speak a phoneme balanced sentence or phrase. The first audio transfer function may then be determined when the user speaks the phoneme balance sentence.

According to another aspect of the present disclosure, there is provided an apparatus comprising: a headset, the headset comprising: an internal microphone configured to generate a first audio signal; an external microphone configured to generate a second audio signal; a speaker; and one or more processors configured to: determining a first audio transfer function between a first audio signal and a second audio signal when the headset is worn by the user and the user is speaking; and equalizing the first audio signal based on a difference between the open-ear transfer function and the closed-ear transfer function to generate an equalized first audio signal.

The one or more processors may be further configured to: the equalized first audio signal is output to the speaker when it is determined that the user is speaking.

The one or more processors may be further configured to: determining that the one or more processors are implementing Active Noise Cancellation (ANC); and adjusting the equalization to account for the ANC.

The one or more processors may be further configured to: requesting the user to speak a phoneme balance sentence or phrase, wherein the first audio transfer function is determined when the user speaks the phoneme balance sentence.

The headset may include one or more of the one or more processors.

According to another aspect of the present disclosure, there is provided an electronic device including the above apparatus.

Throughout this specification the word "comprise", or variations such as "comprises" or "comprising", will be understood to imply the inclusion of a stated element, integer or step, or group of elements, integers or steps, but not the exclusion of any other element, integer or step, or group of elements, integers or steps.

Drawings

Embodiments of the present disclosure will now be described, by way of non-limiting example only, with reference to the accompanying drawings, in which:

FIG. 1 is a schematic view of the sound and bone conduction paths around and through a user's head;

FIG. 2 is a schematic illustration of the sound and bone conduction paths around and through the head of the user shown in FIG. 1 wearing the headset;

fig. 3 is a schematic diagram of a headset according to an embodiment of the present disclosure;

fig. 4a is a schematic view of a module of the headset shown in fig. 3;

FIG. 4b is a block diagram illustrating the electrical conduction paths present in the module shown in FIG. 4 a;

FIG. 5 is a flow diagram illustrating a process for determining and applying EQs in the module of FIG. 4a to restore high frequency attenuation at the eardrum of a user;

FIG. 6 is a schematic view of a sound conduction path between the ear entrance and the tympanic membrane of the user shown in FIG. 1;

fig. 7 is a schematic diagram of a sound conduction path and an electrical conduction path between the ear entrance and the eardrum of the user shown in fig. 2 wearing the headset of fig. 3;

FIG. 8 is a flow chart illustrating a process for determining a transfer function for the acoustic conduction path shown in FIG. 6;

FIG. 9 is a flow chart illustrating a process for determining a transfer function of the electrically conductive path shown in FIG. 7;

FIG. 10a illustrates the estimated open-ear transfer function of the user shown in FIG. 1;

FIG. 10b illustrates the measured transfer function between the output of the error microphone and the output of the reference microphone of the module shown in FIG. 4 a;

FIG. 10c illustrates a measured transfer function between the input of the speaker of FIG. 4a and the output of the error microphone;

FIG. 10d illustrates an exemplary default gain for the module shown in FIG. 4 a;

FIG. 10e illustrates an example of an EQ applied in the module shown in FIG. 4a for recovering HF attenuation;

FIG. 11a illustrates an estimated leakage path transfer function from the input of the speaker to the output of the reference microphone for the module shown in FIG. 4 a;

FIG. 11b illustrates the open loop transfer function of the feedback oscillating system for the module shown in FIG. 4 a;

FIG. 12 is a flow diagram illustrating a process for determining and applying EQs in the module of FIG. 4a to attenuate low frequency boosting due to occlusion effects at the user's eardrum;

FIG. 13 is a schematic illustration of the acoustical and skeletal conduction paths between the ear inlet and the tympanic membrane of the user shown in FIG. 1 when the user is speaking;

fig. 14 is a schematic diagram of a sound conduction path, a bone conduction path, and an electrical conduction path between the ear entrance and the eardrum of the user shown in fig. 2 wearing the headset of fig. 3;

FIG. 15 is a graph comparing a theoretically derived raw EQ and an approximated EQ for attenuating low frequency advance due to occlusion effects, in accordance with an embodiment of the present disclosure; and

FIG. 16 is a flow diagram of a process for dynamically adjusting EQs applied in the modules shown in FIG. 4a based on the voice activity of the user shown in FIG. 2.

Detailed Description

Figures 1 and 2 comparatively illustrate the effect of an ear closure on the user's own voice. Fig. 1 shows a situation where the user 100 is not wearing headphones. There is an acoustic conduction path through air between the mouth and the ear of the user 100 and a bone conduction path between the mouth and the ear inside the head of the user 100. The line on the graph in fig. 1 represents a typical open-ear frequency response of a user 100 from the ear inlet to the tympanic membrane. Fig. 2 shows the gain between the closed-ear frequency response and the open-ear frequency response of a user 100 wearing a headset 102 and speaking.

Isolating the user's 100 eardrums from the external environment has two side effects when the user wants to listen to their own-sound (OV). One of the side effects is Passive Loss (PL) at high frequencies, which results in relatively attenuated high frequency sound at the user's eardrum, as shown in the graph in fig. 2. This attenuation makes the user inaudible to his voice. Another effect of having the ears occluded is the amplification of the user 100's own voice at a low frequency, which makes their sound buzzing. This enlargement is also shown in the graph of fig. 2. Amplification of the user's own voice at low frequencies is commonly referred to as an Occlusion Effect (OE).

Embodiments of the present disclosure relate to methods for a) restoring attenuated high frequency sound, and b) attenuating low frequency components introduced due to occlusion effects, with the aim of restoring the voice of the user 100 such that when wearing a headset, his voice sounds substantially as if he did not wear a headset.

The inventors have also recognized that high frequency attenuation due to passive losses occurs whether or not the user of the headset 200 is speaking, while the low frequency boom occurs only when the user is speaking. Thus, in embodiments of the present disclosure, a method of changing equalization in response to detecting that a user is speaking is presented.

In view of the above, equalization to recover attenuated high frequency sounds may be referred to herein as hearing enhancement equalization (HAEQ). Equalization to recover low frequency components of sound introduced due to occlusion effects may be referred to herein as incremental hearing enhancement equalization (dHAEQ).

Fig. 3 illustrates a headset 200 in which a HAEQ and/or dHAEQ may be implemented. It will be understood that the methods described herein may be implemented on any headset that includes two microphones, one of which is located external to the headset (e.g., a reference microphone) and one of which is located such that when the headset is worn by a user, the microphone is located close to the ear entrance (e.g., an error microphone). The microphone positioned proximate to the ear entrance may be associated with a speaker such that a feedback path exists between the microphone and the speaker.

The headset 200 shown in fig. 3 includes two modules 202 and 204. The modules 202, 204 may be wirelessly connected or otherwise connected. Each module 202, 204 comprises an error microphone 205, 206, a reference microphone 208, 210 and a loudspeaker 209, 211, respectively. The reference microphones 208, 210 may be positioned to pick up ambient noise from outside the ear canal and outside the earpiece. The error microphones 205, 206 may be positioned towards the ear in use in order to sense acoustic sound within the ear canal containing the output of the respective speakers 209, 211. The speakers 209, 211 are provided primarily for delivering sound to the ear canal of the user. The headset 200 may be configured to allow a user to listen to music or audio, make a telephone call, and/or deliver voice commands to a voice recognition system, among other such audio processing functions. The headset 200 may be configured to be worn on the ear, in which case the modules 202, 204 may be configured to be worn on the ear. Likewise, the modules 202, 204 may be configured to be worn in the ear canal.

Fig. 4a is a system diagram of the first module 202 of the headset. The second module 204 may be configured in substantially the same manner as the first module 202 and therefore is not separately shown or described. In other embodiments, the headset 200 may include only the first module 202.

The first module 202 may include a Digital Signal Processor (DSP)212 configured to receive microphone signals from the error microphone 205 and the reference microphone 208. The module 202 may further include a memory 214, which may be provided as a single component or as multiple components. A memory 214 may be provided for storing data and program instructions. The module 202 may also include a transceiver 216 to enable the module 202 to communicate wirelessly with external devices such as the second module 204, smart phones, computers, and the like. In alternative embodiments, such communication between the modules 202, 204 may include wired communication, where a suitable wire is provided between the left and right sides of the headset, either directly, such as within an overhead harness, or via an intermediate device, such as a smartphone. The module 202 may also include a Voice Activity Detector (VAD)218 configured to detect when a user is speaking. The module 202 may be battery powered and may include other sensors (not shown).

Fig. 4b is a block diagram illustrating exemplary electrical conduction paths of the first module 202 between the error microphone 205, the reference microphone 208 and the speaker 209. The conductive paths of the first module 202 shown in fig. 4b will be described in more detail below. Briefly, however, the first module 202 may implement Active Noise Cancellation (ANC) using feedback and feedforward filters, respectively denoted as H in fig. 4bFB(f) And HW2(f) In that respect In addition, the first module 202 may implement a hearing enhancement filter (or equalization block), HHA(f) Configured to recover sound components in the headset 200 of the user 100 that are lost due to high frequency passive loss attenuation and/or low frequency rods. H according to various embodiments of the present disclosure will now be describedHA(f) Determination and application.

FIG. 5 is a flowchart for determining HHA(f) A flow chart of a process 500 of restoring high frequency sound attenuated by passive losses in the headset 200 of fig. 3.

At step 502, an open-ear transfer function (i.e., open-ear Transfer Function (TFOE)) may be determined. The open-ear transfer function may be measured on the user, for example, by an audiologist using microphones located at the ear entrance and tympanic membrane. Alternatively, the open-ear transfer function may be estimated based on the average open-ear transfer function of the general population. Alternatively, the open-ear transfer function of the user may be estimated based on a transfer function measured on a head simulator such as KEMAR (knowledge electronic human body model for acoustic studies). Various methods of determining the open-ear transfer function are known in the art and will therefore not be explained further herein. Where the open-ear transfer function is estimated based on demographic data or the like, the step 502 of determining the open-ear transfer function may be omitted or may simply comprise reading the stored open-ear transfer function from memory.

At step 504, the closed-ear transfer function of the user is determined. The closed-ear transfer function may represent the air conduction and electrical conduction paths that exist for the user 100 wearing the headset 200.

In step 506, a hearing enhancement eq (haeq) may be determined based on a comparison between the open-ear transfer function and the determined closed-ear transfer function of the user 100 wearing the headset 200. For example, the HAEQ may be determined based on a ratio (in the frequency domain) between the open-ear and closed-ear transfer functions or based on a dB spectral difference between the open-ear and closed-ear transfer functions. The EQ represents the difference in sound reaching the eardrum of the user 100 when the user wears the earphone 200 and when the user does not wear the earphone 200 (i.e., an open-ear state).

After the HAEQ is determined at step 506, the HAEQ may be applied to the input signal of the speaker 209 at step 508 in order to recover high frequency sound attenuated by passive losses in the headset 200.

Determining an open-ear transfer function

Determination of an open-ear transfer function according to an exemplary embodiment of the present disclosure will now be described with reference to fig. 6, which fig. 6 shows an open-ear system 600. It is assumed below that the user 100 is not speaking and therefore the osteoinductive path does not contribute to the sound incident at the eardrum.

Referring to fig. 6, a sound signal received at the eardrum may be defined as:

ZED_O(f)=ZEE(f)·HO(f) (1.1)

wherein:

ZED_O(f) opening the sound signal at the drum membrane in the ear;

ZEE(f) in the earSound signals at the mouth (whether open or closed); and

HO(f) an open-ear transfer function from an ear entrance in the open ear to the tympanic membrane.

As mentioned above, in some embodiments, Z may be recorded using a pair of measurement microphones, a first measurement microphone 602 and a second measurement microphone 604ED_O(f) And ZEE(f) In that respect The first measurement microphone 602 may be placed at the entrance of the ear and the second measurement microphone 604 may be placed at the eardrum of the user 100. Preferably, the first microphone 602 and the second microphone 604 are matched, i.e., they have the same characteristics (including frequency response and sensitivity). As described above, this process may be performed specifically on the user, or alternatively, data from the general population relating to the open-ear transfer function may be used to approximate the open-ear transfer function of the user 100.

The recorded electrical signals from the first microphone 602 and the second microphone 604 may be defined as:

XED_O(f)=ZED_O(f)·GMM1(f) (1.2)

XEE(f)=ZEE(f)·GMM2(f) (1.3)

wherein G isMM1(f) And GMM2(f) The frequency responses of the first measurement microphone 602 and the second measurement microphone 604, respectively. For typical measuring microphones, their frequency response is flat and equal to a fixed factor q for frequencies between 10Hz and 20kHzMM(conversion factor from physical sound signal to electrical digital signal). XED_O(f) Is the electrical signal of the first measurement microphone 602 at the tympanic membrane in the open ear. This can be approximated by using the KEMAR's ear with a KEMAR's eardrum microphone. When measuring the open-ear transfer function of a particular user 100, the first measurement microphone 602 may be a probe tube microphone that may be inserted into the ear canal until it contacts the tympanic membrane of the user 100. XEE(f) Is the electrical signal of the second measurement microphone 604 at the entrance of the ear.

Assuming that the first measurement microphone 602 and the second measurement microphone 604 match:

thus, HO(f) Can be composed of XED_O(f) And XEE(f) The estimation is:

whereinIs an estimated open-ear transfer function from an ear entrance in the open ear to the tympanic membrane.

Determining closed-ear transfer function

Referring again to fig. 5, an exemplary method for determining the closed-ear transfer function at step 504 of process 500 will now be described in more detail with reference to fig. 7, which illustrates the closed-ear system 700 when the user 100 is not making any sounds. As mentioned above, the determination of the closed loop transfer function is described herein with respect to a single module 202 of the headset 200. It should be appreciated that if other modules 204 are provided, similar techniques may be employed to determine the closed-loop transfer functions of the other modules 204.

In the closed-ear configuration, i.e. when the user 100 wears the headset, there are air conduction paths (as is the case in the open-ear situation of fig. 6) and electrical conduction paths between the error microphone 205, the reference microphone 208 and the speaker 209 of the module 202. As shown in H in FIG. 7S2(f) An additional air conduction path is shown between the speaker 209 and the error microphone 205.

Note that the electrical configuration of the module 202 shown in fig. 7 is provided as an example only, and different electrical configurations known in the art fall within the scope of the present disclosure.

The sound signal at the eardrum in the closed ear situation can be defined as:

ZED_C(f)=ZEM(f)·HC2(f) (1.6)

wherein:

ZEM(f) sound signal at the ear-closed error microphone 205 position; and

HC2(f) the transfer function of the sound signal from the position of the error microphone 205 to the eardrum of the closed ear. When the error microphone 205 is close to the tympanic membrane, H is presentC2(f)≈1。

Sound signal Z at error microphone 205EM(f) Can be defined as:

wherein:

the component of the sound signal contributed by the air conduction path at the location of the ear closed error microphone 205;

the component of the sound signal at the location of the closed-ear error microphone 205 contributed by the conductive path (taking into account the acoustic coupling between the speaker 209 and the error microphone 205).

Embodiments of the present disclosure are directed to estimating the components of an acoustic signal present due to air conduction by first estimating the componentsSecond estimates the contribution present at the error microphone 205 due to the electrical characteristics of the module 202 (i.e., the processed electrical signal output to the speaker 209)To estimate the sound signal Z present at the error microphone 205EM(f) In that respect The inventors have realized that not only the air conducting component is dependent on the fit of the headset 200 on the user 100, but also the electrically conducting path componentDepending on the fit of the earpiece 200 on the user 100 and the geometry of the ear canal of the user 100.

Determining

The acoustic transfer function from the ear entrance to the tympanic membrane in the closed-ear state (with the user 100 wearing the headset 200) may be defined as:

HC(f)=HP(f)·HC2(f) (1.8)

where is the transfer function of the sound signal from the entrance of the ear to the error microphone 205, which corresponds to the passive loss of sound caused by the earphone 200, and is the transfer function between the error microphone 205 and the tympanic membrane.

H may be made by assuming that error microphone 205 is very close to the eardrumC2(f) 1 and therefore HC(f)≈HP(f) Close to the eardrum simplifies the above equation (1.8).

In view of the above and assuming that the reference microphone 208 is located substantially at the entrance of the ear, the acoustic path transfer function H may be estimated by comparing the sound signal received at the reference microphone 208 with the sound signal received at the live error microphone 205 when the user 100 wears the headset 200C(f) In that respect Referring to fig. 8, at step 802, the earpiece is muted to ensure that the electrical conduction path does not contribute to the sound signal reaching the error microphone 205. In the presence of sound external to the headset 200, the electrical signal produced by the error microphone 205 may be captured at step 804. Sound signal at error microphoneCan be defined as:

captured by error microphone 205Electrical signal ofCan be defined as:

wherein G isEM(f) Is the frequency response of the error microphone 205, which is typically flat and equal to a fixed factor q for MEMS microphones at frequencies between 100Hz and 8kHzEM(conversion factor from physical sound signal to electrical digital signal).

At step 806, the electrical signal X generated by the reference microphone 208 may be capturedRM(f) In that respect Ear entrance sound signal ZEE(f) May be recorded by the reference microphone 208 as:

XRM(f)=ZEE(f)·GRM(f) (1.11)

wherein G isRM(f) Is the frequency response of the reference microphone 208, which is typically flat and equal to a fixed factor q for MEMS microphones at frequencies between 100Hz and 8kHzEM(conversion factor from physical sound signal to electrical digital signal).

Assuming that the frequency responses of the reference microphone 208 and the error microphone 205 match, then:

thus, at step 808, the error microphone 205 and the reference microphone 208 may be based on the captured electrical signals XEM(f)、XRM(f) To determine a user-specific acoustic transfer function H from the entrance of the ear to the eardrum in the closed earC(f) As defined below.

Determining

The inventors have realized that knowing the electrical characteristics of the processing between the reference microphone 208, the error microphone 205 and the speaker 209, the transfer function between the eardrum and the ear entrance due to the electrical conduction path can be determined by comparing the sound output at the speaker 209 and the same sound received at the error microphone 205.

FIG. 9 is a block diagram for determining the component of the sound signal at the location of the error microphone 205 in the near ear contributed by the conductive path (taking into account the acoustic coupling between the speaker 209 and the error microphone 205)Is shown in the flowchart of process 900.

At step 902, a signal is output to the speaker 209, preferably muting any external sound such that there is no external sound contribution at the error microphone 205 due to the closed otoacoustic conduction path between the ear entrance and the tympanic membrane. Loudspeaker input signal XSI(f) Generated by processing electronics within the module 202.

In the case where the external sound is muted, the sound signal at the error microphone 205 is muted by the speaker 209The contribution of (d) may be defined as:

wherein HS2(f) Is a transfer function of the sound signal from the position at the output of the loudspeaker 209 to the position of the error microphone 205, and GSK(f) Is the frequency response of the speaker 209, and XSI(f) Is the loudspeaker input signal.

Accordingly, the electrical signal output from the error microphone 205 may be defined as:

wherein G isEM(f) Is the frequency response of the error microphone 205.

Can be based on speaker input XSI(f) The signal and the frequency response of the speaker 209 estimate the sound signal at the position of the head mounted speaker. The transfer function between the input signal at the speaker 209 and the error microphone 205 output signal can be defined as:

from the above equation, due to GSK(f) And GEM(f) Is fixed, so that for different ear canal geometries and different earpiece fits,will be reacted with HS2(f) Is in direct proportion.

Loudspeaker input signal XSI(f) The back-end processing definition implemented by module 202. Accordingly, at step 906, the electrical characteristics of the module 202 for generating the speaker input signal may be determined. In some embodiments where headset 200 is only noise isolated (i.e., no Active Noise Cancellation (ANC)), the speaker input signal may be substantially unaffected by the processing in module 202. However, in some embodiments, the headset 200 may implement active noise cancellation. In this case, the loudspeaker input signal X is due to an equalization of the loudspeaker input signalSI(f) Will be affected by the feed-forward and feedback filters and by the hearing enhancement. In this case, the loudspeaker input signal XSI(f) Can be defined as:

XSI(f)=XRM(f)HHA(f)-XRM(f)HW1(f)-XCE(f)HFB(f) (1.16)

wherein:

HHA(f) the method comprises the following steps Implementing HAEQ (and dHAEQ below) using a hearing enhancement filter as described herein;

HW1(f) the method comprises the following steps A feed-forward (FF) ANC digital filter;

HFB(f) the method comprises the following steps A Feedback (FB) ANC digital filter;

XPB(f) playback signals (music, internally generated noise, etc.); and

XCE(f) the corrected error signal is used as the input of the FBANC filter.

Thus, at step 908, a transfer function between the error microphone 205 signal, the reference microphone 208 signal, and the speaker input signal is determined based on the determined electrical characteristics of the module 200 and the speaker-to-error microphone 205 acoustic coupling.

Note that if the headset does not implement ANC, there will be no feedback or feedforward filtering, such that XSI(f)=XRM(f)HHA(f)。

Playback of X when HA is enabledPB(f) Will typically be muted so that the user can hear the sound restored to his eardrum from outside the headset. Assuming that playback is muted and equal to zero when the HA function is enabled, equation (1.17) becomes:

combining sound conduction paths with electrically conductive paths

The air conducting and electrically conducting components may be combined as follows:

therefore:

when ANC is perfect, equation (1.20) can be simplified to:

this means that the air conduction contribution of the external sound at the eardrum has been completely cancelled and only the electrical conduction contribution (at the speaker 209) remains.

When ANC is muted, equation (1.20) can be simplified to:

note that whenAndwith similar amplitudes but different phases, their summation will produce a comb filtering effect. To reduce the comb filter effect, it is preferably ensured that the delay between the electrically conductive path and the air conductive path is minimized.

Thus, the method described herein may be used to derive the EQ that takes into account the air conduction path between the headphone inlet and the eardrum (using the reference error microphone ratio, the electrical conduction path within headphone module 202, and the air conduction path between speaker 209 and error microphone 209). Since both air conduction paths depend on the earpiece fit and ear canal geometry, the present embodiment provides a technique for determining in situ the customized EQ for the user 100 of the earpiece 200.

Derivation of HAEQ

Referring to step 506 of process 500 shown in FIG. 5, to restore sound at the eardrum to an open-ear state in a closed-ear configuration, the objective is to derive HHA(f) (i.e., HAEQ) to close the eardrum Z in the earED_C(f) Equal to the sound signal Z in the open earED_O(f) In that respect Therefore, there are:

therefore:

assuming that the error microphone is close to the tympanic membrane, H existsC2(f) 1. Assuming that the reference microphone 205 and the error microphone 208 have similar characteristics,thus, equation (1.24) can be simplified to:

if the ANC is functioning well, then,equation (1.25) can be further simplified to:

thus, when ANC is operating effectively, the reference microphone 208 and the error microphone 205 are matched and the error microphone 205 is close to the eardrum of the user 100, HHA(f) Will only consist ofAndto decide.

Thus, the sound signal at the eardrum of the user is determinedNumber ZED_C(f) HAEQ reverts to the open ear state.

Note that the frequency response H applied at the speaker input can beHA(f) Further decomposed into a default fixed electrical frequency response HHAEE(f) And an adjustable frequency response (or equalizer) HHAEQ(f):

HHA(f)=HHAEE(f)·HHAEQ(f) (1.28)

Wherein HHAEE(f) Is derived from H when all filters (e.g. equalizer, noise cancellation, etc.) are disabledHA(f) Input to output of and HHAEQ(f) Is a balance for restoring an open ear condition at the eardrum of the user 100. Then, the user can use the device to perform the operation,

equation (1.29) above shows the measurement with the user 100 wearing the head-mounted device 200And HHAEE(f) (i.e., in situ measurement) and knowledge of the feedback filter H from the head-mounted device 200W1(f) And HFB(f) Can directly calculate H after the current value ofHAEQ(f)。

The inventors have also recognized that the effect of EQ is not substantially affected when phase is ignored. Thus, the above equation (1.29) can be simplified as follows.

Note that HHA(f) Preferably designed to restore/compensate but not eliminate the sound signal at the eardrum. Therefore | HHAEQ(f) | should preferably not be negative. In the case of the equation (1.30),is always greater than or equal to(whether ANC is on or off), therefore | HHAEQ(f) I is always positive.

Fig. 10a to 10 e. Fig. 10a illustrates the estimated open-ear transfer function of the user 100. Fig. 10b illustrates the measured transfer function between the output of the error microphone 205 and the output of the reference microphone 208 of the first module 202 according to the process 800 described above. Fig. 10c illustrates the transfer function measured between the input of the speaker 209 and the output of the error microphone 205 according to the process 900 described above. Fig. 10d illustrates the default transfer function or gain H of the headset 200HAEE(f)。

In addition to the transfer functions mentioned in equation (1.30), two further transfer functions may be considered. Leakage paths between the first considered error microphone 205 and the reference microphone 208The second approach may take into account the possibility of feedback jitter by estimating the module's open loop transfer function during feedback jitter.

When considering the above referenced paths:

therefore, the first and second electrodes are formed on the substrate,

wherein the content of the first and second substances,is an estimate of the leakage path when the external sound is muted, ANC is disabled, and the playback signal is output to the speaker 209.Being a feedback oscillating systemAn open loop transfer function; the transfer function should be less than 1 to avoid the generation of feedback jitter.

FIGS. 11a and 11b show estimated leakage path transfer functions, respectively, for a feedback oscillator systemAnd an open loop transfer function. It can be seen that the leakage in the exemplary system is small and the open loop transfer function of the feedback oscillator system is much less than 1. Therefore, the derived HAEQ should not cause feedback jitter. However, in systems where the open loop transfer function at some frequencies is close to 1, the HAEQ should be reduced at those frequencies to avoid feedback jitter.

Using HAEQ

Finally, referring back to fig. 5, at step 508 of process 500, the HAEQ may be applied to the speaker input signal to restore the open-ear sound to the user 100 of the headset 200.

Deriving dHAEQ for own speech

As described above, the effect of having the ears occluded with an earphone such as the earphone 200 described herein is to amplify the user's 100 own low frequency speech, which causes their speech to buzz for them. This amplification is due to the transmission of the user's voice through the bones and muscles of their head, the so-called skeletal conduction pathways. The determination of dHAEQ may be made in a similar manner as described above with reference to process 500 for determining HAEQ shown in fig. 5. However, in addition to the acoustic and electrical conduction paths, the bone conduction path must be considered.

An added complication in addressing low frequency amplification of self-speech due to bone conduction is that bone conduction varies with the phenomenon that the user 100 is speaking, as the resonance position in the mouth varies for different phenomena that are speaking. This means that the osteoinductive path is time-varying.

Fig. 12 is a flow diagram of a process 1200 for determining to attenuate the elevation of native speech at the eardrum of user 200 due to a native speech occlusion.

At step 1202, the open-ear transfer function of the user (i.e., the open-ear Transfer Function (TFOE) of the user) may be determined. The open-ear transfer function of the user may be measured, estimated, or otherwise determined in the same manner as described above with reference to fig. 5.

In step 1204, the closed-ear transfer function of the user is determined. The closed-ear transfer function may represent the air, bone and electrical conduction paths that exist for the user 100 wearing the headset 200 and speaking.

At step 1206, a hearing enhancement EQI may be determined based on a comparison between the open-ear transfer function and the determined closed-ear transfer function of the user 100 wearing the headset 200. For example, EQ may be determined based on the ratio between the open-ear and closed-ear transfer functions (in the frequency domain) or based on the dB spectral difference between the open-ear and closed-ear transfer functions. The EQ represents the difference in sound reaching the eardrum of the user 100 when the user is wearing the headphone 200, when the user is speaking and when the user is not wearing the headphone 200 (i.e., the open-ear state).

After dHAEQ is determined in step 1206, dHAEQ may be applied to the input signal to the speaker 209 in step 1208 in order to attenuate low frequency sounds that reach the eardrum due to self voice occlusion.

Determining an open-ear transfer function

Determination of an open-ear transfer function according to an exemplary embodiment of the present disclosure will now be described with reference to fig. 13, with fig. 6 showing an open-ear system 1300. It is assumed below that the user 100 is speaking, and therefore the osteoinductive path contributes to the sound incident at the eardrum.

Referring to fig. 13, for example, an open ear system 1300 may be characterized using three measurement microphones, referred to herein as a first measurement microphone 1302, a second measurement microphone 1304, and a third measurement microphone 1306. The first measurement microphone 1302 may be placed at the tympanic membrane in a manner similar to that described above. The second microphone 1304 may be placed at the entrance to the ear, and the third microphone 1306 may be placed at or near the user's mouth. The position of the third microphone 1306 is hereinafter referred to as the mouth point.

It may be assumed that the sound conduction (AC) path between the mouth and the entrance of the ear of the user is approximately time-invariant. Thus, the sound signal at the entrance to the ear can be defined as:

ZEE(f)=ZMP(f)HA(f) (2.1)

wherein ZEE(f) Is the sound signal at the entrance of the ear, ZMP(f) Is the sound signal of the self-voice at the point of mouth, and HA(f) Is the transfer function of the AC path between the point of mouth and the entrance to the ear when the user 100 speaks.

H may be estimated using a second measurement microphone 1304 and a third measurement microphone 1306 (one at the mouth point and the other at the entrance of the ear of the user 100)A(f) To give:

wherein XEE(f) And XMP(f) Respectively represents ZEE(f) And ZMP(f) Electrical output signals at the microphones 1304 and 1304.

AC and BC contributionsAndat the tympanic membrane may be defined as:

wherein:

an AC component contributing to the self-voice of the sound signal at the eardrum in the open ear;

HB_O(f, k): from mouth toThe BC channel of the eardrum is the transfer function of the own voice; k is the time-varying exponent of the transfer function; the transfer function typically changes according to the phenomenon spoken by the user 100.

In the open ear, the BC speech component contributes to the sound signal on the eardrum.

The transfer function of self-speech from the ear entrance to the tympanic membrane through the reverse of the AC path and then through the BC path in the open ear can be defined as:

therefore, equation (2.4) becomes:

the sum of the AC and BC contributions to the sound at the eardrum can then be defined as:

when Z isED_O(f, k) and ZEE(f) Recorded as X by the first measurement microphone 1302 and the second measurement microphone 1304ED_O(f, k) and XEE(f) And H isO(f) Having been estimated as equation (1.4) above, H can be calculatedAB_O(f, k) estimated as:

when the user 100 is speaking, the ratio between the sound signal at the eardrum and the sound signal at the entrance of the ear may be defined as:

we can also put the ratio R between the AC and BC contributions of the user's own voice at the eardrumZ_ED_O(f, k) is defined as:

RZ_ED_Othe different phonemes of (f, k) have been measured and estimated by previous researchers for the general population. At Reinfeldt, S,P.、B.、&The details of exemplary experimental measurements and estimates are described in "hearing person' S own voice during phoneme utterance-transmission through air and bone conduction" by Stenfelt, S. (2010). Journal of the acoustical society of america, 128(2), 751-762, the contents of which are incorporated herein by reference in their entirety.

Determining self-voice closed-ear transfer function

Referring again to fig. 12, an exemplary method for determining the closed-ear transfer function at step 1204 of process 1200 will now be described. As described above, the determination of the native voice closed loop transfer function is described herein with respect to a single module 202 of the headset 200. It should be appreciated that if other modules 204 are provided, similar techniques may be employed to determine the closed-loop transfer functions of the other modules 204. As noted above, it should also be noted that the electrical configuration of the module 202 shown in fig. 14 is provided as an example only, and that different electrical configurations known in the art fall within the scope of the present disclosure.

As shown in H in FIG. 14S2(f) An additional air conduction path is shown between the speaker 209 and the error microphone 205.

In its voice closed-ear configuration, i.e., when the user 100 wears the headset 200 and speaks, there is an electrically conductive path between the error microphone 205, the reference microphone 208, and the speaker 209 of the module 202, in addition to the air-conductive and bone-conductive paths that also exist in the open-ear situation of fig. 13.

The analysis of the AC and EC path contributions to self-speech are the same as those described above with reference to fig. 5 to 7. The additional Bone Conduction (BC) component for the native voice may be added to the AC component provided by equation (1.21) to provide updated equation (1.21) for interpreting the native voice:

wherein HAB_C1(f, k) is the transfer function of the self-voice from the entrance of the ear through the inverse of the AC path (i.e., entrance of the ear to the mouth point) and then through the closed-ear BC path to the position of the error microphone 205; k is a time-varying index of the transfer function that can change with different phonemes uttered by the user, different manifestations leading to different sounds and mouth shapes.

HAB_C1(f, k) may be defined as:

wherein HB_C1(f, k) is the transfer function of the BC path from the mouth to the position of the error microphone 205 for the own voice; k is a time-varying index of the transfer function that may vary with different phonemes uttered by the user; at frequencies less than about 1kHz, H is due to occlusion effectsB_C1(f, k) is usually in ratio to HB_O(f, k) are much larger.

When the output of the speaker 209 is muted, equation (2.11) becomes:

thus, H can be substitutedAB_C1(f, k) estimated as:

assuming ANC in module 202 works well, equation (2.12) can be simplified to:

this means that both the AC and BC contributions of the user 100 own voice are completely eliminated at the eardrum, and only the EC contribution is left.

When ANC is muted, equation (2.12) can be simplified to:

for frequencies below 1kHz, H due to occlusion effectsAB_C1(f, k) ratio of equation (2.16)Andmuch larger.

Derivation of dHAEQ for self-speech

Referring to step 1206 of process 1200 shown in FIG. 12, to restore the sound at the eardrum to an open-ear state in a closed-ear configuration, the objective is to derive HHA(f) So as to make the eardrum Z in the closed earED_C(f) The sound signal of (A) is equal to Z in the open earED_O(f)。

There are:

therefore:

assuming that the error microphone 205 is located close to the tympanic membrane, HC2(f) 1, assuming that the error microphone 205 and the reference microphone 208 substantially match,

therefore, equation (2.18) can be simplified to:

h, as previously discussed with reference to equation (1.25)HA(f) It is always positive for external sounds (i.e. external sounds that are not the voice from the user). However, in some cases, the H for the own voice calculated by equation (2.19)HA(f) May be negative. This is because HAB_C1(f, k) may be greater than HAB_O(f, k) is 30dB greater. Even when ANC is turned on in the earphone 100, HAB_C1Attenuation on (f, k)Typically less than 30 dB.

Equation (2.19) may be further rewritten to yield one term identical to equation (1.25) above and another term defined as:

wherein H is used for external soundHAforOS(f):HHA(f) As described in equation (1.25).

The product term in equation (2.20) can be defined as:

from equation (2.21) we can see that H is when there is no own voicedHAEQ(f, k) is changed to 1, HHA(f, k) is to become HHAforOS(f) In that respect Thus, HdHAEQ(f, k) represents the additional equalization needed to account for the user's own voice low frequency boost at the eardrum. H since the occlusion effect occurs mainly at low frequenciesdHAEQ(f, k) may be applied only at frequencies below the low frequency threshold. In some embodiments, HdHAEQ(f, k) may be applied at a frequency of less than 2000Hz, or less than 1500Hz, or less than 1000Hz or less than 500 Hz.

When ANC works well, equation (2.21) can be simplified to:

RX_ED_O(f, k) (as defined in equation (2.9)) is the ratio between the output of the error microphone 205 (i.e., the microphone recording at the eardrum) and the output of the reference microphone (i.e., approximately at the entrance of the self-speech in the open ear).

When ANC is performed enough to eliminate the AC path instead of the BC path (which is the most likely case), equation (2.21) can be simplified to:

when ANC and HA are open and HHA(f, k) is set to HHAforOS(f, k), there are:

we can define:

therefore, equation (2.23) can be rewritten as:

HdHAEQ(f,k)≈RX_ED_O(f,k)-RX_EM_ANConHAon(f,k)+1 (2.26)

note that in equation (2.26) RX_ED_O(f, k) and RX_EM_ANConHAon(f, k) is always greater than 1. Furthermore, for different phonemes, RX_ED_O(f, k) and RX_EM_ANConHAon(f, k) are all time-varying. Since R is in the case that the user 100 wears the headset 200X_ED_O(f, k) need to be recorded in the open ear, and RX_EM_ANConHAon(f, k) needs to be recorded in the closed ear, so it is difficult to record both in situ at the same time. Thus, in some embodiments, to approximate RX_ED_O(f, k) and RX_EM_ANConHAon(f, k), during calibration, the user 100 may be required to read sentences, preferably phoneme-balanced sentences, in an open-ear and closed-ear configuration while wearing the headset 200 and with ANC and HA enabled. The ratio can then be determined on the phoneme balanced sentenceAndaverage value of (a).

Thus, HdHAEQ(f, k) may be fixed as:

also note that the HA block is designed to compensate for, but not cancel, the sound signal at the eardrum, soShould be limited to greater than zero, e.g., at least 0.01, as follows:

the inventors have also found that the following equation provides HdHAEQ(f, k) anda good approximation of:

in other words,may be approximated as the ratio between the electrical output of the reference microphone and the electrical output at the error microphone when ANC and HA are switched on.

FIG. 15 provides various R's calculated using equation (2.28)X_ED_OOf (f, k) valueAnd calculated using equation (2.30)Comparison of (1).

It can be seen that equation (2.30) approximates the provided equation (2.28) RX_ED_O(f, k) is known. The approximation of equation (2.30) means that it is not necessary to measure the open-ear function RX_ED_O(f, k); for approximations using equation (2.28)Only need to derive (A) fromFor the closed-ear function.

Application of dHAEQ

Finally, referring back to fig. 12, at step 1208 of process 1200, while the user is speaking, dHAEQ may be applied (in combination with HAEQ for restoring HF attenuation) to the speaker input signal to restore the open-ear sound to the user 100 of the headset 200.

As described above, whether H is useddHAEQ(f,k)、Or an approximation thereof, the equalization is only needed when the user is speaking. Thus, preferably, the headset 200 may be configured to determine when the user 100 is speaking such that the total EQ, i.e., H, applied by the HA blockHA(f) Or HHA(f, k), may be in HHAEQ(f) (i.e., EQ for recovering HF attenuation due to passive loss) and HHAEQ(f)+HdHAEQ(f) (i.e., a combination of EQ for restoring HF decay and EQ for removing LF ringing due to occlusion effects). To this end, Voice Activity Detector (VAD)218 may be configured to provide a determination (e.g., flag or probability) of voice activity to module 202 such that dHAEQ may be turned on and off.

Fig. 16 is a flow diagram of a process 1600 that may be implemented by the first module 202/headset 200 for controlling the HA block, HHA(f)。

At step 1602, a HAEQ may be determined as described above with reference to fig. 5.

At step 1604, dHAEQ may be determined as described above with reference to fig. 12.

At step 1606, DSP 212 may be configured to determine whether user 100 is speaking based on the output received from VAD 218.

If it is determined that the user 100 is not speaking, the process 1600 continues to step 1608, and the DSP 212 implements the HA block HHATo contain only HHAEQIn order to recover the attenuated high frequency sound lost due to passive losses in the closed ear state. The process then continues to step 1606, where the determination of whether the user 100 is speaking is repeated.

However, if it is determined that the user 100 is speaking, the process 1600 continues to step 1610, and the DSP 212 implements the HA block HHATo contain HHAEQAnd HdHAEQTo both recover the attenuated high frequency sound loss due to passive loss in the ear-closed state and suppress it when the user is speakingLow frequency boosting due to occlusion effects is suppressed.

Note that since the occlusion effect occurs only at low frequencies (e.g., below about 1kHz), dHAEQ is preferably applied only to frequencies where it is needed in order to minimize distortion in the signal output to speaker 209.

Note that although it may be preferable to consider high frequency attenuation and low frequency enhancement (due to bone conduction), embodiments of the present disclosure are not limited thereto. For example, in some embodiments, the headset 200 may be configured to implement an HA block to equalize high frequency attenuation rather than low frequency (occlusion effect) enhancement. Likewise, in some embodiments, the earpiece 200 may be configured to implement an HA block to equalize low frequency (occlusion effect) boosting rather than high frequency attenuation.

The embodiments described herein may be implemented in an electronic, portable, and/or battery-powered host device, such as a smartphone, audio player, mobile or cellular telephone, handset. Embodiments may be implemented on one or more integrated circuits provided within such host devices. Alternatively, embodiments may be implemented in a personal audio device, such as a smartphone, mobile or cellular telephone, headset, earphone, or the like, that is configurable to provide audio playback to a single person.

Also, embodiments may be implemented on one or more integrated circuits provided within such personal audio devices. In yet another alternative, embodiments may be implemented in a combination of a host device and a personal audio device. For example, embodiments may be implemented in one or more integrated circuits provided within a personal audio device and one or more integrated circuits provided within a host device.

It will be appreciated by those of ordinary skill in the art having the benefit of the present disclosure that the various operations described herein, and in particular those described in conjunction with the figures, may be implemented by other circuits or other hardware components. The order of performing each operation of a given method can be varied, and various elements of the systems shown herein can be added, reordered, combined, omitted, modified, etc. The disclosure is intended to cover all such modifications and alterations, and therefore, the above description should be regarded as illustrative rather than restrictive.

Similarly, although the present invention has been described with reference to particular embodiments, certain modifications and changes may be made to those embodiments without departing from the scope and coverage of the present invention. Furthermore, any benefits, advantages, or solutions to problems that are described herein with regard to specific embodiments are not intended to be construed as a critical, required, or essential feature or element.

Additional embodiments and implementations will be apparent to those of ordinary skill in the art having the benefit of this disclosure, and such embodiments should be considered included herein. Moreover, those of ordinary skill in the art will recognize that a variety of equivalent techniques may be applied in place of, or in combination with, the discussed embodiments, and all such equivalents should be considered encompassed by the present disclosure.

Those skilled in the art will recognize that some aspects of the apparatus and methods described above, such as the discovery and configuration methods, may be embodied as processor control code, for example, on a non-volatile carrier medium such as a disk, CD or DVD ROM, programmed memory such as read only memory (firmware), or on a data carrier such as an optical or electrical signal carrier. For many applications, embodiments of the present disclosure will be implemented on a DSP (digital signal processor), an ASIC (application specific integrated circuit), or an FPGA (field programmable gate array). Thus, the code may comprise conventional program code or microcode or, for example code for setting up or controlling an ASIC or FPGA. The code may also include code for dynamically configuring a reconfigurable device, such as a reprogrammable array of logic gates. Similarly, the code may include code for a hardware description language such as Verilog (TM) or VHDL (very high speed Integrated Circuit hardware description language). As will be appreciated by those skilled in the art, code may be distributed among multiple coupled components in communication with each other. Embodiments may also be implemented using code running on a field (re) programmable analog array or similar device, where appropriate, to configure analog hardware.

Note that as used herein, the term module will be used to refer to a functional unit or block that may be implemented at least in part by a dedicated hardware component, such as a custom circuit, and/or at least in part by one or more software processors or appropriate code running on an appropriate general purpose processor or the like. The modules themselves may include other modules or functional units. A module may be provided by a plurality of components or sub-modules which need not be co-located and may be provided on different integrated circuits and/or run on different processors.

It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design many alternative embodiments without departing from the scope of the appended claims or embodiments. The word "comprising" does not exclude the presence of elements or steps other than those listed in a claim or embodiment, "a" or "an" does not exclude a plurality, and a single feature or other element may fulfil the functions of several elements recited in the claims or embodiments. Any reference signs or signs in the claims or examples should not be construed as limiting their scope.

Although the present disclosure and certain representative advantages have been described in detail, it should be understood that various changes, substitutions and alterations can be made herein without departing from the spirit and scope of the disclosure as defined by the appended claims or examples. Moreover, the scope of the present disclosure is not intended to be limited to the particular embodiments of the process, machine, manufacture, composition of matter, means, methods and steps, presently existing or later to be developed, that perform substantially the same function or achieve substantially the same result as the corresponding embodiments described herein. Accordingly, the appended claims or embodiments are intended to include within their scope such processes, machines, manufacture, compositions of matter, means, methods, or steps.

38页详细技术资料下载
上一篇:一种医用注射器针头装配设备
下一篇:用于处理音频信号的设备

网友询问留言

已有0条留言

还没有人留言评论。精彩留言会获得点赞!

精彩留言,会给你点赞!

技术分类