Audio processing method and device, mobile terminal and storage medium

文档序号:1506823 发布日期:2020-02-07 浏览:4次 中文

阅读说明:本技术 音频处理方法、装置及移动终端及存储介质 (Audio processing method and device, mobile terminal and storage medium ) 是由 李�浩 陈翔宇 于 2018-07-27 设计创作,主要内容包括:本公开提供一种音频处理方法,所述方法包括以下步骤:获取采集模块采集的第一音频数据;对所述第一音频数据通过并发处理方式进行第一音效处理生成第二音频数据;将所述第二音频数据与用于伴奏的第三音频数据进行混音处理生成第一混音数据,所述第三音频数据为预先存储的音频数据;输出所述第一混音数据。本公开能够将音频的延时控制在一个很低的范围内,以便于在最终直播时可以将音频数据与高质量的伴奏进行混音,提高直播k歌的质量。(The present disclosure provides an audio processing method, the method comprising the steps of: acquiring first audio data acquired by an acquisition module; carrying out first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data; mixing the second audio data with third audio data used for accompanying to generate first mixed audio data, wherein the third audio data is pre-stored audio data; and outputting the first mixed voice data. This openly can be with the time delay control of audio frequency in a very low within range to can carry out the audio mixing with high-quality accompaniment with audio data when finally live, improve the quality of live k song.)

1. An audio processing method, comprising the steps of:

acquiring first audio data acquired by an acquisition module;

carrying out first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data;

performing sound mixing processing on the second audio data and third audio data to generate first sound mixing data, wherein the third audio data is pre-stored audio data;

and outputting the first mixed voice data.

2. The method of claim 1, wherein after acquiring the first audio data acquired by the acquisition module, further comprising:

and performing second sound effect processing on the first audio data to generate fourth audio data.

3. The method of claim 2, wherein the second sound effect processing the first audio data to generate fourth audio data comprises:

and processing the first audio data according to a preset sound effect algorithm.

4. The method of claim 2, wherein after the performing the second sound effect processing on the first audio data to generate fourth audio data, further comprises:

and performing sound mixing processing on the fourth audio data and the third audio data to generate second sound mixing data.

5. The method of claim 4, wherein after the mixing the fourth audio data with the third audio data to generate second mixed data, the method further comprises:

and pushing the second audio mixing data to a distribution server through a streaming media module so that the distribution server distributes the second audio mixing data to each audio receiving output side terminal.

6. The method of claim 4, wherein mixing the fourth audio data with the third audio data to generate second mixed data comprises:

time-proofreading the fourth audio data and the third audio data to time-synchronize the fourth audio data and the third audio data.

7. The method of claim 1, wherein the outputting the first mixing data comprises:

and sending the first mixed sound data to a peripheral sound device of the terminal for receiving the audio backtransmission party.

8. An audio processing apparatus, comprising:

an acquisition module configured to acquire first audio data;

the processing module is configured to perform first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data;

the audio mixing module is configured to perform audio mixing processing on the second audio data and third audio data to generate first audio mixing data, wherein the third audio data is pre-stored audio data;

an output module configured to output the first remix data.

9. A mobile terminal, comprising:

a processor;

a memory for storing processor-executable instructions;

wherein the processor is configured to: the steps of performing the audio processing method of any of claims 1 to 7.

10. A non-transitory computer readable storage medium, wherein instructions in the storage medium, when executed by a processor of a mobile terminal, enable the mobile terminal to perform an audio processing method, the method comprising the steps of the audio processing method of any one of claims 1-7.

Technical Field

The present disclosure relates to the field of information processing, and in particular, to an audio processing method and apparatus, a mobile terminal, and a storage medium.

Background

The rapid development of the internet gradually changes the living style of the contemporary people, the requirements of the contemporary people on the spiritual culture are higher and higher, and singing gradually becomes one of the favorite entertainment activities of people. Especially, various Karaoke software products are popularized, so that more and more people can sing or record the singing voice of the people at any time and any place. The Karaoke software product is characterized in that singing voice of a user is synthesized in accompaniment provided by software, and then the singing voice recording result with better quality is obtained through karaoke audio effect processing and editing.

At present, although karaoke scenes are well-developed in the technical implementation of record-following, real-time rendering sound effects in live broadcasting is lacked. In the prior art, an AVCaptureSession with an iPhone advanced package is generally used for acquiring video and audio data in parallel, the buffer for audio acquisition is large, the output frequency is low, and the time axis is not easy to align when background music is processed.

Therefore, in the prior art, the time delay is large and uncontrollable, the function of the karaoke is limited, the problems that vocal accompaniment cannot be aligned, the quality of accompaniment is poor and the like easily occur, and the quality of live karaoke is influenced.

Disclosure of Invention

In order to solve the problems in the related art, the present disclosure provides an audio processing method and apparatus, and a corresponding mobile terminal, which can control the delay of audio within a very low range, so as to mix audio data with high-quality accompaniment in the final live broadcast, thereby improving the quality of live broadcast of a karaoke song.

In order to achieve the purpose, the technical scheme adopted by the disclosure is as follows:

according to a first aspect of embodiments of the present disclosure, the present disclosure provides an audio processing method, including the steps of:

acquiring first audio data acquired by an acquisition module;

carrying out first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data;

mixing the second audio data with third audio data used for accompanying to generate first mixed audio data, wherein the third audio data is pre-stored audio data;

and outputting the first mixed voice data.

Specifically, the processing of the first audio data by a concurrent processing mode to generate second audio data includes:

and performing noise reduction, reverberation, equalization and sound change processing on the first audio data.

Specifically, after the first audio data acquired by the acquisition module is acquired, the method further includes:

and performing second sound effect processing on the first audio data to generate fourth audio data.

Specifically, the performing of the second sound effect processing on the first audio data to generate fourth audio data includes:

and processing the first audio data according to a preset sound effect algorithm.

Specifically, after performing the second sound effect processing on the first audio data to generate fourth audio data, the method further includes:

and performing sound mixing processing on the fourth audio data and the third audio data to generate second sound mixing data.

Specifically, after the mixing the fourth audio data and the third audio data to generate second mixed data, the method further includes:

and pushing the second audio mixing data to a distribution server through a streaming media module so that the distribution server distributes the second audio mixing data to each audio receiving side terminal.

Optionally, mixing the fourth audio data with the third audio data to generate second mixed data, including:

time-proofreading the fourth audio data and the third audio data to time-synchronize the fourth audio data and the third audio data.

Specifically, the outputting the first mixed sound data includes:

and sending the first mixed sound data to a peripheral sound device of the terminal for receiving the audio backtransmission party.

Specifically, the third audio data is third audio data.

Specifically, the acquisition module is configured to acquire and receive the first audio data input by the peripheral input device of the audio return terminal.

Specifically, the peripheral input device of the audio receiving and returning party terminal includes a microphone, an earphone microphone, and a main sound card.

According to a second aspect of embodiments of the present disclosure, there is provided an audio processing apparatus comprising:

an acquisition module configured to acquire first audio data;

the processing module is configured to perform first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data;

the audio mixing module is configured to perform audio mixing processing on the second audio data and third audio data to generate first audio mixing data, wherein the third audio data is pre-stored audio data;

an output module configured to output the first remix data.

Specifically, the processing module includes:

a first processing unit configured to perform noise reduction, reverberation, equalization, and voicing processing on the first audio data.

Specifically, still include:

and the second sound effect processing module is configured to perform second sound effect processing on the first audio data to generate fourth audio data.

Specifically, the second sound effect processing module includes:

and the second processing unit is configured to process the first audio data according to a preset sound effect algorithm.

Specifically, still include:

and the mixing unit is configured to perform mixing processing on the fourth audio data and the third audio data to generate second mixed data.

Specifically, still include:

and the sending unit is configured to push the second mixed voice data to a distribution server through a streaming media module, so that the distribution server distributes the second mixed voice data to each receiving audio output terminal.

Specifically, the mixing unit includes:

a collation unit configured to time-collate the fourth audio data with the third audio data to time-synchronize the fourth audio data with the third audio data.

Optionally, the output module includes:

an output unit configured to transmit the first mixed sound data to a peripheral sound device of a receiving audio backhauling side terminal.

Specifically, the acquisition module is configured to acquire and receive the first audio data input by the peripheral input device of the audio return terminal.

Specifically, the peripheral input device of the audio receiving and returning party terminal includes a microphone, an earphone microphone, and a main sound card.

According to a third aspect of the embodiments of the present disclosure, there is provided a mobile terminal including:

a processor;

a memory for storing processor-executable instructions;

wherein the processor is configured to: the steps of the audio processing method according to any of the first aspect are performed.

According to a fourth aspect of embodiments of the present disclosure, the present disclosure provides a non-transitory computer-readable storage medium, characterized in that instructions in the storage medium, when executed by a processor of a mobile terminal, enable the mobile terminal to perform an audio processing method, the method comprising the steps of the audio processing method of any one of the first aspects.

According to a fifth aspect of embodiments of the present disclosure, there is provided an application program, which when executed by a processor of a mobile terminal, enables the mobile terminal to execute an audio processing method, the method comprising the steps of the audio processing method according to any one of the claims.

The present disclosure has the following advantages:

the audio processing method comprises the steps of acquiring first audio data acquired by an acquisition module; carrying out first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data; performing sound mixing processing on the second audio data and the third audio data to generate first sound mixing data; and outputting the first mixed voice data. According to the method and the device, the delay control of the audio is controlled in a very low range through a processing mode of concurrently processing the audio data, so that the second audio data and high-quality accompaniment are mixed conveniently during final live broadcasting, the quality of live broadcasting Karaoke is improved, and the Karaoke effect is improved.

2, the first aspect of the disclosure improves the acquisition mechanism of the acquisition module, and can control the buffer acquired by the audio of the acquisition module in a very small range to control the time delay; in the second aspect, the first audio data and the second audio data are processed in real time, so that the data processing time is shortened, the data processing efficiency is improved, and the audio time delay is controlled finally; in a third aspect, the present disclosure controls latency by concurrently processing audio data. This is disclosed can be with audio frequency time delay control in a very low within range through above various measures, realize the alignment of people's voice and accompaniment and will make up the accompaniment sound mixing of quality when being convenient for follow-up mixing sound to promote K song effect.

3, this disclosure can also be right first audio data carry on after the audio effect is handled again with third audio data carry out the audio mixing in order to obtain second audio mixing data to when carrying out the audio mixing and handling, accomplish the alignment of adult's sound and accompaniment, in order to reach the audio frequency treatment effect of optimality.

To sum up, this disclosure obtains lower time delay and better K sings effect under the prerequisite of guaranteeing real-time, multi-functional and high quality audio experience, promotes user experience.

It is to be understood that the foregoing description is of the advantages of the present disclosure and that further description of the advantages will be apparent to those skilled in the art from the following description of the embodiments, and that many more advantages of the present disclosure will be apparent to those skilled in the art from the disclosure herein.

Additional aspects and advantages of the disclosure will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the disclosure.

Drawings

The foregoing and/or additional aspects and advantages of the present disclosure will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:

FIG. 1 is a flow diagram illustrating a method of audio processing according to an exemplary embodiment;

FIG. 2 is an architecture diagram illustrating an iOS audio system stack, according to an exemplary embodiment;

FIG. 3 is a flow diagram illustrating an audio processing device according to an exemplary embodiment;

FIG. 4 is a schematic diagram illustrating the structure of an audio processing device according to an exemplary embodiment;

fig. 5 is a block diagram illustrating a mobile terminal according to an example embodiment.

Detailed Description

Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.

Fig. 1 is a flowchart illustrating an audio processing method, which is used in a mobile terminal, as shown in fig. 1, according to an exemplary embodiment, and includes the following steps.

In step S11: and acquiring first audio data acquired by the acquisition module.

In the embodiment of the disclosure, the acquisition module is configured to acquire and receive the first audio data input by the peripheral input device of the audio return terminal. The peripheral input equipment for receiving the audio and returning the audio to the terminal comprises a microphone, an earphone and a main broadcasting sound card. The terminal for receiving the audio back-transmitting party can be a main broadcasting terminal.

The method utilizes the Audio Unit technology, integrates the advantages of the Audio Unit during recording on the premise of guaranteeing timeliness, performance and experience, and is applied to a live broadcast scene to achieve the optimal audio processing effect.

Specifically, the AudioUnit scheme used in the present disclosure is closer to the bottom layer of the iOS Audio system stack, and is used in cooperation with multiple components such as an AudioUnit Processing Graph Services management IO device (IO Unit), a mixer (Mix Unit), an effector (Effect Unit), and the like, to obtain a lower delay and a better Effect.

Referring to fig. 2, fig. 2 is an architecture diagram of an iOS audio system stack, shown in accordance with an exemplary embodiment. As shown in FIG. 2, the iOS audio system stack includes Low-Level, Mid-Level, and High-Level.

Wherein, (1) Low-Level is mainly used in the audio APP implementation on MAC and under the condition that maximum real-time performance is required, most of the audio APPs do not need to use the services of this layer. Moreover, a high-level API with high real-time performance is provided on iOS to meet your demand. E.g., OpenAL, has real-time audio processing capability in games that is invoked directly with I/O.

And the I/O Kit is responsible for interacting with the hardware driver.

The Audio HAL, which is an Audio hardware abstraction layer, separates API calls from the actual hardware and remains independent.

Core MIDI, which provides a software abstraction layer of work for MIDI streams and devices.

And the Host Time Services is responsible for accessing the hardware clock of the computer.

(2) The Mid-Level layer has complete functions, including audio data format conversion, audio file reading and writing, audio stream analysis, plug-in work support and the like.

The Audio Convert Services are responsible for the conversion of Audio data formats.

The Audio File Services are responsible for reading and writing Audio data.

Audio Unit Services and Audio Processing Graph Services are plug-ins that support digital signal Processing such as equalizers and mixers.

Audio File Screen Services are responsible for stream parsing.

Core Audio Clock Services are responsible for Audio Clock synchronization.

(3) High-Level is a group of High-Level applications combined from low-Level interfaces, and basically much work on audio development can be done at this Level.

Audio Queue Services provide recording, playing, pausing, looping, and synchronizing Audio that automatically processes compressed Audio formats with the necessary codecs.

The AVAudio Player is an Objective-C interface-based audio playing class specially provided for the IOS platform, and can support the playing of all audio supported by the iOS.

Extended Audio File Services are composed of Audio files and Audio converters, and provide read-write capability for compressed and uncompressed Audio files.

OpenAL is the implementation of the CoreAudio to the OpenAL standard, and can play a 3D sound mixing effect.

The AudioUnit is a bottom layer technology of an iOS audio system stack, and the iOS provides audio processing plug-ins such as audio mixing, equalization, format conversion, real-time IO recording, playback, offline rendering, voice over internet protocol (VoIP), and the like, which all belong to different audiounits, and support dynamic loading and use. Audiounits can be created and used individually, but are more often used in combination in AudioProcessing Graph containers to meet diverse processing needs.

Based on the principle, the advantages of the Audio Unit during recording are integrated, and the Audio Unit is applied to a live scene to achieve the optimal audio processing effect.

In step S12: and carrying out first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data.

In the embodiment of the present disclosure, it is right that the first audio data is subjected to a concurrent processing mode through a first audio processing module to perform a first audio processing, the audio processing specifically includes performing noise reduction, reverberation, equalization, and sound variation processing on the first audio data, and the first audio processing module has a higher requirement on time delay.

It should be noted that, in the present disclosure, the first audio data is preferably processed in a concurrent processing manner, so as to shorten the data processing time, improve the data processing efficiency, and finally realize controlling the ear return delay.

In addition, the audio time delay is controlled by the following measures:

in a first aspect, the present disclosure improves the acquisition mechanism of the acquisition module, so that the buffer of audio acquisition of the acquisition module can be controlled in a very small range to control the time delay. This is disclosed solves prior art through this mechanism, and the buffer memory of audio acquisition is great, and the output frequency is lower, is difficult to the problem of alignment time axis when handling background music.

In a second aspect, the present disclosure processes the first audio data and the second audio data in real time to shorten data processing time, improve data processing efficiency, and finally control audio time delay.

Further, according to the present disclosure, after the first audio data is acquired, second sound effect processing is further performed on the first audio data to generate fourth audio data. Specifically, the first audio data is processed according to a preset sound effect algorithm. And the fourth audio data is used for outputting the audio data to a terminal of an audio receiving output party after being mixed with the third audio data subsequently, such as a terminal of each listener, so that each listener can hear the audio of the main broadcasting.

In step S13: and mixing the second audio data and the audio data used for accompaniment to generate first mixed sound data.

In the embodiment of the present disclosure, the second audio data and the third audio data are input to a first audio mixing module to mix the second audio data and the third audio data through the first audio mixing module, and the first audio mixing data is output to a peripheral sound device that receives the audio returning party terminal. The peripheral sound device of the receiving audio back-transmitting terminal includes but is not limited to a speaker, an earphone, and the like.

Furthermore, this disclosure is right after first audio data carries out second audio treatment and generates fourth audio data, it is right fourth audio data with third audio data carries out the audio mixing and generates second audio mixing data, it is right simultaneously fourth audio data with the completion will when third audio data carries out the audio mixing fourth audio data with third audio data carries out the time proofreading, so that fourth audio data with the time synchronization of third audio data to accomplish the alignment of people's voice and accompaniment when realizing the audio mixing, promote the audio treatment effect. The second audio mixing data is used for being pushed to a distribution server through a streaming media module, so that the distribution server distributes the second audio mixing data to each audio receiving side terminal.

This it is right that this a set of audio processing algorithm of preferred its self handles first audio data to reach individualized processing audio data's purpose, and through the better realization of control time delay fourth audio data with the time proofreading of third audio data accomplishes the alignment of voice and accompaniment, reach better K sings the effect, promote user experience.

In step S14: and outputting the first mixed voice data.

In the embodiment of the present disclosure, two sets of processing methods are performed on the input first mixed sound data to finally obtain two output results, that is, the first mixed sound data and the second mixed sound data. And the first audio mixing data is correspondingly output to the peripheral sound equipment of the audio return receiving terminal, so that the anchor can hear the sound of the anchor. And the second mixed sound data is output to a distribution server, so that the distribution server distributes the second mixed sound data to each audio receiving side terminal, and each listener can hear the sound of the main broadcasting.

The present disclosure has performed first sound effect processing and sound mixing on first sound mixing data output to the anchor, has performed second sound effect processing and sound mixing on second sound mixing data output to the listener terminal. Wherein the second sound effect processing is more focused on sound effect processing. In terms of controlling time delay, the present disclosure prefers a concurrent processing method to process the first audio data; in the aspect of sound effect processing, it is right that this disclosure prefers its one set of sound effect processing algorithm of self setting do individualized processing to first audio data to promote the effect of sound effect processing, promoted the effect that the ear of anchor returned on the one hand, on the other hand has promoted listener's experience.

Fig. 3 is a block flow diagram illustrating an audio processing apparatus according to an example embodiment. Referring to fig. 3, the apparatus includes an acquisition module 11, a processing module 12, a mixing module 13, and an output module 14.

The acquisition module 11 is configured to acquire first audio data.

In the embodiment of the disclosure, the acquisition module is configured to acquire and receive the first audio data input by the peripheral input device of the audio return terminal. The peripheral input equipment for receiving the audio and returning the audio to the terminal comprises a microphone, an earphone and a main broadcasting sound card.

The method utilizes the Audio Unit technology, integrates the advantages of the Audio Unit during recording on the premise of guaranteeing timeliness, performance and experience, and is applied to a live broadcast scene to achieve the optimal audio processing effect.

Specifically, the AudioUnit scheme used in the present disclosure is closer to the bottom layer of the iOS Audio system stack, and is used in cooperation with multiple components such as an AudioUnit Processing Graph Services management IO device (IO Unit), a mixer (Mix Unit), an effector (Effect Unit), and the like, to obtain a lower delay and a better Effect.

With continuing reference to fig. 2, fig. 2 is an architectural diagram of an iOS audio system stack, shown in accordance with an exemplary embodiment. As shown in FIG. 2, the iOS audio system stack includes Low-Level, Mid-Level, and High-Level.

Wherein, (1) Low-Level is mainly used in the audio APP implementation on MAC and under the condition that maximum real-time performance is required, most of the audio APPs do not need to use the services of this layer. Moreover, a high-level API with high real-time performance is provided on iOS to meet your demand. E.g., OpenAL, has real-time audio processing capability in games that is invoked directly with I/O.

And the I/O Kit is responsible for interacting with the hardware driver.

The Audio HAL, which is an Audio hardware abstraction layer, separates API calls from the actual hardware and remains independent.

Core MIDI, which provides a software abstraction layer of work for MIDI streams and devices.

And the Host Time Services is responsible for accessing the hardware clock of the computer.

(2) The Mid-Level layer has complete functions, including audio data format conversion, audio file reading and writing, audio stream analysis, plug-in work support and the like.

The Audio Convert Services are responsible for the conversion of Audio data formats.

The Audio File Services are responsible for reading and writing Audio data.

Audio Unit Services and Audio Processing Graph Services are plug-ins that support digital signal Processing such as equalizers and mixers.

Audio File Screen Services are responsible for stream parsing.

Core Audio Clock Services are responsible for Audio Clock synchronization.

(3) High-Level is a group of High-Level applications combined from low-Level interfaces, and basically much work on audio development can be done at this Level.

Audio Queue Services provide recording, playing, pausing, looping, and synchronizing Audio that automatically processes compressed Audio formats with the necessary codecs.

The AVAudio Player is an Objective-C interface-based audio playing class specially provided for the IOS platform, and can support the playing of all audio supported by the iOS.

Extended Audio File Services are composed of Audio files and Audio converters, and provide read-write capability for compressed and uncompressed Audio files.

OpenAL is the implementation of the CoreAudio to the OpenAL standard, and can play a 3D sound mixing effect.

The AudioUnit is a bottom layer technology of an iOS audio system stack, and the iOS provides audio processing plug-ins such as audio mixing, equalization, format conversion, real-time IO recording, playback, offline rendering, voice over internet protocol (VoIP), and the like, which all belong to different audiounits, and support dynamic loading and use. Audiounits can be created and used individually, but are more often used in combination in AudioProcessing Graph containers to meet diverse processing needs.

Based on the principle, the advantages of the Audio Unit during recording are integrated, and the Audio Unit is applied to a live scene to achieve the optimal audio processing effect.

The processing module 12 is configured to perform a first sound effect processing on the first audio data in a concurrent processing manner to generate second audio data.

Referring to fig. 4, fig. 4 is a schematic structural diagram of an audio processing apparatus according to an exemplary embodiment. As shown in fig. 4, the audio processing apparatus includes an input/output module 1, a first sound processing module 2, a second sound processing module 3, a playing module 4, a first mixing module 5, a second mixing module 6, and a streaming media module 7.

Specifically, the input/output module 1 includes an input unit 11 and an output unit 12, the playing module 4 includes an accompaniment playing unit 41 and a vocal playing unit 42, and the first mixing module 5 includes the second audio data input unit 51, an accompaniment/vocal input unit 52 and a mixing input unit 53.

Input unit 11 connects input devices such as microphone, headset and the anchor sound card that receives audio passback side terminal, the acquisition module acquires promptly in input unit 11 first audio data, first audio data transmission extremely first audio processing module 2 generates after the audio is handled second audio data transmission extremely first audio mixing module 5, first audio mixing module 5 is received first audio mixing data and play module 4 is afferent third audio data carries out the first audio mixing data process of audio mixing generation audio mixing output unit 53 exports to output unit 12, finally exports peripheral hardware such as earphone to receiving audio passback side terminal.

First audio data process second audio processing module 3 is handled the back and is generated fourth audio data, and export to second audio mixing module 6, second audio mixing module 6 receives fourth audio data and play module 4 input carry out the audio mixing after the third audio data and generate second audio mixing data and export to streaming media module 7, through streaming media module 7 propelling movement extremely the distribution server, so that the distribution server will second audio mixing data distribution to individual user terminal.

The processing module 12 includes a first processing unit configured to denoise, reverberate, equalize, and voicing the first audio data.

The present disclosure further includes an audio processing module configured to perform a second audio processing on the first audio data to generate fourth audio data. The sound effect processing module comprises: and the second processing unit is configured to process the first audio data according to a preset sound effect algorithm. The sound effect processing module further comprises a mixing unit, and the mixing unit is configured to mix the fourth audio data and the third audio data to generate second mixed data. The sound effect processing module further comprises a sending unit, and the sending unit is configured to push the second mixed sound data to a distribution server through a streaming media module, so that the distribution server distributes the second mixed sound data to each receiving audio output terminal. The mixing unit includes a proofreading unit configured to time-proofread the fourth audio data and the third audio data to time-synchronize the fourth audio data and the third audio data.

It should be noted that, in the present disclosure, the first audio data is preferably processed in a concurrent processing manner, so as to shorten the data processing time, improve the data processing efficiency, and finally realize controlling the ear return delay.

In addition, the audio time delay is controlled by the following measures:

in a first aspect, the present disclosure improves the acquisition mechanism of the acquisition module, so that the buffer of audio acquisition of the acquisition module can be controlled in a very small range to control the time delay. This is disclosed solves prior art through this mechanism, and the buffer memory of audio acquisition is great, and the output frequency is lower, is difficult to the problem of alignment time axis when handling background music.

In a second aspect, the present disclosure processes the first audio data and the second audio data in real time to shorten data processing time, improve data processing efficiency, and finally control audio time delay.

Further, according to the present disclosure, after the first audio data is acquired, second sound effect processing is further performed on the first audio data to generate fourth audio data. Specifically, the first audio data is processed according to a preset sound effect algorithm. And the fourth audio data is used for outputting the audio data to a terminal of a receiving audio output side after being mixed with the third audio data subsequently, so that each listener can hear the audio of the anchor.

The mixing module 13 is configured to mix the second audio data with third audio data for accompanying to generate first mixing data.

In the embodiment of the present disclosure, the second audio data and the third audio data are input to a sound mixing module to mix the second audio data and the third audio data through the sound mixing module, and the first mixed data is output to a peripheral sound device that receives the audio and returns the party terminal. The peripheral sound device of the receiving audio back-transmitting terminal includes but is not limited to a speaker, an earphone, and the like.

Furthermore, this disclosure is right after first audio data carries out second audio treatment and generates fourth audio data, it is right fourth audio data with third audio data carries out the audio mixing and generates second audio mixing data, it is right simultaneously fourth audio data with the completion will when third audio data carries out the audio mixing fourth audio data with third audio data carries out the time proofreading, so that fourth audio data with the time synchronization of third audio data to accomplish the alignment of people's voice and accompaniment when realizing the audio mixing, promote the audio treatment effect. And the second audio mixing data is used for being pushed to a distribution server through a streaming media module, so that the distribution server distributes the second audio mixing data to each audio receiving side terminal.

This it is right that this a set of audio processing algorithm of preferred its self handles first audio data to reach individualized processing audio data's purpose, and through the better realization of control time delay fourth audio data with the time proofreading of third audio data accomplishes the alignment of voice and accompaniment, reach better K sings the effect, promote user experience.

The output module 14 is configured to output the first mixed sound data.

The output module 14 includes an output unit configured to transmit the first mixed data to a peripheral sound device of a receiving audio backhauling terminal.

In the embodiment of the present disclosure, two sets of processing methods are performed on the input first mixed sound data to finally obtain two output results, that is, the first mixed sound data and the second mixed sound data. And the first audio mixing data is correspondingly output to the peripheral sound equipment of the audio return receiving terminal, so that the anchor can hear the sound of the anchor. And the second mixed sound data is output to a distribution server, so that the distribution server distributes the second mixed sound data to each audio receiving side terminal, and each listener can hear the sound of the main broadcasting.

The present disclosure has performed first sound effect processing and sound mixing on first sound mixing data output to the anchor, has performed second sound effect processing and sound mixing on second sound mixing data output to the listener terminal. Wherein the second sound effect processing is more focused on sound effect processing. In terms of controlling time delay, the present disclosure prefers a concurrent processing method to process the first audio data; in the aspect of sound effect processing, it is right that this disclosure prefers its one set of sound effect processing algorithm of self setting do individualized processing to first audio data to promote the effect of sound effect processing, promoted the effect that the ear of anchor returned on the one hand, on the other hand has promoted listener's experience.

Fig. 5 is a block diagram illustrating a method for a mobile terminal 800 according to an example embodiment. For example, the mobile terminal 800 may be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.

Referring to fig. 5, the apparatus 800 may include one or more of the following components: a processing component 802, a memory 804, a power component 806, a multimedia component 808, an audio component 810, an input/output (I/O) interface 812, a sensor component 814, and a communication component 816.

The processing component 802 generally controls overall operation of the device 800, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. The processing components 802 may include one or more processors 820 to execute instructions to perform all or a portion of the steps of the methods described above. Further, the processing component 802 can include one or more modules that facilitate interaction between the processing component 802 and other components. For example, the processing component 802 can include a multimedia module to facilitate interaction between the multimedia component 808 and the processing component 802.

The memory 804 is configured to store various types of data to support operation at the device 800. Examples of such data include instructions for any application or method operating on device 800, contact data, phonebook data, messages, pictures, videos, and so forth. The memory 804 may be implemented by any type or combination of volatile or non-volatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.

Power components 806 provide power to the various components of device 800. The power components 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the apparatus 800.

The multimedia component 808 includes a screen that provides an output interface between the device 800 and a user. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component 808 includes a front facing camera and/or a rear facing camera. The front-facing camera and/or the rear-facing camera may receive external multimedia data when the device 800 is in an operating mode, such as a shooting mode or a video mode. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.

The audio component 810 is configured to output and/or input audio signals. For example, the audio component 810 includes a Microphone (MIC) configured to receive external audio signals when the apparatus 800 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode. The received audio signals may further be stored in the memory 804 or transmitted via the communication component 816. In some embodiments, audio component 810 also includes a speaker for outputting audio signals.

The I/O interface 812 provides an interface between the processing component 802 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.

The sensor assembly 814 includes one or more sensors for providing various aspects of state assessment for the device 800. For example, the sensor assembly 814 may detect the open/closed state of the device 800, the relative positioning of the components, such as a display and keypad of the apparatus 800, the sensor assembly 814 may also detect a change in position of the apparatus 800 or a component of the apparatus 800, the presence or absence of user contact with the apparatus 800, orientation or acceleration/deceleration of the apparatus 800, and a change in temperature of the apparatus 800. Sensor assembly 814 may include a proximity sensor configured to detect the presence of a nearby object without any physical contact. The sensor assembly 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor assembly 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.

The communication component 816 is configured to facilitate communications between the apparatus 800 and other devices in a wired or wireless manner. The apparatus 800 may access a wireless network based on a communication standard, such as WiFi, an operator network (such as 2G, 3G, 4G, or 5G), or a combination thereof. In an exemplary embodiment, the communication component 816 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 816 further includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.

In an exemplary embodiment, the apparatus 800 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, micro-controllers, microprocessors or other electronic components for performing the audio processing methods of the present disclosure, the methods comprising: acquiring first audio data acquired by an acquisition module; carrying out first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data; performing sound mixing processing on the second audio data and the third audio data to generate first sound mixing data; and outputting the first mixed voice data.

In an exemplary embodiment, a non-transitory computer-readable storage medium comprising instructions, such as the memory 804 comprising instructions, executable by the processor 820 of the device 800 to perform the above-described method is also provided. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.

A non-transitory computer readable storage medium is provided in which instructions, when executed by a processor of a mobile terminal, enable the mobile terminal to perform an audio processing method, the method comprising: acquiring first audio data acquired by an acquisition module; carrying out first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data; performing sound mixing processing on the second audio data and the third audio data to generate first sound mixing data; and outputting the first mixed voice data. The processor can realize the functions of the acquisition module, the processing module, the mixing module and the output module of the audio processing device in the embodiment shown in fig. 3.

In an exemplary embodiment, there is also provided an application program, when instructions in the storage medium are executed by a processor of a mobile terminal, enabling the mobile terminal to perform an audio processing method, the method including acquiring first audio data acquired by an acquisition module; carrying out first sound effect processing on the first audio data in a concurrent processing mode to generate second audio data; performing sound mixing processing on the second audio data and the third audio data to generate first sound mixing data; and outputting the first mixed voice data. The processor can realize the functions of the acquisition module, the processing module, the mixing module and the output module of the audio processing device in the embodiment shown in fig. 3. .

Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.

It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

16页详细技术资料下载
上一篇:一种医用注射器针头装配设备
下一篇:声音处理方法、装置及存储介质

网友询问留言

已有0条留言

还没有人留言评论。精彩留言会获得点赞!

精彩留言,会给你点赞!