Device control method and apparatus, storage medium, and electronic apparatus

文档序号:193327 发布日期:2021-11-02 浏览:32次 中文

阅读说明:本技术 设备的控制方法和装置、存储介质及电子装置 (Device control method and apparatus, storage medium, and electronic apparatus ) 是由 姜娜娜 于 2021-06-11 设计创作,主要内容包括:本申请公开了一种设备的控制方法和装置、存储介质、电子装置,其中,该方法包括:在检测到目标对象进入与第一设备对应的第一区域范围的情况下,获取第一设备所监听到的目标语音数据,其中,第一设备为允许免唤醒进行语音交互的设备,第一区域范围为免唤醒进行语音交互的情况下解析第一设备所监听到的语音数据的范围;获取与目标语音数据对应的目标解析结果,其中,目标解析结果是对目标语音数据进行解析所得到的结果,目标解析结果包含待第二设备执行的目标指令;按照目标解析结果,控制第二设备执行目标指令。通过本申请,解决了相关技术中通过免唤醒与家电设备进行语音交互的方式存在的家电设备易被误唤醒的问题。(The application discloses a device control method and device, a storage medium and an electronic device, wherein the method comprises the following steps: the method comprises the steps that under the condition that a target object is detected to enter a first area range corresponding to first equipment, target voice data monitored by the first equipment are obtained, wherein the first equipment is equipment which allows the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out; acquiring a target analysis result corresponding to the target voice data, wherein the target analysis result is obtained by analyzing the target voice data and comprises a target instruction to be executed by the second equipment; and controlling the second equipment to execute the target instruction according to the target analysis result. Through the method and the device, the problem that the household appliance is easy to be mistakenly awaken in a voice interaction mode with the household appliance without awakening in the related technology is solved.)

1. A method of controlling a device, comprising:

the method comprises the steps that under the condition that a target object is detected to enter a first area range corresponding to first equipment, target voice data monitored by the first equipment are obtained, wherein the first equipment is equipment which allows the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out;

acquiring a target analysis result corresponding to the target voice data, wherein the target analysis result is a result obtained by analyzing the target voice data and comprises a target instruction to be executed by second equipment;

and controlling the second equipment to execute the target instruction according to the target analysis result.

2. The method of claim 1, wherein prior to obtaining the target speech data heard by the first device, the method further comprises:

detecting that the target object enters a second area range, wherein the second area range is a range of which the distance between the first device and the target object is smaller than or equal to a target distance threshold value;

and automatically waking up the first equipment, wherein the automatically woken-up first equipment enters a state to be interacted, and the state to be interacted is a state of monitoring voice data.

3. The method of claim 2, wherein after automatically waking the first device, the method further comprises:

determining that the target object enters a target sound pickup range, wherein the target sound pickup range is a range corresponding to a target sound pickup angle of a target sound pickup part on the first device, the first area range is an intersection of the target sound pickup range and the second area range, and the target voice data is voice data picked up by the target sound pickup part.

4. The method of claim 2, wherein after automatically waking the first device, the method further comprises:

sending first prompt information through the first equipment, wherein the first prompt information is used for prompting that the first equipment is awakened.

5. The method of claim 2, wherein after automatically waking the first device, the method further comprises:

and controlling the first equipment to enter a sleep state when the target object is detected to move out of the range of the second area.

6. The method of claim 1, wherein obtaining the target parsing result corresponding to the target speech data comprises:

sending the target voice data to a cloud server, wherein the cloud server is used for analyzing the received voice data to obtain an analysis result;

and receiving the target analysis result obtained by analyzing the target voice data by the cloud server.

7. The method of any of claims 1-6, wherein after controlling the second device to execute the target instruction, the method further comprises:

and sending second prompt information through the first equipment, wherein the second prompt information is used for prompting that the second equipment executes the target instruction.

8. A control apparatus of a device, characterized by comprising:

the device comprises a first obtaining unit, a second obtaining unit and a processing unit, wherein the first obtaining unit is used for obtaining target voice data monitored by first equipment under the condition that a target object is detected to enter a first area range corresponding to the first equipment, the first equipment is equipment which allows the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out;

a second obtaining unit, configured to obtain a target analysis result corresponding to the target voice data, where the target analysis result is a result obtained by analyzing the target voice data, and the target analysis result includes a target instruction to be executed by a second device;

and the first control unit is used for controlling the second equipment to execute the target instruction according to the target analysis result.

9. A computer-readable storage medium, comprising a stored program, wherein the program when executed performs the method of any of claims 1 to 7.

10. An electronic device comprising a memory and a processor, characterized in that the memory has stored therein a computer program, the processor being arranged to execute the method of any of claims 1 to 7 by means of the computer program.

Technical Field

The present application relates to the field of communications, and in particular, to a method and an apparatus for controlling a device, a storage medium, and an electronic apparatus.

Background

At present, when household appliances such as sound boxes, air conditioners, televisions and the like carry out voice interaction, a wake-up word needs to be used for waking up, and the use is troublesome and unnatural. If the wake-up word is not used, i.e. the voice interaction is performed without wake-up, a false wake-up situation may occur. For example, chats between users are often recognized and replied to, disturbing the user's life.

Therefore, in the related technology, the voice interaction with the household appliance is carried out in a wake-up-free mode, so that the problem that the household appliance is easy to be woken up by mistake exists.

Disclosure of Invention

The embodiment of the application provides a device control method and device, a storage medium and an electronic device, which are used for at least solving the problem that the household appliance is easy to be mistakenly awakened in a voice interaction mode with the household appliance without being awakened in the related technology.

According to an aspect of an embodiment of the present application, there is provided a method of controlling a device, including: the method comprises the steps that under the condition that a target object is detected to enter a first area range corresponding to first equipment, target voice data monitored by the first equipment are obtained, wherein the first equipment is equipment which allows the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out; acquiring a target analysis result corresponding to the target voice data, wherein the target analysis result is a result obtained by analyzing the target voice data and comprises a target instruction to be executed by second equipment; and controlling the second equipment to execute the target instruction according to the target analysis result.

In an exemplary embodiment, before obtaining the target speech data heard by the first device, the method further comprises: detecting that the target object enters a second area range, wherein the second area range is a range of which the distance between the first device and the target object is smaller than or equal to a target distance threshold value; and automatically waking up the first equipment, wherein the automatically woken-up first equipment enters a state to be interacted, and the state to be interacted is a state of monitoring voice data.

In one exemplary embodiment, after automatically waking up the first device, the method further comprises: determining that the target object enters a target sound pickup range, wherein the target sound pickup range is a range corresponding to a target sound pickup angle of a target sound pickup part on the first device, the first area range is an intersection of the target sound pickup range and the second area range, and the target voice data is voice data picked up by the target sound pickup part.

In one exemplary embodiment, after automatically waking up the first device, the method further comprises: sending first prompt information through the first equipment, wherein the first prompt information is used for prompting that the first equipment is awakened.

In one exemplary embodiment, after automatically waking up the first device, the method further comprises: and controlling the first equipment to enter a sleep state when the target object is detected to move out of the range of the second area.

In one exemplary embodiment, obtaining the target parsing result corresponding to the target voice data includes: sending the target voice data to a cloud server, wherein the cloud server is used for analyzing the received voice data to obtain an analysis result; and receiving the target analysis result obtained by analyzing the target voice data by the cloud server.

In one exemplary embodiment, after controlling the second device to execute the target instruction, the method further comprises: and sending second prompt information through the first equipment, wherein the second prompt information is used for prompting that the second equipment executes the target instruction.

According to another aspect of the embodiments of the present application, there is also provided a control apparatus of a device, including: the device comprises a first obtaining unit, a second obtaining unit and a processing unit, wherein the first obtaining unit is used for obtaining target voice data monitored by first equipment under the condition that a target object is detected to enter a first area range corresponding to the first equipment, the first equipment is equipment which allows the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out; a second obtaining unit, configured to obtain a target analysis result corresponding to the target voice data, where the target analysis result is a result obtained by analyzing the target voice data, and the target analysis result includes a target instruction to be executed by a second device; and the first control unit is used for controlling the second equipment to execute the target instruction according to the target analysis result.

In one exemplary embodiment, the apparatus further comprises: a first detection unit, configured to detect that the target object enters a second area range before the target voice data monitored by the first device is acquired, where the second area range is a range in which a distance from the first device is smaller than or equal to a target distance threshold; and the awakening unit is used for automatically awakening the first equipment, wherein the automatically awakened first equipment enters a state to be interacted, and the state to be interacted is a state for monitoring voice data.

In one exemplary embodiment, the apparatus further comprises: the determination unit is configured to determine that the target object enters a target sound pickup range after the first device is automatically woken up, where the target sound pickup range is a range corresponding to a target sound pickup angle of a target sound pickup unit on the first device, the first area range is an intersection of the target sound pickup range and the second area range, and the target voice data is voice data picked up by the target sound pickup unit.

In one exemplary embodiment, the apparatus further comprises: the first prompting unit is used for sending first prompting information through the first equipment after the first equipment is automatically awakened, wherein the first prompting information is used for prompting that the first equipment is awakened.

In one exemplary embodiment, the apparatus further comprises: and the second control unit is used for controlling the first equipment to enter a sleep state when the target object is detected to move out of the range of the second area after the first equipment is automatically awakened.

In one exemplary embodiment, the second acquisition unit includes: the sending module is used for sending the target voice data to a cloud server, wherein the cloud server is used for analyzing the received voice data to obtain an analysis result; and the receiving module is used for receiving the target analysis result obtained by analyzing the target voice data by the cloud server.

In one exemplary embodiment, the apparatus further comprises: and the second prompting unit is used for sending second prompting information through the first equipment after controlling the second equipment to execute the target instruction, wherein the second prompting information is used for prompting the second equipment to execute the target instruction.

According to still another aspect of the embodiments of the present application, there is also provided a computer-readable storage medium having a computer program stored therein, wherein the computer program is configured to execute the control method of the above apparatus when running.

According to another aspect of the embodiments of the present application, there is also provided an electronic apparatus, including a memory, a processor, and a computer program stored in the memory and executable on the processor, wherein the processor executes the control method of the above apparatus through the computer program.

In the embodiment of the application, a way of performing wake-up-free voice interaction on a user in a certain area range is adopted, and target voice data monitored by first equipment is acquired under the condition that a target object is detected to enter a first area range corresponding to the first equipment, wherein the first equipment is equipment allowing the wake-up-free voice interaction, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition of the wake-up-free voice interaction; acquiring a target analysis result corresponding to the target voice data, wherein the target analysis result is obtained by analyzing the target voice data and comprises a target instruction to be executed by the second equipment; according to the target analysis result, the second device is controlled to execute the target instruction, and the device only carries out wake-free voice interaction on the object in the specific area range, so that the aim of reducing the possibility of recognizing and replying voice data which does not need to be recognized can be achieved, the technical effects of reducing the false wake-up rate of the device and improving the use experience of a user are achieved, and the problem that the household appliance is easily mistakenly woken up in a voice interaction mode with the household appliance through wake-free in the related technology is solved.

Drawings

The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present application and together with the description, serve to explain the principles of the application.

In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings needed to be used in the description of the embodiments or the prior art will be briefly described below, and it is obvious for those skilled in the art to obtain other drawings without inventive exercise.

FIG. 1 is a schematic diagram of a hardware environment for an alternative method of controlling a device according to an embodiment of the present application;

FIG. 2 is a schematic flow chart diagram illustrating an alternative method of controlling a device according to an embodiment of the present application;

FIG. 3 is a schematic flow chart diagram of an alternative method of controlling a device according to an embodiment of the present application;

FIG. 4 is a block diagram of an alternative apparatus control arrangement according to an embodiment of the present application;

fig. 5 is a block diagram of an alternative electronic device according to an embodiment of the present application.

Detailed Description

In order to make the technical solutions better understood by those skilled in the art, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only partial embodiments of the present application, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.

It should be noted that the terms "first," "second," and the like in the description and claims of this application and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application described herein are capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.

According to an aspect of an embodiment of the present application, there is provided a control method of an apparatus. Alternatively, in the present embodiment, the control method of the above-described device may be applied to a hardware environment formed by the terminal 102 and the server 104 as shown in fig. 1. As shown in fig. 1, the server 104 is connected to the terminal 102 through a network, and may be configured to provide services (e.g., game services, application services, etc.) for the terminal or a client installed on the terminal, and may be configured with a database on the server or separately from the server, and configured to provide data storage services for the server 104.

The network may include, but is not limited to, at least one of: wired networks, wireless networks. The wired network may include, but is not limited to, at least one of: wide area networks, metropolitan area networks, local area networks, which may include, but are not limited to, at least one of the following: WIFI (Wireless Fidelity), bluetooth. The terminal 102 may not be limited to a PC, a mobile phone, a tablet computer, etc.

The control method of the device in the embodiment of the present application may be executed by the server 104, the terminal 102, or both the server 104 and the terminal 102. The terminal 102 may execute the control method of the device according to the embodiment of the present application, or may be executed by a client installed thereon.

Taking the terminal 102 to execute the method for controlling the device in this embodiment as an example, fig. 2 is a schematic flowchart of an optional method for controlling the device according to the embodiment of the present application, and as shown in fig. 2, the method may include the following steps:

step S202, when it is detected that the target object enters a first area range corresponding to the first device, obtaining target voice data monitored by the first device, where the first device is a device that allows wake-up-free voice interaction, and the first area range is a range within which the voice data monitored by the first device is analyzed when wake-up-free voice interaction is performed.

The control method of the device in this embodiment may be applied to a scenario where wake-up-free voice interaction is performed with the device, where wake-up-free voice interaction refers to voice interaction with the device without using a wake-up word or other wake-up devices. The device may be a first device, and the first device may be a smart device, and may include, but is not limited to, a smart home device (e.g., a home device such as a sound box, an air conditioner, a television, etc.), and may also be other functional devices. In the present embodiment, a home appliance is taken as an example for description.

The first device may be a device allowing wake-free voice interaction, which may be placed in a living room, bedroom, etc. If the monitoring state is kept, the chatting among the users is often recognized and replied by the first device, and the life of the users is disturbed. And, also violate user privacy. Optionally, in this embodiment, it may be configured to analyze a range of the voice data monitored by the first device under the condition of wake-up-free voice interaction, that is, a first area range, and under the condition of wake-up-free voice interaction, only when the user is in the first area range, the voice data monitored by the first device is analyzed.

The first device (which may also be other than the first device) may perform human body detection (or living body detection) to determine whether an object enters a first area range corresponding to the first device. If it is detected that an object (i.e., a target object) enters the first area range, the first device may acquire voice data monitored by the device, i.e., target voice data.

The target voice data may be voice data monitored by the first device after the target object enters the first area range, or may be voice data monitored after the first device is voice-woken, for example, the voice automatic wake-up is triggered under a certain condition, and the voice data monitoring can be performed after the voice automatic wake-up. In this embodiment, the target voice data is not limited.

It should be noted that, although the scenario of avoiding waking up for voice interaction is described as an example in the present embodiment, this does not unduly limit the device that executes the control method of the device in the present embodiment. The first device allows for voice interaction without waking up, and may also allow for voice interaction by waking up with a wake-up word, and when the wake-up word is used for waking up, the area range of the first device allowed to wake up may be defined by the configuration information, or the area range of the first device allowed to wake up may not be defined, and in this embodiment, the scene woken up with the wake-up word is not defined.

Step S204, a target analysis result corresponding to the target voice data is obtained, wherein the target analysis result is obtained by analyzing the target voice data, and the target analysis result comprises a target instruction to be executed by the second device.

For the target voice data, the first device may acquire a target parsing result corresponding to the target voice data, the target parsing result being a result of parsing the target voice data. The device for parsing the target voice data may be the first device, or may be another device (e.g., a cloud server) other than the first device.

For example, the first device may be communicatively coupled to other devices. The first device can send the target voice data to other devices through the communication connection between the first device and the second device for voice analysis to obtain a target analysis result, and receives the target analysis result sent by the other devices through the communication connection between the first device and the second device.

The target resolution result may be a target instruction carrying data to be executed by the second device. The second device may be the same device as the first device or may be a different device. For example, the first device may be a smart speaker and the second device may be a smart air conditioner, a smart refrigerator, a humidifier, etc. The target instruction corresponds to the target voice data, and may be any operation instruction that the second device is allowed to execute, such as a voice broadcast instruction, an open instruction, a close instruction, and the like.

And step S206, controlling the second equipment to execute the target instruction according to the target analysis result.

According to the target analysis result, the first device can control the second device to execute the target instruction. The first device may execute the target instruction if the first device and the second device are the same device. If the first device and the second device are different devices, the first device can transmit the target instruction to the second device through the communication connection with the second device so as to control the second device to execute the target instruction.

For example, the first device is an intelligent sound box, the second device and the first device are the same device, the target voice data is "what is so today's weather", the target instruction is to play the today's weather, and the intelligent sound box can execute the target instruction to play the today's weather.

For another example, the first device is a smart speaker, the second device is a dehumidifier, the target voice data is "the family is too humid, the dehumidifier bar is opened", the target instruction is to open the dehumidifier, and the smart speaker can control the dehumidifier to perform an opening operation.

Through the steps, under the condition that a target object is detected to enter a first area range corresponding to first equipment, target voice data monitored by the first equipment is obtained, wherein the first equipment is equipment allowing the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out; acquiring a target analysis result corresponding to the target voice data, wherein the target analysis result is obtained by analyzing the target voice data and comprises a target instruction to be executed by the second equipment; according to the target analysis result, the second device is controlled to execute the target instruction, the problem that the household appliance is easily mistakenly awakened in a voice interaction mode without awakening and with the household appliance in the related technology is solved, the mistaken awakening rate of the appliance is reduced, and the use experience of a user is improved.

As an alternative embodiment, before acquiring the target voice data heard by the first device, the method further includes:

s11, detecting that the target object enters a second area range, wherein the second area range is a range of which the distance between the target object and the first equipment is smaller than or equal to a target distance threshold;

and S12, automatically waking up the first device, wherein the automatically woken up first device enters a to-be-interacted state, and the to-be-interacted state is a state of monitoring voice data.

The first device may be in a voice wake-up state, i.e., a to-be-interacted state (a state in which voice data listening is performed) all the time. In this state, the first device may monitor the voice data all the time, and perform the voice data parsing only when the target object enters the first area range.

Alternatively, in order to reduce resource consumption (e.g., power consumption, consumption of processing resources, etc.) of the device, human body detection may be performed by using a sensor (e.g., an infrared sensor or an ultrasonic sensor) or the like to avoid that the voice is always in a listening state, for example, whether the voice needs to be awakened may be determined according to a location where the user is located. A target sensor may be provided on the first device, which may be used for human activity detection (detecting whether there is human activity). The first device may be in a sleep state (low power consumption state) when human activity is not detected by the target sensor within the second area range (range where the distance to the first device is less than or equal to the target distance threshold).

Here, the distance between the target object and the first device may be a distance between the target object and the target sensor; the absence of human activity detected in the second area may be the absence of any human activity detected by the target sensor, or the absence of human activity detected by the target sensor may be the absence of human activity detected by the target sensor.

When the target object is detected to enter the second area range through the target sensor, the first device is automatically awakened, namely, the voice interaction function of the first device is automatically awakened. And the automatically awakened first equipment enters a state to be interacted.

For example, a hardware sensor may be added to the home appliance and a value (e.g., 1 meter) may be set. When a person enters the range of 1 meter of the sensor, the person is automatically awakened by voice to enter a state to be interacted.

It should be noted that the first area range may be the same as the second area range, or may be a subset of the second area range, for example, after the voice wake-up, the voice data monitored by the first device may be directly analyzed, or it may be further determined whether the position of the target object meets a condition for analyzing the voice data, and only when the condition for analyzing the voice data is met, the voice data is analyzed.

Through the embodiment, the voice awakening area range is configured, and the voice awakening is automatically performed when the user enters the area range, so that the resource consumption of the equipment can be reduced, and the rationality of resource utilization is improved.

As an optional embodiment, after automatically waking up the first device, the method further includes:

and S21, determining that the target object enters a target sound pickup range, wherein the target sound pickup range is a range corresponding to a target sound pickup angle of a target sound pickup part on the first equipment, the target sound pickup range is a first area range, and the target voice data is voice data picked up by the target sound pickup part.

Except that whether need awaken the pronunciation according to the position judgement that the user is located, can also judge whether need to resolve the corpus according to the pickup angle, for example, upload the corpus to the high in the clouds and resolve. Alternatively, a target sound pickup part (e.g., a microphone array) may be provided on the first device, and a sound pickup angle of the target sound pickup part may be set to a target sound pickup angle, e.g., an angle of 60 °, and a sound pickup range of the target sound pickup part is a range corresponding to the target sound pickup angle, i.e., a target sound pickup range.

After the voice auto-wake, the first device may further determine whether the target object enters a sound pickup range of the target sound pickup section. If the target object enters the pick-up range of the target pick-up component, the first device may determine that parsing of the material is required. In this case, the first device may acquire the target voice data that the first device hears.

It is to be understood that the aforementioned first area range is an intersection of the target sound pickup range and the second area range, that is, when it is detected that the target object enters the second area range and is within the sound pickup range of the target sound pickup section, the first device acquires the above-mentioned target voice data. The first device listens for voice data through the target sound pickup part, and correspondingly, the target voice data is the voice data picked up by the target sound pickup part.

For example, the voice pickup range of the home device may be set to an angle (e.g., 60 °), and the recognition reply may be performed only on the voice within the 60 ° angle range without using the wake-up word. Then, when a person enters the home device within 1 meter and within 60 °, the home device can perform a wake-up free voice interaction.

Through this embodiment, be the pickup angle through setting up the pickup part, confirm whether need parse the material according to the pickup angle, can reduce the mistake awakening rate of equipment, improve user's use and experience.

As an optional embodiment, after automatically waking up the first device, the method further includes:

and S31, sending a first prompt message through the first device, wherein the first prompt message is used for prompting that the first device is awakened.

After the voice automatic wake-up, the first device may send a first prompt message to prompt that the first device has been woken up, and voice interaction may be performed. After receiving the first prompt message, the user can enter voice input, and can also ignore the prompt message.

The manner of issuing the first prompt message may be various, and may include, but is not limited to, at least one of the following: voice broadcast, screen prompt. For example, the first device sends out a voice prompt to prompt that the target object can perform voice interaction; for another example, the front panel display of the first device lights up.

Through this embodiment, awaken the back through the suggestion information prompt facility and awaken up automatically at pronunciation, can improve the convenience that the information acquireed, improve user's use simultaneously and experience.

As an optional embodiment, after automatically waking up the first device, the method further includes:

s41, controlling the first device to enter a sleep state when the target object is detected to move outside the second zone range.

As the target object moves, the distance between the target object and the first device changes. If it is detected that the target object moves outside the second area range, that is, the distance between the target object and the first device is greater than the target distance threshold, the first device may continue to remain in the to-be-interacted state.

Alternatively, in order to reduce resource consumption (e.g., power consumption, etc.) of the first device, the first device may be controlled to enter the sleep state upon detecting that the target object moves outside the second area range. The sleep state may be a low power consumption state, for example, where the power supply supplies power only to necessary components (e.g., the object sensor) and does not supply power to other components (e.g., the panel display), thereby reducing power consumption of the first device.

Optionally, in order to avoid continuous monitoring of the voice dialog of the user by the first device, when it is detected that the target object moves outside the range of the second area, the first device may be controlled to enter a sleep state, and the first device is prohibited from monitoring the voice dialog of the user, so that privacy information of the user is prevented from being leaked. The sleep state may be a low power consumption state, for example, where the power supply supplies power only to necessary components (e.g., the object sensor) and does not supply power to other components (e.g., the panel display).

Through the embodiment, when the human body is detected to move beyond the set range, the control equipment enters the sleep state, so that the resource consumption of the equipment can be reduced, and the rationality of resource utilization is improved.

As an alternative embodiment, the obtaining of the target parsing result corresponding to the target voice data includes:

s51, sending the target voice data to a cloud server, wherein the cloud server is used for analyzing the received voice data to obtain an analysis result;

and S52, receiving a target analysis result obtained by analyzing the target voice data by the cloud server.

Parsing the target speech data may be performed by the first device or by a device other than the first device. In this embodiment, a cloud server (an example of other devices) may perform an analysis operation on the target voice data, and the cloud server may be configured to analyze the received voice data to obtain an analysis result of the voice data.

A target communication connection may be established between the first device and the cloud server. The first device can transmit the target voice data to the cloud server through the target communication connection, and the cloud server analyzes the target voice data to obtain a target analysis result. The first device can also receive the target analysis result transmitted by the cloud server through the target communication connection, so that the target analysis result is obtained.

Through this embodiment, upload the high in the clouds with the corpus, carry out the corpus by the high in the clouds and resolve, can reduce the corpus and resolve the occupation to equipment resource, reduce the demand to equipment throughput simultaneously.

As an alternative embodiment, after controlling the second device to execute the target instruction, the method further includes:

and S61, sending a second prompt message by the first device, wherein the second prompt message is used for prompting the second device that the target instruction is executed.

After controlling the second device to execute the target instruction, neither the first device nor the second device may execute any feedback operation of the execution result, or at least one of the first device and the second device may issue a prompt message to prompt the second device that the target instruction has been executed.

When the target voice data is acquired, the target object is located in the first area range corresponding to the first device, and therefore the target object can more easily receive prompt information sent by the first device. In this embodiment, the first device may issue a second prompt message, which may be used to prompt the second device that the target instruction has been executed.

The manner of issuing the second prompt message may be various, and may include, but is not limited to, at least one of the following: voice broadcast, screen prompt. For example, the first device may issue a voice prompt, prompting the second device that the target instruction has been executed; for another example, the second prompt message is displayed on a panel display of the first device.

Through the embodiment, after the instruction is executed, the equipment for acquiring the voice data sends the prompt information to prompt that the instruction is executed, so that a user can conveniently know the execution condition of the instruction, and the convenience of information acquisition is improved.

The following explains a control method of the apparatus in the embodiment of the present application with reference to an alternative example. In this example, the first device and the second device are both home devices; the human body detection is performed by a sensor on the household appliance, and the sound pickup operation is performed by a microphone array on the household appliance; the reminding range of the sensor is 1 meter (an example of a target distance threshold), the target voice pickup angle is 60 degrees, namely when a person enters the range of 1 meter and is within the 60 degrees, the wake-up-free voice interaction can be carried out; and analyzing the corpus by the cloud.

The control method of the device provided in this optional example reduces the false wake-up rate of wake-up avoidance by determining the distance position, performs living body detection within the setting range by the sensor, does not always stay in the monitoring state, and reduces the possibility of recognizing the user's chatting and replying by setting the voice pickup range.

As shown in fig. 3, the flow of the control method of the apparatus in this alternative example may include the steps of:

step S302, a sensor detects human body movement;

step S304, judging whether the human body is within the range of 1 meter, if so, executing the step S306, otherwise, ending;

step S306, automatically waking up the voice to enter a state to be interacted;

step S308, judging whether the human body is in the 60-degree angle range of the microphone array, if so, executing step S310, otherwise, ending;

step S310, the cloud analyzes the voice and returns a result;

in step S312, the content is broadcasted by voice or the device executes the command.

In step S312, different voice data may correspond to different execution operations, for example, for a voice instruction to query weather, the home device may broadcast the query result in a voice, and for example, for a voice instruction to control a device, the controlled device may execute a corresponding command.

The following describes a control method of the above-described apparatus with reference to an optional scenario. When the owner returns home, the sensor of the front door panel senses that someone enters a range of 1 meter, the panel display screen is lightened, and the voice is automatically awakened to enter a state to be interacted. The owner says in front of the panel: "the family is too wet, open the dehumidifier bar". The panel receives the instruction, answers "good owner, which opens the dehumidifier for you", and opens the dehumidifier. The host walks into the living room, the panel screen is displayed, and the panel voice enters the sleep state.

The host walks to the living room and passes through the television, the television sensor detects that a person passes through the television, but the television does not respond when the person does not enter the range of 1 meter. The owner sits the sofa, and the audio amplifier detects someone's activity to within 1 meter, the audio amplifier pronunciation awaken up automatically and get into and wait the interactive state, and owner calls the phone for friend and chats, and the audio amplifier detects not in the pickup range, does not upload the high in the clouds and does speech recognition, and the audio amplifier does not have the feedback.

According to the optional example, whether the voice needs to be awakened or not is judged according to the position of the user, whether the corpus is uploaded to the cloud end for analysis or not is judged according to the pickup angle, and the voice can be prevented from being in a monitoring state all the time by utilizing living body detection; moreover, voice interaction is carried out by utilizing living body detection, distance and sound pickup range, so that the false awakening rate can be reduced.

It should be noted that, for simplicity of description, the above-mentioned method embodiments are described as a series of acts or combination of acts, but those skilled in the art will recognize that the present application is not limited by the order of acts described, as some steps may occur in other orders or concurrently depending on the application. Further, those skilled in the art should also appreciate that the embodiments described in the specification are preferred embodiments and that the acts and modules referred to are not necessarily required in this application.

Through the above description of the embodiments, those skilled in the art can clearly understand that the method according to the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but the former is a better implementation mode in many cases. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (e.g., a ROM (Read-Only Memory)/RAM (Random Access Memory), a magnetic disk, an optical disk) and includes several instructions for enabling a terminal device (e.g., a mobile phone, a computer, a server, or a network device) to execute the methods according to the embodiments of the present application.

According to another aspect of the embodiments of the present application, there is also provided a control apparatus of an apparatus for implementing the control method of the apparatus described above. Fig. 4 is a block diagram of a control apparatus of an alternative device according to an embodiment of the present application, and as shown in fig. 4, the apparatus may include:

a first obtaining unit 402, configured to obtain target voice data monitored by a first device when it is detected that a target object enters a first area range corresponding to the first device, where the first device is a device that allows wake-up-free voice interaction, and the first area range is a range within which the voice data monitored by the first device is analyzed when the wake-up-free voice interaction is performed;

a second obtaining unit 404, connected to the first obtaining unit 402, configured to obtain a target parsing result corresponding to the target voice data, where the target parsing result is a result obtained by parsing the target voice data, and the target parsing result includes a target instruction to be executed by the second device;

and the first control unit 406 is connected to the second obtaining unit 404, and is configured to control the second device to execute the target instruction according to the target analysis result.

It should be noted that the first obtaining unit 402 in this embodiment may be configured to execute the step S202, the second obtaining unit 404 in this embodiment may be configured to execute the step S204, and the first control unit 406 in this embodiment may be configured to execute the step S206.

Through the module, under the condition that a target object is detected to enter a first area range corresponding to first equipment, target voice data monitored by the first equipment is obtained, wherein the first equipment is equipment allowing the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out; acquiring a target analysis result corresponding to the target voice data, wherein the target analysis result is obtained by analyzing the target voice data and comprises a target instruction to be executed by the second equipment; according to the target analysis result, the second device is controlled to execute the target instruction, the problem that the household appliance is easily mistakenly awakened in a voice interaction mode without awakening and with the household appliance in the related technology is solved, the mistaken awakening rate of the appliance is reduced, and the use experience of a user is improved.

As an alternative embodiment, the apparatus further comprises:

the first detection unit is used for detecting that the target object enters a second area range before the target voice data monitored by the first equipment is acquired, wherein the second area range is a range of which the distance between the second area range and the first equipment is smaller than or equal to a target distance threshold value;

and the awakening unit is used for automatically awakening the first equipment, wherein the automatically awakened first equipment enters a state to be interacted, and the state to be interacted is a state of monitoring voice data.

As an alternative embodiment, the apparatus further comprises:

the device comprises a determining unit and a processing unit, wherein the determining unit is used for determining that a target object enters a target sound pickup range after automatically waking up the first device, the target sound pickup range is a range corresponding to a target sound pickup angle of a target sound pickup part on the first device, the target sound pickup range is a first area range, and target voice data is voice data picked up by the target sound pickup part.

As an alternative embodiment, the apparatus further comprises:

the first prompting unit is used for sending first prompting information through the first equipment after the first equipment is automatically awakened, wherein the first prompting information is used for prompting that the first equipment is awakened.

As an alternative embodiment, the apparatus further comprises:

and the second control unit is used for controlling the first equipment to enter a sleep state when the target object is detected to move out of the range of the second area after the first equipment is automatically awakened.

As an alternative embodiment, the second obtaining unit 404 includes:

the transmitting module is used for transmitting the target voice data to the cloud server, wherein the cloud server is used for analyzing the received voice data to obtain an analysis result;

and the receiving module is used for receiving a target analysis result obtained by analyzing the target voice data by the cloud server.

As an alternative embodiment, the apparatus further comprises:

and the second prompting unit is used for sending second prompting information through the first equipment after controlling the second equipment to execute the target instruction, wherein the second prompting information is used for prompting the second equipment to execute the target instruction.

It should be noted here that the modules described above are the same as the examples and application scenarios implemented by the corresponding steps, but are not limited to the disclosure of the above embodiments. It should be noted that the modules described above as a part of the apparatus may be operated in a hardware environment as shown in fig. 1, and may be implemented by software, or may be implemented by hardware, where the hardware environment includes a network environment.

According to still another aspect of an embodiment of the present application, there is also provided a storage medium. Alternatively, in this embodiment, the storage medium may be used to execute a program code of a control method of any one of the apparatuses in this embodiment.

Optionally, in this embodiment, the storage medium may be located on at least one of a plurality of network devices in a network shown in the above embodiment.

Optionally, in this embodiment, the storage medium is configured to store program code for performing the following steps:

s1, acquiring target voice data monitored by first equipment under the condition that a target object is detected to enter a first area range corresponding to the first equipment, wherein the first equipment is equipment allowing the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out;

s2, acquiring a target analysis result corresponding to the target voice data, wherein the target analysis result is obtained by analyzing the target voice data and comprises a target instruction to be executed by the second device;

and S3, controlling the second equipment to execute the target instruction according to the target analysis result.

Optionally, the specific example in this embodiment may refer to the example described in the above embodiment, which is not described again in this embodiment.

Optionally, in this embodiment, the storage medium may include, but is not limited to: various media capable of storing program codes, such as a U disk, a ROM, a RAM, a removable hard disk, a magnetic disk, or an optical disk.

According to still another aspect of the embodiments of the present application, there is also provided an electronic device for implementing the control method of the above apparatus, which may be a server, a terminal, or a combination thereof.

Fig. 5 is a block diagram of an alternative electronic device according to an embodiment of the present application, as shown in fig. 5, including a processor 502, a communication interface 504, a memory 506, and a communication bus 508, wherein the processor 502, the communication interface 504, and the memory 506 are communicated with each other via the communication bus 508, and wherein,

a memory 506 for storing a computer program;

the processor 502, when executing the computer program stored in the memory 506, implements the following steps:

s1, acquiring target voice data monitored by first equipment under the condition that a target object is detected to enter a first area range corresponding to the first equipment, wherein the first equipment is equipment allowing the voice interaction without waking up, and the first area range is a range for analyzing the voice data monitored by the first equipment under the condition that the voice interaction without waking up is carried out;

s2, acquiring a target analysis result corresponding to the target voice data, wherein the target analysis result is obtained by analyzing the target voice data and comprises a target instruction to be executed by the second device;

and S3, controlling the second equipment to execute the target instruction according to the target analysis result.

Alternatively, in this embodiment, the communication bus may be a PCI (Peripheral Component Interconnect) bus, an EISA (Extended Industry Standard Architecture) bus, or the like. The communication bus may be divided into an address bus, a data bus, a control bus, etc. For ease of illustration, only one thick line is shown in FIG. 5, but this is not intended to represent only one bus or type of bus. The communication interface is used for communication between the electronic device and other equipment.

The memory may include RAM, and may also include non-volatile memory (non-volatile memory), such as at least one disk memory. Alternatively, the memory may be at least one memory device located remotely from the processor.

As an example, the memory 506 may include, but is not limited to, a first obtaining unit 402, a second obtaining unit 404, and a first control unit 406 in a control apparatus including the above devices. In addition, other module units in the control device of the above-mentioned apparatus may also be included, but are not limited to these, and are not described in detail in this example.

The processor may be a general-purpose processor, and may include but is not limited to: a CPU (Central Processing Unit), an NP (Network Processor), and the like; but also a DSP (Digital Signal Processing), an ASIC (Application Specific Integrated Circuit), an FPGA (Field Programmable Gate Array) or other Programmable logic device, discrete Gate or transistor logic device, discrete hardware component.

Optionally, the specific examples in this embodiment may refer to the examples described in the above embodiments, and this embodiment is not described herein again.

It can be understood by those skilled in the art that the structure shown in fig. 5 is only an illustration, and the device implementing the control method of the device may be a terminal device, and the terminal device may be a terminal device such as a smart phone (e.g., an Android phone, an iOS phone, etc.), a tablet computer, a palm computer, a Mobile Internet Device (MID), a PAD, and the like. Fig. 5 is a diagram illustrating a structure of the electronic device. For example, the electronic device may also include more or fewer components (e.g., network interfaces, display devices, etc.) than shown in FIG. 5, or have a different configuration than shown in FIG. 5.

Those skilled in the art will appreciate that all or part of the steps in the methods of the above embodiments may be implemented by a program instructing hardware associated with the terminal device, where the program may be stored in a computer-readable storage medium, and the storage medium may include: flash disk, ROM, RAM, magnetic or optical disk, and the like.

The above-mentioned serial numbers of the embodiments of the present application are merely for description and do not represent the merits of the embodiments.

The integrated unit in the above embodiments, if implemented in the form of a software functional unit and sold or used as a separate product, may be stored in the above computer-readable storage medium. Based on such understanding, the technical solution of the present application may be substantially implemented or a part of or all or part of the technical solution contributing to the prior art may be embodied in the form of a software product stored in a storage medium, and including instructions for causing one or more computer devices (which may be personal computers, servers, network devices, or the like) to execute all or part of the steps of the method described in the embodiments of the present application.

In the above embodiments of the present application, the descriptions of the respective embodiments have respective emphasis, and for parts that are not described in detail in a certain embodiment, reference may be made to related descriptions of other embodiments.

In the several embodiments provided in the present application, it should be understood that the disclosed client may be implemented in other manners. The above-described embodiments of the apparatus are merely illustrative, and for example, the division of the units is only one type of division of logical functions, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or may be integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, units or modules, and may be in an electrical or other form.

The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, and may also be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution provided in the embodiment.

In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.

The foregoing is only a preferred embodiment of the present application and it should be noted that those skilled in the art can make several improvements and modifications without departing from the principle of the present application, and these improvements and modifications should also be considered as the protection scope of the present application.

18页详细技术资料下载
上一篇:一种医用注射器针头装配设备
下一篇:联动场景执行方法和装置、存储介质及电子设备

网友询问留言

已有0条留言

还没有人留言评论。精彩留言会获得点赞!

精彩留言,会给你点赞!