Sweeping robot and sweeping task planning method and device thereof

文档序号:666154 发布日期:2021-04-30 浏览:2次 中文

阅读说明:本技术 扫地机器人及其清扫任务规划方法和装置 (Sweeping robot and sweeping task planning method and device thereof ) 是由 檀冲 沈荻 张书新 李贝 于 2020-12-25 设计创作,主要内容包括:本发明适用于智能电器技术领域,提供了一种扫地机器人及其清扫任务规划方法和装置,其中,所述方法包括:获取扫地机器人周围环境的轮廓信息;当识别所述轮廓信息中包括预设的目标物体的轮廓信息时,获取所述目标物体的RGB图像;根据所述RGB图像,获取所述目标物体的语义信息;根据所述目标物体的语义信息,在三维语义地图中规划对所述目标物体的清扫任务。由于获取轮廓信息相比于获取三维语义地图所耗费的资源低很多,因此通过本发明可以有效提高清扫的效率。(The invention is suitable for the technical field of intelligent electric appliances, and provides a sweeping robot and a sweeping task planning method and device thereof, wherein the method comprises the following steps: acquiring contour information of the surrounding environment of the sweeping robot; when the contour information is recognized to comprise preset contour information of a target object, acquiring an RGB image of the target object; acquiring semantic information of the target object according to the RGB image; and planning a cleaning task of the target object in a three-dimensional semantic map according to the semantic information of the target object. Compared with the method for acquiring the three-dimensional semantic map, the method for acquiring the outline information has the advantages that the resource consumption is much lower, so that the cleaning efficiency can be effectively improved.)

1. A cleaning task planning method is characterized by comprising the following steps:

acquiring contour information of the surrounding environment of the sweeping robot;

when the contour information is recognized to comprise preset contour information of a target object, acquiring an RGB image of the target object;

acquiring semantic information of the target object according to the RGB image;

and planning a cleaning task of the target object in a three-dimensional semantic map according to the semantic information of the target object.

2. The cleaning task planning method according to claim 1, wherein the obtaining of the contour information of the environment around the sweeping robot comprises:

acquiring depth data around the sweeping robot;

determining contour information of the environment around the sweeping robot based on the depth image.

3. The cleaning task planning method according to claim 1 or 2, wherein the three-dimensional contour information includes: three-dimensional contour information and two-dimensional contour information.

4. The cleaning task planning method according to claim 1, wherein acquiring an RGB image of the target object includes:

starting an RGB camera on the sweeping robot;

and controlling the RGB camera to shoot to obtain an RGB image containing the target object.

5. The cleaning task planning method according to claim 1 or 4, wherein obtaining semantic information of the target object according to the RGB image specifically includes:

according to the RGB image, semantic information of the surrounding environment of the sweeping robot is obtained, wherein the contour information comprises three-dimensional contour information;

and establishing a three-dimensional semantic map according to the contour information and the semantic information of the surrounding environment.

6. A cleaning task planning apparatus, comprising:

the contour information acquisition module is used for acquiring contour information of the surrounding environment of the sweeping robot;

the image information acquisition module is used for acquiring an RGB image of the target object when the contour information is identified to include the preset contour information of the target object;

the semantic information identification module is used for acquiring the semantic information of the target object according to the RGB image;

and the cleaning task planning module is used for planning a cleaning task of the target object in the three-dimensional semantic map according to the semantic information of the target object.

7. The cleaning task planning device according to claim 6, wherein the contour information obtaining module includes:

the depth data acquisition unit is used for acquiring depth data around the sweeping robot;

and the contour information establishing unit is used for determining the contour information of the surrounding environment of the sweeping robot based on the depth number image.

8. The cleaning task planning device according to claim 6 or 7, wherein the three-dimensional contour information includes: three-dimensional contour information and two-dimensional contour information.

9. The cleaning task planning device according to claim 6, wherein the acquisition image information acquisition module includes:

the camera control unit is used for starting an RGB camera on the sweeping robot;

and the image acquisition unit is used for controlling the RGB camera to shoot to obtain an RGB image containing the target object.

10. The cleaning task planning device according to claim 6 or 9, wherein the semantic information identifying module specifically includes:

the semantic recognition unit is used for acquiring semantic information of the surrounding environment of the sweeping robot according to the RGB image, wherein the contour information comprises three-dimensional contour information;

and the map building unit is used for building a three-dimensional semantic map according to the contour information and the semantic information of the surrounding environment.

11. A sweeping robot comprising:

a sweeping assembly;

an RGB camera;

a structured light sensor;

a memory storing a computer program;

a processor connected to the structured light sensor, the RGB camera, the cleaning assembly and the memory, respectively, wherein the steps of the cleaning task planning method according to any one of claims 1 to 5 are implemented when the processor executes the computer program.

Technical Field

The invention belongs to the technical field of intelligent electric appliances, and particularly relates to a sweeping robot and a sweeping task planning method and device thereof.

Background

The existing sweeping robot inevitably collides with surrounding objects in the process of executing a sweeping task. In practice, some objects are very expensive or fragile and can easily be damaged if the sweeping robot collides with them during the sweeping task. Therefore, the user wants to precisely clean some special objects by the sweeping robot, so that the sweeping robot does not collide with the objects during the cleaning process.

In the prior art, to avoid collision between the sweeping robot and a target object, high-precision sweeping can be achieved only by providing map information with rich information, namely planning a sweeping task in a map with higher precision. However, in practice, it is necessary to consume huge computing resources to realize such high-precision cleaning, and therefore, the cleaning efficiency is reduced, i.e., the cleaning process is slow.

Therefore, how to further improve the high-precision cleaning task planning aiming at the special object while ensuring the cleaning efficiency of the cleaning robot is a technical problem discovered in the research and development process.

Disclosure of Invention

In view of this, embodiments of the present invention provide a sweeping robot and a sweeping task planning method and apparatus thereof, so as to solve the problem in the prior art how to further improve high-precision sweeping task planning for a special object while ensuring sweeping efficiency of the sweeping robot.

In a first aspect of the embodiments of the present invention, a cleaning task planning method is provided, which includes: acquiring contour information of the surrounding environment of the sweeping robot; when the contour information is recognized to comprise preset contour information of a target object, acquiring an RGB image of the target object; acquiring semantic information of the target object according to the RGB image; and planning a cleaning task of the target object in a three-dimensional semantic map according to the semantic information of the target object.

In some alternatives, the obtaining of the contour information of the environment around the sweeping robot includes: acquiring depth data around the sweeping robot; determining contour information of the environment around the sweeping robot based on the depth image.

In some alternatives, the three-dimensional profile information includes: three-dimensional contour information and two-dimensional contour information.

In some alternatives, acquiring an RGB image of the target object includes: starting an RGB camera on the sweeping robot; and controlling the RGB camera to shoot to obtain an RGB image containing the target object.

In some alternatives, obtaining semantic information of the target object according to the RGB image specifically includes: according to the RGB image, semantic information of the surrounding environment of the sweeping robot is obtained, wherein the contour information comprises three-dimensional contour information; and establishing a three-dimensional semantic map according to the contour information and the semantic information of the surrounding environment.

In a second aspect of the embodiments of the present invention, there is provided a cleaning task planning apparatus, including: the contour information acquisition module is used for acquiring contour information of the surrounding environment of the sweeping robot; the image information acquisition module is used for acquiring an RGB image of the target object when the contour information is identified to include the preset contour information of the target object; the semantic information identification module is used for acquiring the semantic information of the target object according to the RGB image; and the cleaning task planning module is used for planning a cleaning task of the target object in the three-dimensional semantic map according to the semantic information of the target object.

In some alternatives, the contour information acquisition module includes: the depth data acquisition unit is used for acquiring depth data around the sweeping robot; and the contour information establishing unit is used for determining the contour information of the surrounding environment of the sweeping robot based on the depth number image.

In some alternatives, the three-dimensional profile information includes: three-dimensional contour information and two-dimensional contour information.

In some alternatives, the acquired image information acquisition module includes: the camera control unit is used for starting an RGB camera on the sweeping robot; and the image acquisition unit is used for controlling the RGB camera to shoot to obtain an RGB image containing the target object.

In some alternatives, the semantic information identifying module specifically includes: the semantic recognition unit is used for acquiring semantic information of the surrounding environment of the sweeping robot according to the RGB image, wherein the contour information comprises three-dimensional contour information; and the map building unit is used for building a three-dimensional semantic map according to the contour information and the semantic information of the surrounding environment.

In a third aspect of the embodiments of the present invention, there is provided a sweeping robot, including: a sweeping assembly; an RGB camera; a structured light sensor; a memory storing a computer program; a processor connected to the structured light sensor, the RGB camera, the cleaning component and the memory, respectively, wherein the processor implements the steps of the cleaning task planning method according to any one of the first aspect when executing the computer program.

Compared with the prior art, the embodiment of the invention has the following beneficial effects: according to the invention, when the sweeping robot encounters a target object, the RGB camera is started to acquire the RGB image, the semantics of the target object are identified, and then the cleaning task is planned for the target object according to the three-dimensional semantic map.

Drawings

In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed for the embodiments or the prior art descriptions will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings without creative efforts.

Fig. 1 is a flowchart of a cleaning task planning method provided in a first embodiment of the present invention;

FIG. 2 is a schematic diagram of a cleaning task planning apparatus according to a first embodiment of the present invention;

fig. 3 is a schematic diagram of a sweeping robot to which some embodiments of the cleaning task planning method and the cleaning task planning device described above can be applied.

Detailed Description

In the following description, for purposes of explanation and not limitation, specific details are set forth, such as particular system structures, techniques, etc. in order to provide a thorough understanding of the embodiments of the invention. It will be apparent, however, to one skilled in the art that the present invention may be practiced in other embodiments that depart from these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present invention with unnecessary detail.

In order to explain the technical means of the present invention, the following description will be given by way of specific examples.

Example one

Fig. 1 is a flowchart of a cleaning task planning method according to a first embodiment of the present invention.

As shown in FIG. 1, the cleaning task planning method includes the following four steps S01-S04.

Step S01: and acquiring the contour information of the surrounding environment of the sweeping robot.

Wherein the contour information includes three-dimensional contour information and two-dimensional contour information.

Specifically, the contour information of the target object can be identified in the following steps by the three-dimensional contour information. The two-dimensional contour information may include projection information of the three-dimensional contour information, and when a cleaning task is planned using the contour information, the cleaning task may be planned using the projection information.

More specifically, the three-dimensional contour information may be obtained by acquiring depth data of the surrounding environment by the sweeping robot. For example, a sweeping robot is provided with a structured light sensor or a laser radar or an RGB-D camera to acquire surrounding depth data, and then the three-dimensional profile information is acquired according to the depth data. In addition, the three-dimensional contour information acquired by other equipment for the surrounding environment of the sweeping robot can be acquired by sending the three-dimensional contour information to the sweeping robot through a network. For example, a device provided with a structured light sensor or a laser radar or an RGB-D camera acquires depth data of the environment around the sweeping robot, then generates three-dimensional contour information from the depth data, and transmits the three-dimensional contour information to the sweeping robot through a network. It should be understood that the present application is not limited to the specific manner of obtaining the three-dimensional contour information.

In an example, the step S01 may specifically include:

step S11: acquiring depth data around the sweeping robot;

step S12: determining profile information of an environment surrounding the sweeping robot based on the depth data.

The depth data includes information on distances from respective positions of the surrounding environment to the image capturing position, and therefore, profile information of the surrounding environment can be determined from the depth data. In practical applications, the method for obtaining the depth data may refer to the method described above, and the present application does not limit this.

Step S02: and when the contour information is recognized to comprise the preset contour information of the target object, acquiring the RGB image of the target object.

The target object is an object for which the sweeping robot needs to perform a precision planning and cleaning task by combining a three-dimensional semantic map.

Specifically, the identification of the contour information includes preset contour information of the target object, the user may specifically photograph the target object in advance to establish corresponding three-dimensional contour information, then compare the three-dimensional contour information around the sweeping robot with the three-dimensional contour information of the target object, and determine whether the preset target contour information exists according to a comparison result.

Or, the identification of the contour information includes preset contour information of the target object, or may be a classification model established by a machine learning algorithm on sample data of the target object, and then the established classification model is used to identify the contour information acquired around the sweeping robot, so as to determine whether the contour information of the preset target object exists.

In addition, in combination with the example in step S01, for the case where it is recognized that the contour information of the preset target object is not included in the contour information, the cleaning task may be planned by directly using the contour information determined from the depth data, for example, three-dimensional contour information or two-dimensional contour information. Since the planning of such cleaning tasks is not a contribution of the present application to the prior art, it will not be described in detail.

In an example, in the step S02, the acquiring the RGB image of the target object may specifically include:

step S211: starting an RGB camera on the sweeping robot;

step S212: and controlling the RGB camera to photograph the surrounding environment of the sweeping robot to obtain an RGB image containing the target object.

This example is applicable to the condition that the robot of sweeping the floor is provided with the RGB camera simultaneously, obtains its RGB image of surrounding by the RGB camera of robot of sweeping the floor.

In another example, in the step S02, the acquiring an RGB image of the target object may further specifically include:

step S221: sending a photographing instruction to external equipment connected with the sweeping robot, wherein the photographing instruction is used for controlling an RGB (red, green and blue) camera to photograph the surrounding environment of the sweeping robot to obtain an RGB image containing the target object, and the external equipment comprises the RGB camera;

step S222: and acquiring the RGB image returned by the external equipment responding to the photographing instruction.

The example is suitable for the situation that the sweeping robot is not provided with the RGB camera, and the RGB images around the sweeping robot can be acquired through external equipment, so that the accuracy of the sweeping task is improved.

Step S03: and acquiring semantic information of the target object according to the RGB image.

Upon acquiring the RGB image, the RGB image may be identified to acquire semantic information of the surrounding environment. Wherein the semantic information includes, but is not limited to, the contour confidence, the object type, etc. of the target object. Here, the semantic information of the target object is acquired in order to perform more accurate cleaning task planning on the target object by using the three-dimensional semantic map in the following steps.

In an example, the step S02 may specifically include:

step S31: obtaining semantic information of the surrounding environment of the sweeping robot according to the RGB image, wherein the semantic information comprises semantic information of the target object;

step S32: and establishing a three-dimensional semantic map according to the three-dimensional outline information and the semantic information of the surrounding environment of the sweeping robot.

The three-dimensional semantic map is built through the acquired RGB images and the contour information, so that the cleaning task planning of the target object is performed with high precision in the following steps.

Step S04: and planning a cleaning task of the target object in a three-dimensional semantic map according to the semantic information of the target object.

Because the information contained in the three-dimensional semantic map is richer and more accurate than the information contained in the three-dimensional outline information, the cleaning task is planned according to the three-dimensional semantic map more accurately, the situation of mistaken collision is not easy to occur, and higher resources are consumed for constructing the three-dimensional semantic map, so that the cleaning efficiency is reduced if the cleaning task is planned by the three-dimensional semantic map all the time in the task execution process of the sweeping robot. Therefore, in the embodiment, the cleaning task is generally planned by using the profile information determined by the depth data, when a target object is encountered, the RGB camera is started to recognize the semantics of the target object, and then the cleaning task is planned according to the semantic information of the target object according to the three-dimensional semantic map.

It should be understood that, the sequence numbers of the steps in the foregoing embodiments do not imply an execution sequence, and the execution sequence of each process should be determined by its function and inherent logic, and should not constitute any limitation to the implementation process of the embodiments of the present invention.

Example two

Fig. 2 is a schematic diagram of a cleaning task planning apparatus according to a first embodiment of the present invention.

The cleaning task planning apparatus provided in this embodiment is the same as the cleaning task planning method, and specifically, as shown in fig. 2, the cleaning task planning apparatus 200 includes: the contour information acquisition module 201 is used for acquiring contour information of the surrounding environment of the sweeping robot; an image information obtaining module 202, configured to obtain an RGB image of a target object when it is identified that the contour information includes preset contour information of the target object; the semantic information identification module 203 is used for acquiring semantic information of the target object according to the RGB image; and the cleaning task planning module 204 is configured to plan a cleaning task for the target object in a three-dimensional semantic map according to the semantic information of the target object.

In some optional examples, the contour information obtaining module includes: the depth data acquisition unit is used for acquiring depth data around the sweeping robot; and the contour information establishing unit is used for determining the contour information of the surrounding environment of the sweeping robot based on the depth number image.

In some optional examples, the three-dimensional contour information includes: three-dimensional contour information and two-dimensional contour information.

In some optional examples, the acquired image information acquisition module includes: the camera control unit is used for starting an RGB camera on the sweeping robot; and the image acquisition unit is used for controlling the RGB camera to shoot to obtain an RGB image containing the target object.

In some optional examples, the semantic information identifying module specifically includes: the semantic recognition unit is used for acquiring semantic information of the surrounding environment of the sweeping robot according to the RGB image, wherein the contour information comprises three-dimensional contour information; and the map building unit is used for building a three-dimensional semantic map according to the contour information and the semantic information of the surrounding environment.

Since the present embodiment and the above-mentioned embodiment belong to the same inventive concept, the modules or units therein correspond to the steps in the above-mentioned method one to one, and for this reason, the two have the same specific technical features. Therefore, for the specific content of the device, reference may be made to the description and illustration in the first embodiment, which is not described herein again.

EXAMPLE III

Fig. 3 is a schematic diagram of a sweeping robot to which some embodiments of the cleaning task planning method and the cleaning task planning apparatus described above can be applied.

Referring to fig. 3, the sweeping robot 300 includes: a sweeping component 303; an RGB camera 305; a structured light sensor 304; a memory 302 storing a computer program 303; a processor 301, respectively connected to the structured light sensor 304, the RGB camera 305, the cleaning component 303 and the memory 302, wherein the processor 301 executes the computer program 303 to implement the steps of the cleaning task planning method according to any one of the embodiments, or the processor 301 executes the computer program to implement the functions of the modules/units in the apparatus embodiments of the second embodiment, such as the functions of the modules 201 to 204 shown in fig. 2.

Illustratively, the computer program 303 may be partitioned into one or more modules/units that are stored in the memory 302 and executed by the processor 301 to implement the present invention. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, which are used to describe the execution process of the computer program 303 in the sweeping robot 300. For example, the computer program 303 may be divided into a synchronization module, a summary module, an acquisition module, and a return module (a module in a virtual device), and each module has the following specific functions:

the sweeping robot 300 may be a desktop computer, a notebook, a palm computer, a cloud server, or other computing devices. The sweeping robot may include, but is not limited to, a processor 301 and a memory 302. Those skilled in the art will appreciate that fig. 3 is merely an example of the sweeping robot 300, and does not constitute a limitation of the sweeping robot 300, and may include more or fewer components than those shown, or some components in combination, or different components, for example, the sweeping robot may further include input and output devices, network access devices, buses, etc.

The Processor 301 may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.

The memory 302 may be an internal storage unit of the cleaning robot 300, such as a hard disk or a memory of the cleaning robot 300. The memory 302 may also be an external storage device of the sweeping robot 300, such as a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), and the like, which are provided on the sweeping robot 300. Further, the memory 302 may also include both an internal memory unit and an external memory device of the sweeping robot 300. The memory 302 is used to store the computer program and other programs and data required by the sweeping robot. The memory 302 may also be used to temporarily store data that has been output or is to be output.

It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-mentioned division of the functional units and modules is illustrated, and in practical applications, the above-mentioned function distribution may be performed by different functional units and modules according to needs, that is, the internal structure of the apparatus is divided into different functional units or modules to perform all or part of the above-mentioned functions. Each functional unit and module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units and modules are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present application. The specific working processes of the units and modules in the system may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.

In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment.

Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.

In the embodiments provided in the present invention, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described apparatus embodiments are merely illustrative, and for example, the division of the modules or units is only one logical division, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or may be integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.

The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.

In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.

The integrated modules/units, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium. Based on such understanding, all or part of the processes in the method according to the above embodiments may be implemented by a computer program, which may be stored in a computer readable storage medium and used by a processor to implement the steps of the cleaning task planning method embodiments. Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM), Random Access Memory (RAM), electrical carrier wave signals, telecommunications signals, software distribution medium, and the like. It should be noted that the computer readable medium may contain content that is subject to appropriate increase or decrease as required by legislation and patent practice in jurisdictions, for example, in some jurisdictions, computer readable media does not include electrical carrier signals and telecommunications signals as is required by legislation and patent practice.

The above-mentioned embodiments are only used for illustrating the technical solutions of the present invention, and not for limiting the same; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; such modifications and substitutions do not substantially depart from the spirit and scope of the embodiments of the present invention, and are intended to be included within the scope of the present invention.

12页详细技术资料下载
上一篇:一种医用注射器针头装配设备
下一篇:基于扫地机的清洁方法、装置、可读存储介质及电子设备

网友询问留言

已有0条留言

还没有人留言评论。精彩留言会获得点赞!

精彩留言,会给你点赞!