Video display method, device, equipment and storage medium

文档序号:1044343 发布日期:2020-10-09 浏览:11次 中文

阅读说明:本技术 视频展示方法、装置、设备及存储介质 (Video display method, device, equipment and storage medium ) 是由 金海善 李勇 王威 裴建军 王欢 邹辉 李睿 于 2019-03-28 设计创作,主要内容包括:本申请公开了一种视频展示方法、装置、设备及存储介质。方法包括:获取视频图像,视频图像由一个或多个摄像机拍摄的图像得到;确定视频图像中的参照物,参照物的数量为多个;获取不同数据采集方式所采集的包括参照物的相关数据;根据相关数据,确定参照物的物理位置信息;根据参照物的物理位置信息及参照物在视频图像中的位置信息,确定视频图像中目标点的物理位置信息;基于视频图像中目标点的物理位置信息关联展示目标点的基础数据。本申请能够将不同类型的数据及应用进行互通,实现了多种数据采集方式的融合,丰富了视频展示方式,进而可提升用户体验。(The application discloses a video display method, a video display device, video display equipment and a storage medium. The method comprises the following steps: acquiring video images, wherein the video images are obtained from images shot by one or more cameras; determining a plurality of reference objects in the video image; acquiring related data including a reference object acquired by different data acquisition modes; determining physical position information of the reference object according to the related data; determining the physical position information of the target point in the video image according to the physical position information of the reference object and the position information of the reference object in the video image; and associating basic data for displaying the target point based on the physical position information of the target point in the video image. According to the method and the device, data and applications of different types can be intercommunicated, fusion of multiple data acquisition modes is achieved, video display modes are enriched, and user experience can be improved.)

1. A method for video presentation, the method comprising:

acquiring video images, wherein the video images are obtained from images shot by one or more cameras;

determining a plurality of reference objects in the video image;

acquiring related data including the reference object acquired by different data acquisition modes;

determining physical position information of the reference object according to the related data;

determining the physical position information of a target point in the video image according to the physical position information of the reference object and the position information of the reference object in the video image;

and associating and displaying the basic data of the target point based on the physical position information of the target point in the video image.

2. The method of claim 1, wherein the obtaining the relevant data including the reference object acquired by the different data acquisition modes comprises:

determining a target data acquisition mode according to the application scene displayed by the video and the type of the reference object;

and acquiring the related data including the reference object acquired by the target data acquisition mode.

3. The method of claim 2, wherein determining the physical location information of the reference object based on the correlation data comprises:

when the target data acquisition mode is a first data acquisition mode, the related data comprises a three-dimensional video obtained according to the first data acquisition mode;

fusing the video image and the three-dimensional video to obtain a fused video to be processed;

determining the position information of the reference object in the video to be processed;

and determining the physical position information of the reference object from the related data according to the position information of the reference object in the video to be processed.

4. The method of claim 2, wherein determining the physical location information of the reference object based on the correlation data comprises:

and matching and determining the physical position information of the reference object from the related data according to the attribute information of the reference object.

5. The method according to any one of claims 1-4, wherein said associating the base data showing the target point based on the physical location information of the target point in the video image comprises:

displaying the video image in a first display area, and displaying the target point information in the video image;

and displaying the basic data of the target point in a second display area.

6. The method of claim 5, wherein said presenting the base data of the target point in the second presentation area comprises:

and if the number of the basic data of the target point is multiple, performing space-time superposition on the multiple basic data of the target point, and displaying the superposed basic data of the target point in the second display area.

7. The method of claim 5, wherein said presenting the base data of the target point in the second presentation area comprises:

if the number of the basic data of the target point is multiple, determining the level relation among the multiple basic data of the target point, and displaying the basic data of the target point in a grading manner according to the level relation among the multiple basic data in the second display area.

8. A video presentation apparatus, said apparatus comprising:

the device comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for acquiring video images, and the video images are obtained from images shot by one or more cameras;

the first determination module is used for determining a plurality of reference objects in the video image;

the second acquisition module is used for acquiring the related data including the reference object acquired by different data acquisition modes;

the second determining module is used for determining the physical position information of the reference object according to the related data;

the third determining module is used for determining the physical position information of the target point in the video image according to the physical position information of the reference object and the position information of the reference object in the video image;

and the display module is used for displaying the basic data of the target point in the video image in a correlation manner based on the physical position information of the target point.

9. A computer device comprising a processor and a memory, the memory having stored therein at least one instruction which, when executed by the processor, implements a video presentation method as claimed in any one of claims 1 to 7.

10. A computer-readable storage medium having stored therein at least one instruction which, when executed, implements a video presentation method as recited in any one of claims 1 to 7.

Technical Field

The embodiment of the application relates to the technical field of data processing, in particular to a video display method, a video display device, video display equipment and a storage medium.

Background

With the continuous development of video application, more and more video acquisition modes and more acquired video types are provided.

Due to the diversification of the acquisition mode and the video type, how to display the acquired video becomes a key problem for video application.

Disclosure of Invention

The embodiment of the application provides a video display method, a video display device, video display equipment and a storage medium, which can be used for solving the problems in the related art. The technical scheme is as follows:

in one aspect, an embodiment of the present application provides a video display method, where the method includes:

acquiring video images, wherein the video images are obtained from images shot by one or more cameras;

determining a plurality of reference objects in the video image;

acquiring related data including the reference object acquired by different data acquisition modes;

determining physical position information of the reference object according to the related data;

determining the physical position information of a target point in the video image according to the physical position information of the reference object and the position information of the reference object in the video image;

and associating and displaying the basic data of the target point based on the physical position information of the target point in the video image.

In a possible implementation manner of the present application, the acquiring related data including the reference object acquired by different data acquisition manners includes:

determining a target data acquisition mode according to the application scene displayed by the video and the type of the reference object;

and acquiring the related data including the reference object acquired by the target data acquisition mode.

In a possible implementation manner of the present application, the determining physical location information of the reference object according to the related data includes:

when the target data acquisition mode is a first data acquisition mode, the related data comprises a three-dimensional video obtained according to the first data acquisition mode;

fusing the video image and the three-dimensional video to obtain a fused video to be processed;

determining the position information of the reference object in the video to be processed;

and determining the physical position information of the reference object from the related data according to the position information of the reference object in the video to be processed.

In a possible implementation manner of the present application, the determining physical location information of the reference object according to the related data includes:

and matching and determining the physical position information of the reference object from the related data according to the attribute information of the reference object.

In a possible implementation manner of the present application, the associating and displaying the basic data of the target point based on the physical location information of the target point in the video image includes:

displaying the video image in a first display area, and displaying the target point information in the video image;

and displaying the basic data of the target point in a second display area.

In a possible implementation manner of the present application, the presenting the basic data of the target point in the second presentation area includes:

and if the number of the basic data of the target point is multiple, performing space-time superposition on the multiple basic data of the target point, and displaying the superposed basic data of the target point in the second display area.

In a possible implementation manner of the present application, the presenting the basic data of the target point in the second presentation area includes:

if the number of the basic data of the target point is multiple, determining the level relation among the multiple basic data of the target point, and displaying the basic data of the target point in a grading manner according to the level relation among the multiple basic data in the second display area.

In a possible implementation manner of the present application, the determining, according to the physical location information of the reference object and the location information of the reference object in the video image, the physical location information of the target point in the video image includes:

establishing a position mapping model according to the physical position information of the reference object and the position information of the reference object in the video image;

determining physical location information of a target point in the video image based on the location mapping model.

In one aspect, a video presentation device is provided, the device comprising:

the device comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for acquiring video images, and the video images are obtained from images shot by one or more cameras;

the first determination module is used for determining a plurality of reference objects in the video image;

the second acquisition module is used for acquiring the related data including the reference object acquired by different data acquisition modes;

the second determining module is used for determining the physical position information of the reference object according to the related data;

the third determining module is used for determining the physical position information of the target point in the video image according to the physical position information of the reference object and the position information of the reference object in the video image;

and the display module is used for displaying the basic data of the target point in the video image in a correlation manner based on the physical position information of the target point.

In a possible implementation manner of the present application, the second obtaining module is configured to determine a target data acquisition manner according to an application scene displayed by a video and a type of the reference object; and acquiring the related data including the reference object acquired by the target data acquisition mode.

In a possible implementation manner of the present application, the second obtaining module is configured to, when the target data acquisition manner is a first data acquisition manner, include, in the related data, a three-dimensional video obtained according to the first data acquisition manner; fusing the video image and the three-dimensional video to obtain a fused video to be processed; determining the position information of the reference object in the video to be processed; and determining the physical position information of the reference object from the related data according to the position information of the reference object in the video to be processed.

In a possible implementation manner of the present application, the second obtaining module is configured to determine, by matching, physical location information of the reference object from the related data according to the attribute information of the reference object.

In a possible implementation manner of the present application, the display module is configured to display the video image in a first display area, and display the target point information in the video image; and displaying the basic data of the target point in a second display area.

In a possible implementation manner of the present application, the display module is configured to perform space-time superposition on multiple pieces of basic data of the target point if the number of pieces of basic data of the target point is multiple, and display the superposed basic data of the target point in the second display area.

In a possible implementation manner of the present application, the display module is configured to determine a level relationship between a plurality of pieces of basic data of the target point if the number of pieces of basic data of the target point is multiple, and hierarchically display the basic data of the target point in the second display area according to the level relationship between the plurality of pieces of basic data.

In a possible implementation manner of the present application, the third determining module is configured to establish a position mapping model according to the physical position information of the reference object and the position information of the reference object in the video image; determining physical location information of a target point in the video image based on the location mapping model.

In one aspect, a computer device is provided, the computer device comprising a processor and a memory, the memory having stored therein at least one instruction, which when executed by the processor, implements the video presentation method as described in any one of the above.

In one aspect, a computer-readable storage medium is provided, having at least one instruction stored therein, which when executed, implements a video presentation method as described in any one of the above.

The technical scheme provided by the embodiment of the application at least has the following beneficial effects:

after the physical position information of the reference object is determined through the related data including the reference object acquired by different data acquisition modes, the physical position information of the target point in the video image is determined based on the physical position information of the reference object and the position information of the reference object in the video image, so that the basic data of the target point is displayed through the physical position information, different types of data and applications can be intercommunicated, the integration of multiple data acquisition modes is realized, the video display modes are enriched, and the user experience can be improved.

Drawings

In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.

FIG. 1 is a schematic illustration of an implementation environment provided by an embodiment of the present application;

fig. 2 is a flowchart of a video display method according to an embodiment of the present application;

FIG. 3 is a schematic view of a video display interface provided in an embodiment of the present application;

FIG. 4 is a schematic view of a video display interface provided in an embodiment of the present application;

fig. 5 is a schematic structural diagram of a video display apparatus according to an embodiment of the present application;

fig. 6 is a schematic structural diagram of a video display apparatus according to an embodiment of the present application.

Detailed Description

To make the objects, technical solutions and advantages of the present application more clear, embodiments of the present application will be described in further detail below with reference to the accompanying drawings.

With the continuous development of video applications, around the digitization of smart cities, a variety of application ecology is derived, such as a Geographic Information System (GIS), a high-precision GIS, a three-dimensional (3D) GIS, a map of each department of government, and the like, all around collecting city data based on a unified Geographic space. The city data has many kinds, large data quantity and high complexity, so that no suitable logic can be unified up to now. And the video intelligent application is more various, virtual reality, augmented reality, mixed reality, various wearable equipment, smart machine etc..

The embodiment of the application provides a video display method aiming at the situation that various data acquisition modes and video data types are diversified. The method can be applied to the video presentation system shown in fig. 1. As shown in fig. 1, the system includes: a data acquisition device 11, a data processing device 12 and a video presentation device 13.

The data acquisition device 11 is used for acquiring video data. For example, the data acquisition device 11 may be a video camera and the acquired data may be video images. The data acquisition device 11 may also be an unmanned aerial vehicle, which may perform 3D oblique photography. The data acquisition device 11 may also be a data acquisition vehicle equipped with various sensors such as high-precision positioning, laser radar, video, etc., and can acquire a high-precision map. The embodiment of the present application does not limit the type of the data acquisition device 11 and the manner in which the data acquisition device 11 acquires data. In addition, in the video display system provided in the embodiment of the present application, the number of the data acquisition devices 11 may be two or more.

Because the high-precision map runs on a road through a vehicle, the video of the target with obvious characteristics along the running path is collected, and the GPS position of the target can be extracted, the collected target positions are all absolute GPS positions, and the collection range is large. However, the collection range of the high-precision map is limited, theoretically, only road data can be collected, and three-dimensional characteristic targets such as buildings are lack of fine description.

And the three-dimensional oblique photography carries a group of cameras forming angles with each other through an unmanned aerial vehicle, 5 cameras or 7 cameras are used as main cameras, one camera vertically irradiates, and the other 4 or 6 cameras irradiate forming a certain angle with each other. From this, unmanned aerial vehicle flies the round along the collection area and can acquire this regional 3D data, acquires the 3D video through the modeling. The 3D scene is closer to the real environment, the data volume collected along the way is large, and corresponding data can still be obtained in an area without a road. However, the cost of three-dimensional oblique photography is relatively high, the problem of limited unmanned aerial vehicle endurance performance in ultra-large range acquisition, communication distance and the like cannot reach ideal expectation, the inconsistent resolution caused by inconsistent flight heights of the unmanned aerial vehicles, the target GPS precision has great influence, the acquired 3D video target position is based on the relative position, and finally, the integration with a unified world coordinate system is difficult.

Based on the above analysis, the three-dimensional oblique photography and the high-precision map acquisition means, the technical forms, and the costs are all different, and for this, the data processing device 12 processes the acquired data in the embodiment of the present application. For example, for different data acquisition means, fusion of data acquired by multiple acquisition means is realized based on the attribute of the target unique space-time feature, integration of multiple acquisition means in different scenes is realized, and a datamation logic is opened.

After fusion of data acquired by various acquisition means is realized based on the unique space-time characteristic attribute of the target, the video can be displayed through the video display equipment 13 by corresponding the position in the video to the real position (namely, the physical position) so as to realize fusion application and display of the video and a BIM (Building Information Modeling), a 3D GIS, a high-precision GIS, an AR (Augmented Reality), a VR (Virtual Reality), a wearable device and the like.

Next, with reference to the above system, a video display method provided by the embodiment of the present application will be described by the following method embodiments.

The embodiment of the present application provides a video presentation method, which may be implemented by the video processing apparatus shown in fig. 1. As shown in fig. 2, the method includes the following steps.

In step 201, a video image is acquired, the video image being derived from images captured by one or more cameras.

The acquired video image can be used as a basis for subsequent video display, and the video image can be obtained through images shot by one or more cameras. For example, an image captured by one camera is acquired, and a video image is obtained. Or acquiring images shot by a plurality of cameras, and splicing the acquired images to obtain a video image.

In step 202, a plurality of reference objects are determined in the video image.

The video picture can be subjected to structuralization processing by an image identification technology aiming at the acquired video image, and a target point in the video image can be automatically identified, wherein the position of the target point can be fixed and can be used for subsequently displaying the target point of the associated information. For example, in the road video image, a marker line, a lane line, a sign, a manhole cover, a camera on the road, a building, or the like in the road video image may be used as the target point. After identifying the plurality of target points, a reference object is selected from the plurality of target points. For example, among the target points, a target point that is easy to be subjected to coordinate conversion is selected as a reference object. The reference object can also be selected in a manner of manual designation, which is not limited in the embodiments of the present application. The number of the selected reference objects may be plural because the reference objects may be used to determine the physical location information of other target points in the video image, and the number of the selected reference objects are not limited in the embodiments of the present application, and at least three are taken as an example.

In step 203, relevant data including the reference object acquired by different data acquisition modes is acquired.

Since the video image is obtained by shooting with the camera, the positions of the respective target points in the video image are relative positions in the video image, and the coordinate system of the video shot by the camera is different from the world coordinate system. The method provided by the embodiment of the application adopts a mode of determining the physical position information of each target point in the video image, so that various types of basic data of the target point are displayed in a relevant manner based on the physical position information of the target point.

In contrast, in the method provided by the embodiment of the application, the reference object is selected from the video image, the physical position information of the reference object in the video image is determined, and then the physical position information of the target point in the video image is determined according to the physical position information of the reference object. When determining the physical position information of the reference object, the method provided by the embodiment of the application acquires the related data including the reference object through different data acquisition modes, so that the physical position information of the reference object in the video image is determined through the related data including the physical position information. The embodiment of the application does not limit different data acquisition modes, and the data acquisition modes can acquire the physical position information of the reference object.

In addition, due to the fact that the types of the reference objects are different, the applicable data acquisition modes are different, the types of data which can be acquired by different data acquisition modes are different, and the requirements for the acquired data are different in different application scenes of video display. For example, a reference object located at a position along a road can be acquired by a high-precision map, and since the positions of the high-precision map acquisition are all absolute GPS positions, that is, physical positions, this type of reference object can be applied to a data acquisition method for a high-precision map. On the other hand, reference objects located in areas where vehicles cannot reach, such as communities and key places, can be acquired by 3D oblique photography.

Therefore, in a possible embodiment of the present application, acquiring relevant data including a reference object acquired by different data acquisition manners includes: determining a target data acquisition mode according to the application scene displayed by the video and the type of the reference object; and acquiring related data including the reference object acquired by the target data acquisition mode. The correlation data including the reference object includes physical position information of the reference object and attribute information of the reference object excluding the physical position information. The attribute information includes, but is not limited to, information such as a length, a width, a size, a distance from another reference object, and an angle of a three-dimensional space of the reference object, which is not limited in the embodiment of the present application.

For example, in a road scene, a data acquisition method of a high-precision map may be used as a target data acquisition method. For example, the method collects relevant data of target points such as a well cover, a street lamp, a garbage can, a green belt, a sign post, a ring island and the like on a road, wherein the relevant data comprises GPS position data, namely physical position information, of each target point and attribute information of each target point. If the street lamp is a reference object, the related data including the street lamp acquired by the data acquisition mode of the high-precision map can be acquired.

For example, in an area where a vehicle cannot conveniently reach such as a building, a community, or an important place, or an area where a three-dimensional space cannot be acquired, the data acquisition method of 3D oblique photography is used as the target data acquisition method to acquire the relevant data including the reference object.

Under the data acquisition mode of 3D oblique photography, before unmanned aerial vehicle takes off the collection data, through demarcating unmanned aerial vehicle's GPS position in advance, the GPS position of a plurality of calibration points in the collection region, later gather the 3D video through 3D oblique photography technique. The GPS positions of other point positions except the pre-calibrated calibration points in the 3D video can be obtained through conversion with the calibration point GPS and the unmanned aerial vehicle GPS, and therefore the 3D video with the position data is obtained. The 3D video includes GPS position information, i.e., physical position information, of the reference object. In addition, attribute information other than the physical position information may be included, such as the length, width, size, distance to another reference object, and angle of a three-dimensional space.

In step 204, the physical position information of the reference object is determined based on the correlation data.

Since the same reference object can appear in the video image or in the data acquired by different data acquisition modes, the attribute information of the same reference object should be consistent. In addition, because the related data includes the physical position information of the reference object, the method provided by the embodiment can match the reference object in the video image with the reference object in the target data acquisition mode according to the attribute information of the reference object. The matching success means that the reference objects are the same, and therefore, the physical position information of the reference objects in the related data can be used as the physical position information of the reference objects in the video image.

In possible embodiments of the present application, the physical location information of the reference object is determined according to the related data, including but not limited to the following two cases:

in the first case: when the target data acquisition mode is a first data acquisition mode, the related data comprises a three-dimensional video obtained according to the first data acquisition mode; fusing the video image and the three-dimensional video to obtain a fused video to be processed; determining the position information of a reference object in a video to be processed; and determining the physical position information of the reference object from the related data according to the position information of the reference object in the video to be processed.

Taking the data acquisition mode that the first data acquisition mode is 3D oblique photography as an example, the relevant data including the reference object acquired by the first data acquisition mode includes the 3D video obtained according to the first data acquisition mode. The video image and the 3D video can be fused through algorithms such as texture mapping and the like, and a fused video to be processed is obtained.

Under the condition, the fused to-be-processed video realizes the correspondence between the 3D video and the video image, and the fused part of the fused to-be-processed video can display the real-time video of the camera. In addition, for a fixed-position object and some regions that do not change over time, the fused content corresponds to the content in the 3D video. And after the fused video to be processed is obtained, determining the position information of the reference object in the video to be processed, and determining the physical position information of the reference object from the related data according to the position information of the reference object in the video to be processed. For example, the fused to-be-processed video includes the reference object B, and the reference object B is also in the 3D video, so that the physical position information of the reference object B can be determined from the related data.

In the second case: and matching and determining the physical position information of the reference object from the related data according to the attribute information of the reference object.

Taking the data acquisition mode that the target data acquisition mode is the high-precision map as an example, the attribute information of the reference object acquired by the data acquisition mode of the high-precision map is matched with the attribute information of the reference object in the video image, and if the matching is successful, the two reference objects are considered to be the same reference object. And the related data of the reference object acquired by the data acquisition mode of the high-precision map comprises the physical position information of the reference object, so the physical position information of the reference object in the related data is corresponding to the video image to obtain the physical position information of the reference object in the video image.

For example, the attribute information of the reference object a acquired by the data acquisition method of the high-precision map is matched with the attribute information of the reference object a 'in the video image, and if the matching is successful, the reference object a and the reference object a' are considered to be the same reference object. And the related data of the reference object A acquired by the data acquisition mode of the high-precision map comprises the physical position information of the reference object A, so the physical position information of the reference object A is corresponded to the video image to obtain the physical position information of the reference object A' in the video image.

It should be noted that, if the target acquisition mode is a data acquisition mode of 3D oblique photography, the mode of matching and determining the physical position information of the reference object from the related data according to the attribute information of the reference object is consistent with the mode principle of the high-precision map, which can be referred to above and is not described here again.

In step 205, the physical position information of the target point in the video image is determined based on the physical position information of the reference object and the position information of the reference object in the video image.

The position information of the reference object in the video image is determined by the coordinate system of the video, the physical position information of the reference object is determined by the world coordinate system, and the mapping relation between the video coordinate system and the world coordinate system can be established based on the physical position information of the reference object and the position information of the reference object in the video image, so that the position mapping model can be obtained. Thereafter, physical location information of the target point in the video image may be determined based on the mapping. In a possible embodiment of the present application, determining the physical position information of the target point in the video image according to the physical position information of the reference object and the position information of the reference object in the video image includes: establishing a position mapping model according to the physical position information of the reference object and the position information of the reference object in the video image; physical location information of a target point in the video image is determined based on the location mapping model.

It should be understood that the video images may be images taken by a camera, and that if the camera is fixed in position, there may be differences between images taken for different times. For example, taking the road image as an example, 2 vehicles may be included in the image taken at 8 am, and 5 vehicles may be included in the image taken at 9 am. However, although the images taken at different times are different, a reference object fixed on the road may be taken in each image. Therefore, the method provided by the embodiment of the application can be applied to video images with continuously changing contents, namely basic data of target points with unfixed positions in the video images can be displayed. By applying the method provided by the embodiment of the application, the physical position information of each target point in the image can be determined based on the physical position information of the reference object.

In step 206, the underlying data showing the target point is correlated based on the physical location information of the target point in the video image.

Since the basic data of the target point is borne in the respective service systems, after the physical position information of the target point in the video image is acquired, the various basic data of the target point can be associated through the physical position information, and thus the basic data of the target point can be associated and displayed during video display. In a possible implementation manner of the present application, the basic data of the target point includes, but is not limited to, business data, AR data, BIM data, map data, and the like of the target point, and the example of the present application does not limit the basic data of the target point.

In a possible embodiment of the present application, associating basic data showing a target point based on physical location information of the target point in a video image includes: displaying a video image in a first display area, and displaying target point information in the video image; and displaying the basic data of the target point in the second display area.

The target point information includes, but is not limited to, content corresponding to the target point, description information of the target point, or an identifier of the target point, which is not limited in this embodiment. For example, taking the display interface shown in fig. 3 as an example, the video image displayed in the first display area on the left side includes 3 target points, which are a target point a, a target point B, and a target point C. The second display area on the right shows the base data of these 3 target points. The types of the basic data of the 3 target points are different, the basic data of the target point A is underground pipe network BIM information, the basic data of the target point B is AR data acquired based on an image shot at the target point B, and the basic data of the target point C is map information including the target point C. The content in the interface shown in fig. 3, in combination with the actual scene, may be as shown in fig. 4.

If the number of the target points is large or the basic data of the target points is large, the basic data of all the target points may not be displayed in the second display area. In a possible embodiment of the present application, when the target point information is displayed in the video image, a control may be set for each target point, the control of which interaction point is detected to be selected is detected, and then the basic information of the target point is displayed in the second display area. Or after detecting that the control of which interaction point is selected, displaying the basic data identification of the target point, wherein each basic data identification corresponds to one type of basic data. And displaying the basic data corresponding to the selected basic data identification in a second display area.

In addition to the above display modes, because there is a spatio-temporal relationship between the basic data of some target points, in a possible embodiment of the present application, the displaying the basic data of the target points in the second display area includes: and if the number of the basic data of the target point is multiple, performing space-time superposition on the multiple basic data of the target point, and displaying the superposed basic data of the target point in a second display area.

For example, the destination point is a building, and the basic data that can be associated includes 3D, BIM information of the building based on the physical location information of the destination point. In addition, each floor in the building is provided with a video shooting device, so the basic data of the target point can also comprise monitoring videos of each floor of the building. The 3D, BIM information of the building has a spatio-temporal relationship with the surveillance video of each floor in the building so that the underlying data can be overlaid spatio-temporally. For example, when the 3D, BIM information of the building is presented in the second presentation area, the monitoring videos of the floors are displayed at the corresponding positions of each floor beside the building. In addition, the basic data of the target point may also include service data, for example, the service data includes the identity information of a certain resident on the layer 9 of the building, and the identity information of the certain resident is displayed in an overlapping manner while the building is displayed.

In a possible embodiment of the present application, the displaying the basic data of the target point in the second display area includes: and if the number of the basic data of the target point is multiple, determining the level relation among the multiple basic data of the target point, and displaying the basic data of the target point in a grading way according to the level relation among the multiple basic data in the second display area.

Still taking the above-mentioned destination point as an example of the building, the building has a hierarchical relationship with each floor, only the 3D, BIM information of the building is displayed in the second display area, the indication point of each floor is displayed, and the monitoring videos of the floors in the building are further displayed in a progressive manner based on the indication point, thereby realizing the hierarchical display of the base data of the destination point according to the hierarchical relationship among the plurality of base data.

According to the method provided by the embodiment of the application, after the physical position information of the reference object is determined through the related data including the reference object acquired by different data acquisition modes, the physical position information of the target point in the video image is determined based on the physical position information of the reference object and the position information of the reference object in the video image, so that the basic data of the target point is displayed in a correlation mode through the physical position information, the integration of various data acquisition modes is realized, the data and the application of different types can be intercommunicated, the video display modes are enriched, and the user experience can be further improved.

Based on the same technical concept, referring to fig. 5, an embodiment of the present application provides a video display apparatus, including:

a first obtaining module 501, configured to obtain video images, where the video images are obtained from images captured by one or more cameras;

a first determining module 502, configured to determine a plurality of reference objects in the video image;

a second obtaining module 503, configured to obtain relevant data including a reference object, which are collected in different data collection manners;

a second determining module 504, configured to determine physical location information of the reference object according to the relevant data;

a third determining module 505, configured to determine physical location information of the target point in the video image according to the physical location information of the reference object and the location information of the reference object in the video image;

and the display module 506 is configured to associate basic data for displaying the target point based on the physical location information of the target point in the video image.

In a possible implementation manner of the present application, the second obtaining module 503 is configured to determine a target data acquisition manner according to an application scene displayed by a video and a type of a reference object; and acquiring related data including the reference object acquired by the target data acquisition mode.

In a possible implementation manner of the present application, the second obtaining module 503 is configured to, when the target data collection manner is a first data collection manner, include, in the related data, a three-dimensional video obtained according to the first data collection manner; fusing the video image and the three-dimensional video to obtain a fused video to be processed; determining the position information of the reference object in the video to be processed; and determining the physical position information of the reference object from the related data according to the position information of the reference object in the video to be processed.

In a possible implementation manner of the present application, the second obtaining module 503 is configured to match and determine the physical location information of the reference object from the related data according to the attribute information of the reference object.

In a possible implementation manner of the present application, the display module 506 is configured to display a video image in a first display area, and display target point information in the video image; and displaying the basic data of the target point in the second display area.

In a possible implementation manner of the present application, the displaying module 506 is configured to perform space-time superposition on multiple pieces of basic data of the target point if there are multiple pieces of basic data of the target point, and display the superposed basic data of the target point in the second display area.

In a possible implementation manner of the present application, the displaying module 506 is configured to determine a level relationship between a plurality of pieces of basic data of the target point if the number of pieces of basic data of the target point is multiple, and hierarchically display the basic data of the target point in the second displaying area according to the level relationship between the plurality of pieces of basic data.

In a possible implementation manner of the present application, the third determining module 505 is configured to establish a position mapping model according to the physical position information of the reference object and the position information of the reference object in the video image; physical location information of a target point in the video image is determined based on the location mapping model.

It should be noted that, when the apparatus provided in the foregoing embodiment implements the functions thereof, only the division of the functional modules is illustrated, and in practical applications, the functions may be distributed by different functional modules according to needs, that is, the internal structure of the apparatus may be divided into different functional modules to implement all or part of the functions described above. In addition, the apparatus and method embodiments provided by the above embodiments belong to the same concept, and specific implementation processes thereof are described in the method embodiments for details, which are not described herein again.

Fig. 6 is a schematic structural diagram of a video display apparatus according to an embodiment of the present application. The device may be a terminal, and may be, for example: a smartphone, a tablet, a laptop, or a desktop computer. A terminal may also be referred to by other names such as user equipment, portable terminal, laptop terminal, desktop terminal, etc.

Generally, a terminal includes: a processor 601 and a memory 602.

The processor 601 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so on. The processor 601 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). The processor 601 may also include a main processor and a coprocessor, where the main processor is a processor for processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 601 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and drawing the content required to be displayed on the display screen. In some embodiments, processor 601 may also include an AI (Artificial Intelligence) processor for processing computational operations related to machine learning.

The memory 602 may include one or more computer-readable storage media, which may be non-transitory. The memory 602 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 602 is used to store at least one instruction for execution by processor 601 to implement the video presentation method provided by the method embodiments of the present application.

In some embodiments, the terminal may further include: a peripheral interface 603 and at least one peripheral. The processor 601, memory 602, and peripheral interface 603 may be connected by buses or signal lines. Various peripheral devices may be connected to the peripheral interface 603 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of a radio frequency circuit 604, a touch screen display 605, a camera 606, an audio circuit 607, a positioning component 608, and a power supply 609.

The peripheral interface 603 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 601 and the memory 602. In some embodiments, the processor 601, memory 602, and peripheral interface 603 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 601, the memory 602, and the peripheral interface 603 may be implemented on a separate chip or circuit board, which is not limited in this embodiment.

The Radio Frequency circuit 604 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 604 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 604 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 604 comprises: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuitry 604 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: metropolitan area networks, various generation mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the rf circuit 604 may further include NFC (Near Field Communication) related circuits, which are not limited in this application.

The display 605 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display screen 605 is a touch display screen, the display screen 605 also has the ability to capture touch signals on or over the surface of the display screen 605. The touch signal may be input to the processor 601 as a control signal for processing. At this point, the display 605 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 605 may be one, providing the front panel of the terminal; in other embodiments, the display 605 may be at least two, respectively disposed on different surfaces of the terminal or in a folding design; in still other embodiments, the display 605 may be a flexible display disposed on a curved surface or on a folded surface of the terminal. Even more, the display 605 may be arranged in a non-rectangular irregular pattern, i.e., a shaped screen. The display 605 may be made of LCD (Liquid crystal display), OLED (Organic Light-Emitting Diode), and the like.

The camera assembly 606 is used to capture images or video. Optionally, camera assembly 606 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 606 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.

Audio circuitry 607 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 601 for processing or inputting the electric signals to the radio frequency circuit 604 to realize voice communication. For the purpose of stereo sound collection or noise reduction, a plurality of microphones can be arranged at different parts of the terminal respectively. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 601 or the radio frequency circuit 604 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, audio circuitry 607 may also include a headphone jack.

The positioning component 608 is used to locate the current geographic Location of the terminal to implement navigation or LBS (Location based service). The positioning component 608 can be a positioning component based on the united states GPS (Global positioning system), the chinese beidou system, the russian graves system, or the european union's galileo system.

The power supply 609 is used to supply power to various components in the terminal. The power supply 609 may be ac, dc, disposable or rechargeable. When the power supply 609 includes a rechargeable battery, the rechargeable battery may support wired or wireless charging. The rechargeable battery may also be used to support fast charge technology.

In some embodiments, the terminal also includes one or more sensors 610. The one or more sensors 610 include, but are not limited to: acceleration sensor 611, gyro sensor 612, pressure sensor 613, fingerprint sensor 614, optical sensor 615, and proximity sensor 616.

The acceleration sensor 611 may detect the magnitude of acceleration on three coordinate axes of a coordinate system established with the terminal. For example, the acceleration sensor 611 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 601 may control the touch screen display 605 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 611. The acceleration sensor 611 may also be used for acquisition of motion data of a game or a user.

The gyro sensor 612 may detect a body direction and a rotation angle of the terminal, and the gyro sensor 66 may cooperate with the acceleration sensor 611 to acquire a 3D motion of the user on the terminal. The processor 601 may implement the following functions according to the data collected by the gyro sensor 612: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.

The pressure sensor 613 may be disposed on a side frame of the terminal and/or on a lower layer of the touch display screen 605. When the pressure sensor 613 is disposed on the side frame of the terminal, a user's holding signal to the terminal can be detected, and the processor 601 performs left-right hand recognition or shortcut operation according to the holding signal collected by the pressure sensor 613. When the pressure sensor 613 is disposed at the lower layer of the touch display screen 605, the processor 601 controls the operability control on the UI interface according to the pressure operation of the user on the touch display screen 605. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.

The fingerprint sensor 614 is used for collecting a fingerprint of a user, and the processor 601 identifies the identity of the user according to the fingerprint collected by the fingerprint sensor 614, or the fingerprint sensor 614 identifies the identity of the user according to the collected fingerprint. Upon identifying that the user's identity is a trusted identity, the processor 601 authorizes the user to perform relevant sensitive operations including unlocking the screen, viewing encrypted information, downloading software, paying, and changing settings, etc. The fingerprint sensor 614 may be provided on the front, back or side of the terminal. When a physical button or vendor Logo is provided on the terminal, the fingerprint sensor 614 may be integrated with the physical button or vendor Logo.

The optical sensor 66 is used to collect the ambient light intensity. In one embodiment, processor 601 may control the display brightness of touch display 605 based on the ambient light intensity collected by optical sensor 66. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 605 is increased; when the ambient light intensity is low, the display brightness of the touch display screen 605 is turned down. In another embodiment, processor 601 may also dynamically adjust the shooting parameters of camera assembly 606 based on the ambient light intensity collected by optical sensor 66.

A proximity sensor 616, also known as a distance sensor, is typically provided on the front panel of the terminal. The proximity sensor 616 is used to collect the distance between the user and the front face of the terminal. In one embodiment, when the proximity sensor 616 detects that the distance between the user and the front face of the terminal gradually decreases, the processor 601 controls the touch display 605 to switch from the bright screen state to the dark screen state; when the proximity sensor 616 detects that the distance between the user and the front face of the terminal gradually becomes larger, the processor 601 controls the touch display 605 to switch from the rest screen state to the bright screen state.

Those skilled in the art will appreciate that the configuration shown in fig. 6 is not intended to be limiting, and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components may be used.

In an example embodiment, a computer device is also provided that includes a processor and a memory having at least one instruction stored therein. The at least one instruction is configured to be executed by one or more processors to implement any of the video presentation methods described above.

In an exemplary embodiment, there is also provided a computer-readable storage medium having stored therein at least one instruction which, when executed by a processor of a computer device, implements any of the video presentation methods described above.

Alternatively, the computer-readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.

It should be understood that reference to "a plurality" herein means two or more. "and/or" describes the association relationship of the associated objects, meaning that there may be three relationships, e.g., a and/or B, which may mean: a exists alone, A and B exist simultaneously, and B exists alone. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.

The above-mentioned serial numbers of the embodiments of the present application are merely for description and do not represent the merits of the embodiments.

The above description is only exemplary of the present application and should not be taken as limiting the present application, and any modifications, equivalents, improvements and the like that are made within the spirit and principle of the present application should be included in the protection scope of the present application.

19页详细技术资料下载
上一篇:一种医用注射器针头装配设备
下一篇:图像处理方法及装置

网友询问留言

已有0条留言

还没有人留言评论。精彩留言会获得点赞!

精彩留言,会给你点赞!