Quick coding method based on multi-feature analysis of coding unit

文档序号:1589962 发布日期:2020-01-03 浏览:29次 中文

阅读说明:本技术 基于编码单元多特征分析的快速编码方法 (Quick coding method based on multi-feature analysis of coding unit ) 是由 刘欣刚 朱超 吴立帅 汪卫彬 代成 李辰琦 于 2019-08-31 设计创作,主要内容包括:本发明公开了一种基于编码单元特征分析的快速编码方法,属于视频编码技术领域。本发明包括:编码单元纹理、边缘、结构特征提取;将提取的多特征输入至SVM分类器中离线学习,得到各深度下SVM的分类模型;通过提取的特征寻找当前关联度最大的编码单元进行深度0预判决;根据特征将编码单元划分为简单、中等和复杂三种情况;在各深度下对划分为简单的编码单元终止其深度判决,对划分为复杂的编码单元跳过当前深度进行下一深度判决,对划分为中等的编码单元则按照原流程进行当前深度的判决。本发明基于视频图像复杂度进行编码单元快速划分的方法可以大大减少编码单元在深度判决过程中的计算复杂度,在保证视频质量的前提下节约了编码时间。(The invention discloses a rapid coding method based on coding unit characteristic analysis, and belongs to the technical field of video coding. The invention comprises the following steps: extracting texture, edge and structural features of the coding unit; inputting the extracted multiple features into an SVM classifier for off-line learning to obtain classification models of the SVM at various depths; searching a coding unit with the maximum current association degree through the extracted features to perform depth 0 pre-judgment; dividing the coding unit into three cases of simple, medium and complex according to the characteristics; and stopping the depth judgment of the simply divided coding units under each depth, skipping the current depth of the complicated coding units for the next depth judgment, and judging the current depth of the medium-sized coding units according to the original flow. The method for rapidly dividing the coding units based on the complexity of the video images can greatly reduce the calculation complexity of the coding units in the depth judgment process, and saves the coding time on the premise of ensuring the video quality.)

1. The quick coding method based on the multi-feature analysis of the coding unit is characterized by comprising the following steps of:

s1: extracting features of the coding units under all depths to obtain texture, edge and structural features of the coding units under all depths;

s2: and (3) offline learning of classifier features:

inputting the multi-features extracted in the step S1 at different depths into an SVM classifier for off-line learning to obtain SVM classification models at each depth, wherein each SVM classification model is used for determining the complexity classification of the coding units at each depth;

s3: determining the CTU with the maximum neighborhood relevance based on the multi-feature relevance;

s4: and (3) judging the maximum association CTU depth 0:

judging the current CTU in advance according to the depth of the CTU with the maximum neighborhood relevance; on the premise that the current coding depth is 0, if the final partition depth of the CTU with the maximum association degree with the current coding unit is 0, terminating the quad-tree partition of the current CTU; otherwise, continuing to execute step S5;

s5: and (3) complexity prediction:

when the CU at each depth is coded, inputting the extracted texture, edge and structure characteristics into an SVM classification model at the corresponding depth, and dividing the complexity of a coding unit into three conditions of simplicity, medium and complexity according to an output result;

s6: and executing corresponding division judgment according to the classification result of the complexity prediction:

if the coding unit image is classified as simple, terminating the quadtree division of the current depth;

if the coding unit image is classified as complex, directly performing quadtree division on the current coding unit at the current depth, and performing complexity prediction and division judgment of the next depth;

if the coding unit image is classified as medium, the coding is performed according to the HEVC standard.

2. The method according to claim 1, wherein in step S1, the texture, edge and texture features are specifically:

texture characteristics: the pixel neighborhood mean square error of each coding unit;

edge characteristics: pixel Sobel gradient values of each coding unit;

the structure is characterized in that: the variance of the prediction residual variance of each coding unit from its four sub-blocks.

3. The method of claim 1, wherein the step S2 includes the steps of:

s21: inputting the texture, edge and structure characteristics of all coding units and corresponding depth information into an SVM classifier for off-line training;

s22: for the training model under each depth, obtaining a classifier model under each depth based on the trained classifier parameters;

s23: and respectively determining the optimal complexity prediction parameters of the coding units under each depth according to the accuracy of the offline training classification.

4. The method according to claim 1, wherein in step S3, the step of determining the CTU with the largest neighborhood relevance includes:

based on texture features, edge features and structural features of the coding unit, determining the CTU with the maximum neighborhood relevance degree from the adjacent space-time domain of the current coding unit:

calculating the difference degree R between the current coding unit and the CTU of the adjacent time-space domaincur-i

Figure FDA0002187279670000021

Wherein A iscur、BcurAnd CcurRespectively representing texture, edge and texture features of the current coding unit, Ai、BiAnd CiRespectively representing texture features, edge features and structural features of CTUs of adjacent time-space domains of a current coding unit;

taking the difference degree Rcur-iAnd the CTU of the minimum adjacent time-space domain is taken as the CTU with the maximum neighborhood relevance of the current coding unit.

5. The method of claim 1, wherein the step S5 includes the steps of:

s51: inputting texture, edge and structural characteristics of the coding unit at the current depth into an SVM classifier for complexity calculation;

s52: if the output result of the prediction coding unit of the SVM classifier which is not divided is less than 0, the image complexity of the coding unit is classified as simple;

s53: if the output result of the direct division of the prediction coding unit classified by the SVM is greater than 0, the image complexity of the coding unit is classified as complex;

s54: if the SVM classification outputs other results, the image complexity of the coding unit is classified as medium.

Technical Field

The invention relates to the technical field of video coding, in particular to a high-efficiency video coding method for coding unit depth division based on video image complexity and multiple characteristics.

Background

The video is composed of a frame of image, but the original video cannot meet daily storage and transmission requirements due to the huge data volume of the original video, so that the original video needs to be compressed. The international telecommunication union (ITU-T) and the international organization for standardization/international electrotechnical commission (ISO/IEC) were again in cooperation with each other to establish the joint working group for video coding (JCT-VC), and a new generation of video coding standard, the high efficiency video coding standard (HEVC/h.265), was published in 2013. HEVC continues to use the hybrid coding framework of the previous generation h.264, introduces a variety of advanced coding techniques, and for the same video sequence, on the premise of ensuring the unchanged coding quality, the HEVC standard saves the coding bit rate by 50% compared with the h.264 standard. Although HEVC has a great improvement in coding efficiency, its computational complexity is very large, and its coding time is almost twice as long as that of the h.264 standard, which also largely hinders the popularization and application of the HEVC standard in daily life.

HEVC employs a flexible block partitioning approach, including Coding Units (CU), Prediction Units (PU), and Transform Units (TU). In the CU layer HEVC, a coded image is divided into four pixel sizes, 64x64, 32x32, 16x16 and 8x8, by means of quadtree recursion, and is represented by four depths of 0, 1, 2 and 3, respectively, where a coding unit of 64x64 is called a Coding Tree Unit (CTU). The final partitioning combination of the CU is determined by cost comparison from maximum depth 3 to minimum depth 0 down and up. Since the partitioning and comparing methods for the CUs are computationally complex, how to reduce unnecessary computation becomes a key to speed up the HEVC coding time.

The complexity characteristic of the image and the final division result are often connected. Generally, areas with simple image texture are generally coded by using larger coding blocks, while areas with complex image texture are coded by using more small blocks. Aiming at the problem of redundancy of the recursive computation of the quadtree at the CU layer, most of the traditional methods fit the two classification curves of the coding unit based on single statistical characteristics, and set a threshold value according to the curve result to perform division judgment of the CU layer. Since the complexity of the image cannot be accurately measured by a single feature and the adaptive requirement cannot be met by only using a single threshold, the prediction efficiency of the coding is very low.

Disclosure of Invention

The invention aims to: the invention provides a rapid coding method based on coding unit multi-feature analysis, aiming at the technical problems of high computational complexity, single selection feature and single division threshold value of the traditional method in the existing high-efficiency video coding technology.

The invention comprises the following steps:

s1: extracting features of the coding units under all depths to obtain texture, edge and structural features of the coding units under all depths;

s2: and (3) offline learning of classifier features:

inputting the multi-features extracted in the step S1 at different depths into an SVM classifier for off-line learning to obtain SVM classification models at each depth, wherein each SVM classification model is used for determining the complexity classification of the coding units at each depth;

s3: determining the CTU with the maximum neighborhood relevance based on the multi-feature relevance;

s4: and (3) judging the maximum association CTU depth 0:

judging the current CTU in advance according to the depth of the CTU with the maximum neighborhood relevance; on the premise that the current coding depth is 0, if the final partition depth of the CTU with the maximum association degree with the current coding unit is 0, terminating the quad-tree partition of the current CTU; otherwise, continuing to execute step S5;

s5: and (3) complexity prediction:

when the CU at each depth is coded, inputting the extracted texture, edge and structure characteristics into an SVM classification model at the corresponding depth, and dividing the complexity of a coding unit into three conditions of simplicity, medium and complexity according to an output result;

s6: and executing corresponding division judgment according to the classification result of the complexity prediction:

if the coding unit image is classified as simple, terminating the quadtree division of the current depth;

if the coding unit image is classified as complex, directly performing current quadtree division (namely, directly performing quadtree division on the current coding unit at the current depth), and performing complexity prediction and division judgment of the next depth;

if the coding unit image is classified as medium, the coding is performed according to the HEVC standard.

Further, in step S1, the texture, edge, and structural features are specifically:

texture characteristics: extracting the mean square error of a pixel neighborhood of each coding unit as a characteristic for measuring the Texture Complexity (TC) of the image;

edge characteristics: extracting a pixel Sobel gradient value of each coding unit as a characteristic for measuring the Edge Complexity (EC) of an image;

the structure is characterized in that: the variance of each coding unit and the variance of the prediction residuals of its four sub-blocks is extracted as a feature to measure the Complexity (SC) of the picture Structure.

Further, the step S2 includes the following steps:

s21: inputting TC, EC and SC of all coding units and corresponding depth information into an SVM classifier for off-line training;

s22: for the training model under each depth, obtaining a threshold value corresponding to the complexity division condition of the coding unit based on the trained classifier parameters, namely corresponding to the classifier model under each depth;

s23: and respectively determining the optimal complexity prediction parameters of the coding units under each depth according to the accuracy of the offline training classification.

Further, the step S5 includes the following steps:

s51: inputting TC, EC and SC of the coding unit under the current depth into an SVM classifier for complexity calculation;

s52: if the output result of the prediction coding unit of the SVM classifier which is not divided is less than 0, the image complexity of the coding unit is classified as simple;

s53: if the output result of the direct division of the prediction coding unit classified by the SVM is greater than 0, the image complexity of the coding unit is classified as complex;

s54: if the SVM classification outputs other results, the image complexity of the coding unit is classified as medium.

In summary, due to the adoption of the technical scheme, the invention has the beneficial effects that: the invention extracts a plurality of characteristics of the coded image, and more accurately measures the complexity condition of the coded image from a plurality of angles. And depth 0 judgment is terminated in advance according to the relevance of the neighborhood images, and the depth 0 judgment is accelerated by utilizing the time-space domain and the depth information of the coded images, so that the coding time is greatly accelerated. A multi-classifier prediction model is obtained by a characteristic off-line learning method, and prediction of multiple classifiers and multiple thresholds is more accurate and flexible.

Drawings

FIG. 1: the invention relates to a quick coding flow chart of multi-feature analysis of a coding unit.

FIG. 2: the invention extracts the horizontal and vertical Sobel gradient template schematic diagrams of the image edge complexity.

FIG. 3: the coding unit quadtree is divided into schematic diagrams.

Detailed Description

In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be described in further detail with reference to the following embodiments and accompanying drawings.

The invention discloses a rapid coding method based on multi-feature analysis of a coding unit, which is a High Efficiency Video Coding (HEVC) method for accelerating depth division of the coding unit based on complexity characteristics of a video image. Firstly, extracting coding unit characteristics under each depth, including texture, edge and structure characteristics of a coding unit; performing offline learning on the classifier features, inputting the extracted multiple features into an SVM classifier for offline learning, and obtaining classification models of SVM at various depths; analyzing the association degree of adjacent coding units in a time-space domain, and searching the coding unit with the maximum association degree at present through the extracted features to perform depth 0 pre-judgment; judging the complexity, namely dividing the coding units into three conditions of simple, medium and complex according to the characteristics; and (3) coding unit depth judgment, namely stopping the depth judgment of the coding units divided into simple depths under each depth, skipping the current depth of the coding units divided into complex depths to judge the next depth, and judging the current depth of the coding units divided into medium depths according to the original flow. The method for rapidly dividing the coding units based on the complexity of the video images can greatly reduce the calculation complexity of the coding units in the depth judgment process, and saves the coding time on the premise of ensuring the video quality.

Referring to fig. 1, the specific implementation process is as follows:

s1: and extracting the features of the coding units at each depth.

S11: extracting the pixel neighborhood mean square error of each coding unit as a characteristic for measuring the Texture Complexity (TC), namely the TC expression is shown as formula (1) and formula (2):

Figure BDA0002187279680000041

Figure BDA0002187279680000042

wherein, N represents the number of the current encoding block pixel points, f (i, j) represents the image pixel value of the current encoding unit at the coordinate (i, j), and

Figure BDA0002187279680000043

then, the pixel mean value of eight neighborhood pixels of the current image block at the pixel point with the coordinate (i, j) is obtained;

s12: extracting a pixel Sobel gradient value of each coding unit as a characteristic for measuring image Edge Complexity (EC), wherein EC expressions are shown as formula (3) and formula (4):

Figure BDA0002187279680000044

Figure BDA0002187279680000045

in the formula, N represents the number of the current coding block pixel points,

Figure BDA0002187279680000046

represents the Sobel gradient mean, E, of the current coding unit at coordinate (i, j)hor(i, j) and EvecAnd (i, j) respectively refer to a horizontal Sobel gradient value and a vertical Sobel gradient value of the current coding unit at the coordinate (i, j). In this embodiment, a horizontal Sobel gradient template (S)hor) And vertical Sobel gradient template (S)vec) As shown in fig. 2;

s13: extracting the variance of the prediction residual variance of each coding unit and its four sub-blocks as the characteristic for measuring the image Structure Complexity (SC), and taking the variance of the prediction residual variance of each coding unit and its four sub-blocks as SCom of the coding unit, as shown in formula (5):

Figure BDA0002187279680000047

in the formula, variDenotes the prediction residual variance of the sub-block numbered i,

Figure BDA0002187279680000048

representing the mean of the variance of the prediction residuals;

s2: the classifier feature offline learning method comprises the following steps of inputting extracted multiple features into an SVM classifier for offline learning to obtain classification models of SVM under various depths, and comprises the following specific steps:

s21: inputting TC, EC and SC of all coding units and corresponding depth information into an SVM classifier for off-line training;

s22: for the training model at each depth, selecting appropriate classifier parameters to obtain a threshold corresponding to the division condition of the coding unit, that is, the classifier model at each depth can be simply represented by formula (6):

Figure BDA0002187279680000051

in the formula, ω and b are parameters obtained after the classifier training, x is the input characteristics of TC, EC, SC, etc., sign is the activation function: when the output value is-1, the predicted coding unit is not divided, when the output value is 1, the predicted coding unit is continuously divided, and when the output value is 0, the predicted coding unit is uncertain;

s23: according to the accuracy of the offline training classification, respectively determining the optimal prediction parameters of the coding units under each depth, which is shown in formula (7):

Figure BDA0002187279680000052

in the formula, ω1、b1Respectively, parameters of the prediction coding unit not being partitioned, ω2、b2Respectively, parameters for direct partitioning of the prediction coding unit;

s3: determining the CTU with the maximum neighborhood relevance based on multi-feature relevance analysis;

s31: performing image feature relevance analysis on CTUs adjacent to a time-space domain of a current coding unit according to TC, EC and SC, and determining the CTU with the maximum relevance (also called as maximum relevance CTU), wherein a calculation expression of the maximum relevance is shown as a formula (8):

wherein R represents the maximum correlation degree, TCcur、ECcur、SCcurTexture complexity, edge complexity and structure complexity of the current coding unit image respectively, nei is the CTU of the adjacent time-space domain, i is the co-located CTU, right left CTU, upper left CTU and right upper CTU of the previous frame respectivelyA square CTU and an upper right CTU;

and obtaining the maximum association CTU of the current coding unit based on the CTU corresponding to the maximum association degree.

S4: judging the depth 0 of the maximum correlation CTU, and judging the current CTU in advance according to the depth of the neighborhood maximum correlation CTU; on the premise that the current coding depth is 0, if the final partition depth of the CTU with the maximum association degree with the current coding unit is 0, the quadtree partition of the current CTU is terminated, and a schematic diagram of the quadtree partition of the coding unit is shown in fig. 3. Otherwise, continuing to execute the following steps;

s5: and (3) complexity prediction, when the CU at each depth is coded, inputting the extracted features into a classification model, dividing a coding unit into three conditions of simplicity, medium and complexity according to an output result, and performing corresponding processing:

s51: inputting TC, EC and SC of a coding unit under the current depth into a classifier to carry out complexity calculation;

s52: if y1If the output result is less than 0, the image complexity of the coding unit is classified as simple;

s53: if y2If the output result is greater than 0, the image complexity of the coding unit is classified as complex;

s54: if other results are output, the image complexity of the coding unit is classified as medium;

s6: and (4) processing the classification result, and executing corresponding division judgment according to the classification result:

s61: if the coding unit image is classified as simple, terminating the quadtree division of the current depth;

s62: if the coding unit image is classified as complex, directly performing current quadtree division and performing next depth judgment;

s63: if the coding unit image is classified as medium, coding according to the original HEVC standard;

the rapid coding method extracts a plurality of characteristics of the image, and judges the division condition of each depth of the coding unit by using a classifier trained by characteristic off-line. The above method reduces the number of comparisons traversing the CU at all depths, thereby greatly reducing the encoding complexity. Through analysis of experimental simulation results, the coding time of the invention can be reduced by 52.97% on the premise of equivalent acceptable quality loss, while the current similar method is 46.5%. Therefore, the invention effectively reduces the coding complexity on the premise that the coding performance is hardly influenced.

While the invention has been described with reference to specific embodiments, any feature disclosed in this specification may be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise; all of the disclosed features, or all of the method or process steps, may be combined in any combination, except mutually exclusive features and/or steps.

10页详细技术资料下载
上一篇:一种医用注射器针头装配设备
下一篇:图像的编码、解码方法及装置、电子设备及系统

网友询问留言

已有0条留言

还没有人留言评论。精彩留言会获得点赞!

精彩留言,会给你点赞!

技术分类