Next Article in Journal
Layerwise Adversarial Learning for Image Steganography
Previous Article in Journal
Elicitation of Content Layout Preferences in Virtual 3D Spaces Based on a Free Layout Creation Task
Previous Article in Special Issue
Artificial Neural Network for Photonic Crystal Band Structure Prediction in Different Geometric Parameters and Refractive Indexes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Visual Perception Based Intra Coding Algorithm for H.266/VVC

1
Department of Electrical Engineering, National Dong Hwa University, Hualien 974301, Taiwan
2
Department of Electrical Engineering, National Taiwan Normal University, Taipei 106308, Taiwan
3
Department of Electrical Engineering, National Sun Yat-sen University, Kaohsiung 804201, Taiwan
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(9), 2079; https://doi.org/10.3390/electronics12092079
Submission received: 28 February 2023 / Revised: 22 April 2023 / Accepted: 25 April 2023 / Published: 1 May 2023

Abstract

:
The latest international video coding standard, H.266/Versatile Video Coding (VVC), supports high-definition videos, with resolutions from 4 K to 8 K or even larger. It offers a higher compression ratio than its predecessor, H.265/High Efficiency Video Coding (HEVC). In addition to the quadtree partition structure of H.265/HEVC, the nested multi-type tree (MTT) structure of H.266/VVC provides more diverse splits through binary and ternary trees. It also includes many new coding tools, which tremendously increases the encoding complexity. This paper proposes a fast intra coding algorithm for H.266/VVC based on visual perception analysis. The algorithm applies the factor of average background luminance for just-noticeable-distortion to identify the visually distinguishable (VD) pixels within a coding unit (CU). We propose calculating the variances of the numbers of VD pixels in various MTT splits of a CU. Intra sub-partitions and matrix weighted intra prediction are turned off conditionally based on the variance of the four variances for MTT splits and a thresholding criterion. The fast horizontal/vertical splitting decisions for binary and ternary trees are proposed by utilizing random forest classifiers of machine learning techniques, which use the information of VD pixels and the quantization parameter. Experimental results show that the proposed algorithm achieves around 47.26% encoding time reduction with a Bjøntegaard Delta Bitrate (BDBR) of 1.535% on average under the All Intra configuration. Overall, this algorithm can significantly speed up H.266/VVC intra coding and outperform previous studies.

1. Introduction

The applications of multimedia and network communication technologies have grown rapidly. High-definition videos have increased the amount of data transmission required for video communication. Therefore, video compression to reduce the amount of video transmission and data storage is essential. H.266/Versatile Video Coding (VVC) [1,2,3] was developed to improve the performance of video compression and can support ultra-high-definition videos, with resolutions from 4 K to 8 K or even larger. H.266/VVC, the latest generation of international video coding standards, was completed in July 2020 [4,5].
In H.265/High Efficiency Video Coding (HEVC), the previous generation video coding standard, the maximum allowed size of a coding tree unit (CTU) is 64 × 64. A picture is divided into non-overlapping CTUs. Then, a recursive quadtree (QT) partition can be applied to each CTU to divide it into four equal-sized coding units (CUs), and each CU can be divided recursively until the CU size is 8 × 8. Besides the existing QT partition structure in HEVC, the new multi-type tree (MTT) in H.266/VVC can be separated into two partition types: binary tree (BT) and ternary tree (TT). The partition direction of the binary and ternary trees can be further classified into horizontal and vertical directions, as displayed in Figure 1. In H.266/VVC, the maximum allowed size of a CTU is 128 × 128 [2]. The QT partition also occurs during the CTU encoding process. In the QT partition process, 128 × 128 corresponds to QtDepth 0, while 64 × 64 corresponds to QtDepth 1, and so on. Then, the nested MTT partition starts if the maximum size of the current prediction configuration that can perform the MTT partition is reached in the partition process. The MttDepth begins at 0 and increases by 1 for each BT or TT split. An H.266/VVC partition example of a CU is illustrated in Figure 2. In addition, H.266/VVC also includes many new coding tools and can be adapted to different videos. Some of these coding tools are used in intra coding, such as intra sub-partitions (ISP) [6], matrix weighted intra prediction (MIP) [7], multiple reference line intra prediction (MRL) [8], cross-component linear model (CCLM) and position dependent intra prediction combination (PDPC) [9], etc. ISP enables the luma CUs equal to 4 × 8 (or 8 × 4) and larger than 4 × 8 (or 8 × 4) to be divided into two and four sub-partitions horizontally or vertically, respectively. All sub-partitions have the same intra mode. In the MIP mode, smaller boundaries are obtained by averaging neighboring boundary samples. The averaged samples are the input of a matrix vector multiplication which is followed by the addition of an offset. Then, the predicted pixels are generated through linear interpolation [2].
However, more flexible partition modes and advanced coding tools result in large computations. This paper presents a fast intra coding algorithm based on visual perception, to efficiently decide the execution of the coding tools and MTT partition in H.266/VVC and save the encoding time. The proposed algorithm detects the visually distinguishable (VD) pixels that may affect visual perception. We observe that the variances computed by the VD pixels among MTT partitions within the coding unit are related to the usage of the coding tools and their corresponding MTT splitting modes. Therefore, the variances representing the perceptual information of human vision are used to be the cues for the decision of turning off coding tools and the input features of machine learning. Since the proposed method applies the factor of average background luminance for just-noticeable-distortion (JND) by using spatial average filtering to identify the VD pixels, we focus on All Intra configuration for the study.

2. Related Work

The fast algorithms of H.266/VVC intra coding have been explored to solve the problem of the high computational requirement of H.266/VVC. The methods can be roughly categorized into probability-based [10,11,12], learning-based [13,14,15,16,17,18,19,20], probability- and learning-based [21,22], texture-based [23,24,25,26], gradient-based [27,28], and texture- and gradient-based [29,30,31] techniques. The related work on the fast intra coding of H.266/VVC is discussed below.
Fu et al. [10] decided to terminate the vertical splits and TT horizontal split early based on the information of the two sub-CUs after the BT horizontal split. Park and Kang [11] determined the TT split direction of the current CU based on the Bayesian approach by comparing the rate-distortion costs (RD costs) of the horizontal and vertical splits of the BT split. If the RD cost of the horizontal direction is greater than or equal to the RD cost in the vertical direction, the horizontal direction split will be skipped during the TT split; otherwise, if the RD cost in the vertical direction is greater than or equal to the RD cost in the horizontal direction, then the TT vertical split will be skipped. Zhang et al. [12] employed the Bayesian principle to make the CU splitting decision, and utilized the improved de-blocking filter to determine the CU splitting mode.
Yang et al. [13] used the decision tree to propose a cascade decision structure for QT and MTT partitions. A fast intra mode decision by gradient descent search quickly found a suitable intra frame prediction mode. Chen et al. [14] utilized visual perception to extract the VD pixels. The distributions acquired by the horizontal and vertical projections of VD pixels and the quantization parameter (QP) were the input features of random forest classifiers to quickly select the MTT partition. Li et al. [15] proposed a multi-stage exit convolutional neural network model with an early-exit mechanism to decide the CU partition. Wu et al. [16] trained two support vector machine classifiers to predict the split or non-split, and horizontal or vertical split for CUs of different sizes. Wang et al. [17] devised a multi-stage early termination convolutional neural network model that can predict all the partition information of a 32 × 32 CU and its sub-CUs. Zouidi et al. [18] proposed an intra mode decision to skip unlikely intra prediction modes by using a multi-task learning convolutional neural network. Taabane et al. [19] used five binary light gradient boosting machine classifiers, which were offline-trained to predict the QT/MTT partitions. Park and Kang [20] developed two types of features which can be derived during encoding. The two lightweight neural network models use the features as input vectors to determine the TT structures subsequent to a quadtree partition.
Zhao et al. [21] proposed a complexity reduction algorithm based on statistical theory and size-adaptive convolutional neural network to decide whether to divide CUs of different sizes. Zhao et al. [22] extracted the spatial-temporal neighboring coding features by the deep convolutional network and fused all reference features acquired by different convolutional kernels to decide an optimal intra coding depth. The probability estimation and the spatial-temporal coherence are employed to further select the candidate partition modes within the optimal coding depth.
Peng et al. [23] adaptively set texture thresholds to classify CUs into three categories: simple, common, and complex. For simple CUs, the partition modes were skipped early, and for complex CUs, the non-partition mode was skipped. By computing the directional features, unnecessary horizontal or vertical partitions are omitted for common and complex CUs. Zhang et al. [24] proposed a fast intra block partition algorithm by utilizing gray level co-occurrence matrix to calculate the texture direction information of the CU and terminate the horizontal or vertical split of the BT and TT. Shu et al. [25] designed a complexity control algorithm for VVC intra coding by using texture entropy at the CU level. Zhang et al. [26] decided whether to divide the CU into sub-CUs according to the texture complexity. The CU splitting mode is determined by the texture direction.
Cui et al. [27] computed the gradients to omit unnecessary CU sizes and judge the BT split or TT split in the horizontal or vertical direction. Gou et al. [28] employed the histogram of the oriented gradient and the intra modes of the upper and left CUs for a fast intra mode decision.
Yoon et al. [29] utilized the gradient calculation used in the adaptive loop filter to propose the activity-based block partitioning decision method. The activity indicates the block texture complexity. Jing et al. [30] proposed a fast QT partition decision based on gradient structure similarity. The texture direction was determined by calculating the standard deviations of the vertical and horizontal directions of the current CU to decide the MTT partition direction. Fan et al. [31] utilized the variances of the pixels in a CU and the gradient features extracted by the Sobel operator to terminate the further splitting of a CU, decide the QT partition and selecting one partition from the five QT/MTT partitions.

3. Proposed Method

Except for the objective pixel distortion, the perceptible interval for gray level variation also affects the decoded video quality. Finding the luminance variation on the regions truly sensitive to human eyes is more feasible than the gradient or texture measurement [14]. People may just distinguish the change once the difference on the gray level between the current pixel and background exceeds the visible threshold. A just-noticeable-distortion (JND) can be utilized to quantify perceptual redundancy [32]. This paper applies the factor of average background luminance for JND [32] to detect the VD pixels within a CU. The average background luminance A ( i , j ) is computed by 3 × 3 average filtering for efficient calculation in this paper, as shown in (1). The I ( i , j ) represents the gray level of the current pixel with the coordinate ( i , j ) . The visible thresholds [32] due to background luminance J N D ( i , j ) are listed in (2), in which T 0 and γ are 17 and 3/128, respectively. VD ( i , j ) indicates whether or not the current pixel, with the coordinate i , j , is the VD pixel according to (3), in which VD ( i , j ) is defined according to whether the absolute value of the difference between I ( i , j ) and A ( i , j ) is greater than or equal to J N D ( i , j ) .
A ( i , j ) = I i 1 , j 1 I ( i , j 1 ) I i + 1 , j 1 I ( i 1 , j ) I ( i , j ) I i + 1 , j I ( i 1 , j + 1 ) I i , j + 1 I i + 1 , j + 1 × 1 9 1 9 1 9 1 9 1 9 1 9 1 9 1 9 1 9
J N D ( i , j ) = T 0 × 1 A ( i , j ) 127 1 2 + 3 , for   A ( i , j ) 127 γ × A ( i , j ) 127 + 3 , for   A ( i , j ) > 127
V D ( i , j ) = 1 , i f I ( i , j ) A ( i , j ) J N D ( i , j ) 0 , o t h e r w i s e
BTH, BTV, TTH, and TTV are the four kinds of MTT partitions in the QT-MTT structure. The MTT partition is allowed for the CUs no larger than 32 × 32, which corresponds to QtDepth 2 under the All Intra configuration. As a foundation for the fast algorithm, we propose calculating the variances of the numbers of VD pixels in various MTT splits of a CU. We compute v a r B T H ,   v a r B T V ,   v a r T T H , and v a r T T V for the four kinds of MTT splits in the square CUs (MttDepth 0), for the 32 × 32 CU (QtDepth 2) and four child 16 × 16 CUs, as shown in (4)–(7). V D s u b C U k represents the number of VD pixels in each sub-CU for the four kinds of MTT splits. In particular, the V D s u b C U k of the middle sub-CU of the TT split is further divided by 2 for normalization. The μ B T H μ B T V , μ T T H , and μ T T V represent the mean values of all V D s u b C U k for each kind of MTT split. The individual variances, v a r B T H , v a r B T V , v a r T T H , and v a r T T V of the four MTT splits, are calculated, and μ v a r is the average of the above four variances, as shown in (8). The v a r M T T , as in (9), denotes the variance of v a r B T H v a r B T V v a r T T H , and v a r T T V .
v a r B T H = 1 2 k = 1 2 V D s u b C U k μ B T H 2
v a r B T V = 1 2 k = 1 2 V D s u b C U k μ B T V 2
v a r T T H = 1 3 k = 1 3 V D s u b C U k μ T T H 2
v a r T T V = 1 3 k = 1 3 V D s u b C U k μ T T V 2
μ v a r = v a r B T H + v a r B T V + v a r T T H + v a r T T V 4
v a r M T T = v a r B T H μ v a r 2 + v a r B T V μ v a r 2 + v a r T T H μ v a r 2 + ( v a r T T V μ v a r ) 2 4
The sets of variances for the parent 32 × 32 CU (QtDepth 2) and four child 16 × 16 CUs are indicated in (10) and (11), respectively.
v a r S e t 32 = { v a r B T H 32 , v a r B T V 32 , v a r T T H 32 , v a r T T V 32 }
v a r S e t 16 = { v a r B T H 16 , s , v a r B T V 16 , s , v a r T T H 16 , s , v a r T T V 16 , s | s = 1,2 , 3,4 }

3.1. Fast Decision for Intra Coding Tools

The intra coding tools in H.266/VVC can ameliorate the compression efficiency, but increase the coding time. Six video sequences, including Class A1 Tango2 (3840 × 2160), Class A2 ParkRunning3 (3840 × 2160), Class B BQTerrace (1920 × 1080), Class C PartyScene (832 × 480), Class D RaceHorses (416 × 240), and Class E KristenAndSara (1280 × 720), are used to observe the contribution of the intra coding tools of ISP and MIP. The statistics of the probabilities, and the performance of Bjøntegaard Delta Bitrate (BDBR) [33] and Time Saving (TS) using ISP and MIP for each MttDepth in VTM10.0 [2,3] and All Intra configuration are shown in Table 1 and Table 2, respectively. The definition of TS is shown in (12). The probability distributions show that the usage of the two coding tools decreases as the MttDepth increases. The BDBR is insignificant when the MttDepth is 3. From the observations, we assume that the ISP and MIP can be turned off when the MttDepth is larger than 2.
Time   Saving T S = 1 4 Q P 22,27,32,37 T V T M 10.0 Q P T p r o p o s e d Q P T V T M 10.0 Q P × 100 %
When the QtDepth is 2 or 3, and the MttDepth of CU is larger than 2, the ISP and MIP are turned off conditionally by our method. The decision to turn off the ISP and MIP tools is determined by the criterion exhibited in (13). The threshold, T h T o o l s , related to the quantization parameter is defined as in (14). The choice of α in (14) is particularly important, and selecting a larger α will lead to an increase in BDBR. Figure 3 shows the relationship between α , and the encoding time ratio and BDBR by the average of the experiments, using the six sequences mentioned above. It can be found that the overall encoding time and BDBR have relatively moderate changes when α is in the range of 0.7 to 1.5. When α exceeds 1.5, the BDBR begins to increase dramatically. Therefore, the final parameter α is chosen as 1.5.
v a r M T T T h T o o l s
T h T o o l s = α × Q P

3.2. Fast Decision for Multi-Type Tree Partition

The proposed algorithm adopts the random forest (RF) model [34] in machine learning for the fast decision of the MTT partition. In the RF model, multiple decision trees are constructed using a random sampling method, and the prediction is made by voting. Compared to a single decision tree model, the RF model has stronger generalization ability and a lower overfitting risk. Two binary classifiers using RF are trained individually for the BT split and TT split to quickly determine the split direction in the 32 × 32 CUs corresponding to QtDepth 2. The input features consist of v a r S e t 32 , v a r S e t 16 , and QP. A total of 21 features are the input to the two RF classifiers, as shown in Figure 4. The horizontal or vertical split direction is the output of the binary classifier.
The video frames in the training dataset consist of two parts and are different from the test video sequences in our experiments. The training dataset for the first part contains still images of four different resolutions (4928 × 3264, 2880 × 1920, 1536 × 1024, 768 × 512) from the CPIH dataset [35]. For the second part of the training dataset, four video sequences with a 4 K (3840 × 2160) resolution are selected from the Ultra Video Group (UVG) dataset [36]. In the UVG dataset, Jockey, FlowerFocus, YachtRide, and RiverBank video sequences are used. The RF classifiers are trained by a total of 177,894 CUs for the horizontal/vertical classifier of BT and a total of 141,654 CUs for the horizontal/vertical classifier of TT. We use OpenCV 4.5.4 [37] to train the RF models. The overall training dataset is divided into a 75% training set and a 25% validation set. In addition, we adjust the hyperparameters of the random forest model and select 35 decision trees in the BT classifier and 25 decision trees in the TT classifier.

3.3. Overall Algorithm

The overall algorithm in this paper mainly consists of two strategies based on visual perception information. The first strategy is the fast algorithm for turning off the ISP and MIP intra picture prediction tools, based on the condition decided by the variance when the QtDepth is 2 or 3 and the MttDepth of CU is larger than 2. The second strategy is to use the information of visual perception and QP to propose horizontal/vertical split models trained by the random forest classifiers of machine learning, to select the splitting direction for BT and TT individually when the QtDepth is 2. In addition, if both v a r T T H and v a r T T V , calculated by (6) and (7), are 0 for the case of 32 × 32 CU (QtDepth 2) or 16 × 16 CU (QtDepth 3), the subsequent TT splits are terminated in the corresponding QtDepth. The flowchart of the overall algorithm is shown in Figure 5.

4. Results

We implement the H.266/VVC reference software version VTM 10.0 [2,3] in our experiments to show the performance under the All Intra configuration, with the QPs of 22, 27, 32, and 37. Class A1, Class A2, Class B, Class C, Class D, Class E, and Class F are included in the test video sequences. The experimental setup follows the common test conditions (CTC) [38]. The information of the test sequences is shown in Table 3. The experiments are executed on Windows 10 64 bit with an Intel i7-7700 K processor @4.20 GHz and a 64 GB RAM. Bjøntegaard Delta Bitrate (BDBR), Bjøntegaard Delta Peak Signal-to-Noise Ratio (BDPSNR), Time Saving (TS), and TS/BDBR are used to evaluate the performance and compare with [16]. The tradeoff between coding efficiency and computation time is evaluated by TS/BDBR, which is computed by the average TS divided by the average BDBR. The greater the ratio, the better the performance.
Table 4 lists the experimental results of the proposed algorithm for all test sequences. The average BDBR, BDPSNR, and TS are 1.535%, −0.0814 dB, and 47.26%, respectively. For comparison with [16], the proposed algorithm results in only 1.454% BDBR on average, lower than the 2.71% in [16]. The average BDBR of each Class sequence for the proposed algorithm is lower than that in [16]. The average BDBRs in Class C and Class D are less than half of those in [16], and the average TSs are also well maintained. Furthermore, our method achieves a TS/BDBR of 32.68, which is better than the 23.31 of [16].
Table 5 lists the performance comparison between the proposed algorithm and [31], in which only Class B, C, D, and E sequences are available. The testing of [31] was conducted using VTM 7.0, as indicated. The coding tools of the ISP, MIP and low-frequency non-separable transform (LFNST) are turned off for the experiments in [31]. Although LFNST is deactivated, the ISP and MIP are conditionally turned off by the proposed criterion in our experiments for comparison with [31], as shown in Table 5. The number of encoded frames is 100, which is the same as that in [31]. The performance of the proposed algorithm has the BDBRs of 1.49% and 2.12% in Class B and E sequences, respectively, which is better than 2.07% and 2.90% in [31]. The TS/BDBR of the proposed algorithm is 33.03, which is better than the 29.66 of [31].
Figure 6 compares the rate-distortion curve (RD curve) between the proposed algorithm and VTM10.0 for Class A2 Parkrunning3 sequences. The curve of the proposed method is close to that of VTM10.0, and the TS of the proposed method reaches 59.75%. Figure 7 demonstrates the QT/MTT partition results of the frame in the class D BQSquare sequence. The partition results are similar regarding both the complex blocks and uniform blocks.

5. Conclusions

In this paper, we present the fast and precise decisions for intra coding tools and MTT partitions based on the variances of the numbers of VD pixels among the MTT splits for the intra coding of H.266/VVC. According to the proposed criterion, conditionally turning off the ISP and MIP intra coding tools can speed up the encoding process. RF classifiers of machine learning based on the visual perception analysis effectively skip unnecessary split directions for the BT and the TT of the nested MTT partition. Our experimental results demonstrate that the proposed algorithm provides better coding efficiency and performance than the previous studies. Conventional methods only consider fast decisions either for coding tools or QT/MTT partitions. Previous studies regarding the H.266/VVC fast coding algorithm rarely utilize perceptual redundancy. Our proposed method takes advantage of visual perception to terminate the processes of both aspects of the coding tools and MTT partition early. Visual perception in the temporal domain can be investigated for applications in low-delay and random access configurations in future works.

Author Contributions

Conceptualization, Y.-H.T. and M.-J.C.; Data curation, Y.-H.T.; Funding acquisition, M.-J.C. and C.-H.Y.; Methodology, Y.-H.T. and M.-J.C.; Software, Y.-H.T.; Supervision, M.-J.C., C.-M.Y. and C.-H.Y.; Validation, C.-R.L. and M.-C.H.; Writing—original draft, Y.-H.T., C.-R.L. and M.-J.C.; Writing—review and editing, M.-C.H., C.-M.Y. and C.-H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Science and Technology, Taiwan, under grant MOST 109-2221-E-259-012-MY3.

Data Availability Statement

Publicly available datasets were analyzed in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bross, B.; Chen, J.; Ohm, J.R.; Sullivan, G.J.; Wang, Y.K. Developments in international video coding standardization after AVC, with an overview of versatile video coding (VVC). Proc. IEEE 2021, 109, 1463–1493. [Google Scholar] [CrossRef]
  2. Chen, J.; Ye, Y.; Kim, S.H. Algorithm Description for Versatile Video Coding and Test Model 10 (VTM 10); doc. JVET-S2002; Joint Video Experts Team: Teleconference, 2020. [Google Scholar]
  3. VVC Reference Software. Available online: https://vcgit.hhi.fraunhofer.de/jvet/VVCSoftware_VTM/-/tree/VTM-10.0 (accessed on 13 August 2020).
  4. Bross, B.; Wang, Y.K.; Ye, Y.; Liu, S.; Chen, J.; Sullivan, G.J.; Ohm, J.R. Overview of the versatile video coding (VVC) standard and its applications. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 3736–3764. [Google Scholar] [CrossRef]
  5. Hamidouche, W.; Biatek, T.; Abdoli, M.; François, E.; Pescador, F.; Radosavljević, M.; Menard, D.; Raulet, M. Versatile video coding standard: A review from coding tools to consumers deployment. IEEE Consum. Electron. Mag. 2022, 11, 10–24. [Google Scholar] [CrossRef]
  6. De-Luxán-Hernández, S.; George, V.; Ma, J.; Nguyen, T.; Schwarz, H.; Marpe, D.; Wiegand, T. CE3: Intra Sub-Partitions Coding Mode (Tests 1.1.1 and 1.1.2); doc. JVET-M0102; Joint Video Experts Team: Marrakech, Morocco, 2019. [Google Scholar]
  7. Pfaff, J.; Stallenberger, B.; Schäfer, M.; Merkle, P.; Helle, P.; Hinz, T.; Schwarz, H.; Marpe, D.; Wiegand, T. CE3: Affine Linear Weighted Intra Prediction (CE3-4.1, CE3-4.2); doc. JVET-N0217; Joint Video Experts Team: Geneva, Switzerland, 2019. [Google Scholar]
  8. Bross, B.; Keydel, P.; Schwarz, H.; Marpe, D.; Wiegand, T.; Zhao, L.; Zhao, X.; Li, X.; Liu, S.; Chang, Y.J.; et al. CE3: Multiple Reference Line Intra Prediction (Test 1.1.1, 1.1.2, 1.1.3 and 1.1.4); doc. JVET-L0283; Joint Video Experts Team: Macao, China, 2018. [Google Scholar]
  9. Van der Auwera, G.; Seregin, V.; Said, A.; Ramasubramonian, A.K.; Karczewicz, M. CE3: Simplified PDPC (Test 2.4.1); doc. JVET-K0063; Joint Video Experts Team: Ljubljana, Slovenia, 2018. [Google Scholar]
  10. Fu, T.; Zhang, H.; Mu, F.; Chen, H. Fast CU partitioning algorithm for H.266/VVC intra-frame coding. In Proceedings of the IEEE International Conference on Multimedia and Expo, Shanghai, China, 8–12 July 2019. [Google Scholar]
  11. Park, S.-h.; Kang, J.W. Context-based ternary tree decision method in versatile video coding for fast intra-coding. IEEE Access 2019, 7, 172597–172605. [Google Scholar] [CrossRef]
  12. Zhang, Q.; Zhao, Y.; Jiang, B.; Wu, Q. Fast CU partition decision method based on Bayes and improved de-blocking filter for H.266/VVC. IEEE Access 2021, 9, 70382–70391. [Google Scholar] [CrossRef]
  13. Yang, H.; Shen, L.; Dong, X.; Ding, Q.; An, P.; Jiang, G. Low-complexity CTU partition structure decision and fast intra mode decision for versatile video coding. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1668–1682. [Google Scholar] [CrossRef]
  14. Chen, M.J.; Lee, C.A.; Tsai, Y.H.; Yang, C.M.; Yeh, C.H.; Kau, L.J.; Chang, C.Y. Efficient partition decision based on visual perception and machine learning for H.266/versatile video coding. IEEE Access 2022, 10, 42141–42150. [Google Scholar] [CrossRef]
  15. Li, T.; Xu, M.; Tang, R.; Chen, Y.; Xing, Q. DeepQTMT: A deep learning approach for fast QTMT-based CU partition of intra-mode VVC. IEEE Trans. Image Process 2021, 30, 5377–5390. [Google Scholar] [CrossRef] [PubMed]
  16. Wu, G.; Huang, Y.; Zhu, C.; Song, L.; Zhang, W. SVM based fast CU partitioning algorithm for VVC intra coding. In Proceedings of the 2021 IEEE International Symposium on Circuits and Systems, Daegu, Korea, 22–28 May 2021. [Google Scholar]
  17. Wang, Y.; Dai, P.; Zhao, J.; Zhang, Q. Fast CU partition decision algorithm for VVC intra coding using an MET-CNN. Electronics 2022, 11, 3090. [Google Scholar] [CrossRef]
  18. Zouidi, N.; Kessentini, A.; Hamidouche, W.; Masmoudi, N.; Menard, D. Multitask learning based intra-mode decision framework for versatile video coding. Electronics 2022, 11, 4001. [Google Scholar] [CrossRef]
  19. Taabane, I.; Menard, D.; Mansouri, A.; Ahaitouf, A. Machine learning based fast QTMTT partitioning strategy for VVenC encoder in intra coding. Electronics 2023, 12, 1338. [Google Scholar] [CrossRef]
  20. Park, S.h.; Kang, J.W. Fast multi-type tree partitioning for versatile video coding using a lightweight neural network. IEEE Trans. Multimed. 2021, 23, 4388–4399. [Google Scholar] [CrossRef]
  21. Zhao, J.; Dai, P.; Zhang, Q. A complexity reduction method for VVC intra prediction based on statistical analysis and SAE-CNN. Electronics 2021, 10, 3112. [Google Scholar] [CrossRef]
  22. Zhao, T.; Huang, Y.; Feng, W.; Xu, Y.; Kwong, S. Efficient VVC intra prediction based on deep feature fusion and probability estimation. IEEE Trans. Multimed. 2022. Early Access. [Google Scholar] [CrossRef]
  23. Peng, S.; Peng, Z.; Ren, Y.; Chen, F. Fast intra-frame coding algorithm for versatile video coding based on texture feature. In Proceedings of the 2019 IEEE International Conference on Real-time Computing and Robotics, Irkutsk, Russia, 4–9 August 2019. [Google Scholar]
  24. Zhang, H.; Yu, L.; Li, T.; Wang, H. Fast GLCM-based intra block partition for VVC. In Proceedings of the 2021 Data Compression Conference (DCC), Snowbird, UT, USA, 23–26 March 2021; p. 382. [Google Scholar]
  25. Shu, Z.; Li, J.; Peng, Z.; Chen, F.; Yu, M. Intra complexity control algorithm for VVC. Electronics 2022, 11, 2572. [Google Scholar] [CrossRef]
  26. Zhang, Q.; Zhao, Y.; Jiang, B.; Huang, L.; Wei, T. Fast CU partition decision method based on texture characteristics for H.266/VVC. IEEE Access 2020, 8, 203516–203524. [Google Scholar] [CrossRef]
  27. Cui, J.; Zhang, T.; Gu, C.; Zhang, X.; Ma, S. Gradient-based early termination of CU partition in VVC intra coding. In Proceedings of the 2020 Data Compression Conference, Snowbird, UT, USA, 24–27 March 2020. [Google Scholar]
  28. Gou, A.; Sun, H.; Katto, J.; Li, T.; Zeng, X.; Fan, Y. Fast intra mode decision for VVC based on histogram of oriented gradient. In Proceedings of the 2022 IEEE International Symposium on Circuits and Systems (ISCAS), Austin, TX, USA, 27 May–1 June 2022; pp. 3028–3032. [Google Scholar]
  29. Yoon, Y.U.; Kim, J.G. Activity-based block partitioning decision method for versatile video coding. Electronics 2022, 11, 1061. [Google Scholar] [CrossRef]
  30. Jing, Z.; Li, P.; Zhao, J.; Zhang, Q. A fast CU partition algorithm based on gradient structural similarity and texture features. Symmetry 2022, 14, 2644. [Google Scholar] [CrossRef]
  31. Fan, Y.; Chen, J.A.; Sun, H.; Katto, J.; Jing, M.E. A fast QTMT partition decision strategy for VVC intra prediction. IEEE Access. 2020, 8, 107900–107911. [Google Scholar] [CrossRef]
  32. Chou, C.H.; Li, Y.C. A perceptually tuned subband image coder based on the measure of just-noticeable-distortion profile. IEEE Trans. Circuits Syst. Video Technol. 1995, 5, 467–476. [Google Scholar] [CrossRef]
  33. Bjøntegaard, G. Calculation of Average PSNR Differences between RD-Curves; doc. VCEG-M33; ITU-T Video Coding Experts Group: Austin, TX, USA, 2001. [Google Scholar]
  34. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  35. Li, T.; Xu, M.; Deng, X. A deep convolutional neural network approach for complexity reduction on intra-mode HEVC. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo, Hong Kong, China, 10–14 July 2017; pp. 1255–1260. [Google Scholar]
  36. Mercat, A.; Viitanen, M.; Vanne, J. UVG dataset: 50/120fps 4K sequences for video codec analysis and development. In Proceedings of the 11th ACM Multimedia Systems Conference, New York, NY, USA, 27 May 2020; pp. 297–302. [Google Scholar]
  37. Bradski, G. The OpenCV Library. Dr. Dobb's J. Softw. Tools. 2000, 25, 120–125. [Google Scholar]
  38. Bossen, F.; Boyce, J.; Suehring, K.; Li, X.; Seregin, V. JVET Common Test Conditions and Software Reference Configurations for SDR Video; doc. JVET-N1010; Joint Video Experts Team: Geneva, Switzerland, 2019. [Google Scholar]
Figure 1. H.266/VVC QT-MTT partition diagram.
Figure 1. H.266/VVC QT-MTT partition diagram.
Electronics 12 02079 g001
Figure 2. Example of H.266/VVC QT-MTT partition.
Figure 2. Example of H.266/VVC QT-MTT partition.
Electronics 12 02079 g002
Figure 3. Selection of α in T h T o o l s .
Figure 3. Selection of α in T h T o o l s .
Electronics 12 02079 g003
Figure 4. Illustration of input features for the horizontal/vertical random forest classifier.
Figure 4. Illustration of input features for the horizontal/vertical random forest classifier.
Electronics 12 02079 g004
Figure 5. The flowchart of the overall fast algorithm.
Figure 5. The flowchart of the overall fast algorithm.
Electronics 12 02079 g005
Figure 6. RD curve comparison between the proposed algorithm and VTM10.0 for Class A2 Parkrunning3 sequence.
Figure 6. RD curve comparison between the proposed algorithm and VTM10.0 for Class A2 Parkrunning3 sequence.
Electronics 12 02079 g006
Figure 7. QT/MTT partition results for the frame in Class D BQSquare video sequence (QP 37).
Figure 7. QT/MTT partition results for the frame in Class D BQSquare video sequence (QP 37).
Electronics 12 02079 g007
Table 1. Probability distributions of ISP and MIP in each MttDepth.
Table 1. Probability distributions of ISP and MIP in each MttDepth.
MttDepthISPMIPAverage
067.18%77.16%72.17%
138.32%42.78%40.55%
227.33%30.13%28.73%
316.89%10.88%13.89%
Table 2. Performance of turning off ISP and MIP in each MttDepth.
Table 2. Performance of turning off ISP and MIP in each MttDepth.
MttDepthBDBR (%)TS (%)
01.2112.70
10.778.48
20.545.14
30.124.32
Table 3. Information of the test sequences.
Table 3. Information of the test sequences.
ClassResolutionSequenceFrame Rate (fps)Bit DepthFrame Count
A13840 × 2160Campfire3010300
3840 × 2160FoodMarket46010300
3840 × 2160Tango26010294
A23840 × 2160CatRobot6010300
3840 × 2160DaylightRoad26010300
3840 × 2160ParkRunning35010300
B1920 × 1080BasketballDrive508500
1920 × 1080BQTerrace608600
1920 × 1080Cactus508500
1920 × 1080MarketPlace6010600
1920 × 1080RitualDance6010600
C832 × 480BasketballDrill508500
832 × 480BQMall608600
832 × 480PartyScene508500
832 × 480RaceHorses308300
D416 × 240BasketballPass508500
416 × 240BlowingBubbles508500
416 × 240BQSquare608600
416 × 240RaceHorses308300
E1280 × 720FourPeople608600
1280 × 720Johnny608600
1280 × 720KristenAndSara608600
F1920 × 1080ArenaOfValor608600
832 × 480BasketballDrillText508500
1280 × 720SlideEditing308300
1280 × 720SlideShow208500
Table 4. Performance comparison between the proposed algorithm and [16].
Table 4. Performance comparison between the proposed algorithm and [16].
ClassSequence[16]Proposed
BDBR (%)TS (%)BDPSNR (dB)BDBR (%)TS (%)
A1Campfire2.6564.74−0.03611.16552.48
FoodMarket41.4746.93−0.05931.80650.08
Tango22.4264.45−0.02661.71350.73
Average2.1858.71−0.04071.56151.10
A2CatRobot3.2763.81−0.04921.69652.68
DaylightRoad22.0270.39−0.05281.24953.11
ParkRunning31.4655.14−0.05590.95559.75
Average2.2563.11−0.05261.30055.18
BBasketballDrive2.3867.81−0.06052.20948.57
BQTerrace2.4364.25−0.05261.23540.90
Cactus2.7866.61−0.05401.54047.86
MarketPlace2.5871.93−0.06411.58656.74
RitualDance4.2164.06−0.09381.86852.15
Average2.8866.93−0.06501.68849.24
CBasketballDrill5.3965.29−0.08851.84244.15
BQMall2.9262.93−0.08871.65644.80
PartyScene1.4058.77−0.03940.55739.94
RaceHorses2.0062.10−0.06101.07346.97
Average2.9362.27−0.06941.28243.97
DBasketballPass2.3461.15−0.09841.61345.00
BlowingBubbles2.2459.94−0.03170.47638.18
BQSquare1.6859.98−0.03200.40935.90
RaceHorses1.6958.98−0.07231.12646.26
Average1.9960.01−0.05860.90641.34
EFourPeople4.3667.14−0.10281.87545.63
Johnny4.3467.01−0.09522.48547.56
KristenAndSara3.5666.21−0.08771.84445.69
Average4.0966.79−0.09522.06846.29
FArenaOfValor--−0.10591.80946.49
BasketballDrillText--−0.09991.80242.99
SlideEditing--−0.32642.25149.60
SlideShow--−0.18212.05744.50
Average--−0.17861.98045.90
Total Average--−0.08141.53547.26
TS/BDBR-30.79
Same Sequence Average2.7163.16−0.06381.45447.51
Same Sequence TS/BDBR23.3132.68
Table 5. Performance comparison between the proposed algorithm and [31].
Table 5. Performance comparison between the proposed algorithm and [31].
ClassSequence[31]Proposed
BDBR (%)TS (%)BDBR (%)TS (%)
BBasketballDrive3.2859.352.1452.17
BQTerrace1.0845.300.7742.78
Cactus1.8452.441.5750.12
Average2.0752.361.4948.36
CBasketballDrill1.8248.481.9348.35
BQMall1.8752.471.5948.27
PartyScene0.2638.620.4242.18
RaceHorses0.8849.051.1650.17
Average1.2147.161.2847.24
DBasketballPass1.9547.701.9147.78
BlowingBubbles0.4740.350.6743.81
BQSquare0.1931.950.1633.14
RaceHorses0.5441.691.1947.37
Average0.7940.420.9843.03
EFourPeople2.7057.571.9649.31
Johnny3.2256.882.3851.24
KristenAndSara2.7855.112.0249.87
Average2.9056.522.1250.14
Total Average1.6348.351.4246.90
TS/BDBR29.6633.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tsai, Y.-H.; Lu, C.-R.; Chen, M.-J.; Hsieh, M.-C.; Yang, C.-M.; Yeh, C.-H. Visual Perception Based Intra Coding Algorithm for H.266/VVC. Electronics 2023, 12, 2079. https://doi.org/10.3390/electronics12092079

AMA Style

Tsai Y-H, Lu C-R, Chen M-J, Hsieh M-C, Yang C-M, Yeh C-H. Visual Perception Based Intra Coding Algorithm for H.266/VVC. Electronics. 2023; 12(9):2079. https://doi.org/10.3390/electronics12092079

Chicago/Turabian Style

Tsai, Yu-Hsiang, Chen-Rung Lu, Mei-Juan Chen, Meng-Chun Hsieh, Chieh-Ming Yang, and Chia-Hung Yeh. 2023. "Visual Perception Based Intra Coding Algorithm for H.266/VVC" Electronics 12, no. 9: 2079. https://doi.org/10.3390/electronics12092079

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop