Audio Video Standard


Audio Video Coding Standard refers to the digital audio and digital video series compression standard formulated by the Audio and Video coding standard workgroup of China. Work began in 2002, and three generations of standards were published.
The first generation AVS standard includes "Information Technology, Advanced Audio Video Coding, Part 2: Video" and "Information Technology, Advanced Audio Video Coding Part 16: Radio Television Video" For the second generation, referred to as AVS2, the primary application target was ultra-high-definition television video, supporting the efficient compression of ultra-high-resolution, high-dynamic-range videos, and was published as IEEE international standard IEEE 1857.4. An industry alliance was established to develop and promote AVS standards. A patent pool charges a small royalty for terminal products excluding content providers and operators.
The AVS3 codec was added to DVB's media delivery toolbox.

Organizations

Workgroup

The AVS workgroup was founded in June 2002 to cooperate with Chinese enterprises and scientific research institutions, to formulate and revise common technical standards such as digital audio and digital video's compression, decompression, processing and representation, thus to provide efficient and economic coding/decoding technologies for digital audio and digital video devices and systems, serving the high-resolution digital broadcasting, high-density digital laser storage media, wireless broadband multimedia communication, Internet broadband streaming media, and other applications.
The workgroup is headed by Gao Wen, the academician of Chinese Academy of Engineering, the professor and Ph.D. supervisor of Peking University, and the deputy director of the National Natural Science Fund Committee, consisting of a requirement group, system group, video group, audio group, test group, intellectual property group and other departments. The first setback was when China did not use AVS for its own digital television broadcast system in 2003.

Patent pool

A patent pool which manages and authorizes the patents was founded on September 20, 2004. The committee was the first patent pool management institution in China. An independent corporate association, the Beijing Haidian District Digital Audio and Video Standard Promotion Center, is registered in the Civil Affairs Bureau of Haidian District of Beijing City. for patent technologies included in the standard, as the expert committee and the main business decision-making institution of the promotion center. The royalty for the first generation AVS standard was only one yuan per terminal. The plan was to charge a small amount of royalty only for the terminal, excluding the contents, as well as software services on the Internet.

Alliance

The AVS Industry Alliance, short for Zhongguancun Audio Visual Industry Technology Innovation Alliance, was formed in May 2005 in Beijing by twelve entities, including TCL Group Co., Ltd., Skyworth Group Research Institute, Huawei Technology Co., Ltd., Hisense Group Co., Ltd., Haier Group Co., Ltd., Beijing Haier Guangke Co., Ltd., Inspur Group Co., Ltd., Joint Source Digital Audio Video Technology Co., Ltd., New Pudong District Mobile Communication Association, Sichuan Changhong Co., Ltd., Shanghai SVA Central Research Institute, Zte Communication Co., Ltd., and Zhongguancun Hi-Tech Industry Association. The organization is also known as AVSA, and it collaborates with "AVS Workgroup" and "AVS Patent Pool Management Committee" as part of the "Three Carriages."

First generation

The first generation AVS standard included Chinese national standard "Information Technology, Advanced Audio Video Coding, Part 2: Video" and "Information Technology, Advanced Audio Video Coding Part 16: Radio Television Video". A test hosted by the Radio and Television Planning Institute of the State Administration of Radio, Film, and Television shows: if the AVS1 bitrate is half of MPEG-2 standard, the coding quality will reach excellent for both standard definition or high definition; if the bitrate is less than 1/3, it also reaches good-excellent levels. The AVS1 standard video part was promulgated as the Chinese national standard in February 2006.
Around this time, AVS was considered for use in the enhanced versatile disc format, although products never reached the market.
During the May 7–11, 2007 meeting of the ITU-T, AVS1 was one of the standards available for Internet Protocol television along with MPEG-2, H.264 and VC-1. On June 4, 2013, the AVS1 video part was issued by the Institute of Electrical and Electronics Engineers as standard IEEE1857-2013,
AVS+ is not only the radio, film and television industry standard GY/T 257.1-2012 "Advanced Audio Video Coding for Radio and Television, Part 1: Video" issued by the SARFT on July 10, 2012, but also the enhanced version of AVS1.

Second generation

The second-generation AVS standard included the series of Chinese national standard "Information Technology, Efficient Multi Media Coding". AVS2 mainly faces the transmission of extra HD TV programs,
The SARFT issued AVS2 video as the industry standard in May 2016, and as the Chinese national standard on December 30, 2016.
AVS2 was published by the Institute of Electrical and Electronics Engineers as standard 1857.4-2018 in August, 2019.
A test showed the coding efficiency of AVS2 more than doubled that of AVS+, and the compression rate surpassed the international standard HEVC. Compared with the first generation AVS standard, the second can save half transmission bandwidth.

Features

AVS2 adopts a hybrid-coding framework, and the whole coding process includes modules such as intra-frame prediction, inter-frame prediction, transformation, quantization, inverse quantization and inverse transformation, loop filter and entropy coding. It owns technical features as followings:

Flexible Coding Structure Partition

In order to satisfy the requirements of HD and Ultra HD resolution videos for the compression efficiency, AVS2 adopts a block partition structure based on the quadtree, including the CU, PU and TU. An image is partitioned into LCU of fixed size, which is iterated and partitioned into a series of CUs in the form of quadtree. Each CU contains a luminance-coding block and two corresponding chrominance-coding blocks. Compared with the traditional macro block, the partition structure based on the quadtree is more flexible, with the CU size extended from 8×8 to 64×64.
The PU stipulates all prediction modes of CU, and it is the basic unit for the prediction, including intra-frame and inter-frame prediction. The maximum size of PU is not permitted to exceed that of the current CU it belongs to. On the basis of AVS1 square intra-frame prediction blocks, the non-square intra-frame prediction block partition is added. Meanwhile, on the basis of the symmetric prediction block partition, the inter-frame prediction also adds 4 asymmetric partition ways.
Besides CU and PU, AVS2 also defines a transformation unit TU for the prediction of residual transformation and quantization. TU is the basic unit of transformation and quantization, defined in CU like PU. Its size selection is related to the corresponding PU shape. If the current CU is partitioned into non-square PU, the non-square partition will be applied to the corresponding TU; otherwise, the square partition type will be applied. The size of TU could be greater than that of the PU, but no more than that of the CU it belongs to.

Intra Prediction Coding

Compared with the AVS1 and H.264/AVC, AVS2 designs 33 modes for the intra-frame prediction coding of luminance blocks, including DC prediction mode, plane prediction mode, bilinear prediction mode and 30 angel prediction modes. There are 5 modes for chrominance blocks: DC mode, horizontal prediction mode, vertical prediction mode, bilinear interpolation mode as well as the luminance derived mode newly added.

Inter Prediction Coding

Compared with AVS1, AVS2 increases the maximum quantity of candidate reference frames to 4, so as to adapt to the multi-level reference frame management, which also takes full advantage of the redundant space of the buffer. In order to satisfy the requirements of multiple reference frame management, AVS2 adopts a kind of multi-level reference frame management mode. In this mode, the frames in each GOP are partitioned into multiple levels according to the reference relationship between frames.

Inter Prediction Mode

On the basis of AVS1's three image types I, P, B, according to the requirements of application, AVS2 adds the forward multi-hypothesis prediction image F. Aiming at the video surveillance, scene play and other specific applications, AVS2 designs scene frames and reference scene frame S. For Frame B, in addition to traditional forward, backward, two-way mode and skip/direct mode, a new symmetric mode is added. In symmetric mode, only forward motion vectors are required to be encoded, and then backward motion vectors will be derived from the forward motion vectors.
In order to fully exert the performance of the skip/direct mode of Frame B, AVS2 also adopts multi-direction skip/direct mode under the premise of retaining the original skip/direct mode of Frame B: two-way skip/direct mode, symmetrical skip/direct mode, backward skip/direct mode and forward skip/direct mode. For the four particular modes, the same prediction mode block between adjacent blocks is discovered according to the prediction mode of the current block, and the motion vectors of adjacent blocks with the same prediction mode, which are found out first, will be considered as that of the current block. For Frame F, coding blocks can refer to the two forward reference blocks, equivalent to the double hypothesis prediction of Frame P.
AVS2 divides the multi-hypothesis prediction into two categories, namely temporal and spatial multi-hypothesis mode. The current encoding block of the time-domain double hypothesis applies the weighted average of prediction blocks as the current prediction value, but there is only one for both the MVD and the reference image index, while another MVD and reference image index are derived from linear scaling based on the distance in the time domain.
The spatial-domain double prediction is also called DMH, which is obtained by fusing two prediction points around the initial prediction point, and the initial point is located in the line between the two prediction points. In addition to the initial prediction point, there are 8 prediction points in total, to be fused only with the two prediction points located in the same straight line with the initial prediction point. Besides four different directions, the adjustment will also be conducted according to the distance, and the four modes with 1/2 pixel distance and 1/4 pixel distance will be respectively calculated, plus the initial prediction point, to work out 9 modes in total for comparison, thus to select out the optimal prediction mode.
The scene frame is proposed by AVS2 based on the surveillance video coding method of background modeling. When the surveillance tool is not opened, Frame I is only for reference for images before the next random access point. When the surveillance tool is opened, AVS2 will apply a certain frame in the video as the scene image frame G, which can be considered as a long-term reference for the subsequent images. AVS2 can generate the scene image frame GB with some frames in the video, and frame GB can also be applied as a long-term reference.
In order to simplify the motion compensation, AVS2 adopts an 8-tap interpolation filter based on DCT transformation, which requires only one filtering, and supports the generation of higher motion vector accuracy than 1/4 pixel.