![]() METHOD FOR DECODING VIDEO
专利摘要:
inter-prediction method, inter-prediction apparatus, video decoding apparatus, video-encoding apparatus, and computer-readable recording medium an inter-prediction method including determining a colocalized block of a current block of a current image among blocks of an image which is restored before the current image; preferably check if a first reference list among reference lists of the colocalized block is referenced and selectively check if a second reference list is referenced according to whether the first reference list is referenced; based on a verification result, determine a single colocalized reference list from the first reference list and the second reference list; determine a reference block from the current block using movement information from the colocalized reference list; and perform inter prediction on the current block using the given reference block. 公开号:BR112013033906B1 申请号:R112013033906-3 申请日:2012-07-02 公开日:2020-09-29 发明作者:Il-koo KIM 申请人:Samsung Electronics Co., Ltd; IPC主号:
专利说明:
Technical Field [001] The present invention relates to a method and apparatus for encoding a video by means of inter prediction and motion compensation and to a method and apparatus to decode video by means of inter prediction and motion compensation. Fundamentals of Technique [002] As hardware to reproduce and store high-resolution or high-quality video content is being developed and delivered, the need for a video codec to effectively encode or decode high-resolution or high-quality video content increases. . According to a conventional video codec, a video is encoded according to a limited encoding method based on a macro block that has a predetermined size. [003] Image data of a spatial region are transformed into coefficients of a frequency region through frequency transformation. According to a video codec, an image is divided into a block that has a predetermined size, the discrete cosine transformation (in the acronym in English for discrete cosene transformation, DCT) is performed for each respective block, and the frequency coefficients are coded in block units, for quick calculation of frequency transformation. Compared to image data from a spatial region, the coefficients of a frequency region are easily compressed. Particularly, as an image pixel value of a spatial region is expressed according to a prediction error through inter-prediction or intra-prediction of a video codec, when the frequency transformation is performed on the prediction error, a large amount of data can be transformed to 0. According to a video codec, a amount of data can be reduced by replacing the data that is generated consecutively, and repeatedly, with the small size data. Revelation Technical problem [004] The present invention provides an inter-prediction method and apparatus for determining a reference image using a colocalized image, a video encoding method and a video decoding method by means of inter prediction, and a method of video decoding and a video decoding device by means of inter prediction. Technical Solution [005] In accordance with an aspect of the present invention, an inter-prediction method is provided including determining a colocalized block of a current block of a current image from among the blocks of an image that is restored before the current image; preferably check if a first reference list among the reference lists of the colocalized block is consulted and selectively check if a second reference list is consulted according to the fact that the first reference list is consulted; based on a verification result, determine a single reference list colocalized from the first reference list and the second reference list; determine a reference block from the current block using movement information from the colocalized reference list; and perform inter-prediction in the current block using the determined reference block. Advantageous Effects [006] Without checking an entire plurality of reference images included in the reference list of the colocalized block to determine the reference image of the current block, the inter-prediction apparatus can preferably check the first reference list including reference images positioned in a direction opposite to a direction from the current block to the colocalized block in the colocalized image. The inter-prediction apparatus can selectively check the remaining reference lists. Thus, an unnecessary process is ignored in a process to determine the image and reference of the current block using the colocalized block, thus increasing the efficiency of a process of determining the reference image for inter-prediction. Description of Drawings [007] Figure 1 is a block diagram of an inter-prediction apparatus according to an embodiment of the present invention; [008] Figure 2 shows a conventional method of determining a reference image using a colocalized block; [009] Figure 3 shows a method of determining a reference image using a colocalized block, according to an embodiment of the present invention; [0010] Figure 4 is a flow chart of an inter-prediction method according to an embodiment of the present invention; [0011] Figure 5 is a flow chart of a method of encoding video by means of inter-prediction according to an embodiment of the present invention; [0012] Figure 6 is a flow chart of a method of decoding video by means of inter-prediction according to an embodiment of the present invention; [0013] Figure 7 is a block diagram of a video coding apparatus based on a coding unit according to a tree structure, according to an embodiment of the present invention; [0014] Figure 8 is a block diagram of a video decoding apparatus based on a coding unit according to a tree structure, according to an embodiment of the present invention; [0015] Figure 9 is a diagram for describing a concept of coding units according to an embodiment of the present invention; [0016] Figure 10 is a block diagram of an image encoder based on the coding units according to an embodiment of the present invention; [0017] Figure 11 is a block diagram of an image decoder based on the coding units according to an embodiment of the present invention; [0018] Figure 12 is a diagram illustrating the deepest coding units according to the depths, and partitions according to an embodiment of the present invention; [0019] Figure 13 is a diagram for describing a relationship between a coding unit and transformation units, according to an embodiment of the present invention; [0020] Figure 14 is a diagram for describing coding information of the coding units corresponding to a coded depth, according to an embodiment of the present invention; [0021] Figure 15 is a diagram of deeper coding units according to depths, according to an embodiment of the present invention; [0022] Figures 16 to 18 are diagrams to describe a relationship between coding units, prediction units and transformation units, according to an embodiment of the present invention; and [0023] Figure 19 is a diagram to describe a relationship between a coding unit, a prediction unit or a partition, and a transformation unit, according to the coding mode information in Table 1. Best Mode [0024] In accordance with an aspect of the present invention, an inter-prediction method is provided including determining a colocalized block of a current block of a current image, among blocks of an image that is preferably restored before the current image , checking if a first reference list among the reference lists of the colocalized block is consulted and selectively checking if a second reference list is consulted according to the fact that the first reference list is consulted, based on the result of the verification , determining a single colocalized reference list from the first reference list and the second reference list, determining a reference block from the current block, using movement information from the colocalized reference list, and performing inter-prediction on the current block using the given reference block. [0025] The first list of references can include images that are positioned in front of a direction from the current image to the colocalized block. [0026] The selective verification of the second reference list may include, when the first reference list is consulted for the inter prediction, skipping an operation to check whether the second reference list is consulted. [0027] The determination of the colocalized reference list may include, when an image order count number (in the acronym for Picture order counting, POC) of an image of the colocalized block is always less than that of the current image, determine a reference list from the current block as the colocalized reference list. [0028] Selective checking of the second reference list may include checking the first reference list or the second reference list according to whether there is movement information on the first reference list or the second reference list. [0029] In accordance with another aspect of the present invention, an inter-prediction apparatus including a colocalized reference list checking unit is provided to determine a colocalized block of a current block of a current image, among blocks of an image that is restored before the current image and, preferably, checking if the first reference list among the reference lists of the colocalized block is consulted and selectively checking if a second reference list is consulted, according to whether the first reference list is consulted; a reference block determiner to, based on a verification result, determine a single reference list colocalized between the first reference list and the second reference list, and determine a reference block of the current block using information from movement of the colocalized reference list; and an inter-prediction unit for performing inter-prediction on the current block using the given reference block. [0030] In accordance with another aspect of the present invention, a video decoding apparatus is provided that includes an analyzer for performing entropy decoding on a bit sequence obtained by analyzing a received bit stream to restore samples; an inverse transformer for performing reverse quantization and reverse transformation of a quantized transformation coefficient among the restored samples to restore samples; an intra predictor for performing intra prediction in blocks in an intra prediction mode between samples restored by the reverse transformer; and a movement compensator to check, preferably, if the first reference list among reference lists of a colocalized block of the current block, selectively checking if a second reference list is consulted, according to whether the first list of references is consulted. references is consulted, determining a single reference list colocalized between the first reference list and the second reference list based on the result of the verification, and making inter prediction in the current block using a reference block from the current block based on the information of movement of the colocalized reference list, to perform inter prediction in a current block in an inter mode among the samples restored by the inverse transformer, and a restorer to restore an image using blocks that are restored through inter prediction or intra prediction. [0031] In accordance with another aspect of the present invention, a video coding apparatus is provided, including an intra predictor for performing intra block prediction in an intra prediction mode between video blocks, an inter predictor for, preferably, check if a first reference list among reference lists of a colocalized block of the current block is consulted, selectively checking if a second reference list is consulted, according to the fact that the first reference list is consulted, determining a single colocalized reference list between the first reference list and the second reference list based on the result of the verification, and performing inter prediction on the current block using a reference block from the current block based on the movement information of the colocalized reference list, for inter prediction of a current block in an inter mode; a transformation quantizer to perform transformation and quantization as a result of intra prediction or inter prediction, and an output unit for emitting a bit stream generated by performing entropy coding on samples including a quantized transformation coefficient generated as a result of transformation and quantization. [0032] In accordance with another aspect of the present invention, a computer-readable recording medium is provided having a program recorded on it to perform the inter-prediction method. Mode for the Invention [0033] Next, an inter-prediction method and apparatus using a reference list from a colocalized block will be described with reference to figures 1 to 5. A video encoding method and a video decoding apparatus by means of inter prediction will be described with reference to figures 5 and 6. In addition, a video encoding method and a video decoding apparatus by means of inter-prediction based on a coding unit having a tree structure will be described with reference to figures 7 to 19. Next, the term "image" can refer to a still image or a moving image, that is, a video itself. [0034] First, with reference to figures 1 to 4, an inter prediction method and an inter prediction apparatus using a reference list of a colocalized block according to an embodiment of the present invention will be described. In addition, with reference to figures 5 and 6, a method of encoding video and a method of decoding video by means of inter-prediction according to an embodiment of the present invention will be described. [0035] Figure 1 is a block diagram of an inter-prediction apparatus 10 according to an embodiment of the present invention. [0036] Inter-prediction apparatus 10 includes a reference list verification unit 12, a reference block determination unit 14, and an inter-prediction unit 16. [0037] Interpreter 10 encodes each video image for each respective block. A block can have a square shape, a rectangular shape, or any geometric shape and is not limited to a data unit, which has a predetermined size. According to an embodiment of the present invention, a block can be a maximum coding unit, a coding unit, a prediction unit, a transformation unit, or the like, among the coding units according to a tree structure. Methods of encoding and decoding video based on encoding units according to a tree structure will be described below with reference to figures 7 to 19. [0038] The reference list verification unit 12 can determine a colocalized block of a current block of a current image between blocks of an image that is restored before the current image. The colocalized block of the current block of the current image can be determined from the blocks of the image that is restored before the current image, and then a colocalized block positioned in one location of the block in a colocalized image, corresponding to a location of block of the current block in the current image, can be determined. [0039] The reference list verification unit 12 may determine a reference list from the current block, using a reference list from the colocalized block. [0040] The reference list verification unit 12 can check whether a first reference list among reference lists of the colocalized block is preferably consulted. The first list of references according to the present modality can include images that are positioned in a direction opposite to the direction from the current image to the colocalized block in the colocalized block. [0041] The reference list verification unit 12 can selectively check whether a second reference list is consulted, according to whether the first reference list is consulted. When the first reference list is consulted, it does not have to be checked whether the second reference list is consulted. [0042] When the first reference list is consulted for the inter prediction of the colocalized block, the reference list verification unit 12 can skip a process of verifying whether the second reference list is consulted. [0043] The reference list verification unit 12 can check whether the first reference list or the second reference list is consulted, according to whether there is movement information from the first reference list or the second reference list. references. [0044] The reference block determination unit 14 can determine a reference block from the current block according to the result of checking whether the first reference list or the second reference list is consulted. [0045] The reference block determination unit 14 can determine a single reference list colocalized between the first reference list and the second reference list. When the reference block determination unit 14 verifies that the first reference list is capable of being consulted, the reference block determination unit 14 determines that the first reference list is the colocalized reference list. When the reference block determination unit 14 verifies that the second reference list is capable of being consulted, the reference block determination unit 14 determines that the second reference list is the colocalized reference list. [0046] The reference block determination unit 14 can determine the reference block of the current block, using movement information from the colocalized reference list. The colocalized reference image can be determined according to the colocalized reference list. The reference image of the current image can be determined according to the direction and distance of the colocalized image to the colocalized reference image. In addition, the movement information of the current block can be determined by modifying the movement information from the colocalized reference list in proportion to the direction and distance from the colocalized image to the colocalized reference image, and the reference block can be determined in the reference image of the current image according to the modified movement information of the colocalized reference list. [0047] However, when an image order count (POC) number of an image of the colocalized block is always less than that of the current image, the reference block determination unit 14 replaces the colocalized reference list with the list of references for the current block. Thus, the reference image of the current block can be determined according to the reference list of the current block. [0048] The reference block determination unit 14 can determine the reference image of the current block according to the reference list of the current block in a low delay condition to prevent the video encoding from being delayed. For example, when a list 0 and list 1 of the reference list of the current block include the same reference images, that is, in generalized P and B (GPB) mode, the reference image can be determined according to the list of current block references. When a current condition for decoding an image satisfies the low delay condition, the reference block determination unit 14 can determine the reference image of the current block according to the reference list of the current block. [0049] Inter prediction unit 16 can perform inter prediction on the current block using the reference block determined by the reference block determination unit 14. [0050] Interpreter 10 can include a central processor (not shown) to generally control the reference list checking unit 12, the reference block determination unit 14, and the inter prediction unit 16. Alternatively , the reference list verification unit 12, the reference block determination unit 14, and the inter prediction unit 16 can be controlled by respective processors (not shown) and the processors can cooperatively interact with each other, in order to control a global operation between prediction devices 10. Alternatively, the reference list verification unit 12, the reference block determination unit 14, and the inter prediction unit 16 can be controlled according to the control of an external processor (not shown) of the inter-prediction apparatus 10. [0051] Interpreter 10 can include at least one data storage unit (not shown) for storing data that is input and output from reference list verification unit 12 of the determination unit reference block 14, and inter-prediction unit 16. Inter-prediction apparatus 10 may include a controller (not shown) for controlling the input / output of a data storage unit (not shown). [0052] Interpreter 10 can preferably check the first reference list, including reference images positioned in a direction opposite to the direction from the current block to the colocalized block in the colocalized image. The inter-prediction apparatus 10 selectively controls the remaining reference lists, without checking all of a plurality of reference images included in the reference list of the colocalized block, to determine the reference image of the current block. [0053] When the inter-prediction apparatus 10 verifies that the first list of references of the colocalized block is used for the inter-prediction of the colocalized image, since the inter-prediction apparatus 10 can determine the reference image of the current block based on the first list of references from the colocalized block, a process for re-checking if the remaining references from the colocalized block are consulted can be ignored. Thus, an unnecessary process is ignored in a process for determining the reference image of the current block using the colocalized block, thus increasing the efficiency of a process for determining the reference image for inter-prediction. [0054] Figure 2 shows a conventional method of determining a reference image, using a colocalized block. [0055] The reference image of a current block 25 of the current image 20 can be determined with reference to a reference list of a colocalized block 27 of the current block 25. [0056] Reference list indexes can be expressed by list 0 28 and list 1 29. According to a POC order of images 22, 20, 21 and 23, a list of references, including reference images in front of the image current 20 can be expressed by list 0 L0 and reference images, including reference images behind the current image 20 can be expressed by List 1 LI. [0057] A 'colDir' value of a colocalized image 21 of the current block 25 indicates a direction for the colocalized image 21. Since the colocalized image 21 is included in a list 1 26 of the current image of 20, the 'colDir' can be 1. As another example, a 'collocated_from_10_flag' value can be used as a parameter to search for the colocalized image 21. The' collocated_from_10_flag' value can indicate that the colocalized image 21 is a list 0 image of the current image 20. Thus , the value 'collocated_from_10_flag' of the current image 20 can be set to 0. [0058] The colocalized block 27 can be positioned at a block location in the colocalized image 21, which corresponds to a block location of the current block 25 in the current image 20. In a conventional method, a reference image of the current block 25 can be determined by checking whether a list 0 28, and a list 1 29, of a reference list from the colocalized block 27, are consulted. [0059] Typically, the reference image of the current block 25 can be determined from the colocalized block 27, in a reference direction through the current image 20. Since the reference direction through the current image 20 from the block colocalized 27 is a direction in the sense of list 0 28, the reference image of the current block 25 is likely to be positioned in the direction of list 0 28. Thus, conventionally, even if a process of verifying that list 1 29 is consulted is probably unnecessary, it needs to be checked whether list 0 28 and list 1 29 of the reference list of colocalized block 27 are consulted. [0060] Figure 3 shows a method of determining a reference image, using a colocalized block, according to an embodiment of the present invention. [0061] Generally, a reference image of a current block 35 can be determined from a colocalized block 37 in a reference direction using a current image 30. That is, if a colocalized image 31 is included in a list 1 36 of the current block 35, the reference image of the current block 35 is likely to be determined from the colocalized block 37 in a reference direction for a list 0 38 through the current image 30. [0062] If another colocalized image is positioned in the reference direction for list 0 38, the reference image of the current block 35 is likely to be determined from the colocalized image in a reference direction for list 1 36 through the image current 30. [0063] Thus, according to the present modality, to determine the reference image of the current block 35, the inter-prediction apparatus 10 can preferably check whether a single reference list is consulted among the reference lists, that is, the lists of 0 38 and 1 39 of a co-located block 37. Whether a corresponding reference list is consulted can be determined according to whether the co-located block 37 has movement information about the corresponding reference list as a result of the fact that the corresponding reference list was previously consulted during the restoration of the colocalized block 37. [0064] If the reference list that is preferably checked has not been used for the inter prediction of the colocalized block 37, the inter prediction apparatus 10 can check whether the remaining reference list of the colocalized block 37 is consulted. [0065] As described above, the reference list can be determined from the colocalized block 37 in the reference direction using the current image 30. Thus, if the colocalized image 31 is included in list 1 36 of the current block 35, the device Interpretation 10 can check whether list 0 38 is queried from the colocalized block 37 along a direction through the current image 30. When it is determined that list 0 38 is consulted, it does not have to be checked whether a list 1 39 is consulted. However, if the images in list 0 38 of colocalized block 36 are not consulted for the inter prediction, the inter prediction apparatus 10 can simply check whether list 1 39 of the colocalized block 36 is consulted. [0066] Likewise, if a colocalized image of a current block is included in a list 0 of the current block, the inter-prediction apparatus 10, can preferably check if a list 1 of a colocalized block is consulted. [0067] Thus, the inter-prediction apparatus 10 determines a list of references that is subject to an operation, preferably checking whether the reference list is consulted, among the reference lists of a colocalized block, based on a reference direction. from a current block to a colocalized image. [0068] That is, the inter-prediction device 10 determines a direction for a reference list that is subject to an operation preferably to check if the reference list is consulted, among reference lists of a colocalized block, as an opposite direction to the reference direction from the current block to the colocalized image. Thus, if the colocalized image is an image of a list 0 of the current image, if a list 1 of the colocalized block is consulted it can preferably be checked. If the image is a colocalized image from list 1 of the current image, if list 0 of the colocalized block is consulted it can preferably be checked. [0069] For example, a reference list, which is subject to an operation preferably to check if the reference list is consulted between reference lists of the colocalized block, can be determined opposite to a reference direction of the current block for the image colocalized. Thus, when the reference direction of the current block for the colocalized image is expressed by 'colDir', the inter-prediction device 10 determines a reference list that is subject to an operation, preferably checking whether the reference list is consulted along of '1-colDir', among reference lists of the colocalized block. [0070] As another example, when an image is colocalized from a list 0 image of a current image, a 'collocated f rom__10_f lag' value of a current block is 1. When the colocalized image is a list 1 image of the current image, the value 'collocated_from_10_flag'is 0. Thus, the prediction device 10 determines a direction for a reference list that is subject to an operation, preferably checking if the reference list is consulted between reference lists in the block colocalized according to the 'collocated from_10_flag' value of the current block. [0071] Thus, prediction apparatus 10 can determine the reference block of the current block, using movement information from a colocalized reference list that is selected based on whether the first reference list is consulted. [0072] However, in a condition of little delay, the inter-prediction device 10 determines the reference image of the current block based on the reference list of the current block, instead of a reference list of the colocalized block. For example, when a POC number of an image in the colocalized block is always less than that of the current image, or when a predetermined condition, including a GPB prediction mode, is satisfied, in which lists 0 and 1 of reference lists of the current block include the same reference images, an image is decoded in the condition of short delay. In the condition of a short delay, the inter-prediction device 10 replaces the colocalized reference list with the reference list of the current block and then can determine the reference block of the current block, using movement information from the colocalized reference list. [0073] Figure 4 is a flow chart of an inter-prediction method according to an embodiment of the present invention. [0074] In operation 41, a colocalized block of a current block of a current image is determined, among the blocks of an image that is restored before the current image. [0075] In operation 42, a first reference list, preferably consulted among reference lists of the colocalized block, is checked, and a second consulted reference list is checked according to whether the first reference list is consulted. [0076] According to the present modality, the first list of references may include images that are positioned opposite a direction from the current image for the colocalized block. When the first reference list is consulted for the inter prediction of the colocalized block, a process of checking whether the second reference list is consulted can be ignored. [0077] In operation 43, based on the result of verification of operation 42, a single colocalized reference list is determined from the first reference list and the second reference list. When a video is decoded in the condition of a short delay, the reference list of the current block is determined as a colocalized reference list; and a reference image can be determined according to the reference list of the current block. [0078] In operation 44, a reference block of the current block is determined by means of movement information from the colocalized reference list. In operation 45, inter prediction is performed in the current block using the reference block determined in operation 44. [0079] Thus, in the method of determining a reference image for the inter-prediction according to the present modality, if it is found that the first list of references of the colocalized block is used for the inter-prediction of the colocalized image, an unnecessary process for further verification of whether the remaining reference lists of the colocalized block are consulted can be ignored, thus increasing the efficiency of the inter prediction. [0080] Figure 5 is a flow chart of a video encoding method by means of inter-prediction according to an embodiment of the present invention. [0081] In operation 51, intra prediction is performed in blocks in an intra prediction mode between blocks of a video. [0082] In operation 52, it is verified whether a first reference list among reference lists of a colocalized block of a current block is preferably consulted for the inter prediction of the current block in the inter mode. The first reference list can include images that are positioned in a direction opposite to a direction from the current image for the colocalized block. [0083] When the first list of references is able to be consulted, it does not have to be checked whether a second list of references is consulted. When the first list of references is not consulted, if the second list of references is consulted it can be checked. Based on the result of the check, a single colocalized reference list can be determined between the first reference list and the second reference list and a reference block from the current block can be determined based on the movement information of the colocalized reference list. Inter-prediction can be performed on the current block using the reference block of the current block to generate a residual value. [0084] In operation 53, transformation and quantization are performed in the sequence of intra prediction or inter prediction to generate a quantized transformation coefficient. In operation 55, a bit stream generated by performing entropy coding on samples, including the quantized transformation coefficient of operation 53 is output. A 'colDir' parameter, indicating a direction for the colocalized image of the current block or a 'collocated_from_10_flag' parameter indicating whether the current image of the colocalized image is a list 0 image can be transmitted. [0085] In addition, during operation prediction 52, when an image is restored in a delay condition, a reference image can be determined according to the reference list of the current block, regardless of the colocalized reference list. [0086] A video encoding apparatus which performs the video encoding method of figure 5 may include the inter-prediction apparatus 10 according to an embodiment of the present invention. The video encoding apparatus, including the inter prediction apparatus 10 can perform intra prediction, inter prediction, transformation, and quantization for each image block to generate samples and can perform entropy encoding on the samples to generate a bit stream. In the video encoding apparatus, including the inter-prediction apparatus 10, the inter-prediction apparatus 10 can interact with a video encoding processor or an external video encoding processor, which is mounted on the video encoding apparatus to performing a video encoding operation, including transformation, to produce a video encoding result. According to an embodiment of the present invention, in an internal video encoding processor of the video encoding apparatus, since a video encoding apparatus, a central processing device, or a graphics processing apparatus may include a video encoding module, as well as a separate processor, a basic video encoding operation can be performed. [0087] Figure 6 is a flow chart of a method of video decoding by means of inter prediction according to a modality of the present invention. [0088] In operation 61, entropy decoding is performed on a bit sequence obtained by analyzing a received bit stream to restore samples. In operation 62, inverse quantization and inverse transformation are performed on a transformation coefficient quantized from the samples to restore the samples. In operation 63, intra prediction is performed on samples in an intra mode. In operation 64, motion compensation is performed on samples in an inter mode. In operation 65, an image is restored using blocks that are restored through the intra prediction of operation 63 or the motion compensation of operation 64. [0089] In operation 64, a colocalized block of a current block is determined from the samples, for the prediction of a current block in the inter mode. A 'colDir' parameter, indicating a direction for the colocalized image of the current block or a 'collocated_from_10_flag' parameter indicating whether the current image of the colocalized image is a list 0 image can be analyzed from a bit stream and restored. The colocalized block of the current block can be determined based on the 'colDir' parameter or the 'collocated_from_10_flag' parameter. [0090] If a first list of references among lists of references in the colocalized block is consulted, it is preferably verified. The first reference list can include images that are positioned in a direction opposite to a direction from the current image for the colocalized block. [0091] When the first list of references is able to be consulted, it is not necessary to check whether a second list of references is consulted. When the first list of references is not consulted, if the second list of references is consulted it can be checked. Based on the result of the check, a single colocalized reference list can be determined between the first reference list and the second reference list and a reference block from the current block can be determined based on the movement information of the colocalized reference list. Motion compensation for the current block can be performed on the current block using the reference block of the current block to generate a pixel value from the sample block. [0092] In addition, during the motion compensation operation 63, when an image is restored to a delay condition, a reference image can be determined according to a reference list of the current block, regardless of the colocalized reference list . [0093] A video decoding apparatus that performs the video decoding method of figure 6 may include the inter-prediction apparatus 10, according to an embodiment of the present invention. The video decoding apparatus, including the inter-prediction apparatus 10, can analyze samples obtained by encoding a bit stream and can perform reverse quantization, reverse transformation, intra prediction and motion compensation for each image block to restore samples. In the video decoding apparatus, the inter-prediction apparatus 10 may interact with a video encoding processor or an external video encoding processor, which is mounted on the video decoding apparatus to perform a video decoding operation, including reverse transformation or prediction / compensation, in order to produce the result of video decoding. In accordance with an embodiment of the present invention, in an internal decoding video processor or video decoding apparatus, since a video decoding apparatus, a central processing device, or a graphics processing apparatus may include a video encoding module, as well as a separate processor, a basic video decoding operation can be performed. [0094] In the inter-prediction apparatus 10, blocks obtained by dividing the video data are divided into coding units having a tree structure and prediction units are used for the inter-prediction of the coding units, as described above. Next, with reference to figures 7 to 19, a method and apparatus for encoding a video, and a method and apparatus for decoding video, based on a coding unit having a tree structure, and a unit of coding, will be described. [0095] Figure 7 is a block diagram of a video coding apparatus 100 based on a coding unit according to a tree structure, according to an embodiment of the present invention. [0096] The video coding apparatus 100 by means of video prediction based on a coding unit according to a tree structure comprises a maximum coding unit divider 110, a coding unit determiner 120, and a coding unit output 130. Next, for convenience of description, the video encoding apparatus 100 by means of video prediction based on an encoding unit, according to a tree structure, is referred to as 'the video encoding apparatus 100' . [0097] The maximum encoding unit divider 110 can divide a current image based on a maximum encoding unit for the current image of an image. The current image is larger than the maximum encoding unit, the image data of the current image can be divided into at least one maximum encoding unit. The maximum coding unit according to an exemplary embodiment can be a data unit having a size of 32x32, 64x64, 128x128, 256x256, etc., wherein a format of the data unit is a square having a width and height in squares of 2. Image data can be output to the encoding unit determinator 120 according to at least one maximum encoding unit. [0098] A coding unit according to an exemplary modality can be characterized by a maximum size and depth. Depth denotes a number of times that the coding unit is spatially divided from the maximum coding unit and, as the depth is deepened or increased, deeper coding units, according to the depths, can be divided from the maximum coding unit to a minimum coding unit. A maximum coding depth is a higher depth and a minimum coding depth is a lower depth. As the size of a coding unit corresponding to each depth decreases as the depth of the maximum coding unit is deepened, a coding unit corresponding to a higher depth may include a plurality of coding units corresponding to the lower depths. [0099] As described above, the image data of the current image is divided into the maximum encoding units according to a maximum encoding unit size, and each of the maximum encoding units can include deeper encoding units that are divided according to the depths. As the maximum coding unit according to an exemplary modality is divided according to the depths, the image data of a spatial domain included in the maximum coding unit can be classified hierarchically according to the depths. [00100] The maximum depth and maximum size of a coding unit, which limit the total number of times the height and width of the maximum coding unit are divided hierarchically, can be predetermined. [00101] The coding unit determinator 120 encodes at least one divided region obtained by dividing a region of the maximum coding unit according to the depths, and determines a depth to produce the finally encoded image data according to the hair least one divided region. In other words, the encoding unit determiner 120 determines an encoded depth by encoding the image data in the deepest encoding units according to the depths, according to the maximum encoding unit of the current image, and selecting a depth having the minimum coding error. The determined coded depth and the image data encoded according to the determined coded depth are output to output unit 130. [00102] The image data in the maximum encoding unit is encoded based on the deepest encoding units corresponding to at least a depth equal to or below the maximum depth, and the results of encoding the image data are compared based on each one of the deepest coding units. A depth having the minimum coding error can be selected after comparing the coding errors of the deepest coding units. At least one coded depth can be selected for each maximum coding unit. [00103] The size of the maximum coding unit is divided as a coding unit is divided hierarchically according to depths, and as the number of coding units increases. In addition, even if the coding units correspond to the same depth in a maximum coding unit, each coding unit corresponding to the same depth can be divided to a lower depth by measuring a coding error in the image data of each coding unit separately. Consequently, even when image data is included in a maximum coding unit, the image data is divided into regions according to depths and coding errors can differ according to regions in a maximum coding unit, and so the coded depths may differ according to the regions in the image data. Thus, one or more coded depths can be determined in a maximum coding unit, and the image data from the maximum coding unit can be divided according to the coding units of at least one coded depth. [00104] Consequently, the coding unit determinator 120 can determine coding units having a tree structure included in the maximum coding unit. The 'coding units having a tree structure' according to an embodiment of the present invention include coding units corresponding to a depth determined to be the coded depth, among all the deepest coding units included in the maximum coding unit. A coding unit of a coded depth can be determined hierarchically according to the depths in the same region as the maximum coding unit, and can be determined independently in different regions. Similarly, a coded depth in one current region can be determined independently from a coded depth in another region. [00105] A maximum depth according to an embodiment of the present invention is an index related to the number of times the division is performed from a maximum coding unit to a minimum coding unit. A first maximum depth according to an embodiment of the present invention can denote the total number of times that the division is performed from the maximum coding unit to the minimum coding unit. A second maximum depth according to an embodiment of the present invention can denote the total number of depth levels from the maximum coding unit to the minimum coding unit. For example, when a maximum coding unit depth is 0, a coding unit depth, in which the maximum coding unit is divided once, can be set to 1, and a depth of a coding unit, in which the maximum coding unit divided twice, can be set to 2. Here, if the minimum coding unit is a coding unit in which the maximum coding unit is divided four times, 5 levels deep, 0 depth , 1, 2, 3 and 4 exist, so the first maximum depth can be adjusted to 4, and the second maximum depth can be adjusted to 5. [00106] Prediction coding, and transformation, can be performed according to the maximum coding unit. Prediction coding and transformation are also performed based on the deepest coding units according to a depth equal to or less than the maximum depth, according to the maximum coding unit. The transformation can be performed according to the orthogonal transform or integer transform method. [00107] As the number of deepest coding units increases whenever the maximum coding unit is divided according to depths, coding including prediction coding and transformation is performed on all the deepest coding units generated by as the depth increases. For convenience of description, prediction coding and transformation will now be described based on the current depth coding unit, in a maximum coding unit. [00108] The video encoding apparatus 100 can select in a varied way a size or format of a data unit to encode the image data. To encode image data, operations such as prediction, transformation, and entropy coding are performed, and this time, the same data unit can be used for all different operations or data units can be used for each operation. [00109] For example, the video encoding apparatus 100 may select not only a coding unit to encode the image data, but also a data unit other than the encoding unit to perform the prediction encoding on the image data in the encoding unit. In order to perform prediction coding in the maximum coding unit, the prediction coding can be carried out based on a coding unit corresponding to a coded depth, that is, based on a coding unit that is no longer divided into coding units corresponding to a lower depth. Then, the coding unit that is no longer divided and becomes a basic unit for prediction coding will now be referred to as a 'prediction unit'. A partition obtained by dividing the prediction unit can include a prediction unit or a data unit obtained by dividing at least one of a height and a width of the prediction unit. The partition is a data unit obtained by dividing the prediction unit from the coding unit and the prediction unit can be a partition the same size as the coding unit. [00110] For example, when a 2Nx2N coding unit (where N is a positive integer) is no longer divided and becomes a 2Nx2N prediction unit, a partition size can be 2Nx2N, 2NxN, Nx2N, or NxN. Examples of a partition type include symmetric partitions that are obtained by symmetrically dividing a height or width of the prediction unit, partitions obtained by asymmetrically dividing the height or width of the prediction unit, such as l: n or n: l, partitions which are obtained by geometric division of the prediction unit, and partitions having arbitrary shapes. [00111] A prediction mode of the prediction unit can be at least one in an intra mode, an inter mode, and a jump mode. For example, the intra mode or the inter mode can be performed on the 2Nx2N, 2NxN, Nx2N or NxN partition. In addition, the jump mode can only be performed on the 2Nx2N partition. Coding is carried out independently in a prediction unit in an encoding unit, thereby selecting a prediction mode having a minimal coding error. [00112] The video encoding apparatus 100 can also perform the transformation in the image data in a coding unit based not only on the coding unit to encode the image data, but also based on a data unit which is different from the encoding unit. To perform the transformation on the coding unit, the transformation can be performed on the basis of a transformation unit having a size less than or equal to that of the coding unit. Examples of a transformation unit may include a data unit for an intra mode and a data unit for an inter mode. [00113] Like the coding unit according to the tree structure according to the present modality, the transformation unit in the coding unit can be divided recursively into smaller regions and residual data in the unit coding can be divided according to the transformation having the tree structure according to the transformation depths. [00114] According to an embodiment of the present invention, a depth of transformation indicating the number of times the division is carried out to reach the transformation unit by dividing the height and width of the coding unit, can also be established in the unit transformation. For example, in a current coding unit of 2Nx2N, a transformation depth can be set to 0. When the size of a transformation unit is also 2Nx2N, a transformation depth can be set to 1. In addition, when the size of the transformation unit is NxN, the transformation depth can be set to 2. That is, the transformation unit according to the tree structure can also be established according to the transformation depth. [00115] Coding information according to the coded depths requires not only information about the coded depths, but also information about prediction and information about transformation. Therefore, the coding unit determiner 120 can not only determine a coding depth that has at least one coding error, but also determine a partition type in a prediction unit, a prediction mode according to the units of prediction, and a size of a transformation unit for the transformation. [00116] Coding units and a prediction / partition unit according to a tree structure, in a maximum coding unit, and a method of determining a transformation unit, according to the modalities of the present invention, will be described in detail later with reference to figures 7 to 19. [00117] The coding unit determinator 120 can measure a coding error of deeper coding units according to the depths using Rate Distortion Optimization based on Lagrangian multipliers. [00118] Output unit 130 produces the image data of the maximum coding unit, which is coded based on at least one coded depth determined by coding unit determiner 120, and information on the coding mode according to coded depth, in bit streams. [00119] The encoded image data can be obtained by encoding residual image data. [00120] Information about the coding mode according to the coded depth can include information about the coded depth, the type of partition in the prediction unit, the prediction mode, and the size of the transformation unit. [00121] The information about the coded depth can be defined by the use of information divided according to the depths, which represents whether the coding is performed in coding units of a lower depth instead of a current depth. If the current depth of the current coding unit is the coded depth, the image data in the current coding unit is encoded and produced, so the split information can be set not to divide the current coding unit to a lesser depth. Alternatively, if the current depth of the current coding unit is not the coded depth, coding is carried out in the coding unit of the lower depth, and so the divided information can be set to divide the current coding unit to obtain the coding units the bottom depth. [00122] If the current depth is not the coded depth, the coding is carried out in the coding unit which is divided into the coding unit of the lower depth. As there is at least one coding unit for the bottom depth in a coding unit for the current depth, coding is performed repeatedly on each coding unit for the bottom depth, and so coding can be performed recursively for the coding units having the same depth. [00123] As the coding units having a tree structure are determined for a maximum coding unit, and information about at least one coding mode is determined for a coding unit of a coded depth, information about at least one mode encoding can be determined for a maximum encoding unit. In addition, an encoded depth of the image data of the maximum encoding unit can be different according to the locations, since the image data is divided hierarchically according to the depths and thus the information about the encoded path, and the encoding mode, can be adjusted for the image data. Consequently, output unit 130 can assign encoding information about a corresponding encoded path and encoding mode for at least one of the encoding unit, prediction unit, and an encoding mode for at least one of the encoding unit. coding, prediction unit, and a minimum unit included in the maximum coding unit. [00125] The minimum unit according to one embodiment of the present invention is a rectangular data unit obtained by dividing the minimum coding unit constituting the lowest depth by 4. Alternatively, the minimum unit can be a maximum rectangular data unit that can be included in all coding units, prediction units, partition units and transformation units included in the maximum coding unit. [00126] For example, the coding information produced through the output unit 130 can be classified into coding information according to the coding units, and coding information according to the prediction units. The encoding information according to the encoding units can include information about the prediction mode and the size of the partitions. The coding information according to the prediction units may include information about an estimated direction in an inter mode, in relation to a reference image index of the inter mode, in relation to a motion vector, in relation to a chroma component. in an intra mode, and in relation to an intra mode interpolation method. [00127] In addition, information about the maximum encoding unit size defined according to images, slices, or GOPs, and information about the maximum depth can be inserted in a header of a bit stream, an SPS (set of sequence parameters), or a set of image parameters (PPS). [00128] In addition, information about the size of a maximum and minimum size of a processing unit available for current video can be transmitted via a bit stream header, an SPS, a set of image parameters or the like . Output unit 130 can encode and output reference information, prediction information, unidirectional prediction information, and information about the types of slice including a fourth type of slice, which are related to the prediction, as described above with reference to figures 1 to 6. [00129] In the video coding apparatus 100, the deepest coding unit can be a coding unit obtained by dividing a height or width of a coding unit of a greater depth, which is one layer above, by two. In other words, when the size of the current depth coding unit is 2Nx2N, the size of the bottom depth coding unit is NxN. In addition, the current depth coding unit having the size of 2Nx2N can include a maximum of 4 of the lower depth coding units. [00130] Consequently, the video encoding apparatus 10 can form the encoding units having the tree structure upon determining the encoding units having an optimal format and an optimum size for each maximum encoding unit, based on the size of the unit maximum coding and maximum depth determined considering the characteristics of the current image. In addition, since coding can be performed on each maximum coding unit using any of the various prediction and transformation modes, an optimal coding mode can be determined by considering the characteristics of the coding unit of various image sizes. [00131] Thus, if an image having high resolution or a large amount of data is encoded in a conventional macro block, a number of macro blocks per image increases excessively. Consequently, a number of pieces of compressed information generated for each macro block increases, and thus it is difficult to transmit the compressed information and decreases the efficiency of data compression. However, by using the video coding apparatus 100, the image compression efficiency can be increased once a coding unit is adjusted while considering the characteristics of an image while increasing the maximum size of a coding unit while considering a image size. [00132] The video coding apparatus 100 of figure 7 can perform the operation of the inter-prediction apparatus 10, as described with reference to figure 1. [00133] The coding unit determinator 120 can perform an operation of the inter prediction apparatus 10. For each maximum coding unit, a prediction unit for the inter prediction can be determined in coding units according to a tree structure and inter prediction can be performed in prediction units. [00134] In particular, if a first reference list is consulted among reference lists of a colocalized block of a current block, it is preferably verified, for the inter prediction of a current prediction unit in a prediction mode. The first reference list can include images that are positioned in a direction opposite to a direction from the current image for the colocalized block. [00135] When the first list of references is able to be consulted, it is not necessary to check whether a second list of references is consulted. When the first list of references is not consulted, if the second list of references is consulted it can be checked. Based on the result of the check, a single colocalized reference list can be determined between the first reference list and the second reference list and a reference block from a current prediction unit can be determined based on the movement information from the list of references. colocalized references. Inter prediction can be performed on the current prediction unit using the reference block of the current prediction unit to generate a residual value. A 'collocated_from_10_flag' parameter or a 'colDir' parameter, indicating a colocalized block of the current prediction unit can be transmitted. [00136] Figure 8 is a block diagram of a video decoding apparatus 200 based on a coding unit according to a tree structure, according to an embodiment of the present invention. [00137] The video decoding apparatus 200 based on the encoding unit according to the tree structure includes a receiver 210, extractor of encoding information and image data 220, and a decoder of image data 230. Next For convenience of description, the video decoding apparatus 200 through video prediction based on a coding unit according to a tree structure will be referred to as a 'video decoding apparatus 200'. [00138] The definitions of various terms, such as an encoding unit, a depth, a prediction unit, a transformation unit, and information on the various encoding modes, for various operations of the video decoding apparatus 200 are identical to those described above with reference to figure 7 and the video encoding apparatus 100. [00139] Receiver 210 receives and analyzes a bit stream from an encoded video. The encoder information and image data extractor 220 extracts the encoded image data for each encoding unit from the analyzed bit stream, where the encoding units have a tree structure according to each maximum encoding unit, and produces the extracted image data for the decoder 230. The coding information and image data extractor 220 can extract information about a maximum size of a coding unit from a current image, from a header over a current image, an SPS, or a PPS. [00140] In addition, encoder information and image data extractor 220 extracts information about an encoded depth and an encoding mode for the encoding units having a tree structure according to each maximum encoding unit from the flow of analyzed bits. The information extracted about an encoded depth and an encoding mode is output to the decoder 230. In other words, the image data in a bit stream is divided into the maximum encoding unit so that the image data decoder 230 decodes the image data for each maximum encoding unit. [00141] Information about the coded depth and the coding mode according to the maximum coding unit can be established for information about at least one coding unit corresponding to the coded depth, and information about a coding mode can include information about a partition type of a corresponding coding unit that corresponds to the coded depth, in relation to the prediction mode, and to a size of a transformation unit. In addition, the split information according to depths can be extracted as the coded depth information. [00142] The information about the coded depth and the coding mode according to each maximum coding unit extracted by the coding information and image data extractor 220 is information about a coded depth and a coding mode determined to generate an error minimum encoding when an encoder, such as video encoding apparatus 100, repeatedly encodes for each deepest encoding unit according to the depths in accordance with each maximum encoding unit. Consequently, the video decoding apparatus 200 can reconstruct an image by decoding the image data according to an encoded depth and an encoding mode that generates the minimum encoding error. [00143] As the encoding information about the encoded depth, and the encoding mode, can be assigned to a predetermined data unit among a corresponding encoding unit, a prediction unit, and a minimal unit, the information extractor of encoding and image data 220 can extract information about encoded depth and encoding mode according to predetermined data units. The predetermined data units to which the same information about the coded depth and the coding mode is assigned can be deduced as being the data units included in the same maximum coding unit. [00144] The image data decoder 230 restores the image by decoding the image data in each of the maximum encoding units based on information about the encoded depth and the encoding mode according to the maximum encoding units. In other words, the decoder 230 image data can decode the encoded image data based on the information extracted about the partition type, the prediction mode, and the transformation unit for each encoding unit among the encoding units having the tree structure included in each maximum coding unit. A decoding process can include a prediction including intra prediction and motion compensation, and a reverse transformation. [00145] image data decoder 230 can perform intra prediction or motion compensation according to a partition and a prediction mode of each encoding unit, based on information on the partition type and prediction mode of the drive prediction of the coding unit according to the coded depths. [00146] In addition, to perform the reverse transformation on the units of maximum coding units, the image data decoder 230 can read the information about the transformation units having a tree structure, in units of coding units and perform the reverse transformation in units of the coding units, based on the transformation units. When performing the reverse transformation, the pixel values of the coding units of a spatial domain can be restored. O [00147] image data decoder 230 can determine at least one coded depth of a maximum coding unit using information divided according to the depths. If the split information represents that image data is no longer divided at the current depth, the current depth is the coded depth. Consequently, the decoder 230 can decode the encoded image data from at least one encoding unit corresponding to each encoded depth in the current maximum encoding unit using information about the partition type of the prediction unit, the prediction mode, and the size of the processing unit for each coding unit corresponding to the coded depth, and outputting the image data from the current maximum coding unit. [00148] In other words, data units containing the coding information including the same division information can be grouped by observing the set of coding information assigned to the predetermined data unit among the coding unit, the prediction unit and the minimum unit, and the grouped data units can be considered to be a data unit to be decoded by decoder 230 in the same encoding mode. For each encoding unit determined as described above, information about the encoding mode can be obtained in order to decode the current encoding unit. [00149] The image data decoder 230 of the video decoding apparatus 200 of figure 8 can perform the operation of the inter-prediction apparatus 10, as described above with reference to figure 1. [00150] The image data decoder 230 can determine a prediction unit for the inter prediction for each coding unit according to a tree structure and can perform inter prediction for each prediction unit, for a maximum coding unit. [00151] In particular, a colocalized block of a current block is determined from restored samples, for the prediction of a current block in an inter mode. A colocalized block of a current prediction unit can be determined based on a 'collocated_from_10_flag' parameter or a 'colDir' parameter which is a current prediction unit obtained by analyzing a bit stream. [00152] If a first reference list is consulted between reference lists of the colocalized block, it is preferably verified. The first list of references can include images that are positioned inside and in a direction opposite to a direction from the current image to the colocalized block. [00153] When the first list of references is able to be consulted, it is not necessary to check whether a second list of references is consulted. When the first list of references is not consulted, if the second list of references is consulted it can be checked. Based on the result of the check, a single colocalized reference list can be determined between the first reference list and the second reference list and a reference block of the current prediction unit can be determined based on the movement information of the reference list colocalized. Motion compensation can be performed with the current prediction unit using the reference block of the current prediction unit to generate a pixel value from the sample block. [00154] In addition, when an image is restored in a short delay condition, a reference image can be determined according to a reference list of the current prediction unit, regardless of the colocalized reference list. [00155] The video decoding apparatus 200 can obtain information about at least one encoding unit that generates the minimum encoding error when encoding is performed recursively for each maximum encoding unit, and can use the information to decode the current image. In other words, coding units having the tree structure determined to be the optimal coding unit in each maximum coding unit can be decoded. In addition, the maximum size of the encoding unit is determined by considering the resolution and amount of image data. [00156] Consequently, even if the image data has high resolution and a large amount of data, the image data can be efficiently decoded and restored using a size of an encoding unit and an encoding mode, which they are determined adaptively according to the characteristics of the image data, using information about an optimal encoding mode received from an encoder. [00157] Figure 9 is a diagram to describe a concept of the coding units according to an embodiment of the present invention. [00158] One size of a coding unit can be expressed in width x height, and can be 64x64, 32x32, 16x16 and 8x8. A 64x64 encoding unit can be divided into 64x64, 64x32, 32x64 or 32x32 partitions, and a 32x32 encoding unit can be divided into 32x32, 32x16, 16x32, or 16x16 partitions, a 16x16 encoding unit can be divided into 16x16, 16x8, 8x16 or 8x8 partitions, and an 8x8 encoding unit can be divided into 8x8, 8x4, 4x8 or 4x4 partitions. [00159] In video data 310, a resolution is 1920x1080, a maximum size of an encoding unit is 64, and a maximum depth is 2. In video data 320, a resolution is 1920x1080, a maximum size encoding is 64, and a maximum depth is 3. In video data 330, a resolution is 352x288, a maximum encoding unit size is 16, and a maximum depth is 1. The maximum depth shown in Figure 9 denotes a total number of divisions from a maximum coding unit to a minimum decoding unit. [00160] If a resolution is high or if the amount of data is large, a maximum size of an encoding unit can be so large not only to increase the efficiency of encoding, but also to accurately reflect the characteristics of an image. Consequently, the maximum encoding unit size of video data 310 and 320 having the highest resolution than video data 330 can be 64. [00161] Since the maximum depth of video data 310 is 2, the encoding units 315 of video data 310 may include a maximum encoding unit having a long axis size of 64, and encoding units having axis sizes 32 and 16 long since the depths are deepened for two layers by dividing the maximum coding unit twice. However, as the maximum depth of video data 330 is 1, the coding units 335 of video data 330 may include a maximum coding unit having a long axis size of 16, and coding units having an axis size. over 8 since the depths are deepened to a layer by dividing once of the maximum coding unit. [00162] As the maximum depth of video data 320 is 3, encoding units 325 of video data 320 may include a maximum encoding unit having a long axis size of 64, and encoding units having axis sizes over 32, 16 and 8 since the depths are deepened to 3 layers by dividing the maximum coding unit three times. As the depth is deepened, detailed information can be accurately expressed. [00163] Figure 10 is a block diagram of an image encoder 400 based on the coding units, according to an embodiment of the present invention. [00164] The image encoder 400 performs operations of the encoding unit determinator 120 of the video encoding equipment 100 to encode the image data. In other words, an intra-410 predictor performs intra-prediction in coding units in an intra mode, within a current image 405, and a motion estimator 420 and a motion compensator 425 perform inter estimation and motion compensation in the coding units in an inter mode between the current image 405 using the current image 405, and a reference image 495. [00165] Data emitted from the intra predictor 410, the motion estimator 420, and the motion compensator 425 are output as a transformation coefficient quantized through transformer 430 and a quantizer 440. The quantized transformation coefficient is restored as data in a spatial domain through an inverse quantizer 460 and an inverse transformer 470, and the data restored in the spatial domain is output as the reference image 495 after being post-processed through an unlocking unit 480 and a filtering unit of loop 490. The quantized transformation coefficient can be output as a bit stream 455 through an entropy encoder 450. [00166] For the image encoder 400 to be applied to the video encoding equipment 100, all elements of the image encoder 400, that is, the intra predictor 410, the motion estimator 420, the motion compensator 425, the transformer 430, quantizer 440, entropy encoder 450, inverse quantizer 460, inverse transformer 470, unlocking unit 480, and loop filtering unit 490 perform operations based on each coding unit among coding units having a tree structure while considering the maximum depth of each maximum coding unit. [00167] Specifically, the intra predictor 410, the motion estimator 420, and the motion compensator 425 determine partitions and a prediction mode for each coding unit among the coding units having a tree structure while considering the maximum size and the maximum depth of a current maximum coding unit, and transformer 430 determines the size of the transformation unit in each coding unit among the coding units having a tree structure. [00168] Specifically, to determine a reference image for the inter-prediction of a current prediction unit, the movement compensator 425 preferably checks whether the first reference list of a colocalized block is consulted, and it is not verified again whether the remaining reference lists of the colocalized block are consulted when there is information about the movement of the first reference list since the first reference list of the colocalized block is preferably consulted. However, when there is no movement information from the first reference list since the first reference list in the colocalized block is not consulted, the movement compensator 425 can check again whether the remaining reference lists in the colocalized block are consulted. The motion compensator 425 can determine a reference list from the current prediction unit using the reference list of the colocalized block in which the verification operation has been performed. [00169] Figure 11 is a block diagram of an image decoder 500 based on the coding units, according to an embodiment of the present invention. [00170] An analyzer 510 analyzes the encoded image data to be decoded and encoding information required for decoding from a 505 bit stream. The encoded image data is output as inverse quantized data via an entropy decoder 520 and an inverse quantizer 530, and inverse quantized data are restored to image data in a spatial domain through an inverse transformer 540. [00171] An intra predictor 550 performs intra prediction in the coding units in an intra mode with respect to image data in the spatial domain, and a motion compensator 560 performs motion compensation in the coding units in an inter mode, using a 585 reference image. [00172] Image data in the spatial domain, which has passed through the intra predictor 550 and the motion compensator 560, can be output as a restored image 595 after being post-processed through an unlocking unit 570 and a loop filtration 580. In addition, the image data that is post-processed via the unlocking unit 570 and the loop filtration unit 580 can be output as the reference image 585. [00173] To decode the image data in the decoder 230 of the video decoding apparatus 200, the image decoder 500 can perform the operations that are performed after the analyzer 510 performs an operation. [00174] For the image decoder 500 to be applied to the video decoding equipment 200, all elements of the image decoder 500, that is, the analyzer 510, the entropy decoder 520, the inverse quantizer 530, the inverse transformer 540, the intra predictor 550, the motion compensator 560, the unlocking unit 570, and the loop filtering unit 580 perform operations based on the coding units having a tree structure for each maximum coding unit. [00175] Specifically, intra prediction 550 and motion compensator 560 perform operations based on partitions and in a prediction mode for each of the coding units having a tree structure, and the reverse transformer 540 performs operations based on one size of a processing unit for each coding unit. [00176] Specifically, to determine a reference image for the inter-prediction of a current prediction unit, the motion compensator 560 preferably checks whether the first reference list of a colocalized block is consulted, and it is not verified again whether the remaining reference lists of the colocalized block are consulted when there is information about the movement of the first reference list since the first reference list of the colocalized block is preferably consulted. However, when there is no movement information for the first reference list, since the first reference list of the colocalized block is not consulted, the movement compensator 560 can check again whether the remaining reference lists of the colocalized block are consulted. . The motion compensator 560 can determine a reference list from the current prediction unit using the reference list of the colocalized block in which the verification operation has been performed. [00177] Figure 12 is a diagram illustrating units of deeper coding according to depths, and partitions, according to an embodiment of the present invention. [00178] The video encoding apparatus 100 and the video decoding apparatus 200 use hierarchical encoding units to consider the characteristics of an image. The maximum height, maximum width, and maximum depth of the coding units can be determined adaptively according to the characteristics of the image, or they can be adjusted differently by a user. Sizes of deeper coding units, according to depths, can be determined according to the maximum predetermined size of the coding unit. [00179] In a hierarchical structure 600 of the coding units, according to an exemplary modality, the maximum height and the maximum width of the coding units are individually 64, and the maximum depth is 4. As a depth is deepened to the along a vertical axis of the hierarchical structure 600, the height and width of the deepest coding unit are individually divided. In addition, a prediction unit and partitions, which are the bases for prediction coding for each deeper coding unit, are shown along a horizontal axis of hierarchical structure 600. [00180] In other words, a coding unit 610 is a maximum coding unit in hierarchical structure 600, in which a depth is 0 and a size, that is, a height by width, is 64x64. The depth is deepened along the vertical axis, and there is a coding unit 620, having a size of 32x32, and a depth of 1; a coding unit 630, having a size of 16x16, and a depth of 2; and there is a coding unit 640, having a size of 8x8, and a depth of 3. The coding unit 640 having a size of 8x8 and depth of 3 is a minimum coding unit. [00181] The prediction unit and partitions of a coding unit are arranged along the horizontal axis according to each depth. In other words, if the encoding unit 610, having the size of 64x64, and the depth of 0, is a prediction unit, the prediction unit can be divided into partitions included in the encoding unit 610, that is, a partition 610 having a size of 64x64, partitions 612 having a size of 64x32, partitions 614 having a size of 32x64, or partitions 616 having a size of 32x32. [00182] Similarly, a prediction unit of the encoding unit 620 having the size of 32x32 and the depth of 1 can be divided into partitions included in the encoding unit 620, i.e., a partition 620 having a size of 32x32, partitions 622 having a size of 32x16, partitions 624 having a size of 16x32, and partitions 62 having a size of 16x16. [00183] Similarly, a prediction unit of the encoding unit 630 having the size of 16x16 and the depth of 2 can be divided into partitions included in the encoding unit 630, that is, a partition having a size of 16x16 included in the unit of encoding 630, partitions 632 having a size of 16x8, partitions 634 having a size of 8x16, and partitions 636 having a size of 8x8. Similarly, a prediction unit of the coding unit 640 having the size of 8x8 and the depth of 3 can be divided into partitions included in the coding unit 640, that is, a partition having a size of 8x8 included in the unit of encoding 640, partitions 642 having a size of 8x4, partitions 644 having a size of 4x8, and partitions 646 having a size of 4x4. [00185] To determine at least one coded depth of the coding units constituting the maximum coding unit 610, the decoding unit determiner 120 of the video coding apparatus 10 performs coding for the coding units corresponding to each depth included in the maximum coding unit 610. [00186] A number of deeper coding units, according to the depths including data in the same variation and in the same size, increases as the depth is deepened. For example, four coding units corresponding to a depth of 2 are required to cover the data that is included in a coding unit corresponding to a depth of 1. Consequently, to compare the results of coding the same data according to the depths , the coding unit corresponding to the depth of 1; and four coding units corresponding to the depth of 2; are individually coded. [00187] To perform the coding for a current depth between the depths, a minimum coding error can be selected for the current depth by performing coding for each prediction unit in the coding units corresponding to the current depth along the horizontal axis of the hierarchical structure 600. Alternatively, the minimum coding error can be searched by comparing the minimum coding errors according to the depths, by performing coding for each depth as the depth is deepened along the vertical axis of the hierarchical structure 600. A depth and a partition having the minimum coding error in the 610 coding unit can be selected as the coded depth and partition type of the 610 coding unit. [00188] Figure 13 is a diagram for describing a relationship between a coding unit 710 and a transformation unit 720, according to an embodiment of the present invention. [00189] The video encoding apparatus 100 or 200 encodes or decodes an image according to the encoding units having sizes less than or equal to a maximum encoding unit for each maximum encoding unit. Transformation unit sizes for transformation during coding can be selected based on data units that are not larger than a corresponding coding unit. [00190] For example, in the video encoding apparatus 100 or 200, or the video decoding apparatus 200, if a size of the encoding unit 710 is 64x64, the transformation can be performed using the transformation units 720 having a size of 32x32. [00191] Furthermore, the data of the encoding unit 710 having the size of 64x64 can be encoded by performing the transformation in each of the transformation units having the size of 32x32, 16x16, 8x8, and 4x4, which are smaller than 64x64, and then a transformation unit having the minimum coding error can be selected. [00192] Figure 14 is a diagram for describing coding information of the coding units corresponding to a coded depth, according to an embodiment of the present invention. [00193] The output unit 130 of the video encoding apparatus 100 can encode and transmit information 800 about a partition type, information 810 about a prediction mode, and information 820 about a size of a transformation unit for each unit of encoding corresponding to an encoded depth, such as information about an encoding mode. [00194] Information 800 indicates information about a partition format obtained by dividing a prediction unit from a current encoding unit, where the partition is a data unit for prediction encoding of the current encoding unit. For example, a current encoding unit CU_0 having a size of 2Nx2N can be divided into any one of an 802 partition having a size of 2Nx2N, an 804 partition having a size of 2NxN, an 806 partition having a size of Nx2N, and an 808 partition having an NxN size. Here, information 800 about a partition type is established to indicate one of partition 804 having a size of 2NxN, partition 806 having a size of Nx2N, and partition 808 having a size of NxN. [00195] Information 810 indicates a prediction mode for each partition. For example, information 810 may indicate a prediction encoding mode performed on a partition indicated by information 800, that is, an intra mode 812, an inter mode 814, or a hop mode 816. [00196] Information 820 indicates a transformation unit to be based on when the transformation is performed in a current coding unit. For example, the transformation unit can be a first intra transformation unit 822, a second intra transformation unit 824, a first inter transformation unit 826, or a second intra transformation unit 828. [00197] The coding information and image data extractor 220 of the video decoding apparatus 200 can extract and use the information 800, 810 and 820 for decoding, according to each deeper coding unit. [00198] Figure 15 is a diagram of deeper coding units according to the depths, according to an embodiment of the present invention. [00199] Division information can be used to indicate a change in depth. The split information indicates whether a coding unit of a current depth is divided into coding units of a lower depth. [00200] A prediction unit 910 for prediction encoding of an encoding unit 900 having a depth of 0 and a size of 2N_0x2N_0 can include partitions of a partition type 912 having a size of 2N_0x2N_0, a partition type 914 having a size of 2N_0xN_0, a partition type 916 having a size of N_0x2N_0, and a partition type 918 having a size of N_0xN_0. Figure 9 illustrates only the partition types 912 to 918 that are obtained by symmetrically dividing the prediction unit 910, but a partition type is not limited to this, and the prediction unit 910 partitions can include asymmetric partitions, partitions having a predetermined shape and partitions having a geometric shape. [00201] Prediction coding is performed repeatedly on a partition having a size of 2N_0x2N_0, two partitions having a size of 2N_0xN_0, two partitions having a size of N_0x2N_0, and four partitions having a size of N OxN 0, according with each partition type. Prediction coding in an intra and an inter mode can be performed on partitions having the sizes of 2N_0x2N_0, N_0x2N_0, 2N OxN 0 and N 0xN_0. Prediction encoding in a jump mode is performed only on the partition having the size of 2N_0x2N 0. [00202] Encoding errors, including prediction encoding in partition types 912 through 918 are compared, and the minor encoding error is determined, between partition types. If a coding error is less in one of the partition types 912 to 916, the prediction unit 910 may not be divided into a smaller depth. [00203] If the coding error is the smallest in partition type 918, a depth is changed from 0 to 1 to divide partition type 918 in operation 920, and coding is performed repeatedly on coding units 930 having a depth 2 and a size of N_0xN 0 to look for a minimum coding error. [00204] A prediction unit 940 for prediction encoding of encoding unit 930 having a depth of 1 and a size of 2N_lx2N_l (= N_0xN_0) can include partitions of a partition type 942 having a size of 2N_lx2N_l, a type of partition 944 having a size of 2N_lxN_l, a partition type 946 having a size of N_lx2N_l, and a partition type 948 having a size of N__lxN 1. [00205] If an encoding error is the smallest in partition type 948, a depth is changed from 1 to 2 to divide partition type 948 In operation 950, and encoding is performed repeatedly on encoding units 960, which have a depth of 2 and a size of N_2xN_2 to look for a minimum coding error. [00206] When a maximum depth is d, the division operation according to each depth can be performed even when a depth becomes d-1; and the split information can be encoded even when a depth is one from 0 to d-2. In other words, when coding is performed even when the depth is d-1 after a coding unit corresponding to a depth of d-2 is divided in operation 970, a prediction unit 990 for prediction coding of a coding unit 980 having a depth of d-1 and a size of 2N_ (d-1) x2N_ (d-1) can include partitions of a partition type 992 having a size of 2N_ (d-1) x2N_ (d-1), a partition type 994 having a size of 2N_ (d-1) xN_ (d-1), a partition type 996 having a size of N_ (d-1) x2N_ (d-1), a partition type 998 having a size of N_ (d-1) xN_ (d-1). [00207] Prediction encoding can be performed repeatedly on a partition having a size of 2N (d- l) x2N_ (dl), two partitions having a size of 2N_ (dl) xN (d- 1), two partitions having a size of N_ (d-1) x2N_ (d-1), four partitions having a size of N_ (d-1) xN_ (d-1) among partition types 992 to 998 to search for a partition type having an error minimum coding. [00208] Even when partition type 998 has the minimum coding error, as a maximum depth is d, a CU_ coding unit (dl), having a depth of d-1 is no longer divided to a lower depth, and a coded depth for the coding units constituting a current maximum coding unit 900 is determined to be d-1 and a partition type of the current maximum coding unit 900 can be determined to be N (d- l) xN_ (dl) . In addition, since the maximum depth is d and a minimum coding unit 980 having the lowest depth of d-1 is no longer divided to a lower depth, the split information for the minimum coding unit 980 is not established. [00209] The data unit 999 can be a minimum unit for the current maximum encoding unit. A minimum unit according to an exemplary modality can be a rectangular data unit obtained by dividing a minimum coding unit 980 by 4. By performing the coding repeatedly, the video coding equipment 100 can select a depth having the error of minimum coding by comparing coding errors according to the depths of the coding unit 900 to determine a coded depth, and to establish a corresponding partition type and prediction mode as a coded depth coding mode. [00210] As such, the minimum coding errors according to the depths are compared at all depths from 1 to d, and the depth having the smallest coding error can be determined as a coded depth. The encoded depth, the partition type of the prediction unit, and the prediction mode can be encoded and transmitted as information about an encoding mode. In addition, as a coding unit is divided from a depth of 0 to a coded depth, only the coded depth division information is set to 0, and the depth division information excluding the coded depth is set to 1 . [00211] The encoding information and image data extractor 220 of the video decoding apparatus 200 can extract and use the information about the encoded depth and the prediction unit of the encoding unit 900 to decode partition 912. The decoding apparatus video decoding 200 can determine a depth, at which the divided information is 0, as an encoded depth using the information divided according to the depths, and use the information on a corresponding depth encoding mode for decoding. [00212] Figures 16 to 18 are diagrams to describe a relationship between coding units 1010, prediction units 1060, and transformation units 1070, according to an embodiment of the present invention. [00213] The coding units 1010 are coding units that have a tree structure, corresponding to the coded depths determined by the video coding apparatus 100, in a maximum coding unit. The prediction units 1060 are partitions of the prediction units of each of the 1010 encoding units, and the transformation units 1070 are transformation units of each of the 1010 encoding units. [00214] When a depth of a maximum coding unit is 0 in coding units 1010, the depths of coding units 1012 and 1054 are 1, the depths of coding units 1014, 1016, 1018, 1028, 1050, and 1052 are 2, the depths of the decoding units 1020, 1022, 1024, 1026, 1030, 1032 and 1048 are 3, and the depths of the encoding units 1040, 1042, 1044 and 1046 are 4. [00215] In the 1060 prediction units, some coding units 1014, 1016, 1022, 1032, 1048, 1050, 1052, 1054 are obtained by dividing the coding units into the 1010 coding units. In other words, partition types in the coding units 1014, 1022, 1050, and 1054 have a size of 2NxN, the partition types in the coding units 1016, 1048 and 1052 have a size of Nx2N, and a partition type of the coding unit 1032 has a size of NxN. The prediction units and partitions of the 1010 coding units are less than or equal to each coding unit. [00216] Transformation or inverse transformation is performed on the image data of the 1052 encoding unit in the 1070 transformation units in a data unit that is smaller than the 1052 encoding unit. In addition, the encoding units 1014, 1016, 1022, 1032, 1048, 1050 and 1052 in transformation units 1070 are different from those in prediction units 1060 in terms of sizes and shapes. In other words, the video encoding and decoding apparatus 100 and 200 can perform intra prediction, motion estimation, motion compensation, transformation, and reverse transformation individually in a data unit in the same encoding unit. [00217] Consequently, encoding is performed recursively in each of the coding units having a hierarchical structure in each region of a maximum coding unit to determine an optimal coding unit, and thus coding units that have a structure of recursive tree can be obtained. The encoding information can include divided information about a coding unit, information about a partition type, information about a prediction mode, and information about a size of a transformation unit. Table 1 shows the encoding information that can be established by the video encoding and video decoding apparatus 100 and 200. Table 1 [00218] Output unit 130 of video encoding apparatus 100 can output encoding information about encoding units having a tree structure, and the encoding information and image data extractor 220 from the decoding apparatus of video 200, can extract the encoding information about the encoding units having a tree structure from a received bit stream. [00219] The split information indicates whether a current coding unit is divided into coding units of a lower depth. If the information divided by a current depth d is 0, a depth, in which a current coding unit is no longer divided into a lower depth, is an encoded depth, and so the information about a partition type, the prediction mode , and a size of a transformation unit, can be defined for the coded depth. If the current coding unit is further divided according to the divided information, coding is carried out independently into four divided coding units of a lower depth. [00220] A prediction mode can be one of: an intra mode, an inter mode, and a jump mode. The intra-mode and the inter-mode can be set on all partition types, and the jump mode is set only on a partition type having a size of 2Nx2N. [00221] The partition type information can indicate symmetric partition types having sizes of 2Nx2N, 2NxN, Nx2N, and NxN, which are obtained by symmetrically dividing a height or width of a prediction unit, and asymmetric partition types having sizes of 2NxnU, 2NxnD, nLx2N and nRx2N, which are obtained by asymmetrically dividing the height or width of the prediction unit. The types of asymmetric partition having the sizes of 2NxnU and 2NxnD can be obtained respectively by dividing the height of the prediction unit in 1: 3 and 3: 1, and the types of asymmetric partition having the sizes of nLx2N and nRx2N can be obtained respectively by dividing the width of the prediction unit into 1: 3 and 3: 1. [00222] The size of the processing unit can be adjusted to be of two types in the intra mode and two types in the inter mode. In other words, if the transformation unit's split information is 0, the size of the transformation unit can be 2Nx2N, which is the size of the current coding unit. If split information from the processing unit is 1, the transformation units can be obtained by dividing the current coding unit. In addition, if a split type of the current coding unit having the size of 2Nx2N is an asymmetric partition type, a size of a transformation unit can be NxN, and if the partition type of the current coding unit is a type Asymmetric partition, the size of the transformation unit can be N / 2xN / 2. [00223] The coding information about the coding units having a tree structure can include at least one of a coding unit corresponding to a coded depth, a prediction unit, and a minimum unit. The coding unit corresponding to the coded depth can include at least one of a prediction unit and a minimum unit containing the same coding information. [00224] Consequently, it is determined whether adjacent data units are included in the same coding unit corresponding to the coded depth by comparing the coding information of the adjacent data units. In addition, a corresponding coding unit corresponding to a coded depth is determined using the coding information from a data unit, and thus a distribution of the coded depths in a maximum coding unit can be determined. [00225] Consequently, if a current coding unit is predicted based on the coding information of the adjacent data units, coding information of the data units in deeper coding units adjacent to the current coding unit can be directly queried and used . [00226] Alternatively, if a current coding unit is predicted based on the coding information of the adjacent data units, the data units adjacent to the current coding unit are searched for using the coding information of the data units, and the adjacent coding units sought can be referred to for prediction of the current coding unit. [00227] Figure 19 is a diagram to describe a relationship between a coding unit, a prediction unit or a partition, and a transformation unit, according to the coding mode information in Table 1. [00228] The maximum coding unit 1300 includes coding units 1302 1304, 1306, 1312, 1314, 1316 and 1318 of coded depths. Here, as the 1318 coding unit is a coded depth coding unit, the divided information can be set to 0. The information about a partition type of the 1318 coding unit having a size of 2Nx2N can be set to be a of: a type of partition 1322 having a size of 2Nx2N, a type of partition 1324 having a size of 2NxN, a type of partition 1326 having a size of Nx2N, a type of partition 32 8 having a size of NxN, a type of partition 1332 having a size of 2NxnU, a partition type 1334 having a size of 2NxnD, a partition type 1336 having a size of nLx2N, and a partition type 1338 having a size of nRx2N. [00229] Division information (TU size indicator (Transformation Unit)) of a transformation unit, that is, a Tu size flag, is a type of a transformation index. The size of a transformation unit corresponding to the transformation index can vary according to a type of prediction unit or a partition type of a coding unit. [00230] For example, when the partition type is set to be symmetric, that is, partition type 1322, 1324, 1326 or 1328, a 1342 transformation unit having a size of 2Nx2N is established when the TU size flag is 0, and a transformation unit 1344 having an NxN size is established when the TU size flag is 1. [00231] When the partition type is established to be asymmetric, that is, partition type 1332, 1334, 1336 or 1338, a 1352 transformation unit having a size of 2Nx2N is established if a TU size flag is 0, and a transformation unit 1354 having a size of N / 2xN / 2 is established if a flag of size TU is 1. [00232] With reference to Figure 19, the TU size flag is a flag having a value of 0 or 1, however the TU size flag is not limited to a 1 bit, and the transformation unit can be divided hierarchically having a tree structure while the TU size flag increases from 0. The division information (TU size flag) of a transformation unit can be an example of a transformation index. [00233] In this case, the size of a transformation unit that was effectively used can be expressed using a TU size indicator of a transformation unit, according to an exemplary modality, together with a maximum size and a size minimum of the processing unit. According to an exemplary embodiment, the video encoding equipment 100 can encode maximum transformation unit size information, minimum transformation unit size information, and a maximum TU size indicator. The result of coding the maximum transformation unit size information, minimum transformation unit size information, and the maximum TU size indicator can be entered in an SPS. According to an exemplary embodiment, the video encoding equipment 200 can decode video using the maximum transformation unit size information, the minimum transformation unit size information, and the maximum TU size indicator. [00234] For example, (a) if the size of a current encoding unit is 64x64 and a maximum transformation unit size is 32x32, (a-1) then the size of a transformation unit can be 32x32 when an indicator TU size is 0, (a-2) can be 16x16 when TU size indicator is 1, and (a-3) can be 8x8 when TU size indicator is 2. [00235] As another example, (b) if the current encoding unit size is 32x32 and a minimum transformation unit size is 32x32, (b-1) then the transformation unit size can be 32x32 when the TU size is 0. Here, the TU size indicator cannot be set to a value other than 0, since the size of the processing unit cannot be less than 32x32. [00236] As another example, (c) if the current encoding unit size is 64x64 and a maximum TU size indicator is 1, then the TU size indicator can be 0 or 1. Here, the TU size indicator does not can be set to a value other than 0 or 1. [00237] Thus, if the maximum TU size indicator is defined as "MaxTransformSizelndex", a minimum transformation unit size is "MinTransformSize", and a transformation unit size is "RootTuSize" when the TU size indicator is 0, then a minimum current transformation unit size "CurrMinTuSize" that can be determined in a current coding unit, can be defined by Equation (D: CurrMinTuSize = max (MinTransformSize, RootTuSize / (2AMaxTransformSizeIndex)) (1) [00238] Compared to the current minimum processing unit size "CurrMinTuSize" that can be determined in the current coding unit, a processing unit size "RootTuSize" when the TU size indicator is 0, can denote a size of maximum processing unit that can be selected in the system. In Equation (1), "RootTuSize (2AMaxTransformSizeIndex)" denotes a transformation unit size when the transformation unit size "RootTuSize", when the TU size indicator is 0, the number of times corresponding to the size indicator is divided Maximum TU, and "MintransformSize" denotes a minimum transformation size. Thus, a smaller value between "RootTuSize / (2AMaxTransformSizeIndex)" and "MinTransformSize" can be the current minimum transformation unit size "CurrMinTuSize" that can be determined in the current coding unit. [00239] According to an exemplary modality, the maximum RootTuSize transformation unit size may vary according to the type of prediction mode. [00240] For example, if a current prediction mode is an inter mode, then RootTuSize can be determined using Equation (2) below. In Equation (2), "MaxTransformSize" denotes a maximum transformation unit size, and "PUSize" denotes a current prediction unit size. RootTuSize = min (MaxTransformSize, PUSize) (2) [00241] That is, if the current prediction mode is inter mode, the "RootTuSize" transformation unit size when the TU size indicator is 0, may be a smaller value between the maximum transformation unit size and the current prediction unit size. [00242] If a prediction mode for a current partition drive is an intra mode, "RootTuSize" can be determined using Equation (3) below. In Equation (3), "Partitionsize" denotes the size of the current partition unit. RootTuSize = min (MaxTransformSize, Partitionsize) .. (3) [00243] That is, if the current prediction mode is intra mode, the "RootTuSize" transformation unit size when the TU size indicator is 0 may be a smaller value between the maximum transformation unit size and the size of the current partition drive. [00244] However, the current maximum root transformation unit size "RootTuSize" which varies according to the type of a prediction mode in a partition unit, is just one example, and the present invention is not limited to it. [00245] According to the video encoding method based on encoding units having a tree structure, as described with reference to figures 7 to 19, the image data of a spatial region are encoded for each encoding unit of a tree structure. According to the video decoding method based on encoding units that have a tree structure, decoding is performed for each maximum encoding unit to restore image data from a spatial region. Thus, an image and a video that is a sequence of images can be restored. The restored video can be played by a playback device, stored on a storage medium, or transmitted over a network. [00246] The modalities of the present invention can be recorded as computer programs and can be implemented in common digital computers that execute programs using a computer-readable recording medium. Examples of the computer-readable recording medium include magnetic storage media (for example, ROM, floppy disks, hard drives, etc.) and optical recording media (for example, CD-ROMs or DVDs). [00247] Although this invention has been particularly shown and described with reference to its preferred modalities, those of ordinary skill in the art will understand that various changes in shape and details can be made in them without departing from the essence and scope of the present invention as defined by the attached claims.
权利要求:
Claims (3) [0001] 1. METHOD FOR DECODING VIDEO, the method characterized by understanding: when the prediction of a colocalized block of a current block is available using a reference list LO of the colocalized block and prediction of the colocalized block is available using a reference list LI of the colocalized block and when the reference images of the current block will be sent before a current image including the current block, select a motion vector corresponding to a reference list of the current block from a LO motion vector of the colocalized block corresponding to an image determined reference from the list LO and a movement vector LI from the colocalized block corresponding to a determined reference image from the list Ll; determine using the selected motion vector the motion vector predictor candidate according to the colocalized block; and obtaining a motion vector predictor from the current block among predictor candidates including the motion vector predictor candidate according to the colocalized block. [0002] 2. METHOD, according to claim 1, characterized by the fact that the colocalized block is a colocalized block with a block location of the current block in a colocalized image which is determined among decoded images before the current image. [0003] 3. METHOD, according to claim 1, characterized by the fact that the current image is divided into a plurality of maximum coding units, a maximum coding unit among the plurality of maximum coding units is hierarchically divided into coding units depth, including a current depth and a lower depth according to the split information, when the split information indicates a split for the current depth, the current depth coding unit is divided into four quadratic coding units of a lower depth regardless of the neighboring coding units, when the division information indicates a non-division for the current depth, at least one prediction unit is obtained from the current depth coding unit is divided, and the current block is one of the at least one prediction unit.
类似技术:
公开号 | 公开日 | 专利标题 BR112013033906B1|2020-09-29|METHOD FOR DECODING VIDEO BR122013019952A2|2020-09-15|VIDEO DECODING DEVICE BR122020014018B1|2022-02-01|Method to decode a video BR122021004619B1|2021-08-24|APPARATUS TO DECODE A VIDEO BR122015013888A2|2020-09-01|METHOD FOR DECODING AN ENCODED VIDEO BR122021004152B1|2022-01-11|DEVICE TO DECODE VIDEO BR122020013760B1|2022-01-11|METHOD TO DECODE A VIDEO THROUGH MESH FILTRATION BASED ON ENCODING UNITS, AND VIDEO DECODING APPARATUS TO DECODE A VIDEO THROUGH MESH FILTRATION BASED ON ENCODING UNITS BR122015021751B1|2022-01-18|METHOD OF DECODING VIDEO DATA BR122015021376A2|2019-08-27|device for decoding a video MX2014000157A|2014-02-19|Video encoding method with intra prediction using checking process for unified reference possibility, video decoding method and device thereof. BR122021004622B1|2022-02-08|METHOD TO DECODE A VIDEO BR122021004831B1|2022-02-01|Method to decode an encoded video BR122021004836B1|2022-02-08|METHOD TO DECODE AN ENCODED VIDEO BR112012025308B1|2022-02-08|METHOD OF ENCODING A VIDEO, METHOD OF DECODING AN ENCRYPTED VIDEO, VIDEO ENCODING EQUIPMENT INCLUDING A PROCESSOR, VIDEO DECODING EQUIPMENT INCLUDING A PROCESSOR, AND COMPUTER-LEABLE RECORDING MEDIA BR112014018115B1|2021-10-26|METHOD OF DECODING A VIDEO BR122020001936B1|2021-05-25|method of decoding a video BR112012017407B1|2021-12-07|METHOD FOR ENCODING A VIDEO BITS STREAM COMPRISING ENCODED VIDEO DATA
同族专利:
公开号 | 公开日 WO2013005963A3|2013-03-14| MX2014000159A|2014-02-19| JP2015156702A|2015-08-27| PH12016501353A1|2016-08-31| US10397601B2|2019-08-27| CN104796698A|2015-07-22| RU2688255C1|2019-05-21| KR101626448B1|2016-06-13| EP2728862A2|2014-05-07| TWI594618B|2017-08-01| CN104796699B|2018-01-16| PH12016501350B1|2016-08-31| US20150201211A1|2015-07-16| KR101626006B1|2016-06-13| JP5997317B2|2016-09-28| BR112013033906A2|2019-09-17| KR20130004435A|2013-01-10| TWI520581B|2016-02-01| TW201313025A|2013-03-16| US9253488B2|2016-02-02| US9232229B2|2016-01-05| ZA201502919B|2017-01-25| KR102041886B1|2019-11-07| RU2636100C1|2017-11-20| TW201735636A|2017-10-01| CN104780380A|2015-07-15| JP5997316B2|2016-09-28| MX336708B|2016-01-28| CN104796698B|2017-03-29| CA2840726A1|2013-01-10| CN104780363B|2018-01-26| CN103765894A|2014-04-30| KR101946575B1|2019-02-11| ZA201502918B|2017-01-25| US20150201210A1|2015-07-16| US20160182914A1|2016-06-23| JP2014521243A|2014-08-25| RU2605915C2|2016-12-27| JP2015130708A|2015-07-16| PH12016501352B1|2016-08-31| US20150172697A1|2015-06-18| US20150172698A1|2015-06-18| CN104780363A|2015-07-15| JP5832646B2|2015-12-16| MX337234B|2016-02-18| RU2665230C1|2018-08-28| AU2017268677A1|2017-12-21| AU2016201369B2|2016-11-03| US9253502B2|2016-02-02| KR20140085386A|2014-07-07| KR101639335B1|2016-07-22| AU2012278479A1|2014-02-13| ZA201502916B|2017-01-25| JP5832683B2|2015-12-16| TW201541940A|2015-11-01| PH12016501351A1|2016-08-31| CN104780380B|2017-01-18| US20140133571A1|2014-05-15| ZA201400765B|2015-12-23| SG10201503685RA|2015-06-29| JP2015149776A|2015-08-20| CN105100798A|2015-11-25| CA2982301A1|2013-01-10| SG10201600009WA|2016-02-26| CA2840726C|2017-11-28| US20180332300A1|2018-11-15| KR20150000854A|2015-01-05| AU2017268677B2|2019-01-24| PH12016501353B1|2016-08-31| SG10201708530WA|2017-12-28| US9313517B2|2016-04-12| TW201540046A|2015-10-16| MY166149A|2018-06-06| JP2015130707A|2015-07-16| US9118926B2|2015-08-25| US10034014B2|2018-07-24| CN104796699A|2015-07-22| US9762924B2|2017-09-12| TWI643489B|2018-12-01| KR20190014045A|2019-02-11| SG10201503690QA|2015-06-29| AU2017200249B2|2017-09-07| KR20150060617A|2015-06-03| TWI542198B|2016-07-11| US20170238007A1|2017-08-17| EP2728862A4|2015-03-25| KR20150060619A|2015-06-03| KR101519818B1|2015-05-13| PH12016501351B1|2016-08-31| TW201541941A|2015-11-01| ZA201502917B|2017-01-25| PH12016501352A1|2016-08-31| JP5832684B2|2015-12-16| CA2982301C|2020-03-10| RU2014103484A|2015-08-10| SG10201503692UA|2015-06-29| CN105100798B|2018-08-07| PH12016501350A1|2016-08-31| KR101495725B1|2015-02-25| CN103765894B|2017-05-10| AU2012278479B2|2016-01-14| TWI526052B|2016-03-11| AU2017200249A1|2017-02-02| WO2013005963A2|2013-01-10| AU2016201369A1|2016-03-24| KR20150060618A|2015-06-03|
引用文献:
公开号 | 申请日 | 公开日 | 申请人 | 专利标题 DE10113880B4|2001-03-21|2004-04-29|T-Mobile Deutschland Gmbh|Method for compressing and decompressing video data| ES2687176T3|2001-11-06|2018-10-24|Panasonic Intellectual Property Corporation Of America|Encoding method of moving images and decoding method of moving images| JP2005510984A|2001-11-21|2005-04-21|ジェネラル・インストルメント・コーポレーション|Picture-level adaptive frame / field coding for digital video content| US6980596B2|2001-11-27|2005-12-27|General Instrument Corporation|Macroblock level adaptive frame/field coding for digital video content| KR100508798B1|2002-04-09|2005-08-19|엘지전자 주식회사|Method for predicting bi-predictive block| JP2004208258A|2002-04-19|2004-07-22|Matsushita Electric Ind Co Ltd|Motion vector calculating method| HUE044616T2|2002-04-19|2019-11-28|Panasonic Ip Corp America|Motion vector calculating method| EP2271106B1|2002-04-19|2016-05-25|Panasonic Intellectual Property Corporation of America|Motion vector calculating method| JP2004023458A|2002-06-17|2004-01-22|Toshiba Corp|Moving picture encoding/decoding method and apparatus| US7970058B2|2002-07-15|2011-06-28|Hitachi Consumer Electronics Co., Ltd.|Moving picture encoding method and decoding method| KR100506864B1|2002-10-04|2005-08-05|엘지전자 주식회사|Method of determining motion vector| WO2005002233A1|2003-06-27|2005-01-06|Koninklijke Philips Electronics N.V.|Method of encoding for handheld apparatuses| US20050013498A1|2003-07-18|2005-01-20|Microsoft Corporation|Coding of motion vector information| EP1683361B1|2003-10-27|2007-07-18|Koninklijke Philips Electronics N.V.|Power optimized collocated motion estimation method| US7400681B2|2003-11-28|2008-07-15|Scientific-Atlanta, Inc.|Low-complexity motion vector prediction for video codec with two lists of reference pictures| US7480334B2|2003-12-23|2009-01-20|Genesis Microchip Inc.|Temporal motion vector filtering| US8036271B2|2004-02-24|2011-10-11|Lsi Corporation|Method and apparatus for determining a second picture for temporal direct-mode block prediction| ZA200607434B|2004-03-09|2008-08-27|Thomson Res Funding Corp|Reduced resolution update mode for advanced video coding| US8731054B2|2004-05-04|2014-05-20|Qualcomm Incorporated|Method and apparatus for weighted prediction in predictive frames| US20060133495A1|2004-12-22|2006-06-22|Yan Ye|Temporal error concealment for video communications| CN101253777A|2005-07-01|2008-08-27|极速决件公司|Method, apparatus and system for use in multimedia signal encoding| JP2007043651A|2005-07-05|2007-02-15|Ntt Docomo Inc|Dynamic image encoding device, dynamic image encoding method, dynamic image encoding program, dynamic image decoding device, dynamic image decoding method, and dynamic image decoding program| JP2007028393A|2005-07-20|2007-02-01|Matsushita Electric Ind Co Ltd|Method for calculating motion vector of direct mode block in bidirectionally predictive coded picture| RU2368095C1|2005-07-22|2009-09-20|Мицубиси Электрик Корпорейшн|Image coder and image decoder, method of image coding and method of image decoding, program of image coding and program of image decoding and computer-readable record medium, on which image coding program is recorded, and computer-readable record medium, on which image decoding program is recorded| JP4702943B2|2005-10-19|2011-06-15|キヤノン株式会社|Image processing apparatus and method| KR100775176B1|2006-03-10|2007-11-12|엘지전자 주식회사|Thumbnail recording method for providing information of video data and terminal using the same| JP5055354B2|2006-03-30|2012-10-24|エルジーエレクトロニクスインコーポレイティド|Video signal decoding / encoding method and apparatus| MX2009000205A|2006-07-06|2009-03-23|Thomson Licensing|Method and apparatus for decoupling frame number and/or picture order count for multi-view video encoding and decoding.| TW200820791A|2006-08-25|2008-05-01|Lg Electronics Inc|A method and apparatus for decoding/encoding a video signal| KR101031022B1|2006-10-20|2011-06-29|노키아 코포레이션|Virtual decoded reference picture marking and reference picture list| US8218640B2|2006-10-31|2012-07-10|Sony Computer Entertainment Inc.|Picture decoding using same-picture reference for pixel reconstruction| CN101647279A|2007-01-24|2010-02-10|Lg电子株式会社|A method and an apparatus for processing a video signal| CN101669367A|2007-03-02|2010-03-10|Lg电子株式会社|A method and an apparatus for decoding/encoding a video signal| CN101682784A|2007-04-19|2010-03-24|汤姆逊许可证公司|Adaptive reference picture data generation for intra prediction| JP4884290B2|2007-05-07|2012-02-29|パナソニック株式会社|Moving picture decoding integrated circuit, moving picture decoding method, moving picture decoding apparatus, and moving picture decoding program| CN101415115B|2007-10-15|2011-02-02|华为技术有限公司|Method for encoding and decoding video based on movement dancing mode, and encoder and decoder thereof| KR101691199B1|2008-04-11|2016-12-30|톰슨 라이센싱|Method and apparatus for template matching prediction in video encoding and decoding| JP5210091B2|2008-08-29|2013-06-12|キヤノン株式会社|Image processing apparatus, control method therefor, imaging apparatus, and program| US8634457B2|2008-09-26|2014-01-21|Qualcomm Incorporated|Determining availability of video data units| US8831087B2|2008-10-06|2014-09-09|Qualcomm Incorporated|Efficient prediction mode selection| US20100111166A1|2008-10-31|2010-05-06|Rmi Corporation|Device for decoding a video stream and method thereof| KR101279573B1|2008-10-31|2013-06-27|에스케이텔레콤 주식회사|Motion Vector Encoding/Decoding Method and Apparatus and Video Encoding/Decoding Method and Apparatus| US8687685B2|2009-04-14|2014-04-01|Qualcomm Incorporated|Efficient transcoding of B-frames to P-frames| KR20110000914A|2009-06-29|2011-01-06|조영식|Em-containing interior construction material made of charcoal and method of making same| US9161057B2|2009-07-09|2015-10-13|Qualcomm Incorporated|Non-zero rounding and prediction mode selection techniques in video encoding| US20110216640A1|2009-07-17|2011-09-08|James Curtis|Universal multimedia distribution, storage, and playback systems, and methods| US8462852B2|2009-10-20|2013-06-11|Intel Corporation|Methods and apparatus for adaptively choosing a search range for motion estimation| KR20110068792A|2009-12-16|2011-06-22|한국전자통신연구원|Adaptive image coding apparatus and method| KR20120009861A|2010-07-22|2012-02-02|에스케이 텔레콤주식회사|Method and Apparatus for Encoding/Decoding of Video Data Using Expanded Skip Mode| MX2014000159A|2011-07-02|2014-02-19|Samsung Electronics Co Ltd|Sas-based semiconductor storage device memory disk unit.| MX2013009864A|2011-03-03|2013-10-25|Panasonic Corp|Video image encoding method, video image decoding method, video image encoding device, video image decoding device, and video image encoding/decoding device.| MX338669B|2011-06-28|2016-04-27|Samsung Electronics Co Ltd|Video encoding method using offset adjustments according to pixel classification and apparatus therefor, video decoding method and apparatus therefor.| WO2013157783A1|2012-04-15|2013-10-24|삼성전자 주식회사|Inter prediction method in which reference picture lists can be changed and apparatus for same| EP2932712A4|2013-06-11|2016-11-30|Hfi Innovation Inc|Method of inter-view residual prediction with reduced complexity in three-dimensional video coding| WO2015055143A1|2013-10-17|2015-04-23|Mediatek Inc.|Method of motion information prediction and inheritance in multi-view and three-dimensional video coding|JP3296876B2|1993-03-26|2002-07-02|オリンパス光学工業株式会社|Zoom lens| MX2014000159A|2011-07-02|2014-02-19|Samsung Electronics Co Ltd|Sas-based semiconductor storage device memory disk unit.| EP2804374A1|2013-02-22|2014-11-19|Thomson Licensing|Coding and decoding methods of a picture block, corresponding devices and data stream| JP2016528792A|2013-07-12|2016-09-15|サムスン エレクトロニクス カンパニー リミテッド|Inter-layer video decoding apparatus, inter-layer video encoding apparatus, and block-based disparity vector prediction method for the method| US20160205409A1|2013-09-04|2016-07-14|Samsung Electronics Co., Ltd.|Screen image encoding method and apparatus therefor, and screen image decoding method and apparatus therefor| WO2017082636A1|2015-11-11|2017-05-18|삼성전자 주식회사|Method for encoding/decoding image, and device therefor| KR20180075517A|2015-11-24|2018-07-04|삼성전자주식회사|Video encoding method and apparatus, video decoding method and apparatus| KR20180075518A|2015-11-24|2018-07-04|삼성전자주식회사|Video encoding method and apparatus, video decoding method and apparatus| CN108432252A|2015-12-22|2018-08-21|真实网络公司|Motion vector selection and forecasting system in Video coding and method| EP3439303B1|2016-04-28|2020-12-30|LG Electronics Inc. -1-|Inter prediction mode-based image processing method and apparatus therefor| CN106507111B|2016-11-17|2019-11-15|上海兆芯集成电路有限公司|Method for video coding using residual compensation and the device using this method| WO2018128228A1|2017-01-03|2018-07-12|엘지전자 주식회사|Image decoding method and device in image coding system| CN112219400A|2018-11-06|2021-01-12|北京字节跳动网络技术有限公司|Location dependent storage of motion information| CN111164976A|2019-01-03|2020-05-15|北京大学|Video processing method and device| CN110213590A|2019-06-25|2019-09-06|浙江大华技术股份有限公司|Time-domain motion vector acquisition, inter-prediction, Video coding method and apparatus|
法律状态:
2019-10-01| B06F| Objections, documents and/or translations needed after an examination request according art. 34 industrial property law| 2020-05-05| B09A| Decision: intention to grant| 2020-09-29| B16A| Patent or certificate of addition of invention granted|Free format text: PRAZO DE VALIDADE: 20 (VINTE) ANOS CONTADOS A PARTIR DE 02/07/2012, OBSERVADAS AS CONDICOES LEGAIS. |
优先权:
[返回顶部]
申请号 | 申请日 | 专利标题 US201161504177P| true| 2011-07-02|2011-07-02| US61/504,177|2011-07-02| US201161548415P| true| 2011-10-18|2011-10-18| US61/548,415|2011-10-18| PCT/KR2012/005247|WO2013005963A2|2011-07-02|2012-07-02|Method and apparatus for coding video, and method and apparatus for decoding video accompanied by inter prediction using collocated image| 相关专利
Sulfonates, polymers, resist compositions and patterning process
Washing machine
Washing machine
Device for fixture finishing and tension adjusting of membrane
Structure for Equipping Band in a Plane Cathode Ray Tube
Process for preparation of 7 alpha-carboxyl 9, 11-epoxy steroids and intermediates useful therein an
国家/地区
|