Method and apparatus for image coding/decoding

Saved in:
Bibliographic Details
Title: Method and apparatus for image coding/decoding
Patent Number: 11470,342
Publication Date: October 11, 2022
Appl. No: 17/214040
Application Filed: March 26, 2021
Abstract: A method for decoding an image according to the present invention comprises the steps of: decoding a residual block by quantizing and inverse transforming an entropy-decoded residual block; generating a prediction block via motion compensation; and decoding an image by adding the decoded residual block to the prediction block, wherein on the basis of the maximum number of motion vector candidates of the motion vector candidate list related to the prediction block, a motion vector candidate list is adjusted by adding a particular motion vector candidate or by discarding a portion from among the motion vector candidates, and in the prediction block generation step, a prediction motion vector of the prediction block is determined on the basis of the adjusted motion vector candidate list. Accordingly, the complexity of arithmetic operations is reduced during encoding/decoding of an image.
Inventors: Electronics and Telecommunications Research Institute (Daejeon, KR)
Assignees: Electronics and Telecommunications Research Institute (Daejeon, KR)
Claim: 1. An image decoding method comprising: configuring a motion vector candidate list; modifying the motion vector candidate list based on a number of motion vector candidates in the motion vector candidate list; and determining a prediction motion vector based on the modified motion vector candidate list, wherein the modified motion vector candidate list comprises any one or any combination of any two or more of a spatial motion vector candidate, a temporal motion vector candidate and a (0,0) motion vector, wherein the configuring the motion vector candidate list comprises: deriving the spatial motion vector candidate, deriving the temporal motion vector candidate except when two derived spatial motion vector candidates are present and different from each other, and adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list, wherein when the number of motion vector candidates in the motion vector candidate list is smaller than a maximum number of motion vector candidates, the modifying the motion vector candidate list includes repeatedly adding the (0,0) motion vector candidate until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates, based on the maximum number of motion vector candidates and the number of motion vector candidates in the motion vector candidate list, wherein the adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list includes carrying out an operation of checking the same motion vector candidate only on the spatial motion vector candidates for removing the same motion vector candidate, and wherein the maximum number of motion vector candidates is 2 and when no motion vector candidate is present in the motion vector candidate list, two (0,0) motion vector candidates are added to the motion vector candidate list.
Claim: 2. The image decoding method of claim 1 , wherein in a case where the number of motion vector candidates in the motion vector candidate list is smaller than the maximum number of motion vector candidates, the modifying the motion vector candidate list includes iteratively adding the (0,0) motion vector candidate until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates irrespective of whether a motion vector candidate is present in the motion vector candidate list or whether the (0,0) motion vector candidate is present in the motion vector candidate list.
Claim: 3. The image decoding method of claim 1 , wherein when one (0,0) motion vector candidate is present in the motion vector candidate list, one additional (0,0) motion vector candidate is added to the motion vector candidate list.
Claim: 4. An image encoding method comprising: generating a prediction block by performing an inter prediction; and performing an entropy encoding of a residual block corresponding to a difference between a current block and the prediction block predicted by the inter prediction, wherein a prediction motion vector corresponding to the prediction block is included in a motion vector candidate list, wherein the motion vector candidate list includes any one or any combination of any two or more of a spatial motion vector candidate, a temporal motion vector candidate and a (0,0) motion vector candidate, wherein the motion vector candidate list is configured by: deriving the spatial motion vector candidate, deriving the temporal motion vector candidate except when two derived spatial motion vector candidates are present and different from each other, and adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list, wherein when a number of motion vector candidates in the motion vector candidate list is smaller than a maximum number of motion vector candidates, the (0,0) motion vector candidate is repeatedly added to the motion vector candidate list until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates, based on the maximum number of motion vector candidates and the number of motion vector candidates in the motion vector candidate list, wherein the adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list includes carrying out an operation of checking the same motion vector candidate only on the spatial motion vector candidates for removing the same motion vector candidate, and wherein the maximum number of motion vector candidates is 2 and when no motion vector candidate is present in the motion vector candidate list, two (0,0) motion vector candidates are added to the motion vector candidate list.
Claim: 5. The image decoding method of claim 4 , wherein when the number of motion vector candidates in the motion vector candidate list is smaller than the maximum number of motion vector candidates, the motion vector candidate list is iteratively modified by adding the (0,0) motion vector candidate until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates irrespective of whether a motion vector candidate is present in the motion vector candidate list or whether the (0,0) motion vector candidate is present in the motion vector candidate list.
Claim: 6. A non-transitory computer-readable storage medium storing a bitstream, the bitstream comprising: information for determining a prediction motion vector based on a motion vector candidate list, wherein the motion vector candidate list includes any one or any combination of any two or more of a spatial motion vector candidate, a temporal motion vector candidate and a (0,0) motion vector candidate, wherein the motion vector candidate list is configured by: deriving the spatial motion vector candidate, deriving the temporal motion vector candidate except when two derived spatial motion vector candidates are present and different from each other, and adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list, wherein when a number of motion vector candidates in the motion vector candidate list is smaller than a maximum number of motion vector candidates, the (0,0) motion vector candidate is repeatedly added to the motion vector candidate list until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates, based on the maximum number of motion vector candidates and the number of motion vector candidates in the motion vector candidate list, wherein the adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list includes carrying out an operation of checking the same motion vector candidate only on the spatial motion vector candidates for removing the same motion vector candidate, and wherein the maximum number of motion vector candidates is 2 and when no motion vector candidate is present in the motion vector candidate list, two (0,0) motion vector candidates are added to the motion vector candidate list.
Patent References Cited: 6343100 January 2002 Fujiwara et al.
9319657 April 2016 Kang et al.
9392300 July 2016 Kim et al.
10218999 February 2019 Lim et al.
2004/0247031 December 2004 Hagai et al.
2008/0075169 March 2008 Ugur et al.
2010/0026903 February 2010 Tetsukawa et al.
2011/0176612 July 2011 Tsai et al.
2011/0176615 July 2011 Lee et al.
2011/0194608 August 2011 Rusert et al.
2011/0194609 August 2011 Rusert et al.
2012/0195368 August 2012 Chien et al.
2012/0269268 October 2012 Kim et al.
2012/0269270 October 2012 Chen et al.
2012/0275522 November 2012 Kim et al.
2012/0320984 December 2012 Zhou
2013/0107963 May 2013 Wahadaniah et al.
2013/0242046 September 2013 Zhang et al.
2014/0133570 May 2014 Lee
1523896 August 2004
1960496 May 2007
101102503 January 2008
101102512 January 2008
101653009 February 2010
102027746 April 2011
102215386 October 2011
102273206 December 2011
2013-85235 May 2013
6449203 January 2019
10-2010-0048435 May 2010
10-2011-0008653 January 2011
10-2011-0027480 March 2011
10-2011-0045908 May 2011
10-2011-0085896 July 2011
2010/050706 May 2010
2011/010858 January 2011
2011/031030 March 2011
2011/052897 May 2011
2012/173415 December 2012
2013/002557 January 2013
2013/067924 May 2013

















Other References: Bici Oguz et al., “Non-CE13: Simplification of merge mode,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG1, 7.sup.th Meeting: Geneva, CH, Nov. 8, 2011, Document: JCTVC-G593. cited by applicant
Brass et al., “WD5: Working Draft 5 of High-Efficiency Video Coding,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G1103_d4, 7th Meeting: Geneva, CH, Nov. 21-30, 2011. cited by applicant
Chen, Jianle. “MVP index parsing with fixed number of candidates.” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-F402, 6th Meeting, Torino, IT., Jul. 14-22, 2011. cited by applicant
European Patent Office Action for Application No. 13738327.9 dated Jun. 17, 2019 (5 pages). cited by applicant
Extended European Search Report dated Aug. 21, 2015 in counterpart European Application No. 13738327.9. cited by applicant
Hiroya Nakamura et al., “Non-CE9: Construction of MVP list without using scaling operation”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G219, 7.sup.th Meeting: Geneva, CH, Nov. 21-30, 2011. cited by applicant
International Search Report dated Apr. 12, 2013 in counterpart Application No. PCT/KR2013/000418 (5 pages, in Korean, with English language translation). cited by applicant
ITU-T, “Recommendation H.264: Advanced video coding for generic audiovisual services,” Series H: Audiovisual and Multimedia Systems, Infrastructure of audiovisual services—Coding of Moving video, 2010, pp. 158-160. cited by applicant
Japanese Office Action dated Aug. 18, 2015 in counterpart Japanese Application No. 2014-553258 (5 pages in Japanese). cited by applicant
Na Zhang et al., “CE9: Improvement in temporal candidate of merge mode and AMVP”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G343, 7.sup.th Meeting: Geneva, CH, Nov. 21-30, 2011. cited by applicant
Nakamura, Hiroya, et al. “Unification of derivation process for merge mode and MVP.” document JCTVC-F419 (2011) (10 pages). cited by applicant
Thomas Wiegand, et al., “Overview of the H.264/AVC Video Coding Standard,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 13, No. 7, Jul. 2003, pp. 560-576 (17 pages, in English). cited by applicant
Toshiyasu Sugio et al., “Non-CE9/Non-CE13: Simplification on AMVP/Merge”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G542, 7.sup.th Meeting: Geneva, CH, Nov. 21-30, 2011. cited by applicant
Toshiyasu Sugio, et al., “Non-CE9: Simplified scaling calculation method for temporal/spatial MVP of AMVP/Merge,” Proceedings of the 7.sup.th Meeting of the Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G541 Geneva, CH, Nov. 2011, pp. 1-5 (5 pages, in English). cited by applicant
Zhuo et al., “A study on HM3.0 parsing throughput issue,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-F068, 6th Meeting: Torino, IT, Jul. 14-22, 2011. cited by applicant
Toshiyasu Sugio et al., “Parsing Robustness for Merge/AMVP”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-F470, 6th Meeting: Torino, IT, Jul. 14-22, 2011. cited by applicant
Sung-Chang Lim et al., “Non-CE9: Simplification on AMVP list construction,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-H0239, 8th Meeting: San Jose, CA, USA, Feb. 1-10, 2012. cited by applicant
European Patent Office Search Report for Application No. 21202076.2 dated Feb. 10, 2022 (5 pages). cited by applicant
Assistant Examiner: Brumfield, Shanika M
Primary Examiner: Czekaj, Dave
Attorney, Agent or Firm: Michael Best & Friedrich LLP
Accession Number: edspgr.11470342
Database: USPTO Patent Grants
Description
Abstract:A method for decoding an image according to the present invention comprises the steps of: decoding a residual block by quantizing and inverse transforming an entropy-decoded residual block; generating a prediction block via motion compensation; and decoding an image by adding the decoded residual block to the prediction block, wherein on the basis of the maximum number of motion vector candidates of the motion vector candidate list related to the prediction block, a motion vector candidate list is adjusted by adding a particular motion vector candidate or by discarding a portion from among the motion vector candidates, and in the prediction block generation step, a prediction motion vector of the prediction block is determined on the basis of the adjusted motion vector candidate list. Accordingly, the complexity of arithmetic operations is reduced during encoding/decoding of an image.