Method and apparatus for image coding/decoding
Saved in:
| Title: | Method and apparatus for image coding/decoding |
|---|---|
| Patent Number: | 11470,342 |
| Publication Date: | October 11, 2022 |
| Appl. No: | 17/214040 |
| Application Filed: | March 26, 2021 |
| Abstract: | A method for decoding an image according to the present invention comprises the steps of: decoding a residual block by quantizing and inverse transforming an entropy-decoded residual block; generating a prediction block via motion compensation; and decoding an image by adding the decoded residual block to the prediction block, wherein on the basis of the maximum number of motion vector candidates of the motion vector candidate list related to the prediction block, a motion vector candidate list is adjusted by adding a particular motion vector candidate or by discarding a portion from among the motion vector candidates, and in the prediction block generation step, a prediction motion vector of the prediction block is determined on the basis of the adjusted motion vector candidate list. Accordingly, the complexity of arithmetic operations is reduced during encoding/decoding of an image. |
| Inventors: | Electronics and Telecommunications Research Institute (Daejeon, KR) |
| Assignees: | Electronics and Telecommunications Research Institute (Daejeon, KR) |
| Claim: | 1. An image decoding method comprising: configuring a motion vector candidate list; modifying the motion vector candidate list based on a number of motion vector candidates in the motion vector candidate list; and determining a prediction motion vector based on the modified motion vector candidate list, wherein the modified motion vector candidate list comprises any one or any combination of any two or more of a spatial motion vector candidate, a temporal motion vector candidate and a (0,0) motion vector, wherein the configuring the motion vector candidate list comprises: deriving the spatial motion vector candidate, deriving the temporal motion vector candidate except when two derived spatial motion vector candidates are present and different from each other, and adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list, wherein when the number of motion vector candidates in the motion vector candidate list is smaller than a maximum number of motion vector candidates, the modifying the motion vector candidate list includes repeatedly adding the (0,0) motion vector candidate until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates, based on the maximum number of motion vector candidates and the number of motion vector candidates in the motion vector candidate list, wherein the adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list includes carrying out an operation of checking the same motion vector candidate only on the spatial motion vector candidates for removing the same motion vector candidate, and wherein the maximum number of motion vector candidates is 2 and when no motion vector candidate is present in the motion vector candidate list, two (0,0) motion vector candidates are added to the motion vector candidate list. |
| Claim: | 2. The image decoding method of claim 1 , wherein in a case where the number of motion vector candidates in the motion vector candidate list is smaller than the maximum number of motion vector candidates, the modifying the motion vector candidate list includes iteratively adding the (0,0) motion vector candidate until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates irrespective of whether a motion vector candidate is present in the motion vector candidate list or whether the (0,0) motion vector candidate is present in the motion vector candidate list. |
| Claim: | 3. The image decoding method of claim 1 , wherein when one (0,0) motion vector candidate is present in the motion vector candidate list, one additional (0,0) motion vector candidate is added to the motion vector candidate list. |
| Claim: | 4. An image encoding method comprising: generating a prediction block by performing an inter prediction; and performing an entropy encoding of a residual block corresponding to a difference between a current block and the prediction block predicted by the inter prediction, wherein a prediction motion vector corresponding to the prediction block is included in a motion vector candidate list, wherein the motion vector candidate list includes any one or any combination of any two or more of a spatial motion vector candidate, a temporal motion vector candidate and a (0,0) motion vector candidate, wherein the motion vector candidate list is configured by: deriving the spatial motion vector candidate, deriving the temporal motion vector candidate except when two derived spatial motion vector candidates are present and different from each other, and adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list, wherein when a number of motion vector candidates in the motion vector candidate list is smaller than a maximum number of motion vector candidates, the (0,0) motion vector candidate is repeatedly added to the motion vector candidate list until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates, based on the maximum number of motion vector candidates and the number of motion vector candidates in the motion vector candidate list, wherein the adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list includes carrying out an operation of checking the same motion vector candidate only on the spatial motion vector candidates for removing the same motion vector candidate, and wherein the maximum number of motion vector candidates is 2 and when no motion vector candidate is present in the motion vector candidate list, two (0,0) motion vector candidates are added to the motion vector candidate list. |
| Claim: | 5. The image decoding method of claim 4 , wherein when the number of motion vector candidates in the motion vector candidate list is smaller than the maximum number of motion vector candidates, the motion vector candidate list is iteratively modified by adding the (0,0) motion vector candidate until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates irrespective of whether a motion vector candidate is present in the motion vector candidate list or whether the (0,0) motion vector candidate is present in the motion vector candidate list. |
| Claim: | 6. A non-transitory computer-readable storage medium storing a bitstream, the bitstream comprising: information for determining a prediction motion vector based on a motion vector candidate list, wherein the motion vector candidate list includes any one or any combination of any two or more of a spatial motion vector candidate, a temporal motion vector candidate and a (0,0) motion vector candidate, wherein the motion vector candidate list is configured by: deriving the spatial motion vector candidate, deriving the temporal motion vector candidate except when two derived spatial motion vector candidates are present and different from each other, and adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list, wherein when a number of motion vector candidates in the motion vector candidate list is smaller than a maximum number of motion vector candidates, the (0,0) motion vector candidate is repeatedly added to the motion vector candidate list until the number of motion vector candidates in the motion vector candidate list reaches the maximum number of motion vector candidates, based on the maximum number of motion vector candidates and the number of motion vector candidates in the motion vector candidate list, wherein the adding either one or both of the derived spatial motion vector candidate and the derived temporal motion vector candidate to the motion vector candidate list includes carrying out an operation of checking the same motion vector candidate only on the spatial motion vector candidates for removing the same motion vector candidate, and wherein the maximum number of motion vector candidates is 2 and when no motion vector candidate is present in the motion vector candidate list, two (0,0) motion vector candidates are added to the motion vector candidate list. |
| Patent References Cited: | 6343100 January 2002 Fujiwara et al. 9319657 April 2016 Kang et al. 9392300 July 2016 Kim et al. 10218999 February 2019 Lim et al. 2004/0247031 December 2004 Hagai et al. 2008/0075169 March 2008 Ugur et al. 2010/0026903 February 2010 Tetsukawa et al. 2011/0176612 July 2011 Tsai et al. 2011/0176615 July 2011 Lee et al. 2011/0194608 August 2011 Rusert et al. 2011/0194609 August 2011 Rusert et al. 2012/0195368 August 2012 Chien et al. 2012/0269268 October 2012 Kim et al. 2012/0269270 October 2012 Chen et al. 2012/0275522 November 2012 Kim et al. 2012/0320984 December 2012 Zhou 2013/0107963 May 2013 Wahadaniah et al. 2013/0242046 September 2013 Zhang et al. 2014/0133570 May 2014 Lee 1523896 August 2004 1960496 May 2007 101102503 January 2008 101102512 January 2008 101653009 February 2010 102027746 April 2011 102215386 October 2011 102273206 December 2011 2013-85235 May 2013 6449203 January 2019 10-2010-0048435 May 2010 10-2011-0008653 January 2011 10-2011-0027480 March 2011 10-2011-0045908 May 2011 10-2011-0085896 July 2011 2010/050706 May 2010 2011/010858 January 2011 2011/031030 March 2011 2011/052897 May 2011 2012/173415 December 2012 2013/002557 January 2013 2013/067924 May 2013 |
| Other References: | Bici Oguz et al., “Non-CE13: Simplification of merge mode,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG1, 7.sup.th Meeting: Geneva, CH, Nov. 8, 2011, Document: JCTVC-G593. cited by applicant Brass et al., “WD5: Working Draft 5 of High-Efficiency Video Coding,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G1103_d4, 7th Meeting: Geneva, CH, Nov. 21-30, 2011. cited by applicant Chen, Jianle. “MVP index parsing with fixed number of candidates.” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-F402, 6th Meeting, Torino, IT., Jul. 14-22, 2011. cited by applicant European Patent Office Action for Application No. 13738327.9 dated Jun. 17, 2019 (5 pages). cited by applicant Extended European Search Report dated Aug. 21, 2015 in counterpart European Application No. 13738327.9. cited by applicant Hiroya Nakamura et al., “Non-CE9: Construction of MVP list without using scaling operation”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G219, 7.sup.th Meeting: Geneva, CH, Nov. 21-30, 2011. cited by applicant International Search Report dated Apr. 12, 2013 in counterpart Application No. PCT/KR2013/000418 (5 pages, in Korean, with English language translation). cited by applicant ITU-T, “Recommendation H.264: Advanced video coding for generic audiovisual services,” Series H: Audiovisual and Multimedia Systems, Infrastructure of audiovisual services—Coding of Moving video, 2010, pp. 158-160. cited by applicant Japanese Office Action dated Aug. 18, 2015 in counterpart Japanese Application No. 2014-553258 (5 pages in Japanese). cited by applicant Na Zhang et al., “CE9: Improvement in temporal candidate of merge mode and AMVP”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G343, 7.sup.th Meeting: Geneva, CH, Nov. 21-30, 2011. cited by applicant Nakamura, Hiroya, et al. “Unification of derivation process for merge mode and MVP.” document JCTVC-F419 (2011) (10 pages). cited by applicant Thomas Wiegand, et al., “Overview of the H.264/AVC Video Coding Standard,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 13, No. 7, Jul. 2003, pp. 560-576 (17 pages, in English). cited by applicant Toshiyasu Sugio et al., “Non-CE9/Non-CE13: Simplification on AMVP/Merge”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G542, 7.sup.th Meeting: Geneva, CH, Nov. 21-30, 2011. cited by applicant Toshiyasu Sugio, et al., “Non-CE9: Simplified scaling calculation method for temporal/spatial MVP of AMVP/Merge,” Proceedings of the 7.sup.th Meeting of the Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-G541 Geneva, CH, Nov. 2011, pp. 1-5 (5 pages, in English). cited by applicant Zhuo et al., “A study on HM3.0 parsing throughput issue,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-F068, 6th Meeting: Torino, IT, Jul. 14-22, 2011. cited by applicant Toshiyasu Sugio et al., “Parsing Robustness for Merge/AMVP”, Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-F470, 6th Meeting: Torino, IT, Jul. 14-22, 2011. cited by applicant Sung-Chang Lim et al., “Non-CE9: Simplification on AMVP list construction,” Joint Collaborative Team on Video Coding (JCT-VC) of ITU-T SG16 WP3 and ISO/IEC JTC1/SC29/WG11, Document: JCTVC-H0239, 8th Meeting: San Jose, CA, USA, Feb. 1-10, 2012. cited by applicant European Patent Office Search Report for Application No. 21202076.2 dated Feb. 10, 2022 (5 pages). cited by applicant |
| Assistant Examiner: | Brumfield, Shanika M |
| Primary Examiner: | Czekaj, Dave |
| Attorney, Agent or Firm: | Michael Best & Friedrich LLP |
| Accession Number: | edspgr.11470342 |
| Database: | USPTO Patent Grants |
| Abstract: | A method for decoding an image according to the present invention comprises the steps of: decoding a residual block by quantizing and inverse transforming an entropy-decoded residual block; generating a prediction block via motion compensation; and decoding an image by adding the decoded residual block to the prediction block, wherein on the basis of the maximum number of motion vector candidates of the motion vector candidate list related to the prediction block, a motion vector candidate list is adjusted by adding a particular motion vector candidate or by discarding a portion from among the motion vector candidates, and in the prediction block generation step, a prediction motion vector of the prediction block is determined on the basis of the adjusted motion vector candidate list. Accordingly, the complexity of arithmetic operations is reduced during encoding/decoding of an image. |
|---|