US20100208796A1 - Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system - Google Patents

Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system Download PDF

Info

Publication number
US20100208796A1
US20100208796A1 US12/734,023 US73402308A US2010208796A1 US 20100208796 A1 US20100208796 A1 US 20100208796A1 US 73402308 A US73402308 A US 73402308A US 2010208796 A1 US2010208796 A1 US 2010208796A1
Authority
US
United States
Prior art keywords
view
equal
max
parameters
function block
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/734,023
Inventor
Jiancong Luo
Peng Yin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Thomson Licensing LLC
Original Assignee
Thomson Licensing LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing LLC filed Critical Thomson Licensing LLC
Priority to US12/734,023 priority Critical patent/US20100208796A1/en
Assigned to THOMSON LICENSING reassignment THOMSON LICENSING ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: YIN, PENG, LUO, JIANCONG
Publication of US20100208796A1 publication Critical patent/US20100208796A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

There are provided methods and apparatus for incorporating video usability information (VUI) within multi-view video coding (MVC). An apparatus includes an encoder for encoding multi-view video content by specifying video usability information for at least one selected from: individual views, individual temporal levels in a view, and individual operating points. Further, an apparatus includes a decoder for decoding multi-view video content by specifying video usability information for at least one selected from: individual views, individual temporal levels in a view, and individual operating points.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims the benefit of U.S. Provisional Application Ser. No. 60/977,709, filed Oct. 5, 2007, which is incorporated by reference herein in its entirety. Further, this application is related to the non-provisional application, Attorney Docket No. PU070239, entitled “METHODS AND APPARATUS FOR INCORPORATING VIDEO USABILITY INFORMATION (VUI) WITHIN A MULTI-VIEW VIDEO (MVC) CODING SYSTEM”, which also claims the benefit of U.S. Provisional Application Ser. No. 60/977,709, filed Oct. 5, 2007, and which is commonly assigned, incorporated by reference herein, and concurrently filed herewith.
  • TECHNICAL FIELD
  • The present principles relate generally to video encoding and decoding and, more particularly, to methods and apparatus for incorporating video usability information (VUI) within multi-view video coding (MVC).
  • BACKGROUND
  • The International Organization for Standardization/International Electrotechnical Commission (ISO/IEC) Moving Picture Experts Group-4 (MPEG-4) Part 10 Advanced Video Coding (AVC) standard/International Telecommunication Union, Telecommunication Sector (ITU-T) H.264 recommendation (hereinafter the “MPEG-4 AVC standard”) specifies syntax and semantics of video usability information (VUI) parameters of sequence parameter sets. Video usability information includes information of aspect ratio, over-scanning, video signal type, chroma location, timing, network abstraction layer (NAL) hypothetical reference decoder (HRD) parameters, video coding layer (VCL) hypothetical reference decoder parameters, bitstream restriction, and so forth. Video usability information provides extra information for a corresponding bitstream to permit a wider application for a user. For example, in bitstream restriction information, video usability information specifies: (1) if the motion is over a picture boundary; (2) the maximal bytes per picture; (3) the maximal bits per macroblock; (4) the maximal motion vector length (horizontal and vertical); (5) the number of reordering frames; and (6) the maximal decoded frame buffer size. When the decoder sees the information, instead of using the “level” information to set the decoding requirement, which in general is higher than what the bitstream actually requires, the decoder can customize its decoding operation based on the tighter limit.
  • Multi-view video coding (MVC) is an extension to the MPEG-4 AVC Standard. In multi-view video coding, video images for multiple views can be encoded by exploiting the correlation between views. Among all views, one view is the base view, which is MPEG-4 AVC Standard compatible and cannot be predicted from the other views. The other views are referred to as non-base views. Non-base views can be predictively encoded from the base view and other non-base views. Each view can be temporally sub-sampled. A temporal subset of a view can be identified by a temporal_id syntax element. A temporal level of a view is one representation of the video signal. There can be different combinations of views and temporal levels in a multi-view video coded bitstream. Each combination is called an operation point. Sub-bitstreams, corresponding to the operation points, may be extracted from the bitstream.
  • SUMMARY
  • These and other drawbacks and disadvantages of the prior art are addressed by the present principles, which are directed to methods and apparatus for incorporating video usability information (VUI) within multi-view video coding (MVC).
  • According to an aspect of the present principles, there is provided an apparatus. The apparatus includes an encoder for encoding multi-view video content by specifying video usability information for at least one of individual views, individual temporal levels in a view, and individual operating points.
  • According to another aspect of the present principles, there is provided a method. The method includes encoding multi-view video content by specifying video usability information for at least one of individual views, individual temporal levels in a view, and individual operating points.
  • According to yet another aspect of the present principles, there is provided an apparatus. The apparatus includes a decoder for decoding multi-view video content by specifying video usability information for at least one of individual views, individual temporal levels in a view, and individual operating points.
  • According to still another aspect of the present principles, there is provided a method. The method includes decoding multi-view video content by specifying video usability information for at least one of individual views, individual temporal levels in a view, and individual operating points.
  • These and other aspects, features and advantages of the present principles will become apparent from the following detailed description of exemplary embodiments, which is to be read in connection with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present principles may be better understood in accordance with the following exemplary figures, in which:
  • FIG. 1 is a block diagram for an exemplary Multi-view Video Coding (MVC) encoder to which the present principles may be applied, in accordance with an embodiment of the present principles;
  • FIG. 2 is a block diagram for an exemplary Multi-view Video Coding (MVC) decoder to which the present principles may be applied, in accordance with an embodiment of the present principles;
  • FIG. 3 is a flow diagram for an exemplary method for encoding bitstream restriction parameters for each view, using a mvc_vui_parameters_extension( ) syntax element, in accordance with an embodiment of the present principles;
  • FIG. 4 is a flow diagram for an exemplary method for decoding bitstream restriction parameters for each view, using a mvc_vui_parameters_extension( ) syntax element, in accordance with an embodiment of the present principles;
  • FIG. 5 is a flow diagram for an exemplary method for encoding bitstream restriction parameters for each temporal level in each view, using a mvc_vui_parameters_extension( ) syntax element, in accordance with an embodiment of the present principles;
  • FIG. 6 is a flow diagram for an exemplary method for decoding bitstream restriction parameters for each temporal level in each view, using a mvc_vui_parameters_extension( ) syntax element, in accordance with an embodiment of the present principles;
  • FIG. 7 is a flow diagram for an exemplary method for encoding bitstream restriction parameters for each operation point, using a view_scalability_parameters_extension( ) syntax element, in accordance with an embodiment of the present principles; and
  • FIG. 8 is a flow diagram for an exemplary method for decoding bitstream restriction parameters for each operation point, using a view_scalability_parameters_extension( ) syntax element, in accordance with an embodiment of the present principles.
  • DETAILED DESCRIPTION
  • The present principles are directed to methods and apparatus for incorporating video usability information (VUI) within multi-view video coding (MVC).
  • The present description illustrates the present principles. It will thus be appreciated that those skilled in the art will be able to devise various arrangements that, although not explicitly described or shown herein, embody the present principles and are included within its spirit and scope.
  • All examples and conditional language recited herein are intended for pedagogical purposes to aid the reader in understanding the present principles and the concepts contributed by the inventor(s) to furthering the art, and are to be construed as being without limitation to such specifically recited examples and conditions.
  • Moreover, all statements herein reciting principles, aspects, and embodiments of the present principles, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally; it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future, i.e., any elements developed that perform the same function, regardless of structure.
  • Thus, for example, it will be appreciated by those skilled in the art that the block diagrams presented herein represent conceptual views of illustrative circuitry embodying the present principles. Similarly, it will be appreciated that any flow charts, flow diagrams, state transition diagrams, pseudocode, and the like represent various processes which may be substantially represented in computer readable media and so executed by a computer or processor, whether or not such computer or processor is explicitly shown.
  • The functions of the various elements shown in the figures may be provided through the use of dedicated hardware as well as hardware capable of executing software in association with appropriate software. When provided by a processor, the functions may be provided by a single dedicated processor, by a single shared processor, or by a plurality of individual processors, some of which may be shared. Moreover, explicit use of the term “processor” or “controller” should not be construed to refer exclusively to hardware capable of executing software, and may implicitly include, without limitation, digital signal processor (“DSP”) hardware, read-only memory (“ROM”) for storing software, random access memory (“RAM”), and non-volatile storage.
  • Other hardware, conventional and/or custom, may also be included. Similarly, any switches shown in the figures are conceptual only. Their function may be carried out through the operation of program logic, through dedicated logic, through the interaction of program control and dedicated logic, or even manually, the particular technique being selectable by the implementer as more specifically understood from the context.
  • In the claims hereof, any element expressed as a means for performing a specified function is intended to encompass any way of performing that function including, for example, a) a combination of circuit elements that performs that function or b) software in any form, including, therefore, firmware, microcode or the like, combined with appropriate circuitry for executing that software to perform the function. The present principles as defined by such claims reside in the fact that the functionalities provided by the various recited means are combined and brought together in the manner which the claims call for. It is thus regarded that any means that can provide those functionalities are equivalent to those shown herein.
  • Reference in the specification to “one embodiment” or “an embodiment” of the present principles means that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment of the present principles. Thus, the appearances of the phrase “in one embodiment” or “in an embodiment” appearing in various places throughout the specification are not necessarily all referring to the same embodiment.
  • It is to be appreciated that the use of the terms “and/or” and “at least one of”, for example, in the cases of “A and/or B” and “at least one of A and B”, is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of both options (A and B). As a further example, in the cases of “A, B, and/or C” and “at least one of A, B, and C”, such phrasing is intended to encompass the selection of the first listed option (A) only, or the selection of the second listed option (B) only, or the selection of the third listed option (C) only, or the selection of the first and the second listed options (A and B) only, or the selection of the first and third listed options (A and C) only, or the selection of the second and third listed options (B and C) only, or the selection of all three options (A and B and C). This may be extended, as readily apparent by one of ordinary skill in this and related arts, for as many items listed.
  • Multi-view video coding (MVC) is the compression framework for the encoding of multi-view sequences. A Multi-view Video Coding (MVC) sequence is a set of two or more video sequences that capture the same scene from a different view point.
  • As interchangeably used herein, “cross-view” and “inter-view” both refer to pictures that belong to a view other than a current view.
  • Moreover, as used herein, “high level syntax” refers to syntax present in the bitstream that resides hierarchically above the macroblock layer. For example, high level syntax, as used herein, may refer to, but is not limited to, syntax at the slice header level, Supplemental Enhancement Information (SEI) level, Picture Parameter Set (PPS) level, Sequence Parameter Set (SPS) level and Network Abstraction Layer (NAL) unit header level.
  • Also, it is to be appreciated that while one or more embodiments of the present principles are described herein for illustrative purposes with respect to the multi-view video coding extension of the MPEG-4 AVC standard, the present principles are not limited to solely this extension and/or this standard and, thus, may be utilized with respect to other video coding standards, recommendations, and extensions thereof, while maintaining the spirit of the present principles.
  • Additionally, it is to be appreciated that while one or more embodiments of the present principles are described herein for illustrative purposes with respect to bitstream restriction information, the present principles are not limited to solely using bitstream restriction information as a type of video usability information and, thus, other types of video usability information that may be extended for use with respect to multi-video video coding may also be used in accordance with the present principles, while maintaining the spirit of the present principles.
  • Turning to FIG. 1, an exemplary Multi-view Video Coding (MVC) encoder is indicated generally by the reference numeral 100. The encoder 100 includes a combiner 105 having an output connected in signal communication with an input of a transformer 110. An output of the transformer 110 is connected in signal communication with an input of quantizer 115. An output of the quantizer 115 is connected in signal communication with an input of an entropy coder 120 and an input of an inverse quantizer 125. An output of the inverse quantizer 125 is connected in signal communication with an input of an inverse transformer 130. An output of the inverse transformer 130 is connected in signal communication with a first non-inverting input of a combiner 135. An output of the combiner 135 is connected in signal communication with an input of an intra predictor 145 and an input of a deblocking filter 150. An output of the deblocking filter 150 is connected in signal communication with an input of a reference picture store 155 (for view i). An output of the reference picture store 155 is connected in signal communication with a first input of a motion compensator 175 and a first input of a motion estimator 180. An output of the motion estimator 180 is connected in signal communication with a second input of the motion compensator 175
  • An output of a reference picture store 160 (for other views) is connected in signal communication with a first input of a disparity/illumination estimator 170 and a first input of a disparity/illumination compensator 165. An output of the disparity/illumination estimator 170 is connected in signal communication with a second input of the disparity/illumination compensator 165.
  • An output of the entropy decoder 120 is available as an output of the encoder 100. A non-inverting input of the combiner 105 is available as an input of the encoder 100, and is connected in signal communication with a second input of the disparity/illumination estimator 170, and a second input of the motion estimator 180. An output of a switch 185 is connected in signal communication with a second non-inverting input of the combiner 135 and with an inverting input of the combiner 105. The switch 185 includes a first input connected in signal communication with an output of the motion compensator 175, a second input connected in signal communication with an output of the disparity/illumination compensator 165, and a third input connected in signal communication with an output of the intra predictor 145.
  • A mode decision module 140 has an output connected to the switch 185 for controlling which input is selected by the switch 185.
  • Turning to FIG. 2, an exemplary Multi-view Video Coding (MVC) decoder is indicated generally by the reference numeral 200. The decoder 200 includes an entropy decoder 205 having an output connected in signal communication with an input of an inverse quantizer 210. An output of the inverse quantizer is connected in signal communication with an input of an inverse transformer 215. An output of the inverse transformer 215 is connected in signal communication with a first non-inverting input of a combiner 220. An output of the combiner 220 is connected in signal communication with an input of a deblocking filter 225 and an input of an intra predictor 230. An output of the deblocking filter 225 is connected in signal communication with an input of a reference picture store 240 (for view i). An output of the reference picture store 240 is connected in signal communication with a first input of a motion compensator 235.
  • An output of a reference picture store 245 (for other views) is connected in signal communication with a first input of a disparity/illumination compensator 250.
  • An input of the entropy decoder 205 is available as an input to the decoder 200, for receiving a residue bitstream. Moreover, an input of a mode module 260 is also available as an input to the decoder 200, for receiving control syntax to control which input is selected by the switch 255. Further, a second input of the motion compensator 235 is available as an input of the decoder 200, for receiving motion vectors. Also, a second input of the disparity/illumination compensator 250 is available as an input to the decoder 200, for receiving disparity vectors and illumination compensation syntax.
  • An output of a switch 255 is connected in signal communication with a second non-inverting input of the combiner 220. A first input of the switch 255 is connected in signal communication with an output of the disparity/illumination compensator 250. A second input of the switch 255 is connected in signal communication with an output of the motion compensator 235. A third input of the switch 255 is connected in signal communication with an output of the intra predictor 230. An output of the mode module 260 is connected in signal communication with the switch 255 for controlling which input is selected by the switch 255. An output of the deblocking filter 225 is available as an output of the decoder.
  • In the MPEG-4 AVC Standard, syntax and semantic parameters of the sequence parameter sets are specified for video usability information (VUI). This represents additional information that may be inserted into a bitstream to enhance the usability of the video for a wide variety of purposes. Video usability information includes information of aspect ratio, over-scanning, video signal type, chroma location, timing, network abstraction layer (NAL) hypothetical reference decoder (HRD) parameters, video coding layer (VCL) hypothetical reference decoder parameters, bitstream restriction, and so forth.
  • In accordance with one or more embodiments of the present principles, we use this existing video usability information field for new and different purposes than in the prior art and, further, extend its use to multi-view video coding (MVC). In our multi-view video coding scheme, the video usability information is extended so that it may be different between, for example, different views, different temporal levels in a view, or different operation points. Thus, in accordance with an embodiment, we specify video usability information according to one or more of, but not limited to, the following: specifying the video usability information for individual views; specifying the video usability information for individual temporal levels in a view; and specifying the video usability information for individual operation points separately.
  • In the MPEG-4 AVC Standard, a set that includes Video Usability Information (VUI) can be transmitted in a sequence parameter set (SPS). In accordance with an embodiment, we extend the concept of video usability information for use within a multi-view video coding (MVC) context. Advantageously, this allows different video usability information to be specified for different views, different temporal levels in a view, or different operation points in multi-view video coding. In an embodiment, we provide a novel approach in considering, modifying, and using bitstream restriction information in video usability information for multi-view video coding.
  • The bitstream restriction information in the MPEG-4 AVC Standard is specified in the vui_parameters( ) syntax element which is a part of the sequence_parameter_set( ) TABLE 1 illustrates the MPEG-4 AVC Standard syntax of vui_parameters( )
  • TABLE 1
    vui_parameters( ) { C Descriptor
    aspect_ratio_info_present_flag 0 u(1)
    ...
    bitstream_restriction_flag 0 u(1)
    if( bitstream_restriction_flag ) {
    motion_vectors_over_pic_boundaries_flag 0 u(1)
    max_bytes_per_pic_denom 0 ue(v)
    max_bits_per_mb_denom 0 ue(v)
    log2_max_mv_length_horizontal 0 ue(v)
    log2_max_mv_length_vertical 0 ue(v)
    num_reorder_frames 0 ue(v)
    max_dec_frame_buffering 0 ue(v)
    }
    }
  • The semantics of the syntax elements of bitstream restriction information are as follows:
      • bitstream_restriction_flag equal to 1 specifies that the following coded video sequence bitstream restriction parameters are present. bitstream_restriction_flag equal to 0 specifies that the following coded video sequence bitstream restriction parameters are not present.
      • motion_vectors_over_pic_boundaries_flag equal to 0 indicates that no sample outside the picture boundaries and no sample at a fractional sample position whose value is derived using one or more samples outside the picture boundaries are used to inter predict any sample. motion_vectors_over_pic_boundaries_flag equal to 1 indicates that one or more samples outside picture boundaries may be used in inter prediction.
      • When the motion_vectors_over_pic_boundaries_flag syntax element is not present, motion_vectors_over_pic_boundaries_flag value shall be inferred to be equal to 1.
      • max_bytes_per_pic_denom indicates a number of bytes not exceeded by the sum of the sizes of the virtual coding layer (VCL) network abstraction layer (NAL) units associated with any coded picture in the coded video sequence.
      • The number of bytes that represent a picture in the network abstraction layer unit stream is specified for this purpose as the total number of bytes of virtual coding layer network abstraction layer unit data (i.e., the total of the NumByteslnNALunit variables for the virtual coding layer network abstraction layer units) for the picture. The value of max_bytes_per_pic_denom shall be in the range of 0 to 16, inclusive.
  • Depending on max_bytes_per_pic_denom the following applies:
      • If max_bytes_per_pic_denom is equal to 0, then no limits are indicated.
      • Otherwise (max_bytes_per_pic_denom is not equal to 0), no coded picture shall be represented in the coded video sequence by more than the following number of bytes:

  • (PicSizeInMbs*RawMbBits)÷(8*max_bytes_per_pic_denom)
  • When the max_bytes_per_pic_denom syntax element is not present, the value of max_bytes_per_pic_denom shall be inferred to be equal to 2. The variable PicSizeInMbs is the number of macroblocks in the picture. The variable RawMbBits is derived as in sub-clause 7.4.2.1 of MPEG-4 AVC Standard.
      • max_bits_per_mb_denom indicates the maximum number of coded bits of macroblocklayer( ) data for any macroblock in any picture of the coded video sequence. The value of max_bits_per_mb_denom shall be in the range of 0 to 16, inclusive.
  • Depending on max_bits_per_mb_denom the following applies:
      • —If max_bits_per_mb_denom is equal to 0, then no limit is specified.
      • Otherwise (max_bits_per_mb_denom is not equal to 0), no coded macroblock_layer( ) shall be represented in the bitstream by more than the following number of bits.

  • (128+RawMbBits)÷max_bits_per_mb_denom
  • Depending on entropy_coding_mode_flag, the bits of macroblock_layer( ) data are counted as follows:
      • If entropy_coding_mode_flag is equal to 0, then the number of bits of macroblock_layer( ) data is given by the number of bits in the macroblock_layer( ) syntax structure for a macroblock.
      • Otherwise (entropy_coding_mode_flag is equal to 1), the number of bits of macroblock_layer( ) data for a macroblock is given by the number of times read_bits (1) is called in sub-clauses 9.3.3.2.2 and 9.3.3.2.3 of he MPEG-4 AVC Standard when parsing the macroblock_layer( ) associated with the macroblock.
      • When the max_bits_per_mb_denom is not present, the value of max_bits_per_mb_denom shall be inferred to be equal to 1.
      • log 2_max_mv_length_horizontal and log 2_max_mv_length_vertical indicate the maximum absolute value of a decoded horizontal and vertical motion vector component, respectively, in ¼ luma sample units, for all pictures in the coded video sequence. A value of n asserts that no value of a motion vector component shall exceed the range from −2n to 2n−1, inclusive, in units of ¼ luma sample displacement. The value of log 2_max_mv_length_horizontal shall be in the range of 0 to 16, inclusive. The value of log 2_max_mv_length_vertical shall be in the range of 0 to 16, inclusive. When log 2_max_mv_length_horizontal is not present, the values of log 2_max_mv_length_horizontal and log 2_max_mv_length_vertical shall be inferred to be equal to 16. It is to be noted that the maximum absolute value of a decoded vertical or horizontal motion vector component is also constrained by profile and level limits as specified in Annex A of the MPEG-4 AVC Standard.
      • num_reorder_frames indicates the maximum number of frames, complementary field pairs, or non-paired fields that respectively precede any frame, complementary field pair, or non-paired field in the coded video sequence in decoding order and follow it in output order. The value of num_reorder_frames shall be in the range of 0 to max_dec_frame_buffering, inclusive. When the num_reorder_frames syntax element is not present, the value of num_reorder_frames shall be inferred as follows:
        • —If profile_idc is equal to 44, 100, 110, 122, or 244 and constraint_set3_flag is equal to 1, then the value of num_reorder_frames shall be inferred to be equal to 0.
        • Otherwise (profile_idc is not equal to 44, 100, 110, 122, or 244 or constraint_set3_flag is equal to 0), the value of num_reorder_frames shall be inferred to be equal to max_dec_frame_bufferingMaxDpbSize.
      • max_dec_frame_buffering specifies the required size of the hypothetical reference decoder decoded picture buffer (DPB) in units of frame buffers. The coded video sequence shall not require a decoded picture buffer with size of more than Max(1, max_dec_frame_buffering) frame buffers to enable the output of decoded pictures at the output times specified by dpb_output_delay of the picture timing Supplemental Enhancement Information (SEI) messages. The value of max_dec_frame_buffering shall be in the range of num_ref_frames to MaxDpbSize (as specified in subclause A.3.1 or A.3.2 of the MPEG-4 AVC Standard), inclusive. When the max_dec_frame_buffering syntax element is not present, the value of max_dec_frame_buffering shall be inferred as follows:
        • —If profile_idc is equal to 44 or 244 and constraint_set3_flag is equal to 1, then the value of max_dec_frame_buffering shall be inferred to be equal to 0.
        • Otherwise (profile_idc is not equal to 44 or 244 or constraint_set3_flag is equal to 0), the value of max_dec_frame_buffering shall be inferred to be equal to MaxDpbSize.
  • In multi-view video coding, the bitstream restriction parameters customize the decoding operation of a sub-stream based on tighter limits. Therefore, the bitstream restriction parameters shall be allowed to be specified for each extractable sub-stream of a multi-view video coded bitstream. In accordance with an embodiment, we propose to specify bitstream restriction information for each view, for each temporal level in a view, and/or for each operation point.
  • Specifying Bitstream Restriction Parameters for Each View.
  • Bitstream restriction parameters can be specified for each view. We propose the syntax of mvc_vui_parameters_extension, which is a part of subset_sequence_parameter_set. TABLE 2 illustrates the syntax of mvc_vui_parameters_extension.
      • mvc_vui_parameters_extension( ) loops over all the views that are associated to this subset_sequence_parameter_set. The view_id of each view and the bitstream restriction parameters of each view are specified inside the loop.
  • TABLE 2
    mvc_vui_parameters_extension( ) { C Descriptor
    num_views_minus1 0 ue(v)
    for( i = 0; i <= num_views_minus1; i++ ) {
    view_id[ i ] 0 u(3)
    bitstream_restriction_flag[ i ] 0 u(1)
    if( bitstream_restriction_flag[ i ] ) {
    motion_vectors_over_pic_boundaries_flag[ i ] 0 u(1)
    max_bytes_per_pic_denom[ i ] 0 ue(v)
    max_bits_per_mb_denom[ i ] 0 ue(v)
    log2_max_mv_length_horizontal[ i ] 0 ue(v)
    log2_max_mv_length_vertical[ i ] 0 ue(v)
    num_reorder_frames[ i ] 0 ue(v)
    max_dec_frame_buffering[ i ] 0 ue(v)
    }
    }
    }
  • The semantics of the bitstream restriction syntax elements are as follows:
      • bitstream_restriction_flag [i] specifies the value of bitstream restriction_flag of the view having view_id[i] equal to view_id. motion_vectors_over_pic_boundaries_flag [i] specifies the value of motion_vectors_over_pic_boundaries_flag of the view having view_id[i] equal to view_id. When the motion_vectors_over_pic_boundaries_flag[i] syntax element is not present, the value of motion_vectors_over_pic_boundaries_flag for the view having view_id[i] equal to view_id shall be inferred to be equal to 1.
      • max_bytes_per_pic_denom [i] specifies the max_bytes_per_pic_denom value of the view having view_id[i] equal to view_id. When the max_bytes_per_pic_denom[i] syntax element is not present, the value of max_bytes_per_pic_denom of the view having view_id[i] equal to view_id shall be inferred to be equal to 2.
      • max_bits_per_mb_denom [i] specifies the max_bits_per_mb_denom value of the view having view_id[i] equal to view_id. When the max_bits_per_mb_denom[i] is not present, the value of max_bits_per_mb_denom of the view having view_id[i] equal to view_id shall be inferred to be equal to 1.
      • log 2_max_mv_length_horizontal [i] and log 2_max_mv_length_vertical [i] respectively specify the values of log 2_max_mv_length_horizontal and log 2_max_mv_length_vertical of the view having view_id[i] equal to view_id. When log 2_max_mv_length_horizontal[i] is not present, the values of log 2_max_mv_length_horizontal and log 2_max_mv_length_vertical of the view having view_id[i] equal to view_id shall be inferred to be equal to 16.
      • num_reorder_frames [i] specifies the value of num_reorder_frames of the view having view_id[i] equal to view_id. The value of num_reorder_frames[i] shall be in the range of 0 to max_dec_frame_buffering, inclusive. When the num_reorder_frames[i] syntax element is not present, the value of num_reorder_frames of the view having view_id[i] equal to view_id shall be inferred to be equal to max_dec_frame_buffering.
      • max_dec_frame_buffering [i] specifies the value of max_dec_frame_buffering of the view having view_id[i] equal to view_id. The value of max_dec_frame_buffering[i] shall be in the range of num_ref_frames[i] to MaxDpbSize (as specified in sub-clause A.3.1 or A.3.2 in the MPEG-4 AVC Standard), inclusive. When the max_dec_frame_buffering[i] syntax element is not present, the value of max_dec_frame_buffering of the view having view_id[i] equal to view_id shall be inferred to be equal to MaxDpbSize.
  • Turning to FIG. 3, an exemplary method for encoding bitstream restriction parameters for each view, using a mvc_vui_parameters_extension( ) syntax element, is indicated generally by the reference numeral 300.
  • The method 300 includes a start block 305 that passes control to a function block 310. The function block 310 sets a variable M equal to a number of views minus one, and passes control to a function block 315. The function block 315 writes the variable M to a bitstream, and passes control to a function block 320. The function block 320 sets a variable i equal to zero, and passes control to a function block 325. The function block 325 writes a view_id[i] syntax element, and passes control to a function block 330. The function block 330 writes a bitstream_restriction_flag[i] syntax element, and passes control to a decision block 335. The decision block 335 determines whether or not the bitstream restriction_flag[i] syntax element is equal to zero. If so, then control is passed to a decision block 345. Otherwise, control is passed to a function block 340.
  • The function block 340 writes the bitstream restriction parameters of view i, and passes control to the decision block 345. The decision block 345 determines whether or not the variable i is equal to the variable M. If so, then control is passed to an end block 399. Otherwise, control is passed to a function block 350.
  • The function block 350 sets the variable i equal to i plus one, and returns control to the function block 325.
  • Turning to FIG. 4, an exemplary method for decoding bitstream restriction parameters for each view, using a mvc_vui_parameters_extension( ) syntax element, is indicated generally by the reference numeral 400.
  • The method 400 includes a start block 405 that passes control to a function block 407. The function block 407 reads a variable M from a bitstream, and passes control to a function block 410. The function block 410 sets the number of views equal to the variable M plus one, and passes control to a function block 420. The function block 420 sets a variable i equal to zero, and passes control to a function block 425. The function block 425 reads a view_id[i] syntax element, and passes control to a function block 430. The function block 430 reads a bitstream_restriction_flag[i] syntax element, and passes control to a decision block 435. The decision block 435 determines whether or not the bitstream restriction_flag[i] syntax element is equal to zero. If so, then control is passed to a decision block 445. Otherwise, control is passed to a function block 440.
  • The function block 440 reads the bitstream restriction parameters of view i, and passes control to the decision block 445. The decision block 445 determines whether or not the variable i is equal to the variable M. If so, then control is passed to an end block 499. Otherwise, control is passed to a function block 450.
  • The function block 450 sets the variable i equal to i plus one, and returns control to the function block 425.
  • Specifying Bitstream Restriction Parameters for Each Temporal Level of Each View.
  • Bitstream restriction parameters can be specified for each temporal level of each view. We propose the syntax of mvc_vui_parameters_extension, which is a part of subset_sequence_parameter_set. TABLE 3 illustrates the syntax of mvc_vui_parameters_extension.
  • TABLE 3
    mvc_vui_parameters_extension( ) { C Descriptor
    num_views_minus1 0 ue(v)
    for( i = 0; i <= num_views_minus1; i++ ) {
    view_id[ i ] 0 u(3)
    num_temporal_layers_in_view_minus1[ i ] 0 ue(v)
    for( j = 0; j <= num_temporal_level_in_view_minus1; j++ ) {
    temporal_id[ i ][ j ]
    bitstream_restriction_flag[ i ][ j ] 0 u(1)
    if( bitstream_restriction_flag[ i ][ j ] ) {
    motion_vectors_over_pic_boundaries_flag[ i ][ j ] 0 u(1)
    max_bytes_per_pic_denom[ i ][ j ] 0 ue(v)
    max_bits_per_mb_denom[ i ][ j ] 0 ue(v)
    log2_max_mv_length_horizontal[ i ][ j ] 0 ue(v)
    log2_max_mv_length_vertical[ i ][ j ] 0 ue(v)
    num_reorder_frames[ i ][ j ] 0 ue(v)
    max_dec_frame_buffering[ i ][ j ] 0 ue(v)
    }
    }
    }
    }
  • The semantics of the bitstream restriction syntax elements are as follows:
      • bitstream_restriction_flag [i][j] specifies the value of bitstream restriction_flag of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id. motion_vectors_over_pic_boundaries_flag [i][j] specifies the value of motion_vectors_over_pic_boundaries_flag of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id. When the motion_vectors_over_pic_boundaries_flag[i] syntax element is not present, motion_vectors_over_pic_boundaries_flag value of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id shall be inferred to be equal to 1.
      • max_bytes_per_pic_denom [i][j] specifies the max_bytes_per_pic_denom value of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id. When the max_bytes_per_pic_denom[i] syntax element is not present, the value of max_bytes_per_pic_denom of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id shall be inferred to be equal to 2.
      • max_bits_per_mb_denom [i][j] specifies the max_bits_per_mb_denom value of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id. When the max_bits_per_mb_denom[i] is not present, the value of max_bits_per_mb_denom of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id shall be inferred to be equal to 1.
      • log 2_max_mv_length_horizontal [i][j] and log 2_max_mv_length_vertical [i][j] respectively specify the values of log 2_max_mv_length_horizontal and log 2_max_mv_length_vertical of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id. When log 2_max_mv_length_horizontal[i] is not present, the values of log 2_max_mv_length_horizontal and log 2_max_mv_length_vertical of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id shall be inferred to be equal to 16.
      • num_reorder_frames [i][j] specifies the value of num_reorder_frames of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id. The value of num_reorder_frames[i] shall be in the range of 0 to max_dec_frame_buffering, inclusive. When the num_reorder_frames[i] syntax element is not present, the value of num_reorder_frames of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id shall be inferred to be equal to max_dec_frame_buffering.
      • max_dec_frame_buffering [i][j] specifies the value of max_dec_frame_buffering of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id. The value of max_dec_frame_buffering[i] shall be in the range of num_ref_frames[i] to MaxDpbSize (as specified in sub-clause A.3.1 or A.3.2 in the MPEG-4 AVC Standard), inclusive. When the max_dec_frame_buffering[i] syntax element is not present, the value of max_dec_frame_buffering of the temporal level having temporal_id[i][j] equal to temporal_id in the view having view_id[i] equal to view_id shall be inferred to be equal to MaxDpbSize.
  • In mvc_vui_parameters_extension( ) two loops are executed. The outer loop loops over all the views associated to the subset_sequence_parameter_set. The view_id for the number of temporal levels of each view is specified in the outer loop. The inner loop loops over all the temporal levels of a view. The bitstream restriction information is specified in the inner loop.
  • Turning to FIG. 5, an exemplary method for encoding bitstream restriction parameters for each temporal level in each view, using a mvc_vui_parameters_extension( ) syntax element, is indicated generally by the reference numeral 500.
  • The method 500 includes a start block 505 that passes control to a function block 510. The function block 510 sets a variable M equal to a number of views minus one, and passes control to a function block 515. The function block 515 writes the variable M to a bitstream, and passes control to a function block 520. The function block 520 sets a variable i equal to zero, and passes control to a function block 525. The function block 525 writes a view_id[i] syntax element, and passes control to a function block 530. The function block 530 sets a variable N equal to a number of temporal levels in view i minus 1, and passes control to a function block 535. The function block 535 writes the variable N to the bitstream, and passes control to a function block 540. The function block 540 sets a variable j equal to zero, and passes control to a function block 545. The function block 545 writes a temporal_id[i][j] syntax element, and passes control to a function block 550. The function block 550 writes a bitstream_restriction_flag[i][j] syntax element, and passes control to a decision block 555. The decision block 555 determines whether or not the bitstream restriction_flag[i][j] syntax element is equal to zero. If so, then control is passed to a decision block 565. Otherwise, control is passed to a function block 560.
  • The function block 560 writes the bitstream restriction parameters of temporal level j in view i, and passes control to the decision block 565. The decision block 565 determines whether or not the variable j is equal to the variable N. If so, then control is passed to a decision block 570. Otherwise, control is passed to a function block 575.
  • The decision block 570 determines whether or not the variable i is equal to the variable M. If so, then control is passed to an end block 599. Otherwise, control is passed to a function block 580.
  • The function block 580 sets the variable i equal to i plus one, and returns control to the function block 525.
  • The function block 575 sets the variable j equal to j plus one, and returns control to the function block 545.
  • Turning to FIG. 6, an exemplary method for decoding bitstream restriction parameters for each temporal level in each view, using a mvc_vui_parameters_extension( ) syntax element, is indicated generally by the reference numeral 600.
  • The method 600 includes a start block 605 that passes control to a function block 607. The function block 607 reads a variable M from a bitstream, and passes control to a function block 610. The function block 610 sets a number of views equal to M plus one, and passes control to a function block 620. The function block 620 sets a variable i equal to zero, and passes control to a function block 625. The function block 625 reads a view_id[i] syntax element, and passes control to a function block 627. The function block 627 reads a variable N from the bitstream, and passes control to a function block 630. The function block 630 sets a number of temporal levels in view i equal to N plus 1, and passes control to a function block 640. The function block 640 sets a variable j equal to zero, and passes control to a function block 645. The function block 645 reads a temporal_id[i][j] syntax element, and passes control to a function block 650. The function block 650 reads a bitstream_restriction_flag[i][j] syntax element, and passes control to a decision block 655. The decision block 655 determines whether or not the bitstream restriction_flag[i][j] syntax element is equal to zero. If so, then control is passed to a decision block 665. Otherwise, control is passed to a function block 660.
  • The function block 660 reads the bitstream restriction parameters of temporal level j in view i, and passes control to the decision block 665. The decision block 665 determines whether or not the variable j is equal to the variable N. If so, then control is passed to a decision block 670. Otherwise, control is passed to a function block 675.
  • The decision block 670 determines whether or not the variable i is equal to the variable M. If so, then control is passed to an end block 699. Otherwise, control is passed to a function block 680.
  • The function block 680 sets the variable i equal to i plus one, and returns control to the function block 625.
  • The function block 675 sets the variable j equal to j plus one, and returns control to the function block 645.
  • Specifying Bitstream Restriction Information for Each Operation Point
  • Bitstream restriction parameters can be specified for each operation point. We propose to convey the bitstream restriction parameters of each operation point in the view scalability information SEI message. The syntax of view scalability information SEI message can be modified as in TABLE 4. The syntax for bitstream restriction information is inserted in a loop that loops over all the operation points.
  • TABLE 4
    view_scalability_info( payloadSize ) { C Descriptor
    num_operation_points_minus1 5 ue(v)
    for( i = 0; i <= num_operation_points_minus1; i++ ) {
    operation_point_id[ i ] 5 ue(v)
    priority_id[ i ] 5 u(5)
    temporal_id[ i ] 5 u(3)
    num_active_views_minus1[ i ] 5 ue(v)
    for( j = 0; j <= num_active_views_minus1[ i ]; j++ )
    view_id[ i ][ j ] 5 ue(v)
    profile_level_info_present_flag[ i ] 5 u(1)
    bitrate_info_present_flag[ i ] 5 u(1)
    frm_rate_info_present_flag[ i ] 5 u(1)
    op_dependency_info_present_flag[ i ] 5 u(1)
    init_parameter_sets_info_present_flag[ i ] 5 u(1)
    bitstream_restriction_flag[ i ]
    if (profile_level_info_present_flag[ i ]) {
    op_profile_idc[ i ] 5 u(8)
    op_constraint_set0_flag[ i ] 5 u(1)
    op_constraint_set1_flag[ i ] 5 u(1)
    op_constraint_set2_flag[ i ] 5 u(1)
    op_constraint_set3_flag[ i ] 5 u(1)
    reserved_zero_4bits /* equal to 0 */ 5 u(4)
    op_level_idc[ i ] 5 u(8)
    } else
    profile_level_info_src_op_id_delta[ i ] ue(v)
    if( bitrate_info_present_flag[ i ]) {
    avg_bitrate[ i ] 5 u(16)
    max_bitrate[ i ] 5 u(16)
    max_bitrate_calc_window[ i ] 5 u(16)
    }
    if( frm_rate_info_present_flag[ i [ ) {
    constant_frm_rate_idc[ i ] 5 u(2)
    avg_frm_rate[ i ] 5 u(16)
    } else
    frm_rate_info_src_op_id_delta[ i ] 5 ue(v)
    if(op_dependency_info_present_flag[ i ]) {
    num_directly_dependent_ops[ i ] 5 ue(v)
    for( j = 0; j < num_directly_dependent_ops[ i ]; j++ ) {
    directly_dependent_op_id_delta_minus1[ i ][ j ] 5 ue(v)
    } else
    op_dependency_info_src_op_id_delta[ i ] 5 ue(v)
    if( init_parameter_sets_info_present_flag[ i ] ) {
    num_init_seq_parameter_set_minus1[ i ] 5 ue(v)
    for( j = 0; j <= num_init_seq_parameter_set_minus1[ i ]; j++ )
    init_seq_parameter_set_id_delta[ i ][ j ] 5 ue(v)
    num_init_pic_parameter_set_minus1 [ i ] 5 ue(v)
    for( j = 0; j <= num_init_pic_parameter_set_minus1 [ i ]; j++ )
    init_pic_parameter_set_id_delta[ i ][ j ] 5 ue(v)
    } else
    init_parameter_sets_info_src_op_id_delta[ i ] 5 ue(v)
    if ( bitstream_restriction_flag[ i ] ) {
    motion_vectors_over_pic_boundaries_flag[ i ] 0 u(1)
    max_bytes_per_pic_denom[ i ] 0 ue(v)
    max_bits_per_mb_denom[ i ] 0 ue(v)
    log2_max_mv_length_horizontal[ i ] 0 ue(v)
    log2_max_mv_length_vertical[ i ] 0 ue(v)
    num_reorder_frames[ i ] 0 ue(v)
    max_dec_frame_buffering[ i ] 0 ue(v)
    }
    }
    }
  • The semantics of the bitstream restriction syntax elements are as follows:
      • bitstream_restriction_flag [i] specifies the value of bitstream restriction_flag of the operation point having operation_point_id[i] equal to operation_point_id.
      • motion_vectors_over_pic_boundaries_flag [i] specifies the value of motion_vectors_over_pic_boundaries_flag of the operation point having operation_point_id[i] equal to operation_point_id. When the motion_vectors_over_pic_boundaries_flag[i] syntax element is not present, motion_vectors_over_pic_boundaries_flag value of the operation point having operation_point_id[i] equal to operation_point_id shall be inferred to be equal to 1.
      • max_bytes_per_pic_denom [i] specifies the max_bytes_per_pic_denom value of the operation point having operation_point_id[i] equal to operation_point_id. When the max_bytes_per_pic_denom[i] syntax element is not present, the value of max_bytes_per_pic_denom of the operation point having operation_point_id[i] equal to operation_point_id shall be inferred to be equal to 2.
      • max_bits_per_mb_denom [i] specifies the max_bits_per_mb_denom value of the operation point having operation_point_id[i] equal to operation_point_id. When the max_bits_per_mb_denom[i] is not present, the value of max_bits_per_mb_denom of the operation point having operation_point_id[i] equal to operation_point_id shall be inferred to be equal to 1.
      • log 2_max_mv_length_horizontal [i] and log 2_max_mv_length_vertical [i] respectively specify the value of log 2_max_mv_length_horizontal and the value of log 2_max_mv_length_vertical of the operation point having operation_point_id[i] equal to operation_point_id. When log 2_max_mv_length_horizontal[i] is not present, the values of log 2_max_mv_length_horizontal and log 2_max_mv_length_vertical of the operation point having operation_point_id[i] equal to operation_point_id shall be inferred to be equal to 16.
      • num_reorder_frames [i] specifies the value of num_reorder_frames of the operation point having operation_point_id[i] equal to operation_point_id. The value of num_reorder_frames[i] shall be in the range of 0 to max_dec_frame_buffering, inclusive. When the num_reorder_frames[i] syntax element is not present, the value of num_reorder_frames of the operation point having operation_point_id[i] equal to operation_point_id shall be inferred to be equal to max_dec_frame_buffering.
      • max_dec_frame_buffering [i] specifies the value of max_dec_frame_buffering of the operation point having operation_point_id[i] equal to operation_point_id. The value of max_dec_frame_buffering[i] shall be in the range of num_ref_frames[i] to MaxDpbSize (as specified in sub-clause A.3.1 or A.3.2 in the MPEG-4 AVC Standard), inclusive. When the max_dec_frame_buffering[i] syntax element is not present, the value of max_dec_frame_buffering of the operation point having operation_point_id[i] equal to operation_point_id shall be inferred to be equal to MaxDpbSize.
  • Turning to FIG. 7, an exemplary method for encoding bitstream restriction parameters for each operation point, using a view_scalability_parameters_extension( ) syntax element, is indicated generally by the reference numeral 700.
  • The method 700 includes a start block 705 that passes control to a function block 710. The function block 710 sets a variable M equal to a number of operation points minus one, and passes control to a function block 715. The function block 715 writes the variable M to a bitstream, and passes control to a function block 720. The function block 720 sets a variable i equal to zero, and passes control to a function block 725. The function block 725 writes an operation_point_id[i] syntax element, and passes control to a function block 730. The function block 730 writes a bitstream_restriction_flag[i] syntax element, and passes control to a decision block 735. The decision block 735 determines whether or not the bitstream restriction_flag[i] syntax element is equal to zero. If so, then control is passed to a decision block 745. Otherwise, control is passed to a function block 740.
  • The function block 740 writes the bitstream restriction parameters of operation point i, and passes control to the decision block 745. The decision block 745 determines whether or not the variable i is equal to the variable M. If so, then control is passed to an end block 799. Otherwise, control is passed to a function block 750.
  • The function block 750 sets the variable i equal to i plus one, and returns control to the function block 725.
  • Turning to FIG. 8, an exemplary method for decoding bitstream restriction parameters for each operation point, using a view_scalability_parameters_extension( ) syntax element, is indicated generally by the reference numeral 800.
  • The method 800 includes a start block 805 that passes control to a function block 807. The function block 807 reads a variable M from a bitstream, and passes control to a function block 810. The function block 810 sets a number of operation points equal to M plus one, and passes control to a function block 820. The function block 820 sets a variable i equal to zero, and passes control to a function block 825. The function block 825 reads an operation_point_id[i] syntax element, and passes control to a function block 830. The function block 830 reads a bitstream_restriction_flag[i] syntax element, and passes control to a decision block 835. The decision block 835 determines whether or not the bitstream restriction_flag[i] syntax element is equal to zero. If so, then control is passed to a decision block 845. Otherwise, control is passed to a function block 840.
  • The function block 840 reads the bitstream restriction parameters of operation point i, and passes control to the decision block 845. The decision block 845 determines whether or not the variable i is equal to the variable M. If so, then control is passed to an end block 899. Otherwise, control is passed to a function block 850.
  • The function block 850 sets the variable i equal to i plus one, and returns control to the function block 825.
  • A description will now be given of some of the many attendant advantages/features of the present invention, some of which have been mentioned above. For example, one advantage/feature is an apparatus that includes an encoder for encoding multi-view video content by specifying video usability information for at least one of individual views, individual temporal levels in a view, and individual operating points.
  • Another advantage/feature is the apparatus having the encoder as described above, wherein the parameters are specified in at least one high level syntax element.
  • Moreover, another advantage/feature is the apparatus having the encoder as described above, wherein the at least high level syntax element includes at least one of a mvc_vui_parameters_extension( ) syntax element, a mvc_scalability_info supplemental enhancement information syntax message, at least a portion of a sequence parameters set, a picture parameters set, and supplemental enhancement information.
  • Further, another advantage/feature is the apparatus having the encoder as described above, wherein at least a portion of the video usability information comprises bitstream restriction parameters.
  • These and other features and advantages of the present principles may be readily ascertained by one of ordinary skill in the pertinent art based on the teachings herein. It is to be understood that the teachings of the present principles may be implemented in various forms of hardware, software, firmware, special purpose processors, or combinations thereof.
  • Most preferably, the teachings of the present principles are implemented as a combination of hardware and software. Moreover, the software may be implemented as an application program tangibly embodied on a program storage unit. The application program may be uploaded to, and executed by, a machine comprising any suitable architecture. Preferably, the machine is implemented on a computer platform having hardware such as one or more central processing units (“CPU”), a random access memory (“RAM”), and input/output (“I/O”) interlaces. The computer platform may also include an operating system and microinstruction code. The various processes and functions described herein may be either part of the microinstruction code or part of the application program, or any combination thereof, which may be executed by a CPU. In addition, various other peripheral units may be connected to the computer platform such as an additional data storage unit and a printing unit.
  • It is to be further understood that, because some of the constituent system components and methods depicted in the accompanying drawings are preferably implemented in software, the actual connections between the system components or the process function blocks may differ depending upon the manner in which the present principles are programmed. Given the teachings herein, one of ordinary skill in the pertinent art will be able to contemplate these and similar implementations or configurations of the present principles.
  • Although the illustrative embodiments have been described herein with reference to the accompanying drawings, it is to be understood that the present principles is not limited to those precise embodiments, and that various changes and modifications may be effected therein by one of ordinary skill in the pertinent art without departing from the scope or spirit of the present principles. All such changes and modifications are intended to be included within the scope of the present principles as set forth in the appended claims.

Claims (12)

1. An apparatus, comprising:
a decoder for decoding multi-view video content by specifying video usability information for at least one selected from: individual views, individual temporal levels in a view, and individual operating points.
2. The apparatus of claim 1, wherein the parameters are specified in at least one high level syntax element.
3. The apparatus of claim 2, wherein the at least one high level syntax element comprises at least one of a mvc_vui_parameters_extension( ) syntax element, a mvc_scalability_info supplemental enhancement information syntax message, at least a portion of a sequence parameters set, a picture parameters set, and supplemental enhancement information.
4. The apparatus of claim 1, wherein at least a portion of the video usability information comprises bitstream restriction parameters.
5. A method, comprising:
decoding multi-view video content by specifying video usability information for at least one selected from: individual views, individual temporal levels in a view, and individual operating points.
6. The method of claim 5, wherein the parameters are specified in at least one high level syntax element.
7. The method of claim 6, wherein the at least one high level syntax element comprises at least one of a mvc_vui_parameters_extension( ) syntax element, a mvc_scalability_info supplemental enhancement information syntax message, at least a portion of a sequence parameters set, a picture parameters set, and supplemental enhancement information.
8. The method of claim 5, wherein at least a portion of the video usability information comprises bitstream restriction parameters.
9. A video signal structure for video encoding, decoding, and transport; comprising:
multi-view video content encoded by specifying video usability information for at least one selected from: individual views, individual temporal levels in a view, and individual operating points.
10. The video signal structure of claim 9, wherein the parameters are specified in at least one high level syntax element.
11. The video signal structure of claim 10, wherein the at least one high level syntax element comprises at least one of a mvc_vui_parameters_extension( ) syntax element, a mvc_scalability_info supplemental enhancement information syntax message, at least a portion of a sequence parameters set, a picture parameters set, and supplemental enhancement information.
12. The video signal structure of claim 9, wherein at least a portion of the video usability information comprises bitstream restriction parameters.
US12/734,023 2007-10-05 2008-09-16 Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system Abandoned US20100208796A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/734,023 US20100208796A1 (en) 2007-10-05 2008-09-16 Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US97770907P 2007-10-05 2007-10-05
US12/734,023 US20100208796A1 (en) 2007-10-05 2008-09-16 Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system
PCT/US2008/010775 WO2009048502A2 (en) 2007-10-05 2008-09-16 Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system

Publications (1)

Publication Number Publication Date
US20100208796A1 true US20100208796A1 (en) 2010-08-19

Family

ID=40404801

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/734,023 Abandoned US20100208796A1 (en) 2007-10-05 2008-09-16 Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system
US12/733,979 Abandoned US20110038424A1 (en) 2007-10-05 2008-09-16 Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system

Family Applications After (1)

Application Number Title Priority Date Filing Date
US12/733,979 Abandoned US20110038424A1 (en) 2007-10-05 2008-09-16 Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system

Country Status (8)

Country Link
US (2) US20100208796A1 (en)
EP (2) EP2198619A2 (en)
JP (2) JP5264920B2 (en)
KR (3) KR101682322B1 (en)
CN (4) CN101889448B (en)
BR (10) BR122012021797A2 (en)
TW (6) TWI400957B (en)
WO (2) WO2009048503A2 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130114694A1 (en) * 2011-11-08 2013-05-09 Qualcomm Incorporated Parameter set groups for coded video data
US20130182755A1 (en) * 2012-01-14 2013-07-18 Qualcomm Incorporated Coding parameter sets and nal unit headers for video coding
US20140169448A1 (en) * 2012-12-19 2014-06-19 Qualcomm Incorporated Low-delay buffering model in video coding
US20150229945A1 (en) * 2012-09-28 2015-08-13 Sony Corporation Image processing device and method
US20150373356A1 (en) * 2014-06-18 2015-12-24 Qualcomm Incorporated Signaling hrd parameters for bitstream partitions
US9241158B2 (en) 2012-09-24 2016-01-19 Qualcomm Incorporated Hypothetical reference decoder parameters in video coding
US9467700B2 (en) 2013-04-08 2016-10-11 Qualcomm Incorporated Non-entropy encoded representation format
US20170048523A1 (en) * 2011-07-11 2017-02-16 Sharp Kabushiki Kaisha Method for encoding video
US9729898B2 (en) 2011-06-30 2017-08-08 Mircosoft Technology Licensing, LLC Reducing latency in video encoding and decoding
US10104386B2 (en) 2013-01-04 2018-10-16 Ge Video Compression, Llc Efficient scalable coding concept
US11582473B2 (en) 2013-04-08 2023-02-14 Ge Video Compression, Llc Coding concept allowing efficient multi-view/layer coding

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8948241B2 (en) * 2009-08-07 2015-02-03 Qualcomm Incorporated Signaling characteristics of an MVC operation point
KR101682137B1 (en) 2010-10-25 2016-12-05 삼성전자주식회사 Method and apparatus for temporally-consistent disparity estimation using texture and motion detection
KR20130058584A (en) 2011-11-25 2013-06-04 삼성전자주식회사 Method and apparatus for encoding image, and method and apparatus for decoding image to manage buffer of decoder
US10158873B2 (en) 2011-11-30 2018-12-18 Qualcomm Incorporated Depth component removal for multiview video coding (MVC) compatible three-dimensional video coding (3DVC)
WO2013151814A1 (en) 2012-04-06 2013-10-10 Vidyo, Inc. Level signaling for layered video coding
KR20130116782A (en) 2012-04-16 2013-10-24 한국전자통신연구원 Scalable layer description for scalable coded video bitstream
US9912941B2 (en) * 2012-07-02 2018-03-06 Sony Corporation Video coding system with temporal layers and method of operation thereof
US10110890B2 (en) 2012-07-02 2018-10-23 Sony Corporation Video coding system with low delay and method of operation thereof
US20140003534A1 (en) * 2012-07-02 2014-01-02 Sony Corporation Video coding system with temporal scalability and method of operation thereof
US9648352B2 (en) 2012-09-24 2017-05-09 Qualcomm Incorporated Expanded decoding unit definition
US9521393B2 (en) 2013-01-07 2016-12-13 Qualcomm Incorporated Non-nested SEI messages in video coding
CN104053008B (en) * 2013-03-15 2018-10-30 乐金电子(中国)研究开发中心有限公司 Video coding-decoding method and Video Codec based on composograph prediction
US20140301477A1 (en) * 2013-04-07 2014-10-09 Sharp Laboratories Of America, Inc. Signaling dpb parameters in vps extension and dpb operation
JP6669617B2 (en) * 2016-09-12 2020-03-18 ルネサスエレクトロニクス株式会社 Video processing system
CN106678778B (en) * 2017-02-08 2018-08-10 安徽中企能源管理有限公司 A kind of efficient cyclone environment-protection boiler
CN108933768B (en) * 2017-05-27 2021-06-08 成都鼎桥通信技术有限公司 Method and device for acquiring sending frame rate of video frame

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070211796A1 (en) * 2006-03-09 2007-09-13 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding multi-view video to provide uniform picture quality
US20080095228A1 (en) * 2006-10-20 2008-04-24 Nokia Corporation System and method for providing picture output indications in video coding
US20080175325A1 (en) * 2007-01-08 2008-07-24 Nokia Corporation System and method for providing and using predetermined signaling of interoperability points for transcoded media streams
US20090220010A1 (en) * 2006-09-07 2009-09-03 Seung Wook Park Method and Apparatus for Decoding/Encoding of a Video Signal
US20090279608A1 (en) * 2006-03-30 2009-11-12 Lg Electronics Inc. Method and Apparatus for Decoding/Encoding a Video Signal
US20100046619A1 (en) * 2006-08-25 2010-02-25 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US20100195738A1 (en) * 2007-04-18 2010-08-05 Lihua Zhu Coding systems

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1415479A1 (en) * 2001-08-02 2004-05-06 Koninklijke Philips Electronics N.V. Video coding method
WO2005013623A1 (en) 2003-08-05 2005-02-10 Koninklijke Philips Electronics N.V. Multi-view image generation
JP2005348093A (en) * 2004-06-03 2005-12-15 Sony Corp Image processor, program and method thereof
US20060146734A1 (en) * 2005-01-04 2006-07-06 Nokia Corporation Method and system for low-delay video mixing
EP1869891A4 (en) * 2005-04-13 2014-06-11 Coding, storage and signalling of scalability information
US8902989B2 (en) * 2005-04-27 2014-12-02 Broadcom Corporation Decoder system for decoding multi-standard encoded video
US7974517B2 (en) * 2005-10-05 2011-07-05 Broadcom Corporation Determination of decoding information
JP5192393B2 (en) * 2006-01-12 2013-05-08 エルジー エレクトロニクス インコーポレイティド Multi-view video processing
KR100754205B1 (en) * 2006-02-07 2007-09-03 삼성전자주식회사 Multi-view video encoding apparatus and method
AU2007243935A1 (en) * 2006-03-29 2007-11-08 Thomson Licensing Multi view video coding method and device
CN100471278C (en) * 2007-04-06 2009-03-18 清华大学 Multi-view video compressed coding-decoding method based on distributed source coding
CN100559877C (en) * 2007-04-27 2009-11-11 北京大学 A kind of network flow-medium player and method of supporting that multi-view point video is synthetic
WO2010017166A2 (en) 2008-08-04 2010-02-11 Dolby Laboratories Licensing Corporation Overlapped block disparity estimation and compensation architecture

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070211796A1 (en) * 2006-03-09 2007-09-13 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding multi-view video to provide uniform picture quality
US8411744B2 (en) * 2006-03-30 2013-04-02 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US20100215100A1 (en) * 2006-03-30 2010-08-26 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US8565319B2 (en) * 2006-03-30 2013-10-22 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US20090279608A1 (en) * 2006-03-30 2009-11-12 Lg Electronics Inc. Method and Apparatus for Decoding/Encoding a Video Signal
US20090296811A1 (en) * 2006-03-30 2009-12-03 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US20100020870A1 (en) * 2006-03-30 2010-01-28 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US8565303B2 (en) * 2006-03-30 2013-10-22 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US8170108B2 (en) * 2006-03-30 2012-05-01 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal
US20100260265A1 (en) * 2006-03-30 2010-10-14 Byeong Moon Jeon Method and Apparatus for Decoding/Encoding a Video Signal
US20100046619A1 (en) * 2006-08-25 2010-02-25 Han Suh Koo Method and apparatus for decoding/encoding a video signal
US8532178B2 (en) * 2006-08-25 2013-09-10 Lg Electronics Inc. Method and apparatus for decoding/encoding a video signal with inter-view reference picture list construction
US20090220010A1 (en) * 2006-09-07 2009-09-03 Seung Wook Park Method and Apparatus for Decoding/Encoding of a Video Signal
US20080095228A1 (en) * 2006-10-20 2008-04-24 Nokia Corporation System and method for providing picture output indications in video coding
US20080175325A1 (en) * 2007-01-08 2008-07-24 Nokia Corporation System and method for providing and using predetermined signaling of interoperability points for transcoded media streams
US20100195738A1 (en) * 2007-04-18 2010-08-05 Lihua Zhu Coding systems

Cited By (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10003824B2 (en) 2011-06-30 2018-06-19 Microsoft Technology Licensing, Llc Reducing latency in video encoding and decoding
EP4246967A3 (en) * 2011-06-30 2023-12-20 Microsoft Technology Licensing, LLC Reducing latency in video encoding and decoding
US9729898B2 (en) 2011-06-30 2017-08-08 Mircosoft Technology Licensing, LLC Reducing latency in video encoding and decoding
US9743114B2 (en) 2011-06-30 2017-08-22 Microsoft Technology Licensing, Llc Reducing latency in video encoding and decoding
US20170048523A1 (en) * 2011-07-11 2017-02-16 Sharp Kabushiki Kaisha Method for encoding video
US10390013B2 (en) * 2011-07-11 2019-08-20 Velos Media, Llc Method for encoding video
US11805253B2 (en) 2011-07-11 2023-10-31 Velos Media, Llc Processing a video frame having slices and tiles
US11451776B2 (en) 2011-07-11 2022-09-20 Velos Media, Llc Processing a video frame having slices and tiles
US10812799B2 (en) 2011-07-11 2020-10-20 Velos Media, Llc Method for encoding video
US20130114694A1 (en) * 2011-11-08 2013-05-09 Qualcomm Incorporated Parameter set groups for coded video data
US9451252B2 (en) * 2012-01-14 2016-09-20 Qualcomm Incorporated Coding parameter sets and NAL unit headers for video coding
JP2015507428A (en) * 2012-01-14 2015-03-05 クゥアルコム・インコーポレイテッドQualcomm Incorporated Coding parameter set and NAL unit header for video coding
KR101760165B1 (en) 2012-01-14 2017-07-20 퀄컴 인코포레이티드 Coding parameter sets and nal unit headers for video coding
US20130182755A1 (en) * 2012-01-14 2013-07-18 Qualcomm Incorporated Coding parameter sets and nal unit headers for video coding
RU2633117C2 (en) * 2012-01-14 2017-10-11 Квэлкомм Инкорпорейтед Encoding of parameter sets and nal unit headers for video encoding
US9351005B2 (en) 2012-09-24 2016-05-24 Qualcomm Incorporated Bitstream conformance test in video coding
US10021394B2 (en) 2012-09-24 2018-07-10 Qualcomm Incorporated Hypothetical reference decoder parameters in video coding
US9241158B2 (en) 2012-09-24 2016-01-19 Qualcomm Incorporated Hypothetical reference decoder parameters in video coding
AU2017251760B2 (en) * 2012-09-28 2019-10-10 Sony Corporation Image processing device and method
AU2020200093B2 (en) * 2012-09-28 2021-05-20 Sony Corporation Image processing device and method
US20220124361A1 (en) * 2012-09-28 2022-04-21 Sony Group Corporation Image processing device and method
US20150229945A1 (en) * 2012-09-28 2015-08-13 Sony Corporation Image processing device and method
US9924189B2 (en) * 2012-09-28 2018-03-20 Sony Corporation Image processing device and method
US20180152725A1 (en) * 2012-09-28 2018-05-31 Sony Corporation Image processing device and method
CN104854870A (en) * 2012-12-19 2015-08-19 高通股份有限公司 Low-delay buffering model in video coding
US20140169448A1 (en) * 2012-12-19 2014-06-19 Qualcomm Incorporated Low-delay buffering model in video coding
US9374585B2 (en) * 2012-12-19 2016-06-21 Qualcomm Incorporated Low-delay buffering model in video coding
US11677966B2 (en) 2013-01-04 2023-06-13 Ge Video Compression, Llc Efficient scalable coding concept
US10104386B2 (en) 2013-01-04 2018-10-16 Ge Video Compression, Llc Efficient scalable coding concept
CN110062240A (en) * 2013-01-04 2019-07-26 Ge视频压缩有限责任公司 Efficient scalable coding concept
US11025928B2 (en) 2013-01-04 2021-06-01 Ge Video Compression, Llc Efficient scalable coding concept
US10609396B2 (en) 2013-01-04 2020-03-31 Ge Video Compression, Llc Efficient scalable coding concept
US9485508B2 (en) 2013-04-08 2016-11-01 Qualcomm Incorporated Non-entropy encoded set of profile, tier, and level syntax structures
US9467700B2 (en) 2013-04-08 2016-10-11 Qualcomm Incorporated Non-entropy encoded representation format
US9473771B2 (en) 2013-04-08 2016-10-18 Qualcomm Incorporated Coding video data for an output layer set
US11582473B2 (en) 2013-04-08 2023-02-14 Ge Video Compression, Llc Coding concept allowing efficient multi-view/layer coding
US9565437B2 (en) 2013-04-08 2017-02-07 Qualcomm Incorporated Parameter set designs for video coding extensions
US20150373356A1 (en) * 2014-06-18 2015-12-24 Qualcomm Incorporated Signaling hrd parameters for bitstream partitions
US10063867B2 (en) 2014-06-18 2018-08-28 Qualcomm Incorporated Signaling HRD parameters for bitstream partitions
US9819948B2 (en) 2014-06-18 2017-11-14 Qualcomm Incorporated Signaling HRD parameters for bitstream partitions
US9813719B2 (en) * 2014-06-18 2017-11-07 Qualcomm Incorporated Signaling HRD parameters for bitstream partitions

Also Published As

Publication number Publication date
KR101682322B1 (en) 2016-12-05
CN105979270A (en) 2016-09-28
WO2009048502A3 (en) 2009-06-25
CN101889448B (en) 2016-08-03
KR101558627B1 (en) 2015-10-07
EP2198619A2 (en) 2010-06-23
TWI400958B (en) 2013-07-01
BRPI0817420A2 (en) 2013-06-18
KR101703019B1 (en) 2017-02-06
JP2010541470A (en) 2010-12-24
WO2009048502A2 (en) 2009-04-16
KR20100061715A (en) 2010-06-08
TWI517718B (en) 2016-01-11
BR122012021948A2 (en) 2015-08-11
JP5264920B2 (en) 2013-08-14
BR122012021949A2 (en) 2015-08-11
TWI520616B (en) 2016-02-01
KR20150086553A (en) 2015-07-28
TW200922332A (en) 2009-05-16
KR20100085078A (en) 2010-07-28
BR122012021799A2 (en) 2015-08-04
CN101889448A (en) 2010-11-17
JP2010541471A (en) 2010-12-24
CN105979270B (en) 2019-05-28
BR122012021797A2 (en) 2015-08-04
BRPI0817508A2 (en) 2013-06-18
BR122012021947A2 (en) 2015-08-04
TWI401966B (en) 2013-07-11
JP5264919B2 (en) 2013-08-14
TW201244496A (en) 2012-11-01
EP2198620A2 (en) 2010-06-23
BR122012021796A2 (en) 2015-08-04
CN105812826A (en) 2016-07-27
WO2009048503A2 (en) 2009-04-16
BR122012021950A2 (en) 2015-08-04
US20110038424A1 (en) 2011-02-17
BR122012021801A2 (en) 2015-08-04
TW201246935A (en) 2012-11-16
CN101971630A (en) 2011-02-09
TWI400957B (en) 2013-07-01
WO2009048503A3 (en) 2009-05-28
TW200926831A (en) 2009-06-16
TW201244495A (en) 2012-11-01
TWI530195B (en) 2016-04-11
TW201244483A (en) 2012-11-01

Similar Documents

Publication Publication Date Title
US20100208796A1 (en) Methods and apparatus for incorporating video usability information (vui) within a multi-view video (mvc) coding system
US10511845B2 (en) Methods and apparatus for video usability information (VUI) for scalable video coding (SVC)
US20190037241A1 (en) Methods and apparatus for multi-view video coding
US9241168B2 (en) Methods and apparatus for illumination and color compensation for multi-view video coding
US20090279612A1 (en) Methods and apparatus for multi-view video encoding and decoding
US20090147860A1 (en) Method and apparatus for signaling view scalability in multi-view video coding

Legal Events

Date Code Title Description
AS Assignment

Owner name: THOMSON LICENSING, FRANCE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LUO, JIANCONG;YIN, PENG;SIGNING DATES FROM 20081027 TO 20081103;REEL/FRAME:024204/0596

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION