Patent application number | Description | Published |
20120317303 | System and Method of Media Content Streaming with a Multiplexed Representation - In accordance with an embodiment, a method of operating a streaming media client includes electronically receiving from network information regarding component access points of a first multiplexed representation of a streaming content. The first multiplexed representation of the streaming content includes a plurality of media components, and each component access point defines an independently coded access point for one of the plurality of media components. The method further includes selecting a first segment of the first multiplexed representation of the streaming content based on the information regarding the component access points, and requesting the selected first segment of the first multiplexed representation of the streaming content from the network. | 12-13-2012 |
20130016769 | SIGNALING PICTURE SIZE IN VIDEO CODING - A video encoder is configured to determine a picture size for one or more pictures included in a video sequence. The picture size associated with the video sequence may be a multiple of an aligned coding unit size for the video sequence. In one example, the aligned coding unit size for the video sequence may comprise a minimum coding unit size where the minimum coding unit size is selected from a plurality of smallest coding unit sizes corresponding to different pictures in the video sequence. A video decoder is configured to obtain syntax elements to determine the picture size and the aligned coding unit size for the video sequence. The video decoder decodes the pictures included in the video sequence with the picture size, and stores the decoded pictures in a decoded picture buffer. | 01-17-2013 |
20130034170 | CODING PARAMETER SETS FOR VARIOUS DIMENSIONS IN VIDEO CODING - In one example, a device for coding video data includes a video coder configured to code, for a bitstream, information representative of which of a plurality of video coding dimensions are enabled for the bitstream, and code values for each of the enabled video coding dimensions, without coding values for the video coding dimensions that are not enabled, in a network abstraction layer (NAL) unit header of a NAL unit comprising video data coded according to the values for each of the enabled video coding dimensions. In this manner, NAL unit headers may have variable lengths, while still providing information for scalable dimensions to which the NAL units correspond. | 02-07-2013 |
20130057646 | SLICE HEADER THREE-DIMENSIONAL VIDEO EXTENSION FOR SLICE HEADER PREDICTION - In one example, a video coder is configured to code one or more blocks of video data representative of texture information of at least a portion of a frame of video data, process a texture slice for a texture view component of a current view associated, the texture slice comprising the coded one or more blocks and a texture slice header comprising a set of syntax elements representative of characteristics of the texture slice, code depth information representative of depth values for at least the portion of the frame, and process a depth slice for a depth view component corresponding to the texture view component of the view, the depth slice comprising the coded depth information and a depth slice header comprising a set of syntax elements representative of characteristics of the depth slice, wherein process the texture slice or the depth slice comprises predict at least one syntax element. | 03-07-2013 |
20130077677 | REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - Techniques are described related to constructing reference picture lists. The reference picture lists may be constructed from reference picture subsets of a reference picture set. In some examples, the reference picture subsets may be ordered in a particular manner to form the reference picture lists. | 03-28-2013 |
20130077678 | REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - Techniques are described related to constructing reference picture lists. The reference picture lists may be constructed from reference picture subsets of a reference picture set. In some examples, the techniques may repeatedly list reference pictures identified in the reference picture subsets until the number of entries in the reference picture list is equal to the maximum number of allowable entries in the reference picture list. | 03-28-2013 |
20130077679 | VIDEO CODING WITH SUBSETS OF A REFERENCE PICTURE SET - Techniques are described related to deriving a reference picture set. A reference picture set may identify reference pictures that can potentially be used to inter-predict a current picture and picture following the current picture in decoding order. In some examples, deriving the reference picture set may include constructing a plurality of reference picture subsets that together form the reference picture set. | 03-28-2013 |
20130077680 | DECODED PICTURE BUFFER MANAGEMENT - Techniques are described related to output and removal of decoded pictures from a decoded picture buffer (DPB). The example techniques may remove a decoded picture from the DPB prior to coding a current picture. For instance, the example techniques may remove the decoded picture if that decoded picture is not identified in the reference picture set of the current picture. | 03-28-2013 |
20130077681 | REFERENCE PICTURE SIGNALING AND DECODED PICTURE BUFFER MANAGEMENT - Techniques are described related to performing random access starting from a random access point picture that is not an instantaneous decoder refresh picture. Some techniques are also related to reducing the amount of information that is signaled for long-term reference pictures of a reference picture set. Additional techniques are also related to decoded picture buffer management, such as removing decoded pictures based on a temporal identification value. | 03-28-2013 |
20130077685 | REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - Techniques are described related to modifying an initial reference picture list. The example techniques may identify a reference picture in at least one of the reference picture subsets used to construct the initial reference picture. The example techniques may list the identified reference picture in a current entry of the initial reference picture list to construct a modified reference picture list. | 03-28-2013 |
20130077687 | CODING REFERENCE PICTURES FOR A REFERENCE PICTURE SET - Techniques are described related to coding of long-term reference pictures for a reference picture set. In some examples, a video coder may code candidate long-term reference pictures in a parameter set. The video coder also code syntax elements that indicate which long-term reference pictures from the candidate long-term reference pictures belong in the reference picture set. | 03-28-2013 |
20130089134 | EFFICIENT SIGNALING OF REFERENCE PICTURE SETS - A video coder can select which reference pictures should be signaled in a parameter set such as a picture parameter set (PPS) and which reference pictures should be signaled in a slice header such that when a video decoder constructs a reference picture set, the video decoder does not need to reorder the reference picture set to construct an initial reference picture list for a slice of video data. | 04-11-2013 |
20130089135 | ADAPTIVE FRAME SIZE SUPPORT IN ADVANCED VIDEO CODECS - Techniques are described related to receiving a first decoded frame of video data, wherein the first decoded frame is associated with a first resolution, determining whether a decoded picture buffer is available to store the first decoded frame based on the first resolution, and in the event the decoded picture buffer is available to store the first decoded frame, storing the first decoded frame in the decoded picture buffer, and determining whether the decoded picture buffer is available to store a second decoded frame of video data, wherein the second decoded frame is associated with a second resolution, based on the first resolution and the second resolution, wherein the first decoded frame is different than the second decoded frame. | 04-11-2013 |
20130089152 | SIGNALING PICTURE IDENTIFICATION FOR VIDEO CODING - In one example, a video coder, such as a video encoder or video decoder, is configured to determine a number of least significant bits of picture identifying information for a picture of video data, determine a value of the picture identifying information for the picture, and code information indicative of the determined number of least significant bits of the value of the picture identifying information for the picture. | 04-11-2013 |
20130089154 | ADAPTIVE FRAME SIZE SUPPORT IN ADVANCED VIDEO CODECS - Techniques are described related to receiving first and second sub-sequences of video, wherein the first sub-sequence includes one or more frames each having a first resolution, and the second sub-sequence includes one or more frames each having a second resolution, receiving a first sequence parameter set and a second sequence parameter set for the coded video sequence, wherein the first sequence parameter set indicates the first resolution of the one or more frames of the first sub-sequence, and the second sequence parameter set indicates the second resolution of the one or more frames of the second sub-sequence, and wherein the first sequence parameter set is different than the second sequence parameter set, and using the first sequence parameter set and the second sequence parameter set to decode the coded video sequence. | 04-11-2013 |
20130101035 | GROUPING OF TILES FOR VIDEO CODING - Techniques described herein for coding video data include techniques for coding pictures partitioned into tiles, in which each of the plurality of tiles in a picture is assigned to one of a plurality of tile groups. One example method for coding video data comprising a picture that is partitioned into a plurality tiles comprises coding video data in a bitstream, and coding, in the bitstream, information that indicates one of a plurality of tile groups to which each of the plurality of tiles is assigned. The techniques for grouping tiles described herein may facilitate improved parallel processing for both encoding and decoding of video bitstreams, improved error resilience, and more flexible region of interest (ROI) coding. | 04-25-2013 |
20130107942 | FRAGMENTED PARAMETER SET FOR VIDEO CODING | 05-02-2013 |
20130107952 | UNIFIED DESIGN FOR PICTURE PARTITIONING SCHEMES | 05-02-2013 |
20130107953 | RANDOM ACCESS WITH ADVANCED DECODED PICTURE BUFFER (DPB) MANAGEMENT IN VIDEO CODING | 05-02-2013 |
20130107973 | LOOP FILTERING CONTROL OVER TILE BOUNDARIES | 05-02-2013 |
20130114694 | PARAMETER SET GROUPS FOR CODED VIDEO DATA - A video coding device, such as a video encoder or a video decoder, may be configured to code a parameter set group representing a first parameter set of a first type and a second parameter set of a second, different type, and code a slice of video data using information of the parameter set group, information of the first parameter set, and information of the second parameter set, wherein the slice includes information referring to the parameter set group. The video coding device may further code the first and second parameter sets. | 05-09-2013 |
20130114735 | VIDEO CODING WITH NETWORK ABSTRACTION LAYER UNITS THAT INCLUDE MULTIPLE ENCODED PICTURE PARTITIONS - A video encoder generates a Network Abstraction Layer (NAL) unit that contains a plurality of encoded picture partitions of the video data. The video encoder generates a bitstream that includes a variable-length value that represents an entropy-encoded first syntax element, a variable-length value that represents an entropy-encoded second syntax element, and fixed-length values that represent offset syntax elements. Lengths of each of the offset syntax elements are determinable based on the first syntax element. A video decoder uses the first syntax element, the second syntax element, and the offset syntax elements when decoding the encoded picture partitions. | 05-09-2013 |
20130114736 | PADDING OF SEGMENTS IN CODED SLICE NAL UNITS - A video encoder divides a picture into a plurality of picture partitions, such as tiles or wavefront parallel processing (WPP) waves. The picture partitions are associated with non-overlapping subsets of the treeblocks of the picture. The video encoder generates a coded slice network abstraction layer (NAL) unit that includes encoded representations of the treeblocks associated with a slice of the picture. The coded treeblocks are grouped within the coded slice NAL unit into segments associated with different ones of the picture partitions. The video encoder pads one or more of the segments such that each of the segments begins on a byte boundary. | 05-09-2013 |
20130135431 | SEQUENCE LEVEL INFORMATION FOR MULTIVIEW VIDEO CODING (MVC) COMPATIBLE THREE-DIMENSIONAL VIDEO CODING (3DVC) - In general, techniques are described for separately coding depth and texture components of video data. A video coding device for processing the video data comprising one or more processors may perform the techniques. The one or more processors may be configured to determine first sequence level information describing characteristics of the depth components, and determine second sequence level information describing characteristics of an operation point of the video data. | 05-30-2013 |
20130135433 | DEPTH COMPONENT REMOVAL FOR MULTIVIEW VIDEO CODING (MVC) COMPATIBLE THREE-DIMENSIONAL VIDEO CODING (3DVC) - In general, techniques are described for separately coding depth and texture components of video data. A video coding device configured to code video data may perform the techniques. The video coding device may comprise a decoded picture buffer and a processor configured to store a depth component in the decoded picture buffer, analyze a view dependency to determine whether the depth component is used for inter-view prediction and remove the depth component from the decoded picture buffer in response to determining that the depth component is not used for inter-view prediction. for processing video data including a view component comprised of a depth component and a texture component | 05-30-2013 |
20130135434 | NESTED SEI MESSAGES FOR MULTIVIEW VIDEO CODING (MVC) COMPATIBLE THREE-DIMENSIONAL VIDEO CODING (3DVC) - In general, techniques are described for separately processing depth and texture components of video data. A device configured to process video data including a view component comprised of a depth component and a texture component may perform various aspects of the techniques. The device may comprise a processor configured to determine a supplemental enhancement information message that applies when processing the view component of the video data, and determine a nested supplemental enhancement information message that applies in addition to the supplemental enhancement information message when processing the depth component of the view component. | 05-30-2013 |
20130136176 | ACTIVATION OF PARAMETER SETS FOR MULTIVIEW VIDEO CODING (MVC) COMPATIBLE THREE-DIMENSIONAL VIDEO CODING (3DVC) - In general, techniques are described for separately coding depth and texture components of video data. A video coding device for coding video data that includes a view component comprised of a depth component and a texture component may perform the techniques. The video coding device may comprise, as one example, a processor configured to activate a parameter set as a texture parameter set for the texture component of the view component, and code the texture component of the view component based on the activated texture parameter set. | 05-30-2013 |
20130142256 | CODING LEAST SIGNFICANT BITS OF PICTURE ORDER COUNT VALUES IDENTIFYING LONG-TERM REFERENCE PICTURES - In general, techniques are described for coding picture order count values identifying long-term reference pictures. A video decoding device comprising a processor may perform the techniques. The processor may be configured to determine a number of bits used to represent least significant bits of the picture order count value that identifies a long-term reference picture to be used when decoding at least a portion of a current picture and parse the determined number of bits from a bitstream representative of the encoded video data. The parsed bits represent the least significant bits of the picture order count value. The processor retrieves the long-term reference picture from a decoded picture buffer based on the least significant bits, and decodes at least the portion of the current picture using the retrieved long-term reference picture. | 06-06-2013 |
20130142257 | CODING PICTURE ORDER COUNT VALUES IDENTIFYING LONG-TERM REFERENCE FRAMES - In general, techniques are described for coding picture order count values identifying long-term reference pictures. A video decoding device comprising a processor may perform the techniques. The processor may determine least significant bits (LSBs) of a picture order count (POC) value that identifies a long-term reference picture (LTRP). The LSBs do not uniquely identify the POC value with respect to the LSBs of any other POC value identifying any other picture in a decoded picture buffer (DPB). The processor may determine most significant bits (MSBs) of the POC value. The MSBs combined with the LSBs is sufficient to distinguish the POC value from any other POC value that identifies any other picture in the DPB. The processor may retrieve the LTRP from the decoded picture buffer based on the LSBs and MSBs of the POC value, and decode a current picture of the video data using the retrieved LTRP. | 06-06-2013 |
20130155184 | REFERENCE PICTURE LIST CONSTRUCTION FOR MULTI-VIEW AND THREE-DIMENSIONAL VIDEO CODING - A video encoder generates, based on a reference picture set of a current view component, a reference picture list for the current view component. The reference picture set includes an inter-view reference picture set. The video encoder encodes the current view component based at least in part on one or more reference pictures in the reference picture list. In addition, the video encoder generates a bitstream that includes syntax elements indicating the reference picture set of the current view component. A video decoder parses, from the bitstream, syntax elements indicating the reference picture set of the current view component. The video decoder generates, based on the reference picture set, the reference picture list for the current view component. In addition, the video decoder decodes at least a portion of the current view component based on one or more reference pictures in the reference picture list. | 06-20-2013 |
20130163668 | PERFORMING MOTION VECTOR PREDICTION FOR VIDEO CODING - In general, techniques are described for performing motion vector prediction for video coding. A video coding device comprising a processor may perform the techniques. The processor may be configured to determine a plurality of candidate motion vectors for a current block of the video data so as to perform the motion vector prediction process and scale one or more of the plurality of candidate motion vectors determined for the current block of the video data to generate one or more scaled candidate motion vectors. The processor may then be configured to modify the scaled candidate motion vectors to be within a specified range. | 06-27-2013 |
20130176389 | SIGNALING VIEW SYNTHESIS PREDICTION SUPPORT IN 3D VIDEO CODING - In one example, a video coder is configured to code information indicative of whether view synthesis prediction is enabled for video data. When the information indicates that view synthesis prediction is enabled for the video data, the video coder may generate a view synthesis picture using the video data and code at least a portion of a current picture relative to the view synthesis picture. The at least portion of the current picture may comprise, for example, a block (e.g., a PU, a CU, a macroblock, or a partition of a macroblock), a slice, a tile, a wavefront, or the entirety of the current picture. On the other hand, when the information indicates that view synthesis prediction is not enabled for the video data, the video coder may code the current picture using at least one of intra-prediction, temporal inter-prediction, and inter-view prediction without reference to any view synthesis pictures. | 07-11-2013 |
20130177084 | MOTION VECTOR SCALING IN VIDEO CODING - This disclosure proposes techniques for motion vector scaling. In particular, this disclosure proposes that both an implicit motion vector scaling process (e.g., the POC-based motion vector scaling process described above), as well as an explicit motion vector (e.g., a motion vector scaling process using scaling weights) may be used to perform motion vector scaling. This disclosure also discloses example signaling methods for indicating the type of motion vector scaling used. | 07-11-2013 |
20130182755 | CODING PARAMETER SETS AND NAL UNIT HEADERS FOR VIDEO CODING - In one example, a video coder, such as a video encoder or video decoder, is configured to code a video parameter set (VPS) for one or more layers of video data, wherein each of the one or more layers of video data refer to the VPS, and code the one or more layers of video data based at least in part on the VPS. The video coder may code the VPS for video data conforming to High-Efficiency Video Coding, Multiview Video Coding, Scalable Video Coding, or other video coding standards or extensions of video coding standards. The VPS may include data specifying parameters for corresponding sequences of video data within various different layers (e.g., views, quality layers, or the like). The parameters of the VPS may provide indications of how the corresponding video data is coded. | 07-18-2013 |
20130182774 | INDICATION OF USE OF WAVEFRONT PARALLEL PROCESSING IN VIDEO CODING - A video encoder generates a bitstream that includes a syntax element that indicates whether a picture is encoded according either a first coding mode or a second coding mode. In the first coding mode, the picture is entirely encoded using wavefront parallel processing (WPP). In the second coding mode, each tile of the picture is encoded without using WPP and the picture may have one or more tiles. A video decoder may parse the syntax element from the bitstream. In response to determining that the syntax element has a particular value, the video decoder decodes the picture entirely using WPP. In response to determining that the syntax element does not have the particular value, the video decoder decodes each tile of the picture without using WPP. | 07-18-2013 |
20130182775 | SUB-STREAMS FOR WAVEFRONT PARALLEL PROCESSING IN VIDEO CODING - A video encoder signals whether WPP is used to encode a picture of a sequence of video picture. If WPP is used to encode the picture, the video encoder generates a coded slice NAL unit that includes a plurality of sub-streams, each of which includes a consecutive series of bits that represents one encoded row of coding tree blocks (CTBs) in a slice of the picture. A video decoder receives a bitstream that includes the coded slice NAL unit. Furthermore, the video decoder may determine, based on a syntax element in the bitstream, that the slice is encoded using WPP and may decode the slice using WPP. | 07-18-2013 |
20130188733 | SIGNALING OF DEBLOCKING FILTER PARAMETERS IN VIDEO CODING - This disclosure describes techniques for signaling deblocking filter parameters for a current slice of video data with reduced bitstream overhead. Deblocking filter parameters may be coded in one or more of a picture layer parameter set and a slice header. The techniques reduce a number of bits used to signal the deblocking filter parameters by coding a first syntax element that indicates whether deblocking filter parameters are present in both the picture layer parameter set and the slice header, and only coding a second syntax element in the slice header when both sets of deblocking filter parameters are present. Coding the second syntax element is eliminated when deblocking filter parameters are present in only one of the picture layer parameter set or the slice header. The second syntax element indicates which set of deblocking filter parameters to use to define a deblocking filter applied to a current slice. | 07-25-2013 |
20130195171 | METHOD OF CODING VIDEO AND STORING VIDEO CONTENT - A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include a dedicated array for parameter sets. | 08-01-2013 |
20130195172 | METHOD OF CODING VIDEO AND STORING VIDEO CONTENT - A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include an indicator identifying a number of temporal layers of the video stream. | 08-01-2013 |
20130195173 | METHOD OF CODING VIDEO AND STORING VIDEO CONTENT - A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include an indicator identifying a number of parameter sets stored within one or more access units of the video stream. | 08-01-2013 |
20130195205 | METHOD OF CODING VIDEO AND STORING VIDEO CONTENT - A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include stream properties associated with the video stream. | 08-01-2013 |
20130202035 | REFERENCE PICTURE LIST MODIFICATION FOR VIDEO CODING - A video coder may, in some cases, signal whether one or more initial reference picture lists are to be modified. When an initial list is to be modified, the video coder can signal information indicating a starting position in the initial reference picture list. When the starting position signaled by the video coder is less than a number of pictures included in the initial reference picture list, then the video coder signals the number of pictures to be inserted into the initial reference picture list, and a reference picture source from which a picture can be retrieved to insert into the initial reference picture list to construct a modified reference picture list. | 08-08-2013 |
20130222537 | BITSTREAM EXTRACTION IN THREE-DIMENSIONAL VIDEO - To extract a sub-bitstream from a 3-dimensional video (3DV) bitstream, a device determines a texture target view list that indicates views in the 3DV bitstream that have texture view components that are required for decoding pictures in a plurality of target views. The target views are a subset of the views in the bitstream that are to be decodable from the sub-bitstream. In addition, the device determines a depth target view list that indicates views in the 3DV bitstream that have depth view components that are required for decoding pictures in the plurality of target views. The device determines the sub-bitstream based at least in part on the texture target view list and the depth target view list. | 08-29-2013 |
20130222538 | NETWORK ABSTRACTION LAYER (NAL) UNIT HEADER DESIGN FOR THREE-DIMENSIONAL VIDEO CODING - A video encoder generates a network abstraction layer (NAL) unit that includes at least a first syntax element and a second syntax element. The first syntax element indicates that the NAL unit belongs to a particular NAL unit type. Coded slices of texture view components and depth view components are encapsulated within NAL units that belong to the particular NAL unit type. The second syntax element indicates whether a NAL unit header of the NAL unit includes an Advanced Video Coding (AVC)-compatible 3-dimensional video (3DV) header extension or includes a Multiview Video Coding (MVC)-compatible 3DV header extension. The video encoder outputs a bitstream that includes the NAL unit. A video decoder receives the NAL unit and determines whether the second syntax element indicates that the NAL unit header of the NAL unit includes the AVC-compatible 3DV header extension or the MVC-compatible 3DV header extension. | 08-29-2013 |
20130243081 | HIGH-LEVEL SYNTAX EXTENSIONS FOR HIGH EFFICIENCY VIDEO CODING - In one example, a device includes a video coder configured to code a picture order count (POC) value for a first picture of video data, code a second-dimension picture identifier for the first picture, and code, in accordance with a base video coding specification or an extension to the base video coding specification, a second picture based at least in part on the POC value and the second-dimension picture identifier of the first picture. The video coder may comprise a video encoder or a video decoder. The second-dimension picture identifier may comprise, for example, a view identifier, a view order index, a layer identifier, or other such identifier. The video coder may code the POC value and the second-dimension picture identifier during coding of a motion vector for a block of the second picture, e.g., during advanced motion vector prediction or merge mode coding. | 09-19-2013 |
20130243093 | MOTION VECTOR CODING AND BI-PREDICTION IN HEVC AND ITS EXTENSIONS - In one example, a device includes a video coder (e.g., a video encoder or a video decoder) configured to determine that a block of video data is to be coded in accordance with a three-dimensional extension of High Efficiency Video Coding (HEVC), and, based the determination that the block is to be coded in accordance with the three-dimensional extension of HEVC, disable temporal motion vector prediction for coding the block. The video coder may be further configured to, when the block comprises a bi-predicted block (B-block), determine that the B-block refers to a predetermined pair of pictures in a first reference picture list and a second reference picture list, and, based on the determination that the B-block refers to the predetermined pair, equally weight contributions from the pair of pictures when calculating a predictive block for the block. | 09-19-2013 |
20130266075 | LOW-DELAY VIDEO BUFFERING IN VIDEO CODING - As one example, a method of coding video data includes storing one or more decoding units of video data in a picture buffer. The method further includes obtaining a respective buffer removal time for the one or more decoding units, wherein obtaining the respective buffer removal time comprises receiving a respective signaled value indicative of the respective buffer removal time for at least one of the decoding units. The method further includes removing the decoding units from the picture buffer in accordance with the obtained buffer removal time for each of the decoding units. The method further includes coding video data corresponding to the removed decoding units, wherein coding the video data comprises decoding the at least one of the decoding units. | 10-10-2013 |
20130266076 | LOW-DELAY VIDEO BUFFERING IN VIDEO CODING - As one example, a method of coding video data includes storing one or more decoding units of video data in a coded picture buffer (CPB). The method further includes obtaining a respective buffer removal time for the one or more decoding units. The method further includes removing the decoding units from the CPB in accordance with the obtained buffer removal time for each of the decoding units. The method further includes determining whether the CPB operates at access unit level or sub-picture level. The method further includes coding video data corresponding to the removed decoding units. If the CPB operates at access unit level, coding the video data comprises coding access units comprised in the decoding units. If the CPB operates at sub-picture level, coding the video data comprises coding subsets of access units comprised in the decoding units. | 10-10-2013 |
20130272370 | WAVEFRONT PARALLEL PROCESSING FOR VIDEO CODING - In one example, a video coder may be configured to determine that a slice of a picture of video data begins in a row of coding tree units (CTUs) in the picture at a position other than a beginning of the row. Based on the determination, the video coder may be further configured to determine that the slice ends within the row of CTUs. The video coder may be further configured to code the slice based on the determination that the slice ends within the row of CTUs. | 10-17-2013 |
20130272403 | REFERENCE PICTURE SET PREDICTION FOR VIDEO CODING - In one example, a device for decoding video data includes a video decoder configured to decode one or more syntax elements of a current reference picture set (RPS) prediction data structure, wherein at least one of the syntax elements represents a picture order count (POC) difference between a POC value associated with the current RPS and a POC value associated with a previously decoded RPS, form a current RPS based at least in part on the RPS prediction data structure and the previously decoded RPS, and decode one or more pictures using the current RPS. A video encoder may be configured to perform a substantially similar process during video encoding. | 10-17-2013 |
20130279564 | VIDEO CODING WITH ENHANCED SUPPORT FOR STREAM ADAPTATION AND SPLICING - Various techniques for enhanced support of stream adaptation and splicing based on clean random access (CRA) pictures are described. Instead of using a flag in the slice header to indicate that a broken link picture is present, a distinct network abstraction layer (NAL) unit type can be used to indicate the presence of a broken link picture. In some implementations, a first distinct NAL unit type may be used to indicate the presence of a broken link picture with leading pictures, while a second distinct NAL unit type indicates the presence of a broken link picture without leading pictures. In some implementations, a third distinct NAL unit type may be used to indicate the presence of a broken link picture with decodable leading pictures. | 10-24-2013 |
20130279575 | MARKING REFERENCE PICTURES IN VIDEO SEQUENCES HAVING BROKEN LINK PICTURES - Systems, methods, and devices for processing video data are disclosed. Some examples determine that a current picture is a broken-link access (BLA) picture. These examples may also mark a reference picture in a picture storage buffer as unused for reference. In some examples, this may be done prior to decoding the BLA picture. | 10-24-2013 |
20130279576 | VIEW DEPENDENCY IN MULTI-VIEW CODING AND 3D CODING - This disclosure described techniques for coding layer dependencies for a block of video data. According to these techniques, a video encoder generates layer dependencies associated with a given layer. The video encoder also generates a type of prediction associated with one or more of the layer dependencies. In some examples, the video encoder generates a first syntax element to signal layer dependencies and a second syntax element to signal a type of prediction associated with one or more of the layer dependencies. A video decoder may obtain the layer dependencies associated with a given layer and the type of prediction associated with one or more of the layer dependencies. | 10-24-2013 |
20130279599 | DECODED PICTURE BUFFER PROCESSING FOR RANDOM ACCESS POINT PICTURES IN VIDEO SEQUENCES - Systems, methods, and devices for processing video data are disclosed. Some examples receive a slice of a current picture to be decoded for a sequence of video data. These examples may also receive, in a slice header of the slice, at least one entropy coded syntax element and at least one non-entropy coded syntax element, wherein the non-entropy coded syntax element is before the entropy coded syntax element in the slice header and indicates whether pictures prior to the current picture in decoding order are to be emptied from a decode picture buffer without being output. They may decode the slice based on the non-entropy coded syntax element. | 10-24-2013 |
20130287115 | PARAMETER SET CODING - Systems, methods, and devices for processing video data are disclosed. Some examples relate to receiving or forming a parameter set having an identifier that is fixed length coded, wherein a parameter set identification (ID) for the parameter set is before any syntax element in the parameter set that is entropy coded and using the parameter set having the identifier that is fixed length coded to decode or encode video data. Other examples determine whether a first parameter set ID of a first parameter set of a first bitstream is the same as a second parameter set ID of a second parameter set of a second bitstream. In response to determining that the second parameter set ID is the same as the first parameter set ID, changing the second parameter set ID to a unique parameter set ID. A parameter set associated with the unique parameter set ID may be transmitted. | 10-31-2013 |
20130287366 | IDENTIFYING PARAMETER SETS IN VIDEO FILES - An apparatus is configured to store coded video data including a number of sequences of coded video pictures in an electronic file. The apparatus includes at least one processor configured to determine whether a sample description associated with at least one sample includes all parameter sets of a particular type associated with the at least one sample. The at least one sample comprises at least a portion of the plurality of sequences of coded video pictures. The particular type is one of a plurality of different particular types of parameter sets. The at least one processor is also configured to provide, in the electronic file, an indication indicating whether the sample description includes all parameter sets of the particular type based on the determination. | 10-31-2013 |
20130294499 | PARAMETER SET UPDATES IN VIDEO CODING - Techniques of this disclosure provide an indication of whether a parameter set update can occur in a portion of a bitstream. The indication may enable a video decoder to determine whether an update of a stored parameter set can occur without performing a content comparison between the stored parameter set and a new parameter set of the same type with the same identification value. A parameter set update includes storing a current parameter set with a given identification value to replace a previous parameter set of the same type and having the same identification value. When a parameter set update cannot occur, the video decoder may store and activate a single parameter set of a given type for the entire portion of the bitstream. When a parameter set update can occur, the video decoder may automatically update a stored parameter set, or may determine whether to update the stored parameter. | 11-07-2013 |
20130294500 | FULL RANDOM ACCESS FROM CLEAN RANDOM ACCESS PICTURES IN VIDEO CODING - Techniques of this disclosure provide an indication of whether performing random access from a particular access unit in a bitstream requires fetching of parameter sets from previous access units. A clean random access (CRA) picture can be positioned at any point within a coded video sequence and does not clean a decoded picture buffer (DPB) of a video decoder. In order to perform random access decoding from the CRA picture, a video decoder may need to fetch one or more parameter sets included in unavailable access units that precede the CRA picture. The techniques provide an indication, for each CRA picture, that indicates whether parameter sets included in previous access units are needed to perform random access from the picture. When no parameter sets from previous access units are needed for random access from a particular CRA picture, a video decoder may determine to perform random access from that picture. | 11-07-2013 |
20130322531 | EXTERNAL PICTURES IN VIDEO CODING - A video encoder generates a syntax element that indicates whether a video unit of a current picture is predicted from an external picture. The external picture is in a different layer than the current picture. Furthermore, the video encoder outputs a video data bitstream that includes a representation of the syntax element. The video data bitstream may or may not include a coded representation of the external picture. A video decoder obtains the syntax element from the video data bitstream. The video decoder uses the syntax element in a process to reconstruct video data of a portion of the video unit. | 12-05-2013 |
20130329787 | SIGNALING DATA FOR LONG TERM REFERENCE PICTURES FOR VIDEO CODING - A video coder codes a slice header for a slice of video data. The slice header includes a syntax element comprising identifying information for a long term reference picture, wherein the identifying information is explicitly signaled in the slice header or derived from a sequence parameter set corresponding to the slice. When the syntax element indicates that the identifying information for the long term reference picture is explicitly signaled, to code the slice header, the video coder is further configured to code a value for the identifying information for the long term reference picture in the slice header. | 12-12-2013 |
20130342762 | DEVICE AND METHOD FOR MULTIMEDIA COMMUNICATIONS WITH PICTURE ORIENTATION INFORMATION - Systems, devices, and methods for capturing and displaying picture data including picture orientation information are described. In one innovative aspect, a method for transmitting media information is provided. The method includes obtaining picture or video information, said picture or video information including image data and orientation information of a media capture unit when the picture or video information is obtained. The method further includes encoding said picture or video information, wherein the orientation information is included in a first portion and the image data is included in a second portion, the second portion being encoded and the first portion being distinct from the second portion. The method also includes transmitting the first portion and the second portion. | 12-26-2013 |
20130343465 | HEADER PARAMETER SETS FOR VIDEO CODING - An example method of decoding video data includes determining a header parameter set that includes one or more syntax elements specified individually by each of one or more slice headers, the header parameter set being associated with a header parameter set identifier (HPS ID), and determining one or more slice headers that reference the header parameter set to inherit at least one of the syntax elements included in the header parameter set, where the slice headers are each associated with a slice of the encoded video data, and where the slice headers each reference the header parameter set using the HPS ID. | 12-26-2013 |
20140003491 | VIDEO PARAMETER SET FOR HEVC AND EXTENSIONS | 01-02-2014 |
20140003492 | VIDEO PARAMETER SET FOR HEVC AND EXTENSIONS | 01-02-2014 |
20140003493 | VIDEO PARAMETER SET FOR HEVC AND EXTENSIONS | 01-02-2014 |
20140003506 | SIGNALING OF LONG-TERM REFERENCE PICTURES FOR VIDEO CODING | 01-02-2014 |
20140003531 | TILES AND WAVEFRONT PARALLEL PROCESSING | 01-02-2014 |
20140003536 | STREAMING ADAPTION BASED ON CLEAN RANDOM ACCESS (CRA) PICTURES | 01-02-2014 |
20140003537 | RANDOM ACCESS AND SIGNALING OF LONG-TERM REFERENCE PICTURES IN VIDEO CODING | 01-02-2014 |
20140003538 | SIGNALING LONG-TERM REFERENCE PICTURES FOR VIDEO CODING | 01-02-2014 |
20140010277 | SUPPLEMENTAL ENHANCEMENT INFORMATION (SEI) MESSAGES HAVING A FIXED-LENGTH CODED VIDEO PARAMETER SET (VPS) ID - Systems, methods, and devices are disclosed that code a supplemental enhancement information (SEI) message. In some examples, the SEI message may contain an identifier of an active video parameter set (VPS). In some examples, the identifier may be fixed-length coded. | 01-09-2014 |
20140016697 | CODING RANDOM ACCESS PICTURES FOR VIDEO CODING - In one example, a device for decoding video data includes a processor configured to decapsulate a slice of a random access point (RAP) picture of a bitstream from a network abstraction layer (NAL) unit, wherein the NAL unit includes a NAL unit type value that indicates whether the RAP picture is of a type that can have associated leading pictures and whether the RAP picture is an instantaneous decoder refresh (IDR) picture or a clean random access (CRA) picture, determine whether the RAP picture can have associated leading pictures based on the NAL unit type value, and decode video data of the bitstream following the RAP picture based on the determination of whether the RAP picture can have associated leading pictures. | 01-16-2014 |
20140016707 | CODING SEI NAL UNITS FOR VIDEO CODING - In one example, a device for decoding video data includes a processor configured to determine, for a supplemental enhancement information (SEI) network abstraction layer (NAL) unit of a bitstream, whether a NAL unit type value for the SEI NAL unit indicates that the NAL unit comprises a prefix SEI NAL unit including a prefix SEI message or a suffix SEI NAL unit including a suffix SEI message, and decode video data of the bitstream following the SEI NAL unit based on whether the SEI NAL unit is the prefix SEI NAL unit or the suffix SEI NAL unit and data of the SEI NAL unit. | 01-16-2014 |
20140016708 | CODING TIMING INFORMATION FOR VIDEO CODING - In one example, a device for presenting video data includes a processor configured to determine an integer value for the video data, determine a difference value between a presentation time of a first picture and a presentation time of a second picture, wherein the difference value is equal to the integer value multiplied by a clock tick value, and present the first picture and the second picture according to the determined difference value. | 01-16-2014 |
20140049603 | COMPATIBLE THREE-DIMENSIONAL VIDEO COMMUNICATIONS - Information for a video stream indicating whether the video stream includes stereoscopic three-dimensional video data can be provided to a display device. This information allows the device to determine whether to accept the video data and to properly decode and display the video data. This information can be made available for video data regardless of the codec used to encode the video. Systems, devices, and methods for transmission and reception of compatible video communications including stereoscopic three-dimensional picture information are described. | 02-20-2014 |
20140064384 | NETWORK ABSTRACTION LAYER HEADER DESIGN - A video processing device can receive in an encoded bitstream of video data a network abstraction layer (NAL) unit and parse a first syntax element in a header of the NAL unit to determine a temporal identification (ID) for the NAL unit, wherein a value of the first syntax element is one greater than the temporal identification. | 03-06-2014 |
20140078249 | INDICATION OF FRAME-PACKED STEREOSCOPIC 3D VIDEO DATA FOR VIDEO CODING - This disclosure describes techniques for signaling and using an indication that video data is in a frame-packed stereoscopic 3D video data format. In one example of the disclosure, a method for decoding video data comprises receiving video data, receiving an indication that indicates whether any pictures in the received video data contain frame-packed stereoscopic 3D video data, and decoding the received video data in accordance with the received indication. The received video data may be rejected if the video decoder is unable to decode frame-packed stereoscopic 3D video data. | 03-20-2014 |
20140079116 | INDICATION OF INTERLACED VIDEO DATA FOR VIDEO CODING - This disclosure proposes techniques for encoding and video data. The techniques of the disclosure receiving a first indication that indicates whether all pictures in received video data are progressive frames coded as frame pictures. If a video decoder is unable to decode progressive frames, the video data may be rejected based on the first indication. | 03-20-2014 |
20140079140 | VIDEO CODING WITH IMPROVED RANDOM ACCESS POINT PICTURE BEHAVIORS - This disclosure describes techniques for selection of coded picture buffer (CPB) parameters used to define a CPB for a video coding device for clean random access (CRA) pictures and broken link access (BLA) pictures in a video bitstream. A video coding device receives a bitstream including one or more CRA pictures or BLA pictures, and also receives a message indicating whether to use an alternative set of CPB parameters for at least one of the CRA pictures or BLA pictures. The message may be received from an external means, such as a processing means included in a streaming server or network entity. The video coding device sets a variable defined to indicate the set of CPB parameters for a given one of the pictures based on the received message, and selects the set of CPB parameters for the given one of the pictures based on the variable for the picture. | 03-20-2014 |
20140086303 | BITSTREAM CONFORMANCE TEST IN VIDEO CODING - A device performs a decoding process as part of a bitstream conformance test. As part of the decoding process, the device performs a bitstream extraction process to extract, from a bitstream, an operation point representation of an operation point defined by a target set of layer identifiers and a target highest temporal identifier. The target set of layer identifiers contains values of layer identifier syntax elements present in the operation point representation, the target set of layer identifiers being a subset of values of layer identifier syntax elements of the bitstream. The target highest temporal identifier is equal to a greatest temporal identifier present in the operation point representation, the target highest temporal identifier being less than or equal to a greatest temporal identifier present in the bitstream. The device decodes network abstraction layer (NAL) units of the operation point representation. | 03-27-2014 |
20140086317 | INDICATION AND ACTIVATION OF PARAMETER SETS FOR VIDEO CODING - In some examples, a video encoder includes multiple sequence parameter set (SPS) IDs in an SEI message, such that multiple active SPSs can be indicated to a video decoder. In some examples, a video decoder activates a video parameter set (VPS) and/or one or more SPSs through referencing an SEI message, e.g., based on the inclusion of the VPS ID and one or more SPS IDs in the SEI message. The SEI message may be, as examples, an active parameter sets SEI message or a buffering period SEI message. | 03-27-2014 |
20140086324 | LONG-TERM REFERENCE PICTURE SIGNALING IN VIDEO CODING - A video encoder signals, in a slice header for a current slice of a current picture, a first long-term reference picture (LTRP) entry, the first LTRP entry indicating that a particular reference picture is in a long-term reference picture set of the current picture. Furthermore, the video encoder signals, in the slice header, a second LTRP entry only if second LTRP entry does not indicate that the particular reference picture is in the long-term reference picture set of the current picture. | 03-27-2014 |
20140086331 | HYPOTHETICAL REFERENCE DECODER PARAMETERS IN VIDEO CODING - A device performs a hypothetical reference decoder (HRD) operation that determines conformance of a bitstream to a video coding standard or determines conformance of a video decoder to the video coding standard. As part of performing the HRD operation, the device determines a highest temporal identifier of a bitstream-subset associated with a selected operation point of the bitstream. Furthermore, as part of the HRD operation, the device determines, based on the highest temporal identifier, a particular syntax element from among an array of syntax elements. The device then uses the particular syntax element in the HRD operation. | 03-27-2014 |
20140086332 | ACCESS UNIT INDEPENDENT CODED PICTURE BUFFER REMOVAL TIMES IN VIDEO CODING - A video coding device, such as a video encoder or a video decoder, may be configured to code a duration between coded picture buffer (CPB) removal time of a first decoding unit (DU) in an access unit (AU) and a second DU, wherein the second DU is subsequent to the first DU in decoding order and in the same AU as the first DU. The video coding device may further determine a removal time of the DU based at least on the coded duration. The coding device may also code a sub-picture timing supplemental enhancement information (SEI) message associated with the first DU. The video coding device may further determine a removal time of the DU based at least in part on the sub-picture timing SEI message. | 03-27-2014 |
20140086333 | BITSTREAM PROPERTIES IN VIDEO CODING - A device signals a property of a bitstream. The bitstream comprises a plurality of coded video sequences (CVSs). When the property has a particular value, all the CVSs of the bitstream conform to the same profile. A video processing device is able to determine, based on the property, whether to process the bitstream. | 03-27-2014 |
20140086336 | HYPOTHETICAL REFERENCE DECODER PARAMETERS IN VIDEO CODING - A computing device selects, from among a set of hypothetical reference decoder (HRD) parameters in a video parameter set and a set of HRD parameters in a sequence parameter set, a set of HRD parameters applicable to a particular operation point of a bitstream. The computing device performs, based at least in part on the set of HRD parameters applicable to the particular operation point, an HRD operation on a bitstream subset associated with the particular operation point. | 03-27-2014 |
20140086337 | INDICATION AND ACTIVATION OF PARAMETER SETS FOR VIDEO CODING - In some examples, a video encoder includes multiple sequence parameter set (SPS) IDs in an SEI message, such that multiple active SPSs can be indicated to a video decoder. In some examples, a video decoder activates a video parameter set (VPS) and/or one or more SPSs through referencing an SEI message, e.g., based on the inclusion of the VPS ID and one or more SPS IDs in the SEI message. The SEI message may be, as examples, an active parameter sets SEI message or a buffering period SEI message. | 03-27-2014 |
20140086340 | EXPANDED DECODING UNIT DEFINITION - A video coding device, such as a video encoder or a video decoder, may be configured to decode a duration between coded picture buffer (CPB) removal time of a first decoding unit (DU) in an access unit (AU) and CPB removal time of a second DU, wherein the first DU comprises a non-video coding layer (VCL) network abstraction layer (NAL) unit with nal_unit_type equal to UNSPEC0, EOS_NUT, EOB_NUT, in the range of RSV_NVCL44 to RSV_NVCL47 or in the range of UNSPEC48 to UNSPEC63. The video decoder determines a removal time of the first DU based at least in part on the decoded duration and decodes video data of the first DU based at least in part on the removal time. | 03-27-2014 |
20140086341 | CODED PICTURE BUFFER REMOVAL TIMES SIGNALED IN PICTURE AND SUB-PICTURE TIMING SUPPLEMENTAL ENHANCEMENT INFORMATION MESSAGES - A video coding device, such as a video encoder or a video decoder, may be configured to code a sub-picture timing supplemental enhancement information (SEI) message associated with a first decoding unit (DU) of an access unit (AU). The video coding device may further code a duration between coded picture buffer (CPB) removal time of a second DU of the AU in decoding order and CPB removal time of the first DU in the sub-picture SEI message. The coding device may also derive a CPB removal time of the first DU based at least in part on the sub-picture timing SEI message. | 03-27-2014 |
20140086342 | SEQUENCE LEVEL FLAG FOR SUB-PICTURE LEVEL CODED PICTURE BUFFER PARAMETERS - A video coding device, such as a video encoder or a video decoder, may be configured to decode a sequence level flag to determine the presence of one or more sub-picture level coded picture buffer (CPB) parameters for a decoding unit (DU) of an access unit (AU) in either in a picture timing supplemental enhancement information (SEI) message or a sub-picture timing SEI message associated with the DU. The coding device may also decode the one or more sub-picture level CPB parameters from the picture timing SEI message or the sub-picture timing SEI message based on the sequence level flag. | 03-27-2014 |
20140086343 | BUFFERING PERIOD AND RECOVERY POINT SUPPLEMENTAL ENHANCEMENT INFORMATION MESSAGES - A video coding device, such as a video decoder, may be configured to decode a buffering period supplemental enhancement information (SEI) message associated with an access unit (AU). The video decoder is further configured to decode a duration between coded picture buffer (CPB) removal time of a first decoding unit (DU) in the AU and CPB removal time of a second DU from the buffering period SEI message, wherein the AU has a TemporalId equal to 0. The video decoder is configured to determine a removal time of the first DU based at least in part on the decoded duration and decode video data of the first DU based at least in part on the removal time. | 03-27-2014 |
20140086344 | CODED PICTURE BUFFER ARRIVAL AND NOMINAL REMOVAL TIMES IN VIDEO CODING - A video coding device, such as a video decoder, may be configured to derive at least one of a coded picture buffer (CPB) arrival time and a CPB nominal removal time for an access unit (AU) at both an access unit level and a sub-picture level regardless of a value of a syntax element that defines whether a decoding unit (DU) is the entire AU. The video coding device may further be configured to determine a removal time of the AU based at least in part on one of the CPB arrival time and a CPB nominal removal time and decode video data of the AU based at least in part on the removal time. | 03-27-2014 |
20140092213 | SUB-BITSTREAM EXTRACTION FOR MULTIVIEW, THREE-DIMENSIONAL (3D) AND SCALABLE MEDIA BITSTREAMS - Techniques are described for modal sub-bitstream extraction. For example, a network entity may select a sub-bitstream extraction mode from a plurality of sub-bitstream extraction modes. Each sub-bitstream extraction mode may define a particular manner in which to extract coded pictures from views or layers to allow a video decoder to decode target output views or layers for display. In this manner, the network entity may adaptively select the appropriate sub-bitstream extraction technique, rather than a rigid, fixed sub-bitstream extraction technique. | 04-03-2014 |
20140092955 | SIGNALING LAYER IDENTIFIERS FOR OPERATION POINTS IN VIDEO CODING - Techniques described herein are related to coding layer identifiers for operation points in video coding. In one example, a method of decoding video data is provided. The method comprises decoding syntax elements in a video parameter set (VPS) within a conforming bitstream indicating a first operation point having a first set of content. The method further comprises decoding, if present, syntax elements in the VPS within the conforming bitstream indicating hypothetical reference decoder (HRD) parameter information having a second set of content for the first operation point, wherein the conforming bitstream does not include syntax elements in the VPS that duplicate at least one of the first or second set of content for a second operation point, and wherein decoding syntax elements comprises decoding the syntax elements indicating the first operation point and the HRD parameter information only within conforming bitstreams. | 04-03-2014 |
20140092963 | SIGNALING OF REGIONS OF INTEREST AND GRADUAL DECODING REFRESH IN VIDEO CODING - During a coding process, systems, methods, and apparatus may code information indicating whether gradual decoder refresh (GDR) of a picture is enabled. When GDR is enabled, the coding process, systems, methods, and apparatus may code information that indicates whether one or more slices of the picture belong to a foreground region of the picture. In another example, during a coding process, systems, methods, and apparatus may decode video data corresponding to an ISP identification (ISP ID) for one of the ISPs for slices of a picture. The systems, methods, and apparatus may decode video data corresponding to an ROI using the ISP. | 04-03-2014 |
20140092993 | ERROR RESILIENT DECODING UNIT ASSOCIATION - Techniques are described for signaling decoding unit identifiers for decoding units of an access unit. The video decoder determines which network abstraction layer (NAL) units are associated with which decoding units based on the decoding unit identifiers. Techniques are also described for including one or more copies of supplemental enhancement information (SEI) messages in an access unit. | 04-03-2014 |
20140092994 | SUPPLEMENTAL ENHANCEMENT INFORMATION MESSAGE CODING - Techniques are described for signaling decoding unit identifiers for decoding units of an access unit. The video decoder determines which network abstraction layer (NAL) units are associated with which decoding units based on the decoding unit identifiers. Techniques are also described for including one or more copies of supplemental enhancement information (SEI) messages in an access unit. | 04-03-2014 |
20140092996 | SIGNALING OF LAYER IDENTIFIERS FOR OPERATION POINTS - A device for processing video data receives an indication of a maximum layer identification (ID) value for a bitstream; receives a flag for a first layer with a layer ID value less than the maximum layer ID value; and, based on a value of the flag, determines if the first layer is included in an operation point. | 04-03-2014 |
20140098860 | FILE FORMAT FOR VIDEO DATA - A device generates a file that stores coded samples that contain coded pictures of the video data. The file also includes a sample entry that includes an element that indicates whether all sequence parameter sets (SPSs) that are activated when a stream to which the sample entry applies is decoded have syntax elements that indicate that temporal sub-layer up-switching to any higher temporal sub-layer can be performed at any sample associated with the SPSs. The same or different device determines, based on the element in the sample entry, that all SPSs that are activated when the stream to which the sample entry applies is decoded have syntax elements that indicate that temporal sub-layer up-switching to any higher temporal sub-layer can be performed at any sample associated with the SPSs. | 04-10-2014 |
20140098868 | FILE FORMAT FOR VIDEO DATA - A device generates a file that comprises a plurality of samples that contain coded pictures. In addition, the file contains a box that identifies a sample group that contains one or more samples from among the plurality of samples, wherein the box further indicates that each sample in the sample group is a step-wise temporal sub-layer access (STSA) sample. The same or different device identifies, based on data in the box that identifies the sample group, STSA samples from among the samples in the file that contains the box. | 04-10-2014 |
20140098894 | IDENTIFICATION OF OPERATION POINTS APPLICABLE TO NESTED SEI MESSAGE IN VIDEO CODING - A device decodes, from a scalable nesting supplemental enhancement information (SEI) message in an encoded video bitstream, a plurality of syntax elements that identify a plurality of operation points to which a nested SEI message encapsulated by the scalable nesting SEI message applies. Furthermore, the device uses one or more syntax elements of the nested SEI message to perform an operation regarding any of the operation points to which the nested SEI message applies. | 04-10-2014 |
20140098895 | HYPOTHETICAL REFERENCE DECODER PARAMETER SYNTAX STRUCTURE - A video encoder signals, in an encoded video bitstream, a video parameter set (VPS) that includes a plurality of Hypothetical Reference Decoder (HRD) parameter syntax structures that each include HRD parameters. For each respective HRD parameter syntax structure in the plurality of HRD parameter syntax structures, the VPS further includes a syntax element indicating whether the HRD parameters of the respective HRD parameter syntax structure include a common set of HRD parameters in addition to a set of sub-layer-specific HRD parameter information specific to a particular sub-layer of the encoded video bitstream. The common set of HRD parameters is common to all sub-layers of the encoded video bitstream. A video decoder or other device decodes, from the encoded video bitstream, the VPS and performs an operation using the HRD parameters of at least one of the HRD parameter syntax structures. | 04-10-2014 |
20140098896 | SUB-BITSTREAM APPLICABILITY TO NESTED SEI MESSAGES IN VIDEO CODING - A device determines, based at least in part on a syntax element in a scalable nesting supplemental enhancement information (SEI) message encapsulated by an SEI Network Abstraction Layer (NAL) unit, whether a nested SEI message encapsulated by the scalable nesting SEI message is applicable to a default sub-bitstream. The default sub-bitstream is an operation point representation of an operation point defined by a layer identifier specified in a NAL unit header of the SEI NAL unit and a temporal identifier specified in the NAL unit header. When the nested SEI message is applicable to the default sub-bitstream, the device uses the nested SEI message in an operation on the default sub-bitstream. | 04-10-2014 |
20140119437 | TARGET OUTPUT LAYERS IN VIDEO CODING - In one example, a device includes a video coder configured to code a multilayer bitstream comprising a plurality of layers of video data, where the plurality of layers of video data are associated with a plurality of layer sets, and where each layer set contains one or more layers of video data of the plurality of layers, and to code on one or more syntax elements of the bitstream indicating one or more output operation points, where each output operation point is associated with a layer set of the plurality of layer sets and one or more target output layers of the plurality of layers. | 05-01-2014 |
20140119447 | REFERENCE PICTURE STATUS FOR VIDEO CODING - The techniques of this disclosure may be generally related to reference statues of pictures. The techniques may store the reference status information of reference pictures of a picture, at an instance when the picture is being coded. The techniques may then utilize the reference status information of the reference pictures of the picture, at the instance when the picture was coded, to inter-predict video blocks of a subsequent picture. | 05-01-2014 |
20140133556 | MPEG FRAME COMPATIBLE VIDEO CODING - In an example, a video coder may determine a first layer component of a first layer of video data, wherein the first layer of video data is associated with a layer identifier. The video coder may generate at least one filtered layer component by filtering the first layer component, and assign the layer identifier of the first layer and a filtered layer component index to the at least one filtered layer component, where the filtered layer component index is different than a layer component index of the first layer component. The video coder may also add the at least one filtered layer component to a reference picture set for performing inter-layer prediction of a layer other than the first layer of video data. | 05-15-2014 |
20140140399 | LOW-COMPLEXITY SUPPORT OF MULTIPLE LAYERS FOR HEVC EXTENSIONS IN VIDEO CODING - A method of coding video data includes receiving video information associated with a first layer and a second layer and determining whether at least one of the first layer and the second layer is a default layer. The method can include at least partially restricting inter-layer prediction when neither the first layer nor the second layer is the default layer. A default layer can be a base layer or an enhancement layer. A flag can be received that indicates that inter-layer prediction is to be restricted. In addition, the method can include determining whether inter-layer prediction is allowed for the video information associated with the first layer, and determining whether inter-layer prediction is partially allowed for the video information associated with the second layer such that motion compensation is not used with the second layer video information. | 05-22-2014 |
20140169448 | LOW-DELAY BUFFERING MODEL IN VIDEO CODING - Techniques for low-delay buffering in a video coding process are disclosed. Video decoding techniques may include receiving a first decoded picture buffer (DPB) output delay and a second DPB output delay for a decoded picture, determining, for the decoded picture, a first DPB output time using the first DPB output delay in the case a hypothetical reference decoder (HRD) setting for a video decoder indicates operation at a picture level, and determining, for the decoded picture, a second DPB output time using the second DPB output delay in the case that the HRD setting for the video decoder indicates operation at a sub-picture level. | 06-19-2014 |
20140185670 | PROGRESSIVE REFINEMENT WITH TEMPORAL SCALABILITY SUPPORT IN VIDEO CODING - An example method of decoding video data includes receiving information indicating a picture order count (POC) value of a last picture of a plurality of pictures in a progressive refinement segment in an encoded bitstream, and decoding at least some of the pictures in the progressive refinement segment according to a progressive refinement operation based on the received information. | 07-03-2014 |
20140185682 | MOTION VECTOR PREDICTION FOR VIDEO CODING - In one example, a device for coding video data includes a video coder configured to determine a first target reference index representing a reference picture to which a candidate motion vector predictor for a current motion vector refers, determine, for the candidate motion vector predictor and during temporal motion vector prediction (TMVP) of the current motion vector, a value for an additional target reference index that is equal to a predetermined value, and code the current motion vector based at least in part on the value of the additional reference index. | 07-03-2014 |
20140192149 | NON-NESTED SEI MESSAGES IN VIDEO CODING - A device obtains, from a bitstream that includes an encoded representation of the video data, a non-nested Supplemental Enhancement Information (SEI) message that is not nested within another SEI message in the bitstream. Furthermore, the device determines a layer of the bitstream to which the non-nested SEI message is applicable. The non-nested SEI message is applicable to layers for which video coding layer (VCL) network abstraction layer (NAL) units of the bitstream have layer identifiers equal to a layer identifier of a SEI NAL unit that encapsulates the non-nested SEI message. A temporal identifier of the SEI NAL unit is equal to a temporal identifier of an access unit containing the SEI NAL unit. Furthermore, the device processes, based in part on one or more syntax elements in the non-nested SEI message, video data of the layer of the bitstream to which the non-nested SEI message is applicable. | 07-10-2014 |
20140192151 | INDICATION OF PRESENCE OF TEXTURE AND DEPTH VIEWS IN TRACKS FOR MULTIVIEW CODING PLUS DEPTH - Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of video data, wherein the track includes one or more views. The method further includes parsing information to determine whether the track includes only texture views, only depth views, or both texture and depth views. Another example method includes composing a track of video data, wherein the track includes one or more views and composing information that indicates whether the track includes only texture views, only depth views, or both texture and depth views. | 07-10-2014 |
20140192152 | INDICATION OF CURRENT VIEW DEPENDENCY ON REFERENCE VIEW IN MULTIVIEW CODING FILE FORMAT - Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of video data, wherein the track includes one or more views. The method further includes parsing information to determine whether a texture view or a depth view of a reference view is required for decoding at least one of the one or more views in the track. Another example method includes composing a track of video data, wherein the track includes one or more views and composing information that indicates whether a texture view or a depth view of a reference view is required for decoding at least one of the one or more views in the track. | 07-10-2014 |
20140192153 | SIGNALING OF SPATIAL RESOLUTION OF DEPTH VIEWS IN MULTIVIEW CODING FILE FORMAT - Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of multiview video data, wherein the track includes at least one depth view. The method further includes parsing information to determine a spatial resolution associated with the depth view, wherein decoding the spatial resolution does not require parsing of a sequence parameter set of the depth view. Another example method includes composing a track of multiview video data, wherein the track includes the one or more views. The example method further includes composing information to indicate a spatial resolution associated with the depth view, wherein decoding the spatial resolution does not require parsing of a sequence parameter set of the depth view. | 07-10-2014 |
20140192157 | VIEW SYNTHESIS IN 3D VIDEO - In an example, a method of decoding video data includes determining whether a reference index for a current block corresponds to an inter-view reference picture, and when the reference index for the current block corresponds to the inter-view reference picture, obtaining, from an encoded bitstream, data indicating a view synthesis prediction (VSP) mode of the current block, where the VSP mode for the reference index indicates whether the current block is predicted with view synthesis prediction from the inter-view reference picture. | 07-10-2014 |
20140192882 | VIDEO BUFFERING OPERATIONS FOR RANDOM ACCESS IN VIDEO CODING - In an example, the present disclosure provides for receiving in a video bitstream an access unit having a first random access point (RAP) picture and receiving in the video bitstream, after the access unit in the bitstream, a subsequent access unit having a second RAP picture. In a case that one or more random access skipped leading (RASL) pictures for the subsequent access unit are not present in the received video bitstream, shifting a picture buffer removal time earlier based on a picture buffer removal delay offset. Another example provides for receiving an access unit after an earlier initialization of the hypothetical reference decoder (HRD), the access unit having a RAP picture, wherein associated access units containing RASL pictures are not received and initializing a picture buffer removal time and a picture buffer removal delay offset in response to receiving the access unit and not receiving the associated access units containing RASL pictures. | 07-10-2014 |
20140192895 | MULTI-RESOLUTION DECODED PICTURE BUFFER MANAGEMENT FOR MULTI-LAYER VIDEO CODING - This disclosure describes various methods and techniques for decoded picture buffer (DPB) management when multiple decoded layer components with different resolutions need to be stored. In one example, a method of coding video data comprises decoding video data to produce a plurality of decoded layer components, storing the decoded layer components in one or more sub-units of a DPB, and performing a DPB management process on the one or more sub-units, wherein the DPB management process is managed separately for each of the one or more sub-units. | 07-10-2014 |
20140192896 | GRADUAL DECODING REFRESH WITH TEMPORAL SCALABILITY SUPPORT IN VIDEO CODING - An example method of decoding video data includes receiving a plurality of pictures, receiving, in a message associated with a first picture of the plurality of pictures, information indicating a picture order count (POC) value of a recovery point picture of a gradual decoder refresh (GDR) set, when a picture that follows the first picture in decoding order has the POC value that is equal to a POC value of the recovery point picture, identifying the picture having the POC value equal to the POC value of the recovery point picture, as a recovery point picture, and when none of the pictures that follow the first picture has the POC value that is equal to the POC value of the recovery point picture, identifying one of the pictures that has a POC value greater than the POC value of the recovery point picture as the recovery point picture. | 07-10-2014 |
20140192897 | GRADUAL DECODING REFRESH WITH TEMPORAL SCALABILITY SUPPORT IN VIDEO CODING - An example method of decoding video data includes receiving, in a message associated with a picture, information indicating a refreshed region of the picture, determining whether the picture comprises a last picture in a gradual decoder refresh (GDR) set, determining whether the picture comprises a recovery point picture, and responsive to determining that the picture comprises the last picture in the GDR set and the recovery point picture, determining that the message indicates that the entire picture belongs to the refreshed region of the picture. | 07-10-2014 |
20140192901 | CONDITIONAL SIGNALING OF PICTURE ORDER COUNT TIMING INFORMATION FOR VIDEO TIMING IN VIDEO CODING - In an example, the disclosure provides for encoding pictures of a video sequence to generate a coded video sequence comprising the encoded pictures and signaling timing parameters for the coded video sequence by directly signaling a condition for signaling a number of clock ticks corresponding to a difference of picture order count (POC) values equal to 1 in at least one of a video parameter set (VPS) syntax structure referenced by the coded video sequence and a sequence parameter set (SPS) syntax structure referenced by the coded video sequence. Another example provides for receiving timing parameters for a coded video sequence that include a condition for signaling a number of clock ticks corresponding to a difference of POC values equal to 1 directly in at least one of a VPS syntax structure referenced by the coded video sequence and an SPS syntax structure referenced by the coded video sequence. | 07-10-2014 |
20140192902 | SIGNALING OF CLOCK TICK DERIVATION INFORMATION FOR VIDEO TIMING IN VIDEO CODING - In an example, the disclosure provides for receiving a coded video sequence comprising encoded pictures of a video sequence and receiving timing parameters for the coded video sequence that include a time scale and a number of units in a clock tick at most once in a video parameter set (VPS) syntax structure referenced by the coded video sequence and at most once in a video usability information (VUI) part of a sequence parameter set (SPS) syntax structure referenced by the coded video sequence. Another example provides for encoding pictures of a video sequence to generate a coded video sequence and signaling timing parameters for the coded video sequence by at least in part signaling a time scale and a number of units in a clock tick at most once in a VPS syntax structure and at most once in a VUI part of a SPS syntax structure. | 07-10-2014 |
20140192903 | SIGNALING OF PICTURE ORDER COUNT TO TIMING INFORMATION RELATIONS FOR VIDEO TIMING IN VIDEO CODING - In an example, the disclosure provides for receiving a coded video sequence comprising encoded pictures of a video sequence and receiving timing parameters for the coded video sequence that include an indication of whether a picture order count (POC) value for each picture in the coded video sequence that is not a first picture in the coded video sequence according to a decoding order is proportional to an output time of the picture relative to an output time of the first picture in the coded video sequence in a video parameter set (VPS) syntax structure referenced by the coded video sequence. Another example provides for encoding pictures of a video sequence to generate the coded video sequence comprising the encoded pictures and signaling timing parameters for the coded video sequence by signaling the indication in the VPS syntax structure referenced by the coded video sequence. | 07-10-2014 |
20140193139 | SEPARATE TRACK STORAGE OF TEXTURE AND DEPTH VIEWS FOR MULTIVIEW CODING PLUS DEPTH - Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of multiview video data, wherein the track includes one or more views, including only one of a texture view of a particular view and a depth view of the particular view. The method further includes parsing a track reference to determine a dependency of the track to a referenced track indicated in the track reference. Track reference types include ‘deps’ that indicates that the track includes the depth view of the particular view and the reference track includes the texture, ‘tref’ that indicates that the track depends on the texture view which is stored in the referenced track, and ‘dref’ that indicates that the track depends on the depth view which is stored in the referenced track. | 07-10-2014 |
20140241437 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a reference layer (RL) and an enhancement layer, the RL comprising an RL picture having an output region that includes a portion of the RL picture. The processor is configured to determine whether a condition indicates that information outside of the output region is available to predict a current block in the enhancement layer. The processor may encode or decode the video information. | 08-28-2014 |
20140254666 | PARALLEL PROCESSING FOR VIDEO CODING - In one example, a device for coding video data includes a video coder configured to code data representative of whether a tile of an enhancement layer picture can be predicted using inter-layer prediction, and predict data of the tile using inter-layer prediction only when the data indicates that the tile can be predicted using inter-layer prediction. | 09-11-2014 |
20140254667 | PARALLEL PROCESSING FOR VIDEO CODING - In one example, a device for coding video data includes a video coder configured to code data indicating whether tile boundaries of different layers of video data are aligned and whether inter-layer prediction is allowed along or across tile boundaries of enhancement layer blocks, code an enhancement layer block in an enhancement layer tile of the video data without using inter-layer prediction from a collocated base layer block for which inter-layer filtering or reference layer filtering across tile boundaries in a reference layer picture in an access unit including both the enhancement layer tile and the base layer block is enabled, and code the collocated base layer block. | 09-11-2014 |
20140254668 | PARALLEL PROCESSING FOR VIDEO CODING - In one example, a device for coding video data includes a video coder configured to code a value for a syntax element representative of whether any two reference layer samples, collocated with two respective enhancement layer picture samples within a common enhancement layer tile, must be within a common reference layer tile, and code the enhancement layer picture samples based at least in part on the value of the syntax element. | 09-11-2014 |
20140254669 | PARALLEL PROCESSING FOR VIDEO CODING - In one example, a device for coding video data includes a video coder configured to code, for a tile of an enhancement layer picture, data indicating a number of tiles in a base layer picture that need to be coded before the tile of the enhancement layer picture can be coded, code tiles of the base layer picture corresponding to the number of tiles, and, after coding the tiles of the base layer picture, code the tile of the enhancement layer picture substantially in parallel with at least one other tile in the base layer picture. | 09-11-2014 |
20140254679 | INTER-LAYER REFERENCE PICTURE CONSTRUCTION FOR SPATIAL SCALABILITY WITH DIFFERENT ASPECT RATIOS - A method of coding video data includes upsampling at least a portion of a reference layer picture to an upsampled picture having an upsampled picture size. The upsampled picture size has a horizontal upsampled picture size and a vertical upsampled picture size. At least one of the horizontal or vertical upsampled picture sizes may be different than a horizontal picture size or vertical picture size, respectively, of an enhancement layer picture. In addition, position information associated with the upsampled picture may be signaled. An inter-layer reference picture may be generated based on the upsampled picture and the position information. | 09-11-2014 |
20140294062 | INTER-LAYER REFERENCE PICTURE RESTRICTION FOR HIGH LEVEL SYNTAX-ONLY SCALABLE VIDEO CODING - In one implementation, an apparatus is provided for encoding or decoding video information. The apparatus comprises a memory unit configured to store reference layer pictures associated with a reference layer, an enhancement layer, or both. The apparatus further comprises a processor operationally coupled to the memory unit. In one embodiment, the processor is configured to restrict usage of at most one reference layer pictures that has been resampled as an inter-layer reference picture, and predict a current picture using inter-layer prediction and the inter-layer reference picture. | 10-02-2014 |
20140294063 | INTER-LAYER REFERENCE PICTURE RESTRICTION FOR HIGH LEVEL SYNTAX-ONLY SCALABLE VIDEO CODING - In one implementation, an apparatus is provided for encoding or decoding video information. The apparatus comprises a memory unit configured to store reference layer pictures associated with a reference layer, an enhancement layer, or both. The apparatus further comprises a processor operationally coupled to the memory unit. In one embodiment, the processor is configured to restrict usage of at most one reference layer pictures that has a different spatial resolution than a current picture as an inter-layer reference picture, and predict the current picture using inter-layer prediction and the inter-layer reference picture. | 10-02-2014 |
20140294064 | RTP PAYLOAD FORMAT DESIGNS - A device for processing video data includes a memory; a receiver configured to real-time transport protocol (RTP) packets; and one or more processors configured to receive a first fragmentation unit comprising a subset of a fragmented network abstraction layer (NAL) unit; parse a start bit of the fragmentation unit to determine if the first fragmentation unit comprises a start of the fragmented NAL unit; in response to the first fragmentation unit comprising the start of the fragmented NAL unit and one or both of a transmission mode for the first fragmentation unit being a multi-session transmission mode and a first parameter being greater than a first value, parse a second parameter to determine a decoding order for the fragmented NAL unit; decode the fragmented NAL unit based on the determined decoding order. | 10-02-2014 |
20140294092 | RTP PAYLOAD FORMAT DESIGNS - A device for processing video data includes a memory; a receiver configured to real-time transport protocol (RTP) packets; and one or more processors configured to: receive a first real-time transport protocol (RTP) packet comprising a first network abstraction layer (NAL) unit, and in response to a transmission mode for the first RTP packet being a single session transmission mode and a first parameter being equal to a first value, determine a decoding order number for the first NAL unit based on a transmission order of the first NAL unit. | 10-02-2014 |
20140294093 | RTP PAYLOAD FORMAT DESIGNS - A device for processing video data includes a memory, a receiver configured to real-time transport protocol (RTP) packets, and one or more processors configured to receive a first aggregation packet according to a real-time transfer protocol (RTP), wherein the first aggregation packet comprises a payload header and one or more aggregation units; parse a first aggregation unit that is the first aggregation unit of the first aggregation packet to determine a value for a first parameter, wherein the first parameter specifies a decoding order number for a NAL unit included in the first aggregation packet; parse a second aggregation unit to determine a value for a second parameter, wherein the second aggregation unit follows the first aggregation unit in the first aggregation packet; and, based on the first parameter and the second parameter, determine a decoding order for a NAL unit included in the second aggregation unit. | 10-02-2014 |
20140301436 | CROSS-LAYER ALIGNMENT IN MULTI-LAYER VIDEO CODING - An apparatus for coding video information according to certain aspects includes a memory unit and a video processor in communication with the memory unit. The video processor is configured to identify a first picture included in the first set of pictures, wherein pictures within the first set of pictures having an output position after the output position of the first picture also have a decoding position after the decoding position of the first picture. The video processor is further configured to identify a second picture included in the second set of pictures, wherein pictures within the second set of pictures having an output position after the output position of the second pictures also have a decoding position after the decoding position of the second picture. The video processor is also configured to code the identified first picture and the identified second picture via one syntax element into one access unit. | 10-09-2014 |
20140301437 | PICTURE ALIGNMENTS IN MULTI-LAYER VIDEO CODING - A video encoder generates an encoded video bitstream comprising a plurality of layers. At least one access unit of the encoded video bitstream includes one or more Instantaneous Decoding Refresh (IDR) pictures and one or more non-IDR pictures. The IDR pictures include only slices that are decoded using intra prediction only. The video encoder outputs the encoded video bitstream. A video decoder decodes the encoded video bitstream. | 10-09-2014 |
20140301439 | CROSS-LAYER POC ALIGNMENT FOR MULTI-LAYER BITSTREAMS THAT MAY INCLUDE NON-ALIGNED IRAP PICTURES - In one example, a video coder is configured to code a value for a syntax element indicating whether at least a portion of a picture order count (POC) value of a picture is to be reset to a value of zero, when the value for the syntax element indicates that the portion of the POC value is to be reset to the value of zero, reset at least the portion of the POC value such that the portion of the POC value is equal to zero, and code video data using the reset POC value. Coding video data using the reset POC value may include inter-predicting a block of a subsequent picture relative to the picture, where the block may include a motion parameter that identifies the picture using the reset POC value. The block may be coded using temporal inter-prediction or inter-layer prediction. | 10-09-2014 |
20140301441 | NON-ENTROPY ENCODED SET OF PROFILE, TIER, AND LEVEL SYNTAX STRUCTURES - Systems, methods, and devices for coding multilayer video data are disclosed that may include, encoding, decoding, transmitting, or receiving a non-entropy encoded set of profile, tier, and level syntax structures, potentially at a position within a video parameter set (VPS) extension. The systems, methods, and devices may refer to one of the profile, tier, and level syntax structures for each of a plurality of output layer sets. The systems, methods, and devices may encode or decode video data of one of the output layer sets based on information from the profile, tier, and level syntax structure referred to for the output layer set. | 10-09-2014 |
20140301456 | INTER-LAYER PICTURE SIGNALING AND RELATED PROCESSES - In one implementation, an apparatus is provided for encoding or decoding video information. The apparatus comprises a memory configured to store inter-layer reference pictures associated with a current picture that is being coded. The apparatus further comprises a processor operationally coupled to the memory. In one embodiment, the processor is configured to indicate a number of inter-layer reference pictures to use to predict the current picture using inter-layer prediction. The processor is also configured to indicate which of the inter-layer reference pictures to use to predict the current picture using inter-layer prediction. The processor is also configured to determine an inter-layer reference picture set associated with the current picture using the indication of the number of inter-layer reference pictures and the indication of which of the inter-layer reference pictures to use to predict the current picture using inter-layer prediction. | 10-09-2014 |
20140301466 | GENERALIZED RESIDUAL PREDICTION IN HIGH-LEVEL SYNTAX ONLY SHVC AND SIGNALING AND MANAGEMENT THEREOF - An apparatus for decoding video information according to certain aspects includes a memory unit and a processor operationally coupled to the memory unit. The memory unit is configured to store at least one reference picture list of an enhancement layer, the at least one reference picture list comprising residual prediction reference picture information. The processor is configured to: decode signaled information about residual prediction reference picture generation; generate a residual prediction reference picture based on an enhancement layer reference picture and the decoded signaled information such that the generated residual prediction reference picture has the same motion field and the same picture order count (POC) as the enhancement layer reference picture from which it is generated; and store the generated residual prediction reference picture in the at least one reference picture list of the enhancement layer. | 10-09-2014 |
20140301469 | CODING VIDEO DATA FOR AN OUTPUT LAYER SET - Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving multilayer video data. The systems, methods, and devices may receive or transmit a first output layer set for a layer set and receive or transmit a second output layer set for the layer set. The systems, methods, and devices may code (encode or decode) video data for at least one of the first output layer set and the second output layer set. | 10-09-2014 |
20140301483 | NON-ENTROPY ENCODED REPRESENTATION FORMAT - Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving multilayer video data. The systems, methods, and devices may receive or transmit a non-entropy coded representation format within a video parameter set (VPS). The systems, methods, and devices may code (encode or decode) video data based on the non-entropy coded representation format within the VPS, wherein the representation format includes one or more of chroma format, whether different colour planes are separately coded, picture width, picture height, luma bit depth, and chroma bit depth. | 10-09-2014 |
20140301484 | PARAMETER SET DESIGNS FOR VIDEO CODING EXTENSIONS - Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving multilayer video data. The systems, methods, and devices may transmit or receive a video parameter set (VPS) including information for a series of layers, each layer including visual signal information. The systems, methods, and devices may code (encode or decode) video data based on the visual signal information signaled per layer in the VPS. | 10-09-2014 |
20140301485 | IRAP ACCESS UNITS AND BITSTREAM SWITCHING AND SPLICING - In one example, a device for coding video data includes a video coder configured to code an intra random access point (IRAP) picture of a partially aligned IRAP access unit of video data, and code data that indicates, when performing random access from the partially aligned IRAP access unit, at least one picture of a video coding layer that is not correctly decodable. When the video coder comprises a video decoder, the video decoder may skip decoding of the pictures that are not correctly decodable, assuming random access has been performed starting from the partially aligned IRAP access unit. | 10-09-2014 |
20140307803 | NON-ENTROPY ENCODED LAYER DEPENDENCY INFORMATION - Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving a non-entropy encoded layer dependency information at a position within a video parameter set (VPS) extension prior to syntax elements of the VPS extension that are entropy encoded. The systems, methods, and devices may encode or decode the non-entropy encoded layer dependency information before an entropy encoded syntax element. The systems, methods, and devices may encode or decode video data of one or more of the layers of video data based on the non-entropy encoded layer dependency information. The layer dependency information indicates whether one of the layers is a direct reference layer for another of the layers. | 10-16-2014 |
20140314140 | INDICATION OF CROSS-LAYER PICTURE TYPE ALIGNMENT IN MULTI-LAYER VIDEO CODING - In one example, the disclosure is directed to techniques that include receiving a bitstream comprising at least a syntax element, a first network abstraction layer unit type, and a coded access unit comprising a plurality of pictures. The techniques further include determining a value of the syntax element which indicates whether the access unit was coded using cross-layer alignment. The techniques further include determining the first network abstraction layer unit type for a picture in the access unit and determining whether the first network abstraction layer unit type equals a value in a range of type values. The techniques further include setting a network abstraction layer unit type for all other pictures in the coded access unit to equal the value of the first network abstraction layer unit type if the first network abstraction layer unit type is equal to a value in the range of type values. | 10-23-2014 |
20140341275 | SIGNALING PICTURE SIZE IN VIDEO CODING - A video encoder is configured to determine a picture size for one or more pictures included in a video sequence. The picture size associated with the video sequence may be a multiple of an aligned coding unit size for the video sequence. In one example, the aligned coding unit size for the video sequence may comprise a minimum coding unit size where the minimum coding unit size is selected from a plurality of smallest coding unit sizes corresponding to different pictures in the video sequence. A video decoder is configured to obtain syntax elements to determine the picture size and the aligned coding unit size for the video sequence. The video decoder decodes the pictures included in the video sequence with the picture size, and stores the decoded pictures in a decoded picture buffer. | 11-20-2014 |
20140355616 | SINGLE NETWORK ABSTRACTION LAYER UNIT PACKETS WITH DECODING ORDER NUMBER FOR VIDEO CODING - A method for processing video data in a real-time transport protocol (RTP) payload includes encapsulating video data in a single network abstraction layer (NAL) unit packet for an RTP session. The single NAL unit packet contains a single NAL unit. The method may also include encapsulating decoding order number information in the single NAL unit packet based on at least one of: the RTP session being in a multi-stream transmission (MST) mode, or a maximum number of NAL units that may precede the NAL unit in a de-packetization buffer in reception order and follow the NAL unit in decoding order being greater than 0. | 12-04-2014 |
20140355692 | HYPOTHETICAL REFERENCE DECODER MODEL AND CONFORMANCE FOR CROSS-LAYER RANDOM ACCESS SKIPPED PICTURES - A device may determine, based on a value, whether all cross-layer random access skipped (CL-RAS) pictures of an intra random access point (IRAP) access unit are present in a video data bitstream. In addition, the device may reconstruct pictures of the video data based at least in part on syntax elements decoded from the video data bitstream. | 12-04-2014 |
20140362902 | INTER-LAYER PREDICTION TYPES IN MULTI-LAYER VIDEO CODING - An apparatus for coding video information according to certain aspects includes computing hardware. The computing hardware is configured to: identify a current picture to be predicted using at least one type of inter layer prediction (ILP), the type of ILP comprising one or more of inter layer motion prediction (ILMP) or inter layer sample prediction (ILSP); and control: (1) a number of pictures that may be resampled and used to predict the current picture using ILMP and (2) a number of pictures that may be resampled and used to predict the current picture using ILSP, wherein the computing hardware is configured to control the number of pictures that may be resampled and used to predict the current picture using ILMP independent of the number of pictures that may be resampled and used to predict the current picture using ILSP. | 12-11-2014 |
20140362910 | PROCESSING BITSTREAM CONSTRAINTS RELATING TO INTER-LAYER PREDICTION TYPES IN MULTI-LAYER VIDEO CODING - An apparatus for coding video information may include computing hardware configured to: when a current picture is to be predicted using at least inter layer motion prediction (ILMP): process a collocated reference index value associated with the current picture, wherein the collocated reference index value indicates a first reference picture that is used in predicting the current picture using inter layer prediction (ILP); and determine whether the first reference picture indicated by the collocated reference index value is enabled for ILMP; when the current picture is to be predicted using at least inter layer sample prediction (ILSP): process a reference index value associated with a block in the current picture, wherein the reference index value indicates a second reference picture that is used in predicting the block in the current picture using ILP; and determine whether the second reference picture indicated by the reference index value is enabled for ILSP. | 12-11-2014 |
20140369404 | SIGNALING OF DEBLOCKING FILTER PARAMETERS IN VIDEO CODING - This disclosure describes techniques for signaling deblocking filter parameters for a current slice of video data with reduced bitstream overhead. Deblocking filter parameters may be coded in one or more of a picture layer parameter set and a slice header. The techniques reduce a number of bits used to signal the deblocking filter parameters by coding a first syntax element that indicates whether deblocking filter parameters are present in both the picture layer parameter set and the slice header, and only coding a second syntax element in the slice header when both sets of deblocking filter parameters are present. Coding the second syntax element is eliminated when deblocking filter parameters are present in only one of the picture layer parameter set or the slice header. The second syntax element indicates which set of deblocking filter parameters to use to define a deblocking filter applied to a current slice. | 12-18-2014 |
20150016500 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a first layer and a second layer. The processor is configured to decode first layer pictures of the first layer, store the decoded first layer pictures in a decoded picture buffer, determine whether second layer pictures having no corresponding first layer pictures are to be coded, and in response to determining that second layer pictures having no corresponding first layer pictures are to be coded, process an indication that one or more decoded first layer pictures stored in the decoded picture buffer are to be removed. The processor may encode or decode the video information. | 01-15-2015 |
20150016502 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a current layer and an enhancement layer, the current layer having a current picture. The processor is configured to determine whether the current layer may be coded using information from the enhancement layer, determine whether the enhancement layer has an enhancement layer picture corresponding to the current picture, and in response to determining that the current layer may be coded using information from the enhancement layer and that the enhancement layer has an enhancement layer picture corresponding to the current picture, code the current picture based on the enhancement layer picture. The processor may encode or decode the video information. | 01-15-2015 |
20150016503 | TILES AND WAVEFRONT PROCESSING IN MULTI-LAYER CONTEXT - A video encoder may generate a bitstream that includes a syntax element that indicates whether inter-layer prediction is enabled for decoding a tile of a picture of the video data. Similarly, a video decoder may obtain, from a bitstream, a syntax element that indicates whether inter-layer prediction is enabled. The video decoder may determine, based on the syntax element, whether inter-layer prediction is enabled for decoding a tile of a picture of the video data, and decode the tile based on the determination. | 01-15-2015 |
20150016520 | CROSS-LAYER PARALLEL PROCESSING AND OFFSET DELAY PARAMETERS FOR VIDEO CODING - In an example, a method of coding video data includes coding data of a video parameter set (VPS) of a multi-layer bitstream, including at least one of data that indicates whether any layers of the multi-layer bitstream have an inter-layer prediction restriction or data that indicates whether tile boundaries are aligned between at least two of the layers of the multi-layer bitstream, and coding the multi-layer bitstream in accordance with the data of the VPS. | 01-15-2015 |
20150016532 | SELECTION OF TARGET OUTPUT LAYERS IN HIGH EFFICIENCY VIDEO CODING EXTENSIONS - A method, apparatus, and manufacture for processing video data. A list of output layer sets in a video bitstream is received, and an index to at least one target output layer set in the list of output layer sets is received. Next, target output layers in the at least one target output layer set is determined based on the index. At least the target output layers from the video bitstream are decoded. Then, the decoded target output layers are output without outputting layers that are not targeted for output. | 01-15-2015 |
20150016534 | SIGNALING OF VIEW ID BIT DEPTH IN PARAMETER SETS - An apparatus for encoding video information according to certain aspects includes a memory and computing hardware. The memory is configured to store video information. The computing hardware is configured to determine a bit depth of one or more view identifiers to signal, wherein each of the one or more view identifiers is associated with a layer to be encoded. The computing hardware is further configured to signal the bit depth of the one or more view identifiers in a bitstream. | 01-15-2015 |
20150016540 | CROSS-LAYER PARALLEL PROCESSING AND OFFSET DELAY PARAMETERS FOR VIDEO CODING - In an example, a method of coding video data includes determining a location of a reference sample associated with a reference picture of video data based on one or more scaled offset values, where the reference picture is included in a first layer of a multi-layer bitstream and the one or more scaled offset values indicate a difference in scale between the first layer and a second, different layer. The method also includes determining a location of a collocated reference block of video data in the first layer based on the location of the reference sample, and coding a current block of video data in the second layer relative to the collocated reference block. | 01-15-2015 |
20150016543 | BITSTREAM RESTRICTIONS ON PICTURE PARTITIONS ACROSS LAYERS - Bitstream restrictions or constraints on the partitioning of pictures across layers of video data are described. In some examples, the number of tiles per picture for each layer of a plurality of layers is constrained based on a maximum number of tiles per picture for the layer. In some examples, the number of tiles per picture for each layer of the plurality of layers is no greater than the maximum number of tiles per picture for the layer. In some examples, a sum of the numbers of tiles per picture for the plurality of layers is no greater than a sum of the maximum numbers of tiles per picture for the plurality of layers. In some examples, a second largest coding unit (LCU) or coding tree block (CTB) size for a second layer is constrained based on, e.g., to be equal to, a first LCU size for a first layer. | 01-15-2015 |
20150016544 | SIGNALING OF BIT RATE INFORMATION AND PICTURE RATE INFORMATION IN VPS - An apparatus for coding video information according to certain aspects includes a memory and computing hardware. The memory is configured to store video information. The computing hardware is configured to process at least one of a first signal indicating whether at least one sublayer of one or more layer sets has bit rate information to signal or a second signal indicating whether at least one sublayer of the one or more layer sets has picture rate information to signal. | 01-15-2015 |
20150016545 | DECODED PICTURE BUFFER OPERATIONS FOR VIDEO CODING - A method of decoding video data comprising partitioning a decoded picture buffer (DPB) into a plurality of sub-DPBs, receiving at least one indication of a sub-DPB size for the plurality of sub-DPBs for one or more operation points of a multi-layer video coding process, and allocating memory space for the plurality of sub-DPBs based on the at least one indication. | 01-15-2015 |
20150016546 | CROSS-LAYER ALIGNMENT OF INTRA RANDOM ACCESS POINT PICTURES - A video processing device includes a memory storing video data and one or more processors configured to: receive a first network abstraction layer (NAL) unit comprising a first picture of an access unit; in response to determining the first NAL unit comprises an intra random access point (IRAP) picture and in response to a NAL unit type for the first NAL unit indicating the presence of an instantaneous decoding refresh (IDR) picture without any associated leading pictures for a second NAL unit of the access unit comprising another IRAP picture, determine a NAL unit type for the second NAL unit to be a NAL unit type indicating the presence of an IDR picture without any associated leading pictures; and, process the first NAL unit and the second NAL unit based on the NAL unit type for the second NAL unit. | 01-15-2015 |
20150078457 | REPRESENTATION FORMAT SIGNALING IN MULTI-LAYER VIDEO CODING - Techniques are described for signaling of representation format information in multi-layer bitstreams. Representation format information is signaled using representation format syntax structures included in a video parameter set (VPS) for a video sequence in a multi-layer bitstream. When syntax elements associated with the representation format syntax structures are not present in the VPS, a mapping of representation formats to layers in the multi-layer bitstream may be inferred. According to the techniques, in the absence of the syntax elements, a video decoder infers which of the representation format syntax structures is applied to which of the layers in the bitstream based on a number of the representation format syntax structures included in the VPS for the video sequence. By basing the inference on the number of representation format syntax structures for the video sequence, the inference may be accurate for the type of multi-layer video extension used in the multi-layer bitstream. | 03-19-2015 |