US20060018378A1 - Method and system for delivery of coded information streams, related network and computer program product therefor - Google Patents

Method and system for delivery of coded information streams, related network and computer program product therefor Download PDF

Info

Publication number
US20060018378A1
US20060018378A1 US11/179,012 US17901205A US2006018378A1 US 20060018378 A1 US20060018378 A1 US 20060018378A1 US 17901205 A US17901205 A US 17901205A US 2006018378 A1 US2006018378 A1 US 2006018378A1
Authority
US
United States
Prior art keywords
transcoding
cost
stream
link
frame rate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/179,012
Inventor
Emiliano Piccinelli
Gianluca Filippini
Fabrizio Rovati
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
STMicroelectronics SRL
Original Assignee
STMicroelectronics SRL
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by STMicroelectronics SRL filed Critical STMicroelectronics SRL
Assigned to STMICROELECTRONICS S.R.L. reassignment STMICROELECTRONICS S.R.L. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FILIPPINI, GIANLUCA, PICCINELLI, EMILIANO MARIO ANGELO, ROVATI, FABRIZIO SIMONE
Publication of US20060018378A1 publication Critical patent/US20060018378A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/132Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/15Data rate or code amount at the encoder output by monitoring actual compressed data size at the memory before deciding storage at the transmission buffer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/152Data rate or code amount at the encoder output by measuring the fullness of the transmission buffer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/164Feedback from the receiver or from the transmission channel
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/40Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Definitions

  • the invention relates to coding techniques, for instance for video signals.
  • the invention may be applied to the dynamic adaptation of pre-encoded video streams to an error prone wireless channel, having an available bandwidth changing in a non deterministic way, while preserving the optimal quality of the streamed content.
  • the invention was developed by paying specific attention to the possible application to MPEG-2 coded video streams.
  • the MPEG video standard is based on a video compression procedure that exploits the high degree of spatial and temporal correlation in natural video sequences.
  • the MPEG standard defines the syntax and semantics of the transmitted bit-stream and the functionality of the decoder. However, the encoder is not strictly standardized: any encoder that produces a valid MPEG bit-streams is acceptable.
  • Motion estimation is used to compute similarities among successive pictures, in order to remove temporal redundancy, i.e., to transmit only the difference among successive pictures.
  • Block Matching Motion Estimation (BM-ME) is a common way to extract the similarities among pictures and it is the method selected by the MPEG-2 standard.
  • FIG. 1 shows an example of a hybrid Differential Pulse Code Modulation/Discrete Cosine Transform (DPCM/DCT)) coding loop (MPEG encoder).
  • DPCM/DCT Differential Pulse Code Modulation/Discrete Cosine Transform
  • Such an encoder system removes temporal redundancy from the video Input Sequence, indicated with reference I, which is initially elaborated by a Frame Reorder block 200 . Temporal redundancy is then removed in a block 210 that performs an inter-frame Motion Estimation (ME) operation.
  • ME inter-frame Motion Estimation
  • a summation node 212 operates on the output signal coming from the Motion Estimation block 210 , and also receives a signal representative of a previous picture coming from a block 280 that represents the Anchor Frames Buffer.
  • the residual error images obtained as output of the summation node 212 , are further processed by a block 220 in which a Discrete Cosine Transform (DCT) operation is performed, which reduces spatial redundancy by de-correlating the pixels within a block and concentrating the energy of the block itself into a few low order coefficients.
  • DCT Discrete Cosine Transform
  • blocks 230 and 240 representing respectively a Scalar Quantization (Quant) operation and a Variable Length Coding (VLC) operation, produce a bit-stream with good statistical compression efficiency.
  • the quantized sequence outgoing from block 230 is also fed to an Inverse Quantization (IQ) block 290 , and followed by a cascaded Inverse Discrete Cosine Transform (IDCT) block 295 .
  • IQ Inverse Quantization
  • IDCT Inverse Discrete Cosine Transform
  • the final bit-stream outputted by the Variable Length Coding block 240 is produced at variable bit-rate; hence, it has to be transformed to constant bit-rate by the insertion of a suitable Multiplexer 250 and an Output Buffer 260 .
  • Such an Output Buffer 260 feeds a control signal to a Feedback Bit-rate Controller, which decides the granularity of scalar quantization applied in the Scalar Quantization block 230 .
  • the Multiplexer 250 works on bit-stream signals coming from the Variable Length Coding block 240 and Motion Estimation block 210 .
  • a block 270 represents the already mentioned Feedback Bit-rate Controller and elaborates signals coming from Frame Reorder block 200 and Output Buffer block 260 .
  • the output signal from the Feedback Bit-rate Controller 270 is fed to the Quantization block 230 .
  • block 280 represents the Anchor Frames Buffer, which receives and elaborates inputs coming from Motion Estimation block 210 and from the summation node 214 .
  • the summation node 214 sums signals coming from the Anchor Frame Buffer block 280 and Inverse Discrete Cosine Transform (IDCT) block 295 .
  • the signal coming from the Anchor Frame Buffer block 280 is obtained from the motion vectors calculated by the Motion Estimation block 210 .
  • the coding technique just described must be considered in view of the fact that, in recent times, it has become increasingly important to be able to adapt the multimedia content to the client devices, and it widens increasingly the range of transformations needed.
  • a general access to the multimedia content can be provided in two basic ways; the first is by storing, managing, selecting, and delivering different versions of the media objects (images, video, audio, graphics and text) that include the multimedia presentations.
  • the second is by manipulating the media objects on the fly, by using, for example, methods for text-to-speech translation, image and video transcoding, media conversion, and summarization. This allows for adapting the multimedia content delivery to the wide diversity of client device capabilities in communication, processing storage and display.
  • transcoder The device that performs such an operation.
  • Such a device could be placed in the network to help relaying transmissions between these different bit rates or could be used as a pre-processing tool to create the various versions of the media objects that were mentioned earlier.
  • DVD Digital Video Disk
  • CIF Common Intermediate Format
  • the main issue is, therefore, to adapt the bit-rate and the picture resolution of a compressed data stream compliant to a video standard (e.g., MPEG-2) to another one.
  • the most widely adopted procedure is to decode the incoming bit-stream, optionally down-sampling the decoded images to generate a sequence with a reduced picture size, and re-encode with a new encoder, suitably set for achieving the required bit-rate.
  • FIG. 2 shows a block diagram of a transcoding chain, where reference numeral 300 indicates a sequence of blocks collectively named MPEG decoder and reference numeral 350 indicates a sequence of blocks collectively named MPEG encoder.
  • a video Input Bit-stream I′ representing information in the form of MPEG coded frames, is led to the MPEG decoder 300 .
  • video Input Bit-stream I′ is first fed to a Headers Detection block 302 .
  • the signal coming from the Headers Detection block 302 is then fed to a Buffer block 304 to be subsequently processed by a Demultiplexer Variable Length Coding Run-level Decoder block indicated with reference 306 in FIG. 2 .
  • Block 312 represents an Anchor Frame Buffer that operates by exploiting Motion Vector Values 314 and a Motion Vector Modes 316 that receives as input signals, originated from the Decoder block 306 .
  • the output signal from the Anchor Frame Buffer 312 is in turn fed to a summation node 320 that receives also the processed signal from the Inverse Discrete Cosine Transform block 310 .
  • the output signal from the summation node 320 is then fed to the MPEG encoder 350 , and also fed back to the Anchor Frame Buffer 312 .
  • the overall scheme of the MPEG encoder block 350 is substantially analogous to the scheme of the encoder 200 shown in FIG. 1 , but the signal outputted from the MPEG decoder block 300 is here sent preliminarily to a Chroma Filtering block 360 .
  • the filtered signal coming from block 360 is then sent to a Frame Reorder block 370 , and a cascaded Motion Estimation block 380 .
  • the Motion Vectors 385 originated from such Motion Estimation block 380 are fed to an Anchor Frame Buffer block 390 and to a Multiplexer 430 .
  • the signal coming from the Anchor Frame Buffer 390 is fed to two summation nodes, 382 and 392 , and fed back to the Motion Estimation block 380 .
  • the summation node 382 receives signals from the Frame Reorder block 370 and from the Anchor Frame Buffer block 390 .
  • the output signal from the summation node 382 is fed to a Discrete Cosine Transform block 400 to be subsequently quantized in a Quantizer block 410 .
  • the quantized signal from the block 410 is fed to a Variable Length Coding block 420 and to an Inverse Quantizer block 450 .
  • the signal coming out from Variable Length Coding block 420 is fed to a Multiplexer 430 .
  • the multiplexed signal is in turn fed to an Output Buffer 440 that returns the Output MPEG Bit-stream O.
  • the output signal from the block 440 is also fed to a Feedback Bit-rate Controller 470 , that controls the Quantizer block 410 .
  • the signal generated from the Inverse Quantizer block 450 is fed to an Inverse Discrete Cosine Transform block 460 .
  • the inverse transformed signal from block 460 is fed to the summation node 392 .
  • the output signal from the summation node 392 is fed to Anchor Frame Buffer block 390 .
  • FIG. 3 shows an example of such a transcoder system.
  • the video Input Bit-stream indicated with reference I′ is fed to a Sequence & Group Of Pictures headers block 500 that separates the signal and feeds a Sequence & Group Of Pictures Data delay memory block 510 and a Picture Header block 520 .
  • the output signal from block 520 is fed to a Local Cache memory block 530 .
  • a Multiplexer 540 operates on signals coming from block 520 and from block 530 .
  • the multiplexed signal from block 540 is fed to an Inverse-Variable Length Coding block 560 .
  • the signal coming from block 560 is fed to an Inverse-Run Length block 570 .
  • the resultant output signal is fed to a Re-quantizer block 585 that is made up with an Inverse Quantizer block 580 followed by a Quantizer block 590 .
  • the signal from block 585 is fed to a Run Length block 600 followed by a Variable Length Coding block 610 .
  • the output signal form block 610 is fed to a Bit Usage Profile block 620 followed by a Rate Control block 630 that controls the Quantizer block 590 .
  • the Multiplexer 640 elaborates the signals from the block 510 , the block 530 and the block 610 and returns the Output Bit-stream O.
  • Such a kind of system is able to operate only on a single level of transcoding. In order to operate with more levels, additional complexity should be included in the system.
  • explicit transcoding is a long and complex sequential chain, and involves various sources of inefficiency, because the motion information coming from the source stream is completely discarded.
  • the transcoder shown in FIG. 3 instead, is able to reuse inasmuch as possible the incoming information, to the expense of losses in terms of output bit-rate precision and drift being introduced.
  • WO02080518 is known a system for simultaneous transmission of multiple media streams in a fixed bandwidth network, wherein a state machine detects if the network bandwidth is close to saturation and executes a process of selecting at least one stream as a target for lowering total bandwidth usage.
  • the amount of data is reduced by a gradual degradation of the precision of the data, resulting in a greater potential for data compression, and/or by gradually reducing the resolution of the data of the target stream.
  • One embodiment of the present invention provides a method for dynamically adapting pre-encoded video streams to an error prone wireless channel, in which the available bandwidth changes in non-deterministic way, by matching the available bandwidth and preserving at the same time the optimal quality of the streamed content.
  • One embodiment of the present invention is a method having the features set forth in the claims that follow.
  • the invention also relates to a corresponding system, a related network as well as a related computer program product, loadable in the memory of at least one computer and including software code portions for performing the steps of the method of the invention when the product is run on a computer.
  • a computer program product is intended to be equivalent to reference to a computer-readable medium containing instructions for controlling a computer system to coordinate the performance of the method of the invention.
  • Reference to “at least one computer” is evidently intended to highlight the possibility for the present invention to be implemented in a distributed/modular fashion.
  • the arrangement described herein proposes a method that, substantially, provides for operating on a plurality of complexity levels, in particular four, and for choosing the best level among such a plurality of complexity levels by evaluating the available output bit rate and the complexity of the incoming compressed sequence. Such an evaluation is based on the available information contained in the incoming bit-stream, like the motion vector fields and the error level.
  • Adaptive video transcoding capabilities are becoming more and more important in a broad range of applications in the consumer electronics field; the proposed method be easily applied in such field, enhancing the performances of the transcoding system.
  • the proposed method enhances the quality achievable by a system able to transcode from one (or a plurality on multimedia stream to another one (or a plurality of), introducing very little computational complexity.
  • the arrangement described herein thus allows for deciding among re-quantization, frame rate down-sampling, frame size down-sampling procedures, that are very easy to be implemented, and allows the system to achieve better final quality, avoiding the drawbacks of the prior art.
  • FIG. 1 shows a block diagram of an MPEG encoder system
  • FIG. 2 shows a block diagram of an explicit transcoding chain
  • FIG. 3 shows a block diagram of a transcoder
  • FIG. 4 shows a basic block diagram of a typical Wireless Home Network scenario
  • FIG. 5 shows a diagram representative of a transcoding adaptation procedure according to the invention.
  • FIG. 6 shows a schematic diagram of a system for implementing a transcoding adaptation procedure according to the invention.
  • FIG. 4 shows a home access point 10 connected with a Wide Area Cellular Network 20 and with an Internet network 30 .
  • the home access point 10 is connected, by means of a Wireless Bridge 40 , with a Gateway 50 .
  • the Gateway 50 is in turn connected with several mobile terminals, like: Desktop Computer 60 , Printer 65 , Monitor device 70 , Digital Video Disk (DVD) reader 75 , Camcorder 80 , Personal Digital Assistant (PDA) 85 , Laptop Computer 90 , Digital Camera 95 , TV device 100 , and Audio device 105 .
  • Desktop Computer 60 Printer 65
  • Monitor device 70 Digital Video Disk (DVD) reader 75
  • Camcorder 80 Personal Digital Assistant (PDA) 85
  • Laptop Computer 90 Digital Camera 95
  • TV device 100 and Audio device 105 .
  • PDA Personal Digital Assistant
  • the available bandwidth for an 802.11 a/b/gin wireless LAN card is nominally 11 (11b), 54 (11a and 11g), >100(11n) Mbit/s respectively, but, depending on the number of active cards, the relative motions between server and clients and the obstacles between them, the real channel bandwidth could be reduced down to hundreds of Kbits per seconds. For this reason, there is the necessity to find a way to modify dynamically the incoming content to the variable transmission conditions.
  • the transcoder is a system able to convert a MPEG high quality compressed video stream into another one, that is less bandwidth consuming, through a dynamic re-quantization performed in the compressed domain.
  • dynamic re-quantization is not able to achieve compression bit rates lower than few Mbit/sec at full resolution with acceptable quality, because of the limitation of the standard and the intrinsic complexity of the task.
  • a method of controlling a network that operates at a plurality of transcoding levels, in particular at four different levels of complexity.
  • the proposed method provides for choosing the best level depending on the available output bit rate and the complexity of the incoming compressed sequence, i.e., basing the choice on the available information contained in the stream, like the motion vector fields and the error level.
  • the four level of transcoding are the following:
  • the resolution is varied by spatial downsampling, i.e., frame size downsampling, whereas the frame rate is varied by temporal downsampling.
  • Each of such levels can furthermore include sub-levels.
  • Each sublevel is characterized by a different quantization parameter.
  • the method described herein is able to change the frame size (full or reduced), frame rate (full or reduced) and quantization parameter (the whole MPEG-2 range).
  • such a stream generally contains I, P and B pictures.
  • I pictures are coded with reference to no other picture, so are self-decodable.
  • P pictures are instead coded starting from a Prediction coming from the previous I or P pictures.
  • B pictures are generally coded starting from one previous I or P picture, and one following I or P picture. B picture cannot be taken as reference for other picture.
  • the method proposed deals with procedures for using the above mentioned transcoding processes to dynamically adapt the stream to instantaneous channel bandwidth variations.
  • the proposed method envisages down-sampling operations for reducing both the frame rate and the frame size. This arises from the observation that, if the quantizer values are increased in order to decrease the bit-rate or bandwidth, above a certain value of compression factor/quantizer, it is better, in terms of perceived quality by the user, to watch a sequence that is reduced in frame size and/or frame rate, but less quantized, rather than a full size and/or frame rate sequence that undergoes a heavy coarse quantization.
  • the proposed method further, in order to optimize the visual quality at any given bit-rate, provides for operating on such three parameters, frame rate, frame size and quantizer, in a combined and adaptive manner, depending not only on the available bandwidth but also on the picture content.
  • the compressed bit-stream has a variable bit-rate depending on the picture content.
  • different sequences coded at the same bit-rate have different perceived quality, depending on the sequence content. Therefore, a solution using fixed thresholds in terms of bit-rate for shifting frame size and rate is unsuitable and a basic control scheme, such as shown in the following example of pseudo-code, based on assigning fixed thresholds to the bandwidth, would fail to obtain the best possible visual quality given any video content and any channel condition:
  • Threshold 1 and Threshold 2 represent rate threshold that trigger reduction of the frame rate or of the frame size respectively.
  • the proposed method adopts instead a strategy aimed at getting the best possible video quality in every channel condition and for every video content.
  • a strategy is based on the above mentioned fact that the perceived quality is strongly dependent on a Quantization parameter Q of the stream, i.e., the more it is quantized, the less quality is left on the stream.
  • Control of the quantization parameter Q is driven by the transcoder rate control 630 , which selects the appropriate standard quantization parameter Qp, depending on the video content and desired bit-rate.
  • H 1 , H 2 are two real numbers, and Qp and m_quant are quantities in the bit-stream defined in the MPEG- 2 standard that allow bit rate control.
  • the m_quant is the quantizer scale factor, while Qp, as mentioned, is the standard Quantization Parameter.
  • the standard Quantization Parameter Op determines the step size for associating the DCT transformed coefficients with a finite set of steps. Large values of Quantization Parameter Qp represent big steps that roughly approximate the spatial transform, so that most of the signal can be captured by only a few coefficients. Small values of Quantization Parameter Qp more accurately approximate the block's spatial frequency spectrum, but at the cost of more bits.
  • a metric to assess the level of motion content in a sequence and the level of details of the same sequence is also provided in association with the proposed method.
  • a second cost function MOTION_COST is computed starting from entities already available in the incoming bit-stream such as the motion vectors and the number of intra, P and B macroblocks of the incoming bit-stream to assess the level of motion content in the sequence.
  • other typical metrics used in the art like pixels/blocks/macroblock MAE (Mean Absolute Error), MSE (Mean Square Error), activity or variances could be used as well in connection with the proposed method.
  • the second cost function MOTION_COST includes computing an average modulus of the motion vectors average_MV modulus, a number of intra macroblocks number_of_intra_MB, and a number of B macroblocks number_of_birectionally_predicted_MB in the picture, as well as the motion entropy motion_entropy.
  • K 1 , K 2 , K 3 , K 4 represent constants that can assume any real value (positive or negative), including zero.
  • Motion entropy motion_entropy is a measure of the variability of the motion vectors modulus and directions within the image. For example, for each motion vector, the absolute value of the difference in x and y direction of the vector itself with respect to all the neighbors is accumulated. These values are summed for all the vectors in the picture. The higher the results, the more uncorrelated the vectors will be.
  • a third cost function DETAILS_COST is computed that represents a metric for the level of details in the bit-stream; once again, such a third cost function DETAILS_COST can be computed starting from information contained in the incoming bit-stream, such as number of Discrete Cosine Transform coefficients different from zero or above a certain threshold, before or after quantization, before or after motion compensation. All other quantities already used above for the second cost function MOTION_COST can be reused, in a different fashion, for the third cost function DETAILS_COST evaluation.
  • DETAILS_COST J 1 *number_of_hi_freq — DCT _count+ J 2 *number_of intra MB+J 3 *motion_entropy where again J 1 , J 2 , J 3 are constants assuming real values.
  • Number_of_hi_freq_DCT count is defined as the count in an entire frame of the number of hi-frequency Discrete Cosine Transform coefficients that are, after dequantization, higher than a certain threshold TH_DCT.
  • the hi-frequency DCT coefficients are the Discrete Cosine Transform coefficients that come, in a zig-zag-scan order, after the N th position.
  • the threshold TH_DCT and the position N are threshold values arbitrarily chosen.
  • the proposed method makes available two metrics that allow for understanding if the content being processed has a high amount of motion, i.e., the second cost function MOTION_COST, or has many details, i.e., the third cost function DETAILS_COST.
  • the proposed method supplies an adaptive selection rule to choose the operation to perform, decrease or increase bit-rate, that is thus based on the comparison of the three cost functions identified in the foregoing:
  • Such three cost functions are continuously re-computed and compared among each other, in order to establish which is the next action, among the possible actions, i.e., reduce frame rate, reduce frame size, increase quantization, to perform in order to adapt the output bit-rate to the actual channel condition.
  • the action most suitable for decreasing the bandwidth needed is identified as the one involving the smaller cost according to such three cost functions.
  • the proposed method thus provides for comparing such three cost functions.
  • the quantization is increased.
  • the first cost function QUANTIZATION_COST will become greater in value than either the second cost function MOTION_COST or the third cost function DETAILS_COST and this will trigger a change in transcoding level.
  • line T 1 indicates adaptation to different values of bit-rate R by increasing (decreasing) the Quantization Parameter Q and maintaining the full frame rate and full frame size.
  • Line T 1 is represented as a straight line for simplicity's sake, but this does not imply necessarily a linear relationship between Quantization Parameter Q and bit-rate R.
  • the proposed method provides for applying frame rate reduction.
  • Line T 2 indicates the corresponding reduced frame rate, full frame size quantization trend, showing that, in changing from line T 1 to line T 2 , the value of the Quantization Parameter Q necessary for maintaining the same bit rate R drops, since the frame rate is now reduced and thus, as a consequence, the first cost function QUANTIZATION_COST drops.
  • Such a new value assumed by the first cost function QUANTIZATION_COST is registered as a “switch off” point SO for frame rate reduction.
  • the frame rate reduction will be then switched off only when the first cost function QUANTIZATION_COST will be lower or equal than such a value of “switch off” point SO. This conditions applies, if also frame size reduction, that will be now discussed, is already disabled, if it was switched on when the frame size reduction, better illustrated in the following, was on
  • the proposed method After passing from the transcoding level indicated from line T 1 to the transcoding level indicated from line T 2 , to reduce further the bit rate, the proposed method provides again for increasing the quantization, thus the Quantization Parameter Q and the first cost function QUANTIZATION_COST increase again.
  • the first cost function QUANTIZATION_COST will surmount also the third cost function DETAILS_COST value.
  • the proposed method provides for reducing frame size as well.
  • the reduced frame size transcoding level is represented by line T 3
  • line T 4 represents a reduced frame rate reduced frame size transcoding level.
  • the first function QUANTIZATION_COST drops, and the proposed method provides for recording this new value as a second switch off point SO 2 to switch the frame size sub sampling off.
  • third function DETAILS_COST and second function MOTION_COST costs are not compared any more, but attention will be paid only to monitor when the first cost function QUANTIZATION_OST will go below the second switch off point SO 2 .
  • the reduction schemes above illustrated can be recursive, e.g., when a resolution reduction to one fourth is already enabled and the first cost function QUANTIZATION_COST gets higher than the function DETAIL_COST, then a one sixteenth frame size reduction is enabled, and so on.
  • simplified sub cases can envisage for choosing fixed quantization thresholds to switch on and off the sub-sampling, and only the values of the second cost function MOTION_COST and the third function DETAIL_COST are used to select which sub sampling to enable first.
  • the proposed method can additionally envisage letting the “switch off” point SO of the frame rate and the “switch off” point SO 2 of size down-sampling not fixed, but instead variable, because the sequence content can vary over time, e.g., a complex scene can follow a simple one. In this case, an update of the switch off points SO and SO 2 will improve visual quality.
  • the relative position of lines T 2 and T 3 as indicated by the arrows in FIG. 5 , thus can change with respect to the stream content.
  • the quantities with the prefix ‘recorded_’ are the ones sampled when the switch on took place, while the quantities with the prefix ‘current_’ are the latest computed values.
  • M 1 , M 2 , M 3 indicate integer constant values.
  • the proposed method operates without any difference both when the bit-rate must be reduced and when must be increased, adjusting the bit-rate along the lines and the thresholds indicated in FIG. 5 . Therefore, it is not necessary to provide another method to increase the bandwidth.
  • a communication network 700 is shown in FIG. 6 .
  • the network 700 includes a communication link 702 exposed to variable operating conditions; and a control system 704 for controlling delivery of a coded information stream to a user via the communication link 702 .
  • the control system 700 includes a controller module 706 configured for monitoring operating conditions of the link 702 by evaluating a set of cost functions related to the available output bit rate and the complexity of the information stream.
  • the control system 700 also includes a transcoder 708 configured for selectively transcoding the coded information stream by selectively varying a transcoding parameter as a function of said operating conditions monitored.
  • the transcoder 708 is also configured for selectively varying the transcoding parameter by selecting among a plurality of transcoding levels associated with different values of quantization, resolution, and frame rate.
  • the controller module 706 includes a quantization cost analyzer 710 that computes the QUANTIZATION_COST, a motion cost analyzer 712 that computes the MOTION_COST, and a details cost analyzer 714 that computes the DETAILS_COST.
  • the control module 706 also includes a comparator 716 that compares the QUANTIZATION_COST, MOTION_COST, and DETAILS_COST and provides the result of the comparison to a transcoder level modifier 718 .
  • the transcoder level modifier 718 uses the result of the comparison to modify the transcoder level as described above. It will be appreciated that although the controller module 706 and transcoder 708 are shown schematically in FIG. 6 as separate modules, such modules can be incorporated in a single device.
  • the proposed method and circuit can be hardware implemented at least partially through an ASIC, e.g., by a suitable mixing of hardware and software on a embedded processor.

Abstract

A method controls operation of a network wherein at least one coded information stream is delivered to at least one user via at least one link exposed to variable operating conditions. The method includes the steps of: monitoring the operating conditions of the at least one link, and selectively transcoding the at least one coded information stream by selectively varying at least one transcoding parameter as a function of the operating conditions monitored. The monitoring operating conditions of the at least one link includes evaluating a set of cost functions related to an available bit rate and to complexity of said coded information stream and the selective variation of at least one transcoding parameter includes selecting among a plurality of transcoding levels associated to different values of quantization, resolution and frame rate.

Description

    BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The invention relates to coding techniques, for instance for video signals. The invention may be applied to the dynamic adaptation of pre-encoded video streams to an error prone wireless channel, having an available bandwidth changing in a non deterministic way, while preserving the optimal quality of the streamed content. The invention was developed by paying specific attention to the possible application to MPEG-2 coded video streams.
  • However, reference to this specific possible application is not to be construed in a limiting sense of the scope of the invention.
  • 2. Description of the Related Art
  • The MPEG video standard is based on a video compression procedure that exploits the high degree of spatial and temporal correlation in natural video sequences.
  • The MPEG standard defines the syntax and semantics of the transmitted bit-stream and the functionality of the decoder. However, the encoder is not strictly standardized: any encoder that produces a valid MPEG bit-streams is acceptable.
  • Motion estimation is used to compute similarities among successive pictures, in order to remove temporal redundancy, i.e., to transmit only the difference among successive pictures. In particular, Block Matching Motion Estimation (BM-ME) is a common way to extract the similarities among pictures and it is the method selected by the MPEG-2 standard.
  • FIG. 1 shows an example of a hybrid Differential Pulse Code Modulation/Discrete Cosine Transform (DPCM/DCT)) coding loop (MPEG encoder).
  • Such an encoder system removes temporal redundancy from the video Input Sequence, indicated with reference I, which is initially elaborated by a Frame Reorder block 200. Temporal redundancy is then removed in a block 210 that performs an inter-frame Motion Estimation (ME) operation.
  • A summation node 212 operates on the output signal coming from the Motion Estimation block 210, and also receives a signal representative of a previous picture coming from a block 280 that represents the Anchor Frames Buffer.
  • The residual error images, obtained as output of the summation node 212, are further processed by a block 220 in which a Discrete Cosine Transform (DCT) operation is performed, which reduces spatial redundancy by de-correlating the pixels within a block and concentrating the energy of the block itself into a few low order coefficients.
  • Finally, blocks 230 and 240, representing respectively a Scalar Quantization (Quant) operation and a Variable Length Coding (VLC) operation, produce a bit-stream with good statistical compression efficiency.
  • The quantized sequence outgoing from block 230, is also fed to an Inverse Quantization (IQ) block 290, and followed by a cascaded Inverse Discrete Cosine Transform (IDCT) block 295.
  • Due to the intrinsic structure of MPEG standard, the final bit-stream outputted by the Variable Length Coding block 240 is produced at variable bit-rate; hence, it has to be transformed to constant bit-rate by the insertion of a suitable Multiplexer 250 and an Output Buffer 260. Such an Output Buffer 260 feeds a control signal to a Feedback Bit-rate Controller, which decides the granularity of scalar quantization applied in the Scalar Quantization block 230.
  • The Multiplexer 250 works on bit-stream signals coming from the Variable Length Coding block 240 and Motion Estimation block 210.
  • A block 270 represents the already mentioned Feedback Bit-rate Controller and elaborates signals coming from Frame Reorder block 200 and Output Buffer block 260. The output signal from the Feedback Bit-rate Controller 270 is fed to the Quantization block 230.
  • As mentioned, block 280 represents the Anchor Frames Buffer, which receives and elaborates inputs coming from Motion Estimation block 210 and from the summation node 214. The summation node 214 sums signals coming from the Anchor Frame Buffer block 280 and Inverse Discrete Cosine Transform (IDCT) block 295. The signal coming from the Anchor Frame Buffer block 280 is obtained from the motion vectors calculated by the Motion Estimation block 210.
  • Finally, the Output MPEG Bit Stream of this encoder is indicated O in FIG. 1.
  • The coding technique just described must be considered in view of the fact that, in recent times, it has become increasingly important to be able to adapt the multimedia content to the client devices, and it widens increasingly the range of transformations needed.
  • A general access to the multimedia content can be provided in two basic ways; the first is by storing, managing, selecting, and delivering different versions of the media objects (images, video, audio, graphics and text) that include the multimedia presentations.
  • The second is by manipulating the media objects on the fly, by using, for example, methods for text-to-speech translation, image and video transcoding, media conversion, and summarization. This allows for adapting the multimedia content delivery to the wide diversity of client device capabilities in communication, processing storage and display.
  • In both of the basic ways, the need arises of converting a compressed signal into another compressed signal format. The device that performs such an operation is called a transcoder. Such a device could be placed in the network to help relaying transmissions between these different bit rates or could be used as a pre-processing tool to create the various versions of the media objects that were mentioned earlier.
  • By way of example, one may consider the case of a user that desires to watch a Digital Video Disk (DVD) movie, that is MPEG-2 encoded at 8 Mbit/s at standard definition (Main Profile at Main Level), by using a portable wireless device assisted by a Common Intermediate Format (CIF) display.
  • In order to allow viewing of such a movie, a MPEG-2 decoding has to take place, the picture resolution must be changed from standard to Common Intermediate Format, and then again the movie must be MPEG encoded. The resulting bit-stream can be transmitted over a limited bandwidth error-prone channel, received by the portable device and decoded for related display.
  • The main issue is, therefore, to adapt the bit-rate and the picture resolution of a compressed data stream compliant to a video standard (e.g., MPEG-2) to another one. The most widely adopted procedure is to decode the incoming bit-stream, optionally down-sampling the decoded images to generate a sequence with a reduced picture size, and re-encode with a new encoder, suitably set for achieving the required bit-rate.
  • FIG. 2 shows a block diagram of a transcoding chain, where reference numeral 300 indicates a sequence of blocks collectively named MPEG decoder and reference numeral 350 indicates a sequence of blocks collectively named MPEG encoder.
  • As shown in FIG. 2, a video Input Bit-stream I′, representing information in the form of MPEG coded frames, is led to the MPEG decoder 300. Specifically, such video Input Bit-stream I′ is first fed to a Headers Detection block 302. The signal coming from the Headers Detection block 302 is then fed to a Buffer block 304 to be subsequently processed by a Demultiplexer Variable Length Coding Run-level Decoder block indicated with reference 306 in FIG. 2.
  • The processed signal from the block 306 is sent to an Inverse Quantization block 308, with a Quantization level 318 extracted by the same block 306, to be then fed to an Inverse Discrete Cosine Transform block 310. Block 312 represents an Anchor Frame Buffer that operates by exploiting Motion Vector Values 314 and a Motion Vector Modes 316 that receives as input signals, originated from the Decoder block 306.
  • The output signal from the Anchor Frame Buffer 312 is in turn fed to a summation node 320 that receives also the processed signal from the Inverse Discrete Cosine Transform block 310. The output signal from the summation node 320 is then fed to the MPEG encoder 350, and also fed back to the Anchor Frame Buffer 312.
  • The overall scheme of the MPEG encoder block 350 is substantially analogous to the scheme of the encoder 200 shown in FIG. 1, but the signal outputted from the MPEG decoder block 300 is here sent preliminarily to a Chroma Filtering block 360. The filtered signal coming from block 360 is then sent to a Frame Reorder block 370, and a cascaded Motion Estimation block 380. The Motion Vectors 385 originated from such Motion Estimation block 380 are fed to an Anchor Frame Buffer block 390 and to a Multiplexer 430.
  • The signal coming from the Anchor Frame Buffer 390 is fed to two summation nodes, 382 and 392, and fed back to the Motion Estimation block 380.
  • The summation node 382 receives signals from the Frame Reorder block 370 and from the Anchor Frame Buffer block 390. The output signal from the summation node 382 is fed to a Discrete Cosine Transform block 400 to be subsequently quantized in a Quantizer block 410. The quantized signal from the block 410 is fed to a Variable Length Coding block 420 and to an Inverse Quantizer block 450.
  • The signal coming out from Variable Length Coding block 420 is fed to a Multiplexer 430. The multiplexed signal is in turn fed to an Output Buffer 440 that returns the Output MPEG Bit-stream O. The output signal from the block 440 is also fed to a Feedback Bit-rate Controller 470, that controls the Quantizer block 410.
  • The signal generated from the Inverse Quantizer block 450 is fed to an Inverse Discrete Cosine Transform block 460. The inverse transformed signal from block 460 is fed to the summation node 392. The output signal from the summation node 392 is fed to Anchor Frame Buffer block 390.
  • To perform the transcoding procedure above outlined alternative methods have been developed, that are able to operate directly in the Discrete Cosine Transform domain, joining together the decoder and the encoder, and reusing the available useful information, like for example, motion vectors: these systems are able to remove the unnecessary redundancies present in the system.
  • FIG. 3 shows an example of such a transcoder system. The video Input Bit-stream indicated with reference I′ is fed to a Sequence & Group Of Pictures headers block 500 that separates the signal and feeds a Sequence & Group Of Pictures Data delay memory block 510 and a Picture Header block 520.
  • The output signal from block 520 is fed to a Local Cache memory block 530. A Multiplexer 540 operates on signals coming from block 520 and from block 530. The multiplexed signal from block 540 is fed to an Inverse-Variable Length Coding block 560. The signal coming from block 560 is fed to an Inverse-Run Length block 570. The resultant output signal is fed to a Re-quantizer block 585 that is made up with an Inverse Quantizer block 580 followed by a Quantizer block 590. The signal from block 585 is fed to a Run Length block 600 followed by a Variable Length Coding block 610. The output signal form block 610 is fed to a Bit Usage Profile block 620 followed by a Rate Control block 630 that controls the Quantizer block 590.
  • Finally, the Multiplexer 640 elaborates the signals from the block 510, the block 530 and the block 610 and returns the Output Bit-stream O.
  • Conventional systems always provide, as in the transcoder just described with reference to FIG. 3, a de-quantization followed by a re-quantization step, together with an output rate control step.
  • Such a kind of system is able to operate only on a single level of transcoding. In order to operate with more levels, additional complexity should be included in the system.
  • In particular, as described with reference to the encoder of FIG. 2, explicit transcoding is a long and complex sequential chain, and involves various sources of inefficiency, because the motion information coming from the source stream is completely discarded. The transcoder shown in FIG. 3, instead, is able to reuse inasmuch as possible the incoming information, to the expense of losses in terms of output bit-rate precision and drift being introduced.
  • Solutions as depicted in the foregoing are disclosed by published patents WO03047264, WO03047265, EP1374593, EP1231794, and EP1294195.
  • From published patent WO02080518 is known a system for simultaneous transmission of multiple media streams in a fixed bandwidth network, wherein a state machine detects if the network bandwidth is close to saturation and executes a process of selecting at least one stream as a target for lowering total bandwidth usage. In one embodiment, the amount of data is reduced by a gradual degradation of the precision of the data, resulting in a greater potential for data compression, and/or by gradually reducing the resolution of the data of the target stream.
  • BRIEF SUMMARY OF THE INVENTION
  • One embodiment of the present invention provides a method for dynamically adapting pre-encoded video streams to an error prone wireless channel, in which the available bandwidth changes in non-deterministic way, by matching the available bandwidth and preserving at the same time the optimal quality of the streamed content.
  • One embodiment of the present invention is a method having the features set forth in the claims that follow. The invention also relates to a corresponding system, a related network as well as a related computer program product, loadable in the memory of at least one computer and including software code portions for performing the steps of the method of the invention when the product is run on a computer. As used herein, reference to such a computer program product is intended to be equivalent to reference to a computer-readable medium containing instructions for controlling a computer system to coordinate the performance of the method of the invention. Reference to “at least one computer” is evidently intended to highlight the possibility for the present invention to be implemented in a distributed/modular fashion.
  • The arrangement described herein proposes a method that, substantially, provides for operating on a plurality of complexity levels, in particular four, and for choosing the best level among such a plurality of complexity levels by evaluating the available output bit rate and the complexity of the incoming compressed sequence. Such an evaluation is based on the available information contained in the incoming bit-stream, like the motion vector fields and the error level.
  • In particular, a scenario where a home access point is connected with several mobile wireless terminals is considered.
  • Adaptive video transcoding capabilities are becoming more and more important in a broad range of applications in the consumer electronics field; the proposed method be easily applied in such field, enhancing the performances of the transcoding system.
  • The proposed method enhances the quality achievable by a system able to transcode from one (or a plurality on multimedia stream to another one (or a plurality of), introducing very little computational complexity. The arrangement described herein thus allows for deciding among re-quantization, frame rate down-sampling, frame size down-sampling procedures, that are very easy to be implemented, and allows the system to achieve better final quality, avoiding the drawbacks of the prior art.
  • BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
  • The invention will now be described, by way of example only, with reference to the enclosed figures of drawing, wherein:
  • FIG. 1 shows a block diagram of an MPEG encoder system,
  • FIG. 2 shows a block diagram of an explicit transcoding chain,
  • FIG. 3 shows a block diagram of a transcoder,
  • FIG. 4 shows a basic block diagram of a typical Wireless Home Network scenario,
  • FIG. 5 shows a diagram representative of a transcoding adaptation procedure according to the invention, and
  • FIG. 6 shows a schematic diagram of a system for implementing a transcoding adaptation procedure according to the invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • A scenario where a home access point is connected with several mobile wireless terminals is shown in FIG. 4. FIG. 4 shows a home access point 10 connected with a Wide Area Cellular Network 20 and with an Internet network 30. The home access point 10 is connected, by means of a Wireless Bridge 40, with a Gateway 50. The Gateway 50 is in turn connected with several mobile terminals, like: Desktop Computer 60, Printer 65, Monitor device 70, Digital Video Disk (DVD) reader 75, Camcorder 80, Personal Digital Assistant (PDA) 85, Laptop Computer 90, Digital Camera 95, TV device 100, and Audio device 105.
  • The available bandwidth for an 802.11 a/b/gin wireless LAN card is nominally 11 (11b), 54 (11a and 11g), >100(11n) Mbit/s respectively, but, depending on the number of active cards, the relative motions between server and clients and the obstacles between them, the real channel bandwidth could be reduced down to hundreds of Kbits per seconds. For this reason, there is the necessity to find a way to modify dynamically the incoming content to the variable transmission conditions.
  • The transcoder, as seen with reference to FIGS. 2 and 3, is a system able to convert a MPEG high quality compressed video stream into another one, that is less bandwidth consuming, through a dynamic re-quantization performed in the compressed domain. However dynamic re-quantization is not able to achieve compression bit rates lower than few Mbit/sec at full resolution with acceptable quality, because of the limitation of the standard and the intrinsic complexity of the task.
  • In order to overcome such limitations a method of controlling a network is proposed that operates at a plurality of transcoding levels, in particular at four different levels of complexity. The proposed method provides for choosing the best level depending on the available output bit rate and the complexity of the incoming compressed sequence, i.e., basing the choice on the available information contained in the stream, like the motion vector fields and the error level.
  • In a preferred embodiment, the four level of transcoding are the following:
  • full resolution, full frame rate
  • full resolution, reduced frame rate
  • reduced resolution, full frame rate
  • reduced resolution, reduced frame rate
  • The resolution is varied by spatial downsampling, i.e., frame size downsampling, whereas the frame rate is varied by temporal downsampling.
  • Each of such levels can furthermore include sub-levels. Each sublevel is characterized by a different quantization parameter. In other words, the method described herein is able to change the frame size (full or reduced), frame rate (full or reduced) and quantization parameter (the whole MPEG-2 range).
  • It must be noted that methods for transcoding a MPEG-2 stream changing bit rate, operating on quantization level and frame rate, are known per se from the prior art.
  • In particular, when a MPEG-2 stream is considered, such a stream generally contains I, P and B pictures.
  • I pictures are coded with reference to no other picture, so are self-decodable.
  • P pictures are instead coded starting from a Prediction coming from the previous I or P pictures.
  • B pictures are generally coded starting from one previous I or P picture, and one following I or P picture. B picture cannot be taken as reference for other picture.
  • Assuming that the standard sequence of picture type for transmission is IBB PBB PBB PBB, repeated, it can be seen that a frame rate reduction to one third can be obtained by just dropping the bits relating to the B pictures. More complex frame rate reduction processes can be envisaged by people skilled in the art, including partial decoding of the pictures.
  • The method proposed, however, deals with procedures for using the above mentioned transcoding processes to dynamically adapt the stream to instantaneous channel bandwidth variations.
  • In the European patent application n. 04007275.3 in the name of the STMicroelectronics SrI, the assignee of the present invention, it is disclosed a method for measuring in real time the network bandwidth variations, so it can be possible to have an exact estimation of the channel conditions. The results about the channel conditions obtained by such a method can be used as input values for selecting the use of the transcoding levels and the extent of their use.
  • The proposed method, as mentioned above, envisages down-sampling operations for reducing both the frame rate and the frame size. This arises from the observation that, if the quantizer values are increased in order to decrease the bit-rate or bandwidth, above a certain value of compression factor/quantizer, it is better, in terms of perceived quality by the user, to watch a sequence that is reduced in frame size and/or frame rate, but less quantized, rather than a full size and/or frame rate sequence that undergoes a heavy coarse quantization.
  • The proposed method, further, in order to optimize the visual quality at any given bit-rate, provides for operating on such three parameters, frame rate, frame size and quantizer, in a combined and adaptive manner, depending not only on the available bandwidth but also on the picture content.
  • At a defined quantization level, the compressed bit-stream has a variable bit-rate depending on the picture content. Conversely, different sequences coded at the same bit-rate have different perceived quality, depending on the sequence content. Therefore, a solution using fixed thresholds in terms of bit-rate for shifting frame size and rate is unsuitable and a basic control scheme, such as shown in the following example of pseudo-code, based on assigning fixed thresholds to the bandwidth, would fail to obtain the best possible visual quality given any video content and any channel condition:
  • IF(rate>Threshold 1)
  • THEN full_rate
  • ELSE reduce_rate;
  • IF (rate>Threshold 2)
  • THEN full_size
  • ELSE reduce_size;
  • where Threshold 1 and Threshold 2 represent rate threshold that trigger reduction of the frame rate or of the frame size respectively.
  • The proposed method adopts instead a strategy aimed at getting the best possible video quality in every channel condition and for every video content. Such a strategy is based on the above mentioned fact that the perceived quality is strongly dependent on a Quantization parameter Q of the stream, i.e., the more it is quantized, the less quality is left on the stream.
  • Control of the quantization parameter Q, as seen for instance with reference to FIG. 3, is driven by the transcoder rate control 630, which selects the appropriate standard quantization parameter Qp, depending on the video content and desired bit-rate.
  • The proposed method thus provides for making available a first cost function QUANTIZATION_COST, that is a function that take into account how the image has been quantized, according to the following relation:
    QUANTIZATION_COST= H 1*Qp+H 2 *m_quant
  • H1, H2 are two real numbers, and Qp and m_quant are quantities in the bit-stream defined in the MPEG-2 standard that allow bit rate control. The m_quant is the quantizer scale factor, while Qp, as mentioned, is the standard Quantization Parameter. The standard Quantization Parameter Op determines the step size for associating the DCT transformed coefficients with a finite set of steps. Large values of Quantization Parameter Qp represent big steps that roughly approximate the spatial transform, so that most of the signal can be captured by only a few coefficients. Small values of Quantization Parameter Qp more accurately approximate the block's spatial frequency spectrum, but at the cost of more bits.
  • A metric to assess the level of motion content in a sequence and the level of details of the same sequence is also provided in association with the proposed method. Such a metric, a second cost function MOTION_COST, is computed starting from entities already available in the incoming bit-stream such as the motion vectors and the number of intra, P and B macroblocks of the incoming bit-stream to assess the level of motion content in the sequence. In addition, other typical metrics used in the art, like pixels/blocks/macroblock MAE (Mean Absolute Error), MSE (Mean Square Error), activity or variances could be used as well in connection with the proposed method.
  • It is underlined that for what concerns these metrics, and also other metrics mentioned in the foregoing and in the following, a person skilled in the art could supply different definitions, without departing substantially from the scope of the invention.
  • Other variables that can be introduced in the metrics could be, by way of example, the size of the motion compensated block, in case the video coding standard (like H.264) allows selecting different motion compensation bases (16×16 pixels down to 4×4 pixels).
  • In the present embodiment the second cost function MOTION_COST includes computing an average modulus of the motion vectors average_MV modulus, a number of intra macroblocks number_of_intra_MB, and a number of B macroblocks number_of_birectionally_predicted_MB in the picture, as well as the motion entropy motion_entropy. Such values are combined as follows:
    MOTION_COST= K 1*average MV_modulus+K 2*number_of_intra MB+K 3*number_of_birectionally_predicted MB+K 4* motion_entropy
    where K1, K2, K3, K4 represent constants that can assume any real value (positive or negative), including zero. Many variation are possible also on the second cost function MOTION_COST for the person skilled in the art, for instance by including other parameters extracted from the incoming bit-stream(e.g., DCT coefficient values and numbers). Such a second cost function MOTION_COST, working on already available compressed information, is computationally very inexpensive, therefore it is well suited for real-time implementation. Motion entropy motion_entropy is a measure of the variability of the motion vectors modulus and directions within the image. For example, for each motion vector, the absolute value of the difference in x and y direction of the vector itself with respect to all the neighbors is accumulated. These values are summed for all the vectors in the picture. The higher the results, the more uncorrelated the vectors will be.
  • Further, according to the proposed method, a third cost function DETAILS_COST is computed that represents a metric for the level of details in the bit-stream; once again, such a third cost function DETAILS_COST can be computed starting from information contained in the incoming bit-stream, such as number of Discrete Cosine Transform coefficients different from zero or above a certain threshold, before or after quantization, before or after motion compensation. All other quantities already used above for the second cost function MOTION_COST can be reused, in a different fashion, for the third cost function DETAILS_COST evaluation.
  • In the present embodiment such a third cost function DETAILS_COST is:
    DETAILS_COST= J 1*number_of_hi_freq DCT_count+J 2*number_of intra MB+J 3*motion_entropy
    where again J1, J2, J3 are constants assuming real values.
  • Number_of_hi_freq_DCT count is defined as the count in an entire frame of the number of hi-frequency Discrete Cosine Transform coefficients that are, after dequantization, higher than a certain threshold TH_DCT. The hi-frequency DCT coefficients are the Discrete Cosine Transform coefficients that come, in a zig-zag-scan order, after the Nth position. The threshold TH_DCT and the position N are threshold values arbitrarily chosen.
  • In addition, as above, well-known quantities in the art can be used, such as prediction error MAE, macroblock variance or activity to be included in the above metrics.
  • Therefore, the proposed method makes available two metrics that allow for understanding if the content being processed has a high amount of motion, i.e., the second cost function MOTION_COST, or has many details, i.e., the third cost function DETAILS_COST.
  • Although it is known to provide at the receiver side, after the decoding operation and before displaying the video, means and methods to up-sample both the frame size and the frame rate, therefore reversing the action of downsampling frame size and frame rate, such means and methods however cannot recreate information lost during the down-sampling, thus the up-sampling will produce acceptable results only if in the preceding down-sample procedure it was avoided in some way to discard a too large amount of information. Typical examples of cases where the up-sampling is not effective are situations where reduced frame rate in a sequence with a great amount of motion is considered, or reduced frame size in a sequence with many details is considered.
  • In case no up-sampling is available in the post-processing at the display, these impairments will be even more evident. Therefore, it is crucial to correctly understand when to apply down-sampling, and of which type, spatial or temporal. In a sequence with high detail but slow motion, it is correct to reduce the frame rate; in a sequence where a great amount of chaotic motion is present, but little details (e.g., big objects bouncing around) is correct to safely down-sample image size.
  • The proposed method supplies an adaptive selection rule to choose the operation to perform, decrease or increase bit-rate, that is thus based on the comparison of the three cost functions identified in the foregoing:
  • MOTION_COST,
  • DETAILS_COST,
  • QUANTIZATION_COST.
  • Such three cost functions are continuously re-computed and compared among each other, in order to establish which is the next action, among the possible actions, i.e., reduce frame rate, reduce frame size, increase quantization, to perform in order to adapt the output bit-rate to the actual channel condition. The action most suitable for decreasing the bandwidth needed is identified as the one involving the smaller cost according to such three cost functions.
  • The proposed method thus provides for comparing such three cost functions.
  • Assuming that the smaller cost function initially is the first cost function QUANTIZATION_COST, the quantization is increased. By keeping on quantizing, at some point of such a quantization increase, however, the first cost function QUANTIZATION_COST will become greater in value than either the second cost function MOTION_COST or the third cost function DETAILS_COST and this will trigger a change in transcoding level.
  • More specifically, as shown in the diagram of FIG. 5, where the Quantization Parameter Q is plotted against the bit-rate R, line T1 indicates adaptation to different values of bit-rate R by increasing (decreasing) the Quantization Parameter Q and maintaining the full frame rate and full frame size. Line T1 is represented as a straight line for simplicity's sake, but this does not imply necessarily a linear relationship between Quantization Parameter Q and bit-rate R.
  • As the Quantization Parameter Q reaches a value indicated by the horizontal line HL, where the second cost function MOTION_COST becomes smaller than the first cost function QUANTIZATION_COST, the proposed method provides for applying frame rate reduction. Line T2 indicates the corresponding reduced frame rate, full frame size quantization trend, showing that, in changing from line T1 to line T2, the value of the Quantization Parameter Q necessary for maintaining the same bit rate R drops, since the frame rate is now reduced and thus, as a consequence, the first cost function QUANTIZATION_COST drops. Such a new value assumed by the first cost function QUANTIZATION_COST, is registered as a “switch off” point SO for frame rate reduction.
  • In the case that the output bit-rate can be increased again because of improved bandwidth conditions on the network, the frame rate reduction will be then switched off only when the first cost function QUANTIZATION_COST will be lower or equal than such a value of “switch off” point SO. This conditions applies, if also frame size reduction, that will be now discussed, is already disabled, if it was switched on when the frame size reduction, better illustrated in the following, was on
  • After passing from the transcoding level indicated from line T1 to the transcoding level indicated from line T2, to reduce further the bit rate, the proposed method provides again for increasing the quantization, thus the Quantization Parameter Q and the first cost function QUANTIZATION_COST increase again.
  • During such an increment, the first cost function QUANTIZATION_COST will surmount also the third cost function DETAILS_COST value. In this case the proposed method provides for reducing frame size as well. The reduced frame size transcoding level is represented by line T3, while line T4 represents a reduced frame rate reduced frame size transcoding level.
  • Once again, after such an action is taken, the first function QUANTIZATION_COST drops, and the proposed method provides for recording this new value as a second switch off point SO2 to switch the frame size sub sampling off. In this case, third function DETAILS_COST and second function MOTION_COST costs are not compared any more, but attention will be paid only to monitor when the first cost function QUANTIZATION_OST will go below the second switch off point SO2.
  • The reduction schemes above illustrated can be recursive, e.g., when a resolution reduction to one fourth is already enabled and the first cost function QUANTIZATION_COST gets higher than the function DETAIL_COST, then a one sixteenth frame size reduction is enabled, and so on.
  • In other possible embodiments, simplified sub cases can envisage for choosing fixed quantization thresholds to switch on and off the sub-sampling, and only the values of the second cost function MOTION_COST and the third function DETAIL_COST are used to select which sub sampling to enable first.
  • The proposed method can additionally envisage letting the “switch off” point SO of the frame rate and the “switch off” point SO2 of size down-sampling not fixed, but instead variable, because the sequence content can vary over time, e.g., a complex scene can follow a simple one. In this case, an update of the switch off points SO and SO2 will improve visual quality. The relative position of lines T2 and T3, as indicated by the arrows in FIG. 5, thus can change with respect to the stream content.
  • Such an update can easily be obtained by linear combinations of the “fixed” value of the switch off point SO and the variations of the functions MOTION_COST and DETAILS_COST with regards to their value when the quality switch off point was taken. An expression for the quality switch off point QUALITY_COST switch_off_point SO can be:
    Current_QUALITY_COST_switch_off_point SO=
    M1*recorded_QUALITY_COST_switch_off_point+M 2*(current_VECTOR_COST−recorded_VECTOR_COST)+M 3*(current_DETAILS_COST−recorded_DETAILS_COST)
  • The quantities with the prefix ‘recorded_’ are the ones sampled when the switch on took place, while the quantities with the prefix ‘current_’ are the latest computed values. M1, M2, M3 indicate integer constant values.
  • It must be noted that the proposed method operates without any difference both when the bit-rate must be reduced and when must be increased, adjusting the bit-rate along the lines and the thresholds indicated in FIG. 5. Therefore, it is not necessary to provide another method to increase the bandwidth.
  • A communication network 700 according to one embodiment of the invention is shown in FIG. 6. The network 700 includes a communication link 702 exposed to variable operating conditions; and a control system 704 for controlling delivery of a coded information stream to a user via the communication link 702. The control system 700 includes a controller module 706 configured for monitoring operating conditions of the link 702 by evaluating a set of cost functions related to the available output bit rate and the complexity of the information stream. The control system 700 also includes a transcoder 708 configured for selectively transcoding the coded information stream by selectively varying a transcoding parameter as a function of said operating conditions monitored. The transcoder 708 is also configured for selectively varying the transcoding parameter by selecting among a plurality of transcoding levels associated with different values of quantization, resolution, and frame rate.
  • The controller module 706 includes a quantization cost analyzer 710 that computes the QUANTIZATION_COST, a motion cost analyzer 712 that computes the MOTION_COST, and a details cost analyzer 714 that computes the DETAILS_COST. The control module 706 also includes a comparator 716 that compares the QUANTIZATION_COST, MOTION_COST, and DETAILS_COST and provides the result of the comparison to a transcoder level modifier 718. The transcoder level modifier 718 uses the result of the comparison to modify the transcoder level as described above. It will be appreciated that although the controller module 706 and transcoder 708 are shown schematically in FIG. 6 as separate modules, such modules can be incorporated in a single device.
  • The proposed method and circuit can be hardware implemented at least partially through an ASIC, e.g., by a suitable mixing of hardware and software on a embedded processor.
  • Consequently, without prejudice to the underlying principles of the invention, the details and the embodiments may vary, also appreciably, with reference to what has been described by way of example only, without departing from the scope of the invention as defined by the annexed claims.
  • All of the above U.S. patents, U.S. patent application publications, U.S. patent applications, foreign patents, foreign patent applications and non-patent publications referred to in this specification and/or listed in the Application Data Sheet are incorporated herein by reference, in their entirety.

Claims (39)

1. A method of controlling delivery of a coded information stream to a user via a link exposed to variable operating conditions, the method comprising the steps of:
monitoring operating conditions of said link, said monitoring operating conditions of said link includes evaluating a set of cost functions related to an available bit rate and to a complexity of said coded information stream; and
selectively transcoding said coded information stream by selectively varying a transcoding parameter as a function of said operating conditions monitored, wherein
said selectively varying the transcoding parameter includes selecting among a plurality of transcoding levels associated with different values of quantization, resolution, and frame rate.
2. The method of claim 1, wherein said evaluating a set of cost functions includes comparing a first cost function of said set of cost functions against the other functions of said set of cost functions and select among a plurality of transcoding levels based on a result of said comparing.
3. The method of claim 2, wherein said set of cost functions includes a quantization cost function.
4. The method of claim 3, wherein said set of cost functions includes a motion cost function.
5. The method of claim 2 wherein said first cost function is a quantization cost function.
6. The method of claim 2 wherein said transcoding levels(T1, T2, T3, T4) include sublevels.
7. The method of claim 6, wherein said sublevels are associated to different quantization parameters.
8. The method of claim 2, wherein comparing includes storing one or more switch-off values of quantization representative of a switching among said different values of resolution and frame rate.
9. The method of claim 8, wherein said one or more switch-off values include a first switch-off value for the frame rate and a second switch-off value for the resolution and further includes updating said first switch-off value and second switch-off value as a function of the complexity of said coded information stream.
10. The method of claim 1 wherein said set of cost functions includes a details cost function.
11. The method of claim 1 wherein said transcoding levels include at least four levels that include:
full resolution, full frame rate,
full resolution, reduced frame rate,
reduced resolution, full frame rate, and
reduced resolution, reduced frame rate.
12. The method of claim 1 wherein said link is a wireless link.
13. The method of claim 1 wherein said link is included in a local area network.
14. The method of claim 1 wherein said at least one link is included in a wireless LAN.
15. The method of claim 1 wherein said information stream is a media stream.
16. The method of claim 1 wherein said stream is selected out of an audio stream, a video stream and an audio/video stream.
17. The method of claim 16, wherein said stream is an MPEG encoded stream.
18. A system for controlling delivery of a coded information stream to a user via a link exposed to variable operating conditions, the system comprising:
a controller module configured for monitoring operating conditions of said link, by evaluating a set of cost functions related to the available output bit rate and the complexity of the information stream; and
a transcoder configured for selectively transcoding said coded information stream by selectively varying a transcoding parameter as a function of said operating conditions monitored, wherein
said at least one transcoder is configured for selectively varying the transcoding parameter by selecting among a plurality of transcoding levels associated with different values of quantization, resolution, and frame rate.
19. The system of claim 18, wherein the controller module is configured for evaluating a set of cost functions by comparing a first cost function of said set of cost functions against the other functions of said set of cost functions and said transcoder is configured for selecting among a plurality of transcoding levels based on a result of said comparison.
20. The system of claim 19, wherein said set of cost functions includes a quantization cost function.
21. The system of claim 19wherein said set of cost functions includes a motion cost function.
22. The method of 19 wherein said first cost function is a quantization cost function.
23. The system of claim 19, wherein said transcoder is configured for storing one or more switch-off values of quantization representative of a switching among said different values of resolution and frame rate.
24. The method of claim 23, wherein said one or more switch-off values include a first switch-off value for the frame rate and a second switch-off value for the resolution and transcoder is configured further for updating said first switch-off value and second switch-off value as a function of the complexity of said coded information stream.
25. The system of claim 18 wherein said set of cost functions includes a details cost function.
26. The system of claim 18 wherein said transcoding levels include sublevels.
27. The system of claim 26 wherein said sublevels are associated with different quantization parameters.
28. The system of claim 18 wherein said transcoding levels include at least four levels that include:
full resolution, full frame rate,
full resolution, reduced frame rate,
reduced resolution, full frame rate, and
reduced resolution, reduced frame rate.
29. The system of claim 18 wherein said link is a wireless link.
30. The system of claim 18 wherein said link is included in a local area network.
31. The system of claim 30, wherein said link is included in a wireless LAN.
32. The system of claim 18 wherein said coded information stream is a media stream.
33. The system of claim 18 wherein said coded information stream is selected out of an audio stream, a video stream and an audio/video stream.
34. The system of claim 33, wherein said stream is an MPEG or H.264 encoded stream.
35. A communication network, comprising:
a communication link exposed to variable operating conditions; and
a system for controlling delivery of a coded information stream to a user via the communication link, the system including:
a controller module configured for monitoring operating conditions of said link, by evaluating a set of cost functions related to the available output bit rate and the complexity of the information stream; and
a transcoder configured for selectively transcoding said coded information stream by selectively varying a transcoding parameter as a function of said operating conditions monitored, wherein said at least one transcoder is configured for selectively varying the transcoding parameter by selecting among a plurality of transcoding levels associated with different values of quantization, resolution, and frame rate.
36. The network of claim 35, wherein the controller module is configured for evaluating a set of cost functions by comparing a first cost function of said set of cost functions against the other functions of said set of cost functions and said transcoder is configured for selecting among a plurality of transcoding levels based on a result of said comparison.
37. A computer-readable medium whose contents cause a computing device to control delivery of a coded information stream to a user via a link exposed to variable operating conditions, by performing a method comprising:
monitoring operating conditions of said link, said monitoring operating conditions of said link includes evaluating a set of cost functions related to an available bit rate and to a complexity of said coded information stream; and
selectively transcoding said coded information stream by selectively varying a transcoding parameter as a function of said operating conditions monitored, wherein said selectively varying the transcoding parameter includes selecting among a plurality of transcoding levels associated with different values of quantization, resolution, and frame rate.
38. The computer-readable medium of claim 37 wherein the computer-readable medium is a memory of a computing device.
39. The computer-readable medium of claim 37 wherein the computer-readable medium is a data transmission medium transmitting a generated data signal containing the contents.
US11/179,012 2004-07-09 2005-07-11 Method and system for delivery of coded information streams, related network and computer program product therefor Abandoned US20060018378A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP04016172.1 2004-07-09
EP04016172.1A EP1615447B1 (en) 2004-07-09 2004-07-09 Method and system for delivery of coded information streams, related network and computer program product therefor

Publications (1)

Publication Number Publication Date
US20060018378A1 true US20060018378A1 (en) 2006-01-26

Family

ID=34925679

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/179,012 Abandoned US20060018378A1 (en) 2004-07-09 2005-07-11 Method and system for delivery of coded information streams, related network and computer program product therefor

Country Status (2)

Country Link
US (1) US20060018378A1 (en)
EP (1) EP1615447B1 (en)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070172211A1 (en) * 2006-01-26 2007-07-26 Prasanjit Panda Adaptive filtering to enhance video bit-rate control performance
US20070171988A1 (en) * 2006-01-26 2007-07-26 Prasanjit Panda Adaptive filtering to enhance video encoder performance
US20080084927A1 (en) * 2006-09-18 2008-04-10 Elemental Technologies, Inc. Real-time network adaptive digital video encoding/decoding
US20080104520A1 (en) * 2006-11-01 2008-05-01 Swenson Erik R Stateful browsing
US20080104652A1 (en) * 2006-11-01 2008-05-01 Swenson Erik R Architecture for delivery of video content responsive to remote interaction
US20080101466A1 (en) * 2006-11-01 2008-05-01 Swenson Erik R Network-Based Dynamic Encoding
US20080184128A1 (en) * 2007-01-25 2008-07-31 Swenson Erik R Mobile device user interface for remote interaction
US20090125538A1 (en) * 2007-11-13 2009-05-14 Elemental Technologies, Inc. Video encoding and decoding using parallel processors
US20100074325A1 (en) * 2008-09-22 2010-03-25 Avot Media, Inc. Video streaming apparatus with quantization and method thereof
US20100074327A1 (en) * 2008-09-22 2010-03-25 Avot Media, Inc. Video quantizer unit and method thereof
US20100074326A1 (en) * 2008-09-22 2010-03-25 Avot Media, Inc. Transcoder unit and method
US20100111497A1 (en) * 2006-09-06 2010-05-06 Zhikai Song Media playing tool with a multiple media playing model
WO2010034020A3 (en) * 2008-09-22 2010-07-08 Avot Media, Inc. Video streaming apparatus with quantization and method thereof
US20100299453A1 (en) * 2009-05-19 2010-11-25 Fox Brian J System and method for dynamically transcoding data requests
US20110080946A1 (en) * 2009-10-06 2011-04-07 Junlin Li Locally variable quantization and hybrid variable length coding for image and video compression
US20120027295A1 (en) * 2009-04-14 2012-02-02 Koninklijke Philips Electronics N.V. Key frames extraction for video content analysis
US20120120043A1 (en) * 2010-11-16 2012-05-17 Samsung Electronics Co., Ltd. Image data compressing and decompressing methods and display driving device using the same
US20120124633A1 (en) * 2010-11-15 2012-05-17 International Business Machines Corporation Wireless Video Streaming Quality Management for Bandwidth Constrained Cellular Network
US8184715B1 (en) 2007-08-09 2012-05-22 Elemental Technologies, Inc. Method for efficiently executing video encoding operations on stream processor architectures
US20130094564A1 (en) * 2011-10-17 2013-04-18 Google Inc. Rate-distortion-complexity optimization of video encoding
US20130265401A1 (en) * 2012-03-30 2013-10-10 Olympus Medical Systems Corp. Endoscope apparatus
US20130301727A1 (en) * 2012-05-14 2013-11-14 Qualcomm Incorporated Programmable and scalable integer search for video encoding
US20140244805A1 (en) * 2013-02-26 2014-08-28 Samsung Electronics Co., Ltd. Bandwidth estimation-based streaming content processing method and apparatus of terminal
US8885702B2 (en) 2011-10-17 2014-11-11 Google Inc. Rate-distortion-complexity optimization of video encoding guided by video description length
WO2015022685A1 (en) * 2013-08-13 2015-02-19 I.C.V.T Ltd. Quality driven video re-encoding
US9247260B1 (en) 2006-11-01 2016-01-26 Opera Software Ireland Limited Hybrid bitmap-mode encoding
WO2016077712A1 (en) * 2014-11-14 2016-05-19 Videopura, Llc System and method for adaptive video streaming with quality equivalent segmentation and delivery
US9460729B2 (en) 2012-09-21 2016-10-04 Dolby Laboratories Licensing Corporation Layered approach to spatial audio coding
US9554083B2 (en) 2013-08-12 2017-01-24 Samsung Electronics Co., Ltd. Methods for selecting resolution with minimum distortion value and devices performing the methods
US9621951B2 (en) 2008-06-25 2017-04-11 Dominic M. Kotab Methods for receiving and sending video to a handheld device
US9936208B1 (en) * 2015-06-23 2018-04-03 Amazon Technologies, Inc. Adaptive power and quality control for video encoders on mobile devices
US10075768B1 (en) * 2008-01-30 2018-09-11 Dominic M. Kotab Systems and methods for creating and storing reduced quality video data

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8780957B2 (en) 2005-01-14 2014-07-15 Qualcomm Incorporated Optimal weights for MMSE space-time equalizer of multicode CDMA system
CL2006000541A1 (en) * 2005-03-10 2008-01-04 Qualcomm Inc Method for processing multimedia data comprising: a) determining the complexity of multimedia data; b) classify multimedia data based on the complexity determined; and associated apparatus.
EP1921866A3 (en) * 2005-03-10 2010-07-28 QUALCOMM Incorporated Content adaptive multimedia processing
US8879856B2 (en) 2005-09-27 2014-11-04 Qualcomm Incorporated Content driven transcoder that orchestrates multimedia transcoding using content information
US8654848B2 (en) 2005-10-17 2014-02-18 Qualcomm Incorporated Method and apparatus for shot detection in video streaming
US8948260B2 (en) 2005-10-17 2015-02-03 Qualcomm Incorporated Adaptive GOP structure in video streaming
US9131164B2 (en) 2006-04-04 2015-09-08 Qualcomm Incorporated Preprocessor method and apparatus
FR2919779B1 (en) * 2007-08-02 2010-02-26 Canon Kk METHOD AND DEVICE FOR ENCODING LOSS OF A DIGITAL SIGNAL
US8270739B2 (en) 2007-12-03 2012-09-18 Ecole De Technologie Superieure System and method for quality-aware selection of parameters in transcoding of digital images
WO2009055896A1 (en) 2007-11-02 2009-05-07 Vantrix Corporation System and method for predicting the file size of images subject to transformation by scaling and a change of quality-controlling parameters
US8300961B2 (en) 2008-12-12 2012-10-30 Ecole De Technologie Superieure Method and system for low complexity transcoding of images with near optimal quality
US8836857B2 (en) * 2010-11-17 2014-09-16 Motorola Mobility Llc System and method for selectively transcoding signal from one format to one of plurality of formats
US11284133B2 (en) 2012-07-10 2022-03-22 Avago Technologies International Sales Pte. Limited Real-time video coding system of multiple temporally scaled video and of multiple profile and standards based on shared video coding information
US9426498B2 (en) 2012-07-10 2016-08-23 Broadcom Corporation Real-time encoding system of multiple spatially scaled video based on shared video coding information
US9338450B2 (en) 2013-03-18 2016-05-10 Ecole De Technologie Superieure Method and apparatus for signal encoding producing encoded signals of high fidelity at minimal sizes
US10609405B2 (en) 2013-03-18 2020-03-31 Ecole De Technologie Superieure Optimal signal encoding based on experimental data
US9661331B2 (en) 2013-03-18 2017-05-23 Vantrix Corporation Method and apparatus for signal encoding realizing optimal fidelity
FR3006539B1 (en) * 2013-05-31 2016-12-09 Aviwest METHOD OF TRANSFERRING AT LEAST TWO AUDIOVISUAL DATA FLOWS.

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020019984A1 (en) * 2000-01-14 2002-02-14 Rakib Selim Shlomo Headend cherrypicker with digital video recording capability
US20030227977A1 (en) * 2002-05-29 2003-12-11 Canon Kabushiki Kaisha Method and device for selecting a transcoding method from a set of transcoding methods
US20050213502A1 (en) * 2004-03-26 2005-09-29 Stmicroelectronics S.R.I. Method and system for controlling operation of a network, such as a WLAN, related network and computer program product therefor

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020019984A1 (en) * 2000-01-14 2002-02-14 Rakib Selim Shlomo Headend cherrypicker with digital video recording capability
US20030227977A1 (en) * 2002-05-29 2003-12-11 Canon Kabushiki Kaisha Method and device for selecting a transcoding method from a set of transcoding methods
US20050213502A1 (en) * 2004-03-26 2005-09-29 Stmicroelectronics S.R.I. Method and system for controlling operation of a network, such as a WLAN, related network and computer program product therefor

Cited By (62)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070171988A1 (en) * 2006-01-26 2007-07-26 Prasanjit Panda Adaptive filtering to enhance video encoder performance
US8009963B2 (en) 2006-01-26 2011-08-30 Qualcomm Incorporated Adaptive filtering to enhance video bit-rate control performance
US7903733B2 (en) * 2006-01-26 2011-03-08 Qualcomm Incorporated Adaptive filtering to enhance video encoder performance
US20070172211A1 (en) * 2006-01-26 2007-07-26 Prasanjit Panda Adaptive filtering to enhance video bit-rate control performance
US20100111497A1 (en) * 2006-09-06 2010-05-06 Zhikai Song Media playing tool with a multiple media playing model
US20080084927A1 (en) * 2006-09-18 2008-04-10 Elemental Technologies, Inc. Real-time network adaptive digital video encoding/decoding
US8250618B2 (en) * 2006-09-18 2012-08-21 Elemental Technologies, Inc. Real-time network adaptive digital video encoding/decoding
US20080104652A1 (en) * 2006-11-01 2008-05-01 Swenson Erik R Architecture for delivery of video content responsive to remote interaction
US8711929B2 (en) * 2006-11-01 2014-04-29 Skyfire Labs, Inc. Network-based dynamic encoding
US8443398B2 (en) 2006-11-01 2013-05-14 Skyfire Labs, Inc. Architecture for delivery of video content responsive to remote interaction
US9247260B1 (en) 2006-11-01 2016-01-26 Opera Software Ireland Limited Hybrid bitmap-mode encoding
US8375304B2 (en) 2006-11-01 2013-02-12 Skyfire Labs, Inc. Maintaining state of a web page
US20080101466A1 (en) * 2006-11-01 2008-05-01 Swenson Erik R Network-Based Dynamic Encoding
US20080104520A1 (en) * 2006-11-01 2008-05-01 Swenson Erik R Stateful browsing
US20080181498A1 (en) * 2007-01-25 2008-07-31 Swenson Erik R Dynamic client-server video tiling streaming
US20080184128A1 (en) * 2007-01-25 2008-07-31 Swenson Erik R Mobile device user interface for remote interaction
US8630512B2 (en) 2007-01-25 2014-01-14 Skyfire Labs, Inc. Dynamic client-server video tiling streaming
US8437407B2 (en) 2007-08-09 2013-05-07 Elemental Technologies, Inc. Method for efficiently executing video encoding operations on stream processor architectures
US8184715B1 (en) 2007-08-09 2012-05-22 Elemental Technologies, Inc. Method for efficiently executing video encoding operations on stream processor architectures
US10678747B2 (en) 2007-11-13 2020-06-09 Amazon Technologies, Inc. Video encoding and decoding using parallel processors
US9747251B2 (en) 2007-11-13 2017-08-29 Amazon Technologies, Inc. Video encoding and decoding using parallel processors
US8121197B2 (en) 2007-11-13 2012-02-21 Elemental Technologies, Inc. Video encoding and decoding using parallel processors
US20090125538A1 (en) * 2007-11-13 2009-05-14 Elemental Technologies, Inc. Video encoding and decoding using parallel processors
US10075768B1 (en) * 2008-01-30 2018-09-11 Dominic M. Kotab Systems and methods for creating and storing reduced quality video data
US9621951B2 (en) 2008-06-25 2017-04-11 Dominic M. Kotab Methods for receiving and sending video to a handheld device
US20100074326A1 (en) * 2008-09-22 2010-03-25 Avot Media, Inc. Transcoder unit and method
US8295345B2 (en) 2008-09-22 2012-10-23 Smith Micro Software, Inc. Transcoder unit and method
US8345746B2 (en) 2008-09-22 2013-01-01 Smith Micro Software, Inc. Video quantizer unit and method thereof
WO2010034020A3 (en) * 2008-09-22 2010-07-08 Avot Media, Inc. Video streaming apparatus with quantization and method thereof
US8279925B2 (en) 2008-09-22 2012-10-02 Smith Micro Software, Inc. Video streaming apparatus with quantization and method thereof
US20100074327A1 (en) * 2008-09-22 2010-03-25 Avot Media, Inc. Video quantizer unit and method thereof
US20100074325A1 (en) * 2008-09-22 2010-03-25 Avot Media, Inc. Video streaming apparatus with quantization and method thereof
US20120027295A1 (en) * 2009-04-14 2012-02-02 Koninklijke Philips Electronics N.V. Key frames extraction for video content analysis
US20100299453A1 (en) * 2009-05-19 2010-11-25 Fox Brian J System and method for dynamically transcoding data requests
US9131007B2 (en) * 2009-05-19 2015-09-08 Vitrual World Computing, Inc. System and method for dynamically transcoding data requests
US20110080946A1 (en) * 2009-10-06 2011-04-07 Junlin Li Locally variable quantization and hybrid variable length coding for image and video compression
US8340448B2 (en) * 2009-10-06 2012-12-25 Cisco Technology, Inc. Locally variable quantization and hybrid variable length coding for image and video compression
US20120124633A1 (en) * 2010-11-15 2012-05-17 International Business Machines Corporation Wireless Video Streaming Quality Management for Bandwidth Constrained Cellular Network
US9071838B2 (en) * 2010-11-16 2015-06-30 Samsung Electronics Co., Ltd. Image data compressing and decompressing methods and display driving device using the same
US20120120043A1 (en) * 2010-11-16 2012-05-17 Samsung Electronics Co., Ltd. Image data compressing and decompressing methods and display driving device using the same
US8934538B2 (en) * 2011-10-17 2015-01-13 Google Inc. Rate-distortion-complexity optimization of video encoding
EP2769552A4 (en) * 2011-10-17 2015-06-17 Google Inc Rate-distortion-complexity optimization of video encoding guided by video description length
US8885702B2 (en) 2011-10-17 2014-11-11 Google Inc. Rate-distortion-complexity optimization of video encoding guided by video description length
US9225983B2 (en) 2011-10-17 2015-12-29 Google Inc. Rate-distortion-complexity optimization of video encoding guided by video description length
US20130094564A1 (en) * 2011-10-17 2013-04-18 Google Inc. Rate-distortion-complexity optimization of video encoding
CN103975601A (en) * 2011-10-17 2014-08-06 谷歌公司 Rate-distortion-complexity optimization of video encoding
CN103619233A (en) * 2012-03-30 2014-03-05 奥林巴斯医疗株式会社 Endoscopic device
US8885032B2 (en) * 2012-03-30 2014-11-11 Olympus Medical Systems Corp. Endoscope apparatus based on plural luminance and wavelength
US20130265401A1 (en) * 2012-03-30 2013-10-10 Olympus Medical Systems Corp. Endoscope apparatus
US20130301727A1 (en) * 2012-05-14 2013-11-14 Qualcomm Incorporated Programmable and scalable integer search for video encoding
CN104285446A (en) * 2012-05-14 2015-01-14 高通股份有限公司 Programmable and scalable searching for candidate blocks for inter-coding or/intra coding
US9495970B2 (en) 2012-09-21 2016-11-15 Dolby Laboratories Licensing Corporation Audio coding with gain profile extraction and transmission for speech enhancement at the decoder
US9502046B2 (en) 2012-09-21 2016-11-22 Dolby Laboratories Licensing Corporation Coding of a sound field signal
US9460729B2 (en) 2012-09-21 2016-10-04 Dolby Laboratories Licensing Corporation Layered approach to spatial audio coding
US9858936B2 (en) 2012-09-21 2018-01-02 Dolby Laboratories Licensing Corporation Methods and systems for selecting layers of encoded audio signals for teleconferencing
US20140244805A1 (en) * 2013-02-26 2014-08-28 Samsung Electronics Co., Ltd. Bandwidth estimation-based streaming content processing method and apparatus of terminal
US9860742B2 (en) * 2013-02-26 2018-01-02 Samsung Electronics Co., Ltd. Bandwidth estimation-based streaming content processing method and apparatus of terminal
US9554083B2 (en) 2013-08-12 2017-01-24 Samsung Electronics Co., Ltd. Methods for selecting resolution with minimum distortion value and devices performing the methods
US10003799B2 (en) 2013-08-13 2018-06-19 Beamr Imaging Ltd Quality driven video re-encoding
WO2015022685A1 (en) * 2013-08-13 2015-02-19 I.C.V.T Ltd. Quality driven video re-encoding
WO2016077712A1 (en) * 2014-11-14 2016-05-19 Videopura, Llc System and method for adaptive video streaming with quality equivalent segmentation and delivery
US9936208B1 (en) * 2015-06-23 2018-04-03 Amazon Technologies, Inc. Adaptive power and quality control for video encoders on mobile devices

Also Published As

Publication number Publication date
EP1615447A1 (en) 2006-01-11
EP1615447B1 (en) 2016-03-09

Similar Documents

Publication Publication Date Title
EP1615447B1 (en) Method and system for delivery of coded information streams, related network and computer program product therefor
US7072393B2 (en) Multiple parallel encoders and statistical analysis thereof for encoding a video sequence
US7301999B2 (en) Quantization method and system for video MPEG applications and computer program product therefor
US7706447B2 (en) Switching between bit-streams in video transmission
US8311095B2 (en) Method and apparatus for transcoding between hybrid video codec bitstreams
US6700935B2 (en) Stream based bitrate transcoder for MPEG coded video
US5870146A (en) Device and method for digital video transcoding
US6466623B1 (en) Method and apparatus for motion estimation for high performance transcoding
US6526099B1 (en) Transcoder
US20080056354A1 (en) Transcoding Hierarchical B-Frames with Rate-Distortion Optimization in the DCT Domain
US7933336B2 (en) Image down-sampling transcoding method and device
US9667961B2 (en) Video encoding and decoding apparatus, method, and system
US20050175101A1 (en) Apparatus and method for video communication
US20070071096A1 (en) Transcoder and transcoding method operating in a transform domain for video coding schemes possessing different transform kernels
KR20030073254A (en) Method for adaptive encoding motion image based on the temperal and spatial complexity and apparatus thereof
KR20100120691A (en) Fast macroblock delta qp decision
US6252905B1 (en) Real-time evaluation of compressed picture quality within a digital video encoder
US6947486B2 (en) Method and system for a highly efficient low bit rate video codec
KR100843080B1 (en) Video transcoding method and apparatus thereof
Wang et al. Adaptive downsampling video coding with spatially scalable rate-distortion modeling
US20010014125A1 (en) Motion image coding device and decoding device
US7809061B1 (en) Method and system for hierarchical data reuse to improve efficiency in the encoding of unique multiple video streams
US20030138048A1 (en) Macroblock coding using luminance data in analyzing temporal redundancy of a picture, biased by chrominance data
US7236529B2 (en) Methods and systems for video transcoding in DCT domain with low complexity
KR20070081949A (en) Transcoding apparatus and method thereof

Legal Events

Date Code Title Description
AS Assignment

Owner name: STMICROELECTRONICS S.R.L., ITALY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PICCINELLI, EMILIANO MARIO ANGELO;FILIPPINI, GIANLUCA;ROVATI, FABRIZIO SIMONE;REEL/FRAME:016620/0354;SIGNING DATES FROM 20050831 TO 20050901

STCB Information on status: application discontinuation

Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION