|Publication number||USRE38079 E1|
|Application number||US 09/113,615|
|Publication date||15 Apr 2003|
|Filing date||10 Jul 1998|
|Priority date||21 Apr 1993|
|Also published as||EP0997039A1, EP0997039A4, WO1997027704A1|
|Publication number||09113615, 113615, US RE38079 E1, US RE38079E1, US-E1-RE38079, USRE38079 E1, USRE38079E1|
|Inventors||Kinya Washino, Barry H. Schwab|
|Original Assignee||Muti-Format, Inc.|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (53), Non-Patent Citations (18), Referenced by (39), Classifications (89), Legal Events (3)|
|External Links: USPTO, USPTO Assignment, Espacenet|
This application is a continuation-in-part of U.S. patent application Ser. No. 08/050,861, filed Apr. 21, 1993.
This invention relates generally to video production, photographic image processing, and computer graphics design, and, more particularly, to a multi-format video production system capable of professional quality editing and manipulation of images intended for television and other applications, including HDTV programs.
As the number of television channels available through various program delivery methods (cable TV, home video, broadcast, etc.) continues to proliferate, the demand for programming, particularly high-quality HDTV-format programming, presents special challenges, both technical and financial, to program producers. While the price of professional editing and image manipulation equipment continues to increase, due to the high cost of research and development and other factors, general-purpose hardware, including personal computers, can produce remarkable effects at a cost well within the reach of non-professionals, even novices. As a result, the distinction between these two classifications of equipment has become less well defined.
The parent to this application, for example, describes a video production system which integrates equipment supplied by various manufacturers, enabling a consumer to produce and edit video material using an enhanced personal computer. An adapter unit interfaced to each camera in use with the system connects to a camera interface module, and each camera interface module, in turn, feeds a computer interface unit. These computer interface units communicate with a personal computer over a standard interconnect, allowing an operator to control the various cameras while viewing individual video programs which appear in separate “windows” on the computer monitor.
This related invention solves many of the problems associated with combining commercially available hardware to create an economical personal-computer-based system capable of very high quality audio/video production. However, the variety of available and planned program standards and delivery methods places further demands on video production equipment, including the editing and manipulation of images not only from a variety of sources, but in differing pixel formats, frame rates, and so forth. Although general-purpose PC-based equipment may never allow professional-style rendering of images at full resolution in real-time, each new generation of microprocessors enables progressively faster, higher-resolution applications. In addition, as the price of memory circuits and other data storage hardware continues to fall, the capacity of such devices has risen dramatically, thereby improving the prospects for enhancing PC-based image manipulation systems for such applications.
In terms of dedicated equipment, attention has traditionally focused on the development of two kinds of professional image-manipulation systems: those intended for the highest quality levels to support film effects, and those intended for television broadcast to provide “full 35 mm theatrical film quality,” within the realities and economics of present broadcasting systems. Conventional thinking holds that 35 mm theatrical film quality is equivalent to 1200 or more lines of resolution, whereas camera negatives present 2500 or more lines. As a result, image formats under consideration have been directed towards video systems having 2500 or more scan lines for high-level production (such as the Kodak “Electronic Intermediate” system described by Hunt et al.), with hierarchies of production, HDTV broadcast, and NTSC and PAL compatible standards which are derived by down-converting these formats. Several techniques have been described, including those of Bretyl (“3×NTSC ‘Leapfrog’ Production Standard for HDTV”, SMPTE Journal, March 1989), Demos (“An Example Hierarchy of Formats for HDTV”, SMPTE Journal, September 1992), and Lim (“A Proposal for an HDTV/ATV Standard with Multiple Transmission Formats”, SMPTE Journal, August 1993). Most proposals employ progressive scanning, although interlace is considered an acceptable alternative as part of an evolutionary process. In particular, Demos addresses the important issue of compatibility to computer-graphics-compatible formats, although he begins with an 1152-line format, and only considers progressive scanning. And, as pointed out by Thorpe et al., progressive scanning also has drawbacks, and as shown by Kaiser et al. (“Resolution Requirements for HDTV Based Upon the Performance of 35 mm Motion-Picture Films for Theatrical Viewing”, SMPTE Journal, June 1985), even 35 mm theatrical film quality is a misnomer since the realities of mechanical projection systems restrict the typical screen display to less than 700 TV lines/picture height.
Current technology directions in computers and image processing should allow production equipment based upon fewer than 1200 scan lines, with picture expansions to create a hierarchy of upward-converted formats for theatrical projection, film effects, and film recording. In addition general-purpose hardware enhancements should be capable of addressing the economic aspects of production, a subject not considered in detail by any of the available references.
The present invention takes advantage of general-purpose hardware where possible to provide an economical multi-format video production system. In the preferred embodiment, specialized graphics processing capabilities are included in a high-performance personal computer or workstation, enabling the user to edit and manipulate an input video program and produce an output version of the program in a final format which may have a different frame rate, pixel dimensions, or both. An internal production format is chosen which provides the greatest compatibility with existing and planned formats associated with standard and widescreen television, high-definition television, and film. For compatibility with film, the frame rate of the internal production format is preferably 24 fps. Images are re-sized by the system to larger or smaller dimensions so as to fill the particular needs of individual applications, and frame rates are adapted by inter-frame interpolation or by traditional schemes, including “3:2 pull-down” for 24-to-30 fps conversions, or by manipulating the frame rate itself for 24 to 25 fps for a PAL-compatible display. The enhancement to a general-purpose platform preferably takes the form of a graphics processor connected to receive a video signal in an input format. The processor comprises a plurality of interface units, including a standard/widescreen interface unit operative to convert the video program in the input format into an output signal representative of a standard/widescreen formatted image, and output the signal to an attached display device. A high-definition television interface unit is operative to convert the video program in the input format into an output signal representative of an HDTV-formatted image, and output the signal to the display device. A centralized controller in operative communication with the video program input, the graphics processor, and an operator interface, enables commands entered by an operator to cause the graphics processor to perform one or more of the conversions using the television interfaces. The present invention thus encourages production at relatively low pixel dimensions to make use of lower-cost general-purpose hardware and to maintain high signal-to-noise, then subsequently expands the result into a higher-format final program. This is in contrast to competing approaches, which recommend operating at higher resolution, then down-sizing, if necessary, to less expensive formats which has led to the high-cost, dedicated hardware, the need for which the present invention seeks to eliminate.
FIGS. 1A-1D show the preferred and alternative image aspect ratios in pixels;
FIG. 2A shows the mechanical design for a digital camera configured to execute the preferred embodiment;
FIG. 2B shows a digital camera configured to execute the preferred embodiment for several different formats;
FIG. 2C shows a low-cost digital camera configured to execute the preferred embodiment for several different formats;
FIG. 3 shows a functional diagram for disk-based video recording;
FIG. 4 shows the components comprising the multi-format audio/video production system;
FIG. 5 depicts an approach for reducing the chrominance bandwidth of wide-band analog RGB output signals without decreasing the luminance resolution;
FIG. 6 shows the inter-relationship of the multi-format audio/video production system to many of the various existing and planned video formats; and
FIG. 7 shows the implementation of a complete television production system, based on one possible choice for image sizes and aspect ratios.
The present invention builds upon and extends certain of the concepts introduced in the parent to this application, “Personal-Computer-Based Video Production System.” Ser. No. 08/050,861 filed Apr. 21, 1993. The system described in that application allows an operator to control equipment supplied by various manufacturers at a centralized personal computer to produce, edit and record a video program. Each camera to be used with the system described in this previously filed application feeds a signal to the personal computer through a custom adapter unit, cable and camera interface module the latter containing cable compensation and gain circuitry. The interface modules feed a common video switcher, audio mixer and display means, all of which may be provided by a variety of sources, including different manufacturers. In the preferred embodiment, the display is the monitor of a programmed personal computer, and computer interface modules connected between each camera interface module and the computer allow video images generated by the cameras to appear in different windows on the computer monitor. Control signals entered at the computer are routed to the cameras in order to control their functioning.
The present invention is primarily concerned with a different but related aspect of facilitating professional quality audio/video production; namely, the conversion of disparate graphics or television formats, including requisite frame-rate conversions, to establish an interrelated family of aspect ratios, resolutions, and frame rates, while remaining compatible with available and future graphics/TV formats. These formats include images of pixel dimensions capable of being displayed on currently available multi-scan computer monitors, and custom hardware will be described whereby frames of higher pixel-count beyond the capabilities of these monitors may be viewed. Images are re-sized by the system to larger or smaller dimensions so as to fill the particular needs of individual applications, and frame rates are adapted by inter-frame interpolation or by traditional schemes such as using “3:2 pull-down” (for 24 to 30 frame-per-second film-to-NTSC conversions) or by speeding up the frame rate itself (as for 24 to 25 fps for PAL television display). The resizing operations may involve preservation of the image aspect ratio, or may change the aspect ratio by “cropping” certain areas, by performing non-linear transformations, such as “squeezing” the picture, or by changing the vision center for “panning,” “scanning” and so forth. Inasmuch as film is often referred to as “the universal format,” primarily because 35-mm film equipment is standardized and used throughout the world, the preferred internal or “production” frame rate is preferably 24 fps. This selection also has an additional benefit, in that the 24 fps rate allows the implementation of cameras having greater sensitivity than at 30 fps, which is even more critical in systems using progressive scanning, for which the rate will be 48 fields per second vs. 60 fields per second in some other proposed systems.
The image dimensions chosen allow the use of conventional CCD-type cameras, but the use of digital processing directly through the entire signal chain is preferred, and this is implemented by replacing the typical analog RGB processing circuitry with fully digital circuitry. Production effects may be conducted in whatever image size is appropriate, and then re-sized for recording. Images are recorded by writing the digital data to storage devices employing removable hard-disk drives, disk drives with removable media, optical or magneto-optical based drives, or tape-based drives, preferably in compressed-data form. As data rates for image processing and reading-from or writing-to disk drives increase, many processes that currently require several seconds will soon become attainable in real-time, which will eliminate the need to record film frames at slower rates. Other production effects, such as slow-motion or fast-motion may be incorporated, and it is only the frame rates of these effects that are limited in any way by the technology of the day. In particular, techniques such as non-linear-editing, animation, and special-effects will benefit from the implementation of this system. In terms of audio, the data rate requirements are largely a function of sound quality. The audio signals may be handled separately, as in an “interlocked” or synchronized system for production, or the audio data may be interleaved within the video data stream. The method selected will depend on the type of production manipulations desired, and by the limitations of the current technology.
Although a wide variety of video formats and apparatus configurations are applicable to the present invention, the system will be described in terms of the alternatives most compatible with currently available equipment and methods. FIG. 1A illustrates one example of a compatible system of image sizes and pixel dimensions. The selected frame rate is preferably 24 per second (2:1 interlaced), for compatibility with film elements; the selected picture dimension in pixels is preferably 1024×576 (0.5625 Mpxl), for compatibility with the 16:9 “widescreen” aspect ratio anticipated for all HDTV systems, and the conventional 4:3 aspect ratio used for PAL systems [768×576 (0.421875 Mpxl)]. All implementations preferably rely on square pixels, though other pixel shapes may be used. Re-sizing (using the well known, sophisticated sampling techniques available in many image-manipulation software packages or, alternatively, using hardware circuitry described herein below) to 2048×1152 (2.25 Mpxl) provides an image suitable for HDTV displays or even theatrical projection systems, and a further re-sizing to 4096×2304 (9.0 Mpxl) is appropriate for even the most demanding production effects. Images may be data compressed 5:1 for 16:9 “wide-screen” TV frames, or 10:1 for HDTV; the data files may then be stored on conventional disk drives, requiring only approximately 8.1 MB/sec for wide-screen frames in RGB, and only 16.1 MB/sec for HDTV frames in RGB.
An alternative embodiment of the invention is shown in FIG. 1B. In this case, the user would follow a technique commonly used in film production, in which the film is exposed as a 4:3 aspect ratio image. When projected as a wide-screen format image, the upper and lower areas of the frame may be blocked by an aperture plate, so that the image shows the desired aspect ratio (typically 1.85:1 or 1.66:1). If the original image format were recorded at 24 frames per second, with a 4:3 ratio and with a dimension in pixels of 1024×768, all image manipulations would preserve these dimensions. Complete compatibility with the existing formats would result, with NTSC and PAL images produced directly from these images by re-scaling, and the aforementioned wide-screen images would be provided by excluding 96 rows of pixels from the top of the image and 96 rows of pixels from the bottom of the image, resulting in the 1024×576 image size as disclosed above. The data content of each of these frames would be 0.75 Mpxls, and the data storage requirements disclosed above would be affected accordingly.
Another embodiment of the invention is depicted in FIG. 1C. In this alternative, the system would follow the image dimensions suggested in several proposed digital HDTV formats under consideration by the Advanced Television Study Committee of the Federal Communications Commission. The format to be adopted is expected to assume a wide-screen image having dimensions of 1280×720 pixels. Using these image dimensions (but at 24 fps with 2:1 interlace), compatibility with the existing formats would be available, with NTSC and PAL images derived from this frame size by excluding 160 columns of pixels from each side of the image, thereby resulting in an image having a dimension in pixels of 960×720. This new image would then be re-scaled to produce images having pixel dimensions of 640×480 for NTSC, or 768×576 for PAL; the corresponding wide-screen formats would be 854×480 and 1024×576, respectively. In this case, an image having a dimension in pixels of 1280×720 would contain 0.87890625 Mpxl, with 1,000 TV lines of resolution; furthermore, the systems under evaluation by the ATSC of the FCC also assume a decimation of the two chrominance signals, with detail of only 640×360 pixels retained. The data storage requirements disclosed above would be affected accordingly. The development path to 24 fps with progressive scanning is both well-defined and practical, as is the use of the previously described methods to produce images having a dimension in pixels of 2048×1152.
A further alternative embodiment of the invention is shown in FIG. 1D. As with the system described with reference to FIG. 1B, the user follows the technique commonly used in film production, wherein the film is exposed as a 4:3 aspect ratio image. When projected as a wide-screen format image, the upper and lower areas of the frame area again blocked by an aperture plate, so that the image shows the desired aspect ratio (typically 1.85:1 or 1.66:1). For an original image format recorded at 24 frames per second, with 4:3 ratio and with pixel dimensions of 1280×960, all image manipulations preserve these dimensions. Complete compatibility with the existing formats results, with NTSC and PAL images produced directly from these images by rescaling, and the aforementioned wide-screen images are provided by excluding 120 rows of pixels from the top of the image and 120 rows of pixels from the bottom of the image, thereby resulting in the 1280×720 image size as described above. The data content of each of these frames is 0.87890625 Mpxls, and the data storage requirements disclosed above are affected accordingly.
Currently available CCD elements for PAL/HDTV dual-use cameras provide 600,000 pixels, typically as arrays of 1024×592 or similar dimensions. By modifying the camera circuitry, the optical and CCD-driver circuitry may be adapted for use by the present invention, thereby allowing for economical implementation of the preferred configuration. FIG. 2A shows a camera as modified for this application. A lens 2 and viewfinder 4 are mounted upon the body of the camera frame. The usual optical-splitter, CCD-sensors and driver circuitry, and the inventive all-digital signal processing circuitry are located at 6, with optional battery-pack capability at 10. The various analog and digital output signals and any input audio, video or control signals, all shown generally at 16, are interfaced through appropriate connectors disposed on the rear-panel 12 and sub-panel 14. Provisions are included as shown for the input of analog audio signals, and for the output of both analog and digital audio signals. Preferably fiber-optic cabling is employed to carry the necessary signals. Internal video recording facilities 8 are described herein below.
Conventional CCD-element cameras of the type described above produce images of over 800 TV Lines horizontal Luminance (Y) resolution, with a sensitivity of 2,000 lux at f8, and with a signal-to-noise ratio of 62 dB. However, typical HDTV cameras, at 1,000 TV Lines resolution and with similar sensitivity, produce an image with only a 54 dB signal-to-noise ratio, due to the constraints of the wideband analog amplifiers and the smaller physical size of the CCD-pixel-elements. By employing the more conventional CCD-elements in the camera systems of this invention, and by relying upon the computer to create the HDTV-type image by image re-sizing, the improved signal-to-noise ratio is retained. In the practical implementation of cameras conforming to this new design approach, there will be less of a need for extensive lighting provisions, which in turn, means less demand upon the power generators in remote productions, and for AC-power in studio applications.
FIG. 2B shows the configuration of a digital video camera implementing the preferred embodiment of the invention. A lens assembly 20 is coupled to an optical beam-splitter 22, which focuses red, green and blue images onto CCD-elements 24a, 24b, and 24c, respectively. The output signals from each of these CCD-elements is directed to its respective analog-to-digital converter 26a, 26b, and 26c. The output of these three analog-to-digital converters is carried to digital signal processor 28, which provides digital signal outputs 34, configured as RGB, Y/R-Y/B-Y, YUV, YIQ, or any other format, as desired. In addition, these digital output signals are also provided to digital-to-analog converters 30a, 30b, and 30c, and from these converters to the analog signal processor 32. This processor provides the analog output signals 36 in the format desired, including the RGB, Y/R-Y/B-Y, YUV, YIQ, or other formats as described above, or additionally, in the composite video or Y/C formats commonly employed in conventional video production equipment and VTRs. A fiber-optic interface 38 accepts digital video signals from the digital signal processor 28 and provides these signals through the fiber-optic cable 40. Control signals are received from the fiber-optic cable 40 and carried through to the digital signal processor 28; other camera operational and status signals, such as tally signals, remote lens controls, return video signals, and so forth, are carried in the reverse direction along this same path from the digital signal processor 28, through the fiber-optic interface 38, to the fiber-optic cable 40.
In practice, the implementation of this design using three 600,000-element CCDs and the commonly employed technique of the spatial-shift for the green CCD-element (as described below) will produce Y/R-Y/B-Y signals with 800 TV lines of resolution, and will provide a luminance bandwidth of 15 MHz and a Chrominance bandwidth of 7.5 MHz. The RGB video signal outputs will provide a full 15 MHz bandwidth for each channel, and the camera will be suitable for the conventional/widescreen application described herein. However, for HDTV production, a higher performance level is desired. Accordingly, the system of FIG. 2B, as described above, is implemented with three of the latest 2.4 Mpxl CCD-elements, providing images of pixel dimension 2048×1152. In the digital realm, the resultant image is 6.75 MB per frame, and the data rate of 162 MB/sec is subjected to a 10:1 data-compression to 16.2 MB/sec for recording and production. The resulting image exhibits over 1,000 TV lines of resolution, again relying upon the spatial shift of the green CCD-element as described herein below. For Y/R-Y/B-Y signals, the Luminance bandwidth will be 60 MHz, and the Chrominance bandwidth will be 30 MHz. The RGB video signal outputs will provide a full 60 MHz bandwidth for each channel. In this case, it will be possible to re-size the picture image to be as large as 8192×4608, which would even enable the system to be used for special optical effects, or with other specialized film formats, such as IMAX and those employing 65 mm camera negatives.
A more economical alternative implementation of the camera system is shown in FIG. 2C. In this case, the camera employs a single 1.2 Mpxl CCD-element, using color filters to produce the color signals. As shown, the camera lens assembly 42 is coupled to the color-filter assembly 44. The Luminance signal 46, and the Chrominance signals 48 are provided to the inputs of their respective analog-to-digital converters 50 and 52. The outputs of these converters are provided to the digital signal processor 54, which produces the digital video output signals 62. These signals may be in any of a number of alternative formats, including, for example, RGB, Y/R-Y/B-Y, YUV, or YIQ. These signals are additionally provided to digital-to-analog converters 56a, 56b, and 56c, respectively, and then to the analog signal processor 60, which provides analog output signals 64 in the format desired, including the RGB, Y/R-Y/B-Y, YUV, YIQ, or other formats as described above, or additionally in the composite video or Y/C formats commonly employed in conventional video production equipment and VTRs. In this case, the image size will be 1024×576 for the luminance channel (producing approximately 600 TV Lines of resolution), and 512×576 for each of the chrominance channels. In this case, it is not possible to introduce the green spatial-shift approach, because only a single CCD-element is employed. However, the luminance channel bandwidth achieved will be 15 MHz, and the chrominance channel bandwidth will be 7.5 MHz.
In CCD-based cameras, it is a common technique to increase the apparent resolution by mounting the red and blue CCD-elements in registration, but offsetting the green CCD-element by one-half pixel width horizontally. In this case, picture information is in-phase, but spurious information due to aliasing is out-of-phase. When the three color signals are mixed, the picture information is intact, but most of the alias information will be canceled out. This technique will evidently be less effective when objects are of solid colors, so it is still the usual practice to include low-pass optical filters mounted on each CCD-element to suppress the alias information. In addition, this technique cannot be applied to computer-based graphics, in which the pixel images for each color are always in registration. However, in general-use video, the result of the application of this spatial-shift offset is to raise the apparent luminance (Y) horizontal resolution to approximately 800 television lines.
The availability of hard-disk drives of progressively higher capacity and data transmission rates is allowing successively longer and higher resolution image displays in real-time. At the previously cited data rates, wide-screen frames would require 486 MB/min, so that currently available 10 GB disk drives will store more than 21 minutes of video. When the anticipated 100 GB disk drives (2.5-inch or 3.5-inch disks using Co-Cr, barium ferrite, or other high-density recording magnetic materials) become available, these units will store 210 minutes, or 3½ hours of video. For this application, a data storage unit 8 is provided to facilitate editing and production activities, and it is anticipated that these units would be employed in much the same way as video cassettes are currently used in Betacam and other electronic news gathering (ENG) cameras and in video productions. This data storage unit may be implemented by use of a magnetic, optical, or magneto-optical disk drive with removable storage media, or by a removable disk-drive unit, such as those based on the PCMCIA standards. Although PCMCIA media are 1.8-inches in dimension, alternative removable media storage units are not restricted to this limit, and could employ larger media, such as 2.5-inch or 3.5-inch disks; this, in turn, will lead to longer duration program data storage, or could be applied to lower ratios of data compression or higher-pixel-count images within the limits of the same size media.
FIG. 3 shows the functional diagram for the storage-device-based digital recorder employed in the video camera, or separately in editing and production facilities. As shown, a removable hard disk drive 70 is interfaced through a bus controller 72; in practice, alternative methods of storage such as optical or magneto-optical drives could be used, based on various interface bus standards such as SCSI-2 or PCMCIA. This disk drive system currently achieves data transfer rates of 20 MB/sec, and higher rates on these or other data storage devices, such as high-capacity removable memory modules, is anticipated. The microprocessor 74 controls the 64-bit or wider data bus 80, which integrates the various components. Currently available microprocessors include the Alpha 21064 by Digital Equipment Corporation, or the MIPS R4400 by MIPS Technologies, Inc.; future implementations would rely on the already announced P6 by Intel Corp. or the PowerPC 620, which is capable of sustained data transfer rates of 100 MB/sec. Up to 256 MB of ROM, shown at 76, is anticipated for operation, as is 256 MB or more of RAM, shown at 78. Current PC-based video production systems are equipped with at least 64 MB of RAM, to allow sophisticated editing effects. The graphics processor 82 represents dedicated hardware that performs the various manipulations required to process the input video signals 84 and the output video signals 86; although shown using an RGB format, either the inputs or outputs could be configured in alternative formats, such as Y/R-Y/B-Y, YIQ, YUV or other commonly used alternatives. In particular, while a software-based implementation of the processor 82 is possible, a hardware-based implementation preferred, with the system employing a compression ratio of 5:1 for the conventional/widescreen signals (“NTSC/PAL/Widescreen”), and a 10:1 compression ratio for HDTV signals (2048×1152, as described herein above). An example of one of the many available options for this data compression is the currently available Motion-JPEG system. Image re-sizing may alternatively be performed by dedicated microprocessors, such as the gm865×1 or gm833×3 by Genesis Microchip, Inc. Audio signals may be included within the data stream, as proposed in the several systems for digital television transmission already under evaluation by the Federal Communications Commission, or by one of the methods available for integrating audio and video signals used in multi-media recording schemes, such as the Microsoft “.AVI” (Audio/Video Interleave) file format. As an alternative, an independent system for recording audio signals may be implemented, either by employing separate digital recording provisions controlled by the same system and electronics, or by implementing completely separate equipment external to the camera system described herein above.
FIG. 4 shows the components that comprise a multi-format audio/video production system. As in the case of the computer disk-based recording system of FIG. 3, an interface bus controller 106 provides access to a variety of storage devices, preferably including an internal hard-disk drive 100, a tape-back-up drive 102, and a hard-disk drive with removable media or a removable hard-disk drive 104. The interface bus standards implemented could include, among others, SCSI-2 or PCMCIA. Data is transmitted to and from these devices under control of microprocessor 110. Currently, data bus 108 would operate as shown as 64-bits wide, employing microprocessors such as those suggested for the computer-disk-based video recorder of FIG. 3; as higher-powered microprocessors become available, such as the PowerPC 620, the data bus may be widened to accommodate 128 bits, and the use of multiple parallel processors may be employed, with the anticipated goal of 1,000 MIPS per processor. Up to 256 MB of ROM 112 is anticipated to support the requisite software, and at least 1,024 MB of RAM 114 will allow for the sophisticated image manipulations, inter-frame interpolation, and intra-frame interpolation necessary for sophisticated production effects, and for conversions between the various image formats.
A key aspect of the system is the versatility of the graphics processor shown generally as 116. Eventually, dedicated hardware will allow the best performance for such operations as image manipulations and re-scaling, but it is not a requirement of the system that it assume these functions. Three separate sections are employed to process the three classifications of signals. Although the video input and output signals described herein below are shown, by example, as RGB, any alternative format for video signals, such as Y/R-Y/B-Y, YIQ, YUV, or other alternatives may be employed as part of the preferred embodiment. One possible physical implementation would be to create a separate circuit board for each of the sections as described below, and manufacture these boards so as to be compatible with existing or future PC-based electrical and physical interconnect standards.
A standard/widescreen video interface 120, intended to operate within the 1024×576 or 1024×768 image sizes, accepts digital RGB signals for processing and produces digital RGB outputs in these formats, as shown generally at 122. Conventional internal circuitry comprising D/A converters and associated analog amplifiers are employed to convert the internal images to a second set of outputs, including analog RGB signals and composite video signals. These outputs may optionally be supplied to either a conventional multi-scan computer video monitor or a conventional video monitor having input provisions for RGB signals (not shown). A third set of outputs supplies analog Y/C video signals. The graphics processor may be configured to accept or output these signals in the standard NTSC, PAL, or SECAM formats, and may additionally be utilized in other formats as employed in medical imaging or other specialized applications, or for any desired format for computer graphics applications. Conversion of these 24 frame-per-second images to the 30 fps (actually, 29.97 fps) NTSC and 25 fps PAL formats may be performed in a similar manner to that used for scanned film materials, that is to NTSC by using the conventional 3:2 “pull-down” field-sequence, or to PAL by running the images at the higher 25 fps rate. For other HDTV frame rates, aspect ratios, and line rates, intra-frame and inter-frame interpolation and image conversions may be performed by employing comparable techniques well known in the art of computer graphics and television.
The management of 25 fps (PAL-type) output signals in a system configured for 24 fps production applications presents technical issues which must be addressed, however. Simple playback of signals to produce PAL output is not a serious problem, since any stored video images may be replayed at any frame rate desired, and filmed material displayed at 25 fps is not objectionable. Indeed, this is the standard method for performing film-to-tape transfers used PAL- and SECAM-television countries. However, it is not practical to produce both PAL and NTSC signals concurrently from a single source running at 24 fps. Simultaneous output of both NTSC and film-rate images is performed by exploiting the 3:2 field-interleaving approach: 5×24=2×60; that is, two film frames are spread over five video fields. This makes it possible to concurrently produce film images at 24 fps and video images at 30 fps. The difference between 30 fps and the exact 29.97 fps rate of NTSC may be palliated by slightly modifying the system frame rate to 23.976 fps. This is not noticeable in normal film projection, and is an acceptable deviation from the normal film rate. However, if the system frame rate is adjusted to 25 fps to produce PAL or SECAM output, there is no convenient technique to produce 30 fps NTSC concurrently, unless multiple-frame storage with motion-interpolation is employed, which tends to create udesirable artifacts in the image produced. Commercial standards-converters are available to perform this function, however, from companies such as Snell & Wilcox. This system is primarily directed towards production of video-based film and high-definition TV images, for which 24 fps and 30 fps, respectively, are the established frame rate for film and the proposed frame rate for HDTV (in NTSC-countries). The conversion to 25 fps is performed without difficulties in any application in which there is no requirement for the simultaneous production of images at other frame rates. Using this approach, the adjustment of frame rates for playback of the images by the system is sufficient for all of the normal production applications.
An HDTV video interface 124, intended to operate within the 2048×1152 or 2048×1536 image sizes (with re-sizing as necessary), accepts digital RGB (or alternative) signals for processing and produces digital outputs in the same image format, as shown generally at 126. As is the case for the Standard/Widescreen interface 120, conventional internal circuitry comprising D/A converters and associated analog amplifiers are employed to convert the internal images to a second set of outputs, for analog RGB signals and composite video signals. In normal practice, these outputs would have a full 15 MHz bandwidth for each of the three R, G, and B signals. However, by applying the technique shown in FIG. 5, it is possible to produce a signal having a 15 MHz luminance bandwidth, but only 7.5 MHz chrominance bandwidth. In effect, the circuitry shown simulates the results of applying a 4:2:2 sampling technique (as is commonly used in the Television Industry) without employing the step of creating the two chrominance components for sub-sampling, for example, I and Q for NTSC, U and V for PAL, or R-Y and B-Y. As shown, analog R, G, and B signals 140a, 140b, and 140c are supplied to low-pass filters 142a, 142b, and 142c, respectively, which are designed to remove frequencies above 7.5 MHz. In addition, these R, G, and B signals are applied to a standard RGB-to-Y matrix 144 to produce a standard luminance Y signal, which is carried to high-pass filter 146 which is designed to remove signal components below 7.5 MHz. This filtered luminance signal is then carried to a standard Y-to-RGB Matrix 148, in which the signal is proportionately split into R, G, and B components, and then supplied to mixers 150a, 150b, and 150c, wherein the luminance signal is mixed with R, G, and B signals from the three low-pass filters 142a, 142b, and 142c. The resulting analog R, G, and B outputs now have the full 15 MHz luminance bandwidth, but the chrominance bandwidth has been limited to 7.5 MHz. It is anticipated that different applications may require modification of the luminance bandwidth from 15 MHz, and of the chrominance bandwidth from 7.5 MHz, and the application of these techniques should be considered to be within the scope of this invention.
The third section of the graphics processor 116 shown in FIG. 4 is the film output video interface 128, which comprises a special set of video outputs 130 intended for use with devices such as laser film recorders. These outputs are preferably configured to provide a 4096×2304 or 4096×3072 image size from the image sizes employed internally, using re-sizing techniques discussed herein as necessary for the format conversions. Although 24 fps is the standard frame rate for film, some productions employ 30 fps, especially when used with NTSC materials, and these alternative frame rates, as well as alternative image sizes, are anticipated as suitable applications of the invention.
Several additional features of this system are disclosed in FIG. 4. The graphics processor includes a special output 132 for use with a color printer. In order to produce the highest quality prints from the screen display it is necessary to adjust the printer resolution to match the image resolution, and this is automatically optimized by the graphics processor for the various image sizes produced by the system. In addition, provisions are included for an image scanner 134, which may be implemented as a still image scanner or a film scanner, thereby enabling optical images to be integrated into the system. An optional audio processor 136 includes provisions for accepting audio signals in either analog or digital form, and outputting signals in either analog or digital form, as shown in the area generally designated as 138. For materials including audio intermixed with the video signals as described herein above, these signals are routed to the audio processor for editing effects and to provide an interface to other equipment.
It is important to note that although FIG. 4 shows only one set of each type of signal inputs, the system is capable of handling signals simultaneously from a plurality of sources and in a variety of formats. Depending on the performance level desired and the image sizes and frame rates of the signals, the system may be implemented with multiple hard disk units and bus controllers, and multiple graphics processors, thereby allowing integration of any combination of live camera signals, prerecorded materials, and scanned images. Improved data compression schemes and advances in hardware speed will allow progressively higher frame rates and image sizes to be manipulated in real-time.
FIG. 6 shows the inter-relationship of the various film and video formats compatible with the invention, though not intended to be inclusive of all possible implementations. In typical operations, the multi-format audio/video production system 162 would receive film-based elements 160 and combine them with locally produced materials already in the preferred internal format of 24 frames-per-second. In practice, materials May be converted from any other format including video at any frame rate or standard. After the production effects have been performed, the output signals may be configured for any use required, including, but not limited to, HDTV at 30 fps shown as 164, NTSC/widescreen at 30 fps shown as 166, PAL-SECAM/widescreen at 25 fps shown as 170, or HDTV at 25 fps shown as 172. In addition, output signals at 24 fps are available for use in a film-recording unit 168.
FIG. 1A shows the preferred family of aspect ratios and image frame sizes in pixels. The internal production storage format 180 has frame size 1024×576 with aspect ratio 16:9, and may be trimmed of side panels to use as a 768×576 image frame with aspect ratio of 4:3 in conventional television formats such as NTSC or PAL. After a 2:1 expansion/re-sizing, the HDTV format 182 is available, with frame size 2048×1152 and the same 16:9 aspect ratio. A further 2:1 expansion/re-sizing to the film format 184, with frame size 4096×2304 and the same 16:9 aspect ratio, allows for recording of film via currently available technology.
FIG. 1B shows an alternative family of aspect ratios and image frame sizes in pixels. The internal production storage format 190 has frame size 1024×768 with aspect ratio 4:3 as employed in conventional television formats such as NTSC, or PAL, and may be trimmed of top and bottom panels to use as a 1024×576 image frame with aspect ratio of 16:9. After a 2:1 expansion/resizing, the intermediate format 192 is available, with frame size 2048×1536 and the same 4:3 aspect ratio. A further 2:1 expansion/re-sizing to the alternative Film format 194, with frame size 4096×3072 and the same 4:3 aspect ratio, allows for recording of film via currently available technology.
FIG. 1C shows another alternative family of aspect ratios and image frame sizes in pixels, based on compatibility with several of the proposed digital HDTV formats. The internal production storage format 200 has frame size 1280×720 with aspect ratio 16:9, and may be trimmed of side panels to use as a 960×720 image frame with aspect ratio of 4:3 in conventional television formats such as NTSC or PAL. After a 2:1 expansion/re-sizing, the HDTV format 202 is available, with frame size 2560×1440 and the same 16:9 aspect ratio. A further 2:1 expansion/re-sizing to the film format 204, with frame size 5120×2880 and the same 16:9 aspect ratio, allows for recording of film via currently available technology.
FIG. 1D shows another alternative family of aspect ratios and image frame sizes in pixels. The internal production storage format 206 has frame size 1280×960 with aspect ratio 4:3 as employed in conventional television formats such as NTSC or PAL, and may be trimmed of top and bottom panels to use as a 1280×720 image frame with aspect ratio of 16:9. After a 2:1 expansion/re-sizing, the intermediate format 208 is available, sizing to the alternative film format 209, with frame size 5120×3840 and the same 4:3 aspect ratio, allows for recording of film via currently available technology.
Alternative implementations may employ different frame size (in pixels), aspect ratios, or frame rates, and these variations should be considered to be within the scope of the invention.
FIG. 7 shows an implementation involving one possible choice for image sizes, aspect ratios, and frame rates to provide a universal television production system. As shown, signals are provided from any of several sources, including conventional broadcast signals 210, satellite receivers 212, and interfaces to a high bandwidth data network 214. These signals would be provided to the digital tuner 218 and an appropriate adapter unit 220 for the data network or “information superhighway” before being supplied to the decompression processor 222. The processor 222 provides any necessary data de-compression and signal conditioning for the various signal sources, and preferably is implemented as a plug-in circuit board for a general-purpose computer, though the digital tuner 218 and the adapter 220 optionally may be included as part of the existing hardware.
The output of processor 222 is provided to the internal data bus 226. The system microprocessor 228 controls the data bus, and is provided with 16 to 64 MB of RAM 230 ad up to 64 Mb of ROM 232. This microprocessor could be implemented using one of the units previously described, such as the PowerPC 604 or PowerPC 620. A hard disk drive controller 234 provides access to various storage means, including, for example, an internal hard disk drive unit 236, a removable hard disk drive unit 238, or a tape drive 240; these storage units also enable the PC to function as a video recorder, as described above. A graphic processor 242, comprising dedicated hardware which optionally be implemented as a separate plug-in circuit board, performs the image manipulations required to convert between the various frame sizes (in pixels), aspect ratios, and frame rates. This graphics processor uses 16 to 32 MB of DRAM, and 2 to 8 MB of VRAM, depending on the type of display output desired. For frame size of 1280×720 with an aspect ratio 16:9, the lower range of DRAM and VRAM will be sufficient, but for a frame size of 2048×1152, the higher range of DRAM and VRAM is required. In general, the 1280×720 size is sufficient for conventional “multi-sync”, computer display screens up to 20 inches, and the 2048×1152 size is appropriate for conventional “multi-sync” computer display screens up to 35 inches. Analog video outputs 244 are available for these various display units. Using this system, various formats may be displayed, including (for 25 fps, shown by speeding up 24 fps signals) 768×576 PAL/SECAM, 1024×576 wide-screen, and 2048×1152 HDTV, and (for 30 fps, shown by utilizing the well-known “3:2 pull-down” technique, and for 29.97 fps, shown by a slight slow-down in 30 fps signals) 640×480 NTSC and 854×480 wide-screen, and 1280×720 USA and 1920×1080 NHK (Japan) HDTV. While most NTSC monitors will synchronize to a 30 fps signal, possibly requiring that the color subcarrier frequency be adjusted, many PAL and SECAM monitors will not accept a 24 fps signal. In this case, more sophisticated frame-rate conversion techniques may be required for viewing live broadcasts, since the 24 fps input signal rate cannot keep pace with the 25 fps display rate. However, in practice it is anticipated that future television sets will incorporate “multi-sync” designs that eliminate this potential problem.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US3617626 *||16 May 1969||2 Nov 1971||Technicolor||High-definition color picture editing and recording system|
|US3882539||12 Feb 1973||6 May 1975||Faroudja Y C||Method and apparatus for improved skip field recording|
|US4357624 *||20 Mar 1981||2 Nov 1982||Combined Logic Company||Interactive video production system|
|US4633293 *||13 Aug 1984||30 Dec 1986||Rca Corporation||High definition television signal for film-television standards conversion system|
|US4936816||4 Apr 1989||26 Jun 1990||Winkler & Dunnebier Maschinenfabrik Und Eisengiesserei Kg||Process and apparatus for opening the sealing flaps of a carton|
|US4982280||18 Jul 1989||1 Jan 1991||Yves C. Faroudja||Motion sequence pattern detector for video|
|US4998167||14 Nov 1989||5 Mar 1991||Jaqua Douglas A||High resolution translation of images|
|US5027206||13 Sep 1989||25 Jun 1991||U.S. Philips Corporation||High-definition television systems|
|US5045932 *||29 Jun 1989||3 Sep 1991||Eastman Kodak Company||Method and apparatus for generating a high definition electronic signal from a line scan of a color original|
|US5111292||27 Feb 1991||5 May 1992||General Electric Company||Priority selection apparatus as for a video signal processor|
|US5221966||16 Jan 1991||22 Jun 1993||Avesco Plc||Video signal production from cinefilm originated material|
|US5243433 *||6 Jan 1992||7 Sep 1993||Eastman Kodak Company||Digital image interpolation system for zoom and pan effects|
|US5291280||5 May 1992||1 Mar 1994||Faroudja Y C||Motion detection between even and odd fields within 2:1 interlaced television standard|
|US5317398||17 Aug 1992||31 May 1994||Rca Thomson Licensing Corporation||Video/film-mode (3:2 pulldown) detector using patterns of two-field differences|
|US5327235 *||27 Jan 1993||5 Jul 1994||Sony United Kingdom Limited||Video conversions of video signal formats|
|US5329309 *||19 Jul 1993||12 Jul 1994||Sony United Kingdom Limited||Method of integrating format material and an interlace scan format signal|
|US5331346||7 Oct 1992||19 Jul 1994||Panasonic Technologies, Inc.||Approximating sample rate conversion system|
|US5337154||19 Jul 1993||9 Aug 1994||Sony United Kingdom Limited||Format conversion of digital video signals, integration of digital video signals into photographic film material and the like, associated signal processing, and motion compensated interpolation of images|
|US5365273||23 Apr 1993||15 Nov 1994||Deutsche Thomson-Brandt Gmbh||Method and device for film-mode detection|
|US5384598||20 Oct 1992||24 Jan 1995||International Business Machines Corporation||System and method for frame differencing video compression and decompression with frame rate scalability|
|US5444491||6 Dec 1994||22 Aug 1995||Massachusetts Institute Of Technology||Television system with multiple transmission formats|
|US5446497||30 Sep 1993||29 Aug 1995||Sony United Kingdom Ltd.||Method and apparatus for processing an input 60 field/second video signal generated by 3232 pulldown to produce an output video signal|
|US5461420||17 Sep 1993||24 Oct 1995||Sony Corporation||Apparatus for coding and decoding a digital video signal derived from a motion picture film source|
|US5517248||26 May 1995||14 May 1996||Victor Company Of Japan, Ltd.||Frame-frequency converting apparatus for a video signal resulting from 2-3 conversion of original picture information|
|US5519438||30 Jun 1994||21 May 1996||Intel Corporation||Computer with a video subsystem that contains timers which are used to create calibration tables correlating time intervals with the decoding and converting of video input signals|
|US5532749||5 Oct 1994||2 Jul 1996||Goldstar Co., Ltd.||Sample rate conversion device for processing non-standard television signal|
|US5563660||30 Nov 1993||8 Oct 1996||Sony Corporation||Decoder for a compressed digital video signal using a common memory for decoding and 2/3 pull-down conversion|
|US5565998||22 Feb 1994||15 Oct 1996||U.S. Philips Corporation||Identifying film frames in a video sequence|
|US5594552||13 May 1994||14 Jan 1997||Sony Corporation||Apparatus and method for producing downwards compatible video signals with increased vertical resolution, and apparatus for reproducing and displaying same|
|US5600377||9 Sep 1993||4 Feb 1997||Sony Corporation||Apparatus and method for motion compensating video signals to produce interpolated video signals|
|US5606539||31 Aug 1994||25 Feb 1997||U.S. Philips Corporation||Method and apparatus for encoding and decoding an audio and/or video signal, and a record carrier for use with such apparatus|
|US5608464||9 Aug 1994||4 Mar 1997||Scitex Corporation Ltd.||Digital video effects generator|
|US5617218||6 May 1993||1 Apr 1997||Advanced Television Test Center||Bi-directional television and motion picture film to magnetic tape format digital signal converter|
|US5649048 *||20 Jun 1994||15 Jul 1997||Matsushita Electric Industrial Co., Ltd.||Video taperecorder with television standards converter|
|US5666461||30 May 1995||9 Sep 1997||Sony Corporation||High efficiency encoding and decoding of picture signals and recording medium containing same|
|US5724101||7 Feb 1991||3 Mar 1998||Prevail, Inc.||System for conversion of non standard video signals to standard formats for transmission and presentation|
|US5742351||11 Oct 1995||21 Apr 1998||U.S. Philips Corporation||Device for encoding sequences of frames constituted by film-type images and video-type images, and corresponding decoding device|
|US5754248 *||15 Apr 1996||19 May 1998||Faroudja; Yves C.||Universal video disc record and playback employing motion signals for high quality playback of non-film sources|
|US5768469||29 Oct 1996||16 Jun 1998||Sony Corporation||Apparatus for coding and decoding a digital video signal having duplicate pictures and frames with fields originating from different film source frames|
|US5771073||7 Jun 1995||23 Jun 1998||Massachusetts Institute Of Technology||Advanced television system using a different encoding technique for non-image areas|
|US5812204||25 Jul 1996||22 Sep 1998||Brooktree Corporation||System and method for generating NTSC and PAL formatted video in a computer system|
|US5832085||25 Mar 1997||3 Nov 1998||Sony Corporation||Method and apparatus storing multiple protocol, compressed audio video data|
|US5835150||18 Oct 1995||10 Nov 1998||Lg Electronics Inc.||Image format converter for HDTV|
|US5835672||20 Sep 1993||10 Nov 1998||Sony Corporation||Apparatus for coding and decoding a digital video signal having duplicate pictures and frames with fields originating from different film source frames|
|US5838381||20 Dec 1996||17 Nov 1998||Hitachi, Ltd.||Image display apparatus capable of displaying personal computer signals and television signal by conversion into a signal of a higher number of pixels|
|US5930445 *||24 Feb 1995||27 Jul 1999||Avid Technology, Inc.||Electronic film editing system using both film and videotape format|
|US5999220 *||7 Apr 1997||7 Dec 1999||Washino; Kinya||Multi-format audio/video production system with frame-rate conversion|
|EP0314873A1 *||30 Jun 1988||10 May 1989||GRUNDIG E.M.V. Elektro-Mechanische Versuchsanstalt Max Grundig holländ. Stiftung & Co. KG.||Device for automatic application of tuning voltage to tunable components of the intermediate frequency amplifier for television receivers|
|EP0514012A2 *||15 Apr 1992||19 Nov 1992||Vistek Electronics Limited||Method and apparatus for the standard conversion of an image signal|
|JPH0437846A *||Title not available|
|WO1993015586A1 *||29 Jan 1993||5 Aug 1993||Eastman Kodak Co||A partial interpolation method and apparatus for frame rate conversion|
|WO1993023954A1 *||8 May 1992||25 Nov 1993||Tintoretto Corp||Method of and apparatus for transfer of film to video|
|WO1994001971A2 *||1 Jul 1993||20 Jan 1994||Avid Technology Inc||Electronic film editing system using both film and videotape format|
|1||*||A.Kaiser,H.W.Mahler, R.H.McMann, "Resolution Requirements for HDTV Based Upon the Performance of 35mm Motion-Picture Films for Theatrical Viewing", SMPTE Journal, Jun. 1985, pp. 654-659.*|
|2||*||B.Hunt, G. Kennel, L.DeMarsh,S.Kristy, High-Resolution Electronic Intermediate System for Motion-Picture Film, SMPTE Journal, Mar. 1991, pp. 156-161.*|
|3||D. Bancroft, Technology Council of the Motion Picture-Televison Industry Newsletter, Oct. 1993.|
|4||D. Bancroft, Technology Council of the Motion Picture—Televison Industry Newsletter, Oct. 1993.|
|5||D.J. Bancroft, "Pixels and Halide-A Natural Partnership?," SMPTE Journal May 1994, pp. 306-311.|
|6||D.J. Bancroft, "Pixels and Halide—A Natural Partnership?," SMPTE Journal May 1994, pp. 306-311.|
|7||*||G. Demos, "An Example of Hierarchy of Formats for HDTV", SMPTE Journal, Sep. 1992, pp. 609-617.*|
|8||*||G.Reitmeier,C. Carlson,E.Geiger,D.Westerkamp, "The Digital Hierarchy-A Blueprint for television in the 21st Century", SMPTE Journal, Jul. 1992, pp. 466-470.*|
|9||G.Reitmeier,C. Carlson,E.Geiger,D.Westerkamp, "The Digital Hierarchy—A Blueprint for television in the 21st Century", SMPTE Journal, Jul. 1992, pp. 466-470.*|
|10||*||J.S. Lim, "A Proposal for an HDTV/ATV Standard with Multiple Transmission Formats", SMPTE Journal, Aug. 1993, pp/. 699-702.*|
|11||*||L.J.Thorpe,T.Hanabusa, "If Progressive Scanning is So Good, How Bad is Interlace?", SMPTE Journal, Dec. 1990, pp. 972-986.|
|12||M. Adams, "5/96 WHITE PAPER-A Broadband Interactive Cable Gateway," May 1996. 14 pages.|
|13||M. Adams, "5/96 WHITE PAPER—A Broadband Interactive Cable Gateway," May 1996. 14 pages.|
|14||M. Adams; "Network Design and Implementation of a Large-Scale, ATM, Multimedia Network," Time Warner Cable, Dec. 1994.|
|15||R. Brown J. Callahan, "5/95 WHITE PAPER Software Architecture for Broadband CATV Interactive Systems," May 1995, 14 pages.|
|16||*||W.E. Bret1, "3XNTSC-A "Leapfrog' Production Standard for HDTV", SMPTE Journal, Mar. 1989, pp. 173-178.*|
|17||W.E. Bret1, "3XNTSC—A ‘Leapfrog’ Production Standard for HDTV", SMPTE Journal, Mar. 1989, pp. 173-178.*|
|18||*||Y. Ide,M.Sasuga,N.Harada,T.Nishizawa, "A Three-CCD HDTV Color Camera", SMPTE Journal, Jul. 1990, pp. 532-537.*|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US6678740 *||23 Jun 2000||13 Jan 2004||Terayon Communication Systems, Inc.||Process carried out by a gateway in a home network to receive video-on-demand and other requested programs and services|
|US6724420||21 Dec 2000||20 Apr 2004||Dfr2000, Inc.||Portable film conversion device|
|US6829012 *||21 Dec 2000||7 Dec 2004||Dfr2000, Inc.||Method and apparatus for a digital parallel processor for film conversion|
|US6891562||21 Dec 2000||10 May 2005||Stuart T. Spence||Optical design for film conversion device|
|US6891569 *||5 Sep 2000||10 May 2005||Victor Company Of Japan Limited||Wide angle image pickup apparatus|
|US7202893||3 Jan 2005||10 Apr 2007||Microsoft Corporation||Method and apparatus for the display of still images from image files|
|US7206025 *||23 Mar 2001||17 Apr 2007||Lg Electronics Inc.||Device and method for converting format in digital TV receiver|
|US7292273 *||26 Aug 2002||6 Nov 2007||Sanyo Electric Co., Ltd.||Digital camera accommodating recording media from other digital cameras|
|US7403212||30 Dec 2005||22 Jul 2008||Microsoft Corporation||Method and apparatus for the display of still images from image files|
|US7432920||30 Nov 2004||7 Oct 2008||Microsoft Corporation||Method and apparatus for the display of still images from image files|
|US7453519 *||25 Mar 2005||18 Nov 2008||Olympus Corporation||Method and apparatus for converting images at a given frame or field rate to second and third frame or field rates while maintaining system synchronism|
|US7697064 *||12 Dec 2005||13 Apr 2010||Nec Electronics Corporation||Video signal processing apparatus and video signal processing method|
|US7716702 *||10 Jun 2004||11 May 2010||Time Warner Cable, Inc.||Technique for delivering via a communications network data for image display with a desired aspect ratio|
|US7830967||12 Jul 2010||9 Nov 2010||Red.Com, Inc.||Video camera|
|US7843508||29 Aug 2007||30 Nov 2010||Mediostream, Inc.||Method and system for direct recording of video information onto a disk medium|
|US7948556 *||8 Jul 2009||24 May 2011||Kabushiki Kaisha Toshiba||Electronic apparatus and display control method|
|US8064752||9 Dec 2003||22 Nov 2011||Apple Inc.||Video encoding|
|US8135261||9 Dec 2003||13 Mar 2012||Apple Inc.||Insertion and usage of metadata in digital video|
|US8174560||8 May 2012||Red.Com, Inc.||Video camera|
|US8237830||13 Apr 2009||7 Aug 2012||Red.Com, Inc.||Video camera|
|US8314861 *||6 Oct 2005||20 Nov 2012||Sony Corporation||Imaging apparatus and capture assist mark usage control method|
|US8358357||3 Aug 2012||22 Jan 2013||Red.Com, Inc.||Video camera|
|US8619188||19 Oct 2010||31 Dec 2013||Mediostream, Inc.||Method and system for direct recording of video information onto a disk medium|
|US8666222||14 Oct 2011||4 Mar 2014||Apple Inc.||Video encoding|
|US8811802||12 Dec 2011||19 Aug 2014||Aplle, Inc.||Insertion and usage of metadata in digital video|
|US8839110||25 Aug 2011||16 Sep 2014||Apple Inc.||Rate conform operation for a media-editing application|
|US8872933||4 May 2012||28 Oct 2014||Red.Com, Inc.||Video camera|
|US8878952||3 Aug 2012||4 Nov 2014||Red.Com, Inc.||Video camera|
|US8907961||30 May 2008||9 Dec 2014||Microsoft Corporation||Method and apparatus for the display of still images from image files|
|US9019393||16 Sep 2014||28 Apr 2015||Red.Com, Inc.||Video processing system and method|
|US20010030709 *||21 Dec 2000||18 Oct 2001||Tarnoff Harry L.||Method and apparatus for a digital parallel processor for film conversion|
|US20040071445 *||22 Jul 2003||15 Apr 2004||Tarnoff Harry L.||Method and apparatus for synchronization of ancillary information in film conversion|
|US20040209034 *||3 May 2004||21 Oct 2004||Flexplay Technologies, Inc.||Limited play optical devices with interstitial reactive layer and methods of making same|
|US20050039211 *||17 Sep 2003||17 Feb 2005||Kinya Washino||High-quality, reduced data rate streaming video production and monitoring system|
|US20050099385 *||2 Dec 2004||12 May 2005||Microsoft Corporation||Method and apparatus for the creation of digital photo albums|
|US20050155086 *||30 Nov 2004||14 Jul 2005||Microsoft Corporation||Method and apparatus for the display of still images from image files|
|US20050163463 *||3 Jan 2005||28 Jul 2005||Microsoft Corporation||Method and apparatus for the display of still images from image files|
|US20120050575 *||29 Aug 2011||1 Mar 2012||Samsung Electronics Co., Ltd.||Method and apparatus for capturing image in portable terminal|
|USRE40675 *||20 May 2004||24 Mar 2009||Infocus Corporation||Method and apparatus for automatic pixel clock phase and frequency correction in analog to digital video signal conversion|
|U.S. Classification||348/722, 348/554, 348/556, 348/445, 348/911|
|International Classification||H04N5/44, H04N9/804, H04N7/10, G11B27/00, H04N5/77, G06F3/14, G11B27/031, H04N5/765, H04N7/14, G11B27/034, H04N7/18, H04N1/00, H04N5/222, H04N7/15, H04N7/01, H04N5/781, H04N5/85, G11B27/34, G11B27/024, H04N5/926, H04N5/915, H04N9/64, H04N5/232, G11B27/032, G06T9/00|
|Cooperative Classification||H04N5/765, G11B2220/913, G11B2220/455, H04N7/0122, H04N5/4401, H04N1/00283, G06T9/007, H04N5/781, H04N5/23203, H04N7/10, G11B2220/20, H04N9/8042, H04N9/642, G11B27/34, H04N5/232, G06F3/1454, G11B2220/90, G09G2340/0407, H04N9/641, G06F3/14, G09G2340/0442, H04N5/222, G11B2220/2525, H04N7/147, H04N7/0112, H04N7/181, G11B2220/41, G11B27/032, G11B2220/2516, G11B2220/61, H04N7/0125, G11B27/031, H04N5/9261, H04N5/85, G11B27/024, G09G2340/02, G11B27/034, H04N7/142, H04N5/772, H04N7/15, H04N5/915, G11B27/002|
|European Classification||H04N7/01F, H04N5/222, H04N7/18C, H04N7/10, G11B27/031, H04N7/01H, H04N1/00C7B, H04N5/232, H04N5/765, H04N7/15, H04N7/14A2, G06T9/00T, H04N9/804B, H04N7/14A3, H04N5/232C, G06F3/14, G11B27/34|
|30 Dec 2003||CC||Certificate of correction|
|31 Dec 2003||FPAY||Fee payment|
Year of fee payment: 8
|15 Jan 2008||FPAY||Fee payment|
Year of fee payment: 12