WO2001001348A1 - Image conversion and encoding techniques - Google Patents

Image conversion and encoding techniques Download PDF

Info

Publication number
WO2001001348A1
WO2001001348A1 PCT/AU2000/000700 AU0000700W WO0101348A1 WO 2001001348 A1 WO2001001348 A1 WO 2001001348A1 AU 0000700 W AU0000700 W AU 0000700W WO 0101348 A1 WO0101348 A1 WO 0101348A1
Authority
WO
WIPO (PCT)
Prior art keywords
depth
line
image
information
operator
Prior art date
Application number
PCT/AU2000/000700
Other languages
French (fr)
Inventor
Philip Victor Harman
Andrew Millin
Original Assignee
Dynamic Digital Depth Research Pty Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dynamic Digital Depth Research Pty Ltd. filed Critical Dynamic Digital Depth Research Pty Ltd.
Priority to AU52027/00A priority Critical patent/AU5202700A/en
Publication of WO2001001348A1 publication Critical patent/WO2001001348A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/261Image signal generators with monoscopic-to-stereoscopic image conversion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/194Transmission of image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0092Image segmentation from stereoscopic image signals

Definitions

  • the present invention is generally directed towards stereoscopic image synthesis and more particularly toward an improved method of converting two dimensional (2D) images for further encoding, transmission and decoding for the purpose of stereoscopic display.
  • a method of encoding a depth map including the steps of: a.) allocating an object number to an object; b.) allocating the object with a depth; and c.) defining the object outline
  • the object or object outline can be defined by a series of co-ordinates, and/or curves, and in particular bezier curves where found to produce desirable results.
  • This system could also be assisted through the use of generic libraries, both in the identification of objects within an image and also the allocation of depth of each object. Further, the Applicants disclosed a method of transmission of the depth map information whereby the information was included in the vertical Blanking Interval or MPEG data stream.
  • Substitute Sheet The Applicants' prior developments enabled an operator to simply outline an object and assign a depth to that object. This information was then processed to determine the amount of stretching required in order to place the object at the assigned depth. From an end user point of view this system is intuitive and allows the user to interactively alter the depth of an object to obtain an artistically pleasing result. From a processing point of view, the use of Bezier curves proves a highly efficient means of compressing the necessary data. However, whilst the use of Bezier curves is an efficient compressing technique, it is a relatively complex system to implement. Accordingly, a simplified technique of generating and processing depth maps is desired. OBJECTIVE OF INVENTION
  • the present invention provides in one aspect a method of encoding a depth map including: identifying and outlining an object within an image; allocating an object identification symbol to the object; using the allocated object symbol to represent the shape of the object; and allocating the object with a depth.
  • the present invention will include the further steps of: compressing the information representing the object and its depth; transmitting and/or storing the compressed information; and decompressing the information.
  • the present invention provides a method of encoding a depth map including: identifying and outlining an object within an image; allocating an object identification symbol to the object; defining the object by drawing a plurality of lines across the image, and determining the start and finish positions of each said line, wherein a new line is commenced each time an object boundary is reached;
  • Substitute Sheet allocating the object with a depth
  • the said lines will extend horizontally across the screen.
  • the lines could equally extend vertically across (or up and down) the screen.
  • Figure 1 shows a group of objects that have been identified, as previously disclosed and allocated a number.
  • Figure 2 shows how the shape of each identified object may be defined using the object number defined in Figure 1.
  • Figure 3 shows how the information representing this group of objects can be compressed.
  • Figure 4 shows an alternative method for representing each object and its depth.
  • the image conversion technique includes the following steps:
  • Objects in the 2D image to be converted may be identified using any of the methods previously disclosed in application PCT/AU98/01005, the contents of which are hereby included herein by reference.
  • the image to be converted is as per Figure 1.
  • This image consists of four objects, namely a disk which has been allocated object number 1 , a triangle that has been allocated object number 2, a square that has been allocated object number 3 and the background that has been allocated object number 4. It will be understood that the objects need not be allocated numerals, and that alphanumeric characters or any other symbol could also be used.
  • each of these objects within the image may be defined by using each object identification symbol, in this case a number, to "paint" the shape of the object.
  • each object identification symbol in this case a number
  • the shape of the object may still be determined such that a user or viewer may still be able to identify the painted object.
  • each line of Figure 2 is processed to produce Figure 3.
  • Figure 3 the reverse process is applied.
  • This compression technique can effectively reduce the data required to represent Figure 1 to 20% of its uncompressed value. If desired, additional compression may be applied.
  • each object needs to be allocated a depth.
  • the manner of defining the depth of an object is as previously disclosed in PCT/AU98/01005.
  • each object identified in Figure 1 will have associated with it a depth identifier in the form:
  • ⁇ depth flag> indicates the type of depth information that follows and ⁇ depth operator> are any parameters associated with the specific type of depth flag.
  • This object depth information may be added to the object data stream either before or after compression.
  • the image conversion technique includes the following steps: OBJECT IDENTIFICATION Objects in the 2D image to be converted to 3D are identified as previously described. For purposes of explanation we will use the objects identified in Figure 1. OBJECT REPRESENTATION
  • each object can be represented by drawing a series of horizontal lines across the image such that a new line is started each time an object boundary is crossed. This is illustrated in Figure 4.
  • the x, y co-ordinates of the starting point of the first line, marked Z1 in Figure 4 will be known and for illustrative purposes will be assigned 0,0.
  • the end point of the line, marked Z2 in Figure 4 will be known and for illustrative purposes will be assigned 0,255.
  • Lines (Z1 ,Z2) (Z3.Z4) Z5.Z6) represent the first few lines of object 4.
  • the values of Zn indicate the depth at which the object is to appear in the final
  • Substitute Sheet stereoscopic image For illustrative purposes we will assume that there are 255 possible depth values, such that depth 0 is closest to the viewer and depth 255 farthest from the viewer.
  • Z1 could be set to 50 and Z2 to 100.
  • the 2D to 3D conversion process previously disclosed in PCT/AU98/01005 would then interpolate the depth data and produce the appropriate 3D images.
  • Zn-1 could be set to 75 and Zn set to 100.
  • the x, y location of points Z1 to Z7 and Z12 will be known, since they border the image, the location of points Z8 to Z11 are determined by the size of the objects being described, and can be determined from the outline of the object(s).
  • the allocation of depths Za and Zb will allow a linear depth ramp to be applied to an object.
  • the centre of the radial ramp (x, y) and its radius (z) will also need to be known.
  • start_depth is relative to either the left hand edge of the image or the previous object identification and length
  • centre and radius may be required depending upon the length of the line (i.e. if the length is not equal to the image width) and assigned depth characteristics.
  • An alternative method of assigning the depth of an object is to "paint" the object using a graphics paint brush or air brush, such as used in the Adobe Photoshop graphics software, with a colour equal to the depth that is to be assigned to the object.
  • a graphics paint brush or air brush such as used in the Adobe Photoshop graphics software
  • FIG. 1 assume that object 1 is to be set to depth 10. The operator selects depth 10 from a pallet which determines a corresponding shade of grey to assign to the graphics paint brush or air brush.
  • the depth of the object can be altered if necessary by altering the shade of grey.
  • Standard techniques such as found in graphics drawing packages can be used to change the shade of grey.
  • Substitute Sheet Using a mouse, or other locating device, it is possible to add variations in object depth within an object.
  • the variations in depth are applied by selectively changing the shade of grey within the object. Examples of the depth variations include, but are not limited to, linear ramps, non-linear ramps and radial ramps. Paint the object with a shade of grey has been chosen for illustrative purposes only. In practice any colour, shape or symbol may be used to enable the operator to paint the selected object.
  • a preferred embodiment includes separating the steps of painting the object and selecting the depth of the object.
  • An effective process is to cause the object being painted to take on transparent or glass like texture. This enables the operator to see areas of the object that have already been painted rather than obscuring them with a solid colour. Once the object has been painted in this manner a solid depth colour can be applied.
  • DDC data and original 2D images can be stored and transported separately if desired.
  • the DDC data could be accessed from the same server or an entirely different server in a geographically different location. The user could then download the video sequence and DDC data relating
  • Substitute Sheet Rule 26 to the video sequence independently and combine them either at the time of viewing or prior to viewing.
  • the application of the DDC data to the original video sequence to create 3D can be undertaken in real time, as the viewer looks at the images, or off-line prior to viewing.
  • the creation of 3D from the DDC data and the original 2D image can be undertaken in either software or hardware as previously disclosed.
  • a convenient way to achieve this on a Personal Computer is to provide a software plug-in for an existing 2D viewer.
  • a plug in for the Apple QuickTime movie viewer would be suitable for general PC viewing as would plug-ins for Internet Web browser such as Internet Explorer and Netscape Communicator. These plug-ins could be most conveniently provided to the viewer by downloading over the
  • custom software application could be developed and conveniently provided to the user via a download from the Internet or other efficient method.
  • a subset of the use of DDC data for the conversion of 2D moving images into 3D is the conversion of still images.
  • the original image may be in any digital format and may have originated from either a film camera, with the resulting image being scanned to produce a digital file, or directly from a digital still or moving camera.
  • DDC data is imbedded in a standard image format.
  • Such formats are know to those skilled in the art and include, but are not limited to,
  • JPEG and GIFF are compressed formats that also enable the insertion of private data within the image files.
  • the DDC data is included in the private data area of these compressed image formats. This process therefore enables an image containing DDC data to be viewed in either 2D, if the user does not have either a suitable viewer or plug-in, or 3D if a viewer or plug-in is available.
  • DDC data may be used to create both a stereo pair of images and also a number of stereo pairs suitable for use with autostereoscopic displays using lenticular lenses.
  • An example of such a display is that manufactured by Philips which requires 7 or 9 pairs of stereo images.
  • the DDC data can be used to produce multiple monoscopic or stereoscopic views from an original 2D image. In a preferred embodiment such images could be viewed on a Personal Computer using a Windows interface with the viewpoint selected by moving a horizontal slider bar.

Abstract

A method of encoding a depth map including the steps of identifying and outlining an object within an image, allocating an object identification symbol to the object, using the allocated object symbol to represent the shape of the object, and allocating the object with a depth.

Description

IMAGE CONVERSION AND ENCODING TECHNIQUES
FIELD OF INVENTION
The present invention is generally directed towards stereoscopic image synthesis and more particularly toward an improved method of converting two dimensional (2D) images for further encoding, transmission and decoding for the purpose of stereoscopic display. BACKGROUND
The applicants have previously described in PCT/AU96/00820, a method of producing left and right eye images for a stereoscopic display from an original 2D image including the steps of: a.) identifying at least one object within the original image; b.) outlining each object; c.) defining a depth characteristic for each object; and d.) respectively displacing selected areas of each object by a determined amount in a lateral direction as a function of the depth characteristic of each object, to form two stretched images for viewing by the left and right eyes of the viewer.
These steps can be individually and collectively referred to as Dynamic Depth Cuing or DDC.
Additionally, the Applicants have previously described in PCT/AU98/0155, in one aspect a method of encoding a depth map including the steps of: a.) allocating an object number to an object; b.) allocating the object with a depth; and c.) defining the object outline
The object or object outline can be defined by a series of co-ordinates, and/or curves, and in particular bezier curves where found to produce desirable results. This system could also be assisted through the use of generic libraries, both in the identification of objects within an image and also the allocation of depth of each object. Further, the Applicants disclosed a method of transmission of the depth map information whereby the information was included in the vertical Blanking Interval or MPEG data stream.
Substitute Sheet The Applicants' prior developments enabled an operator to simply outline an object and assign a depth to that object. This information was then processed to determine the amount of stretching required in order to place the object at the assigned depth. From an end user point of view this system is intuitive and allows the user to interactively alter the depth of an object to obtain an artistically pleasing result. From a processing point of view, the use of Bezier curves proves a highly efficient means of compressing the necessary data. However, whilst the use of Bezier curves is an efficient compressing technique, it is a relatively complex system to implement. Accordingly, a simplified technique of generating and processing depth maps is desired. OBJECTIVE OF INVENTION
It is an objective of the present invention to provide a relatively simple technique of generating and processing depth maps, and further improve the operations of the Applicants earlier image conversion and encoding techniques. SUMMARY OF INVENTION
With the above objectives in mind, the present invention provides in one aspect a method of encoding a depth map including: identifying and outlining an object within an image; allocating an object identification symbol to the object; using the allocated object symbol to represent the shape of the object; and allocating the object with a depth.
Advantageously, the present invention will include the further steps of: compressing the information representing the object and its depth; transmitting and/or storing the compressed information; and decompressing the information.
In another aspect the present invention provides a method of encoding a depth map including: identifying and outlining an object within an image; allocating an object identification symbol to the object; defining the object by drawing a plurality of lines across the image, and determining the start and finish positions of each said line, wherein a new line is commenced each time an object boundary is reached;
Substitute Sheet allocating the object with a depth;
Preferably, the said lines will extend horizontally across the screen. However, it will be appreciated that the lines could equally extend vertically across (or up and down) the screen. BRIEF DESCRIPTION OF THE DRAWINGS
To provide a better understanding of the present invention, reference is made to the accompanying drawings which illustrate a preferred embodiment of the present invention. IN THE DRAWINGS: Figure 1 shows a group of objects that have been identified, as previously disclosed and allocated a number.
Figure 2 shows how the shape of each identified object may be defined using the object number defined in Figure 1.
Figure 3 shows how the information representing this group of objects can be compressed.
Figure 4 shows an alternative method for representing each object and its depth. DETAILED DESCRIPTION OF INVENTION
In the preferred embodiment, the image conversion technique includes the following steps:
OBJECT IDENTIFICATION
Objects in the 2D image to be converted may be identified using any of the methods previously disclosed in application PCT/AU98/01005, the contents of which are hereby included herein by reference. For illustrative purposes only, assume that the image to be converted is as per Figure 1. This image consists of four objects, namely a disk which has been allocated object number 1 , a triangle that has been allocated object number 2, a square that has been allocated object number 3 and the background that has been allocated object number 4. It will be understood that the objects need not be allocated numerals, and that alphanumeric characters or any other symbol could also be used. OBJECT REPRESENTATION
As shown in Figure 2 the shape of each of these objects within the image may be defined by using each object identification symbol, in this case a number, to "paint" the shape of the object. Alternatively, it can be considered that the object is "filled" with the object identification symbol. In either event the shape of the object may still be determined such that a user or viewer may still be able to identify the painted object.
As can be seen from Figure 2, object 1 has been reproduced using the symbol 1 , object 2 the symbol 2, object 3 the symbol 3 and object 4 the symbol 4. OBJECT COMPRESSION
In order to efficiently store or transmit the information representing the objects shown in Figure 1 it is necessary to compress the information. The compression and decompression process should be able to be undertaken both in software and/or hardware and operate in a fast and efficient manner. In the case of the 2D to 3D conversion of video, where the object information is included in the Vertical Blanking Interval or MPEG stream, it is desirable that the decompression should operate sufficiently rapidly to enable real time 2D to 3D conversion of the associated 2D image.
A simple compression technique is to run length encode for each line of Figure 2. This is shown in Figure 3. The first line of Figure 2 consists of the symbol 4 repeated 50 times. Using run length encoding this would compress to
4(50) which indicates that the symbol 4 is to be repeated 50 times along that line.
In a similar manner, each line of Figure 2 is processed to produce Figure 3. To decompress Figure 3 the reverse process is applied.
This compression technique can effectively reduce the data required to represent Figure 1 to 20% of its uncompressed value. If desired, additional compression may be applied.
It will be appreciated by those skilled in the art that many other data compression techniques are suitable for this application. This would include, although not limited to, Lenple Ziv, Huffman and Shannon-Fano. Once an object has been defined using the above process a differential encoding technique can be used to eliminate the need to transfer or store data that is consistent in consecutive images. The differentially encoded data can then be compressed as previously described. DEPTH MAPS
As previously disclosed, in order to use this representation of an object to form a depth map each object needs to be allocated a depth. The manner of defining the depth of an object is as previously disclosed in PCT/AU98/01005.
Therefore, each object identified in Figure 1 will have associated with it a depth identifier in the form:
<object number> <depth flag> <depth operator>
Where <depth flag> indicates the type of depth information that follows and <depth operator> are any parameters associated with the specific type of depth flag. This object depth information may be added to the object data stream either before or after compression.
In an alternative embodiment, the image conversion technique includes the following steps: OBJECT IDENTIFICATION Objects in the 2D image to be converted to 3D are identified as previously described. For purposes of explanation we will use the objects identified in Figure 1. OBJECT REPRESENTATION
The shape of each object can be represented by drawing a series of horizontal lines across the image such that a new line is started each time an object boundary is crossed. This is illustrated in Figure 4.
The x, y co-ordinates of the starting point of the first line, marked Z1 in Figure 4 will be known and for illustrative purposes will be assigned 0,0. Likewise the end point of the line, marked Z2 in Figure 4, will be known and for illustrative purposes will be assigned 0,255.
Lines (Z1 ,Z2) (Z3.Z4) Z5.Z6) represent the first few lines of object 4. The values of Zn indicate the depth at which the object is to appear in the final
Substitute Sheet stereoscopic image. For illustrative purposes we will assume that there are 255 possible depth values, such that depth 0 is closest to the viewer and depth 255 farthest from the viewer.
If, for example, object 4 is to appear at a constant depth 100 from the viewer then the values of Z1 and Z2 would be 100.
If, however, it was desired that object 4 should appear to ramp away from the viewer then Z1 could be set to 50 and Z2 to 100. The 2D to 3D conversion process previously disclosed in PCT/AU98/01005 would then interpolate the depth data and produce the appropriate 3D images. Conversely, should it be desired that object 4 should appear to both ramp away from the viewer and tilt away then Zn-1 could be set to 75 and Zn set to 100.
Referring to Figure 4, the line identified as starting at Z7 would therefore consist of 3 segments as follows:
Z7 to Z8 - which defines the segment of object 4; Z9 to Z10 - which defines the segment of object 3;
Z11 to Z12 - which defines the segment of object 4
Whilst the x, y location of points Z1 to Z7 and Z12 will be known, since they border the image, the location of points Z8 to Z11 are determined by the size of the objects being described, and can be determined from the outline of the object(s). The allocation of depths Za and Zb will allow a linear depth ramp to be applied to an object. However, since it is desirable to add other than linear ramps to an object, for example radial ramps, the centre of the radial ramp (x, y) and its radius (z) will also need to be known.
Thus the general format for a line describing both an objects length and assigned depth is start_depth, end_depth (length, centre (x, y), radius)
Where start_depth is relative to either the left hand edge of the image or the previous object identification and length, centre and radius may be required depending upon the length of the line (i.e. if the length is not equal to the image width) and assigned depth characteristics.
For example, using start_depths relative to the previous object, and assuming a 0 to 255 by 1 to 255 image with object 4 at a constant depth 50 and
Substitute Sheet object 3 at a constant depth 20, and assuming that object 3 runs from co-ordinate (76,4) to (200,4^ the data for the first 4 lines of Figure 4 will be: Line 1 50
Line 2 50 Line 3 50
Line 4 50, 50, 75 20, 20 125 50, 50, 55
This data can be compressed as previously described or using other compression techniques familiar to those skilled in the art. Similarly, the allocation of a depth for each object is as previously disclosed. ALTERNATIVE PROCESS FOR DEPTH ASSIGNMENT
An alternative method of assigning the depth of an object is to "paint" the object using a graphics paint brush or air brush, such as used in the Adobe Photoshop graphics software, with a colour equal to the depth that is to be assigned to the object. For illustrative purposes assume that there are 256 different depth levels available with 0 representing an object closest to the observer and 255 furthest away from the observer. Assume that depth 0 is assigned to white and 255 to black then intermediate depths will be assigned a shade of grey.
Considering Figure 1 , assume that object 1 is to be set to depth 10. The operator selects depth 10 from a pallet which determines a corresponding shade of grey to assign to the graphics paint brush or air brush.
The operator then paints the area of object 1 with the brush, thus allocating depth 10 to the object. Since the operator can see the results of his actions on the computer screen it is possible to accurately select objects. Different size brushes, or air brush patterns, may be selected to allow fine detail or rapid fill of large areas. Errors can be corrected using a graphical eraser such as found, for example, in Adobe Photoshop.
Once the shape of an object has been painted using the brush the depth of the object can be altered if necessary by altering the shade of grey. Standard techniques such as found in graphics drawing packages can be used to change the shade of grey.
Substitute Sheet Using a mouse, or other locating device, it is possible to add variations in object depth within an object. The variations in depth are applied by selectively changing the shade of grey within the object. Examples of the depth variations include, but are not limited to, linear ramps, non-linear ramps and radial ramps. Painting the object with a shade of grey has been chosen for illustrative purposes only. In practice any colour, shape or symbol may be used to enable the operator to paint the selected object.
A preferred embodiment includes separating the steps of painting the object and selecting the depth of the object. An effective process is to cause the object being painted to take on transparent or glass like texture. This enables the operator to see areas of the object that have already been painted rather than obscuring them with a solid colour. Once the object has been painted in this manner a solid depth colour can be applied.
ALTERNATIVE PROCESS FOR CREATING BEZIER CURVES Since the peripheral edges of the object are known it is possible to locate the boundary of the object and automatically construct Bezier curves that trace the peripheral of the object. The Bezier curves may be adjusted as necessary to cause exact alignment of the object and also adjusted to allow for movement of the object as it moves in successive frames. ALTERNATIVE PROCESS FOR TRANSPORT OF DDC DATA
In previous applications, it has been assumed that the Dynamic Depth
Cueing (DDC) data has been imbedded in the original 2D image in either the
Vertical Blanking Interval or MPEG stream.
In practice this is a convenient way to transport and store the DDC encoded 2D images. However, it will be understood by to those skilled in the art that the
DDC data and original 2D images can be stored and transported separately if desired.
For example, if the original 2D image where a video sequence that was held on a data server accessible via the Internet then the DDC data could be accessed from the same server or an entirely different server in a geographically different location. The user could then download the video sequence and DDC data relating
Substitute Sheet Rule 26 to the video sequence independently and combine them either at the time of viewing or prior to viewing.
The application of the DDC data to the original video sequence to create 3D can be undertaken in real time, as the viewer looks at the images, or off-line prior to viewing. The creation of 3D from the DDC data and the original 2D image can be undertaken in either software or hardware as previously disclosed.
In relation to software conversion from 2D to 3D, a convenient way to achieve this on a Personal Computer (PC) is to provide a software plug-in for an existing 2D viewer. For example, a plug in for the Apple QuickTime movie viewer would be suitable for general PC viewing as would plug-ins for Internet Web browser such as Internet Explorer and Netscape Communicator. These plug-ins could be most conveniently provided to the viewer by downloading over the
Internet.
Alternatively a custom software application could be developed and conveniently provided to the user via a download from the Internet or other efficient method.
USE OF DDC FOR THE 2D TO 3D CONVERSION OF STILL IMAGES
A subset of the use of DDC data for the conversion of 2D moving images into 3D is the conversion of still images. The original image may be in any digital format and may have originated from either a film camera, with the resulting image being scanned to produce a digital file, or directly from a digital still or moving camera.
For transfer between Personal Computers and in particular transfer over the
Internet, it is desirable that the DDC data is imbedded in a standard image format. Such formats are know to those skilled in the art and include, but are not limited to,
JPEG, BMP, GIFF and TIFF.
JPEG and GIFF are compressed formats that also enable the insertion of private data within the image files. In a preferred embodiment the DDC data is included in the private data area of these compressed image formats. This process therefore enables an image containing DDC data to be viewed in either 2D, if the user does not have either a suitable viewer or plug-in, or 3D if a viewer or plug-in is available.
Substitute Sheet As previously disclosed, DDC data may be used to create both a stereo pair of images and also a number of stereo pairs suitable for use with autostereoscopic displays using lenticular lenses. An example of such a display is that manufactured by Philips which requires 7 or 9 pairs of stereo images. For use with a still image, the DDC data can be used to produce multiple monoscopic or stereoscopic views from an original 2D image. In a preferred embodiment such images could be viewed on a Personal Computer using a Windows interface with the viewpoint selected by moving a horizontal slider bar.
Whilst the method and system of the present invention has been summarised and explained by illustrative examples, it will be appreciated by those skilled in the art that many widely varying embodiments and applications are within the teaching and scope of the present invention, and that the examples presented herein are by way of illustration only and should not be construed as limiting the scope of this invention.
Substitute Sheet

Claims

THE CLAIMS DEFINING THE INVENTION ARE AS FOLLOWS:
1. A method of encoding a depth map including: identifying and outlining an object within an image; allocating an object identification symbol to said object; using the allocated object symbol to represent the shape of said object; and allocating said object with a depth.
2. A method as claimed in claim 1 wherein said object identification symbol is an alphanumeric character.
3. A method as claimed in claim 1 wherein said object identification symbol is a symbol.
4. A method as claimed in any one of the preceding claims wherein said object is filled with said object identification symbol.
5. A method as claimed in any preceding claim further including the step of: compressing the information representing said object and its depth.
6. A method as claimed in claim 5 further including the steps of: transmitting and/or storing said compressed information; and decompressing said compressed information.
7. A method as claimed in claim 5 or 6 wherein said decompressing of the information is in real time.
8. A method as claimed in any one of claims 5 to 7 wherein said step of compressing the information uses run length encode for each line of said image.
Substitute Sh
9. A method as used in any one of claims 5 to 7 wherein each line of said image is compressed by representing the information in the following format:
A(X)... wherein A represents the object identification symbol, and X represents the number of consecutive pixels that A covers.
10. A method as claimed in any one of claims 5 to 9 wherein said compressed information is further compressed.
11. A method as claimed in claim 10 wherein said further compression uses Lenple Ziv, Huffman or Shannon-Fano techniques.
12. A method as claimed in any preceding claim further including the step of: assigning each said object a depth identifier.
13. A method as claimed in claim 12 wherein said depth identifier is of the form: <object number> <depth flag> <depth operator> wherein object number is the number allocated to the object, depth flag indicates the type of depth information, and depth operator are parameters associated with said depth information.
14. A method of encoding a depth map including the steps of: identifying and outlining an object within an image; allocating an object identification symbol to said object; defining said object by drawing a plurality of lines across said image, and determining the start and finish positions of each said line, wherein a new line is commenced each time an object boundary is reached; and allocating said object with a depth;
15. A method as claimed in claim 14, wherein said lines extend horizontally across said screen.
Substit
16. A method as claimed in claim 14, wherein said lines extend vertically down said screen.
17. A method as claimed in any one of claims 14 to 16 wherein each said line is defined in the following format:
A,B(x) wherein A is the depth at the start of the line, B is the depth at the end of the line, and x is the length of the line.
18. A method as claimed in any one of claims 14 to 16 wherein each said line is defined in the following format:
A,B(x,C(y,z),r) wherein A is the depth at the start of the line, B is the depth at the end of the line, x is the length of the line, C(y,z) is the centre point of a radial ramp, and r is the radius of said radial ramp.
19. A method as claimed in any preceding claim wherein said objects are painted a colour by an operator to assign a depth to said object, said colour being predefined for a certain depth.
20. A method as claimed in claim 19 wherein the colour assigned by said operator is a shade of grey ranging from white to black.
21. A method as claimed in any one of claims 1 to 18 wherein said objects are painted a colour by an operator, and then said colour is assigned a depth value.
22. A method as claimed in any preceding claim wherein said depth map is attached to or embedded in said image.
Sub
3. A method or system substantially as hereinbefore described with reference the accompanying drawings.
Substitute Sheet
PCT/AU2000/000700 1999-06-25 2000-06-22 Image conversion and encoding techniques WO2001001348A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU52027/00A AU5202700A (en) 1999-06-25 2000-06-22 Image conversion and encoding techniques

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
AUPQ1197 1999-06-25
AUPQ1197A AUPQ119799A0 (en) 1999-06-25 1999-06-25 Ddc/3 image conversion and encoding techniques

Publications (1)

Publication Number Publication Date
WO2001001348A1 true WO2001001348A1 (en) 2001-01-04

Family

ID=3815387

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/AU2000/000700 WO2001001348A1 (en) 1999-06-25 2000-06-22 Image conversion and encoding techniques

Country Status (2)

Country Link
AU (1) AUPQ119799A0 (en)
WO (1) WO2001001348A1 (en)

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004014086A1 (en) * 2002-07-31 2004-02-12 Koninklijke Philips Electronics N.V. Method and apparatus for encoding a digital video signal
US7116324B2 (en) 1998-05-27 2006-10-03 In-Three, Inc. Method for minimizing visual artifacts converting two-dimensional motion pictures into three-dimensional motion pictures
WO2009113761A1 (en) 2008-03-12 2009-09-17 Samsung Electronics Co., Ltd. Image processing method and apparatus, image reproducing method and apparatus, and recording medium
EP2357836A3 (en) * 2002-03-27 2012-03-07 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8254668B2 (en) 2002-03-27 2012-08-28 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8831273B2 (en) 2010-09-10 2014-09-09 Reald Inc. Methods and systems for pre-processing two-dimensional image files to be converted to three-dimensional image files
US9007365B2 (en) 2012-11-27 2015-04-14 Legend3D, Inc. Line depth augmentation system and method for conversion of 2D images to 3D images
US9007404B2 (en) 2013-03-15 2015-04-14 Legend3D, Inc. Tilt-based look around effect image enhancement method
US9241147B2 (en) 2013-05-01 2016-01-19 Legend3D, Inc. External depth map transformation method for conversion of two-dimensional images to stereoscopic images
US9282321B2 (en) 2011-02-17 2016-03-08 Legend3D, Inc. 3D model multi-reviewer system
US9288476B2 (en) 2011-02-17 2016-03-15 Legend3D, Inc. System and method for real-time depth modification of stereo images of a virtual reality environment
US9286941B2 (en) 2001-05-04 2016-03-15 Legend3D, Inc. Image sequence enhancement and motion picture project management system
US9407904B2 (en) 2013-05-01 2016-08-02 Legend3D, Inc. Method for creating 3D virtual reality from 2D images
US9438878B2 (en) 2013-05-01 2016-09-06 Legend3D, Inc. Method of converting 2D video to 3D video using 3D object models
US9485497B2 (en) 2010-09-10 2016-11-01 Reald Inc. Systems and methods for converting two-dimensional images into three-dimensional images
US9547937B2 (en) 2012-11-30 2017-01-17 Legend3D, Inc. Three-dimensional annotation system and method
US9609307B1 (en) 2015-09-17 2017-03-28 Legend3D, Inc. Method of converting 2D video to 3D video using machine learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4947347A (en) * 1987-09-18 1990-08-07 Kabushiki Kaisha Toshiba Depth map generating method and apparatus
US5793900A (en) * 1995-12-29 1998-08-11 Stanford University Generating categorical depth maps using passive defocus sensing
WO1999030280A1 (en) * 1997-12-05 1999-06-17 Dynamic Digital Depth Research Pty. Ltd. Improved image conversion and encoding techniques

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4947347A (en) * 1987-09-18 1990-08-07 Kabushiki Kaisha Toshiba Depth map generating method and apparatus
US5793900A (en) * 1995-12-29 1998-08-11 Stanford University Generating categorical depth maps using passive defocus sensing
WO1999030280A1 (en) * 1997-12-05 1999-06-17 Dynamic Digital Depth Research Pty. Ltd. Improved image conversion and encoding techniques

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7116324B2 (en) 1998-05-27 2006-10-03 In-Three, Inc. Method for minimizing visual artifacts converting two-dimensional motion pictures into three-dimensional motion pictures
US9286941B2 (en) 2001-05-04 2016-03-15 Legend3D, Inc. Image sequence enhancement and motion picture project management system
US8577128B2 (en) 2002-03-27 2013-11-05 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
EP2357836A3 (en) * 2002-03-27 2012-03-07 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8724886B2 (en) 2002-03-27 2014-05-13 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8879824B2 (en) 2002-03-27 2014-11-04 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8559703B2 (en) 2002-03-27 2013-10-15 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8254668B2 (en) 2002-03-27 2012-08-28 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8577127B2 (en) 2002-03-27 2013-11-05 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8369607B2 (en) 2002-03-27 2013-02-05 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8417024B2 (en) 2002-03-27 2013-04-09 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
US8472702B2 (en) 2002-03-27 2013-06-25 Sanyo Electric Co., Ltd. Method and apparatus for processing three-dimensional images
WO2004014086A1 (en) * 2002-07-31 2004-02-12 Koninklijke Philips Electronics N.V. Method and apparatus for encoding a digital video signal
US8270477B2 (en) 2002-07-31 2012-09-18 Koninklijke Philips Electronics N.V. Method and apparatus for encoding a digital video signal
KR101008525B1 (en) * 2002-07-31 2011-01-14 코닌클리케 필립스 일렉트로닉스 엔.브이. Method of encoding a digital video sequence, a computer-readable recording medium having recorded thereon a computer program for an encoder, a computer-readable recording medium having recorded thereon a computer program for a computer, an encoder for encoding a digital video sequence, and a video communication system
CN1672432B (en) * 2002-07-31 2011-06-15 皇家飞利浦电子股份有限公司 Method for encoding a digital video signal and apparatus
EP2253145A4 (en) * 2008-03-12 2013-04-17 Samsung Electronics Co Ltd Image processing method and apparatus, image reproducing method and apparatus, and recording medium
US8849009B2 (en) 2008-03-12 2014-09-30 Samsung Electronics Co., Ltd. Image processing method and apparatus, image reproducing method and apparatus, and recording medium
EP2253145A1 (en) * 2008-03-12 2010-11-24 Samsung Electronics Co., Ltd. Image processing method and apparatus, image reproducing method and apparatus, and recording medium
WO2009113761A1 (en) 2008-03-12 2009-09-17 Samsung Electronics Co., Ltd. Image processing method and apparatus, image reproducing method and apparatus, and recording medium
US9485497B2 (en) 2010-09-10 2016-11-01 Reald Inc. Systems and methods for converting two-dimensional images into three-dimensional images
US8831273B2 (en) 2010-09-10 2014-09-09 Reald Inc. Methods and systems for pre-processing two-dimensional image files to be converted to three-dimensional image files
US9282321B2 (en) 2011-02-17 2016-03-08 Legend3D, Inc. 3D model multi-reviewer system
US9288476B2 (en) 2011-02-17 2016-03-15 Legend3D, Inc. System and method for real-time depth modification of stereo images of a virtual reality environment
US9007365B2 (en) 2012-11-27 2015-04-14 Legend3D, Inc. Line depth augmentation system and method for conversion of 2D images to 3D images
US9547937B2 (en) 2012-11-30 2017-01-17 Legend3D, Inc. Three-dimensional annotation system and method
US9007404B2 (en) 2013-03-15 2015-04-14 Legend3D, Inc. Tilt-based look around effect image enhancement method
US9241147B2 (en) 2013-05-01 2016-01-19 Legend3D, Inc. External depth map transformation method for conversion of two-dimensional images to stereoscopic images
US9407904B2 (en) 2013-05-01 2016-08-02 Legend3D, Inc. Method for creating 3D virtual reality from 2D images
US9438878B2 (en) 2013-05-01 2016-09-06 Legend3D, Inc. Method of converting 2D video to 3D video using 3D object models
US9609307B1 (en) 2015-09-17 2017-03-28 Legend3D, Inc. Method of converting 2D video to 3D video using machine learning

Also Published As

Publication number Publication date
AUPQ119799A0 (en) 1999-07-22

Similar Documents

Publication Publication Date Title
EP3751857A1 (en) A method, an apparatus and a computer program product for volumetric video encoding and decoding
WO2001001348A1 (en) Image conversion and encoding techniques
US7054478B2 (en) Image conversion and encoding techniques
CN1462416A (en) Image conversion and encoding technique
CN1331822A (en) System and method for creating 3D models from 2D sequential image data
EP1034661B1 (en) Low bandwidth encoding scheme for video transmission
KR20010074927A (en) Hierarchical foveation and foveated coding of images based on wavelets
WO2021136878A1 (en) A method, an apparatus and a computer program product for volumetric video encoding and decoding
WO2021207582A1 (en) System and method for generating light field images
JP7344988B2 (en) Methods, apparatus, and computer program products for volumetric video encoding and decoding
US7439976B2 (en) Visual communication signal
WO2021053261A1 (en) A method, an apparatus and a computer program product for video encoding and video decoding
AU738692B2 (en) Improved image conversion and encoding techniques
EP3703013A1 (en) Apparatus and method of generating an image signal
WO2022074286A1 (en) A method, an apparatus and a computer program product for video encoding and video decoding
JPH01181292A (en) Color picture transmitter
JPH10234035A (en) Image encoding/decoding method and device
JP2001331246A (en) Picture associated data display device
Henry Using JPEG2000 compressed images for foveated texture mapping in limited memory environments
LOIS et al. Multiview Video Presentation and Encoding
MXPA00005355A (en) Improved image conversion and encoding techniques

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

WWE Wipo information: entry into national phase

Ref document number: 10018781

Country of ref document: US

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP