US9792012B2 - Method relating to digital images - Google Patents

Method relating to digital images Download PDF

Info

Publication number
US9792012B2
US9792012B2 US13/499,711 US201013499711A US9792012B2 US 9792012 B2 US9792012 B2 US 9792012B2 US 201013499711 A US201013499711 A US 201013499711A US 9792012 B2 US9792012 B2 US 9792012B2
Authority
US
United States
Prior art keywords
digital image
image representation
sensitive display
touch
touch sensitive
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/499,711
Other versions
US20120262490A1 (en
Inventor
Sami Niemi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nokia Technologies Oy
Original Assignee
Mobile Imaging in Sweden AB
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mobile Imaging in Sweden AB filed Critical Mobile Imaging in Sweden AB
Assigned to SCALADO AB reassignment SCALADO AB ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NIEMI, SAMI
Assigned to MOBILE IMAGING IN SWEDEN AB reassignment MOBILE IMAGING IN SWEDEN AB ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SCALADO AB
Publication of US20120262490A1 publication Critical patent/US20120262490A1/en
Application granted granted Critical
Publication of US9792012B2 publication Critical patent/US9792012B2/en
Assigned to NOKIA TECHNOLOGIES OY reassignment NOKIA TECHNOLOGIES OY ASSET TRANSFER AGREEMENT Assignors: MOBILE IMAGING IN SWEDEN AB
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/041Indexing scheme relating to G06F3/041 - G06F3/045
    • G06F2203/04104Multi-touch detection in digitiser, i.e. details about the simultaneous detection of a plurality of touching locations, e.g. multiple fingers or pen and finger
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04805Virtual magnifying lens, i.e. window or frame movable on top of displayed information to enlarge it for better reading or selection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen

Definitions

  • the present invention relates to a method for selecting an image and a method for facilitating generation or capturing of a desired image.
  • a great number of the graphical or photographical images of today are generated digitally. Generally this results in that more images are created and many times a greater amount of undesirable images are created.
  • One of the problems of today is that even undesired images are stored and thereby occupy storage capacity.
  • One simple solution to this problem is to delete all undesired images.
  • the likelihood of undesired images still occupying storage capacity becomes greater as the time pass from the time an image was created.
  • a method for selecting an image comprises forming a group of digital image representations, displaying a first digital image representation of the group of digital image representations on a touch sensitive display, generating a position signal in response to a detection of a pointing device on the touch sensitive display, said position signal indicating a touch position, identifying a selected position within the displayed first digital image representation based on the position signal, generating a zoom-in signal in response to a detection of the pointing device sliding away from the touch position on the touch display, said zoom signal indicating a sliding distance from the touch position, displaying an enlarged representation of the first digital image representation in response to the zoom-in signal, generating a shift signal in response to a detection of a second sliding motion of the pointing device on the touch sensitive display, and displaying a second digital image representation and an enlarged representation of the second digital image representation in response to the shift signal, the enlargement of the second digital image representation being based on the zoom signal generated during displaying of the first digital
  • the advantage of forming a group of images is that the chance of having at least one good image is increased.
  • the selection method the selection of the best image is facilitated and thereby it becomes easier for a user to discard less desirable images.
  • the zoom-in facilitates viewing of details in the high resolution image on a smaller preview screen.
  • the sliding direction of the second sliding motion is along a trajectory that is substantially circular. This is an advantage in that the sliding motion determines the switching between images and if the number of images in the group of image representations are large, the touch sensitive display may not be big enough. However, by making the sliding motion circular there is no longer any limitations as it becomes possible to slide the pointing device many turns.
  • the touch sensitive display is a multi-touch sensitive display, wherein said generating of a position signal further includes a detection of a second pointing device on the multi-touch sensitive display, said position signal indicating a touch position which is based on each position of the two pointing devices respectively, wherein said zoom-in signal is generated in response to a detection of the two pointing devices sliding on the multi-touch display away from each other, and wherein said shift signal is generated in response to a second sliding motion of the two pointing devices at substantially constant distance between the pointing devices.
  • the touch position is a position between the two detected pointing devices.
  • the act of displaying an enlarged representation of the first image includes displaying an enlarged subarea of the first digital image representation in response to the zoom-in signal, the position of the subarea within the first digital image representation being based on the selected position and the enlargement of the subarea being based on the zoom-in signal.
  • the size of the subarea to be enlarged is based on the distance between the two pointing devices at the generation of the position signal.
  • said forming of a group of digital image representations includes capturing a plurality of different digital image representations of essentially the same view at different points in time.
  • said forming of a group of digital image representations includes capturing a plurality of different digital image representations of essentially the same view at different exposure settings.
  • said forming of a group of digital image representations includes capturing a plurality of different digital image representations of essentially the same view having different focus distances.
  • said forming of a group of digital image representations includes capturing a plurality of different digital image representations of essentially the same view being exposed for different transforms.
  • said forming of a group of digital image representations includes generating a plurality of digital image representations from one single original image by manipulating the original image differently for each digital, the manipulation includes applying a transform or a parameter to the original image.
  • a pointing device is a fingertip.
  • FIG. 1 a is a schematic view of a display side of an image presentation device according to one embodiment of the invention
  • FIG. 1 b is a schematic view of a lens side of an image presentation device incorporating a camera according to one embodiment of the invention
  • FIG. 1 c is a schematic block diagram of the image presentation device in FIGS. 1 a - b,
  • FIG. 2 shows a group of image representations according to one embodiment of the invention
  • FIG. 3 a - d schematically shows acts of a method according to one embodiment of the invention.
  • FIGS. 4 a - d are representations of example displays according example embodiments.
  • an image presentation device 10 for implementing the invention is showed.
  • the image presentation device 10 includes housing 12 and a touch sensitive display 14 , see FIG. 1 a . It may also include a lens 16 , see FIG. 1 b , for focusing light to be captured as an image on an image sensor, not showed.
  • the image presentation device 10 may, thus, be included in a camera or have a camera incorporated.
  • the image presentation device 10 includes processor 18 , a volatile memory 20 , a non volatile memory 22 , a display driver 24 , a touch screen driver 26 , touch screen circuitry 28 , and camera circuitry 30 including an image sensor.
  • the processor 18 , the volatile memory 20 and the non volatile memory 22 may be arranged and connected in a way known to the skilled person for operation of the image presentation device and execution of applications stored in the non-volatile memory 22 .
  • the design and implementation of the touch screen circuitry 28 depends on the type of touch sensitive display that is to be used.
  • the implementation of the touch screen driver 26 depends on the type of touch sensitive display and the operating system of the image presentation device 10 .
  • touch sensitive display or touch screen is used for a display that is arranged to detect the presence, location, and/or movement of a “touch” within the display area.
  • the touch screen may be designed to detect presence, location, and/or movement on the display by a finger, a hand, a stylus, a pen, etc.
  • the touch screen be a resistive touch screen, a touch screen based on surface acoustic wave technology, a capacitive touch screen, a touch screen using surface capacitance, a touch screen based on projected capacitive touch technology, a system based on infrared LEDs and photo sensors, a system based on a strain gauge configuration, a touch screen based on dispersive signal technology, a touch screen based on acoustic pulse recognition technology, etc.
  • a method for selecting images is part of a greater scheme of achieving a desired image having specific characteristics.
  • the embodiment relates to an image selecting method operating on a group of digital image representations in order to achieve this result.
  • the images forming the group of image representations may be images retrieved from a storage device, e.g. a hard drive, the non volatile memory 22 , an image server accessed via a network, etc.
  • the images may alternatively be acquired by means of a camera arranged in the device 10 or by means of transforming one original image retrieved from a storage device or one original image acquired by said camera.
  • the image sequence may also be calculated from one or more source images, and the image itself may be virtual representation based on one or more mathematical schemes applied on one or more original images.
  • One example of how to generate the group of image representations are to bracket, i.e. to take photographs at more than one exposure in order to ensure that the desired exposure is obtained in at least one exposure.
  • Other examples are to take a plurality of photographs at different points in time, different depth of field, at different focus distances, or by varying any other setting of the camera.
  • the camera used in these examples may well be a camera implemented in the same device or system as the image presentation device 10 .
  • the group of image representation may be generated from applying different transforms to the images.
  • the number of images in a group of image representations may be as few as two and as many as hundreds, it much depends on the application in which the method is planned to be used.
  • the group of images may be separate images/photos or a sequence of frames in a video.
  • FIG. 2 a group of image representations 50 including three image representations is showed.
  • this particular sequence of images represents photographs 52 , 54 , 56 , also referred to as first, second and third photographs, taken at different points in time.
  • the selection method may be started in many ways, e.g. by pressing a button on the image presentation device 10 , by touching the touch sensitive display 14 by means of a pointing device either within a displayed image representation from the group of image representations or within an area on the display indicating a button for activiating this functionality.
  • FIG. 3 shows one embodiment of the invention by depicting subsequent acts of the selection method of this embodiment.
  • the image presentation device 10 displays a representation of the second photograph 54 , see FIG. 3 a , from the group of image representations 50 .
  • the pointing device 70 is positioned to touch the touch sensitive display 14 at position 72 , see FIG. 3 b , and a position signal is generated, including the touch position, and is sent from the touch sensitive display 14 to the processor 18 , i.e. from the touch screen circuitry 28 and the touch screen driver 26 to the processor 18 .
  • the touch sensitive display 14 detects a sliding motion 74 , performed by means of the pointing device 70 , along the display 14 away from the touch position 72 .
  • This detection results in the generation of a zoom signal that is sent to the processor 18 .
  • the zoom signal includes an indication of the distance of the sliding movement 74 , referred to as zoom-value.
  • zoom-value an indication of the distance of the sliding movement 74 .
  • the displayed image representation 54 is enlarged to a degree that is based on the zoom value, see FIG. 3 c . In the embodiment of FIG. 3 c the enlarging of the image representation is centered at the touch position 72 .
  • the pointing device is moved in another direction 76 , in this specific embodiment essentially perpendicular to previous movement 74 .
  • This second movement 76 may advantageously be following a circular trajectory, as indicated by 76 in FIG. 3 d , in order to facilitating shifting through a large group of image representation without running out of touch sensitive display surface to slide on.
  • the second sliding motion 76 is detected by the touch sensitive display 14 and a shift signal is generated and sent to the processor 18 .
  • the image displayed 54 is shifted to another image representation 56 in the group of image representations 50 .
  • the third photograph 56 is shifted in for display.
  • the same enlargement is applied to the newly shifted image representation as for the previous.
  • the number of images shifted may be proportional to the length of the second sliding motion 76 .
  • the shift would continue to present the image representation of the first photograph if the second sliding motion 76 is continued.
  • the enlargement is not applied to the entire image representation as depicted in FIGS. 3 c and 3 d , but rather to a predetermined area surrounding the touch position indicated by the position signal.
  • the area may be defined by a radius and may be substantially circular.
  • two pointing devices are used, e.g. a finger and the thumb of a users hand.
  • a position signal is generated when the two pointing devices are detected on the touch sensitive display, the touch position being indicated as a position between the detection points of the two pointing devices.
  • a zoom signal is generated and in response to the zoom signal an enlarged image representation of the image representation presently displayed is presented on the display.
  • the degree of enlargement is based on the distance the two pointing devices have been sliding away from each other.
  • not the entire image representation is zoomed but only a subarea.
  • the size of this sub area may correspond to an area defined by the initial positions of the pointing devices, i.e. when the touch position is indicated.
  • a shift signal is generated and in response to the shift signal the image representation displayed is shifted to another image representation from the group of image representations.
  • they are rotated substantially around a position in-between the two pointing devices and at a substantially constant distance from each other, e.g. following a substantially circular trajectory.
  • the length of the sliding motion determines which image representation from the group of image representations to display. The enlargement applied to the initial image is displayed in the shifted images as well.
  • portions of the initially displayed image representation is shifted.
  • the portion to be shifted may for instance be indicated manually by tracing the contours of the area and then the shifting results in that the corresponding area of another image from the group of image representations is displayed.
  • the contours of the area can also be automatically computed by tracing where the two images', aligned to substantially same positions within the particular area, pixels are substantially similar around the traced contour.
  • the method may advantageously be used for browsing images.
  • the zoom-in step may be skipped and a rotational/circular motion using one or two pointing devices may be used to switch images.
  • the group of image representations probably includes the images of a folder in a file system or a database or of a particular category in a database.
  • the image presentation device 10 is a mobile telephone equipped with a camera.
  • FIG. 4 yet another embodiment is shown in FIG. 4 .
  • the image presentation device 10 displays a representation of the second photograph 54 , see FIG. 4 a , from the group of image representations 50 .
  • the pointing device 70 is positioned to touch the touch sensitive display 14 at a first position 82 , see FIG. 4 b , and a position signal is generated, including the touch position, and is sent from the touch sensitive display 14 to the processor 18 , i.e. from the touch screen circuitry 28 and the touch screen driver 26 to the processor 18 .
  • the displayed image representation 54 is enlarged to a degree that is based on a predetermined zoom value, see FIG. 4 b .
  • the enlarging of the image representation is centered at the touch position.
  • the enlargement is not applied to the entire image representation but rather to a predetermined area 81 surrounding the touch position indicated by the position signal.
  • the predetermined area 81 is defined by a radius and is substantially circular.
  • the touch position including the position signal is updated.
  • the updated position signal is sent from the touch sensitive display 14 to the processor 18 , i.e. from the touch screen circuitry 28 and the touch screen driver 26 to the processor 18 .
  • the predetermined area 81 of the image representation that is enlarged in this case image representation of the second photograph 54 , is also updated so that the enlargement is applied to the predetermined area 81 surrounding the touch position indicated by the position signal.
  • the predetermined area 81 of the image representation that is enlarged may be live updated whilst the pointing device 70 is moved from the first position 82 to the second position 84 .
  • the predetermined area 81 of the image representation that is enlarged is live updated.
  • the size of the predetermined area 81 may be reduced or enlarged upon detection of two pointing devices sliding towards each other or away from each other, respectively. This may e.g. be done by pointing the two pointing devices on to opposite sections on the border of the predetermined area 81 , respectively, and then sliding the two pointing devices towards each other or away from each other.
  • the image representation shown in the predetermined area 81 may be shifted.
  • the pointing device 70 is moved in a direction 86 along the border of the enlarged area of the image representation.
  • a substantially circular trajectory as indicated by a sliding motion 86 in FIG. 4 d , this in order to facilitating shifting through a large group of image representation without running out of touch sensitive display surface to slide on.
  • the sliding motion 86 is detected by the touch sensitive display 14 and a shift signal is generated and sent to the processor 18 .
  • the image displayed in the enlarged predetermined area 81 is shifted to another image representation in the group of image representations 50 , in this case to the image representation of the third photograph 56 .
  • the same enlargement is applied to the newly shifted image representation as for the previous.
  • the number of images shifted may be proportional to the length of the sliding motion 86 .
  • the not selected digital image representations of the group of digital image representations may be discarded.
  • the shifting of the image representations may be performed by moving the pointing device 70 in a substantially linear manner.
  • This embodiment may be developed even further by showing small miniatures of the image representations at e.g. the bottom of the touch sensitive display 14 .
  • the shifting of the image representations may then be performed by moving the small miniatures by means of the pointing device 70 in a substantially linear manner along a path that is substantially parallel with the lower border of the touch sensitive display 14 .

Abstract

The present invention is related to a method for selecting an image comprises forming a group of digital image representations, displaying a first digital image representation of the group of digital image representations on a touch sensitive display, generating a position signal in response to a detection of a pointing device on the touch sensitive display, said position signal indicating a touch position, identifying a selected position within the displayed first digital image representation based on the position signal, generating a zoom-in signal in response to a detection of the pointing device sliding away from the touch position on the touch display, said zoom signal indicating a sliding distance from the touch position, displaying an enlarged representation of the first digital image representation in response to the zoom-in signal, generating a shift signal in response to a detection of a second sliding motion of the fingertip on the touch sensitive display, and displaying a second digital image representation and an enlarged representation of the second digital image representation in response to the shift signal, the enlargement of the subarea being based on the zoom signal generated during displaying of the first digital image representation.

Description

TECHNICAL FIELD OF INVENTION
The present invention relates to a method for selecting an image and a method for facilitating generation or capturing of a desired image.
BACKGROUND TO THE INVENTION
A great number of the graphical or photographical images of today are generated digitally. Generally this results in that more images are created and many times a greater amount of undesirable images are created. One of the problems of today is that even undesired images are stored and thereby occupy storage capacity. One simple solution to this problem is to delete all undesired images. However, the likelihood of undesired images still occupying storage capacity becomes greater as the time pass from the time an image was created.
SUMMARY OF THE INVENTION
It is an object of the invention to improve operations on images and to improve the experience for the user of continuous operations on images.
This object is achieved by means of a method for selecting an image according to claim 1. Further embodiments of the invention are disclosed in the dependent claims.
In particular, according to a first aspect of the invention, a method for selecting an image comprises forming a group of digital image representations, displaying a first digital image representation of the group of digital image representations on a touch sensitive display, generating a position signal in response to a detection of a pointing device on the touch sensitive display, said position signal indicating a touch position, identifying a selected position within the displayed first digital image representation based on the position signal, generating a zoom-in signal in response to a detection of the pointing device sliding away from the touch position on the touch display, said zoom signal indicating a sliding distance from the touch position, displaying an enlarged representation of the first digital image representation in response to the zoom-in signal, generating a shift signal in response to a detection of a second sliding motion of the pointing device on the touch sensitive display, and displaying a second digital image representation and an enlarged representation of the second digital image representation in response to the shift signal, the enlargement of the second digital image representation being based on the zoom signal generated during displaying of the first digital image representation.
The advantage of forming a group of images is that the chance of having at least one good image is increased. By implementing the selection method the selection of the best image is facilitated and thereby it becomes easier for a user to discard less desirable images. Moreover, the zoom-in facilitates viewing of details in the high resolution image on a smaller preview screen.
According to one embodiment the sliding direction of the second sliding motion is along a trajectory that is substantially circular. This is an advantage in that the sliding motion determines the switching between images and if the number of images in the group of image representations are large, the touch sensitive display may not be big enough. However, by making the sliding motion circular there is no longer any limitations as it becomes possible to slide the pointing device many turns.
According to yet another embodiment the touch sensitive display is a multi-touch sensitive display, wherein said generating of a position signal further includes a detection of a second pointing device on the multi-touch sensitive display, said position signal indicating a touch position which is based on each position of the two pointing devices respectively, wherein said zoom-in signal is generated in response to a detection of the two pointing devices sliding on the multi-touch display away from each other, and wherein said shift signal is generated in response to a second sliding motion of the two pointing devices at substantially constant distance between the pointing devices.
In another embodiment the touch position is a position between the two detected pointing devices.
In yet another embodiment the act of displaying an enlarged representation of the first image includes displaying an enlarged subarea of the first digital image representation in response to the zoom-in signal, the position of the subarea within the first digital image representation being based on the selected position and the enlargement of the subarea being based on the zoom-in signal.
In one embodiment the size of the subarea to be enlarged is based on the distance between the two pointing devices at the generation of the position signal.
In another embodiment said forming of a group of digital image representations includes capturing a plurality of different digital image representations of essentially the same view at different points in time.
According to another embodiment said forming of a group of digital image representations includes capturing a plurality of different digital image representations of essentially the same view at different exposure settings.
According to yet another embodiment said forming of a group of digital image representations includes capturing a plurality of different digital image representations of essentially the same view having different focus distances.
According to a further embodiment said forming of a group of digital image representations includes capturing a plurality of different digital image representations of essentially the same view being exposed for different transforms.
According to yet another embodiment said forming of a group of digital image representations includes generating a plurality of digital image representations from one single original image by manipulating the original image differently for each digital, the manipulation includes applying a transform or a parameter to the original image.
According to another embodiment a pointing device is a fingertip.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention will now be described in further detail by way of example under reference to the accompanying drawings, on which:
FIG. 1a is a schematic view of a display side of an image presentation device according to one embodiment of the invention,
FIG. 1b is a schematic view of a lens side of an image presentation device incorporating a camera according to one embodiment of the invention,
FIG. 1c is a schematic block diagram of the image presentation device in FIGS. 1a -b,
FIG. 2 shows a group of image representations according to one embodiment of the invention,
FIG. 3a-d schematically shows acts of a method according to one embodiment of the invention.
FIGS. 4a-d are representations of example displays according example embodiments.
DETAILED DESCRIPTION OF A PREFERRED EMBODIMENT
The present invention will now be described more fully hereinafter with reference to the accompanying drawings, in which certain embodiments are shown. Like numbers refer to like elements throughout.
In FIG. 1 an image presentation device 10 for implementing the invention is showed. According to one embodiment the image presentation device 10 includes housing 12 and a touch sensitive display 14, see FIG. 1a . It may also include a lens 16, see FIG. 1b , for focusing light to be captured as an image on an image sensor, not showed. The image presentation device 10 may, thus, be included in a camera or have a camera incorporated.
Moreover, now referring to FIG. 1c , the image presentation device 10 includes processor 18, a volatile memory 20, a non volatile memory 22, a display driver 24, a touch screen driver 26, touch screen circuitry 28, and camera circuitry 30 including an image sensor.
The processor 18, the volatile memory 20 and the non volatile memory 22 may be arranged and connected in a way known to the skilled person for operation of the image presentation device and execution of applications stored in the non-volatile memory 22.
The design and implementation of the touch screen circuitry 28 depends on the type of touch sensitive display that is to be used. The implementation of the touch screen driver 26 depends on the type of touch sensitive display and the operating system of the image presentation device 10.
In the present application the term touch sensitive display or touch screen is used for a display that is arranged to detect the presence, location, and/or movement of a “touch” within the display area. The touch screen may be designed to detect presence, location, and/or movement on the display by a finger, a hand, a stylus, a pen, etc.
Depending on the usage of the image presentation device one of a plurality of types of touch screens may be selected. For example may the touch screen be a resistive touch screen, a touch screen based on surface acoustic wave technology, a capacitive touch screen, a touch screen using surface capacitance, a touch screen based on projected capacitive touch technology, a system based on infrared LEDs and photo sensors, a system based on a strain gauge configuration, a touch screen based on dispersive signal technology, a touch screen based on acoustic pulse recognition technology, etc.
According to one embodiment a method for selecting images is part of a greater scheme of achieving a desired image having specific characteristics. The embodiment relates to an image selecting method operating on a group of digital image representations in order to achieve this result. The images forming the group of image representations may be images retrieved from a storage device, e.g. a hard drive, the non volatile memory 22, an image server accessed via a network, etc. The images may alternatively be acquired by means of a camera arranged in the device 10 or by means of transforming one original image retrieved from a storage device or one original image acquired by said camera. The image sequence may also be calculated from one or more source images, and the image itself may be virtual representation based on one or more mathematical schemes applied on one or more original images.
One example of how to generate the group of image representations are to bracket, i.e. to take photographs at more than one exposure in order to ensure that the desired exposure is obtained in at least one exposure. Other examples are to take a plurality of photographs at different points in time, different depth of field, at different focus distances, or by varying any other setting of the camera. The camera used in these examples may well be a camera implemented in the same device or system as the image presentation device 10. Moreover, the group of image representation may be generated from applying different transforms to the images.
The number of images in a group of image representations may be as few as two and as many as hundreds, it much depends on the application in which the method is planned to be used. The group of images may be separate images/photos or a sequence of frames in a video. In FIG. 2 a group of image representations 50 including three image representations is showed. In order to facilitate the depiction of the different images, this particular sequence of images represents photographs 52, 54, 56, also referred to as first, second and third photographs, taken at different points in time. Hence these images could very well have been taken by the camera of the device shortly before the user desides to use the selection method. The selection method may be started in many ways, e.g. by pressing a button on the image presentation device 10, by touching the touch sensitive display 14 by means of a pointing device either within a displayed image representation from the group of image representations or within an area on the display indicating a button for activiating this functionality.
FIG. 3 shows one embodiment of the invention by depicting subsequent acts of the selection method of this embodiment. The image presentation device 10 displays a representation of the second photograph 54, see FIG. 3a , from the group of image representations 50. Then the pointing device 70 is positioned to touch the touch sensitive display 14 at position 72, see FIG. 3b , and a position signal is generated, including the touch position, and is sent from the touch sensitive display 14 to the processor 18, i.e. from the touch screen circuitry 28 and the touch screen driver 26 to the processor 18.
Then the touch sensitive display 14 detects a sliding motion 74, performed by means of the pointing device 70, along the display 14 away from the touch position 72. This detection results in the generation of a zoom signal that is sent to the processor 18. The zoom signal includes an indication of the distance of the sliding movement 74, referred to as zoom-value. Based on the zoom signal and the zoom value the displayed image representation 54 is enlarged to a degree that is based on the zoom value, see FIG. 3c . In the embodiment of FIG. 3c the enlarging of the image representation is centered at the touch position 72.
Now referring to FIG. 3d , the pointing device is moved in another direction 76, in this specific embodiment essentially perpendicular to previous movement 74. This second movement 76 may advantageously be following a circular trajectory, as indicated by 76 in FIG. 3d , in order to facilitating shifting through a large group of image representation without running out of touch sensitive display surface to slide on. The second sliding motion 76 is detected by the touch sensitive display 14 and a shift signal is generated and sent to the processor 18. In response to the shift signal the image displayed 54 is shifted to another image representation 56 in the group of image representations 50. In FIG. 3d the third photograph 56 is shifted in for display. In this embodiment the same enlargement is applied to the newly shifted image representation as for the previous.
The number of images shifted may be proportional to the length of the second sliding motion 76. Hence, in the example above, in which the group of image representations 50 only includes three photographs 52, 54, 56, the shift would continue to present the image representation of the first photograph if the second sliding motion 76 is continued.
According to another embodiment the enlargement is not applied to the entire image representation as depicted in FIGS. 3c and 3d , but rather to a predetermined area surrounding the touch position indicated by the position signal. The area may be defined by a radius and may be substantially circular.
According to yet another embodiment two pointing devices are used, e.g. a finger and the thumb of a users hand. In this embodiment a position signal is generated when the two pointing devices are detected on the touch sensitive display, the touch position being indicated as a position between the detection points of the two pointing devices.
Then, upon detection of the two pointing devices sliding away from each other, a zoom signal is generated and in response to the zoom signal an enlarged image representation of the image representation presently displayed is presented on the display. The degree of enlargement is based on the distance the two pointing devices have been sliding away from each other. Moreover, according to one embodiment, not the entire image representation is zoomed but only a subarea. The size of this sub area may correspond to an area defined by the initial positions of the pointing devices, i.e. when the touch position is indicated.
Then, in response to detection of a second sliding motion by the two pointing devices, wherein the two pointing devices are sliding at substantially constant distance from each other, a shift signal is generated and in response to the shift signal the image representation displayed is shifted to another image representation from the group of image representations. In one embodiment they are rotated substantially around a position in-between the two pointing devices and at a substantially constant distance from each other, e.g. following a substantially circular trajectory. The length of the sliding motion determines which image representation from the group of image representations to display. The enlargement applied to the initial image is displayed in the shifted images as well.
According to one specific embodiment only portions of the initially displayed image representation is shifted. The portion to be shifted may for instance be indicated manually by tracing the contours of the area and then the shifting results in that the corresponding area of another image from the group of image representations is displayed. The contours of the area can also be automatically computed by tracing where the two images', aligned to substantially same positions within the particular area, pixels are substantially similar around the traced contour. By means of this embodiment combined with a group of image representations being a bracketed image sequence it is possible to generate HDR images, High Dynamic Range images.
According to one embodiment the method may advantageously be used for browsing images. In such an application the zoom-in step may be skipped and a rotational/circular motion using one or two pointing devices may be used to switch images. In this embodiment the group of image representations probably includes the images of a folder in a file system or a database or of a particular category in a database.
According to one particular embodiment the image presentation device 10 is a mobile telephone equipped with a camera.
According to another aspect of the present invention yet another embodiment is shown in FIG. 4. The image presentation device 10 displays a representation of the second photograph 54, see FIG. 4a , from the group of image representations 50. Then the pointing device 70 is positioned to touch the touch sensitive display 14 at a first position 82, see FIG. 4b , and a position signal is generated, including the touch position, and is sent from the touch sensitive display 14 to the processor 18, i.e. from the touch screen circuitry 28 and the touch screen driver 26 to the processor 18.
Then the displayed image representation 54 is enlarged to a degree that is based on a predetermined zoom value, see FIG. 4b . In the embodiment of FIG. 4b the enlarging of the image representation is centered at the touch position. According to this embodiment the enlargement is not applied to the entire image representation but rather to a predetermined area 81 surrounding the touch position indicated by the position signal. According to the in FIG. 4b shown embodiment the predetermined area 81 is defined by a radius and is substantially circular.
By moving the pointing device 70 on the touch sensitive display 14 from the first position 82 to a second position 84, see FIG. 4c , the touch position including the position signal is updated. The updated position signal is sent from the touch sensitive display 14 to the processor 18, i.e. from the touch screen circuitry 28 and the touch screen driver 26 to the processor 18. In response to the updated position signal the predetermined area 81 of the image representation that is enlarged, in this case image representation of the second photograph 54, is also updated so that the enlargement is applied to the predetermined area 81 surrounding the touch position indicated by the position signal. The predetermined area 81 of the image representation that is enlarged, in this case image representation of the second photograph 54, may be live updated whilst the pointing device 70 is moved from the first position 82 to the second position 84. Thus, as long as the pointing device 70 is moved along the trajectory between first position 82 to the second position 84 the predetermined area 81 of the image representation that is enlarged is live updated.
The size of the predetermined area 81 may be reduced or enlarged upon detection of two pointing devices sliding towards each other or away from each other, respectively. This may e.g. be done by pointing the two pointing devices on to opposite sections on the border of the predetermined area 81, respectively, and then sliding the two pointing devices towards each other or away from each other.
As a next step the image representation shown in the predetermined area 81 may be shifted. Now referring to FIG. 4d , the pointing device 70 is moved in a direction 86 along the border of the enlarged area of the image representation. Thus, according to this specific embodiment, along a substantially circular trajectory, as indicated by a sliding motion 86 in FIG. 4d , this in order to facilitating shifting through a large group of image representation without running out of touch sensitive display surface to slide on. The sliding motion 86 is detected by the touch sensitive display 14 and a shift signal is generated and sent to the processor 18. In response to the shift signal the image displayed in the enlarged predetermined area 81 is shifted to another image representation in the group of image representations 50, in this case to the image representation of the third photograph 56. In this embodiment the same enlargement is applied to the newly shifted image representation as for the previous. The number of images shifted may be proportional to the length of the sliding motion 86. Hence, in the example above, in which the group of image representations 50 only includes three photographs 52, 54, 56, the shift would continue to present the image representation of the first photograph 52 if the sliding motion 86 is continued.
Furthermore according to the above embodiment when a specific image representation has been selected among the group of image representations by performing the sliding motion 86 the not selected digital image representations of the group of digital image representations may be discarded.
It is recognized that the embodiment of FIG. 4 may be altered in various ways. For example, according to one embodiment the shifting of the image representations may be performed by moving the pointing device 70 in a substantially linear manner. This embodiment may be developed even further by showing small miniatures of the image representations at e.g. the bottom of the touch sensitive display 14. The shifting of the image representations may then be performed by moving the small miniatures by means of the pointing device 70 in a substantially linear manner along a path that is substantially parallel with the lower border of the touch sensitive display 14.

Claims (19)

The invention claimed is:
1. A method for selecting a digital image representation comprising:
forming a group of digital image representations;
displaying on a touch sensitive display a first digital image representation from the group of digital image representations on the touch sensitive display;
in response to a detection of a first user input at the touch sensitive display, generating a position signal indicative of a touch position;
identifying an area at the touch position indicated by the position signal as a first portion of the first digital image representation within a second portion of the first digital image representation;
receiving a second user input at the touch sensitive display and, in response to the received second user input, generating a first new composite image by displaying on the touch sensitive display the second portion of the first digital image representation but not the first portion of the first digital image representation and displaying on the touch sensitive display, in the area at the touch position, a third portion extracted from a second digital image representation from the group of digital image representations, which third portion is automatically identified by comparing the first portion of the first digital image representation to the second digital image representation; and
receiving a third user input at the touch sensitive display and, in response to the received third user input, generating a second new composite image by displaying on the touch sensitive display the second portion of the first digital image representation but not the first portion of the first digital image representation nor the third portion of the second digital image representation and displaying on the touch sensitive display, in the area at the touch position, a fourth portion extracted from a third digital image representation from the group of digital image representations, which fourth portion is automatically identified by comparing the first portion of the first digital image representation to the third digital image representation.
2. A method as claimed in claim 1, wherein the first user input comprises a sliding motion of a pointing device at the touch sensitive display in a direction along a border of the area at the touch position.
3. A method as claimed in claim 2, wherein the second user input comprises a further sliding motion of a pointing device at the touch sensitive display in a direction along the border of the area at the touch position.
4. A method as claimed in claim 1, wherein the third portion extracted from the second digital image representation corresponds or is similar to the first portion of the first digital image representation.
5. A method as claimed in claim 4, wherein the third portion extracted from the second digital image representation comprises pixels that are substantially similar to pixels of the first portion of the first digital image representation.
6. A method according to claim 1, wherein the first portion of the first digital image representation is displayed as an enlarged representation of the first portion.
7. A method according to claim 1 comprising generating a zoom signal in response to at least a further touch at the touch sensitive display, and displaying a representation of a portion of a digital image representation dependent upon the zoom signal.
8. A method according to claim 1 comprising selecting the third portion extracted from the second digital image representation from the group of digital image representations to replace the first portion of the displayed first digital image representation.
9. A method according to claim 8, further comprising discarding the not selected digital image representations from the group of digital image representations.
10. A method according to claim 1, comprising displaying the other digital image representations, or comprising displaying miniatures of the digital image representations.
11. A method according to claim 1, wherein the group of digital image representations comprises any of or any combination of:
images retrieved from a storage device; or
images of essentially the same view, captured at different points in time; or
images of essentially the same view, captured with different exposure settings; or
images of essentially the same view, captured with different focus distances; or
images of essentially the same view, exposed for different transforms.
12. A method according to claim 1 wherein the group of digital image representations comprise images captured of a same scene at a same field of view.
13. A method as claimed in claim 1, wherein comparing the first portion of the first digital image representation to the second digital image representation comprises comparing a content of the first portion of the first digital image representation to a content of the second digital image representation.
14. A method as claimed in claim 1, wherein the area at the touch position comprises an area within a predetermined radius of the touch position.
15. An image presentation device comprising:
memory for storing a group of digital image representations,
a touch sensitive display for receiving user inputs and displaying a first digital image representation from the group of digital image representations, and
a processor for, in response to a detection of a first user input at the touch sensitive display, generating a position signal indicative of a touch position,
wherein the processor is configured to:
identify an area at the touch position indicated by the position signal as a first portion of the first digital image representation within a second portion of the first digital image representation;
generate a first new composite image, in response to a second user input at the touch sensitive display, by displaying on the touch sensitive display the second portion of the first digital image representation but not the first portion of the first digital image representation and displaying on the touch sensitive display, in the area at the touch position, a third portion extracted from a second digital image representation from the group of digital image representations, which third portion is automatically identified by the processor by comparing the first portion of the first digital image representation to the second digital image representation; and
generate a second new composite image, in response to a third user input at the touch sensitive display, by displaying on the touch sensitive display the second portion of the first digital image representation but not the first portion of the first digital image representation nor the third portion of the second digital image representation and displaying on the touch sensitive display, in the area at the touch position, a fourth portion extracted from a third digital image representation from the group of digital image representations, which fourth portion is automatically identified by the processor by comparing the first portion of the first digital image representation to the third digital image representation.
16. An image presentation device according to claim 15, wherein the first user input comprises a sliding motion of a pointing device at the touch sensitive display in a direction along a border of the area at the touch position.
17. An image presentation device according to claim 15 wherein the display is any of or any combination of:
a resistive touch screen; or
a touch screen based on surface acoustic wave technology; or
a capacitive touch screen; or
a touch screen using surface capacitance; or
a touch screen based on projected capacitive touch technology; or
a system based on infrared LEDs and photo sensors; or
a system based on a strain gauge configuration; or
a touch screen based on dispersive signal technology; or
a touch screen based on acoustic pulse recognition technology.
18. An image presentation device according to claim 15 wherein the group of digital image representations comprise images captured of a same scene at a same field of view.
19. An image presentation device comprising:
storage means for storing a group of digital image representations;
a touch sensitive display for receiving user inputs and displaying a first digital image representation from the group of digital image representations;
processor means for, in response to a detection of a first user input at the touch sensitive display, generating a position signal indicative of a touch position;
means for identifying an area at the touch position indicated by the position signal as a first portion of the first digital image representation within a second portion of the first digital image representation;
means for generating a first new composite image, in response to a second user input at the touch sensitive display, by displaying on the touch sensitive display the second portion of the first digital image representation but not the first portion of the first digital image representation and displaying on the touch sensitive display, in the area at the touch position, a third portion extracted from a second digital image representation from the group of digital image representations, which third portion is automatically identified by the processor means by comparing the first portion of the first digital image representation to the second digital image representation; and
means for generating a second new composite image, in response to a third user input at the touch sensitive display, by displaying on the touch sensitive display the second portion of the first digital image representation but not the first portion of the first digital image representation nor the third portion of the second digital image representation and displaying on the touch sensitive display, in the area at the touch position, a fourth portion extracted from a third digital image representation from the group of digital image representations, which fourth portion is automatically identified by the processor means by comparing the first portion of the first digital image representation to the third digital image representation.
US13/499,711 2009-10-01 2010-09-22 Method relating to digital images Active 2033-02-12 US9792012B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
SE0901263 2009-10-01
SE0901263-4 2009-10-01
PCT/SE2010/051019 WO2011040864A1 (en) 2009-10-01 2010-09-22 Method relating to digital images

Publications (2)

Publication Number Publication Date
US20120262490A1 US20120262490A1 (en) 2012-10-18
US9792012B2 true US9792012B2 (en) 2017-10-17

Family

ID=43826514

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/499,711 Active 2033-02-12 US9792012B2 (en) 2009-10-01 2010-09-22 Method relating to digital images

Country Status (3)

Country Link
US (1) US9792012B2 (en)
EP (1) EP2483767B1 (en)
WO (1) WO2011040864A1 (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2483767B1 (en) 2009-10-01 2019-04-03 Nokia Technologies Oy Method relating to digital images
SE534551C2 (en) 2010-02-15 2011-10-04 Scalado Ab Digital image manipulation including identification of a target area in a target image and seamless replacement of image information from a source image
SE1150505A1 (en) 2011-05-31 2012-12-01 Mobile Imaging In Sweden Ab Method and apparatus for taking pictures
JP5760742B2 (en) * 2011-06-27 2015-08-12 ヤマハ株式会社 Controller and parameter control method
EP2718896A4 (en) * 2011-07-15 2015-07-01 Mobile Imaging In Sweden Ab Method of providing an adjusted digital image representation of a view, and an apparatus
JP2015172836A (en) * 2014-03-11 2015-10-01 キヤノン株式会社 Display control unit and display control method
JP2016111652A (en) * 2014-12-10 2016-06-20 オリンパス株式会社 Imaging apparatus, imaging method and program

Citations (127)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5138460A (en) 1987-08-20 1992-08-11 Canon Kabushiki Kaisha Apparatus for forming composite images
US5657402A (en) * 1991-11-01 1997-08-12 Massachusetts Institute Of Technology Method of creating a high resolution still image using a plurality of images and apparatus for practice of the method
US5689611A (en) 1992-10-09 1997-11-18 Sony Corporation Panorama image producing method and apparatus
JPH1188811A (en) 1997-09-04 1999-03-30 Sony Corp Camcorder and photographing method
WO1999051027A1 (en) 1998-03-31 1999-10-07 Intel Corporation Method of and apparatus for creating panoramic or surround images using a motion sensor equipped camera
US6075905A (en) 1996-07-17 2000-06-13 Sarnoff Corporation Method and apparatus for mosaic image construction
US6105045A (en) 1995-10-18 2000-08-15 Fuji Xerox Co., Ltd. Image processing apparatus and image attribute altering method
US6249616B1 (en) 1997-05-30 2001-06-19 Enroute, Inc Combining digital images based on three-dimensional relationships between source image data sets
US20010010546A1 (en) 1997-09-26 2001-08-02 Shenchang Eric Chen Virtual reality camera
WO2001059709A1 (en) 2000-02-11 2001-08-16 Make May Toon, Corp. Internet-based method and apparatus for generating caricatures
US20010020978A1 (en) 2000-03-08 2001-09-13 Seiichi Matsui Electronic camera
US20010048802A1 (en) 2000-04-19 2001-12-06 Nobuyoshi Nakajima Method, apparatus, and recording medium for generating album
US20020025796A1 (en) 2000-08-30 2002-02-28 Taylor William Stuart System and method conducting cellular POS transactions
US20020110286A1 (en) * 2001-02-10 2002-08-15 Cheatle Stephen Philip Method of selectively storing digital images
US20020159632A1 (en) 2001-02-23 2002-10-31 Chui Charles K. Graphic image re-encoding and distribution system and method
US6529627B1 (en) 1999-06-24 2003-03-04 Geometrix, Inc. Generating 3D models by combining models from a video-based technique and data from a structured light technique
US6542645B1 (en) 1997-07-15 2003-04-01 Silverbrook Research Pty Ltd Adaptive tracking of dots in optical storage system using ink dots
US20030071908A1 (en) 2001-09-18 2003-04-17 Masato Sannoh Image pickup device, automatic focusing method, automatic exposure method, electronic flash control method and computer program
EP1309171A2 (en) 2001-10-23 2003-05-07 STMicroelectronics, Inc. Still picture format for subsequent picture stitching for forming a panoramic image
US20030147000A1 (en) 2002-01-30 2003-08-07 Kenji Shiraishi Digital still camera, reproduction device, and image processor
US6621524B1 (en) 1997-01-10 2003-09-16 Casio Computer Co., Ltd. Image pickup apparatus and method for processing images obtained by means of same
US20030190090A1 (en) 2002-04-09 2003-10-09 Beeman Edward S. System and method for digital-image enhancement
US20030189647A1 (en) * 2002-04-05 2003-10-09 Kang Beng Hong Alex Method of taking pictures
WO2003105466A1 (en) 2002-06-07 2003-12-18 Koninklijke Philips Electronics N.V. Method of imaging an object and mobile imaging device
CN1471375A (en) 2000-10-24 2004-01-28 ��ʽ���綫֥ Image processing device and image processing method
US6724386B2 (en) 2001-10-23 2004-04-20 Sony Corporation System and process for geometry replacement
US20040080661A1 (en) 2000-12-22 2004-04-29 Sven-Ake Afsenius Camera that combines the best focused parts from different exposures to an image
US20040097206A1 (en) 2001-05-11 2004-05-20 Christian Grewing Circuit configuration for the frequency conversion of an oscillator frequency into a carrier frequency
US6750903B1 (en) 1998-03-05 2004-06-15 Hitachi, Ltd. Super high resolution camera
WO2004068865A1 (en) 2003-01-24 2004-08-12 Micoy Corporation Steroscopic panoramic image capture device
US20040165788A1 (en) 2003-02-25 2004-08-26 Microsoft Corporation Image blending by guided interpolation
US20040174434A1 (en) 2002-12-18 2004-09-09 Walker Jay S. Systems and methods for suggesting meta-information to a camera user
US20040189849A1 (en) 2003-03-31 2004-09-30 Hofer Gregory V. Panoramic sequence guide
US20040201699A1 (en) 2001-07-17 2004-10-14 Eastman Kodak Company Revised recapture camera and method
US20040201755A1 (en) 2001-12-06 2004-10-14 Norskog Allen C. Apparatus and method for generating multi-image scenes with a camera
US20040218833A1 (en) 1997-09-10 2004-11-04 Koichi Ejiri System and method for displaying an image indicating a positional relation between partially overlapping images
US20040223649A1 (en) 2003-05-07 2004-11-11 Eastman Kodak Company Composite imaging method and system
US20040239767A1 (en) 2003-05-29 2004-12-02 Stavely Donald J. Systems and methods for providing tactile feedback
US20050007483A1 (en) * 1999-04-08 2005-01-13 Zimmermann Steven Dwain Method and apparatus for providing virtual processing effects for wide-angle video images
US20050031214A1 (en) 2000-10-27 2005-02-10 Microsoft Corporation Rebinning methods and arrangements for use in compressing image-based rendering (IBR) data
WO2005036780A1 (en) 2003-10-07 2005-04-21 Matsushita Electric Industrial Co., Ltd. Search supporting apparatus and search supporting method
US20050099514A1 (en) 2003-11-08 2005-05-12 Cozier Robert P. Automated zoom control
WO2005050567A1 (en) 2003-11-18 2005-06-02 Scalado Ab Method for processing a digital image and image representation format
US20050122412A1 (en) 2002-04-17 2005-06-09 Seiko Epson Corporation Digital camera
US20050168594A1 (en) 2004-02-04 2005-08-04 Larson Brad R. Digital camera and method for in creating still panoramas and composite photographs
US6927874B1 (en) 1999-04-02 2005-08-09 Canon Kabushiki Kaisha Image processing method, apparatus and storage medium therefor
US6930703B1 (en) 2000-04-29 2005-08-16 Hewlett-Packard Development Company, L.P. Method and apparatus for automatically capturing a plurality of images during a pan
WO2005112437A1 (en) 2004-05-13 2005-11-24 Sony Corporation Image pickup system, image pickup device and image pickup method
US6975352B2 (en) 2000-12-18 2005-12-13 Xerox Corporation Apparatus and method for capturing a composite digital image with regions of varied focus and magnification
EP1613060A1 (en) 2004-07-02 2006-01-04 Sony Ericsson Mobile Communications AB Capturing a sequence of images
US20060001650A1 (en) * 2004-06-30 2006-01-05 Microsoft Corporation Using physical objects to adjust attributes of an interactive display application
US6985172B1 (en) 1995-12-01 2006-01-10 Southwest Research Institute Model-based incident detection system with motion classification
WO2006002796A1 (en) 2004-07-02 2006-01-12 Sony Ericsson Mobile Communications Ab Capturing a sequence of images
US20060022961A1 (en) * 2004-07-28 2006-02-02 Tomomi Kaminaga Reproduction apparatus, camera, and display switch method of reproduction apparatus
US20060028579A1 (en) 2004-08-05 2006-02-09 Yutaka Sato Image pickup apparatus, method of controlling image pickup and program
US20060038908A1 (en) * 2004-08-18 2006-02-23 Canon Kabushiki Kaisha Image processing apparatus, image processing method, program, and storage medium
US20060044444A1 (en) 2004-08-30 2006-03-02 Pentax Corporation Digital camera
US20060050152A1 (en) 2004-09-03 2006-03-09 Rai Barinder S Method for digital image stitching and apparatus for performing the same
CN1750593A (en) 2005-10-13 2006-03-22 上海交通大学 Digital camera with image split function
US20060061845A1 (en) 2004-09-17 2006-03-23 Ulead Systems, Inc. Image composition systems and methods
US20060078224A1 (en) 2002-08-09 2006-04-13 Masashi Hirosawa Image combination device, image combination method, image combination program, and recording medium containing the image combination program
US20060171687A1 (en) 2003-04-17 2006-08-03 Seiko Epson Corporation Generation of still image from a plurality of frame images
US20060181619A1 (en) 2005-02-11 2006-08-17 Creative Technology, Ltd. Method and apparatus for forming a panoramic image
US20060181614A1 (en) 2005-02-17 2006-08-17 Jonathan Yen Providing optimized digital images
US20060187321A1 (en) 1998-06-22 2006-08-24 Koichi Sakamoto Image sensing apparatus and method for synthesizing a composite image
US7099510B2 (en) 2000-11-29 2006-08-29 Hewlett-Packard Development Company, L.P. Method and system for object detection in digital images
US20060235765A1 (en) 2005-04-15 2006-10-19 David Clifford R Interactive Image Activation and Distribution System and Associated Methods
US7133069B2 (en) 2001-03-16 2006-11-07 Vision Robotics, Inc. System and method to increase effective dynamic range of image sensors
US7133169B2 (en) 1996-11-05 2006-11-07 Yoshiki Tsuchiyama Apparatus equipped with removable scanner unit
WO2007006075A1 (en) 2005-07-14 2007-01-18 Canon Information Systems Research Australia Pty Ltd Image browser
US20070024721A1 (en) 2005-07-29 2007-02-01 Rogers Sean S Compensating for improperly exposed areas in digital images
US20070025723A1 (en) 2005-07-28 2007-02-01 Microsoft Corporation Real-time preview for panoramic images
US20070030363A1 (en) 2005-08-05 2007-02-08 Hewlett-Packard Development Company, L.P. Image capture method and apparatus
US20070058064A1 (en) 2005-09-14 2007-03-15 Sony Corporation Image processing apparatus and method, and program therefor
WO2007038198A2 (en) 2005-09-26 2007-04-05 Eastman Kodak Company Image capture method and device also capturing audio
US20070081081A1 (en) 2005-10-07 2007-04-12 Cheng Brett A Automated multi-frame image capture for panorama stitching using motion sensor
US20070097206A1 (en) 2005-11-02 2007-05-03 Houvener Robert C Multi-user stereoscopic 3-D panoramic vision system and method
US20070224980A1 (en) 2006-03-24 2007-09-27 Sony Ericsson Mobile Communications Ab Methods, systems, and devices for detecting and indicating loss of proximity between mobile devices
US20070237421A1 (en) 2006-03-29 2007-10-11 Eastman Kodak Company Recomposing photographs from multiple frames
US20070274563A1 (en) 2005-06-02 2007-11-29 Searete Llc, A Limited Liability Corporation Of State Of Delaware Capturing selected image objects
US7339580B2 (en) 1998-01-26 2008-03-04 Apple Inc. Method and apparatus for integrating manual input
US20080062141A1 (en) 2006-09-11 2008-03-13 Imran Chandhri Media Player with Imaged Based Browsing
WO2008038883A1 (en) 2006-09-29 2008-04-03 Lg Electronics Inc. Method of generating key code in coordinate recognition device and video device controller using the same
WO2008064349A1 (en) 2006-11-22 2008-05-29 Nik Software, Inc. Method for dynamic range editing
US20080143744A1 (en) 2006-12-13 2008-06-19 Aseem Agarwala Gradient-domain compositing
US20080152258A1 (en) 2006-12-20 2008-06-26 Jarno Tulkki Digital mosaic image construction
EP1942401A1 (en) 2007-01-05 2008-07-09 Apple Inc. Multimedia communication device with touch screen responsive to gestures for controlling, manipulating and editing of media files
US20080218613A1 (en) 2007-03-09 2008-09-11 Janson Wilbert F Camera using multiple lenses and image sensors operable in a default imaging mode
US20080218611A1 (en) 2007-03-09 2008-09-11 Parulski Kenneth A Method and apparatus for operating a dual lens camera to augment an image
US20090019399A1 (en) 2007-07-10 2009-01-15 Brother Kogyo Kabushiki Kaisha Image displaying device, and method and computer readable medium for the same
US20090021576A1 (en) 2007-07-18 2009-01-22 Samsung Electronics Co., Ltd. Panoramic image production
US20090046943A1 (en) * 2006-03-31 2009-02-19 Nikon Corporation Image processing method
US20090073285A1 (en) 2007-09-14 2009-03-19 Sony Corporation Data processing apparatus and data processing method
US20090141046A1 (en) 2007-12-03 2009-06-04 Apple Inc. Multi-dimensional scroll wheel
US20090190803A1 (en) 2008-01-29 2009-07-30 Fotonation Ireland Limited Detecting facial expressions in digital images
US20090244301A1 (en) * 2008-04-01 2009-10-01 Border John N Controlling multiple-image capture
US20090245685A1 (en) 2008-03-25 2009-10-01 Sony Corporation Image processing apparatus, image processing method, and program
US20090251591A1 (en) 2008-04-04 2009-10-08 Whitham Charles L Digital Camera with High Dynamic Range Mode of Operation
EP2124186A1 (en) 2007-02-15 2009-11-25 Sony Corporation Image processing device and image processing method
US20090290759A1 (en) 2008-05-22 2009-11-26 Matrix Electronic Measuring, L.P. Stereoscopic measurement system and method
US20090295830A1 (en) * 2005-12-07 2009-12-03 3Dlabs Inc., Ltd. User interface for inspection of photographs
US20090303338A1 (en) 2008-06-06 2009-12-10 Texas Instruments Incorporated Detailed display of portion of interest of areas represented by image frames of a video signal
US20090309990A1 (en) 2008-06-11 2009-12-17 Nokia Corporation Method, Apparatus, and Computer Program Product for Presenting Burst Images
US7639897B2 (en) 2006-01-24 2009-12-29 Hewlett-Packard Development Company, L.P. Method and apparatus for composing a panoramic photograph
WO2009156561A1 (en) 2008-06-27 2009-12-30 Nokia Corporation Method, apparatus and computer program product for providing image modification
US20090322926A1 (en) 2008-06-25 2009-12-31 Tetsuo Ikeda Image processing apparatus and image processing method
JP2010020581A (en) 2008-07-11 2010-01-28 Shibaura Institute Of Technology Image synthesizing system eliminating unnecessary objects
US7659923B1 (en) 2005-06-24 2010-02-09 David Alan Johnson Elimination of blink-related closed eyes in portrait photography
US20100045608A1 (en) * 2008-08-20 2010-02-25 Sony Ericsson Mobile Communications Ab Multidimensional navigation for touch sensitive display
US20100053353A1 (en) 2008-08-27 2010-03-04 Micron Technology, Inc. Method and system for aiding user alignment for capturing partially overlapping digital images
US20100079498A1 (en) * 2008-09-26 2010-04-01 Microsoft Corporation Multi-modal interaction for a screen magnifier
EP2175635A1 (en) 2008-10-10 2010-04-14 Samsung Electronics Co., Ltd. Method and apparatus for creating high dynamic range image
US20100149367A1 (en) 2008-12-17 2010-06-17 Samsung Digital Imaging Co., Ltd. Digital image signal processing apparatus and method of displaying scene recognition
WO2010072587A1 (en) 2008-12-23 2010-07-01 Scalado Ab Subimage extraction from jpeg encoded data
US20100268729A1 (en) 2007-11-12 2010-10-21 Megachips Corporation Multimedia synthetic data generating apparatus
WO2011040864A1 (en) 2009-10-01 2011-04-07 Scalado Ab Method relating to digital images
EP2323102A1 (en) 2009-10-23 2011-05-18 ST-Ericsson (France) SAS Image capturing aid
US20110141227A1 (en) 2009-12-11 2011-06-16 Petronel Bigioi Stereoscopic (3d) panorama creation on handheld device
US20110200259A1 (en) 2010-02-15 2011-08-18 Lindskog Alexander Digital image manipulation
US8127232B2 (en) 1998-12-31 2012-02-28 Flashpoint Technology, Inc. Method and apparatus for editing heterogeneous media objects in a digital imaging device
US8160152B2 (en) 2006-03-01 2012-04-17 Mitsubishi Electric Corporation Moving image decoding apparatus and moving image coding apparatus
US20120105601A1 (en) 2010-10-27 2012-05-03 Samsung Electronics Co., Ltd. Apparatus and method for creating three-dimensional panoramic image by using single camera
US8463020B1 (en) * 2008-07-08 2013-06-11 Imove, Inc. Centralized immersive image rendering for thin client
US8494306B2 (en) 2007-12-13 2013-07-23 Samsung Electronics Co., Ltd. Method and an apparatus for creating a combined image
US20130300822A1 (en) 2003-09-12 2013-11-14 Sensormatic Electronics, LLC Video user interface system and method
US20140184852A1 (en) 2011-05-31 2014-07-03 Mobile Imaging In Sweden Ab Method and apparatus for capturing images
US9432583B2 (en) 2011-07-15 2016-08-30 Mobile Imaging In Sweden Ab Method of providing an adjusted digital image representation of a view, and an apparatus

Patent Citations (148)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5138460A (en) 1987-08-20 1992-08-11 Canon Kabushiki Kaisha Apparatus for forming composite images
US5657402A (en) * 1991-11-01 1997-08-12 Massachusetts Institute Of Technology Method of creating a high resolution still image using a plurality of images and apparatus for practice of the method
US5689611A (en) 1992-10-09 1997-11-18 Sony Corporation Panorama image producing method and apparatus
US6105045A (en) 1995-10-18 2000-08-15 Fuji Xerox Co., Ltd. Image processing apparatus and image attribute altering method
US6985172B1 (en) 1995-12-01 2006-01-10 Southwest Research Institute Model-based incident detection system with motion classification
US6075905A (en) 1996-07-17 2000-06-13 Sarnoff Corporation Method and apparatus for mosaic image construction
US7133169B2 (en) 1996-11-05 2006-11-07 Yoshiki Tsuchiyama Apparatus equipped with removable scanner unit
US6621524B1 (en) 1997-01-10 2003-09-16 Casio Computer Co., Ltd. Image pickup apparatus and method for processing images obtained by means of same
US6249616B1 (en) 1997-05-30 2001-06-19 Enroute, Inc Combining digital images based on three-dimensional relationships between source image data sets
US6542645B1 (en) 1997-07-15 2003-04-01 Silverbrook Research Pty Ltd Adaptive tracking of dots in optical storage system using ink dots
JPH1188811A (en) 1997-09-04 1999-03-30 Sony Corp Camcorder and photographing method
US20040218833A1 (en) 1997-09-10 2004-11-04 Koichi Ejiri System and method for displaying an image indicating a positional relation between partially overlapping images
US20010010546A1 (en) 1997-09-26 2001-08-02 Shenchang Eric Chen Virtual reality camera
US7339580B2 (en) 1998-01-26 2008-03-04 Apple Inc. Method and apparatus for integrating manual input
US6750903B1 (en) 1998-03-05 2004-06-15 Hitachi, Ltd. Super high resolution camera
WO1999051027A1 (en) 1998-03-31 1999-10-07 Intel Corporation Method of and apparatus for creating panoramic or surround images using a motion sensor equipped camera
TW401687B (en) 1998-03-31 2000-08-11 Intel Corp Method of and apparatus for creating panoramic or surround images using a motion sensor equipped camera
US6304284B1 (en) 1998-03-31 2001-10-16 Intel Corporation Method of and apparatus for creating panoramic or surround images using a motion sensor equipped camera
US20060187321A1 (en) 1998-06-22 2006-08-24 Koichi Sakamoto Image sensing apparatus and method for synthesizing a composite image
US8127232B2 (en) 1998-12-31 2012-02-28 Flashpoint Technology, Inc. Method and apparatus for editing heterogeneous media objects in a digital imaging device
US6927874B1 (en) 1999-04-02 2005-08-09 Canon Kabushiki Kaisha Image processing method, apparatus and storage medium therefor
US20050007483A1 (en) * 1999-04-08 2005-01-13 Zimmermann Steven Dwain Method and apparatus for providing virtual processing effects for wide-angle video images
US6529627B1 (en) 1999-06-24 2003-03-04 Geometrix, Inc. Generating 3D models by combining models from a video-based technique and data from a structured light technique
WO2001059709A1 (en) 2000-02-11 2001-08-16 Make May Toon, Corp. Internet-based method and apparatus for generating caricatures
US7199829B2 (en) 2000-03-08 2007-04-03 Fuji Photo Film Co., Ltd. Device and method for processing unprocessed image data based on image property parameters
US20010020978A1 (en) 2000-03-08 2001-09-13 Seiichi Matsui Electronic camera
US20010048802A1 (en) 2000-04-19 2001-12-06 Nobuyoshi Nakajima Method, apparatus, and recording medium for generating album
US6930703B1 (en) 2000-04-29 2005-08-16 Hewlett-Packard Development Company, L.P. Method and apparatus for automatically capturing a plurality of images during a pan
US20020025796A1 (en) 2000-08-30 2002-02-28 Taylor William Stuart System and method conducting cellular POS transactions
CN1471375A (en) 2000-10-24 2004-01-28 ��ʽ���綫֥ Image processing device and image processing method
US20050031214A1 (en) 2000-10-27 2005-02-10 Microsoft Corporation Rebinning methods and arrangements for use in compressing image-based rendering (IBR) data
US7099510B2 (en) 2000-11-29 2006-08-29 Hewlett-Packard Development Company, L.P. Method and system for object detection in digital images
US6975352B2 (en) 2000-12-18 2005-12-13 Xerox Corporation Apparatus and method for capturing a composite digital image with regions of varied focus and magnification
US20040080661A1 (en) 2000-12-22 2004-04-29 Sven-Ake Afsenius Camera that combines the best focused parts from different exposures to an image
US20020110286A1 (en) * 2001-02-10 2002-08-15 Cheatle Stephen Philip Method of selectively storing digital images
US20020159632A1 (en) 2001-02-23 2002-10-31 Chui Charles K. Graphic image re-encoding and distribution system and method
US7133069B2 (en) 2001-03-16 2006-11-07 Vision Robotics, Inc. System and method to increase effective dynamic range of image sensors
US20040097206A1 (en) 2001-05-11 2004-05-20 Christian Grewing Circuit configuration for the frequency conversion of an oscillator frequency into a carrier frequency
US20040201699A1 (en) 2001-07-17 2004-10-14 Eastman Kodak Company Revised recapture camera and method
US20030071908A1 (en) 2001-09-18 2003-04-17 Masato Sannoh Image pickup device, automatic focusing method, automatic exposure method, electronic flash control method and computer program
EP1309171A2 (en) 2001-10-23 2003-05-07 STMicroelectronics, Inc. Still picture format for subsequent picture stitching for forming a panoramic image
US6724386B2 (en) 2001-10-23 2004-04-20 Sony Corporation System and process for geometry replacement
US20040201755A1 (en) 2001-12-06 2004-10-14 Norskog Allen C. Apparatus and method for generating multi-image scenes with a camera
US20030147000A1 (en) 2002-01-30 2003-08-07 Kenji Shiraishi Digital still camera, reproduction device, and image processor
US20030189647A1 (en) * 2002-04-05 2003-10-09 Kang Beng Hong Alex Method of taking pictures
US20030190090A1 (en) 2002-04-09 2003-10-09 Beeman Edward S. System and method for digital-image enhancement
US20050122412A1 (en) 2002-04-17 2005-06-09 Seiko Epson Corporation Digital camera
WO2003105466A1 (en) 2002-06-07 2003-12-18 Koninklijke Philips Electronics N.V. Method of imaging an object and mobile imaging device
US20060078224A1 (en) 2002-08-09 2006-04-13 Masashi Hirosawa Image combination device, image combination method, image combination program, and recording medium containing the image combination program
US7623733B2 (en) 2002-08-09 2009-11-24 Sharp Kabushiki Kaisha Image combination device, image combination method, image combination program, and recording medium for combining images having at least partially same background
US20040174434A1 (en) 2002-12-18 2004-09-09 Walker Jay S. Systems and methods for suggesting meta-information to a camera user
WO2004068865A1 (en) 2003-01-24 2004-08-12 Micoy Corporation Steroscopic panoramic image capture device
US20040165788A1 (en) 2003-02-25 2004-08-26 Microsoft Corporation Image blending by guided interpolation
US20040189849A1 (en) 2003-03-31 2004-09-30 Hofer Gregory V. Panoramic sequence guide
US20060171687A1 (en) 2003-04-17 2006-08-03 Seiko Epson Corporation Generation of still image from a plurality of frame images
US20040223649A1 (en) 2003-05-07 2004-11-11 Eastman Kodak Company Composite imaging method and system
US20040239767A1 (en) 2003-05-29 2004-12-02 Stavely Donald J. Systems and methods for providing tactile feedback
US20130300822A1 (en) 2003-09-12 2013-11-14 Sensormatic Electronics, LLC Video user interface system and method
WO2005036780A1 (en) 2003-10-07 2005-04-21 Matsushita Electric Industrial Co., Ltd. Search supporting apparatus and search supporting method
US20070055651A1 (en) 2003-10-07 2007-03-08 Matsushita Electric Industrial Co., Ltd. Apparatus and method for search support
US20050099514A1 (en) 2003-11-08 2005-05-12 Cozier Robert P. Automated zoom control
WO2005050567A1 (en) 2003-11-18 2005-06-02 Scalado Ab Method for processing a digital image and image representation format
US20050168594A1 (en) 2004-02-04 2005-08-04 Larson Brad R. Digital camera and method for in creating still panoramas and composite photographs
US20070274705A1 (en) 2004-05-13 2007-11-29 Kotaro Kashiwa Image Capturing System, Image Capturing Device, and Image Capturing Method
WO2005112437A1 (en) 2004-05-13 2005-11-24 Sony Corporation Image pickup system, image pickup device and image pickup method
US20060001650A1 (en) * 2004-06-30 2006-01-05 Microsoft Corporation Using physical objects to adjust attributes of an interactive display application
US8077213B2 (en) 2004-07-02 2011-12-13 Sony Ericsson Mobile Communications Ab Methods for capturing a sequence of images and related devices
WO2006002796A1 (en) 2004-07-02 2006-01-12 Sony Ericsson Mobile Communications Ab Capturing a sequence of images
EP1613060A1 (en) 2004-07-02 2006-01-04 Sony Ericsson Mobile Communications AB Capturing a sequence of images
US20060022961A1 (en) * 2004-07-28 2006-02-02 Tomomi Kaminaga Reproduction apparatus, camera, and display switch method of reproduction apparatus
US20060028579A1 (en) 2004-08-05 2006-02-09 Yutaka Sato Image pickup apparatus, method of controlling image pickup and program
US20060038908A1 (en) * 2004-08-18 2006-02-23 Canon Kabushiki Kaisha Image processing apparatus, image processing method, program, and storage medium
US7508438B2 (en) 2004-08-30 2009-03-24 Hoya Corporation Digital camera having a bracketing capability
US20060044444A1 (en) 2004-08-30 2006-03-02 Pentax Corporation Digital camera
US20060050152A1 (en) 2004-09-03 2006-03-09 Rai Barinder S Method for digital image stitching and apparatus for performing the same
US20060061845A1 (en) 2004-09-17 2006-03-23 Ulead Systems, Inc. Image composition systems and methods
US20060181619A1 (en) 2005-02-11 2006-08-17 Creative Technology, Ltd. Method and apparatus for forming a panoramic image
US20060181614A1 (en) 2005-02-17 2006-08-17 Jonathan Yen Providing optimized digital images
US20060235765A1 (en) 2005-04-15 2006-10-19 David Clifford R Interactive Image Activation and Distribution System and Associated Methods
US20070274563A1 (en) 2005-06-02 2007-11-29 Searete Llc, A Limited Liability Corporation Of State Of Delaware Capturing selected image objects
US7659923B1 (en) 2005-06-24 2010-02-09 David Alan Johnson Elimination of blink-related closed eyes in portrait photography
WO2007006075A1 (en) 2005-07-14 2007-01-18 Canon Information Systems Research Australia Pty Ltd Image browser
US20090204920A1 (en) 2005-07-14 2009-08-13 Aaron John Beverley Image Browser
US20070025723A1 (en) 2005-07-28 2007-02-01 Microsoft Corporation Real-time preview for panoramic images
US20070024721A1 (en) 2005-07-29 2007-02-01 Rogers Sean S Compensating for improperly exposed areas in digital images
US20070030363A1 (en) 2005-08-05 2007-02-08 Hewlett-Packard Development Company, L.P. Image capture method and apparatus
US20070058064A1 (en) 2005-09-14 2007-03-15 Sony Corporation Image processing apparatus and method, and program therefor
WO2007038198A2 (en) 2005-09-26 2007-04-05 Eastman Kodak Company Image capture method and device also capturing audio
US20070081081A1 (en) 2005-10-07 2007-04-12 Cheng Brett A Automated multi-frame image capture for panorama stitching using motion sensor
CN1750593A (en) 2005-10-13 2006-03-22 上海交通大学 Digital camera with image split function
US20070097206A1 (en) 2005-11-02 2007-05-03 Houvener Robert C Multi-user stereoscopic 3-D panoramic vision system and method
US20090295830A1 (en) * 2005-12-07 2009-12-03 3Dlabs Inc., Ltd. User interface for inspection of photographs
US7639897B2 (en) 2006-01-24 2009-12-29 Hewlett-Packard Development Company, L.P. Method and apparatus for composing a panoramic photograph
US8160152B2 (en) 2006-03-01 2012-04-17 Mitsubishi Electric Corporation Moving image decoding apparatus and moving image coding apparatus
US20070224980A1 (en) 2006-03-24 2007-09-27 Sony Ericsson Mobile Communications Ab Methods, systems, and devices for detecting and indicating loss of proximity between mobile devices
US7787664B2 (en) 2006-03-29 2010-08-31 Eastman Kodak Company Recomposing photographs from multiple frames
US20070237421A1 (en) 2006-03-29 2007-10-11 Eastman Kodak Company Recomposing photographs from multiple frames
US20090046943A1 (en) * 2006-03-31 2009-02-19 Nikon Corporation Image processing method
US20080062141A1 (en) 2006-09-11 2008-03-13 Imran Chandhri Media Player with Imaged Based Browsing
US20100025123A1 (en) 2006-09-29 2010-02-04 Jae Kyung Lee Method of generating key code in coordinate recognition device and video device controller using the same
WO2008038883A1 (en) 2006-09-29 2008-04-03 Lg Electronics Inc. Method of generating key code in coordinate recognition device and video device controller using the same
WO2008064349A1 (en) 2006-11-22 2008-05-29 Nik Software, Inc. Method for dynamic range editing
US20080143744A1 (en) 2006-12-13 2008-06-19 Aseem Agarwala Gradient-domain compositing
US20080152258A1 (en) 2006-12-20 2008-06-26 Jarno Tulkki Digital mosaic image construction
EP1942401A1 (en) 2007-01-05 2008-07-09 Apple Inc. Multimedia communication device with touch screen responsive to gestures for controlling, manipulating and editing of media files
US20080165141A1 (en) 2007-01-05 2008-07-10 Apple Inc. Gestures for controlling, manipulating, and editing of media files using touch sensitive devices
EP2124186A1 (en) 2007-02-15 2009-11-25 Sony Corporation Image processing device and image processing method
US20080218611A1 (en) 2007-03-09 2008-09-11 Parulski Kenneth A Method and apparatus for operating a dual lens camera to augment an image
US20080218613A1 (en) 2007-03-09 2008-09-11 Janson Wilbert F Camera using multiple lenses and image sensors operable in a default imaging mode
US20090019399A1 (en) 2007-07-10 2009-01-15 Brother Kogyo Kabushiki Kaisha Image displaying device, and method and computer readable medium for the same
US20090021576A1 (en) 2007-07-18 2009-01-22 Samsung Electronics Co., Ltd. Panoramic image production
US20090073285A1 (en) 2007-09-14 2009-03-19 Sony Corporation Data processing apparatus and data processing method
US20100268729A1 (en) 2007-11-12 2010-10-21 Megachips Corporation Multimedia synthetic data generating apparatus
US20090141046A1 (en) 2007-12-03 2009-06-04 Apple Inc. Multi-dimensional scroll wheel
US8494306B2 (en) 2007-12-13 2013-07-23 Samsung Electronics Co., Ltd. Method and an apparatus for creating a combined image
US20090190803A1 (en) 2008-01-29 2009-07-30 Fotonation Ireland Limited Detecting facial expressions in digital images
US20090245685A1 (en) 2008-03-25 2009-10-01 Sony Corporation Image processing apparatus, image processing method, and program
US20090244301A1 (en) * 2008-04-01 2009-10-01 Border John N Controlling multiple-image capture
US20090251591A1 (en) 2008-04-04 2009-10-08 Whitham Charles L Digital Camera with High Dynamic Range Mode of Operation
US8249332B2 (en) 2008-05-22 2012-08-21 Matrix Electronic Measuring Properties Llc Stereoscopic measurement system and method
US20090290759A1 (en) 2008-05-22 2009-11-26 Matrix Electronic Measuring, L.P. Stereoscopic measurement system and method
US20090303338A1 (en) 2008-06-06 2009-12-10 Texas Instruments Incorporated Detailed display of portion of interest of areas represented by image frames of a video signal
US8497920B2 (en) 2008-06-11 2013-07-30 Nokia Corporation Method, apparatus, and computer program product for presenting burst images
US20090309990A1 (en) 2008-06-11 2009-12-17 Nokia Corporation Method, Apparatus, and Computer Program Product for Presenting Burst Images
US20090322926A1 (en) 2008-06-25 2009-12-31 Tetsuo Ikeda Image processing apparatus and image processing method
WO2009156561A1 (en) 2008-06-27 2009-12-30 Nokia Corporation Method, apparatus and computer program product for providing image modification
US8463020B1 (en) * 2008-07-08 2013-06-11 Imove, Inc. Centralized immersive image rendering for thin client
JP2010020581A (en) 2008-07-11 2010-01-28 Shibaura Institute Of Technology Image synthesizing system eliminating unnecessary objects
US20100045608A1 (en) * 2008-08-20 2010-02-25 Sony Ericsson Mobile Communications Ab Multidimensional navigation for touch sensitive display
US20100053353A1 (en) 2008-08-27 2010-03-04 Micron Technology, Inc. Method and system for aiding user alignment for capturing partially overlapping digital images
US20100079498A1 (en) * 2008-09-26 2010-04-01 Microsoft Corporation Multi-modal interaction for a screen magnifier
EP2175635A1 (en) 2008-10-10 2010-04-14 Samsung Electronics Co., Ltd. Method and apparatus for creating high dynamic range image
US20100149367A1 (en) 2008-12-17 2010-06-17 Samsung Digital Imaging Co., Ltd. Digital image signal processing apparatus and method of displaying scene recognition
WO2010072587A1 (en) 2008-12-23 2010-07-01 Scalado Ab Subimage extraction from jpeg encoded data
WO2011040864A1 (en) 2009-10-01 2011-04-07 Scalado Ab Method relating to digital images
US20120262490A1 (en) 2009-10-01 2012-10-18 Scalado Ab Method Relating To Digital Images
EP2323102A1 (en) 2009-10-23 2011-05-18 ST-Ericsson (France) SAS Image capturing aid
US20110141227A1 (en) 2009-12-11 2011-06-16 Petronel Bigioi Stereoscopic (3d) panorama creation on handheld device
US20140101590A1 (en) 2010-02-15 2014-04-10 Mobile Imaging In Sweden Ab Digital image manipulation
US8594460B2 (en) 2010-02-15 2013-11-26 Mobile Imaging In Sweden Ab Digital image manipulation
US20110200259A1 (en) 2010-02-15 2011-08-18 Lindskog Alexander Digital image manipulation
US20140177975A1 (en) 2010-02-15 2014-06-26 Mobile Imaging In Sweden Ab Digital image manipulation
US9196069B2 (en) 2010-02-15 2015-11-24 Mobile Imaging In Sweden Ab Digital image manipulation
US9396569B2 (en) 2010-02-15 2016-07-19 Mobile Imaging In Sweden Ab Digital image manipulation
US20120105601A1 (en) 2010-10-27 2012-05-03 Samsung Electronics Co., Ltd. Apparatus and method for creating three-dimensional panoramic image by using single camera
US20140184852A1 (en) 2011-05-31 2014-07-03 Mobile Imaging In Sweden Ab Method and apparatus for capturing images
US9344642B2 (en) 2011-05-31 2016-05-17 Mobile Imaging In Sweden Ab Method and apparatus for capturing a first image using a first configuration of a camera and capturing a second image using a second configuration of a camera
US9432583B2 (en) 2011-07-15 2016-08-30 Mobile Imaging In Sweden Ab Method of providing an adjusted digital image representation of a view, and an apparatus

Non-Patent Citations (90)

* Cited by examiner, † Cited by third party
Title
A.A.Isola, et al., Motion compensated iterative reconstruction of a region of Interest in cardiac cone-beam CT, Computerized Medical Imaging and Graphics, Mar. 1, 2010, vol. 34, No. 2, pp. 149-159, ISSN:0895-6111, abstract.
AGARWALA ET AL.: "interactive digital photomontage", ACM TRANSACTIONS ON GRAPHICS (TOG), ACM, US, vol. 3, no. 23, 1 August 2004 (2004-08-01), US, pages 294 - 302, XP002359799, ISSN: 0730-0301, DOI: 10.1145/1015706.1015718
Aseem Agarwala, et al., Interactive Digital Photomontage, ACM Transactions on Graphics, Aug. 1, 2004, vol. 23, No. 3, pp. 294-302; ISSN: 0730-0301; XP-002359799; figure 1; abstract.
Atallah, M. J. et al., An Optimal Algorithm for Shortest Paths on Weighted Interval and Circular-Arc Graphs, With Applications, Algorithms ESA '93 Sep. 30, 1993 Springer, Berlin, Heidelberg, vol. 726, pp. 13-24.
Baudisch, P. et al., Panoramic Viewfinder: Shooting Panoramic Pictures with the Help of a Real-Time Preview; UIST '05 (Oct. 2005) pp. 1-2.
Baudisch, P., et al., Panoramic Viewfinder: Providing a Real-Time Preview to Help Users Avoid Flaws in Panoramic Pictures, Proceedings of OZCHI, 2005, 10 pages.
Canon: "Powershot S95 Camera User Guide" [online] [retrieved Jun. 1, 2015]. Retrieved from the Internet: UR: http://gdlp01.c-wss.com/gds/4/0300003994/02/PSS95—CUG—EN—02.pdf>. (dated 2010) 196 pages.
Communication and extended search report issued by the European Patent Office in corresponding European patent application No. 07748329.5, dated Mar. 25, 2011.
Communication issued by the European Patent Office in corresponding European patent application No. 07748329.5, dated Mar. 1, 2016.
Communication issued by the European Patent Office in corresponding European patent application No. 07748329.5, dated Mar. 26, 2012.
Communication issued by the European Patent Office in corresponding European patent application No. 10820903.2 dated Jan. 24, 2017.
Decision to Grant a European Patent Pursuant to Article 97(1) EPC for European Patent Application No. 07748329.5 dated Aug. 18, 2017, 2 pages.
Extended European Search Report for corresponding European Application No. 11153998.7 Aug. 12, 2015, 9 pages.
Extended European Search Report for corresponding European Application No. 12814902.8 dated Jun. 1, 2015, 9 pages.
Extended European Search Report for corresponding European Application No. 16177376.7 dated Nov. 14, 2016, 5 pages.
Extended European Search Report for corresponding European Application No. 16177377.5 dated Nov. 14, 2016, 7 pages.
Extended European Search Report from corresponding European Patent Application No. 11827056.0 dated Apr. 11, 2014.
Farin, et al., Shortest Circular Paths on Planar Graphs, 27thSymposium on Information Theory in the Benelux (2006) pp. 117-124.
International Preliminary Report on Patentability from corresponding International Patent Application No. PCT/SE2007/000680 dated Nov. 18, 2008.
International Search Report and Written Opinion for Application No. PCT/SE2012/050688 dated Oct. 26, 2012.
International Search Report for Application No. PCT/SE2007/000680 dated Oct. 30, 2007.
International Search Report for Application No. PCT/SE2011/051124, dated Mar. 12, 2012.
International Search Report.
International Search Report/Written Opinion for Application No. PCT/SE2012/050584 dated Sep. 6, 2012.
International Search Report: mailed Jul. 20, 2010; Search Request No. ITS/SE10/00034.
International Type Search Report and Written Opinion for Application No. ITS/SE10/00290 dated May 2, 2011.
Itti, et al., A Model of Saliency-based Visual Attention for Rapid Scene Analysis, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 20, No. 11 (1998) pp. 1254-1259.
Jiebo Luo et al. Highly Automated Image Recomposition: The Picture You Wish You Had Taken, 2006 Western New York Image Processing Workshop, Sep. 29, 2006, Rochester Institute of Technology, Chester F. Carlson Center for Imaging Science Auditorium, Building 76, Rochester, NY 14623. Download from Internet http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.102.8032&rep=rep1&type=pdf#p.=27.
Ling., H. et al., Diffusion Distance for Histogram Comparison, IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (2006) 8 pages.
Matthews, K. E. et al., Sumiltaneous Motion Parameter Estimation and Image Segmentation using the EM Algorithm, Oct. 23-26, 1995, download from internet: http://ieeexplore.ieee.org/xpl/login.jsp?reload=true&tp=&arnumber=531423&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs—all.jsp%3Farnumber%3D531423.
Notice of Allowability for U.S. Appl. No. 14/233,053 dated May 10, 2016.
Notice of Allowance for U.S. Appl. No. 11/634,264, dated Dec. 2, 2010.
Notice of Allowance for U.S. Appl. No. 13/026,500 dated Jul. 24, 2013.
Notice of Allowance for U.S. Appl. No. 13/026,500 dated May 21, 2013.
Notice of Allowance for U.S. Appl. No. 13/825,214 dated Aug. 24, 2016.
Notice of Allowance for U.S. Appl. No. 13/825,214 dated Aug. 4, 2016.
Notice of Allowance for U.S. Appl. No. 14/037,563 dated Aug. 14, 2014.
Notice of Allowance for U.S. Appl. No. 14/037,563 dated Dec. 12, 2014.
Notice of Allowance for U.S. Appl. No. 14/037,563 dated Jul. 20, 2015.
Notice of Allowance for U.S. Appl. No. 14/037,563 dated Mar. 31, 2015.
Notice of Allowance for U.S. Appl. No. 14/037,708 dated Mar. 22, 2016.
Notice of Allowance for U.S. Appl. No. 14/118,493 dated Dec. 30, 2015.
Notice of Allowance for U.S. Appl. No. 14/118,493 dated Jan. 21, 2016.
Notice of Allowance for U.S. Appl. No. 14/233,053 dated Apr. 29, 2016.
Notice of Allowance from corresponding U.S. Appl. No. 13/825,214 dated Sep. 8, 2016, 13 pages.
Notification of Reason for Rejection in Japanese Patent Application No. 2009-520706 mailed Aug. 30, 2011 (7 pages).
Office Action for Canadian Application No. 2,841,910 dated Apr. 17, 2015.
Office Action for European Application No. 11 15 3998 dated Mar. 9, 2015.
Office Action for European Application No. 12 792 286.1 dated Feb. 15, 2016.
Office Action for European Application No. 12 793 286.1 dated Oct. 2, 2014.
Office Action for European Application No. 12 814 902.8 dated Aug. 2, 2016.
Office Action for U.S. Appl. No. 11/634,264, dated Aug. 20, 2010.
Office Action for U.S. Appl. No. 11/634,264, dated Jan. 26, 2010.
Office Action for U.S. Appl. No. 11/634,264, dated Mar. 24, 2010.
Office Action for U.S. Appl. No. 11/634,264, dated May 22, 2009.
Office Action for U.S. Appl. No. 11/634,264, dated Sep. 4, 2009.
Office Action for U.S. Appl. No. 12/805,850, dated Apr. 8, 2011.
Office Action for U.S. Appl. No. 12/805,850, dated Oct. 12, 2011.
Office Action for U.S. Appl. No. 13/026,500 dated Jan. 2, 2013.
Office Action for U.S. Appl. No. 13/026,500 dated Jul. 31, 2012.
Office Action for U.S. Appl. No. 13/499,711 dated Oct. 5, 2015.
Office Action for U.S. Appl. No. 13/825,214 dated Mar. 24, 2016.
Office Action for U.S. Appl. No. 13/825,214 dated Oct. 21, 2015.
Office Action for U.S. Appl. No. 13/856,843 dated Apr. 22, 2015.
Office Action for U.S. Appl. No. 13/856,843 dated Aug. 26, 2017, 109 pages.
Office Action for U.S. Appl. No. 13/856,843 dated May 2, 2014.
Office Action for U.S. Appl. No. 13/856,843 dated Nov. 10, 2014.
Office Action for U.S. Appl. No. 13/856,843 dated Sep. 6, 2013.
Office Action for U.S. Appl. No. 14/037,708 dated Oct. 19, 2015.
Office Action for U.S. Appl. No. 14/118,493 dated Aug. 4, 2015.
Office Action for U.S. Appl. No. 14/118,493 dated Feb. 6, 2015.
Office Action for U.S. Appl. No. 14/233,053 dated Sep. 11, 2015.
Office Action from corresponding Chinese Patent Application No. 201180052544.7 dated Jan. 29, 2015.
Office Action from corresponding Chinese Patent Application No. 201310132668.7 dated Sep. 5, 2016.
Ojala, T. et al., Multiresolution Gray-Scale and Rotation Invariant Texture Classification With Local Binary Patterns, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, No. 7 (Jul. 2002) 971-987.
Oliva, A. et al., Modeling the Shape of the Scene: A Holistic Representation of the Spatial Envelope, International Journal of Computer Vision 42(3), (2001) 145-175.
Peleg, S. et al., Stereo Panorama With a Single Camera, Proceedings of the 1999 IEEE Computer society conference on Computer Vision and Pattern Recognition, IEEE, vol. 1 (Jun. 1999) pp. 395-401.
Peleg, S., et al., Stereo Panorama with a Single Camera, Proceedings of the 1999 IEEE Computer Society Conference on Computer Vision and Patter Recognition; Jun. 23-25, 1999; Fort Collins, Colorado, IEEE, vol. 1, pp. 395-401.
Shum, H. Y., et al.; "Rendering with Concentric Mosaics;" SIGGRAPH 99; pp. 299-206; dated 1999.
Shum, H-Y. et al., Rendering with Concentric Mosaics, SIGGRAPH 99, ACM (1999) 299-306.
Shutao Li et al., Multifocus Image Fusion Using Region Segmentation and Spatial Frequency, ScienceDirect, Image and Vision Computing, vol. 26 (2008) pp. 971-979.
Summons to attend oral proceedings pursuant to Rule 115(1) EPC in corresponding European patent application No. 07748329.5, dated Nov. 21, 2016.
Supplementary European Search Report for Application No. EP 10 82 0903 dated Dec. 18, 2015.
Supplementary European Search Report for Application No. EP 12 79 3286 dated Feb. 4, 2015.
Swedish Office Action dated Jul. 20, 2010; Ref: SE-21046020.
Third-Party Submission for U.S. Appl. No. 14/037,708 dated Jun. 12, 2015.
Wandell, B. et al., Multiple Capture Image Architecture With a CMOS Sensor, Proceedings of the International Symposium on Multispectral Imaging and Color Reproduction for Digital Archives (Society of Multispectral Imaging of Japan), (Oct. 21-22, 1999) 11-17.
Written Opinion from corresponding International Patent Application No. PCT/SE2007/000680 dated Oct. 30, 2007.
Written Opinion from corresponding International Patent Application No. PCT/SE2011/051124 dated Mar. 12, 2012.
Xiong, Y. et al., Fast Panorama Stitching n Mobile Devices, Consumer Electronics (ICCE), 2010 Digest of Technical Papers International Conference, (Jan. 2010) 319-320.

Also Published As

Publication number Publication date
EP2483767B1 (en) 2019-04-03
WO2011040864A1 (en) 2011-04-07
EP2483767A1 (en) 2012-08-08
US20120262490A1 (en) 2012-10-18
EP2483767A4 (en) 2016-01-20

Similar Documents

Publication Publication Date Title
US10788967B2 (en) Gesture mapping for image filter input parameters
US9792012B2 (en) Method relating to digital images
US20230094025A1 (en) Image processing method and mobile terminal
US11550420B2 (en) Quick review of captured image data
JP5384190B2 (en) Method and apparatus for performing touch adjustment in an imaging device
US9001230B2 (en) Systems, methods, and computer-readable media for manipulating images using metadata
KR101893275B1 (en) Method and apparatus for zooming in on selected area of preview interface
US20090309990A1 (en) Method, Apparatus, and Computer Program Product for Presenting Burst Images
US20150264253A1 (en) Display control apparatus and display control method
US10241660B2 (en) Display control apparatus, method for controlling the same, and storage medium
JP2011050038A (en) Image reproducing apparatus and image sensing apparatus
JP2009500884A (en) Method and device for managing digital media files
JP2013070303A (en) Photographing device for enabling photographing by pressing force to screen, photographing method and program
CN105607825B (en) Method and apparatus for image processing
KR20110062383A (en) Method and apparatus for managing an album
US11442613B2 (en) Electronic apparatus, control method of electronic apparatus, and non-transitory computer readable medium
EP3997558A1 (en) Method for operating an electronic device in order to browse through photos
JP6362110B2 (en) Display control device, control method therefor, program, and recording medium
RU2792413C1 (en) Image processing method and mobile terminal
JP2009077237A (en) Image display device and image display program
CN117528250A (en) Multimedia file processing method, multimedia file processing device and electronic equipment
CN116980759A (en) Shooting method, terminal, electronic device and readable storage medium
KR101477540B1 (en) Photographing apparatus with automatic panning shot function and photographing method for panning shot
JP2017152841A (en) Imaging apparatus, operation control method, and program

Legal Events

Date Code Title Description
AS Assignment

Owner name: SCALADO AB, SWEDEN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NIEMI, SAMI;REEL/FRAME:028177/0140

Effective date: 20120330

AS Assignment

Owner name: MOBILE IMAGING IN SWEDEN AB, SWEDEN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SCALADO AB;REEL/FRAME:028499/0047

Effective date: 20120629

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: NOKIA TECHNOLOGIES OY, FINLAND

Free format text: ASSET TRANSFER AGREEMENT;ASSIGNOR:MOBILE IMAGING IN SWEDEN AB;REEL/FRAME:047900/0899

Effective date: 20171219

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4