WO2004091224A1 - Imaging apparatus and calibration method therefor - Google Patents

Imaging apparatus and calibration method therefor Download PDF

Info

Publication number
WO2004091224A1
WO2004091224A1 PCT/JP2004/005157 JP2004005157W WO2004091224A1 WO 2004091224 A1 WO2004091224 A1 WO 2004091224A1 JP 2004005157 W JP2004005157 W JP 2004005157W WO 2004091224 A1 WO2004091224 A1 WO 2004091224A1
Authority
WO
WIPO (PCT)
Prior art keywords
image data
imaging apparatus
brightness
output
calibration factor
Prior art date
Application number
PCT/JP2004/005157
Other languages
French (fr)
Inventor
Kaneyasu Okawa
Original Assignee
Olympus Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Olympus Corporation filed Critical Olympus Corporation
Priority to EP04726744A priority Critical patent/EP1621024A1/en
Publication of WO2004091224A1 publication Critical patent/WO2004091224A1/en
Priority to US11/244,617 priority patent/US7573506B2/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • H04N17/002Diagnosis, testing or measuring for television systems or their details for television cameras

Definitions

  • the present invention relates to an imaging apparatus evaluation method, an image correction method, and an imaging apparatus .
  • Imaging apparatuses such as a video camera or a digital camera are in heavy usage in general photography, and an output correction such as a dynamic range correction, a visual sensitivity correction or a ⁇ correction is applied to many such imaging apparatuses since a high value is attached to visual characteristics. Therefore, many imaging apparatuses cannot obtain images which correctly reflect the object surface brightness. On the other hand, in imaging apparatuses utilized in some of measurement fields, arrangements are made so that images which correctly reflect the object surface brightness can be obtained. A calibration technique disclosed in Jpn. Pat. Appln. KOKAI No. 6-233333 is one of such techniques.
  • an imaging apparatus evaluation method comprising: obtaining an output function of original image data by comparing a plurality of sets of image data which are obtained by imaging the same spatial brightness distribution surface with different exposure times .
  • an image correction method comprising: obtaining corrected image data by comparing a plurality of sets of image data obtained by imaging the same spatial brightness distribution surface with different exposure times.
  • an image correction method comprising: determining a as a calibration factor and obtaining corrected image data 1 ' (P) relative to original image data I (P) as the following expression:
  • I'(P) I(P) , , w 1 - al(p)
  • an imaging apparatus having a processing portion which obtains corrected image data by comparing a plurality of sets of image data obtained by imaging the same spatial brightness distribution surface with different exposure times.
  • an imaging apparatus having a processing portion which determines a is a calibration factor and obtains corrected image data I'(P) as the following expression with respect to original image data I (P) :
  • FIG. 1 is a block diagram of a first imaging apparatus according to an embodiment of the present invention
  • FIG. 2 is a block diagram of an imaging apparatus evaluation apparatus according to the embodiment of the present invention
  • FIG. 3 is a block diagram of a calibration factor acquisition device according to the embodiment of the present invention.
  • FIG. 4 is a block diagram of a second imaging apparatus according to the embodiment of the present invention.
  • FIG. 5 is a flowchart showing an evaluation processing procedure of the imaging apparatus according to the embodiment of the present invention.
  • FIGS. 6A and 6B are a view showing a characterization factor ⁇ relative to an input brightness before ⁇ correction and a view showing an output brightness relative to the input brightness according to the embodiment of the present invention
  • FIGS. 7A and 7B are a view showing a characterization factor ⁇ relative to an input brightness after ⁇ correction and a view showing an output brightness relative to the input brightness according to the embodiment of the present invention
  • FIG. 8 is a view showing an output sensitivity value (reciprocal of the characterization factor) relative to an input brightness according to the embodiment of the present invention.
  • FIG. 9 is a flowchart showing an image correction processing procedure by the imaging apparatus according to the embodiment of the present invention.
  • FIG. 10 is a flowchart showing a first calibration factor acquisition processing procedure by the imaging apparatus according to the embodiment of the present invention.
  • FIG. 11 is a view showing a calibration factor value relative to a first input brightness according to the embodiment of the present invention.
  • FIG. 12 is a view showing a calibration factor value relative to a second input brightness according to the embodiment of the present invention.
  • FIG. 13 is -a view showing a calibration factor value relative to a third input brightness according to the embodiment of the present invention
  • FIG. 14 is a flowchart showing a second calibration factor acquisition processing procedure by the imaging apparatus according to the embodiment of the present invention
  • FIG. 15 is a view showing a characterization factor value relative to a first input brightness according to the embodiment of the present invention.
  • FIG. 16 is a view showing a characterization factor value relative to a second input brightness according to the embodiment of the present invention.
  • FIG. 17 is a view showing a characterization factor value relative to a third input brightness according to the embodiment of the present invention. Best Mode for Carrying Out the Invention Embodiments according to the present invention will now be described hereinafter with reference to the accompanying drawings.
  • FIG. 1 is a block diagram of a first imaging apparatus according to an embodiment of the present invention.
  • a corrected image processing portion 2 is connected to an imaging apparatus main body 1, and an image output portion 3 is connected to the corrected image processing portion 2.
  • a control portion 10 is connected to the imaging apparatus main body 1, the corrected image processing portion 2 and the image output portion 3.
  • the imaging apparatus main body 1 comprises an imaging element such as a CCD.
  • the corrected image processing portion 2 comprises a computer.
  • the image output portion 3 comprises a memory device.
  • the control portion 10 comprises a CPU.
  • the imaging apparatus main body 1 acquires image data I (P) .
  • the corrected image processing portion 2 receives the image data I (P) , and calculates the following expression, as will be described later. I(P)
  • the image output portion 3 receives the image data I ' (P) , stores it and appropriately outputs it.
  • FIG. 2 is a block diagram of an imaging apparatus evaluation apparatus according to an embodiment of the present invention.
  • like reference numerals denote parts equivalent to those in FIG. 1.
  • an output function calculation portion 11 is connected to an imaging apparatus main body 1
  • an output function output portion 12 is connected to the output function calculation portion 11.
  • a control portion 10 is connected to the imaging apparatus main body 1, the output function calculation portion 11 and the output function output portion 12.
  • FIG. 3 is a block diagram of a calibration factor acquisition device according to an embodiment of the present invention.
  • like reference numerals denote parts equivalent to those in FIGS. 1 and 2.
  • FIG. 3 is a block diagram of a calibration factor acquisition device according to an embodiment of the present invention.
  • an output function calculation portion 11 is connected to an imaging apparatus main body 1
  • a calibration factor calculation portion 21 is connected to the output function calculation portion 11
  • a calibration factor output portion 22 is connected to the calibration factor calculation portion 21.
  • a control portion 10 is connected to the imaging apparatus main body 1, the output function calculation portion 11, the calibration factor calculation portion 21 and the calibration factor output portion 22.
  • FIG. 4 is a block diagram of a second imaging apparatus according to an embodiment of the present invention.
  • like reference numerals denote parts equivalent to those in FIG. 1, FIG. 2 and FIG. 3.
  • an output function calculation portion 11 is connected to an imaging apparatus main body 1
  • a calibration factor calculation portion 21 is connected to the output function calculation portion 11
  • a corrected image processing portion 2 is connected to the calibration factor calculation portion 21, and an image output portion 3 is connected to the corrected image processing portion 2.
  • a control portion 10 is connected to the imaging apparatus main body 1, the output function calculation portion 11, the calibration factor calculation portion 21, the corrected image processing portion 2 and the image output portion 3.
  • first image data is obtained by imaging a spatial brightness distribution surface with a given exposure time and second image data is obtained by imaging the same spatial brightness distribution surface with an exposure time which is k-fold of that of the first image data by using a digital imaging apparatus formed of an imaging apparatus main body 1, wherein an output function of original image data is f(P).
  • This spatial brightness distribution surface may be an object surface dedicated to acquire the later-described output function f (P) and acquire a calibration factor a or it may be an actual object for photography. If the object surface is an actual object, it is sufficient for the imaging apparatus to have the structure shown in FIG.
  • an image picked up with an adequate exposure is determined as an object image, and an image picked up with an exposure time which is k-fold of this adequate exposure is determined as a calibration image used to acquire the calibration factor a. It is to be noted that this calibration image is no longer necessary after the calibration factor a is acquired, and hence it is desirable to delete this calibration image in order to save a space in a memory device.
  • Object point brightnesses on the spatial brightness distribution surface are determined as P Q , P]_, P2, P3, -, P n - i n ascending order, and an exposure sensitivity R n is defined by using the following expression:
  • f (P n ) and f(kP n ) are output brightnesses of the first image data and the second image data corresponding to each object point brightness P n , respectively. That is, the exposure sensitivity R n becomes 1 irrespective of k when the output function f (P) is in proportion to the object point brightness P. However, when it is not proportion to the same, the exposure sensitivity becomes an increasing function or a decreasing function in accordance with the output function f(P). Further, if P n -P n -i i s small, the following expression can be achieved approximately:
  • corrected image data I ' (P) obtained based on the expression (7) correctly reflects the object point brightness irrespective of types of the original image data I (P) . Therefore, when the corrected image data I' (P) is represented by using the original image data I (P) and the calibration factor a, the following expression can be attained:
  • FIG. 5 is a flowchart showing an evaluation processing procedure of an imaging apparatus according to this embodiment. This processing is executed under the control of the control portion 10 in the apparatus depicted in FIG. 2. Incidentally, it is assumed that p max i s a maximum gradation value; P m i n , a minimum gradation value; P s tep' a gradation step; t, an exposure time of the first image data; and k, an exposure time ratio in the following description.
  • the imaging apparatus main body 1 acquires first image data I with an exposure time t.
  • the imaging apparatus main body 1 acquires second image data I' with an exposure time t' .
  • the output function calculation portion 11 calculates the following expression:
  • a characterization factor ⁇ n is an inverse number of the output sensitivity value
  • f(P n ) is an output function.
  • n n max i s achieved at step SI07
  • a segment ⁇ g' of an approximate straight line of ⁇ n is calculated at step SI09. Additionally, the output function calculation portion 11 calculates the following expression by using this ⁇ g ' :
  • the output function f(P n ) is stored in the output function output portion 12 comprising a memory device.
  • the characterization factor ⁇ is corrected, and there can be obtained a graph showing the characterization factor ⁇ relative to the input brightness such as shown in FIG. 7A and a graph showing the output brightness relative to the input brightness such as shown in FIG. 7B.
  • FIG. 7A a graph showing an output sensitivity value (inverse value of the characterization factor) relative to the input brightness such as shown in FIG. 8.
  • the characterization factor is always substantially 1 with respect to a change in the input brightness in the graph of FIG. 7A
  • the output sensitivity value is always substantially 1 with respect to a change in the input brightness in the graph of FIG. 8
  • the input brightness and the output brightness are substantially in proportion to each other in the graph of FIG. 7B.
  • the object point brightness is not correctly reflected in this imaging apparatus. Then, correction processing is carried out so as to correctly reflect the object point brightness.
  • FIG. 9 is a flowchart showing an image correction processing procedure performed by the imaging apparatus according to this embodiment. This processing is executed under the control of the control portion 10 in the apparatus shown in FIG. 4.
  • an output function f (P) is acquired by the output function calculation portion 11.
  • a calibration factor a is acquired by the calibra- tion factor calculation portion 21.
  • the corrected image processing portion 2 calculates the following expression:
  • FIG. 10 is a flowchart showing a first calibration factor acquisition processing procedure effected by the imaging apparatus according to this embodiment.
  • This processing corresponds to step S2 in FIG. 9, and it is executed under the control of the control portion 10 in the apparatus shown in FIG. 4. Moreover, this processing can be also executed by the control of the control portion 10 in the apparatus shown in FIG. 3.
  • P max is a maximum gradation value
  • P m in' a minimum gradation value
  • p step' a gradation step
  • t an exposure time of the first image data
  • k an exposure time ratio in the following description.
  • the imaging apparatus main body 1 acquires first image data I with an exposure time t.
  • the imaging apparatus main body 1 acquires second image data I ' with an exposure time t' .
  • the calibration factor calculation portion 21 acquires binarized image data of the first image data I having a threshold value of P n which is represented by the following expression: ⁇ n
  • the calibration factor calculation portion 21 acquires an area S n of a space part of the binarized image data represented by the following expression: ⁇ n
  • a threshold value f (kP n ) with which an area of the space part of the binarized image data represented by the following expression becomes S n: ⁇ n
  • the calibration factor calculation portion 21 calculates the following expression:
  • FIG. 13 are graphs showing a first calibration factor A(P) , a second calibration factor A(P) and a third calibration factor A(P) respectively obtained by the above-described processing, i.e., calibration function values relative to the input brightnesses.
  • is satisfied based on the third calibration function A(P)
  • the calibration function value is constantly substantially 0 with respect to the input brightness, and a at this time becomes the calibration factor.
  • a factor in the conditional expression is not restricted to 0.001, and it can be appropriately changed.
  • FIG. 14 is a flowchart showing a second calibration factor acquisition processing procedure executed by the imaging apparatus according to this embodiment.
  • This processing corresponds to the step S2 in FIG. 9, and it is executed under the control of the control portion 10 in the apparatus shown in FIG. 4. Moreover, this processing can be also executed under the control of the control portion 10 in the apparatus depicted in FIG. 3.
  • p max i s a maximum gradation value
  • P m i n a minimum gradation value
  • P s ep' a gradation step t
  • t an exposure time of first image data
  • k an exposure time ratio in the following description.
  • the imaging apparatus main body 1 acquires first image data I with an exposure time t.
  • the calibration factor calculation portion 21 determines t' - kt.
  • the imaging apparatus main body 1 acquires second image data I' with an exposure time t' .
  • the calibration factor calculation portion 21 acquires binarized image data of first image data I having a threshold value of P n which is represented by the following expression: ⁇ n
  • the calibration factor calculation portion 21 acquires an area S n of a space part of the binarized image data represented by the following expression: ⁇ n Additionally, it acquires a threshold value f (kP n ) with which an area of the space part of the binarized image data represented by the following expression becomes > n ;
  • the calibration factor calculation portion 21 calculates the following expression:
  • the calibration factor calculation portion 21 calculates the following expression:
  • FIG. 15, FIG. 16 and FIG 17 are graphs showing a first characterization factor ⁇ , a second charac- terization factor ⁇ and a third characterization factor ⁇ obtained by the above-described processing, i.e., characterization factor values relative to each input brightness.
  • the condition of I a' I ⁇ 0.001
  • FIG. 10 and the second calibration factor acquisition processing shown in FIG. 14 is executed in the processing to acquire the calibration factor a at the step S2 in FIG. 9.
  • the corrected image processing portion 2 in the apparatus shown in FIG. 4 calculates the following expression at step S3 in FIG. 9:
  • corrected image data I' (P) relative to the original image data I (P) can be obtained.
  • This corrected image data I' (P) is stored in the image output portion 3.
  • the calibration factor a in the expression (8) By assigning the calibration factor a in the expression (8) in this manner, a corrected image in which an object surface brightness is correctly reflected can be obtained in the subsequent photography using the imaging apparatus. That is, an image which correctly reflects the object surface brightness can be obtained without preparing an object surface having a known brightness distribution at the time of regular photography using the imaging apparatus.
  • the calibration factor a is calculated as described above at the time of shipment of the imaging apparatus, and it is assigned in the above-described expression of the corrected image processing portion 2. Therefore, in case of the first imaging apparatus having the structure shown in FIG. 1, just assigning the calibration factor a which is measured by another device in advance in the expression of the corrected image processing portion 2 can perform setting required to obtain a corrected image in which an object surface brightness is correctly reflected.
  • an imaging apparatus evaluation method which correctly evaluates output characteristics of an imaging apparatus without preparing an object surface having a known brightness distribution.
  • an image correction method which obtains an image which correctly reflects an object surface brightness as a corrected image without preparing an object surface having a known brightness distribution.
  • an imaging apparatus which can obtain an image which correctly reflects an object surface brightness without preparing an object surface having a known brightness distribution.

Abstract

An imaging apparatus evaluation method of the present invention comprises obtaining an output function of original image data by comparing a plurality of sets of image data which are obtained by imaging the same spatial brightness distribution surface with different exposure times.

Description

D E S C R I P T I O N
IMAGING APPARATUS AND CALIBRATION METHOD THEREFOR
Technical Field The present invention relates to an imaging apparatus evaluation method, an image correction method, and an imaging apparatus . Background Art
Imaging apparatuses such as a video camera or a digital camera are in heavy usage in general photography, and an output correction such as a dynamic range correction, a visual sensitivity correction or a γ correction is applied to many such imaging apparatuses since a high value is attached to visual characteristics. Therefore, many imaging apparatuses cannot obtain images which correctly reflect the object surface brightness. On the other hand, in imaging apparatuses utilized in some of measurement fields, arrangements are made so that images which correctly reflect the object surface brightness can be obtained. A calibration technique disclosed in Jpn. Pat. Appln. KOKAI No. 6-233333 is one of such techniques.
However, in an imaging apparatus using this calibration technique, a standard light emitting surface is essential. This standard light emitting surface is required for photography every time, quality management is troublesome, and deterioration occurs. Therefore, in this type of imaging apparatus, it cannot be said that an evaluation accuracy of output charac- teristics and a calibration accuracy of images are enough.
Disclosure of Invention It is a first object of the present invention to provide an imaging apparatus evaluation method which correctly evaluates output characteristics of an imaging apparatus without preparing an object surface having a known brightness distribution.
It is a second object of the present invention to provide an image correction method which obtains an image which correctly reflects an object surface brightness as a corrected image without preparing an object surface having a known brightness distribution. It is a third object of the present invention to provide an imaging apparatus which can obtain an image which correctly reflects an object surface brightness without preparing an object surface having a known brightness distribution.
According to the present invention, there is provided an imaging apparatus evaluation method comprising: obtaining an output function of original image data by comparing a plurality of sets of image data which are obtained by imaging the same spatial brightness distribution surface with different exposure times .
According to the present invention, there is provided an image correction method comprising: obtaining corrected image data by comparing a plurality of sets of image data obtained by imaging the same spatial brightness distribution surface with different exposure times.
According to the present invention, there is provided an image correction method comprising: determining a as a calibration factor and obtaining corrected image data 1 ' (P) relative to original image data I (P) as the following expression:
I'(P) = I(P), , w 1 - al(p)
According to the present invention, there is provided an imaging apparatus having a processing portion which obtains corrected image data by comparing a plurality of sets of image data obtained by imaging the same spatial brightness distribution surface with different exposure times.
According to the present invention, there is provided an imaging apparatus having a processing portion which determines a is a calibration factor and obtains corrected image data I'(P) as the following expression with respect to original image data I (P) :
Figure imgf000005_0001
Brief Description of Drawings FIG. 1 is a block diagram of a first imaging apparatus according to an embodiment of the present invention; FIG. 2 is a block diagram of an imaging apparatus evaluation apparatus according to the embodiment of the present invention;
FIG. 3 is a block diagram of a calibration factor acquisition device according to the embodiment of the present invention;
FIG. 4 is a block diagram of a second imaging apparatus according to the embodiment of the present invention;
FIG. 5 is a flowchart showing an evaluation processing procedure of the imaging apparatus according to the embodiment of the present invention;
FIGS. 6A and 6B are a view showing a characterization factor μ relative to an input brightness before μ correction and a view showing an output brightness relative to the input brightness according to the embodiment of the present invention;
FIGS. 7A and 7B are a view showing a characterization factor μ relative to an input brightness after μ correction and a view showing an output brightness relative to the input brightness according to the embodiment of the present invention;
FIG. 8 is a view showing an output sensitivity value (reciprocal of the characterization factor) relative to an input brightness according to the embodiment of the present invention;
FIG. 9 is a flowchart showing an image correction processing procedure by the imaging apparatus according to the embodiment of the present invention;
FIG. 10 is a flowchart showing a first calibration factor acquisition processing procedure by the imaging apparatus according to the embodiment of the present invention;
FIG. 11 is a view showing a calibration factor value relative to a first input brightness according to the embodiment of the present invention;
FIG. 12 is a view showing a calibration factor value relative to a second input brightness according to the embodiment of the present invention;
FIG. 13 is -a view showing a calibration factor value relative to a third input brightness according to the embodiment of the present invention; FIG. 14 is a flowchart showing a second calibration factor acquisition processing procedure by the imaging apparatus according to the embodiment of the present invention;
FIG. 15 is a view showing a characterization factor value relative to a first input brightness according to the embodiment of the present invention;
FIG. 16 is a view showing a characterization factor value relative to a second input brightness according to the embodiment of the present invention; and
FIG. 17 is a view showing a characterization factor value relative to a third input brightness according to the embodiment of the present invention. Best Mode for Carrying Out the Invention Embodiments according to the present invention will now be described hereinafter with reference to the accompanying drawings.
FIG. 1 is a block diagram of a first imaging apparatus according to an embodiment of the present invention. In FIG. 1, a corrected image processing portion 2 is connected to an imaging apparatus main body 1, and an image output portion 3 is connected to the corrected image processing portion 2. A control portion 10 is connected to the imaging apparatus main body 1, the corrected image processing portion 2 and the image output portion 3. The imaging apparatus main body 1 comprises an imaging element such as a CCD.
The corrected image processing portion 2 comprises a computer. The image output portion 3 comprises a memory device. The control portion 10 comprises a CPU. The imaging apparatus main body 1 acquires image data I (P) . The corrected image processing portion 2 receives the image data I (P) , and calculates the following expression, as will be described later. I(P)
I'(P) =
1 - al(p)
The image output portion 3 receives the image data I ' (P) , stores it and appropriately outputs it.
FIG. 2 is a block diagram of an imaging apparatus evaluation apparatus according to an embodiment of the present invention. In FIG. 2, like reference numerals denote parts equivalent to those in FIG. 1. In FIG. 2, an output function calculation portion 11 is connected to an imaging apparatus main body 1, and an output function output portion 12 is connected to the output function calculation portion 11. A control portion 10 is connected to the imaging apparatus main body 1, the output function calculation portion 11 and the output function output portion 12. FIG. 3 is a block diagram of a calibration factor acquisition device according to an embodiment of the present invention. In FIG. 3, like reference numerals denote parts equivalent to those in FIGS. 1 and 2. In FIG. 3, an output function calculation portion 11 is connected to an imaging apparatus main body 1, a calibration factor calculation portion 21 is connected to the output function calculation portion 11, and a calibration factor output portion 22 is connected to the calibration factor calculation portion 21. A control portion 10 is connected to the imaging apparatus main body 1, the output function calculation portion 11, the calibration factor calculation portion 21 and the calibration factor output portion 22.
FIG. 4 is a block diagram of a second imaging apparatus according to an embodiment of the present invention. In FIG. 4, like reference numerals denote parts equivalent to those in FIG. 1, FIG. 2 and FIG. 3. In FIG. 4, an output function calculation portion 11 is connected to an imaging apparatus main body 1, a calibration factor calculation portion 21 is connected to the output function calculation portion 11, a corrected image processing portion 2 is connected to the calibration factor calculation portion 21, and an image output portion 3 is connected to the corrected image processing portion 2. A control portion 10 is connected to the imaging apparatus main body 1, the output function calculation portion 11, the calibration factor calculation portion 21, the corrected image processing portion 2 and the image output portion 3.
Effects of this embodiment will now be described hereinafter. Now, it is assumed that first image data is obtained by imaging a spatial brightness distribution surface with a given exposure time and second image data is obtained by imaging the same spatial brightness distribution surface with an exposure time which is k-fold of that of the first image data by using a digital imaging apparatus formed of an imaging apparatus main body 1, wherein an output function of original image data is f(P).' This spatial brightness distribution surface may be an object surface dedicated to acquire the later-described output function f (P) and acquire a calibration factor a or it may be an actual object for photography. If the object surface is an actual object, it is sufficient for the imaging apparatus to have the structure shown in FIG. 4, an image picked up with an adequate exposure is determined as an object image, and an image picked up with an exposure time which is k-fold of this adequate exposure is determined as a calibration image used to acquire the calibration factor a. It is to be noted that this calibration image is no longer necessary after the calibration factor a is acquired, and hence it is desirable to delete this calibration image in order to save a space in a memory device.
Object point brightnesses on the spatial brightness distribution surface are determined as PQ , P]_, P2, P3, -, Pn - in ascending order, and an exposure sensitivity Rn is defined by using the following expression:
Figure imgf000011_0001
Here, f (Pn) and f(kPn) are output brightnesses of the first image data and the second image data corresponding to each object point brightness Pn, respectively. That is, the exposure sensitivity Rn becomes 1 irrespective of k when the output function f (P) is in proportion to the object point brightness P. However, when it is not proportion to the same, the exposure sensitivity becomes an increasing function or a decreasing function in accordance with the output function f(P). Further, if Pn-Pn-i is small, the following expression can be achieved approximately:
φ%_!) = f(Pn-l) + £(pn) ~ i-l) (tø^ _ p^) pn - pn-l
(2) The following expression can be obtained based on the expression (1) and the expression (2) :
f( )
Figure imgf000012_0001
(3) However, in case of Pfj'÷ 0/ since it can be considered that a value of the output function is equal to the object point brightness, the following expression can be achieved: f(P0 ) = Pθ (4) A data string f (Pι ) , f ( P2 ) , ( P3 ) , •••/ f ( P ) ' - corresponding to P^, P , P3, ..., Pn, ... can be obtained by using the expression (1), the expression (3) and the expression (4), and forming a function of this data string results in the output function f (P) of the original image data. Furthermore, assuming that an output function after calibration is g(P) and a calibration function is A(P) , the following expression can be achieved approximately:
Figure imgf000013_0001
Here, if g(P) = P, i.e., if the output function g(P) after calibration satisfies conditions which correctly reflect the object point brightness, the following expression can be achieved based on the expression (5) :
A(p) = i ) - ? ( 6>
Using the expression (6) can calculate the calibration factor A(P) used for the output function g(P) after calibration to correctly reflect the object point brightness. Now, when an average value of the calibration factor A(P) is defined as the calibration factor a and A(P) in the expression (5) is substituted with a, the following expression can be obtained:
Figure imgf000013_0002
When the right-side member of the expression (7) is substituted with the output function f (P) of the original image data and the calibration factor a, the output function g(P) after calibration which is substantially equal to P can be obtained.
Since the output function f (P) of the original image data remains the same in the same imaging apparatus irrespective types of the spatial brightness distribution surface, corrected image data I ' (P) obtained based on the expression (7) correctly reflects the object point brightness irrespective of types of the original image data I (P) . Therefore, when the corrected image data I' (P) is represented by using the original image data I (P) and the calibration factor a, the following expression can be attained:
Figure imgf000014_0001
FIG. 5 is a flowchart showing an evaluation processing procedure of an imaging apparatus according to this embodiment. This processing is executed under the control of the control portion 10 in the apparatus depicted in FIG. 2. Incidentally, it is assumed that pmax is a maximum gradation value; Pmin, a minimum gradation value; Pstep' a gradation step; t, an exposure time of the first image data; and k, an exposure time ratio in the following description.
At step S101, the imaging apparatus main body 1 acquires first image data I with an exposure time t. The output function calculation portion 11 determines t' = kt. The imaging apparatus main body 1 acquires second image data I' with an exposure time t' . The output function calculation portion 11 calculates nmax = (P ax ~ pmin) / pstep' and determines n = 0. At step SI02, the output function calculation portion 11 calculates Pn = Pmin + nPstep* ^ne output function calculation portion 11 acquires binarized image data of the first image data I having a threshold value of Pn, the binary image data being represented by the following expression: In The output function calculation portion 11 acquires an area Sn of a space part of the binarized image data represented by the following expression: ϊn
Moreover, it acquires a threshold value f (kPn) with which an area of the space part of the binarized image data represented by the following expression becomes
Sn: ϊn
If n = nmax is not achieved at step S103, the output function calculation portion 11 determines n = n + 1 at step S104, and returns to step S102. If n = nmax is achieved at step S103, the output function calculation portion 11 determines f (Prj) = θ -O = 1 and n = 1 at step S105.
At step SI06, the output function calculation portion 11 calculates the following expression:
Figure imgf000015_0001
f(Pn) = Pn-l - Pn + KRn-ifa - Pn-l) f( ^ = _
(k-l)Pn-l f(Pn)
If n = nmax is not achieved at step SI07, the output function calculation portion 11 determines n = n + 1 at step S108, and returns to step S106. It is to be noted that a characterization factor μn is an inverse number of the output sensitivity value, and f(Pn) is an output function. As a result, there can be obtained a graph showing the characterization factor μ relative to the input brightness such as shown in FIG. 6A and a graph showing an output brightness f (P) relative to the input brightness P, i.e., the output brightness such as shown in FIG. 6B.
If n = nmax is achieved at step SI07, a segment μg' of an approximate straight line of μn is calculated at step SI09. Additionally, the output function calculation portion 11 calculates the following expression by using this μg ' :
Figure imgf000016_0001
As a result, f (PQ) = Po and pg = 1 can be obtained. This processing is performed to match the output brightness to the input brightness in the vicinity of the brightness 0 and set the output sensitivity value to 1. Further, the output function calculation portion 11 determines n = 1. At step S110, the output function calculation portion 11 calculates the following expression:
f(Pn) .
Figure imgf000016_0002
If n = nmax is not achieved at step Sill, the output function calculation portion 11 determines n = n + 1 at step S112, and returns to step S110. It is to be noted that p(Pn) is an output sensitivity, and f (Pn) is an output function. If n = nmax is achieved at step Sill, the output function calculation portion 11 terminates the processing. The output function f(Pn) is stored in the output function output portion 12 comprising a memory device. As a result, the characterization factor μ is corrected, and there can be obtained a graph showing the characterization factor μ relative to the input brightness such as shown in FIG. 7A and a graph showing the output brightness relative to the input brightness such as shown in FIG. 7B. Furthermore, there is obtained from FIG. 7A a graph showing an output sensitivity value (inverse value of the characterization factor) relative to the input brightness such as shown in FIG. 8. If the object point brightness is correctly reflected in the imaging apparatus, the characterization factor is always substantially 1 with respect to a change in the input brightness in the graph of FIG. 7A, the output sensitivity value is always substantially 1 with respect to a change in the input brightness in the graph of FIG. 8, and the input brightness and the output brightness are substantially in proportion to each other in the graph of FIG. 7B. However, as apparent from the evaluation result of FIGS. 7A, 7B and FIG. 8, the object point brightness is not correctly reflected in this imaging apparatus. Then, correction processing is carried out so as to correctly reflect the object point brightness.
FIG. 9 is a flowchart showing an image correction processing procedure performed by the imaging apparatus according to this embodiment. This processing is executed under the control of the control portion 10 in the apparatus shown in FIG. 4.
At step SI, an output function f (P) is acquired by the output function calculation portion 11. At step S2, a calibration factor a is acquired by the calibra- tion factor calculation portion 21. At step S3, the corrected image processing portion 2 calculates the following expression:
1 - al(p) This corrected image data I' (P) is stored in the image output portion 3.
FIG. 10 is a flowchart showing a first calibration factor acquisition processing procedure effected by the imaging apparatus according to this embodiment. This processing corresponds to step S2 in FIG. 9, and it is executed under the control of the control portion 10 in the apparatus shown in FIG. 4. Moreover, this processing can be also executed by the control of the control portion 10 in the apparatus shown in FIG. 3. Incidentally, it is assumed that Pmax is a maximum gradation value; Pmin' a minimum gradation value; pstep' a gradation step; t, an exposure time of the first image data; and k, an exposure time ratio in the following description.
At step S201, the imaging apparatus main body 1 acquires first image data I with an exposure time t. The calibration factor calculation portion 21 determines t' = kt. The imaging apparatus main body 1 acquires second image data I ' with an exposure time t' . The calibration factor calculation portion 21 calculates nmax = (Pmax ~ Pmin) / Pstep' anc^ determines a = 0 and A]_ = 0. The calibration factor calculation portion 21 determines n = 0 at step S202, and the calibration factor calculation portion 21 calculates Pn = Pmin + nPs-ep at step S203. The calibration factor calculation portion 21 acquires binarized image data of the first image data I having a threshold value of Pn which is represented by the following expression: ϊn The calibration factor calculation portion 21 acquires an area Sn of a space part of the binarized image data represented by the following expression: ϊn Furthermore, it acquires a threshold value f (kPn) with which an area of the space part of the binarized image data represented by the following expression becomes Sn: ϊn
If n = nmax is not achieved at step S204, the calibration factor calculation portion 21 determines n = n + 1 at step S205, and returns to step S203. If n = nmax is achieved at step S204, the calibration factor calculation portion 21 determines f (PQ) = Pø a' = 0, S = 00 and n = 1 at step S206.
At step S207, the calibration factor calculation portion 21 calculates the following expression:
R η - f(kpn~l)
Figure imgf000020_0001
A(Pn) = - ^~ , a' = a' + A(Pn) τ n) en If n = nmax is not achieved at step S208, the calibration factor calculation portion 21 determines n = n + 1 at step S209, and returns to step S207. If n = nmax is achieved at step S208, the calibration factor calculation portion 21 calculates the following expression:
Figure imgf000020_0002
If |a'| < 0.001 |a| is not achieved at step S211, the calibration factor calculation portion 21 calculates a = a + a', 1 = 1 / (l-al) and I' = I' / (1 - al') at step S212. If | a | < 0.001 |a| is achieved at step S211, the calibration factor calculation portion 21 terminates the processing. As a result, an appropriate calibration factor a is obtained. It is to be noted that the calibration factor a is stored in the calibration factor output portion 22 comprising the memory device in the apparatus shown in FIG. 3. FIG. 11, FIG. 12 and FIG. 13 are graphs showing a first calibration factor A(P) , a second calibration factor A(P) and a third calibration factor A(P) respectively obtained by the above-described processing, i.e., calibration function values relative to the input brightnesses. As shown in FIG. 13, a conditional expression of |a'| < 0.001 |a| is satisfied based on the third calibration function A(P), the calibration function value is constantly substantially 0 with respect to the input brightness, and a at this time becomes the calibration factor. It is to be noted that a factor in the conditional expression is not restricted to 0.001, and it can be appropriately changed.
FIG. 14 is a flowchart showing a second calibration factor acquisition processing procedure executed by the imaging apparatus according to this embodiment. This processing corresponds to the step S2 in FIG. 9, and it is executed under the control of the control portion 10 in the apparatus shown in FIG. 4. Moreover, this processing can be also executed under the control of the control portion 10 in the apparatus depicted in FIG. 3. Incidentally, it is assumed that pmax is a maximum gradation value; Pmin a minimum gradation value; Ps ep' a gradation step; t, an exposure time of first image data; and k, an exposure time ratio in the following description. At step S301, the imaging apparatus main body 1 acquires first image data I with an exposure time t. The calibration factor calculation portion 21 determines t' - kt. The imaging apparatus main body 1 acquires second image data I' with an exposure time t' . The calibration factor calculation portion 21 calculates nmax = (Pmax ~ P in) / pstep' and determines a = 0 and A]_ = 0.
The calibration factor calculation portion 21 determines n = 0 at step S302, and the calibration factor calculation portion 21 calculates Pn = Pmin + nPstep at step S303. The calibration factor calculation portion 21 acquires binarized image data of first image data I having a threshold value of Pn which is represented by the following expression: ϊn
The calibration factor calculation portion 21 acquires an area Sn of a space part of the binarized image data represented by the following expression: ϊn Additionally, it acquires a threshold value f (kPn) with which an area of the space part of the binarized image data represented by the following expression becomes >n ;
-41
If n = nmax is not achieved at step S304, the calibration factor calculation portion 21 determines n = n + 1 at step S305, and returns to step S303. If n = nmax is achieved at a step S304, the calibration factor calculation portion 21 determines f (PQ) = Po^ a' = 0, S = Og and n = 1 at a step S306.
At step S307, the calibration factor calculation portion 21 calculates the following expression:
Figure imgf000023_0001
If n = nmax is not achieved at step S308, the calibration factor calculation portion 21 determines n = n + 1 at step S309, and returns to step S307. If n = nmax is achieved at step S308, at step S310, the calibration factor calculation portion 21 calculates a segment of an approximate straight line of μg' = μn represented by the following expression in order to obtain a segment μg = 1 of an approximate straight line of μn:
Figure imgf000023_0002
Further, the calibration factor calculation portion 21 determines n = 1.
At step S311, the calibration factor calculation portion 21 calculates the following expression:
) - kpn-l ~ pn + kRn-l(Pn ~ pn-l) „ _ ?n ( n} ~ (k - D Pn-1 f(Pn-1 μn - f )
If n = nmax is not achieved at step S312, the calibration factor calculation portion 21 determines n = n + 1 at step S313, and returns to step S311. If n = nmax is achieved at step S312, the calibration factor calculation portion 21 obtains an approximate gradient a' of the approximate straight line of μn and acquires a square-sum of a deviation between the approximate straight line of μn and μn at step S314. It is to be noted that the second calibration factor acquisition processing is executed for multiple times while changing the exposure time ratio k, and k with which the square-sum of the deviation from the approximate straight line represented by the following expression becomes minimum is used, thereby obtaining a further accurate calibration factor:
μ(p) = P f(p)
If I a' I < 0.001 |a| is not achieved at step S315, the calibration factor calculation portion 21 calculates a = a + a' , 1 = 1 / (1 - al) and I' = I " / (1 - al') at step S316. If | a' | < 0.001 |a| is achieved at step S315, the calibration factor calculation portion 21 terminates the processing. As a result, an appropriate calibration factor a can be obtained. It is to be noted that the calibration factor a is stored in the calibration factor output portion 22 in the apparatus shown in FIG. 3.
FIG. 15, FIG. 16 and FIG 17 are graphs showing a first characterization factor μ, a second charac- terization factor μ and a third characterization factor μ obtained by the above-described processing, i.e., characterization factor values relative to each input brightness. As shown in FIG. 17, the condition of I a' I < 0.001 |a| is satisfied based on the third characterization factor, the characterization factor value constantly becomes substantially 1 with respect to a change in the input brightness, and a at this time becomes a calibration factor.
It is to be noted that one of the first calibration factor acquisition processing shown in
FIG. 10 and the second calibration factor acquisition processing shown in FIG. 14 is executed in the processing to acquire the calibration factor a at the step S2 in FIG. 9. Based on the thus obtained calibration factor a, the corrected image processing portion 2 in the apparatus shown in FIG. 4 calculates the following expression at step S3 in FIG. 9:
I(P)
I'(P) =
1 - al(p)
As a result, corrected image data I' (P) relative to the original image data I (P) can be obtained. This corrected image data I' (P) is stored in the image output portion 3.
By assigning the calibration factor a in the expression (8) in this manner, a corrected image in which an object surface brightness is correctly reflected can be obtained in the subsequent photography using the imaging apparatus. That is, an image which correctly reflects the object surface brightness can be obtained without preparing an object surface having a known brightness distribution at the time of regular photography using the imaging apparatus. It is to be noted that the calibration factor a is calculated as described above at the time of shipment of the imaging apparatus, and it is assigned in the above-described expression of the corrected image processing portion 2. Therefore, in case of the first imaging apparatus having the structure shown in FIG. 1, just assigning the calibration factor a which is measured by another device in advance in the expression of the corrected image processing portion 2 can perform setting required to obtain a corrected image in which an object surface brightness is correctly reflected.
It is to be noted that the present invention is not restricted to the foregoing embodiments, and it can be modified and carried out in many ways without departing from the scope of the invention.
Industrial Applicability According to the present invention, there can be provided an imaging apparatus evaluation method which correctly evaluates output characteristics of an imaging apparatus without preparing an object surface having a known brightness distribution. Furthermore, according to the present invention, there can be provided an image correction method which obtains an image which correctly reflects an object surface brightness as a corrected image without preparing an object surface having a known brightness distribution.
Moreover, according to the present invention, there can be provided an imaging apparatus which can obtain an image which correctly reflects an object surface brightness without preparing an object surface having a known brightness distribution.

Claims

C L A I M S
1. An imaging apparatus evaluation method characterized by comprising: obtaining an output function of original image data by comparing a plurality of sets of image data which are obtained by imaging the same spatial brightness distribution surface with different exposure times.
2. The imaging apparatus evaluation method according to claim 1 characterized in that assuming that k is an exposure time ratio, Pn is an object point brightness which is the nth in ascending order, and f (Pn) and f (kPn) are output brightnesses of first image data and second image data with respect to an object point brightness Pn, respectively, an output function f (P) of the original image data is obtained based on the following expressions:
Figure imgf000028_0001
R"-1 " kffe-i)
Figure imgf000028_0002
3. The imaging apparatus evaluation method according to claim 2, characterized in that f (Pn) is the output brightness of the first image data, Pn is a threshold value, and f (kPn) is a threshold value of the second image data with which an area with the output brightness equal to an area with the output brightness of the first image data larger than the threshold value Pn is obtained.
4. An image correction method comprising: obtaining corrected image data by comparing a plurality of sets of image data obtained by imaging the same spatial brightness distribution surface with different exposure times.
5. An image correction method comprising: determining a as a calibration factor and obtaining corrected image data I' (P) relative to original image data I(P) as the following expression:
Figure imgf000029_0001
6. The image correction method according to claim 5, characterized in that assuming that k is an exposure time ratio, Pn is an object point brightness which is the nth in ascending order, and f(Pn) and f (kPn) are output brightnesses of first image data and second image data relative to an object point brightness Pn, respectively, an output function f (P) of the original image data is obtained based on the following expression:
Figure imgf000029_0002
and the calibration factor a is obtained as an average value of the following expression: Φ) p
7. The image correction method according to claim 6, characterized in that the exposure time ratio k is changed, and k with which a square-sum of a deviation from the average value of the following expression becomes minimum is used:
1 1
Φ) P
8. The image correction method according to claim 5, characterized in that assuming that k is an exposure time ratio, Pn is an object point brightness which is the nth in ascending order, and f (Pn) and f (kPn) are output brightnesses of first image data and second image data with respect to the object point brightness Pn, respectively, an output function f(P) of the original image data is obtained based on the following expressions:
f( }
Figure imgf000030_0001
and a calibration factor a is obtained as an approximate gradient of the following expression:
μ(P) Φ)
9. The image correction method according to claim 8, characterized in that the exposure time ratio k is changed, and k with which a square-sum of a deviation from an approximate straight line of the following expression becomes minimum is used:
μ(p) = W)
10. An imaging apparatus having a processing portion which obtains corrected image data by comparing a plurality of sets of image data obtained by imaging the same spatial brightness distribution surface with different exposure times.
11. An imaging apparatus having a processing portion which determines a is a calibration factor and obtains corrected image data I' (P) as the following expression with respect to original image data I (P) :
Figure imgf000031_0001
12. The imaging apparatus according to claim 11, characterized in that assuming that k is an exposure time ratio, Pn is an object point brightness which is the nth in ascending order, and f(Pn) and f (kPn) are output brightnesses of first image data and second image data relative to the object point brightness Pn, respectively, the imaging apparatus has a processing portion which obtains an output function f (P) of the original image data based on the following expressions:
Figure imgf000031_0002
f(p - kpn-l ~ pn + kRn-l(pn ~ pn-l) f(P \ W " (k-l)Pn-l ln~lj and then obtains a calibration factor a as an average value of the following expression:
Figure imgf000032_0001
13. The imaging apparatus according to claim 11, characterized in that assuming that k is an exposure time ratio, Pn is an object point brightness which is the nth in ascending order, and f (Pn) and f (kPn) are output brightnesses of first image data and second image data with respect to the object point brightness
Pn, the imaging apparatus has a processing portion which obtains an output function f (P) of the original image data based on the following expressions:
)
Figure imgf000032_0002
and then obtains a calibration factor a as an approximate gradient of the following expression:
μ(P) = Φ)
PCT/JP2004/005157 2003-04-10 2004-04-09 Imaging apparatus and calibration method therefor WO2004091224A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP04726744A EP1621024A1 (en) 2003-04-10 2004-04-09 Imaging apparatus and calibration method therefor
US11/244,617 US7573506B2 (en) 2003-04-10 2005-10-06 Imaging device-evaluation method, an image correcting method and an imageing device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2003-106978 2003-04-10
JP2003106978A JP3931152B2 (en) 2003-04-10 2003-04-10 Imaging device evaluation method, image correction method, and imaging device

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US11/244,617 Continuation US7573506B2 (en) 2003-04-10 2005-10-06 Imaging device-evaluation method, an image correcting method and an imageing device

Publications (1)

Publication Number Publication Date
WO2004091224A1 true WO2004091224A1 (en) 2004-10-21

Family

ID=33156927

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2004/005157 WO2004091224A1 (en) 2003-04-10 2004-04-09 Imaging apparatus and calibration method therefor

Country Status (4)

Country Link
US (1) US7573506B2 (en)
EP (1) EP1621024A1 (en)
JP (1) JP3931152B2 (en)
WO (1) WO2004091224A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5828793A (en) * 1996-05-06 1998-10-27 Massachusetts Institute Of Technology Method and apparatus for producing digital images having extended dynamic ranges
US20030007077A1 (en) * 2001-07-09 2003-01-09 Logitech Europe S.A. Method and system for custom closed-loop calibration of a digital camera

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06233333A (en) 1993-01-29 1994-08-19 Imaizumi Shokai:Kk Calibration method/system for video camera
US5682033A (en) * 1995-12-08 1997-10-28 Dba Systems, Inc. Digitizing CCD array system
EP0888686B1 (en) * 1996-12-23 2007-09-05 Koninklijke Philips Electronics N.V. X-ray examination apparatus with X-ray image sensor matrix and correction unit
JP2001346218A (en) 2000-03-27 2001-12-14 Nikon Corp Image processor and electronic camera
JP3854101B2 (en) 2000-06-22 2006-12-06 オリンパス株式会社 Imaging device
US6760471B1 (en) * 2000-06-23 2004-07-06 Teradyne, Inc. Compensation system and related techniques for use in a printed circuit board inspection system
US7038727B2 (en) * 2002-10-30 2006-05-02 The University Of Chicago Method to smooth photometric variations across multi-projector displays

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5828793A (en) * 1996-05-06 1998-10-27 Massachusetts Institute Of Technology Method and apparatus for producing digital images having extended dynamic ranges
US20030007077A1 (en) * 2001-07-09 2003-01-09 Logitech Europe S.A. Method and system for custom closed-loop calibration of a digital camera

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
MITSUNAGA ET AL.: "Radiometric Self Calibration", PROCEEDINGS 1999 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, 1999, LOS ALAMITOS, CA, USA, pages 374 - 380, XP002293122 *

Also Published As

Publication number Publication date
US20060039625A1 (en) 2006-02-23
JP3931152B2 (en) 2007-06-13
EP1621024A1 (en) 2006-02-01
JP2004320079A (en) 2004-11-11
US7573506B2 (en) 2009-08-11

Similar Documents

Publication Publication Date Title
US5181118A (en) Method of correcting image signal
US7626614B1 (en) Transfer function and high dynamic range images
US9516290B2 (en) White balance method in multi-exposure imaging system
CN101523897A (en) Electronic image capture with reduced noise
CN105190424A (en) Imaging apparatus with scene adaptive auto exposure compensation
CN108270975B (en) Exposure time determining method for image sensing
CN108881736B (en) Aperture correction method and device
KR101662407B1 (en) Method for vignetting correction of image and apparatus therefor
US8379142B2 (en) Apparatus and method for balancing color of flashed image using depth estimation
JP2009225345A (en) Program, device, and method for correcting image
TWI618450B (en) Illuminance acquisition device, lighting control system and program
CN109357754B (en) Tunnel hole internal and external brightness instrument calibration method and system based on deep learning
KR101961479B1 (en) Apparatus and method for measuring optical center in camera module
CN110673428B (en) Structured light compensation method, device and equipment
EP1621024A1 (en) Imaging apparatus and calibration method therefor
KR101854355B1 (en) Image correction apparatus selectively using multi-sensor
US20150271383A1 (en) Method for generating hdr images using modified weight functions
JP2005309651A (en) Shading processor and shading processing method for imaging element and imaging device
US7953272B2 (en) Image processor, image processing method and computer readable medium for image processing program
CN112532889B (en) Method and device for determining aperture value of image acquisition equipment
US20090153911A1 (en) Apparatus and Methods for Two Point Tone Reproduction Curve Calibration
CN110858877B (en) Method and device for determining exposure time of projector
KR102495491B1 (en) Frequency control system of color burst signal for camera and control method thereof
JP4756831B2 (en) Image photographing apparatus and image processing method
CN110858898B (en) Camera gain determination method and device

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 11244617

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2004726744

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 2004726744

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 11244617

Country of ref document: US