US8593539B2 - Image processing apparatus and image processing method for performance white balance processing - Google Patents
Image processing apparatus and image processing method for performance white balance processing Download PDFInfo
- Publication number
- US8593539B2 US8593539B2 US13/087,791 US201113087791A US8593539B2 US 8593539 B2 US8593539 B2 US 8593539B2 US 201113087791 A US201113087791 A US 201113087791A US 8593539 B2 US8593539 B2 US 8593539B2
- Authority
- US
- United States
- Prior art keywords
- white
- image
- white balance
- detection area
- aperture value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
- 238000012545 processing Methods 0.000 title claims abstract description 162
- 238000003672 processing method Methods 0.000 title claims description 5
- 238000000034 method Methods 0.000 claims abstract description 40
- 238000001514 detection method Methods 0.000 claims description 78
- 230000008859 change Effects 0.000 claims description 28
- 238000003384 imaging method Methods 0.000 claims description 22
- 230000003287 optical effect Effects 0.000 claims description 20
- 230000005484 gravity Effects 0.000 claims description 17
- 238000012937 correction Methods 0.000 claims description 16
- 230000007246 mechanism Effects 0.000 claims description 5
- 239000013598 vector Substances 0.000 claims description 4
- 238000011161 development Methods 0.000 description 19
- 230000006870 function Effects 0.000 description 18
- 230000010354 integration Effects 0.000 description 18
- 230000008569 process Effects 0.000 description 17
- 230000035945 sensitivity Effects 0.000 description 12
- 238000011156 evaluation Methods 0.000 description 7
- 238000005375 photometry Methods 0.000 description 6
- 230000006854 communication Effects 0.000 description 5
- 230000007423 decrease Effects 0.000 description 5
- 238000002834 transmittance Methods 0.000 description 5
- 230000004888 barrier function Effects 0.000 description 4
- 230000005457 Black-body radiation Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000003595 spectral effect Effects 0.000 description 2
- WHXSMMKQMYFTQS-UHFFFAOYSA-N Lithium Chemical compound [Li] WHXSMMKQMYFTQS-UHFFFAOYSA-N 0.000 description 1
- 229910005580 NiCd Inorganic materials 0.000 description 1
- 229910005813 NiMH Inorganic materials 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 239000011248 coating agent Substances 0.000 description 1
- 238000000576 coating method Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000013144 data compression Methods 0.000 description 1
- 230000006837 decompression Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 229910052744 lithium Inorganic materials 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
- H04N23/84—Camera processing pipelines; Components thereof for processing colour signals
- H04N23/88—Camera processing pipelines; Components thereof for processing colour signals for colour balance, e.g. white-balance circuits or colour temperature control
Definitions
- the present invention relates to an image processing apparatus and an image processing method and, more particularly, to an image processing apparatus for suppressing color crosstalk in an image sensed by an image sensor, and a control method thereof.
- Japanese Patent Laid-Open No. 11-194390 discloses a technique of enabling fine light quantity adjustment by a combination of stop driving using a stepping motor and electronic shutter speed control in a video camera or the like without increasing the number of steps of the stop.
- a method of preparing in advance a correction coefficient for correcting generated color crosstalk and correcting an image signal corrects an image signal by applying, to each color pixel, a gain for correcting a relative sensitivity difference caused by color crosstalk.
- a large gain needs to be applied. This may decrease the S/N ratio or degrade the image quality owing to overcorrection and undercorrection.
- Color crosstalk correction changes the exposure of an image to be captured.
- an image processing apparatus which performs white balance processing for a moving image obtained by capturing, by an image sensor, an object image formed by an imaging optical system
- the apparatus comprising: a storage unit which stores white-detection area information available in correspondence with at least two parameters regarding the imaging optical system; a determination unit which sets a parameter regarding the imaging optical system; a calculation unit which calculates a first white balance coefficient using a frame image obtained using the parameter determined by the determination unit and a first white-detection area obtained from the white-detection area information in the storage unit in correspondence with the determined parameter; an estimating unit which calculates a second white balance coefficient using the frame image obtained using the parameter determined by the determination unit and a second white-detection area different from the first white-detection area, and corrects the second white balance coefficient using information about the first white-detection area to calculate an estimated white balance coefficient; an application unit which applies a white balance coefficient to a frame image contained in the moving image; and
- an image processing method of performing white balance processing for a moving image obtained by sensing, by an image sensor, an object image formed by an imaging optical system comprising: a storage step of storing, in a storage unit, white-detection area information available in correspondence with at least two parameters regarding the imaging optical system; a determination step of setting a parameter regarding the imaging optical system; a calculation step of calculating a first white balance coefficient using a frame image obtained using the parameter determined in the determination step and a first white-detection area determined from the white-detection area information in the storage unit in correspondence with the determined parameter; an estimating step of calculating a second white balance coefficient using the frame image obtained using the parameter determined in the determination step and a second white-detection area different from the first white-detection area, and correcting the second white balance coefficient using information about the first white-detection area to calculate an estimated white balance coefficient; an application step of applying a white balance coefficient to a frame image contained
- the present invention can provide an image processing apparatus capable of effectively correcting color crosstalk in a moving image by a simple method, a control method thereof, and an image-capturing apparatus.
- FIG. 1 is a block diagram exemplifying the functional arrangement of an image-capturing apparatus as an example of an image processing apparatus according to an embodiment of the present invention
- FIG. 2 is a flowchart for explaining the overall operation of the image-capturing apparatus according to the embodiment of the present invention
- FIG. 4 is a flowchart for explaining the moving image capturing operation of the image-capturing apparatus according to the embodiment of the present invention.
- FIG. 5 is a graph exemplifying the relationship between the white-detection area and the aperture value that is set in the embodiment of the present invention.
- FIG. 6 is a flowchart for explaining the estimated white balance coefficient calculation operation of the image-capturing apparatus according to the embodiment of the present invention.
- FIG. 7 is a timing chart of the image-capturing apparatus according to the embodiment of the present invention.
- FIG. 1 is a block diagram exemplifying the functional arrangement of an image-capturing apparatus as an example of an image processing apparatus according to an embodiment of the present invention.
- the image-capturing apparatus includes an imaging lens 10 .
- a shutter 12 has a stop mechanism capable of changing the aperture value.
- the imaging lens 10 and shutter 12 form an imaging optical system.
- An image sensor 14 such as a CCD sensor or CMOS sensor senses an object image formed by the imaging optical system, and converts it into an electrical signal.
- An A/D converter 16 converts an analog image signal read out from the image sensor 14 into a digital image signal.
- An image processing unit 20 performs predetermined pixel interpolation processing and color conversion processing for an image signal from the A/D converter 16 or one from the memory control unit 22 .
- the image processing unit 20 executes predetermined calculation processing using a captured image signal, and performs even TTL AWB (Auto White Balance) processing by applying the obtained calculation result (white balance coefficient) to the image signal.
- white balance processing the image processing unit 20 performs block integration to divide an image formed from captured image signals into a plurality of blocks and obtain the integral value of pixels belonging to each block for each color.
- the image processing unit 20 also executes white search integration to expand the value of each pixel into coordinates in the chromaticity space, and obtain the integral value of pixel values falling within a white-detection area defined in advance in the chromaticity space. Note that the image processing unit 20 stores the setting of the white-detection area in the chromaticity space used in white search integration.
- the memory 30 is also available as a work area for the system control unit 50 .
- the exposure control unit 40 controls the shutter 12 having the stop mechanism. In addition, the exposure control unit 40 provides even a flash light control function in cooperation with a flash 48 .
- the exposure control unit 40 controls the stop mechanism of the shutter 12 by setting an aperture value for the shutter 12 .
- the aperture value is determined based on an AE function implemented by the system control unit 50 , and sent to the exposure control unit 40 .
- the focus control unit 42 controls focusing of the imaging lens 10 .
- a zoom control unit 44 controls zooming of the imaging lens 10 .
- a barrier control unit 46 controls the operation of a lens barrier 102 to protect the imaging lens 10 .
- the flash 48 functions as an auxiliary light source in image capturing.
- the flash 48 has even a light control function and AF auxiliary light projection function.
- the exposure control unit 40 and focus control unit 42 are controlled using the TTL method. More specifically, the image processing unit 20 applies predetermined calculation to a captured image signal. Based on the calculation result, the system control unit 50 controls the exposure control unit 40 and focus control unit 42 .
- the system control unit 50 is, for example, a CPU, and controls the overall image-capturing apparatus 100 by executing a program stored in a memory 52 .
- the memory 52 stores constants, variables, programs, and the like for the operation of the system control unit 50 .
- the memory 52 also stores program charts used in AE control.
- the program chart is a table which defines the relationship between the control values of the stop aperture diameter (or aperture value) and shutter speed with respect to the photometric value (EV value).
- a display unit 54 is formed from a combination of output devices such as an LCD, LED, and loudspeaker.
- the display unit 54 outputs an operating state, message, or the like using a text, image, sound, or the like in accordance with execution of a program by the system control unit 50 .
- One or a plurality of display units 54 are arranged at easily recognizable positions near an operation unit 70 of the image-capturing apparatus 100 .
- Part of the display unit 54 is set in an optical viewfinder 104 .
- a nonvolatile memory 56 is an electrically erasable/programmable memory and is, for example, an EEPROM.
- a mode dial 60 , shutter switches 62 and 64 , and the operation unit 70 configure operation controls for inputting various operation instructions to the system control unit 50 .
- These operation controls are formed from one or a combination of a button, switch, dial, touch panel, pointing based on line-of-sight detection, speech recognition device, and the like.
- the second shutter switch SW 2 64 is turned on by completing the operation of the shutter button (not shown) (full stroke), and designates the start of a series of processes including exposure processing, development processing, and recording processing.
- exposure processing an image signal read out from the image sensor 14 is written in the memory 30 via the A/D converter 16 and memory control unit 22 .
- development processing is done using calculation by the image processing unit 20 or memory control unit 22 under the control of the system control unit 50 .
- recording processing an image signal having undergone development processing is read out from the memory 30 , compressed by the codec unit 32 , and written as an image file on a memory card 200 or 210 .
- the operation unit 70 is a user interface having operation members such as a switch, button, rotary dial switch, and touch panel.
- the operation unit 70 allows selecting an image recording mode, compression ratio, image quality, and the like for a captured image.
- the operation unit 70 includes a moving image button which allows the user in the image shooting mode to designate the start/end of moving image capturing.
- the method of designating the start/end of moving image capturing is not particularly limited, and an arbitrary method can be adopted.
- the user may select the moving image shooting mode with the mode dial 60 , and designate the start/end of moving image capturing by turning on/off the second shutter SW 2 .
- a power control unit 80 includes a battery detection circuit, a DC-DC converter, and a switching circuit for switching a block to be energized.
- the power control unit 80 detects loading/unloading of a battery, the type of battery, and the remaining battery level.
- the power control unit 80 controls the DC-DC converter based on the detection results and an instruction from the system control unit 50 .
- the power control unit 80 applies a necessary voltage to respective units including the recording media 200 and 210 for a necessary period.
- a power supply 86 is a primary battery (for example, an alkaline battery or lithium battery), a secondary battery (for example, an NiCd battery, NiMH battery, or Li battery), an AC adapter, or the like.
- the power supply 86 is attached to the image-capturing apparatus 100 via connectors 82 and 84 .
- the recording media 200 and 210 such as a memory card or hard disk include recording units 202 and 212 each formed from a semiconductor memory, magnetic disk, or the like, and interfaces 204 and 214 and connectors 206 and 216 with the image-capturing apparatus 100 .
- the recording media 200 and 210 are mounted in the image-capturing apparatus 100 via the medium-side connectors 206 and 216 and the connectors 92 and 96 on the side of the image-capturing apparatus 100 .
- the connectors 92 and 96 are connected to interfaces 90 and 94 .
- a recording medium attachment/detachment detection unit 98 detects whether the recording media 200 and 210 are mounted.
- the image-capturing apparatus 100 includes two systems of interfaces and connectors for attaching a recording medium.
- two systems of interfaces and connectors for attaching a recording medium can be arranged.
- interfaces and connectors of different standards may be used for respective systems.
- the lens barrier 102 covers an imaging unit including the imaging lens 10 in the image-capturing apparatus 100 to protect the imaging unit from dirt and damage.
- the optical viewfinder 104 is, for example, a TTL viewfinder, and forms a beam having passed through the imaging lens 10 into an image using a prism or mirror.
- the use of the optical viewfinder 104 allows capturing an image without using an electronic viewfinder function on the image display unit 28 .
- the optical viewfinder 104 displays information about some functions of the display unit 54 , for example, focusing display, camera shake warning display, flash charge display, shutter speed display, aperture value display, and exposure correction display.
- a communication unit 110 performs various communication processes including RS232C, USB, IEEE1394, P1284, SCSI, modem, LAN, and wireless communication processes.
- a connector (antenna for wireless communication) 112 connects the image-capturing apparatus 100 to another device via the communication unit 110 .
- FIGS. 2 and 3 are flowcharts for explaining the overall operation of the image-capturing apparatus 100 .
- the system control unit 50 When the image-capturing apparatus 100 is turned on by, for example, exchanging the battery, the system control unit 50 initializes flags, control variables, and the like (step S 201 ), and initializes the image display of the image display unit 28 to an OFF state (step S 202 ).
- the system control unit 50 determines the set position of the mode dial 60 . If the mode dial 60 is set to power-off (step S 203 ), the system control unit 50 executes predetermined end processing (step S 205 ), and then returns the process to step S 203 .
- this end processing includes the following processes: processing of ending the display on each display unit, processing of closing the lens barrier 102 , processing of recording parameters, set values, and a set mode including flags and control variables in the nonvolatile memory 56 , and processing of stopping unnecessary power supply to the respective units of the image-capturing apparatus 100 including the image display unit 28 under the control of the power control unit 80 .
- step S 203 If the mode dial 60 is set to another mode in step S 203 , the system control unit 50 executes processing corresponding to the selected mode (step S 204 ), and returns the process to step S 203 .
- the system control unit 50 controls the power control unit 80 to check the remaining capacity and operating state of the power supply 86 formed from a battery or the like (step S 206 ). The system control unit 50 then determines whether the power supply 86 has a problem in the operation of the image-capturing apparatus 100 . If the power supply 86 has a problem, the system control unit 50 displays a predetermined warning by an image or sound using the display unit 54 (step S 208 ), and returns the process to step S 203 .
- step S 206 the system control unit 50 determines whether the recording medium 200 or 210 has a problem in the operation of the image-capturing apparatus 100 , especially in the recording/playback operation (step S 207 ). If the recording medium 200 or 210 has a problem, the system control unit 50 displays a predetermined warning by an image or sound using the display unit 54 (step S 208 ), and returns the process to step S 203 .
- the system control unit 50 executes initialization processing in step S 210 to cause the image display unit 28 to function as a viewfinder. More specifically, the system control unit 50 sets building components such as the image sensor 14 , image processing unit 20 , and memory control unit 22 necessary to generate a captured image for through display (through image), in a through image display state in which captured image signals are sequentially displayed. That is, the through image display state corresponds to moving image capturing for display.
- the system control unit 50 Upon completion of image capturing preparations, the system control unit 50 starts capturing a moving image, and starts displaying a through image on the image display unit 28 (step S 211 ).
- the system control unit 50 sequentially displays, on the image display unit 28 via the memory control unit 22 and D/A converter 26 , image signals which have been sequentially written in the image display memory 24 via the image sensor 14 , A/D converter 16 , image processing unit 20 , and memory control unit 22 .
- the image display unit 28 can function as an electronic viewfinder (EVF).
- step S 301 the system control unit 50 determines whether the setting of the mode dial 60 has been changed. If the system control unit 50 detects a change of the setting, it returns the process to step S 203 , and performs processing corresponding to the state of the mode dial 60 . If the setting of the mode dial 60 has not been changed, the system control unit 50 determines the state of the moving image button in step S 302 . If the moving image button is ON, the system control unit 50 advances the process to step S 305 ; if it is OFF, to step S 303 .
- the moving image button is a toggle switch which switches between the ON and OFF states every time it is pressed.
- the system control unit 50 performs predetermined processing to continue the through image display in step S 303 .
- the image processing unit 20 performs a predetermined photometry calculation for a signal obtained from the image sensor 14 , and stores the calculation result in the memory 30 .
- the system control unit 50 Based on the photometry calculation result, the system control unit 50 performs AE processing for through image capturing using the exposure control unit 40 , and determines image capturing conditions such as the aperture value and shutter speed (step S 304 ).
- the system control unit 50 then returns the process to step S 301 .
- step S 302 determines in step S 302 that the moving image button is ON, it performs AF processing in step S 305 . More specifically, the image processing unit 20 executes a predetermined focus calculation for a signal obtained from the image sensor 14 , and stores the calculation result in the memory 30 . Based on the focus calculation result, the system control unit 50 performs AF processing using the focus control unit 42 , and adjusts the focus of the imaging lens 10 to an object.
- step S 306 the system control unit 50 executes integration processing and a predetermined WB calculation using the image processing unit 20 for a signal obtained from the image sensor 14 , obtaining a WB coefficient for moving image capturing.
- the system control unit 50 stores the moving image capturing WB coefficient in the memory 30 , and sets it in the image processing unit 20 .
- step S 307 the system control unit 50 executes moving image capturing processing for one frame.
- the memory 30 saves the image signal of a frame image before signal processing that has been read out from the image sensor 14 via the A/D converter 16 .
- the system control unit 50 executes development processing using the image processing unit 20 for the image signal before signal processing.
- the development processing includes color interpolation processing and white balance processing. Depending on the recording format, encoding processing by the codec unit 32 , generation of a recording file, and the like are also performed.
- the image signal having undergone development processing is saved as image data in the memory 30 .
- the system control unit 50 writes the image data saved in the memory 30 as a moving image file in the recording medium 200 .
- step S 308 the system control unit 50 displays the frame image captured in step S 307 as a through image on the image display unit 28 , completing the moving image capturing start processing. After that, the system control unit 50 shifts to processing in FIG. 4 .
- step S 401 the system control unit 50 checks whether the setting of the mode dial 60 has been changed. If the system control unit 50 detects a change of the setting, it returns the process to step S 203 , and performs processing corresponding to the state of the mode dial 60 . If the setting of the mode dial 60 has not been changed, the system control unit 50 determines the state of the moving image button in step S 402 .
- the system control unit 50 ends the moving image capturing processing, and returns the process to step S 203 .
- step S 403 the system control unit 50 advances the process to step S 403 to continue moving image capturing processing.
- step S 403 the system control unit 50 executes moving image capturing processing for one frame.
- the memory 30 saves the image signal of a frame image before signal processing that has been read out from the image sensor 14 via the A/D converter 16 .
- step S 404 the system control unit 50 applies a predetermined photometry calculation for the image signal in the memory 30 using the image processing unit 20 , and stores the calculation result in the memory 30 .
- step S 405 by referring to the photometry calculation result obtained in step S 404 , the system control unit 50 determines whether to change the aperture value out of current exposure set values including the electronic shutter speed and aperture value. This determination can be made by an arbitrary known method, for example, by using a predetermined program chart. If the system control unit 50 determines to change the aperture value, it obtains a target aperture value and advances the process to step S 406 . If the system control unit 50 determines not to change the aperture value, it advances the process to step S 407 .
- step S 406 the system control unit 50 performs WB correction processing using the image processing unit 20 .
- WB correction processing a change of the white balance upon a change of the aperture value is obtained by calculation.
- the system control unit 50 calculates a WB coefficient corresponding to the target aperture value using the image processing unit 20 based on the current WB coefficient stored in the memory 30 , the current aperture value, the target aperture value, and correction information stored in advance in the nonvolatile memory 56 .
- the corrected WB coefficient is set in the image processing unit 20 and stored in the memory 30 . Note that details of the WB correction processing in step S 406 will be described with reference to FIG. 6 .
- step S 407 the system control unit 50 performs AE processing using the exposure control unit 40 based on the photometry calculation result in step S 404 .
- step S 408 the system control unit 50 performs WB processing using the image processing unit 20 based on a frame image.
- the system control unit 50 controls the image processing unit 20 to execute either integration processing or calculation processing in WB processing. If no integration processing has been executed or the system control unit 50 determines in step S 405 to move the stop, the system control unit 50 performs integration processing for WB processing to the image signal before signal processing that is stored in the memory 30 .
- the image processing unit 20 determines that a block having color evaluation values Cx[i] and Cy[i] falling within a white-detection area corresponding to the target aperture value is a white block.
- the image processing unit 20 calculates integral values SumR, SumG, and SumB of red, green, and blue pixels contained in the white block (white search integration), and stores the result in the memory 30 .
- white-detection area information corresponding to the target aperture value may be stored in advance in the nonvolatile memory 56 or the like for each aperture value.
- each white-detection area may be calculated from white-detection area information corresponding to a given aperture value and a calculation equation prepared in advance.
- the integration processing for a frame image takes some time.
- the system control unit 50 advances the process to step S 409 after the start of the integration processing without waiting for its end.
- the system control unit 50 executes WB calculation processing using the image processing unit 20 .
- the image processing unit 20 sets the calculated moving image capturing white balance coefficients (white balance control values) in it, and also stores them in the memory 30 .
- step S 409 the system control unit 50 executes development processing using the image processing unit 20 for the unprocessed image data which has been captured in step S 403 and stored in the memory 30 using the image processing unit 20 .
- the development processing includes color interpolation processing and white balance processing. Depending on the recording format, encoding processing by the codec unit 32 , generation of a recording file, and the like are also performed.
- the image signal having undergone development processing is saved as image data in the memory 30 .
- the system control unit 50 displays a through image on the image display unit 28 using the image data.
- step S 410 the system control unit 50 writes the image data saved in the memory 30 as a moving image file in the recording medium 200 .
- the system control unit 50 returns the process to step S 401 to continue moving image capturing processing.
- FIG. 5 is a graph exemplifying the relationship between the white-detection area and the aperture value that is set in the embodiment.
- the white-detection area is represented as an area in a chromaticity space where the color temperature axis and the magenta-green axis are perpendicular to each other.
- the X-axis indicates the color temperature (color evaluation value Cx[i])
- the positive direction is a low-color temperature side, that is, red direction
- the negative direction is a high-color temperature side, that is, blue direction.
- the Y-axis is the magenta-green axis (color evaluation value Cy[i]), the positive direction is the magenta direction
- the negative direction is the green direction.
- the white-detection area defines an area where the color evaluation values of a white object exist under various light sources based on the blackbody radiation axis.
- the sensitivity of the color filter for each color changes depending on composite elements including the internal arrangement of the image sensor, the image capturing system such as the color filter, the lens structure, and the optical system such as the coating and filter. For this reason, the position and size of the white-detection area in the chromaticity space change.
- the color crosstalk state also changes depending on the incident angle of the beam.
- the angle of light incident on the image sensor especially the maximum incident angle with respect to the optical axis depends on the aperture size.
- the aperture is larger (aperture value or F-number is smaller), a larger quantity of beam obliquely enters the image sensor.
- the relative sensitivity of each color pixel depends on the aperture value.
- Even the position and size of the white-detection area in the chromaticity space depend on the aperture value.
- the sensitivity to the wavelength of a color having high transmittance for a beam increases in a pixel of another color adjacent to a pixel of the color having high transmittance for a beam. Accordingly, the sensitivity of the pixel of the color having high transmittance for a beam relatively decreases.
- the spectral sensitivity characteristic of the image sensor changes between a case in which color crosstalk occurs and a case in which no color crosstalk occurs, and the color balance of an image also changes between these two cases. As the degree of color crosstalk becomes higher, the color balance changes more greatly.
- the embodiment executes the aforementioned white balance processing by setting a white-detection area to correct a change of the color balance caused by color crosstalk, thereby simply, effectively correcting color crosstalk in a captured image. More specifically, the white-detection area used in white balance processing for a captured image is dynamically set in accordance with the target aperture value obtained in step S 405 .
- FIG. 5 exemplifies a white-detection area for F2.0 and that for F8.0.
- the transmittances and transmission wavelengths of red, blue, and green filters which form a color filter differ between F2.0 and F8.0.
- color crosstalk is large, the sensitivity in the blue direction increases, and that in the green direction decreases.
- the white-detection area extends in a high-color temperature direction and is positioned in the magenta direction (near magenta).
- magenta direction near magenta
- entrance of an oblique beam decreases compared to that at F2.0, and generation of color crosstalk is suppressed.
- the sensitivity in the magenta-green direction greatly changes, and at the same time, the sensitivity in the blue direction slightly changes.
- the spectral sensitivity characteristic changes due to color crosstalk, so the position of the blackbody radiation axis serving as the base of the white-detection area changes depending on the aperture value.
- the position of the white-detection area is therefore set in accordance with that of the blackbody radiation axis corresponding to the aperture value.
- the white balance change amount corresponding to the aperture value can be simply obtained from the motion vector of the gravity center of the white-detection area.
- the change amount can be used to correct the white balance change amount corresponding to the aperture value for a light source having a fixed-value white balance called a preset white balance, such as sunlight or a bulb.
- a white-detection area for F2.0 is used in a full-aperture state, and a white-detection area for F8.0 is used at small apertures of F8.0 or more.
- White-detection areas corresponding to aperture values (for example, F4.0 and F5.6) between the full and small apertures may be prepared in advance, or obtained from the detection areas for F2.0 and F8.0 according to a method such as linear interpolation or linear transformation.
- the image-capturing apparatus 100 is of a lens interchangeable type, the color crosstalk characteristic changes depending on even the lens diameter.
- pieces of white-detection area information corresponding to a lens can also be stored for a plurality of aperture values.
- FIG. 6 is a flowchart for explaining an operation of correcting the white balance coefficient in step S 406 .
- step S 501 the system control unit 50 obtains the currently used white balance coefficient and an aperture value (current aperture value) used when the white balance coefficient was obtained.
- step S 503 the system control unit 50 obtains information which is stored in the nonvolatile memory 56 and defines a white-detection area.
- the nonvolatile memory 56 stores only pieces of information (and corresponding gravity center positions) which define white-detection areas corresponding to F2.0 and F8.0 out of aperture values settable for the shutter 12 .
- the system control unit 50 reads out, from the nonvolatile memory 56 , pieces of information which define white-detection areas corresponding to F2.0 and F8.0, and stores them in the memory 30 .
- step S 504 the system control unit 50 obtains a white-detection area corresponding to the current aperture value, and a white-detection area corresponding to the target aperture value. If the current aperture value is a value falling within the range of F2.0 to F8.0, the system control unit 50 obtains a corresponding white-detection area by interpolation calculation using the pieces of information which define white-detection areas for F2.0 and F8.0. Even if the target aperture value is a value falling within the range of F2.0 to F8.0, the system control unit 50 obtains a corresponding white-detection area in the same way. The system control unit 50 stores, in the memory 30 , pieces of white-detection area information obtained by interpolation calculation. Needless to say, if the current aperture value or target aperture value is F2.0 or F8.0, no interpolation calculation is necessary for this aperture value.
- step S 505 the system control unit 50 calculates the gravity center of the white-detection area corresponding to the current aperture value and that of the white-detection area corresponding to the target aperture value.
- the gravity center is obtained as coordinates in the chromaticity space.
- step S 506 the system control unit 50 calculates the motion vectors of the gravity centers of the white-detection areas corresponding to the current aperture value and target aperture value. Based on the moving amounts of the gravity center in the color temperature direction and magenta-green direction, the system control unit 50 corrects the current white balance coefficient. More specifically, the system control unit 50 corrects the white balance coefficient such that a white pixel detected in the white-detection area corresponding to the target aperture value is corrected to be white. The system control unit 50 sets the corrected white balance coefficient in the image processing unit 20 as an estimated white balance coefficient used until a white balance coefficient based on the color evaluation value is calculated. Also, the system control unit 50 stores the estimated white balance coefficient in the memory 30 .
- the estimated white balance coefficient can be calculated by correcting the current white balance coefficient by calculation based on the change amount (moving amount) between the gravity center position of a white-detection area corresponding to the current aperture value and that of a white-detection area corresponding to the target aperture value.
- the estimated white balance coefficient can be used from development processing for the first image captured at a changed (target) aperture value, unlike a white balance coefficient obtained by white balance coefficient resetting processing based on integration processing described above.
- the estimated white balance coefficient is lower in accuracy than the reset white balance coefficient, but has an effect of correcting the influence of color crosstalk caused by a change of the aperture value.
- Development processing is executed using the estimated white balance coefficient until the white balance coefficient is obtained from an image captured after changing the aperture value. This can suppress the influence of color cast.
- FIG. 7 is a timing chart exemplifying the processing timings of image processing for an image signal and stop control in the image-capturing apparatus 100 according to the embodiment.
- VD represents a vertical sync signal, and is used as the basis of exposure and readout timings.
- the period of the vertical sync signal is sometimes called a VD period.
- Exposure operation A charges (image signal) stored in the image sensor 14 are read out in the next VD period.
- Image signal A A/D-converted at the same time as readout undergoes development processing A parallelly to the readout operation with a short delay from the start of readout.
- Display image data A generated by development processing is displayed in the second next VD period.
- WB integration A is executed.
- WB calculation processing A is executed to set a WB coefficient in the image processing unit 20 .
- the set WB coefficient is applied when the image processing unit 20 executes the next development processing.
- development processing B for image signal B of the next frame has already started when WB calculation A is completed.
- the WB coefficient calculated by WB calculation A cannot be reflected in development processing for image signal B.
- the WB coefficient calculated by WB calculation A is reflected in development processing for image signal C. That is, two VD periods are taken for processing of resetting the white balance by WB integration processing and WB calculation processing.
- aperture value control driving of the stop starts immediately after the start of the vertical sync period, and an aperture value after control is reflected in immediately subsequent exposure operation A.
- a change of the color crosstalk state upon a change of the aperture diameter of the stop affects an image captured immediately after changing the aperture value.
- a target aperture value is obtained to correct the current WB coefficient.
- an estimated WB coefficient corresponding to the target aperture value is calculated and set in the image processing unit 20 .
- This correction processing requires neither image signal integration processing nor calculation processing based on the integral value.
- the estimated WB coefficient can therefore be applied to development processing A for image signal A obtained by image capturing processing A which reflects a change of the aperture value.
- the embodiment implements color crosstalk correction by performing development processing using an estimated WB coefficient obtained by correcting the current WB coefficient in accordance with the target aperture value for an image signal captured until the WB coefficient is obtained again after changing the aperture value. Once the aperture value is changed, even WB coefficient obtaining processing is done for an image captured at the changed aperture value. Immediately after the white balance is obtained, it is used. After that, higher-accuracy color crosstalk correction becomes possible.
- the white balance coefficient is obtained again.
- an estimated white balance coefficient obtained by correcting the current white balance coefficient in accordance with the change of the aperture value is used. Color crosstalk can therefore be corrected even for an image captured before the completion of obtaining a white balance coefficient again after changing the aperture value.
- the embodiment has described a method of suppressing color cast caused by color crosstalk along with aperture control during moving image recording.
- this method is applied not only during moving image recording, but also to EVF moving image capturing executed in a moving/still image capturing standby state. That is, the method is applicable to a case in which white balance control corresponding to aperture control is required.
- the aperture value immediately when the target aperture value is determined, the aperture value is changed to the target one.
- the aperture value may be changed stepwise to the target one when smoothly changing the aperture value, or when, for example, the maximum change amount of the aperture value in one VD period is limited.
- an estimated white balance coefficient can be calculated stepwise at the start of each VD period, and the white balance coefficient re-obtaining operation can start after the start of image capturing at the target aperture value.
- the aperture value is changed from F2.0 to F4.0 before exposure in the first VD period. Then, an estimated white balance coefficient for F4.0 is obtained and applied to the frame image of a moving image captured at F4.0. Before exposure in the next VD period, the aperture value is changed from F4.0 to F8.0, and an estimated white balance coefficient for F8.0 is obtained and applied to the frame image of a moving image captured at F8.0. Also, white balance coefficient re-obtaining processing starts for the frame image of a moving image captured at F8.0.
- the coefficient is to be switched as smoothly as possible even at a timing when the white balance coefficient can be obtained again after estimation as in the embodiment.
- the white balance coefficient may be changed gradually to the re-obtained one using a coefficient obtained by weighting the estimated and re-obtained white balance coefficients.
- the above-described embodiment has explained a case in which the image processing apparatus according to the present invention is applied to an image-capturing apparatus, but the image capturing and recording functions are not indispensable for the present invention.
- the present invention can be practiced in an arbitrary device having a function of processing an image captured by an image sensor having a color filter.
- the present invention is applicable to a cell phone, a PDA (Personal Digital Assistant), various information devices having a camera function, and an information processing apparatus capable of loading and processing captured image data.
- aspects of the present invention can also be realized by a computer of a system or apparatus (or devices such as a CPU or MPU) that reads out and executes a program recorded on a memory device to perform the functions of the above-described embodiment(s), and by a method, the steps of which are performed by a computer of a system or apparatus by, for example, reading out and executing a program recorded on a memory device to perform the functions of the above-described embodiment(s).
- the program is provided to the computer for example via a network or from a recording medium of various types serving as the memory device (for example, computer-readable medium).
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Color Television Image Signal Generators (AREA)
- Processing Of Color Television Signals (AREA)
- Studio Devices (AREA)
Abstract
Description
Cx[i]=(R[i]−B[i])/Y[i]×1024
Cy[i]=(R[i]+B[i]−2G[i])/Y[i]×1024
where Y[i]=R[i]+2G[i]+B[i], i=1 . . . m (m is a block count)
WBCo — R=SumY×1024/SumR
WBCo — G=SumY×1024/SumG
WBCo — B=SumY×1024/SumB
where SumY=(SumR+2×SumG+SumB)/4
Claims (3)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010108679A JP5576706B2 (en) | 2010-05-10 | 2010-05-10 | Image processing apparatus, control method therefor, and imaging apparatus |
JP2010-108679 | 2010-05-10 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20110273584A1 US20110273584A1 (en) | 2011-11-10 |
US8593539B2 true US8593539B2 (en) | 2013-11-26 |
Family
ID=44901701
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/087,791 Expired - Fee Related US8593539B2 (en) | 2010-05-10 | 2011-04-15 | Image processing apparatus and image processing method for performance white balance processing |
Country Status (3)
Country | Link |
---|---|
US (1) | US8593539B2 (en) |
JP (1) | JP5576706B2 (en) |
CN (1) | CN102244792A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2876774A1 (en) | 2013-11-25 | 2015-05-27 | Hand Held Products, Inc. | Indicia-reading system |
EP2916259A1 (en) | 2014-03-07 | 2015-09-09 | Hand Held Products, Inc. | Indicia reader for size-limited applications |
EP2927839A1 (en) | 2014-04-01 | 2015-10-07 | Hand Held Products, Inc. | Hand-mounted indicia-reading device with finger motion triggering |
EP2940505A1 (en) | 2014-04-29 | 2015-11-04 | Hand Held Products, Inc. | Autofocus lens system for indicia readers |
US11317063B2 (en) | 2020-01-21 | 2022-04-26 | Samsung Electronics Co., Ltd. | Calibration module of image sensor, image sensor and method of calibrating crosstalk in image sensor |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8199965B1 (en) | 2007-08-17 | 2012-06-12 | Mcafee, Inc. | System, method, and computer program product for preventing image-related data loss |
JP2017505045A (en) * | 2014-01-15 | 2017-02-09 | アナカトゥム デザイン アーベー | Cognitive signal converter |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10271519A (en) | 1997-03-26 | 1998-10-09 | Sony Corp | Solid-state image pickup device |
JPH11194390A (en) | 1997-12-26 | 1999-07-21 | Canon Electron Inc | Image input device provided with light quantity adjusting device |
EP1351524A2 (en) | 2002-04-04 | 2003-10-08 | Eastman Kodak Company | Method for automatic white balance of digital images |
US20080100702A1 (en) * | 2006-10-30 | 2008-05-01 | Pentax Corporation | Electronic endoscope |
JP2008263492A (en) | 2007-04-13 | 2008-10-30 | Elmo Co Ltd | Exposure control method and imaging apparatus |
US20090219408A1 (en) * | 2008-02-28 | 2009-09-03 | Sanyo Electric Co., Ltd. | Electronic camera |
US20090225186A1 (en) * | 2008-03-07 | 2009-09-10 | Sanyo Electric Co., Ltd. | Electronic camera |
US20100026835A1 (en) * | 2006-04-25 | 2010-02-04 | Norio Negishi | Image sensing apparatus and method for controlling the same |
US20100231746A1 (en) * | 2009-03-12 | 2010-09-16 | Kenichiroh Nomura | Imaging device and imaging method |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3554069B2 (en) * | 1995-04-07 | 2004-08-11 | キヤノン株式会社 | Imaging device |
JP4666964B2 (en) * | 2004-07-09 | 2011-04-06 | キヤノン株式会社 | Imaging apparatus and imaging method |
JP4925271B2 (en) * | 2006-08-21 | 2012-04-25 | 株式会社メガチップス | Continuous imaging method for single-lens reflex digital camera |
-
2010
- 2010-05-10 JP JP2010108679A patent/JP5576706B2/en not_active Expired - Fee Related
-
2011
- 2011-04-15 US US13/087,791 patent/US8593539B2/en not_active Expired - Fee Related
- 2011-05-10 CN CN2011101202558A patent/CN102244792A/en active Pending
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10271519A (en) | 1997-03-26 | 1998-10-09 | Sony Corp | Solid-state image pickup device |
JPH11194390A (en) | 1997-12-26 | 1999-07-21 | Canon Electron Inc | Image input device provided with light quantity adjusting device |
EP1351524A2 (en) | 2002-04-04 | 2003-10-08 | Eastman Kodak Company | Method for automatic white balance of digital images |
US20030189650A1 (en) * | 2002-04-04 | 2003-10-09 | Eastman Kodak Company | Method for automatic white balance of digital images |
US20100026835A1 (en) * | 2006-04-25 | 2010-02-04 | Norio Negishi | Image sensing apparatus and method for controlling the same |
US20080100702A1 (en) * | 2006-10-30 | 2008-05-01 | Pentax Corporation | Electronic endoscope |
JP2008263492A (en) | 2007-04-13 | 2008-10-30 | Elmo Co Ltd | Exposure control method and imaging apparatus |
US20090219408A1 (en) * | 2008-02-28 | 2009-09-03 | Sanyo Electric Co., Ltd. | Electronic camera |
US20090225186A1 (en) * | 2008-03-07 | 2009-09-10 | Sanyo Electric Co., Ltd. | Electronic camera |
US20100231746A1 (en) * | 2009-03-12 | 2010-09-16 | Kenichiroh Nomura | Imaging device and imaging method |
Non-Patent Citations (1)
Title |
---|
The above reference was cited in a Jul. 9, 2013 Chinese Office Action, a copy of which is enclosed with an English Translation, that issued in Chinese Patent Application No. 201110120255.8. |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2876774A1 (en) | 2013-11-25 | 2015-05-27 | Hand Held Products, Inc. | Indicia-reading system |
EP2916259A1 (en) | 2014-03-07 | 2015-09-09 | Hand Held Products, Inc. | Indicia reader for size-limited applications |
US10789435B2 (en) | 2014-03-07 | 2020-09-29 | Hand Held Products, Inc. | Indicia reader for size-limited applications |
EP3836002A1 (en) | 2014-03-07 | 2021-06-16 | Hand Held Products, Inc. | Indicia reader for size-limited applications |
US11531825B2 (en) | 2014-03-07 | 2022-12-20 | Hand Held Products, Inc. | Indicia reader for size-limited applications |
EP4280099A2 (en) | 2014-03-07 | 2023-11-22 | Hand Held Products, Inc. | Indicia reader for size-limited applications |
EP2927839A1 (en) | 2014-04-01 | 2015-10-07 | Hand Held Products, Inc. | Hand-mounted indicia-reading device with finger motion triggering |
EP2940505A1 (en) | 2014-04-29 | 2015-11-04 | Hand Held Products, Inc. | Autofocus lens system for indicia readers |
US11317063B2 (en) | 2020-01-21 | 2022-04-26 | Samsung Electronics Co., Ltd. | Calibration module of image sensor, image sensor and method of calibrating crosstalk in image sensor |
Also Published As
Publication number | Publication date |
---|---|
CN102244792A (en) | 2011-11-16 |
JP2011239157A (en) | 2011-11-24 |
US20110273584A1 (en) | 2011-11-10 |
JP5576706B2 (en) | 2014-08-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8625021B2 (en) | Image capture with region-based adjustment of imaging properties | |
US8593539B2 (en) | Image processing apparatus and image processing method for performance white balance processing | |
US8300142B2 (en) | Imaging apparatus having adjustable power supply based on imager sensitivity | |
US8743239B2 (en) | Image processing apparatus, control method thereof, and image-capturing apparatus | |
JP7195120B2 (en) | Imaging device and its control method | |
US10839497B2 (en) | Image processing apparatus and method | |
JP2007174124A (en) | Imaging apparatus and correction method | |
US8717457B2 (en) | Adaptive spectral imaging for video capture | |
JP7229728B2 (en) | Imaging device, its control method, and program | |
US9013597B2 (en) | Image pickup apparatus that carries out noise compensation, control method therefor, and storage medium | |
JP5948997B2 (en) | Imaging apparatus and imaging method | |
JP2011049641A (en) | Photographing device, control method thereof, and program | |
JP2002135651A (en) | Image processor and method for processing image | |
JP2012090041A (en) | Image processing device, method, and program | |
JP5103678B2 (en) | Continuous imaging method for single-lens reflex digital camera | |
US8373766B2 (en) | Image shooting device and image shooting method | |
JP2006217249A (en) | Electronic camera, electronic camera system and program | |
US10674092B2 (en) | Image processing apparatus and method, and image capturing apparatus | |
US20100259606A1 (en) | Imaging device and method for controlling imaging device | |
JP5482428B2 (en) | Imaging apparatus, camera shake correction method, and program | |
JP5482427B2 (en) | Imaging apparatus, camera shake correction method, and program | |
US10298865B2 (en) | Image processing apparatus, image capturing apparatus, method for controlling image processing apparatus, and non-transitory computer-readable storage medium for correcting image corresponding to change in state of an optical member | |
JP4464143B2 (en) | Image processing method and apparatus | |
JP2014053777A (en) | Imaging apparatus | |
JP5791454B2 (en) | Imaging apparatus and control method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CANON KABUSHIKI KAISHA, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:OGAWA, SHIGEO;REEL/FRAME:026811/0345 Effective date: 20110412 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
REMI | Maintenance fee reminder mailed | ||
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.) |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20171126 |