WO2011099299A1 - 映像抽出装置、撮影装置、プログラム及び記録媒体 - Google Patents
映像抽出装置、撮影装置、プログラム及び記録媒体 Download PDFInfo
- Publication number
- WO2011099299A1 WO2011099299A1 PCT/JP2011/000771 JP2011000771W WO2011099299A1 WO 2011099299 A1 WO2011099299 A1 WO 2011099299A1 JP 2011000771 W JP2011000771 W JP 2011000771W WO 2011099299 A1 WO2011099299 A1 WO 2011099299A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- attribute information
- unit
- evaluation
- user
- Prior art date
Links
- 238000000605 extraction Methods 0.000 title claims description 60
- 239000000284 extract Substances 0.000 claims abstract description 50
- 238000011156 evaluation Methods 0.000 claims description 215
- 238000003860 storage Methods 0.000 claims description 51
- 238000003384 imaging method Methods 0.000 claims description 13
- 230000004044 response Effects 0.000 claims description 3
- 238000001514 detection method Methods 0.000 description 53
- 238000012545 processing Methods 0.000 description 26
- 230000005236 sound signal Effects 0.000 description 25
- 238000000034 method Methods 0.000 description 23
- 230000006835 compression Effects 0.000 description 19
- 238000007906 compression Methods 0.000 description 19
- 230000036544 posture Effects 0.000 description 18
- 238000010586 diagram Methods 0.000 description 15
- 238000006243 chemical reaction Methods 0.000 description 12
- 238000009826 distribution Methods 0.000 description 10
- 230000006837 decompression Effects 0.000 description 8
- 239000003086 colorant Substances 0.000 description 5
- 230000001133 acceleration Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 238000004091 panning Methods 0.000 description 3
- 125000002066 L-histidyl group Chemical group [H]N1C([H])=NC(C([H])([H])[C@](C(=O)[*])([H])N([H])[H])=C1[H] 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 235000019640 taste Nutrition 0.000 description 2
- 241001465754 Metazoa Species 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 230000001747 exhibiting effect Effects 0.000 description 1
- 230000002349 favourable effect Effects 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/77—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
- H04N5/772—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera the recording apparatus and the television camera being placed in the same enclosure
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
- G11B27/034—Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/432—Content retrieval operation from a local storage medium, e.g. hard-disk
- H04N21/4325—Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
- H04N21/4394—Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/8549—Creating video summaries, e.g. movie trailer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/78—Television signal recording using magnetic recording
- H04N5/782—Television signal recording using magnetic recording on tape
- H04N5/783—Adaptations for reproducing at a rate different from the recording rate
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
- H04N5/93—Regeneration of the television signal or of selected parts thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/8042—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
Definitions
- the technology disclosed herein is a video extraction device that extracts a portion to be reproduced from a shadow image, a photographing device including the same, a program that extracts a portion to be reproduced from a shadow image, and a program for recording the program
- the present invention relates to a recording medium.
- the photographing apparatus evaluates a scene based on video metadata and reproduces a digest that narrows down the number of video scenes and clips based on the evaluation result. .
- the content of the preferred digest varies greatly depending on the taste of the viewer.
- the metadata and the evaluation based on the metadata according to Patent Document 1 are unique and do not correspond to user preferences.
- the video extraction device disclosed herein stores a storage unit that stores attribute information about a video to be extracted from the video, and the attribute stored in the storage unit in order to extract a portion to be digest reproduced from the video.
- a control unit that extracts information from the video, and at least a part of the attribute information stored in the storage unit is configured to be input by a user.
- another video extraction device extracts a portion to be digest-reproduced from a storage unit that stores attribute data about the video to be extracted from the video and data corresponding to the evaluation with respect to the attribute information.
- a control unit that extracts the attribute information stored in the storage unit from the video and evaluates a portion having the attribute information based on the correspondence data, and the attribute information stored in the storage unit and It is assumed that at least a part of the evaluation is configured to be input by the user.
- another video extraction device includes a storage unit that stores attribute information about the video to be extracted from the video, an input unit that inputs the attribute information to the storage unit according to a user input operation, And a control unit that extracts a portion for digest reproduction from the inside based on the attribute information stored in the storage unit.
- another video extraction device includes a storage unit that stores attribute data about video to be extracted from the video and data corresponding to the evaluation with respect to the attribute information, and the attribute information and evaluation according to a user input operation.
- An input unit that inputs at least one of the above to the storage unit, and a control unit that extracts a portion for digest playback from the video based on the attribute information and evaluation stored in the storage unit .
- the imaging device disclosed herein includes an imaging system that acquires a video and the video extraction device.
- the program disclosed herein accepts input from at least one user of the attribute information and the evaluation with respect to the attribute information about the video to be extracted from the video and the corresponding data with the evaluation of the attribute information; Extracting the attribute information included in the correspondence data from the video, and extracting a portion for digest playback from the video based on the correspondence data and the extracted attribute information. This is a program to be executed.
- the recording medium disclosed herein is a computer-readable recording medium that records the program.
- the video extraction device since the user can arbitrarily input the attribute information, the attribute information reflecting the user's preference can be extracted. By referring to this attribute information, digest playback corresponding to the user's preference can be realized.
- the user can arbitrarily input at least one of the attribute information and the evaluation, it is possible to evaluate the video reflecting the user's preference. By referring to this evaluation, digest reproduction corresponding to the user's preference can be realized.
- the photographing apparatus it is possible to realize digest reproduction corresponding to the user's preference.
- the recording medium it is possible to realize digest reproduction corresponding to the user's preference.
- FIG. 1 is a perspective view of the video camera according to the present embodiment.
- FIG. 2 is a block diagram showing a schematic configuration of the video camera.
- FIG. 3 is a diagram illustrating a configuration of an image captured by the video camera.
- FIG. 4 is an image diagram showing an example in which a clip is divided into a plurality of scenes.
- FIG. 5 is a diagram showing a table of correspondence data between various attribute information and evaluation for each attribute information.
- FIG. 6 is a diagram showing a result of extracting attribute information from a certain video and assigning an evaluation value.
- FIG. 1 is a perspective view of the video camera according to the present embodiment.
- FIG. 2 is a block diagram showing a schematic configuration of the video camera.
- FIG. 3 is a diagram illustrating a configuration of an image captured by the video camera.
- FIG. 4 is an image diagram showing an example in which a clip is divided into a plurality of scenes.
- FIG. 5 is a diagram showing a table of correspondence data between various attribute
- FIG. 7 is a diagram illustrating an example of an image in which a person's face is photographed, where (A) is an image in which the face of the person A is photographed, (B) is an image in which the face of the person B is photographed, (C) shows an image in which the face of the person A and the face of the person B are photographed, and (D) shows an image in which the face of the person A is photographed with many people.
- FIG. 8 is a diagram showing reproduction information.
- FIG. 9 is a flowchart showing a flow of processing from shooting, scene evaluation, generation of reproduction information, and recording.
- FIG. 10 is a flowchart showing the flow of processing during digest playback.
- FIG. 11 is a table of correspondence data between various types of attribute information and evaluations for each attribute information, showing a relationship between attribute information and evaluation values.
- FIG. 12 is a diagram showing a result of the scene evaluation unit extracting attribute information from a certain video based on the table of FIG. 11 and assigning an evaluation value.
- FIG. 13 shows reproduction information.
- FIG. 14 is a diagram illustrating an example of a selection screen that allows the user to select a mode corresponding to various tables.
- FIG. 15 is a block diagram illustrating a schematic configuration of the video camera according to the second embodiment.
- FIG. 16 is a flowchart showing the flow of processing from shooting to attribute information extraction to recording.
- FIG. 17 is a flowchart showing the flow of processing during digest playback.
- FIG. 18 is a block diagram showing a schematic configuration of a video camera according to another embodiment.
- FIG. 1 is a perspective view of a video camera 100A as an example of a photographing apparatus
- FIG. 2 is a block diagram showing a schematic configuration of the video camera 100A.
- the video camera 100A captures a video, records it, and plays it back. Further, the video camera 100A extracts a portion for digest reproduction from the video, and further performs digest reproduction.
- the video camera 100A includes a lens group 200, an image sensor 201, a video AD converter (Analog-to-Digital Converter) 202, a signal processor 203, a video signal compressor 204, a lens control module 205, an attitude detector 206, and an external input.
- This video camera 100A constitutes a video extraction device and a photographing device.
- the lens group 200 is composed of a plurality of lenses and forms an image of incident light from the subject on the image sensor 201.
- the lens group 200 is configured to be able to adjust the distance between a plurality of lenses, and changes the focal length and zoom magnification (image magnification) by changing the distance between the lenses.
- the adjustment of the focal length and the zoom magnification may be performed manually by the user or automatically performed by the control unit 300 or the like which will be described later.
- the image sensor 201 converts incident light into an electrical signal.
- the image sensor 201 is composed of a CCD or a CMOS.
- the image sensor 201 outputs an electric signal in an arbitrary range on the imaging surface.
- the image sensor 201 also includes chromaticity space information of the three primary color points, white coordinates, and gain information of at least two of the three primary colors, color temperature information, ⁇ uv (delta uv), and 3 Information such as gamma information of primary colors or luminance signals can also be output. Therefore, the output from the image sensor 201 is input not only to the video AD conversion unit 202 but also to an attribute information extraction unit 305 described later of the control unit 300.
- the video AD converter 202 converts an analog electric signal output from the image sensor 201 into a digital signal.
- the signal processor 203 converts the digital signal output from the video AD converter 202 into a predetermined video signal format such as NTSC (National Television System Committee) or PAL (Phase Alternate Line).
- NTSC National Television System Committee
- PAL Phase Alternate Line
- the signal processing unit 203 converts the digital signal from the video AD conversion unit 202 into a digital video signal (video data) that conforms to the number of horizontal lines, the number of scanning lines, and the frame rate specified by NTSC.
- An output from the signal processing unit 203 is input to the control unit 300.
- the signal processing unit 203 is configured by, for example, an IC for video signal conversion.
- the video signal format includes a so-called full high-definition method in which the number of effective pixels in one video frame is 1920 in the horizontal direction and 1080 in the vertical direction, and the number of effective pixels in one video frame is 1280 in the horizontal direction. There is a method that is 720.
- the video signal compression unit 204 performs predetermined coding conversion on the digital video signal output from the signal processing unit 203 to compress the data amount.
- Specific examples of encoding conversion include encoding schemes such as MPEG (Moving Picture Experts Group) 2, MPEG4, and H264.
- An output from the video signal compression unit 204 is input to the control unit 300.
- the video signal compression unit 204 is composed of, for example, an IC for signal compression / decompression.
- the lens control module 205 detects the state of the lens group 200 and operates the lens group 200.
- the lens control module 205 has a lens control motor and a lens position sensor.
- the lens position sensor detects distances or positional relationships between a plurality of lenses constituting the lens group 200.
- the lens position sensor outputs the detection signal to the control unit 300.
- the lens control module 205 has two types of lens control motors. One lens control motor moves the lens group 200 in the optical axis direction based on a control signal from the control unit 300. Thereby, the distance between the plurality of lenses of the lens group 200 is changed, and the focal length and zoom magnification of the lens group 200 are adjusted.
- the other lens control motor moves at least one lens (lens for image blur correction) in the lens group 200 within a plane orthogonal to the optical axis based on a control signal from the control unit 300. Thereby, image blur is corrected.
- the posture detection unit 206 detects the posture of the video camera 100A main body.
- the posture detection unit 206 includes an acceleration sensor, an angular velocity sensor, and an elevation angle / decline angle sensor. With these sensors, it is possible to recognize the posture of the video camera 100A during shooting. Note that it is preferable that the acceleration sensor and the angular velocity sensor can detect the postures in three orthogonal directions (vertical direction, front-rear direction, and left-right direction) of the video camera 100A in order to detect the posture of the video camera 100A in detail.
- a signal from the posture detection unit 206 is input to the control unit 300.
- the posture detection unit 206 is not limited to the above-described configuration, and may include only one of the above-described sensors, or may include a sensor other than the above-described sensor.
- the external input unit 207 is a means for inputting information from the outside to the video camera 100A.
- a signal from the external input unit 207 is input to the control unit 300.
- the signal from the external input unit 207 is input only to the attribute information extraction unit 305 of the control unit 300, but is also input to the part corresponding to the input operation of the lens control unit 310 and the like.
- various information from the outside is input to the video camera 100A via the external input unit 207.
- the external input unit 207 is an input button that is one of input interfaces for receiving input of information from the user, a reception unit that receives shooting index information input from the outside via communication, and the video camera 100A.
- the tripod sensor includes a tripod sensor that detects whether or not it is installed on a tripod. For example, when a user operates an input button, various requests from the user are recorded such as start and end of shooting, insertion of markings into the video being shot, and input and setting of attribute information and its evaluation described later. This can be transmitted to the camera 100A. That is, the external input unit 207 constitutes an input unit that inputs at least one of attribute information and evaluation described later to the storage unit 330 in accordance with a user input operation.
- the shooting index information is, for example, an identification number used for identifying each shooting such as a number for identifying a shooting scene at the time of shooting a movie or a number indicating the number of shootings.
- the tripod sensor is composed of a switch provided at a portion where the tripod of the video camera 100A is fixed. With a tripod sensor, it can be determined whether or not shooting is performed using a tripod.
- the microphone 208 converts sound around the video camera 100A into an electric signal and outputs it as an audio signal.
- the audio AD conversion unit 209 converts an analog audio signal output from the microphone 208 into a digital audio signal (audio data).
- the audio signal compression unit 210 converts the digital audio signal output from the audio AD conversion unit 209 using a predetermined encoding algorithm. For encoding, there are methods such as MP3 (MPEG Audio Layer-3) and AAC (Advanced Audio Coding).
- the audio signal compression unit 210 is constituted by a compression IC, for example.
- the video signal decompression unit 211 combines the video signal output from the control unit 300.
- the output from the video signal expansion unit 211 is input to the video display unit 212.
- the video signal expansion unit 211 is constituted by, for example, an IC for video signal expansion.
- the video display unit 212 displays video recorded in the video camera 100A and video captured in real time by the video camera 100A.
- the video camera 100A displays various types of information such as information relating to shooting and device information.
- the video display unit 212 is configured by, for example, a touch panel type liquid crystal display. As described above, the touch panel video display unit 212 also functions as the external input unit 207.
- the audio signal expansion unit 213 combines the audio signal output from the control unit 300.
- the output from the audio signal expansion unit 213 is input to the audio output unit 214.
- the audio signal expansion unit 213 is configured by an audio signal expansion IC, for example.
- the audio output unit 214 outputs audio accompanying the video. In addition, the audio output unit 214 outputs a warning sound to notify the user from the video camera 100A.
- the audio output unit 214 is constituted by a speaker, for example.
- the output I / F 215 is an interface for outputting a video signal from the video camera 100A to the outside.
- the output I / F 215 is a cable interface when the video camera 100A is connected to an external device with a cable, a memory card interface when a video signal is recorded on a memory card, or the like.
- the control unit 300 controls the entire video camera 100A.
- the control unit 300 includes an image sensor 201, a signal processing unit 203, a video signal compression unit 204, a lens control module 205, an attitude detection unit 206, an external input unit 207, an audio AD conversion unit 209, an audio signal compression unit 210, and a video signal decompression.
- the unit 211, the audio signal expansion unit 213, the output I / F 215, the storage unit 330, and the like are configured to be able to exchange signals.
- the control unit 300 is composed of a CPU.
- the control unit 300 executes various controls of the video camera 100A by reading and executing a program stored in the storage unit 330.
- control executed by the control unit 300 include, for example, control of the focal length and zoom of the lens group 200, processing of input signals from the posture detection unit 206 and the external input unit 207, signal processing unit 203, and video signal.
- This is operation control of ICs such as the compression unit 204, the audio signal compression unit 210, the video signal expansion unit 211, and the audio signal expansion unit 213.
- the signal is appropriately AD-converted or DA-converted between the control unit 300 and the lens control module 205 or the like.
- the control unit 300 can also be configured by an integrated circuit such as an IC.
- the clock 320 outputs a clock signal serving as a reference for processing operation to the control unit 300 or the like operating in the video camera 100A.
- the clock 320 may be a single clock or a plurality of clocks depending on the integrated circuit to be used and data to be handled. Further, an arbitrary multiple of the clock signal of one oscillator may be used.
- the storage unit 330 includes a ROM (Read Only Memory), a RAM (Random Access Memory), and an HDD (Hard Disk Drive).
- the ROM stores a program processed by the control unit 300 and various data for operating the program.
- the RAM is used as a memory area used when executing a program processed by the control unit 300.
- the RAM can also be used as a memory area of the IC.
- the HDD stores various data such as video data and still image data encoded and converted by the video signal compression unit 204. Further, the HDD stores a program executed by the control unit 300. Note that this program is not limited to the HDD, and may be stored in a semiconductor memory, or may be stored in a portable recording medium such as a CD-ROM or DVD.
- control unit 300 will be described in detail. Here, a configuration necessary for performing digest reproduction in the control unit 300 will be described. However, the control unit 300 also has a configuration for exhibiting the same function as that of a general video camera, for example, a configuration for recording or playing back a captured video.
- the control unit 300 includes a lens control unit 301 that controls the lens control module 205, an imaging control unit 302 that controls the imaging element 201, a video analysis unit 303 that analyzes the output from the signal processing unit 203, and an audio AD conversion unit.
- An audio analysis unit 304 that analyzes the output from 209, an attribute information extraction unit 305 that extracts attribute information in the video, a scene evaluation unit 306 that evaluates a scene, and playback information that generates playback information for digest playback It has a generation unit 307, a multiplexing unit 308 that multiplexes video data and audio data, and a digest reproduction unit 309 that performs digest reproduction.
- the control unit 300 implements various processes to be described later by reading and executing a program stored in the storage unit 330.
- the lens control unit 301 receives the detection signal of the lens position sensor of the lens control module 205 and the detection signal of various sensors of the posture detection unit 206.
- the lens control unit 301 outputs a control signal for properly arranging the lens group 200 to the lens control motor based on these detection signals and information from other components such as the image sensor 201.
- the lens control unit 301 performs zoom control, focus control, image blur correction control, and the like.
- the lens control unit 301 outputs a control signal for the lens group 200 to the attribute information extraction unit 305. Note that detection signals from various sensors of the posture detection unit 206 are also output to the attribute information extraction unit 305.
- the imaging control unit 302 controls the operation of the imaging element 201. For the image sensor 201, the amount of exposure at the time of shooting, the shooting speed, sensitivity, and the like are controlled.
- the control signal output from the imaging control unit 302 is output not only to the image sensor 201 but also to the attribute information extraction unit 305.
- the video analysis unit 303 extracts video features based on the video data from the signal processing unit 203.
- the video analysis unit 303 detects video color information (for example, detection of a color distribution included in the video) and white balance information.
- the color distribution can be detected by confirming color information included in the data forming the digital video signal.
- the video analysis unit 303 detects a face from the video when the video includes a human face. Face detection can be realized by using pattern matching or the like.
- the voice analysis unit 304 analyzes the voice data from the voice AD conversion unit 209 and extracts characteristic sounds.
- the characteristic sounds here include, for example, a photographer's voice, pronunciation of a specific word, cheers, gunshots, and the like. These sounds can be extracted by a method of registering in advance a unique frequency of these sounds (speech) and discriminating them by comparison with them. In addition to this, for example, it may be determined that the sound is characteristic when the sound input level is equal to or higher than a predetermined level.
- the attribute information extraction unit 305 extracts attribute information about the video.
- the attribute information is information representing video attributes, and is information relating to shooting (hereinafter also referred to as “shooting information”), external input information, and other information.
- the attribute information extraction unit 305 is input with output signals from the image sensor 201, posture detection unit 206, external input unit 207, lens control unit 301, imaging control unit 302, video analysis unit 303, and audio analysis unit 304. .
- the attribute information extraction unit 305 extracts attribute information based on these output signals.
- the attribute information relating to the video includes attribute information relating to the state of the imaging device at the time of shooting the video and shooting such as camera work, attribute information relating to the video when the video is generated by CG, and attribute information relating to the subject and background included in the video itself. , Attribute information related to audio attached to the video, and attribute information related to video editing contents when editing video by an editing machine or the like.
- examples of attribute information related to the photographing apparatus at the time of photographing include focal length, zoom magnification, exposure, photographing speed, sensitivity, color space information of three primary colors, white balance, gain information of at least two of the three primary colors, color Temperature information, ⁇ uv (delta uv), gamma information of three primary colors or luminance signals, color distribution, face recognition information, camera posture (acceleration, angular velocity, elevation angle, depression angle, etc.), shooting time (shooting start time, end time), shooting Examples include index information, user input, frame rate, sampling frequency, and the like.
- the attribute information extraction unit 305 extracts a focal length and a zoom magnification as attribute information based on a control signal from the lens control unit 301.
- the attribute information extraction unit 305 detects a camera posture (acceleration, angular velocity, elevation angle, depression angle, etc.) based on the detection signal of the posture detection unit 206, and a video camera at the time of shooting pan, tilt, etc. from the camera posture 100A camera work is extracted as attribute information. Furthermore, based on these camera works, a fixed shooting part after camera work (a part shot with the video camera 100 stationary) can be extracted as attribute information. As described above, the attribute information extraction unit 305 may extract the attribute information from the input signal itself, or may extract the attribute information by combining or analyzing the input signal.
- the scene evaluation unit 306 evaluates the portion of the video including the attribute information and assigns the evaluation (value) to the portion. Details of this evaluation will be described later.
- the reproduction information generation unit 307 selects a part (scene) to be reproduced and specifies information (hereinafter referred to as “reproduction information”) for specifying a part to be digest reproduced. Generate. Details of the reproduction information will be described later.
- the multiplexing unit 308 multiplexes and outputs the encoded video data from the video signal compression unit 204, the encoded audio data from the audio signal compression unit 210, and the reproduction information from the reproduction information generation unit 307.
- the data multiplexed by the multiplexing unit 308 is stored in the storage unit 330.
- a multiplexing method for example, there is a technique such as MPEG TS (Transport Stream). However, it is not limited to this. In this embodiment, the case of multiplexing is shown as an example, but it is not always necessary to multiplex.
- the processing of the attribute information extraction unit 305, the scene evaluation unit 306, the reproduction information generation unit 307, and the multiplexing unit 308 is sequentially performed during shooting, specifically during shooting or immediately after shooting.
- the digest playback unit 309 performs digest playback based on user input after the end of shooting. Specifically, the digest reproduction unit 309 reads the multiplexed data stored in the storage unit 330, and in accordance with the reproduction information, the video signal decompression unit 211 respectively encodes the encoded video data and the encoded audio data to be digest-reproduced. And output to the audio signal expansion unit 213. The output encoded video data and encoded audio data are decoded by the video signal expansion unit 211 and the audio signal expansion unit 213, respectively, and output from the video display unit 212 and the audio output unit 214. In this way, digest reproduction in which only a specific part is extracted from the video is executed.
- the digest video may be stored in the storage unit 330.
- FIG. 3 is a diagram showing a configuration of an image captured by the video camera 100A
- FIG. 4 shows an example in which a clip is divided into a plurality of scenes.
- each scene is specified by “start time” and “end time”, but each scene may be specified by a frame number (start frame number and end frame number of a scene) or the like.
- the unit of video shot until the user instructs to start shooting and ends shooting or pauses shooting is “clip”. That is, when the user repeats the start of shooting, the end of shooting, or the pause many times, a plurality of clips are generated.
- One clip is composed of one or more “scenes”.
- a “scene” is a series of logically connected images, and is composed of one or more “frames”.
- a “frame” is an individual image that is a minimum unit constituting a video.
- one “clip” can be regarded as one “scene”.
- a “scene” may be set on the boundary that the screen changes greatly. For example, when the video analysis unit 303 calculates a motion vector between frames and the magnitude of motion (that is, change in motion vector) is larger than a predetermined value, the boundary of the “scene” may be used. That is, the video between the two boundaries set in this way becomes one “scene”.
- the “scene” may be divided based on other shooting information or the like.
- the “scene” may be divided by a button input from the photographer. In this case, the “scene” in the “clip” is configured with the clear intention of the photographer.
- the “clip” includes a continuous “scene” as shown in FIG.
- only a specific part in the “clip” can be regarded as a “scene”.
- an important part of the video can be regarded as a “scene”.
- a part including specific attribute information may be regarded as one “scene”.
- a video having a predetermined time width including attribute information assumed to be important is defined as a “scene”.
- only important parts are extracted as “scenes”.
- the “clip” includes discrete “scenes”.
- the “scene” can be set arbitrarily.
- an important part of the video is regarded as a “scene”.
- FIG. 5 is a table of correspondence data of various attribute information and evaluation for each attribute information used when evaluating a video. This table is stored in the storage unit 330. The scene evaluation unit 306 evaluates the video using this table.
- an evaluation value is set for each attribute information.
- the higher the evaluation value the higher (preferred) the evaluation.
- clip-in (shooting start part) and clip-out (immediately before shooting end part) are introductory and important parts of the video, and it is estimated that the logical meaning of the video is high.
- the evaluation value “100” is set for in (A), and the evaluation value “90” is set for clip-out (F). Since zoom-up (D) and zoom-down (G) as camera work at the time of shooting increase the degree of attention to a specific subject, an evaluation value “30” is set.
- the fact that a face has been detected is often set to an evaluation value “50” because photographing is often performed on a person. Further, among the face detections, the fact that the face of the specific person A is detected (X) is set with an evaluation value of “100”, and the face of the specific person B is detected (Y) The value “80” is set.
- the user's face and the evaluation value for the face can be appropriately set by the user, as will be described in detail later. That is, it is not only that a person is photographed, but a high evaluation value can be assigned to a video image of a specific person with the intention of the user.
- the evaluation may include not only positive evaluation, that is, favorable evaluation, but also negative evaluation, that is, unfavorable evaluation. For example, since the image blur may be a video that is difficult for the viewer to see, a negative evaluation value is assigned to a scene having such attribute information.
- evaluation is digitized, it is not restricted to this.
- codes such as A, B, C,... May be used for evaluation. Superiority or inferiority is determined in advance for the codes used for evaluation (for example, A has the highest evaluation).
- evaluation of codes such as A, B, and C can be freely set by the user's intention.
- the scene evaluation unit 306 assigns an evaluation value corresponding to the attribute information to the part of the video from which the attribute information is extracted by the attribute information extraction unit 305 based on the table.
- the scene evaluation unit 306 extracts a predetermined number of scenes based on the evaluation value.
- the scene evaluation unit 306 extracts in advance a characteristic video portion that can be used for digest playback as a scene more than the number of scenes to be digest played back.
- the scene evaluation unit 306 extracts a video having a predetermined time width including a portion having attribute information with a high evaluation value as one scene.
- the scene evaluation unit 306 extracts a predetermined number of scenes in descending order of evaluation values.
- the predetermined number may be arbitrarily set by the user, or may be set in advance as a fixed value.
- the scene extraction method described above is an example, and the scene may be extracted by a different method.
- a part having attribute information with a high evaluation value may not be extracted as a scene when a part including the same attribute information in a previous video is extracted as a scene. By doing so, it is possible to prevent only scenes having the same attribute information from being extracted.
- a part having specific attribute information for example, face detection of the person A, face detection of the person B, etc. may be preferentially extracted as a scene.
- the scene evaluation unit 306 extracts a scene to be digest-reproduced from the extracted scenes based on a predetermined extraction condition. For example, if the extraction condition is three in descending order of evaluation value, the scene evaluation unit 306 extracts the three scenes with the highest evaluation value. This number can be arbitrarily set by the user. Also, if the extraction condition is such that the evaluation time is extracted in descending order of the evaluation value and the total time becomes the predetermined time, the scene evaluation unit 306 determines that the evaluation value is higher in order so that the total time becomes the predetermined time. Extract sequentially from the scene.
- the predetermined time may be set to a predetermined value in advance, or may be arbitrarily set by the user.
- the scene evaluation unit 306 extracts scenes whose evaluation value is equal to or greater than the predetermined value regardless of the number and the total time.
- the predetermined value can be arbitrarily set by the user.
- the scene evaluation unit 306 can extract a scene from various viewpoints based on the assigned evaluation value.
- the extraction condition may be set by the user as appropriate, or may be set in advance.
- an evaluation value assigned to the content of each attribute information may be added to obtain an evaluation value of the scene.
- the highest evaluation value among a plurality of attribute information may be used as the evaluation value of the scene.
- the table of attribute information and evaluation values is not limited to the one that the video camera 100A has only one. That is, the video camera 100A may have a plurality of tables of attribute information and evaluation values, and appropriately select a table used for scene evaluation. For example, the video camera 100A selects an optimal table from a plurality of tables of attribute information and evaluation values according to the shooting mode (for example, landscape shooting, person (portrait) shooting, sports shooting, still life shooting, etc.). You may comprise so that it may select. In addition, as a configuration for appropriately setting an appropriate table according to the shooting situation in this way, a one-to-one table is not prepared in advance for each shooting situation, but rather than the type of shooting situation.
- the shooting mode for example, landscape shooting, person (portrait) shooting, sports shooting, still life shooting, etc.
- a plurality of tables may be combined (e.g., each evaluation value is added at a constant ratio) according to the shooting situation.
- the table corresponding to the shooting situation may be set by changing the weight of each table at the time of synthesis (for example, the ratio at the time of addition).
- FIG. 6 is a diagram showing a result of the scene evaluation unit 306 extracting attribute information from a certain video and assigning an evaluation value.
- the horizontal axis in FIG. 6 represents time (scene), and the vertical axis represents the evaluation value.
- the portion near time 0 has “clip-in” attribute information A meaning that it is immediately after the start of shooting, and an evaluation value “100” is given.
- the part having attribute information B is a part from which a specific voice is extracted.
- the extraction of the specific voice is performed by the voice analysis unit 304.
- An evaluation value “50” is assigned to the portion having the attribute information B.
- the portion having the attribute information C is a portion where the user is shooting still after panning or tilting the video camera 100A. Since it can be determined that the portion after camera work such as pan and tilt has high value as a video, such still shooting after camera work is set as attribute information. An evaluation value “40” is assigned to the portion having the attribute information C.
- the part having the attribute information D is a part that is photographed by zooming up or down.
- the zoom-up or zoom-down reflects the user's intention regarding shooting and can be determined to be important, so it is set as attribute information.
- An evaluation value “30” is assigned to the portion having the attribute information D.
- the evaluation value may be changed between zoom-up and zoom-down. For example, the evaluation value may be set higher because it is determined that zooming up has a greater intention to gaze at a subject to be photographed than zooming down.
- the part having the attribute information E is a part where the video camera 100A was shot while panning, tilting, or the like. Camera work such as panning and tilting is set as attribute information because it can be determined that the shooting intention of the user who wants to follow the shooting target is reflected. An evaluation value “25” is assigned to the portion having the attribute information E.
- the part having the attribute information I is a part where the image is accompanied by image blurring. In this case, since the video is shaking, the video tends to be difficult for viewers to see. Therefore, a negative evaluation value is given. Specifically, the evaluation value “ ⁇ 20” is assigned to the portion having the attribute information I.
- the part having the attribute information J is a part where the ground is photographed. This is a phenomenon that is likely to occur when the user is walking with the video camera 100A in his / her hand while continuing to shoot without pressing the shooting stop button. In this case, since it can be determined that the user's special intention is not reflected in the video, a negative evaluation value is assigned. Specifically, the evaluation value “ ⁇ 10” is assigned to the portion having the attribute information J.
- the part having the attribute information X is a part in which the face of the person A is reflected.
- the video analysis unit 303 recognizes that the subject in the captured video is a person's face, and the recognized face matches the face of a specific person stored in the storage unit 330 or the like in advance. It is determined whether or not. Based on this result, the scene evaluation unit 306 can extract a part in which the face of a specific person is captured from the video. An evaluation value “100” is assigned to the portion having the attribute information X.
- the part having the attribute information Y is a part in which the face of the person B is reflected. Like the face of the person A, the face of the person B is stored in advance in the storage unit 330 or the like. An evaluation value “80” is set in the portion having the attribute information Y.
- the portion having the attribute information Z is a portion in which a person's face is reflected, but the face does not match any face of any person registered in advance.
- An evaluation value “50” is set in the portion having the attribute information Z.
- FIGS. 7A to 7D there are cases as shown in FIGS. 7A to 7D as examples of images in which a person's face is photographed.
- the scene evaluation unit 306 has an evaluation value set in advance in the table (FIG. 5). In the example, “100” is assigned to the face of the person A, and “80” is assigned to the face of the person B).
- the scene evaluation unit 306 uses the evaluation value with the higher evaluation value of the faces of the persons A and B. It is good also as an evaluation value.
- “100” that is the evaluation value of the face of the person A is the evaluation value. 6 is a portion where the faces of both the person A and the person B are shown.
- an average of both evaluation values may be used as the evaluation value. In the example of FIG. 5, the evaluation value is “90” from (100 + 80) / 2.
- a distribution ratio may be set for each evaluation value and summed up.
- the distribution ratio may be set for each evaluation value such that the larger the face size on the video is, the higher the distribution ratio is.
- the evaluation value is “92” from (100 ⁇ 5 + 80 ⁇ 3) / 8. .5 ".
- the size of the face on the image reflects the distance from the camera to the subject. According to this method, a large distribution ratio can be set for nearby subjects to increase the degree of influence. it can.
- the distribution ratio may be set for each evaluation value according to the center position of the faces of the persons A and B and the distance from the center of the screen or the salient area (Saliant area) on the screen. Specifically, the distribution ratio may be higher as the distance is shorter.
- region is an area
- FIG. 7D when a large number of persons (more than a predetermined number) are photographed as subjects, the face of the person A is photographed, but the size on the screen is small. Because it is taken together with many other people, it may be difficult to differentiate from the faces of other people. In such a case, a distribution ratio may be set and added to the evaluation value of the face of the person A and the evaluation value of the face of another person according to the number of people who are photographed. In the example of FIG. 7D, since there are 10 persons other than the person A who cannot be identified, the evaluation value “54.5” is given from (100 ⁇ 1 + 50 ⁇ 10) / 11.
- For face detection evaluate the position on the screen, face size, face orientation, smile level, eye opening / closing information, face emotion level information, and increase or decrease the evaluation value accordingly. You may let them.
- the scene evaluation unit 306 extracts six scenes in descending order of evaluation value.
- six scenes are labeled # 1 to # 6 in order of time.
- the scene evaluation unit 306 extracts # 1, # 2, and # 5 scenes with the highest evaluation values as scenes to be digest-reproduced based on the extraction condition of three in descending order of evaluation values.
- the reproduction information generation unit 307 generates reproduction information that is information for specifying a scene to be digest reproduced according to the scene extracted by the scene evaluation unit 306.
- the reproduction information may be indicated by the start time and end time of the scene to be reproduced as shown in FIG.
- storing a representative frame in each scene is effective for searching the reference screen.
- the reproduction information is not limited to the above contents, and for example, a scene to be reproduced may be specified by a frame number.
- the position (location) of the corresponding scene in the multiplexed data generated by the multiplexing unit 308 described later may be used as the scene specification.
- reproduction information may be generated using time information such as PTS and DTS.
- time information such as PTS and DTS.
- AVCHD Advanced Video Codec High Definition
- FIG. 9 is a flowchart showing the flow of processing from shooting, scene evaluation, generation of reproduction information, and recording.
- the control unit 300 of the video camera 100A starts shooting in step S101.
- Shooting is started based on an input from the external input unit 207 such as an input button.
- the attribute information extraction unit 305 determines the video attribute based on the detection result of the posture detection unit 206, the control information of the lens control unit 301, the analysis results of the video analysis unit 303 and the audio analysis unit 304, and the like. Extract information.
- step S103 the scene evaluation unit 306 assigns an evaluation value to each part of the video based on the attribute information extracted by the attribute information extraction unit 305. Thereafter, the scene evaluation unit 306 extracts some characteristic scenes, and further extracts a scene to be digest reproduced from them.
- step S104 the reproduction information generation unit 307 generates reproduction information based on the scene to be digest reproduced extracted by the scene evaluation unit 306. Then, the multiplexing unit 314 multiplexes the generated reproduction information together with the encoded video data and the encoded audio data.
- the control unit 300 stores the multiplexed data in the storage unit 330 in step S105.
- step S106 the control unit 300 determines whether or not there is an end of photographing from the external input unit 207. If there is no input for the end of shooting, the process returns to step S102 to continue shooting. On the other hand, if there is an input for the end of shooting, shooting is ended.
- the digest playback unit 309 reads the playback information stored in the storage unit 330, and performs digest playback based on the read information. Specifically, the digest playback unit 309 corresponds to the video and audio information stored in the storage unit 330 based on information such as the start time and end time of each scene to be digest played as shown in FIG. Extract partial scenes.
- FIG. 10 is a flowchart showing the flow of processing during digest playback.
- the digest reproduction unit 309 reads the multiplexed data stored in the storage unit 330 in step S201.
- step S202 the digest reproduction unit 309 decomposes the read multiplexed data and extracts reproduction information.
- step S203 the digest reproduction unit 309 outputs the encoded video data and the encoded audio data to be reproduced to the video signal decompression unit 211 and the audio signal decompression unit 213 based on the extracted reproduction information. Video and audio are reproduced via the display unit 212 and the audio output unit 214.
- digest playback is performed by extracting only a specific scene from the video.
- the face detection of the person A and the face detection of the person B in the above description are additionally set by the user. That is, the user registers the face detection of the person A as new attribute information in the table previously provided in the video camera 100A, registers the evaluation value, and further detects the face detection of another person B with the new attribute. It is registered as information and its evaluation value is also registered. As a result, it is possible not only to detect a person's face but also to detect a face of a specific person in more detail.
- the face recognition table is stored in the storage unit 330 in advance.
- the face recognition table is configured such that ID, face image, person name, and evaluation value are set as one set, and only a plurality of sets (for example, 6 sets) can be registered.
- the ID and the evaluation value are set in advance, and the user registers the face and name of a specific person as appropriate.
- the control unit 300 prompts the user to photograph the face of the person to be registered for a predetermined time (for example, 3 seconds) or more.
- the message is displayed on the video display unit 212.
- the control unit 300 causes the video display unit 212 to display a message asking which ID in the face recognition table the face of the person is to correspond to. An evaluation value has already been set for each ID.
- the evaluation value of ID1 is the highest at “100”, and the evaluation value is set to be smaller as the ID number is larger. Has been. That is, associating a person's face to be registered with an ID is equivalent to setting the evaluation value of the person's face.
- the control unit 300 next causes the video display unit 212 to display a message that prompts the user to input the name of the person.
- the ID, the face image of a specific person, the name, and the evaluation value are set in the face recognition table.
- the evaluation value is set in advance, but the evaluation value may be arbitrarily input by the user. In that case, a message prompting the user to input an evaluation value may be displayed on the video display unit 212 so that the user can input the evaluation value.
- Attribute information and evaluation correspondence data set in this way are used in various ways based on user selection. For example, in the above example, it is set as attribute information that the face of the person A, the face of the person B, and the face of the other person are detected, but the person who is the detected face is identified. Instead, the fact that a person's face has been detected can be extracted as attribute information. That is, the video camera 100A does not specify a person, and simply extracts as a piece of attribute information that a face has been detected, and a specific mode extracts as a piece of attribute information that a face of a specific person has been detected. Have In the specific mode, it is also possible to select a face to be extracted as attribute information from registered human faces.
- control unit 300 causes the video display unit 212 to display a registered person's face image, name, or ID.
- the user operates the video display unit 212 to select a human face to be extracted as attribute information.
- the control unit 300 extracts the face of the selected person as attribute information.
- the conditions for extracting face detection as attribute information may be changed between the normal mode and the specific mode. That is, in the normal mode, it is extracted as attribute information when a face of an unspecified person is captured in the video for a predetermined first time or more.
- the specific mode the face of a specific person (for example, the person A) is captured for a second time or shorter than the first time (for example, is captured only for one frame) during shooting. It is extracted as attribute information. That is, in the normal mode, detection of a face is set as attribute information from the idea that a person is generally more important as a shooting target than a landscape or the like.
- the specific mode is clear and strong for the user who wants to extract a video showing a specific person's face rather than whether the specific person's face is more important than other shooting targets such as landscapes. This mode reflects the intention. Therefore, the importance of face detection in the specific mode is higher than that in the normal mode. Therefore, in the specific face detection mode, the condition for certifying that a face has been detected is relaxed compared to the usual face detection mode.
- the importance of the face of a specific person may be increased by making the evaluation value of the face of a specific person higher than the evaluation value of an unspecified face without changing the face detection conditions.
- FIG. 11 is a table of correspondence data of various attribute information and evaluation for each attribute information used when evaluating a video.
- FIG. 12 shows a scene evaluation unit 306 based on the table of FIG. It is the figure which showed the result of having extracted the attribute information from the image
- the horizontal axis represents time (scene), and the vertical axis represents the evaluation value of each scene.
- FIG. 13 shows reproduction information generated from the evaluation based on the table of FIG.
- the evaluation value of the attribute information for the face detection of the person A is “60”, whereas the evaluation value of the attribute information for the face detection of the person B is “90”.
- the result shown in FIG. 12 is obtained. Specifically, compared with the evaluation using the table of FIG. 5, the evaluation of the scene # 2 is lowered and the evaluation of the scene # 4 is increased.
- scene # 4 is added to the digest instead of scene # 2 in FIG. 8, as shown in FIG.
- the change of the evaluation value as described above may be performed by the user rewriting the evaluation value of the table, or tables having different evaluation values (the table in FIG. 5 and the table in FIG. 11) are prepared in advance and switched. It may be.
- a selection screen for allowing the user to select a mode corresponding to various tables is displayed on the video display unit 212, and the user can select the external input unit 207.
- the mode may be selected via the button.
- various tables may be displayed on the video display unit 212 so that the user can select a table.
- the table prepared in advance may be created in advance by direct input of attribute information or evaluation values by the user.
- digest reproduction reflecting user preference can be performed by allowing the user to arbitrarily input at least one of attribute information and evaluation.
- the user since the user can additionally set attribute information, it is possible to reproduce a video that better matches the user's preference as a digest.
- the video camera 100 includes an external input unit 207 that inputs attribute information about a video, and extracts the attribute information from the video in order to extract a portion to be digest-reproduced. And a control unit 300 that extracts from the control unit 300.
- the user can appropriately input the attribute information used for extracting the portion to be digest-reproduced from the video.
- a video that matches the user's preference can be reproduced as a digest.
- the video camera 100 also includes an external input unit 207 that inputs at least one of the attribute information and the evaluation according to a user input operation with respect to the attribute information about the video and the data corresponding to the evaluation with respect to the attribute information.
- the control unit 300 extracts the attribute information from the video and evaluates the portion having the attribute information based on the corresponding data. Thereby, the user can appropriately input the attribute information and / or the evaluation value used for extracting the portion to be digest reproduced from the video. As a result, a video that matches the user's preference can be reproduced as a digest.
- attribute information of a lower concept called face detection of a specific person with respect to attribute information called face detection (face detection without specifying a person) in advance, the user's deeper preference can be set.
- face detection face detection without specifying a person
- attribute information such as clip-in, clip-out, zoom-up, and the like
- fixed attribute information and user's preference are likely to appear, such as detection of a specific person's face.
- attribute information variable attribute information the processing can be simplified. In other words, if all the attribute information is variable, control contents (input of attribute information, extraction of attribute information, etc.) and memory capacity corresponding to various attribute information must be prepared, and the processing becomes complicated. End up. On the other hand, by narrowing down variable attribute information to some extent, it is possible to reduce control contents and memory capacity prepared in advance, and the process is simplified.
- attribute information such as the attitude of the video camera 100A is difficult to determine from the video afterwards, or is possible but troublesome, and can be easily detected by a detection signal of a sensor at the time of shooting. . That is, some attribute information is easier to detect at the time of shooting. Therefore, such attribute information can be easily extracted by extracting the attribute information at the time of shooting.
- the video recorded on the imaging device such as the video camera 100A is a video just taken without being edited. Therefore, there are many videos with low importance, and the digest playback as described above is very effective.
- FIG. 15 is a block diagram showing a schematic configuration of the video camera 100B.
- the video camera 100B is different from the video camera 100A of the first embodiment in that a scene to be reproduced is selected when performing digest reproduction.
- the basic configuration of the video camera 100B is substantially the same as that of the video camera 100A, and the data flow, that is, the processing order is different from that of the video camera 100A. Therefore, the same configurations as those of the first embodiment are denoted by the same reference numerals, description thereof is omitted, and different portions are mainly described.
- the processing until the attribute information extraction unit 305 extracts the attribute information is the same as that of the first embodiment.
- the attribute information extracted by the attribute information extraction unit 305 is input to the multiplexing unit 308.
- the multiplexing unit 308 multiplexes the encoded video data from the video signal compression unit 204, the encoded audio data from the audio signal compression unit 210, and the attribute information from the attribute information extraction unit 305, and outputs the multiplexed data.
- the multiplexed data is stored in the storage unit 330.
- the scene evaluation unit 306 reads the multiplexed data from the storage unit 330, assigns an evaluation to each part of the video based on the attribute information, and a characteristic scene from the video And a scene to be digest-reproduced is further extracted therefrom.
- the playback information generation unit 307 generates playback information based on the scene extracted by the scene evaluation unit 306 and outputs the playback information to the digest playback unit 309.
- the digest reproduction unit 309 reads out corresponding data from the storage unit 330 based on the reproduction information generated by the reproduction information generation unit 307, and outputs the data to the video signal expansion unit 211 and the audio signal expansion unit 213. Then, the digest video is reproduced by the video display unit 212 and the audio output unit 214.
- FIG. 16 is a flowchart showing the flow of processing from shooting to attribute information extraction to recording.
- FIG. 17 is a flowchart showing the flow of processing during digest playback.
- step S301 the control unit 300 of the video camera 100A starts shooting. Shooting is started based on an input from the external input unit 207 such as an input button.
- the attribute information extraction unit 305 determines the video attribute based on the detection result of the posture detection unit 206, the control information of the lens control unit 301, the analysis results of the video analysis unit 303 and the audio analysis unit 304, and the like. Extract information.
- the multiplexing unit 314 multiplexes the attribute information together with the encoded video data and the encoded audio data.
- the control unit 300 stores these multiplexed data in the storage unit 330.
- step S106 the control unit 300 determines whether or not there is an end of photographing from the external input unit 207. If there is no input for the end of shooting, the process returns to step S302 to continue shooting. On the other hand, if there is an input for the end of shooting, shooting is ended.
- the scene evaluation unit 306 reads the multiplexed data stored in the storage unit 330 in step S401.
- step S402 the scene evaluation unit 306 decomposes the read multiplexed data and reads the attribute information. Subsequently, in step S403, the scene evaluation unit 306 assigns an evaluation value to each part of the video based on the attribute information. In step S404, the scene evaluation unit 306 determines whether evaluation of all parts of the video has been completed. If not completed, the scene evaluation unit 306 returns to step S401 and continues to evaluate the video. On the other hand, if the evaluation has been completed, the scene evaluation unit 306 proceeds to step S405.
- step S405 the scene evaluation unit 306 extracts some characteristic scenes from the video based on the evaluation value, and further extracts a scene to be digest-reproduced from the scenes. Then, the reproduction information generation unit 307 generates reproduction information based on the scene to be digest reproduced extracted by the scene evaluation unit 306.
- step S406 the digest reproduction unit 309 reads out the encoded video data and encoded audio data to be reproduced from the storage unit 330 based on the reproduction information, and sends them to the video signal expansion unit 211 and the audio signal expansion unit 213.
- the video and audio are reproduced via the video display unit 212 and the audio output unit 214.
- the evaluation value for the attribute information can be changed when digest playback is executed after shooting.
- the configuration of the video camera is not limited to the above embodiment.
- the video camera 100 ⁇ / b> C may execute the process up to extracting attribute information from the captured video. Then, the video playback device 500 may execute scene evaluation and digest playback based on the video data to which the attribute information is added. Further, the video camera may execute up to the scene evaluation. Then, the video reproduction device may perform digest reproduction based on the evaluation value.
- the present embodiment is not limited to a photographing apparatus such as a video camera, but can also be applied to a video editing apparatus such as a PC or a video recording / reproducing apparatus such as an HDD recorder.
- the video editing device or the video recording / playback device may extract attribute information from the video, evaluate each part of the video based on the attribute information, and perform digest playback based on the evaluation.
- the attribute information is first extracted from the video, each part is evaluated, the scene is extracted based on the attribute information, and then the digest video is generated based on the scene.
- the video may be divided into a plurality of scenes, attribute information of each scene may be extracted, each scene may be evaluated based on the attribute information, and a digest video may be generated based on the evaluation.
- the scene division may be performed based on a motion vector between frames, may be performed based on an input from the user, or the scene may be included in various attribute information. It may be executed based on the presence / absence of attribute information that influences the division of the image (thing that becomes a segment of the scene such as camera work or shooting of the ground).
- the configuration of the video camera 100A is not limited to the above embodiment.
- all or some of the video AD conversion unit 202, the signal processing unit 203, the video signal compression unit 204, the audio AD conversion unit 209, the audio signal compression unit 210, the video signal decompression unit 211, and the audio image signal decompression unit 213 are simply used. It can also be realized as a single integrated circuit. Also, a part of the processing executed by the control unit 300 can be separately realized as hardware using an FPGA (Field Programmable Gate Gate Array).
- FPGA Field Programmable Gate Gate Array
- the face detection of a person is set as upper attribute information and fixed attribute information
- the face detection of a specific person is set as lower attribute information and variable attribute information.
- the present invention is not limited to this.
- the detection of the face of an animal such as a dog may be set as upper attribute information and fixed attribute information
- the face detection of a specific dog may be input by the user as lower attribute information and variable attribute information.
- detection of transportation means such as trains, cars or airplanes can be input as higher attribute information and fixed attribute information
- detection of a specific train, car or airplane can be input as lower attribute information and variable attribute information. Also good.
- detection of a person's voice may be input as higher-level attribute information and fixed attribute information, and detection of a specific person's voice may be input by the user as lower-level attribute information and variable attribute information.
- a voice of a specific person can be input to the video camera via the microphone 208.
- the evaluation value of the table may be individually input (changed) by the user.
- the evaluation value related to camera work can be increased or decreased.
- the technique disclosed herein is useful for a video extraction apparatus that extracts a portion to be digest-reproduced from an image and a photographing apparatus equipped with the same.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Computer Security & Cryptography (AREA)
- Television Signal Processing For Recording (AREA)
- Studio Devices (AREA)
- Management Or Editing Of Information On Record Carriers (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Indexing, Searching, Synchronizing, And The Amount Of Synchronization Travel Of Record Carriers (AREA)
Abstract
Description
<1.ビデオカメラの構成>
図1は、撮影装置の一例としてのビデオカメラ100Aの斜視図であり、図2は、ビデオカメラ100Aの概略構成を示すブロック図である。
図3は、ビデオカメラ100Aが撮影した映像の構成を示す図であり、図4は、クリップ内を複数のシーンに分割した場合の例を示すものである。図4では、各シーンを「開始時間」と「終了時間」で特定しているが、フレーム番号(シーンの開始フレーム番号及び終了フレーム番号)等で各シーンを特定してもよい。
ダイジェスト再生部309は、記憶部330に記憶された再生情報を読み出し、それに基づいてダイジェスト再生を行う。具体的には、ダイジェスト再生部309は、図8に示されるダイジェスト再生すべき個々のシーンの開始時刻、終了時刻等の情報に基づいて記憶部330に記憶されている映像、音声情報から該当する部分のシーンを抽出する。
続いて、属性情報及び評価値の、ユーザによる入力について説明する。本実施形態に係るビデオカメラ100Aでは、テーブルの属性情報及び評価をユーザが1つ1つ具体的に入力することができる。
したがって、本実施形態によれば、属性情報及び評価の少なくとも一方をユーザが任意に入力可能とすることによって、ユーザの嗜好を反映したダイジェスト再生を行うことができる。
続いて、実施形態2に係るビデオカメラ100Bについて図15を参照しながら説明する。図15は、ビデオカメラ100Bの概略構成を示すブロック図である。ビデオカメラ100Bは、ダイジェスト再生を行うときに再生すべきシーンを選択する点で実施形態1のビデオカメラ100Aと異なる。具体的には、ビデオカメラ100Bの基本的な構成は、ビデオカメラ100Aと略同様であり、データの流れ、即ち、処理の順序がビデオカメラ100Aと異なる。そこで、実施形態1と同様の構成については、同様の符号を付して説明を省略し、異なる部分を中心に説明する。
前記実施形態について、以下のような構成としてもよい。
300 制御部
305 属性情報抽出部
306 シーン評価部
307 再生情報生成部
309 ダイジェスト再生部
Claims (12)
- 映像から抽出すべき、映像に関する属性情報を記憶する記憶部と、
映像の中からダイジェスト再生する部分を抽出するために、前記記憶部に記憶された前記属性情報を映像の中から抽出する制御部とを備え、
前記記憶部に記憶された前記属性情報の少なくとも一部は、ユーザが入力可能に構成されている映像抽出装置。 - 映像から抽出すべき、映像に関する属性情報及び該属性情報に対する評価との対応データを記憶する記憶部と、
映像の中からダイジェスト再生する部分を抽出するために、前記記憶部に記憶された前記属性情報を映像の中から抽出し、前記対応データに基づいて該属性情報を有する部分を評価する制御部とを備え、
前記記憶部に記憶された前記属性情報及び評価の少なくとも一部は、ユーザが入力可能に構成されている映像抽出装置。 - 映像から抽出すべき、映像に関する属性情報を記憶する記憶部と、
ユーザの入力操作に応じて、前記属性情報を前記記憶部に入力する入力部と、
映像の中からダイジェスト再生するための部分を前記記憶部に記憶された前記属性情報に基づいて抽出する制御部とを備える映像抽出装置。 - 映像から抽出すべき、映像に関する属性情報及び該属性情報に対する評価との対応データを記憶する記憶部と、
ユーザの入力操作に応じて、前記属性情報及び評価の少なくとも一方を前記記憶部に入力する入力部と、
映像の中からダイジェスト再生するための部分を前記記憶部に記憶された前記属性情報及び評価に基づいて抽出する制御部とを備える映像抽出装置。 - 請求項1乃至4の何れか1つに記載の映像抽出装置において、
前記記憶部には、複数の特定の人物の顔が記憶されており、
前記記憶部に記憶された人物の顔が検出されたことを、ユーザからの入力によって属性情報として設定可能である映像抽出装置。 - 請求項1乃至5の何れか1つに記載の映像抽出装置において、
前記属性情報は、上位の属性情報と、該上位の属性情報よりも詳細な属性を示す下位の属性情報とを含んでいる映像抽出装置。 - 請求項6に記載の映像抽出装置において、
前記下位の属性情報は、ユーザが入力可能である映像抽出装置。 - 請求項6又は7に記載の映像抽出装置において、
前記上位の属性情報は、人物の顔が検出されたことであり、
前記下位の属性情報は、特定の人物の顔が検出されたことである映像抽出装置。 - 請求項1乃至8の何れか1つに記載の映像抽出装置において、
前記属性情報は、ユーザが入力できない固定の属性情報をと、ユーザが入力可能な可変の属性情報とを含んでいる映像抽出装置。 - 映像を取得する撮像系と、
請求項1乃至9の何れか1つに記載の映像抽出装置とを備えた撮影装置。 - 映像から抽出すべき、映像に関する属性情報及び該属性情報に対する評価との対応データについて、該属性情報及び評価の少なくとも一方のユーザからの入力を受け付けることと、
前記対応データに含まれる前記属性情報を映像の中から抽出することと、
前記対応データと抽出された前記属性情報とに基づいて、映像の中からダイジェスト再生する部分を抽出することとをコンピュータに実行させるためのプログラム。 - 映像から抽出すべき、映像に関する属性情報及び該属性情報に対する評価との対応データについて、該属性情報及び評価の少なくとも一方のユーザからの入力を受け付けることと、
前記対応データに含まれる前記属性情報を映像の中から抽出することと、
前記対応データと抽出された前記属性情報に基づいて、映像の中からダイジェスト再生する部分を抽出することとをコンピュータに実行させるためのプログラムを記録したコンピュータ読み取り可能な記録媒体。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2011537781A JPWO2011099299A1 (ja) | 2010-02-10 | 2011-02-10 | 映像抽出装置、撮影装置、プログラム及び記録媒体 |
US13/578,441 US8964065B2 (en) | 2010-02-10 | 2011-02-10 | Video picture extracting apparatus, imaging apparatus and recording medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010027228 | 2010-02-10 | ||
JP2010-027228 | 2012-02-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011099299A1 true WO2011099299A1 (ja) | 2011-08-18 |
Family
ID=44367585
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/000771 WO2011099299A1 (ja) | 2010-02-10 | 2011-02-10 | 映像抽出装置、撮影装置、プログラム及び記録媒体 |
Country Status (3)
Country | Link |
---|---|
US (1) | US8964065B2 (ja) |
JP (2) | JPWO2011099299A1 (ja) |
WO (1) | WO2011099299A1 (ja) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015104780A1 (ja) * | 2014-01-07 | 2015-07-16 | パナソニックIpマネジメント株式会社 | 映像撮像装置 |
WO2017197817A1 (zh) * | 2016-05-20 | 2017-11-23 | 乐视控股(北京)有限公司 | 一种数据处理方法、装置、电子设备及服务器 |
JP2018007134A (ja) * | 2016-07-06 | 2018-01-11 | 日本放送協会 | シーン抽出装置およびそのプログラム |
JP2019216364A (ja) * | 2018-06-13 | 2019-12-19 | キヤノン株式会社 | 情報処理装置、情報処理方法、及びプログラム |
Families Citing this family (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10296874B1 (en) | 2007-12-17 | 2019-05-21 | American Express Travel Related Services Company, Inc. | System and method for preventing unauthorized access to financial accounts |
US8172135B1 (en) | 2011-06-24 | 2012-05-08 | American Express Travel Related Services Company, Inc. | Systems and methods for gesture-based interaction with computer systems |
US8714439B2 (en) | 2011-08-22 | 2014-05-06 | American Express Travel Related Services Company, Inc. | Methods and systems for contactless payments at a merchant |
US10482753B2 (en) * | 2013-01-04 | 2019-11-19 | Minnesota Imaging And Engineering Llc | Infection control monitoring system |
KR102217186B1 (ko) * | 2014-04-11 | 2021-02-19 | 삼성전자주식회사 | 요약 컨텐츠 서비스를 위한 방송 수신 장치 및 방법 |
CN104284240B (zh) * | 2014-09-17 | 2018-02-02 | 小米科技有限责任公司 | 视频浏览方法及装置 |
US9799376B2 (en) | 2014-09-17 | 2017-10-24 | Xiaomi Inc. | Method and device for video browsing based on keyframe |
US20170112381A1 (en) * | 2015-10-23 | 2017-04-27 | Xerox Corporation | Heart rate sensing using camera-based handheld device |
JP6379307B2 (ja) * | 2015-12-22 | 2018-08-22 | 富士フイルム株式会社 | 撮像装置、合焦制御方法、及び合焦制御プログラム |
JP6878718B2 (ja) * | 2016-07-19 | 2021-06-02 | Seven Dew株式会社 | ダイジェスト映像取得装置、ダイジェスト映像取得方法、およびプログラム |
US10365383B2 (en) | 2016-09-09 | 2019-07-30 | Minnesota Imaging And Engineering Llc | Structured detectors and detector systems for radiation imaging |
CN108012081B (zh) * | 2017-12-08 | 2020-02-04 | 北京百度网讯科技有限公司 | 智能美颜方法、装置、终端和计算机可读存储介质 |
US10701365B2 (en) * | 2017-12-11 | 2020-06-30 | Disney Enterprises, Inc. | Systems and methods for scene categorization |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007060060A (ja) * | 2005-08-23 | 2007-03-08 | Sony Corp | 再生システム、再生装置、再生方法、情報処理装置、情報処理方法、およびプログラム |
JP2008278467A (ja) * | 2007-03-30 | 2008-11-13 | Sanyo Electric Co Ltd | 画像処理装置および画像処理方法 |
JP2009077026A (ja) * | 2007-09-19 | 2009-04-09 | Fujifilm Corp | 撮影装置および方法並びにプログラム |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4230870B2 (ja) | 2003-09-25 | 2009-02-25 | 富士フイルム株式会社 | 動画記録装置、動画記録方法、及びプログラム |
JP2007072520A (ja) | 2005-09-02 | 2007-03-22 | Sony Corp | 映像処理装置 |
US7889244B2 (en) * | 2005-12-27 | 2011-02-15 | Panasonic Corporation | Image processing apparatus |
JP2008103802A (ja) | 2006-10-17 | 2008-05-01 | Sharp Corp | 映像合成装置 |
JP4960121B2 (ja) | 2007-03-12 | 2012-06-27 | パナソニック株式会社 | コンテンツ撮影装置 |
WO2008111308A1 (ja) | 2007-03-12 | 2008-09-18 | Panasonic Corporation | コンテンツ撮影装置 |
US20080240503A1 (en) | 2007-03-30 | 2008-10-02 | Sanyo Electric Co., Ltd. | Image Processing Apparatus And Image Pickup Apparatus Mounting The Same, And Image Processing Method |
JP4360425B2 (ja) | 2007-06-15 | 2009-11-11 | ソニー株式会社 | 画像処理装置、その処理方法およびプログラム |
CN101588450B (zh) | 2008-05-19 | 2013-08-14 | 株式会社日立制作所 | 记录再现装置及方法 |
JP2010272109A (ja) * | 2009-04-20 | 2010-12-02 | Fujifilm Corp | 画像処理装置、画像処理方法およびプログラム |
-
2011
- 2011-02-10 WO PCT/JP2011/000771 patent/WO2011099299A1/ja active Application Filing
- 2011-02-10 JP JP2011537781A patent/JPWO2011099299A1/ja active Pending
- 2011-02-10 US US13/578,441 patent/US8964065B2/en active Active
-
2013
- 2013-10-17 JP JP2013216098A patent/JP5685732B2/ja active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007060060A (ja) * | 2005-08-23 | 2007-03-08 | Sony Corp | 再生システム、再生装置、再生方法、情報処理装置、情報処理方法、およびプログラム |
JP2008278467A (ja) * | 2007-03-30 | 2008-11-13 | Sanyo Electric Co Ltd | 画像処理装置および画像処理方法 |
JP2009077026A (ja) * | 2007-09-19 | 2009-04-09 | Fujifilm Corp | 撮影装置および方法並びにプログラム |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015104780A1 (ja) * | 2014-01-07 | 2015-07-16 | パナソニックIpマネジメント株式会社 | 映像撮像装置 |
WO2017197817A1 (zh) * | 2016-05-20 | 2017-11-23 | 乐视控股(北京)有限公司 | 一种数据处理方法、装置、电子设备及服务器 |
JP2018007134A (ja) * | 2016-07-06 | 2018-01-11 | 日本放送協会 | シーン抽出装置およびそのプログラム |
JP2019216364A (ja) * | 2018-06-13 | 2019-12-19 | キヤノン株式会社 | 情報処理装置、情報処理方法、及びプログラム |
JP7166796B2 (ja) | 2018-06-13 | 2022-11-08 | キヤノン株式会社 | 情報処理装置、情報処理方法、及びプログラム |
Also Published As
Publication number | Publication date |
---|---|
JP2014030260A (ja) | 2014-02-13 |
JP5685732B2 (ja) | 2015-03-18 |
US8964065B2 (en) | 2015-02-24 |
JPWO2011099299A1 (ja) | 2013-06-13 |
US20120307109A1 (en) | 2012-12-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5685732B2 (ja) | 映像抽出装置、プログラム及び記録媒体 | |
US8000558B2 (en) | Thumbnail generating apparatus and image shooting apparatus | |
JP6267961B2 (ja) | 映像提供方法および送信装置 | |
JP5456023B2 (ja) | 画像撮影装置、画像撮影方法、プログラム、及び集積回路 | |
KR100734705B1 (ko) | 동화 기록 장치 및 동화 재생 장치 | |
JP2009156888A (ja) | 音声補正装置及びそれを備えた撮像装置並びに音声補正方法 | |
WO2015098110A1 (ja) | 撮像装置、撮像システムおよび撮像方法 | |
JP4992639B2 (ja) | コンテンツ撮影装置 | |
JP4960121B2 (ja) | コンテンツ撮影装置 | |
JP2007266659A (ja) | 撮像再生装置 | |
JP2010245856A (ja) | 映像編集装置 | |
JP5600405B2 (ja) | 画像処理装置、画像処理方法及びプログラム | |
JP2011119936A (ja) | 撮影装置及び再生方法 | |
JP6295442B2 (ja) | 画像生成装置、撮影装置、画像生成方法及びプログラム | |
JP2011119934A (ja) | 画像撮影装置及び画像撮影方法 | |
JP6314321B2 (ja) | 画像生成装置、撮影装置、画像生成方法及びプログラム | |
JP6295443B2 (ja) | 画像生成装置、撮影装置、画像生成方法及びプログラム | |
WO2015104780A1 (ja) | 映像撮像装置 | |
JP2010263611A (ja) | 映像撮影装置 | |
JP4667411B2 (ja) | コンテンツ撮影装置 | |
JP5836091B2 (ja) | 再生装置及びプログラム | |
JP6332963B2 (ja) | 画像処理装置及び画像処理装置の制御方法 | |
WO2013186962A1 (ja) | 映像処理装置、撮影装置、およびプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 2011537781 Country of ref document: JP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11742048 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13578441 Country of ref document: US |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 11742048 Country of ref document: EP Kind code of ref document: A1 |