WO2007123008A1 - データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 - Google Patents
データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 Download PDFInfo
- Publication number
- WO2007123008A1 WO2007123008A1 PCT/JP2007/057760 JP2007057760W WO2007123008A1 WO 2007123008 A1 WO2007123008 A1 WO 2007123008A1 JP 2007057760 W JP2007057760 W JP 2007057760W WO 2007123008 A1 WO2007123008 A1 WO 2007123008A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- illumination
- data
- photographing
- scene
- video
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/4104—Peripherals receiving signals from specially adapted client devices
- H04N21/4131—Peripherals receiving signals from specially adapted client devices home appliance, e.g. lighting, air conditioning system, metering devices
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
- H04N21/4307—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
- H04N21/43074—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of additional data with content streams on the same device, e.g. of EPG data or interactive icon with a TV program
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/435—Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8126—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
- H04N21/8133—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05B—ELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
- H05B47/00—Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
- H05B47/10—Controlling the light source
- H05B47/155—Coordinated control of two or more light sources
Definitions
- Data transmission device data transmission method, viewing environment control device, viewing environment control system, and viewing environment control method
- the present invention can control illumination light around a video display device when the video is displayed on the video display device in accordance with the atmosphere and scene setting of the shooting scene of the video.
- the present invention relates to a data transmission device, a data transmission method, a viewing environment control device, a viewing environment control system, and a viewing environment control method.
- the surrounding illumination light is adjusted to the display image. It is known to have a technology that provides viewing effects such as enhancing the sense of reality by adjusting the
- a mixed light intensity ratio of three primary colors of a light source is calculated for each frame from a color signal (RG B) and a luminance signal (Y) of a display image of a color television,
- RG B color signal
- Y luminance signal
- a light color variable illumination device that performs dimming control in conjunction with an image.
- This variable light color illumination device extracts color signals (RGB) and luminance signals (Y) from display images on color televisions, and uses the color signals and luminance signals to generate three-color light (red light, green)
- the appropriate dimming illuminance ratio (light, blue light) is calculated, the illuminance of the three-color light is set according to the illuminance ratio, and the three-color light is mixed and output as illumination light.
- a television image is divided into a plurality of parts, and an average hue of the corresponding dividing unit is detected to control illumination around the dividing unit.
- a video effect lighting apparatus to perform is disclosed.
- This video effect lighting device includes illumination means for illuminating the surroundings of the installation location of the color television, divides the image displayed on the color television into a plurality of images, and displays an image corresponding to the portion illuminated by the illumination means. The average hue of the division part is detected, and the illumination means is controlled based on the detected hue.
- Japanese Patent Laid-Open No. 3-184203 simply discloses the entire screen of the image display device.
- the remaining part of the image displayed on the screen of the image display device without the average chromaticity and average brightness of the image is considered to be the background part, and the remaining part of the background part is considered as the background part.
- the average chromaticity and brightness of the back wall of the image display device are calculated by taking out only the RGB signal and luminance signal of the pixel, and the chromaticity and brightness of the back wall of the image display device are the average chromaticity of the whole screen or background excluding human skin color. And a method for controlling the lighting so as to be equal to the average luminance.
- a video scene is shot under a properly set lighting condition as a segmented video based on a series of scene settings, for example, by the intention of a video producer (screenwriter, director, etc.). Created. Therefore, it is desirable to irradiate the viewing space with illumination light according to the lighting conditions at the time of scene shooting of the displayed video in order to increase the presence and atmosphere when viewing the video.
- the feature amount (color signal and luminance signal) for each frame (screen) in the video signal to be displayed is detected and the illumination light is controlled. Therefore, it is difficult to generate illumination light according to the scene condition of the video depending on the content of the displayed video. For example, if an inappropriate color illumination light is irradiated to the surroundings due to the influence of the clothes of the subject person included in the video signal or artifacts in the background of the subject, the atmosphere of each scene is reproduced. Or the realism of each scene cannot be maintained. In other words, the lighting environment power used when shooting each scene greatly detracts from the viewing environment lighting, which impairs the sense of reality.
- the state of the illumination light changes according to the change of the luminance and hue of the video signal for each frame, and particularly when the degree of change of the luminance and hue between frames is large.
- the illumination light changes in a complicated manner, causing the problem that the viewer feels uncomfortable with the flicker.
- FIG. 1 is a diagram for explaining an example of a problem of illumination control according to the above conventional technique. That is, a part of a continuous moving image is shown.
- a scene of an image taken with the scene setting is created in the outdoors and on a clear day.
- This scene consists of images obtained by taking a series of camera work without switching the camera.
- an image is taken of a skier sliding down from the top of the camera toward the vicinity of the camera. Skiers are dressed in red and the sky is clear.
- this video scene is shot under sunlight (natural light) during sunny daytime in winter, and this sunlight is a key light.
- the illumination of this key light is characterized by a spotlight (point light source), the incident direction of the horizontal illumination light is forward light, the incident direction of the vertical illumination light is top light, the illumination intensity is strong, and the illumination color temperature is about The picture was taken at 6000K under the lighting condition.
- FIG. 2 is a diagram for explaining another example of the problem of the illumination control according to the conventional technique.
- a scene of a video shot with the scene setting of moonlight night is created.
- This scene consists of three shots (1, 2, 3) with different camerawork.
- shot 1 the camera takes a long shot of the target ghost.
- shot 2 the ghost was shot in an up shot.
- shot 3 the camera returns to the shot 1 camera position again.
- the illumination light source has a flood light (surface light source) property
- the horizontal illumination light incident direction is forward light
- the vertical illumination light The incident direction of the top light, the illumination intensity is normal
- the nature of the illumination light source is a flood light (surface light source)
- the horizontal light incident direction is a rembrandt light ( Left)
- weak illumination intensity and illumination color temperature of about 7000K illumination light source property spotlight (point light source)
- the incident light direction is side light (right)
- the vertical light incidence direction is forward light
- the illumination intensity is slightly weak
- the touch light (lower stage) is about 5000K.
- illumination light that does not take into account the lighting conditions at the time of shooting is generated and irradiated. This will disturb the viewer and make them uncomfortable. Also, if the illumination light becomes darker or brighter in a single segment of scenes where a single scene setting (atmosphere) is continuous, the atmosphere of that scene will still be disturbed and the viewer will feel uncomfortable.
- the present invention has been made in view of the above-described problems, and is a data transmission device and a data transmission method capable of realizing optimal viewing environment illumination control in accordance with illumination during shooting of a display image. It is an object of the present invention to provide a viewing environment control device, a viewing environment control system, and a viewing environment control method.
- the first invention of the present application is a data transmission device for transmitting video data comprising one or more scene forces.
- the video illumination data indicating the lighting status at the time of shooting each scene of the video data is used as the video. It is characterized by being added to data and transmitted.
- a second invention of the present application is characterized in that the photographing illumination data is attached in units of scenes of the video data.
- a third invention of the present application is characterized in that the photographing illumination data is attached in units of shots of the video data.
- a fourth invention of the present application is characterized in that the photographing illumination data is added in units of frames of the video data.
- the fifth invention of the present application is characterized in that the photographing illumination data includes at least information indicating a lighting type of the illumination used for photographing each scene.
- the sixth invention of the present application is characterized in that the photographic illumination data includes at least information indicating the nature of the illumination used for photographing each scene.
- the seventh invention of the present application is characterized in that the photographing illumination data includes at least information indicating the direction of the illumination light used for photographing each scene.
- the eighth invention of the present application is characterized in that the photographing illumination data includes at least information indicating the intensity of the illumination light used for photographing each scene.
- the ninth invention of the present application is characterized in that the photographic illumination data includes at least information representing the chromaticity of the illumination light used for photographing each scene.
- a tenth invention of the present application is a data transmission device that receives photographing requests and transmits photographing lighting data indicating a lighting state at the time of photographing each scene constituting the video data, and the photographing lighting is provided.
- the data is transmitted together with the start timing of each scene constituting the video data.
- the eleventh invention of the present application is characterized in that the photographing illumination data includes at least information indicating a lighting type of the illumination used for photographing each scene.
- the photographic illumination data includes at least information indicating the nature of the illumination used for photographing each scene.
- the thirteenth invention of the present application is characterized in that the photographing illumination data includes at least information indicating the direction of light of the illumination used for photographing each scene.
- the fourteenth invention of the present application is characterized in that the photographing illumination data includes at least information indicating the intensity of illumination light used for photographing each scene.
- the photographic illumination data includes at least information indicating the chromaticity of the illumination light used for photographing each scene.
- a sixteenth aspect of the present invention is a receiving means for receiving video data to be displayed on a display device, and shooting illumination data indicating a lighting state at the time of shooting each scene constituting the video data, Control means for controlling illumination light of an illumination device installed around the display device based on the photographing illumination data.
- the seventeenth invention of the present application is characterized in that the control means switches and controls the illumination light of the illumination device for each scene of the video data.
- the eighteenth invention of the present application is characterized in that the control means controls illumination light of the illumination device using a feature amount of the video data in addition to the photographing illumination data.
- the nineteenth invention of the present application is characterized in that the photographing illumination data includes at least information indicating a lighting type of illumination used for photographing each scene.
- the twentieth invention of the present application is characterized in that the photographing illumination data includes at least information indicating a property of illumination used for photographing each scene.
- the twenty-first invention of the present application is characterized in that the photographic illumination data includes at least information indicating a direction of illumination light used for photographing each scene.
- a twenty-second invention of the present application is characterized in that the photographing illumination data includes at least information indicating the intensity of illumination light used for photographing each scene.
- the twenty-third invention of the present application is characterized in that the photographic illumination data includes at least information indicating the chromaticity of the illumination light used for photographing each scene.
- a twenty-fourth invention of the present application is a viewing environment control system including the viewing environment control device and a lighting device whose viewing environment illumination light is controlled by the viewing environment control device.
- a twenty-fifth aspect of the present invention in a data transmission method for transmitting video data composed of one or more scenes, photographing illumination data indicating a lighting state at the time of photographing each scene of the video data is obtained.
- the video data is transmitted after being added.
- a twenty-sixth invention of the present application is a data transmission method for transmitting photographing illumination data indicating a lighting state at the time of photographing each scene constituting the video data in response to an external request.
- the photographing illumination data is transmitted together with the start timing of each scene constituting the video data.
- the twenty-seventh invention of the present application receives video data to be displayed on a display device, and shooting illumination data indicating a lighting state at the time of shooting of each scene constituting the video data, and the shooting illumination data Based on the above, the illumination light of the illumination device installed around the display device is controlled.
- FIG. 1 is a diagram for explaining an example of a problem of illumination variation according to the prior art.
- FIG. 2 is a diagram for explaining another example of the problem of illumination variation according to the prior art.
- FIG. 3 is a block diagram showing a schematic configuration of a main part of the video transmission device in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 4 is an explanatory diagram showing an example of an output bit stream of the video transmitting apparatus in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 5 is an explanatory diagram showing an example of photographing illumination data in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 6 is an explanatory diagram for explaining horizontal and vertical divisions of photographing illumination light.
- FIG. 7 is an explanatory diagram for explaining the components of a video.
- FIG. 8 is a block diagram showing a schematic configuration of a main part of the video reception device in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 9 is an explanatory diagram showing an example viewing environment space in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 10 is an explanatory diagram showing an example of control data in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 11 is an explanatory diagram showing a color reproduction range of illumination light in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 12 is an explanatory diagram showing an example of the relationship between the illumination light direction information and the corresponding illumination device ID in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 13 is an explanatory diagram showing photographing illumination data corresponding to the video scene shown in FIG.
- FIG. 14 is an explanatory diagram showing illumination control data generated based on the photographic illumination data shown in FIG. 11 in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 15 An explanatory diagram showing a viewing environment space realized based on the illumination control data shown in FIG. 14 in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 17 is an explanatory diagram showing illumination control data generated based on the photographing illumination data shown in FIG. 16 in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 18 is an explanatory diagram showing a viewing environment space realized based on the illumination control data shown in FIG. 17 in the viewing environment control system according to the first embodiment of the present invention.
- FIG. 19 is a block diagram showing a schematic configuration of a main part of a video reception device in a viewing environment control system according to a second embodiment of the present invention.
- FIG. 20 is a block diagram showing an illumination control data generation unit in FIG.
- FIG. 21 An explanatory diagram showing the color reproduction range of illumination light in the viewing environment control system according to the second embodiment of the present invention.
- FIG. 22 is a block diagram showing a schematic configuration of a main part of the external server device in the viewing environment control system according to the third embodiment of the present invention.
- FIG. 23 is an explanatory diagram showing an example of a shooting illumination data storage table in the viewing environment control system according to the third embodiment of the present invention.
- FIG. 24 is an explanatory diagram showing an example of a photographing illumination data storage table in the viewing environment control system according to the third embodiment of the present invention.
- FIG. 25 is an explanatory diagram showing an example of a photographic illumination data storage table in the viewing environment control system according to the third embodiment of the present invention.
- FIG. 26 is an explanatory diagram showing an example of a photographic illumination data storage table in the viewing environment control system according to the third embodiment of the present invention.
- FIG. 27 is a video receiver in the viewing environment control system according to the third embodiment of the present invention. It is a block diagram which shows the principal part schematic structure of these.
- FIG. 28 is a block diagram showing a schematic configuration of a main part of a video reception device in a viewing environment control system according to a fourth embodiment of the present invention.
- FIG. 29 is a block diagram showing a lighting control data generation unit in FIG. 28.
- FIG. 3 is a block diagram showing a schematic configuration of a main part of a video transmission device (data transmission device) for explaining the viewing environment control system according to the first embodiment of the present invention.
- the transmission device divides and multiplexes each of video data, audio data, and photographing illumination data supplied as additional information into a transport stream packet (TSP) format.
- TSP transport stream packet
- a data multiplexing unit 1 and a transmission unit 2 that performs modulation after adding an error correction code to the output data of the data multiplexing unit 1 and sends it to the transmission line as broadcast data are provided.
- Fig. 4 is an explanatory diagram showing an outline of the structure of Transport Stream Packet (TSP) defined by MPEG2 (Moving Picture Experts Group 2)-Systems.
- TSP Transport Stream Packet
- 11 is the data content of TSP and other MPEG2 —
- 12 is an extension header (advancement 'field), which can describe information determined by the sender
- 13 is a payload composed of data such as video and audio. It is.
- video data and audio data are transmitted in the payload 13
- photographing illumination data that is supplementary information can be transmitted in the extension header (advancement field) 12.
- different data streams of video data, audio data, and photographing illumination data may be multiplexed and transmitted.
- the photographing illumination data indicating the illumination state at the time of photographing each scene of the video data
- shooting illumination data lighting type information indicating the lighting type of the illumination used for shooting each scene
- lighting property information indicating the characteristics of the illumination light source used for shooting each scene
- each scene Illumination light direction (horizontal direction) information that represents the incident direction (horizontal direction) of the illumination light used for shooting
- illumination that represents the incident direction (vertical direction) of the illumination light used for each scene.
- each information is demonstrated.
- keylights, filllights, touch lights, and base lights are known as commonly used lighting types of photographing illumination.
- the key light is the main light source, and is the central light for setting the scene. For example, in the outdoor location during the daytime, it corresponds to ordinary sunlight.
- Fill light is an auxiliary light source that eliminates shadows and brightens or glosses the subject. For example, it reflects sunlight by a reflector at outdoor locations in the daytime. And this light is in the dark.
- the touch light is light for separating the subject from the background or giving a three-dimensional effect, and is effective light that is irradiated obliquely from behind to express the harshness.
- Base light is light that is used to brighten the whole image on average or relieve strong contrast.
- the lighting type information a 2-bit description indicating whether the illumination light at the time of shooting each video scene belongs to the key light / fill light (holding) / touch light / base light level or deviation. The description is included in the photographing illumination data.
- the properties of the illumination light source used for each lighting type are determined.
- Typical properties of illumination light sources during shooting include spotlights (point light sources) and floodlights (surface light sources).
- Spotlights (point light sources) are illuminations that have a clear direction of light, and that produce a clear shadow.
- flood light surface light source is illumination used to obtain uniform brightness over the entire space, and it is an illumination in which shadows are blurred and shading does not appear strongly.
- this illumination light source Normally, since the nature of the illumination light source is determined for each lighting type, this illumination light source The realism of the displayed video scene can be improved by controlling the viewing environment illumination, which will be described later, in accordance with the nature of the display. For example, when reproducing the illumination light from a spotlight (point light source), the illumination range from the viewing environment illumination device is narrowed. When reproducing the illumination light from a floodlight (surface light source), illumination from the viewing environment illumination device is performed. It is possible to widen the range.
- a spotlight point light source
- a floodlight surface light source
- the illumination property information whether the property of the illumination light source used at the time of shooting each video scene belongs to either a spotlight (point light source) or a flood light (surface light source).
- the 1-bit description shown is included in the photographic lighting data.
- the direction of the illumination light is determined for each lighting type, the realism of the display video scene is improved by controlling the viewing environment illumination described later in accordance with the direction of the illumination light. be able to. That is, by independently controlling a plurality of lighting devices arranged on the top, bottom, left, and right of the video display device, it is possible to reproduce the shooting lighting environment at the time of video shooting with great detail.
- ⁇ indicates a subject
- C indicates a camera.
- the horizontal division of the illumination light includes forward light, side light (left, right), backlight, rim light (left, right), rebrandt light ( Left, right).
- Forward light is light from the front that provides the most efficient brightness.
- Sidelight is a light that can enhance the three-dimensional effect, but has a strong shadow and gives a strong impression with a strong contrast.
- Backlight is light from the back that can create a sino-ret effect that raises the outline and produces a fantastic effect.
- Rimlight is a semi-backlit light that is used to enhance the sense of body and bring out the details of the hair.
- Rembrandt light is a 45-degree light that best expresses facial features and makes them look beautiful.
- the illumination light is divided into a normal light, a backlight, a backlight, a tilt, and a Rembrandt light.
- Forward light is light from the front that provides the most efficient brightness.
- Backlight is light from the back that can create a sino-ret effect that raises the outline and produces a fantastic effect.
- the top light is light from almost right above the subject used for characteristic image expression rather than showing people clearly.
- Aori is the light from the lower front of the subject that is used for special effects. is there.
- Rembrandt light is a 45-degree light that best expresses facial features and makes them look beautiful.
- the incident direction (horizontal direction) of the illumination light at the time of shooting each video scene is the following light / side light (left) / side Light (Right) / Backlight / Rim Light (Left) / Rim Light (Right) / Rembrandt Light (Left) Z Rembrandt Light (Right
- the illumination light direction (vertical direction) information indicates whether the incident direction (vertical direction) of the illumination light at the time of shooting each video scene belongs to the following light / backlight / top light / tilt / rembrandt light
- a 3-bit description is included in the photographic lighting data.
- the viewing environment illumination described later is controlled in accordance with the intensity and chromaticity of the illumination light, so that the display video scene is displayed.
- the presence can be improved.
- the illumination intensity information the intensity of the illumination light at the time of shooting each video scene is very weak / weak / somewhat weak / normal / somewhat strong, / strong, / very strong, A 3-bit description indicating whether the image belongs to the deviation is included in the photographic lighting data.
- the illumination color temperature information the color temperature (K) of the illumination light at the time of shooting each video scene is 1000/2000/3000/4000/5000/6000, and a 4-bit description indicating the power belonging to the deviation. Is included in the photographic lighting data.
- FIG. 7 is a diagram for explaining the components of the video, and will be described below with reference to FIG. 7 according to the configuration of the video including the scene yacht.
- the first layer (# 1) constituting the video is a frame.
- a frame is a physical layer and refers to a single two-dimensional image. Frames are usually obtained at a rate of 30 frames per second.
- the second layer (# 2) is a shot.
- a shot is a sequence of frames taken by a single camera.
- the third layer (# 3) is a scene.
- a scene is a sequence of shots that have a stream of connections.
- the above-described photographing illumination data can be added in units of frames, shots, or scenes of video data. If shooting illumination data is added at least in units of scenes, it will be possible to efficiently control the viewing environment lighting described later, but if shooting lighting data is added in units of frames, more detailed viewing environment lighting will be provided. Can be controlled. For example, shooting illumination data may be added only to a specific frame (scene switching frame, etc.) according to the intention of the video producer (screenwriter, director, etc.).
- the photographing illumination data may be added in GOP (Group of Picture) units, which are units of random access of video data.
- a video reception device (data reception device) that receives broadcast data transmitted from the video transmission device, displays and reproduces video's audio, and controls the viewing environment illumination at that time will be described. To do.
- FIG. 8 is a block diagram showing a schematic configuration of a main part of the video reception device in the viewing environment control system according to the first embodiment of the present invention.
- the video reception device in this embodiment is as shown in FIG.
- the broadcast data input from the transmission path is received and demodulated, and error correction is performed, and the output data of the receiver 21 and the output data of the receiver 21 are output to the video display device 25.
- the image data, the audio data output to the audio reproduction device 26, and the photographic illumination data as additional information are separated and extracted, and the photographic illumination data separated by the data separation unit 22 are received.
- the CPU 23 outputs the control data related to the lighting adapted to the scene setting (atmosphere) of each scene constituting the video data, and the lighting control data (RGB data) corresponding to the output control data from the CPU 23 is viewed.
- an illumination control data generation unit 24 that outputs to the illumination device 27 that illuminates the environment space.
- the lighting device 27 can be configured by an LED that is installed around the video display device 25 and emits light of, for example, three primary colors of RGB having a predetermined hue.
- the illumination device 27 is not limited to the combination of LEDs that emit a predetermined color as described above as long as the illumination color and brightness of the surrounding environment of the video display device 25 can be controlled. It may be constituted by an LED and a color filter, or a combination of a white light bulb, a fluorescent tube and a color filter, or a color lamp can be applied. Also, one or more lighting devices 27 should be installed.
- 62 (ID1 to ID62) illumination devices 27 that can be independently controlled are provided in the viewing environment space where the video display device 25 is installed. Shall. An ID is assigned to each lighting device.
- the lighting devices ID1 to ID8 illuminate the left side wall of the video display device 25 from above
- the lighting devices ID9 to ID18 are those of the video display device 25.
- the illumination devices ID19 to ID26 illuminate the right side wall of the image display device 25
- the illumination devices ID27 to ID34 illuminate the left side wall of the image display device 25 from below.
- ID35 to ID44 lighting device irradiates the rear wall of the image display device 25 from below
- ID45 to ID52 lighting device irradiates the right side wall of the image display device 25 from below
- ID53 to ID62 lighting device 62 illuminates the ceiling wall of the image display device 25, and 62 lighting devices are arranged.
- the CPU 23 reproduces the illumination environment at the time of photographing each scene with the illumination light intensity and ID for each illumination device that should be reproduced by the viewing environment illumination by the illumination device 27.
- the color temperature (one point on the black body locus) is determined, and this is output as control data to the illumination control data generator 24.
- the illumination intensity is based on various information included in the photographic illumination data.
- control data indicating the illumination color temperature in 16 steps is generated for each ID of each lighting device.
- the CPU 23 adds information on the lighting status for each lighting type, and generates control data for each ID of each lighting device that reproduces the lighting environment at the time of shooting.
- the black body is an ideal object that completely absorbs energy, and when the temperature rises, the color of the emitted light changes from red to yellow to white.
- the absolute temperature at this time is called the color temperature.
- this color temperature and color locus black body locus
- the temperature of the closest black body that does not match completely is called the “correlated color temperature”.
- the correlated color temperature is expressed along with the deviation ( ⁇ ⁇ ) from the blackbody locus.
- the illumination control data generation unit 24 generates RGB data corresponding to the control data (intensity, color temperature) from the CPU 23 and outputs the RGB data to the illumination device 27, as shown in (C) of FIG. .
- the RGB data power for each lighting device ID is output to the lighting device with the corresponding ID, realizing optimal lighting control in the viewing environment according to the lighting conditions at the time of shooting the display video.
- the ability to do S is possible.
- the CPU 23 has registered in advance information related to the installation position of each lighting device (relative position with respect to the display screen of the video display device). Using this position information and shooting illumination data, the illumination at the time of shooting is recorded. Generate control data for each lighting device that reproduces the environment.
- the illumination property information is a spotlight (point light source)
- a predetermined intensity and color temperature are obtained by N illumination devices centered on an illumination device with an ID located in the incident direction of the illumination light. Irradiate the illumination light.
- the illumination property information is a flood light (surface light source)
- M illumination devices including ID illumination devices located in the incident direction of the illumination light are used to obtain a predetermined intensity.
- illumination light of color temperature is obtained by N illumination devices centered on an illumination device with an ID located in the incident direction of the illumination light.
- FIG. 12 an example of the relationship between the incident direction of the photographic illumination light and the corresponding illumination device ID is as shown in FIG. 12, for example. That is, in (A) of FIG. 12, when the direction of the illumination light in the horizontal direction is the forward light and the direction of the illumination light in the vertical direction is other than the direction, the illumination light of ID9 to ID18 is used to reproduce this illumination light. Use the device.
- the direction of the illumination light in the horizontal direction is side light (left), and the direction of the illumination light in the vertical direction is other than the vertical direction, in order to reproduce this illumination light, the illumination device of ID1 to ID8, the horizontal direction The direction of the illumination light is side light (right), vertical direction If the direction of the illumination light is other than tilt, the illumination devices of ID19 to ID26 are used to reproduce this illumination light.
- the illumination devices of ID5 to ID22 are installed in the horizontal direction.
- Directional force of the illumination light S rim light (left) and when the direction of the illumination light in the vertical direction is other than the tilt, to reproduce this illumination light, the illumination device of ID9 ⁇ ID13, If the direction is the rim light (right) and the direction of the illumination light in the vertical direction is other than the vertical direction, the illumination device of ID14 to ID18 is used to reproduce this illumination light.
- the illumination light of ID5 to ID11 is used to reproduce this illumination light.
- the illumination device of ID16 to ID22 is used to reproduce this illumination light.
- the illumination devices of ID31 to ID48 are connected in the horizontal direction. If the direction of the illumination light is the rim light (left) and the direction of the illumination light in the vertical direction is tilted, in order to reproduce this illumination light, the illumination device of ID35 to ID39 is used, and the direction of the illumination light in the horizontal direction is the rim light. (Right) If the direction of illumination light in the vertical direction is tilted, use the illumination devices ID40 to ID44 to reproduce this illumination light.
- Fig. 13 shows the shooting illumination data corresponding to the video scene described with Fig. 1.
- This video scene has the characteristic power of the illumination light source S spotlight (point light source), the incident direction of the illumination light in the horizontal direction is the following light, The incident direction of the illumination light in the vertical direction is the top light, and the illumination intensity is high.
- S spotlight point light source
- the incident direction of the illumination light in the horizontal direction is the following light
- the incident direction of the illumination light in the vertical direction is the top light
- the illumination intensity is high.
- the CPU 23 illuminates the rear wall of the video display device 25 from above based on the photographing illumination data, and the illumination devices ID9 to ID18 that irradiate the ceiling wall of the video display device 25.
- Control data for controlling the illumination device 27 to emit illumination light with an illumination intensity of about 6000K (indicated by point (b) on the xy chromaticity diagram in Fig. 11). Is generated.
- the illumination control data generation unit 24 receives the control data from the CPU 23 and outputs illumination control data (RGB data) as shown in FIG. As a result, as shown in FIG. 15, light is emitted from ID9 to ID18 and ID53 to ID62 of the illumination device, and the back wall and ceiling of the image display device 25 are illuminated with appropriate illumination light, and the above-mentioned It is possible to reproduce the lighting environment in the viewing space when shooting the displayed video scene, and to increase the sense of reality when viewing the video.
- illumination control data RGB data
- FIG. 16 shows shooting illumination data corresponding to the video scene described with reference to FIG. 2.
- the illumination light source has a flood light (surface light source) property, and the incident direction of the horizontal illumination light. Is the front light, the incident direction of the vertical illumination light is the top light, the illumination intensity is normal, the key color with the illumination color temperature of about 7000K (upper), and the nature of the illumination light source is the flood light ( Surface light source), horizontal light incident direction is Rembrandt light (left), vertical light incident direction is low, illumination intensity is weak and illumination color temperature is about 7000K (middle stage), and illumination light source Spotlight (point light source), horizontal light incident direction is side light (right), vertical light incident direction is forward light, illumination intensity is slightly weaker, illumination color temperature is about 5000K (Lower part) shows that the picture was taken under lighting conditions.
- a flood light surface light source
- the CPU 23 illuminates the rear wall of the video display device 25 from above based on the photographing illumination data, and the illumination devices ID9 to ID18 that irradiate the ceiling wall of the video display device 25.
- the illumination intensity is normal and the illumination color temperature is about 7000K (indicated by point (c) on the xy chromaticity diagram in Fig. 11).
- the illumination device with ID19 to ID26 emits illumination light with a weak illumination intensity and an illumination color temperature of approximately 700,000K (shown as (c) on the xy chromaticity diagram in Fig. 11) and displays the image. Illuminate the left side wall of device 25 from below.
- ID3 Illumination device with ID37: Illumination color temperature is slightly weakened with illumination intensity of about 5000K (shown as point (a) on the xy chromaticity diagram in Fig. 11)
- the control data for controlling the illumination device 27 to generate the illumination light is generated.
- the illumination control data generation unit 24 receives the control data from the CPU 23 and outputs illumination control data (RGB data) as shown in FIG. As a result, as shown in FIG. 18, the ceiling, the back wall, and the left and right side walls of the image display device 25 are illuminated with appropriate illumination light to view the illumination environment at the time of shooting the display video scene described above. This makes it possible to reproduce in space and increase the sense of presence when viewing images.
- illumination control data RGB data
- the illumination control data is used to appropriately reproduce the lighting state (atmosphere) at the time of shooting each scene from various information contents input as shooting lighting data together with video data.
- the illumination light of the illumination device 27 can be controlled. Therefore, it is possible to realize a natural and uncomfortable viewing environment illumination that is not affected by the video content, and to increase the sense of presence when viewing the video.
- shooting illumination data relating to the lighting conditions during shooting of each scene is transmitted and received, so that viewing environment illumination control adapted to various viewing environment spaces can be performed.
- viewing environment illumination control adapted to various viewing environment spaces can be performed.
- the received shooting is performed.
- Appropriate control data corresponding to each video viewing space can be generated from the lighting data.
- the viewing environment lighting may be controlled using only information relating to the keylight. Good.
- the necessary information can be easily selected for shooting lighting data by sending and receiving the lighting status information for each of the keylight, filllight, touchlight, and baselight as a data structure with a hierarchical structure. / Can be processed.
- the viewing environment illumination may be controlled using video data and / or audio data in addition to the shooting illumination data. This will be described below as a second embodiment of the present invention.
- FIG. 19 is a block diagram showing a schematic configuration of the main part of the video reception device in the viewing environment control system of the present embodiment
- FIG. 20 is a block diagram showing the illumination control data generation unit in FIG. 19, and FIG. It is explanatory drawing showing the color reproduction range of the illumination light in the viewing-and-listening environment control system of embodiment.
- the video transmission device (data transmission device) in the present embodiment is the same as that of the first embodiment described with reference to FIG. 3, and the description thereof is omitted here.
- the video receiving device receives and demodulates broadcast data input from a transmission line, and receives a receiving unit 21 that performs error correction, and a receiving unit.
- Video data and TC time code
- Video data and TC time code
- Audio data and TC time code
- photographic lighting data as additional information
- Data separation unit 22 for separating and extracting data
- data separation The CPU 33 outputs the control data adapted to the lighting conditions at the time of shooting of each scene constituting the video data after receiving the shooting lighting data separated by the unit 22, and the lighting control data corresponding to the control data from the CPU 33 ( RGB data) is output to the lighting device 27 that illuminates the viewing environment space, and the video data, audio data, and control data are delayed and output by the processing time in the lighting control data generation unit 34 and the lighting control data generation unit 34.
- Delay generators 35, 36, and 37 are provided.
- the CPU 33 of the present embodiment obtains the intensity and color temperature (one point on the black body locus) of the illumination light for each illumination device ID based on the photographing illumination data, and a predetermined value including the color temperature is obtained.
- the information indicating the color range (function representing the range indicated by the alternate long and short dash line on the xy chromaticity diagram in FIG. 21) is determined and output as control data.
- the lighting control data generation unit 34 corrects the field (atmosphere) estimation result obtained based on the video data and audio data so as to be within the color range obtained based on the photographing illumination data, and performs illumination control.
- Outputs data RGB data).
- the time code (TC) is information added to indicate playback time information of each of the video data and audio data. For example, time (h): minute (m): second ( s): Consists of information indicating the frame (f).
- the illumination control data generation unit 34 of the present embodiment includes a scene segment detection unit 41 that detects the start point TC and end point TC of the scene segment based on the captured illumination data, and a scene Section (atmosphere) estimation unit 42 and field (atmosphere) estimation unit 42 that estimate lighting conditions and scene settings (atmosphere) at the shooting site based on video data and audio data for a predetermined time from the start point TC of the section.
- a lighting control unit 43 that outputs lighting control data for controlling the lighting device 27 from the estimation result and the control data output from the CPU 33 is provided.
- the method of estimating the ambient light state at the time of shooting by the field (atmosphere) estimation unit 42 various techniques including known ones can be used. Note that here, in addition to the feature values of the video data, the feature values of the audio data are also used to estimate the scene (atmosphere) of each scene. This is to improve the situation, so it is possible to estimate the scene (atmosphere) of the shooting scene from only the features of the video data.
- the feature amount of the video data for example, the location of the screen as in the above-described conventional example.
- the color signal and luminance signal in the fixed area can be used as they are, or the color temperature of the ambient light at the time of video recording can be obtained from these. Furthermore, these may be configured to be switchable and output as feature amounts of video data.
- the feature amount of audio data volume, audio frequency, etc. can be used.
- This field (atmosphere) estimation unit 42 uses the force S to estimate the color and brightness of ambient light at the time of shooting based on the feature values of video data and audio data, If the color of light is outside the predetermined color range obtained from the shooting lighting data, it is corrected by the lighting control unit 43 to obtain lighting control data suitable for the lighting conditions at the time of shooting each scene. It is possible.
- the result of estimating the ambient light color by the field (atmosphere) estimation unit 42 is point (A) in the xy chromaticity diagram of Fig. 21, and the illumination color temperature indicated by the control data output from the CPU 33 is
- the lighting control unit 43 determines that the straight line connecting the point (A) and the point (a) on the xy chromaticity diagram, and (a) Find the intersection ( ⁇ ') with the function that shows the color range centered on the point (the function that represents the ellipse shown by the alternate long and short dashed line on the xy chromaticity diagram in Fig. 21).
- the result of the ambient light color estimation by the field (atmosphere) estimation unit 42 is point (B) on the xy chromaticity diagram of Fig. 21, and the illumination color indicated by the control data output from the CPU 33
- the lighting control unit 43 has a straight line connecting the points (B) and (c) on the xy chromaticity diagram ( c) Find the intersection ( ⁇ ') with the function that indicates the color range centered on the point (the function that represents the ellipse indicated by the alternate long and short dash line on the xy chromaticity diagram in Fig. 21), and calculate this as the ambient illumination light
- the estimation result of the brightness of the ambient light by the field (atmosphere) estimation unit 42 is also appropriately corrected according to the illumination intensity indicated by the control data output from the CPU 33. If the ambient light estimation result by the field (atmosphere) estimation unit 42 is within the predetermined illumination range obtained from the captured illumination data, the ambient light estimation result by the field (atmosphere) estimation unit 42 is The illumination control data is output to the illumination device 27 as it is.
- the force for obtaining the illumination control data to be output to the illumination device 27 from the intersection with the function representing the color range indicated by the control data is not limited to this.
- the color difference ⁇ from the point in the predetermined color range where the distance to the estimation result by the field (atmosphere) estimation unit 42 is the minimum or the estimation result by the field (atmosphere) estimation unit 42 The illumination control data may be obtained from a point within a predetermined color range where the wrinkle is minimized.
- an average value or a weighted average value of the estimation result by the field (atmosphere) estimation unit 42 and the illumination color temperature and illumination intensity indicated by the control data output from the CPU 33 is output to the illumination device 27.
- the illumination control data may be used. Needless to say, the above-described illumination control data is generated for each ID of the illumination device.
- the configuration is such that the viewing environment illumination is controlled using the photographic illumination data, the video data, and the Z or audio data. Even for lighting environments that are difficult to perform, it is possible to estimate the lighting environment more accurately by using the estimation results of the scene (atmosphere) based on the feature values of video data and / or audio data. It is possible to perform more appropriate lighting control, and to prevent the presence or the atmosphere from being disturbed by an error in the estimation of the place (atmosphere) based on the feature amount of video data and / or audio data. It is possible to suppress the control and always realize the optimal viewing environment.
- the scene section detection unit 41 that detects the scene section based on the captured illumination data is provided, and the illumination control data is switched and controlled on a scene basis. It can prevent the viewing environment lighting from changing drastically in the scene and impairing the sense of reality.
- information indicating a predetermined color range including the color temperature of illumination light (representing a range indicated by a one-dot chain line in the xy chromaticity diagram of FIG. 21) based on the photographic illumination data. Function), but the size of the predetermined color range may be arbitrarily variable by the user. In other words, if the predetermined color range is set to a large value, the effect of reproducing the spread of the screen can be obtained.If the predetermined color range is set to a small value, the effect of reproducing the illumination environment at the time of shooting can be obtained.
- the size of the color range obtained from the photographic illumination data may be variably set depending on which lighting effect is emphasized by the user.
- the difference between the effect of reproducing the spread of the screen and the effect of reproducing the lighting environment at the time of shooting is desirable depending on the video content displayed on the video display device 25.
- the size of the predetermined color range obtained from the photographic illumination data may be variably set according to the determination result of the type (for example, genre) of the video content.
- shooting illumination data is multiplexed and added to broadcast data.
- shooting illumination data is added to broadcast data. If not, it is possible to realize an optimal viewing environment according to the lighting conditions at the time of video shooting by transmitting / receiving shooting lighting data corresponding to the video data to be displayed from an external server device or the like. This will be described below.
- FIG. 22 is a block diagram showing a schematic configuration of the main part of the external server device in the viewing environment control system of the present embodiment
- FIGS. 23 to 26 show the photographing illumination data storage table in the viewing environment control system of the present embodiment
- FIG. 27 is an explanatory diagram showing an example
- FIG. 27 is a block diagram showing a schematic configuration of a main part of a video reception device in the viewing environment control system of the present embodiment.
- the external server device (data transmission device) in this embodiment makes a transmission request for photographing illumination data related to specific video data (content) from the video reception device (data reception device) side.
- the receiving unit 51 that receives the data, the data storage unit 52 that stores the photographic lighting data for each video data (content), and the photographic lighting data that has received the transmission request is transmitted to the requesting video receiving device (data receiving device).
- a transmission unit 53 a transmission unit 53.
- the shooting illumination data stored in the data storage unit 52 of the present embodiment is associated with the scene number and scene start time code of each video scene, as shown in FIGS. It is described in a table format, and the shooting illumination data corresponding to the video data (program content) for which a transmission request has been received, along with the scene number and scene start TC (time code) of each scene constituting the video data. Transmit from the transmitter 53 to the requesting video receiver.
- the keyline as shown in FIG. Data table describing only the information about the key light, data table describing the information about the key light and fill light as shown in FIG. 24, and data describing the information about the key light, fill light and touch light as shown in FIG.
- the video receiving apparatus in the present embodiment receives and demodulates broadcast data input from the transmission path, and performs error correction, and from the output data of the receiving unit 61.
- the data separation unit 62 separates and extracts each of the video data output to the video display device 25 and the audio data output to the audio playback device 26, and a request for transmission of the photographic illumination data corresponding to the video data (content) to be displayed.
- the photographing illumination data received by the receiving unit 66 is stored, and the illumination for each illumination device ID obtained from the photographing illumination data of the video scene is synchronized with the display timing of each scene of the video data.
- the CPU 63 outputs control data related to intensity and color temperature, and the lighting control data generator 24 outputs the lighting control data (RGB data) corresponding to the output control data from the CPU 63 to the lighting device 27 that illuminates the viewing environment space.
- RGB data RGB data
- the CPU 63 receives the start time code of each scene in the photographic illumination data storage table received from the external server device and stored therein, and the time code of the video data displayed on the video display device 25. Compare them, and when they match, the corresponding lighting
- the control data suitable for the scene setting (atmosphere) of the displayed video scene can be output from this photographic lighting data.
- shooting light data corresponding to the display video data is obtained from the external server device, and this shooting light data is obtained.
- the lighting control data can be generated based on the video, and the display switching timing of the video scene and the switching timing of the viewing environment lighting can be synchronized with a simple configuration. It is possible to realize an optimal viewing environment.
- the viewing environment illumination may be controlled using video data and / or audio data in addition to the photographing illumination data. This will be described below as a fourth embodiment of the present invention.
- FIG. 28 is a block diagram showing a schematic configuration of the main part of the video reception device in the viewing environment control system of the present embodiment
- FIG. 29 is a block diagram showing the illumination control data generation unit in FIG.
- the external server device (data transmission device) in the present embodiment is the same as that of the third embodiment described above with reference to FIG.
- the video reception device receives and demodulates broadcast data input from the transmission line, and also receives the reception unit 61 and the output of the reception unit 61.
- the data separation unit 62 separates and extracts each of the video data output to the video display device 25 and the audio data output to the audio playback device 26 from the data, and the shooting illumination data corresponding to the video data (content) to be displayed.
- a transmission unit 65 that transmits a transmission request to an external server device (data transmission device) via a communication network, and a reception unit 66 that receives the photographing illumination data requested for transmission from the external server device via the communication network. I have.
- the photographing illumination data received by the receiving unit 66 is stored, and the illumination for each illumination device ID obtained from the photographing illumination data of the video scene is synchronized with the output timing of each scene of the video data.
- Lighting device that illuminates the viewing environment space with CPU73 that outputs control data related to intensity, color temperature, color range, and scene start point time code, and lighting control data (RGB data) corresponding to the output control data from CPU73 Illumination control data generation unit 74 that outputs to 27, and delay generation units 35, 36, and 37 that output video data, audio data, and control data with a delay by the processing time in illumination control data generation unit 74.
- the CPU 73 of the present embodiment obtains the intensity and color temperature (one point on the black body locus) of the illumination light for each illumination device ID based on the photographing illumination data, and a predetermined value including the color temperature is obtained.
- the information indicating the color range (function representing the range indicated by the alternate long and short dash line on the xy chromaticity diagram in FIG. 21) is determined and output as control data.
- the lighting control data generation unit 74 corrects the field (atmosphere) estimation result obtained based on the video data and audio data so that it falls within the color range obtained based on the shooting illumination data, and performs illumination control.
- Outputs data RGB data).
- the illumination control data generation unit 74 of the present embodiment is based on video data and audio data for a predetermined time from the scene start point TC from the CPU 73.
- Lighting control data for controlling the lighting device 27 is obtained from the field (atmosphere) estimation unit 82 for estimating the state and scene setting (atmosphere), the estimation result by the field (atmosphere) estimation unit 82 and the control data output from the CPU 73.
- an illumination control unit 43 for outputting.
- This field (atmosphere) estimation unit 82 uses force S to estimate the color and brightness of ambient light at the time of shooting based on the feature values of video data and audio data, If the color of light is outside the predetermined color range obtained from the shooting lighting data, it is corrected by the lighting control unit 43 to obtain lighting control data suitable for the lighting conditions at the time of shooting each scene. It is possible.
- the shooting illumination data corresponding to the display video data is stored in the external server device. Obtained from this shooting lighting data and video data and Z or sound Because it is configured to control viewing environment lighting using voice data, lighting that takes into account the estimation results of the place (atmosphere) based on the feature values of video data and / or audio data in addition to shooting lighting data It is possible to perform control, and lighting control that impairs the sense of presence and atmosphere due to errors in estimation of the place (atmosphere) based on the feature amount of video data and / or audio data It is possible to achieve an optimal viewing environment at all times.
- the lighting control data is switched and controlled on a scene-by-scene basis based on the scene start time code in the photographic lighting data storage table. This can prevent the viewing environment lighting from changing drastically and impairing the sense of reality.
- information indicating a predetermined color range including the color temperature of the illumination light (representing the range indicated by the alternate long and short dash line in the xy chromaticity diagram of FIG. 21) based on the photographing illumination data.
- the power that determines the function The size of this predetermined color range may be arbitrarily variable by the user. In other words, if the predetermined color range is set to a large value, the effect of reproducing the spread of the screen can be obtained. If the predetermined color range is set to a small value, the effect of reproducing the illumination environment at the time of shooting can be obtained.
- the size of the color range obtained from the photographic illumination data may be variably set depending on which lighting effect is emphasized by the user.
- the difference between the effect of reproducing the spread of the screen and the effect of reproducing the lighting environment at the time of shooting is desirable depending on the video content displayed on the video display device 25.
- the size of the predetermined color range obtained from the photographic illumination data may be variably set according to the determination result of the type (for example, genre) of the video content.
- the viewing environment control apparatus, method, and viewing environment control system of the present invention can be realized by various embodiments without departing from the gist of the present invention described above.
- the viewing environment control device may be configured to control an external lighting device based on various information included in input video data that may be provided in the video display device. Les.
- the above-described shooting illumination data is not limited to the case where it is separated from broadcast data or acquired from an external server device.
- an external device a DVD player or a blue
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Automation & Control Theory (AREA)
- Circuit Arrangement For Electric Light Sources In General (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP07741196A EP2018062A4 (en) | 2006-04-21 | 2007-04-06 | DATA TRANSMISSION METHOD AND DEVICE, AND AUDIOVISUAL ENVIRONMENT MANAGEMENT DEVICE, SYSTEM AND METHOD |
US12/297,552 US20090109340A1 (en) | 2006-04-21 | 2007-04-06 | Data Transmission Device, Data Transmission Method, Audio-Visual Environment Control Device, Audio-Visual Environment Control System, And Audio-Visual Environment Control Method |
JP2008512063A JP4950988B2 (ja) | 2006-04-21 | 2007-04-06 | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006117346 | 2006-04-21 | ||
JP2006-117346 | 2006-04-21 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2007123008A1 true WO2007123008A1 (ja) | 2007-11-01 |
Family
ID=38624914
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2007/057760 WO2007123008A1 (ja) | 2006-04-21 | 2007-04-06 | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20090109340A1 (ja) |
EP (1) | EP2018062A4 (ja) |
JP (1) | JP4950988B2 (ja) |
CN (1) | CN101427578A (ja) |
WO (1) | WO2007123008A1 (ja) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009253846A (ja) * | 2008-04-09 | 2009-10-29 | Canon Inc | 放送受信装置、放送受信方法 |
WO2010087155A1 (ja) * | 2009-01-27 | 2010-08-05 | シャープ株式会社 | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御方法および視聴環境制御システム |
WO2010087153A1 (ja) * | 2009-01-27 | 2010-08-05 | シャープ株式会社 | 感覚効果情報のデータ構造、感覚効果情報出力装置、感覚効果情報出力方法、感覚効果情報生成装置、制御装置、システム、プログラムおよび記録媒体 |
EP2315441A1 (en) * | 2008-07-15 | 2011-04-27 | Sharp Kabushiki Kaisha | Data transmission device, data reception device, method for transmitting data, method for receiving data, and method for controlling audio-visual environment |
JP2012511837A (ja) * | 2008-07-14 | 2012-05-24 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート | 感覚再生装置に係るメタデータを用いるマルチメディアアプリケーションシステム及びその方法 |
JP2012524442A (ja) * | 2009-04-15 | 2012-10-11 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | ユーザ環境を適応させるための方法及びシステム |
JP5092015B2 (ja) * | 2008-07-15 | 2012-12-05 | シャープ株式会社 | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 |
JP2015171038A (ja) * | 2014-03-07 | 2015-09-28 | キヤノン株式会社 | 画像処理装置およびその方法 |
JPWO2021250861A1 (ja) * | 2020-06-11 | 2021-12-16 | ||
JP2022500837A (ja) * | 2018-11-01 | 2022-01-04 | シグニファイ ホールディング ビー ヴィSignify Holding B.V. | ビデオ及びオーディオの重みに依存したビデオ及びオーディオ情報に基づく光効果の決定 |
CN118088963A (zh) * | 2024-03-07 | 2024-05-28 | 广东艾罗智能光电股份有限公司 | 一种可自动追光的智能照明控制方法及装置 |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101548551B (zh) | 2006-12-08 | 2011-08-31 | 皇家飞利浦电子股份有限公司 | 环境照明 |
JP4948548B2 (ja) * | 2006-12-28 | 2012-06-06 | シャープ株式会社 | 送信装置、視聴環境制御装置、及び視聴環境制御システム |
WO2011073877A1 (en) * | 2009-12-17 | 2011-06-23 | Koninklijke Philips Electronics N.V. | Ambience cinema lighting system |
WO2012001566A1 (en) * | 2010-06-30 | 2012-01-05 | Koninklijke Philips Electronics N.V. | Methods and apparatus for capturing ambience |
US20140104497A1 (en) * | 2012-10-17 | 2014-04-17 | Adam Li | Video files including ambient light effects |
US8928811B2 (en) * | 2012-10-17 | 2015-01-06 | Sony Corporation | Methods and systems for generating ambient light effects based on video content |
US8928812B2 (en) | 2012-10-17 | 2015-01-06 | Sony Corporation | Ambient light effects based on video via home automation |
JP6334552B2 (ja) * | 2012-11-27 | 2018-05-30 | フィリップス ライティング ホールディング ビー ヴィ | ステージパフォーマンスから導かれるデータに基づきアンビエントライティング効果を生成する方法 |
EP3062519A1 (en) * | 2015-02-27 | 2016-08-31 | Novabase Digital TV Technologies GmbH | Ambient surround information system for a media presentation |
US10368105B2 (en) | 2015-06-09 | 2019-07-30 | Microsoft Technology Licensing, Llc | Metadata describing nominal lighting conditions of a reference viewing environment for video playback |
US10453173B2 (en) * | 2015-07-24 | 2019-10-22 | Robert Bosch Gmbh | Panel transform |
EP3331325A1 (en) | 2016-11-30 | 2018-06-06 | Thomson Licensing | Method and apparatus for creating, distributing and dynamically reproducing room illumination effects |
US10568175B2 (en) * | 2017-05-26 | 2020-02-18 | Cooler Master Technology Inc. | Light control system based on display device and light control method thereof |
CN108184286A (zh) * | 2017-12-27 | 2018-06-19 | 深圳迈睿智能科技有限公司 | 灯具的控制方法和控制系统以及电子设备 |
FR3111724B1 (fr) * | 2020-06-18 | 2022-11-04 | Cgr Cinemas | Méthodes pour la production d’effets d’immersion visuelle pour un contenu audiovisuel |
CN113050915B (zh) * | 2021-03-31 | 2023-12-26 | 联想(北京)有限公司 | 一种电子设备及处理方法 |
US11553577B2 (en) * | 2021-04-30 | 2023-01-10 | Shenzhen Linklite Smart Lighting Co., Ltd | System and method for achieving synchronized audio and image control of lighting |
CN113347754B (zh) * | 2021-08-05 | 2021-11-02 | 杭州米芯微电子有限公司 | 一种用户可编程的led灯光控制方法及系统 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02158094A (ja) | 1988-12-12 | 1990-06-18 | Matsushita Electric Ind Co Ltd | 光色可変形照明装置 |
JPH02253503A (ja) | 1989-03-28 | 1990-10-12 | Matsushita Electric Works Ltd | 映像演出照明装置 |
JPH03184203A (ja) | 1989-12-14 | 1991-08-12 | Matsushita Electric Ind Co Ltd | 光色可変形照明装置 |
JP2005229153A (ja) * | 2004-02-10 | 2005-08-25 | Sony Corp | 調光システムおよび調光方法、配信装置および配信方法、受信装置および受信方法、記録装置および記録方法、ならびに、再生装置および再生方法 |
JP2005531909A (ja) * | 2002-07-04 | 2005-10-20 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | 周辺光および照明ユニットを制御する方法およびシステム |
JP2005310630A (ja) * | 2004-04-23 | 2005-11-04 | Sharp Corp | 高臨場感再生装置及び方法 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1999053728A1 (fr) * | 1998-04-13 | 1999-10-21 | Matsushita Electric Industrial Co., Ltd. | Procede de regulation d'un eclairage et dispositif d'eclairage |
US20050174473A1 (en) * | 1999-11-18 | 2005-08-11 | Color Kinetics, Inc. | Photography methods and systems |
WO2002068777A1 (en) * | 2001-02-26 | 2002-09-06 | Worldstage, Llc | Integrated, adaptable theater, club and multiplex |
GB0111431D0 (en) * | 2001-05-11 | 2001-07-04 | Koninkl Philips Electronics Nv | A real-world representation system and language |
EP1499406A1 (en) * | 2002-04-22 | 2005-01-26 | Intellocity USA, Inc. | Method and apparatus for data receiver and controller |
MXPA05000559A (es) * | 2002-07-15 | 2005-04-19 | Nokia Corp | Metodo para ocultamiento de errores en secuencias de video. |
US8243004B2 (en) * | 2003-03-10 | 2012-08-14 | Fergason Patent Properties, Llc | Apparatus and method for preparing, storing, transmitting and displaying images |
CN100596208C (zh) * | 2004-01-06 | 2010-03-24 | 皇家飞利浦电子股份有限公司 | 环境光脚本命令编码和动态控制环境光源的方法 |
-
2007
- 2007-04-06 WO PCT/JP2007/057760 patent/WO2007123008A1/ja active Search and Examination
- 2007-04-06 CN CNA2007800142640A patent/CN101427578A/zh active Pending
- 2007-04-06 EP EP07741196A patent/EP2018062A4/en not_active Ceased
- 2007-04-06 US US12/297,552 patent/US20090109340A1/en not_active Abandoned
- 2007-04-06 JP JP2008512063A patent/JP4950988B2/ja not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02158094A (ja) | 1988-12-12 | 1990-06-18 | Matsushita Electric Ind Co Ltd | 光色可変形照明装置 |
JPH02253503A (ja) | 1989-03-28 | 1990-10-12 | Matsushita Electric Works Ltd | 映像演出照明装置 |
JPH03184203A (ja) | 1989-12-14 | 1991-08-12 | Matsushita Electric Ind Co Ltd | 光色可変形照明装置 |
JP2005531909A (ja) * | 2002-07-04 | 2005-10-20 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | 周辺光および照明ユニットを制御する方法およびシステム |
JP2005229153A (ja) * | 2004-02-10 | 2005-08-25 | Sony Corp | 調光システムおよび調光方法、配信装置および配信方法、受信装置および受信方法、記録装置および記録方法、ならびに、再生装置および再生方法 |
JP2005310630A (ja) * | 2004-04-23 | 2005-11-04 | Sharp Corp | 高臨場感再生装置及び方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2018062A4 |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009253846A (ja) * | 2008-04-09 | 2009-10-29 | Canon Inc | 放送受信装置、放送受信方法 |
JP2012511837A (ja) * | 2008-07-14 | 2012-05-24 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート | 感覚再生装置に係るメタデータを用いるマルチメディアアプリケーションシステム及びその方法 |
EP2315441A4 (en) * | 2008-07-15 | 2014-07-16 | Sharp Kk | DATA TRANSMISSION DEVICE, DATA RECEIVING DEVICE, DATA TRANSMISSION METHOD, DATA RECEIVING METHOD AND METHOD FOR CONTROLLING THE AUDIOVISUAL ENVIRONMENT |
EP2315441A1 (en) * | 2008-07-15 | 2011-04-27 | Sharp Kabushiki Kaisha | Data transmission device, data reception device, method for transmitting data, method for receiving data, and method for controlling audio-visual environment |
JP5092015B2 (ja) * | 2008-07-15 | 2012-12-05 | シャープ株式会社 | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 |
JP5442643B2 (ja) * | 2009-01-27 | 2014-03-12 | シャープ株式会社 | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御方法および視聴環境制御システム |
CN102292725A (zh) * | 2009-01-27 | 2011-12-21 | 夏普株式会社 | 感官效果信息的数据结构、感官效果信息输出装置、感官效果信息输出方法、感官效果信息生成装置、控制装置、系统、程序以及记录介质 |
WO2010087153A1 (ja) * | 2009-01-27 | 2010-08-05 | シャープ株式会社 | 感覚効果情報のデータ構造、感覚効果情報出力装置、感覚効果情報出力方法、感覚効果情報生成装置、制御装置、システム、プログラムおよび記録媒体 |
WO2010087155A1 (ja) * | 2009-01-27 | 2010-08-05 | シャープ株式会社 | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御方法および視聴環境制御システム |
JP2012524442A (ja) * | 2009-04-15 | 2012-10-11 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | ユーザ環境を適応させるための方法及びシステム |
JP2015171038A (ja) * | 2014-03-07 | 2015-09-28 | キヤノン株式会社 | 画像処理装置およびその方法 |
JP2022500837A (ja) * | 2018-11-01 | 2022-01-04 | シグニファイ ホールディング ビー ヴィSignify Holding B.V. | ビデオ及びオーディオの重みに依存したビデオ及びオーディオ情報に基づく光効果の決定 |
JP7080399B2 (ja) | 2018-11-01 | 2022-06-03 | シグニファイ ホールディング ビー ヴィ | ビデオ及びオーディオの重みに依存したビデオ及びオーディオ情報に基づく光効果の決定 |
US11510300B2 (en) | 2018-11-01 | 2022-11-22 | Signify Holding B.V. | Determinning light effects based on video and audio information in dependence on video and audio weights |
JPWO2021250861A1 (ja) * | 2020-06-11 | 2021-12-16 | ||
WO2021250861A1 (ja) * | 2020-06-11 | 2021-12-16 | 日本電信電話株式会社 | 同期制御装置、同期制御方法、および、同期制御プログラム |
JP7406169B2 (ja) | 2020-06-11 | 2023-12-27 | 日本電信電話株式会社 | 同期制御装置、同期制御方法、および、同期制御プログラム |
CN118088963A (zh) * | 2024-03-07 | 2024-05-28 | 广东艾罗智能光电股份有限公司 | 一种可自动追光的智能照明控制方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
JP4950988B2 (ja) | 2012-06-13 |
CN101427578A (zh) | 2009-05-06 |
JPWO2007123008A1 (ja) | 2009-09-03 |
EP2018062A1 (en) | 2009-01-21 |
EP2018062A4 (en) | 2010-08-04 |
US20090109340A1 (en) | 2009-04-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4950988B2 (ja) | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 | |
JP5058157B2 (ja) | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 | |
JP4950990B2 (ja) | 映像送信装置及び方法、視聴環境制御装置及び方法 | |
US10735668B2 (en) | Synchronized lighting and video active lighting tracks (VALT) with synchronized camera to enable multi-scheme motion picture capture | |
JP4606420B2 (ja) | 補助ビジュアル表示システム | |
WO2007052395A1 (ja) | 視聴環境制御システム | |
US7180529B2 (en) | Immersive image viewing system and method | |
JP5442643B2 (ja) | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御方法および視聴環境制御システム | |
EP1379082B1 (en) | Display apparatus | |
US20030016236A1 (en) | Immersive augmentation for display systems | |
WO2007119277A1 (ja) | 視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 | |
JP2009081822A (ja) | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 | |
JP5074864B2 (ja) | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 | |
JP5166794B2 (ja) | 視聴環境制御装置および視聴環境制御方法 | |
JP4789592B2 (ja) | 視聴環境制御装置及び視聴環境制御方法 | |
JP2009081482A (ja) | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 | |
JP2009060541A (ja) | データ送信装置、データ送信方法、視聴環境制御装置、及び視聴環境制御方法 | |
US11393372B2 (en) | Video lighting apparatus with full spectrum white color | |
JP4709897B2 (ja) | 視聴環境制御システム、視聴環境制御装置、視聴環境照明制御システム、及び視聴環境制御方法 | |
Jaksic et al. | Analysis of the Effects of Front and Back Lights in Chroma Key Effects During Implementation in Virtual TV Studio | |
JP2009060542A (ja) | データ送信装置、データ送信方法、視聴環境制御装置、視聴環境制御システム、及び視聴環境制御方法 | |
WO2012029011A2 (en) | System and method for controlling ambient lighting |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07741196 Country of ref document: EP Kind code of ref document: A1 |
|
DPE1 | Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2008512063 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12297552 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200780014264.0 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007741196 Country of ref document: EP |
|
DPE1 | Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101) |