WO2012111320A1 - 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム - Google Patents
映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム Download PDFInfo
- Publication number
- WO2012111320A1 WO2012111320A1 PCT/JP2012/000974 JP2012000974W WO2012111320A1 WO 2012111320 A1 WO2012111320 A1 WO 2012111320A1 JP 2012000974 W JP2012000974 W JP 2012000974W WO 2012111320 A1 WO2012111320 A1 WO 2012111320A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- stream
- picture
- video stream
- view
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
Definitions
- the present invention relates to video encoding and decoding technology, and more particularly to maintaining playback compatibility.
- the present invention provides a video encoding apparatus capable of encoding high-definition video so as to suppress an increase in the amount of required data while maintaining playback compatibility with a playback apparatus based on the MPEG-2 standard, and video playback
- An object is to provide an apparatus.
- the present invention provides a first encoding means for generating a stream in MPEG-2 format by compressing and encoding a first video obtained by converting an original video to a first quality, MPEG- 4. a second encoding means for generating a base-view video stream and a dependent-view video stream in conformity with the MVC format; and a sending means for transmitting the streams generated by the first and second encoding means.
- the second encoding means includes a stream composed of pseudo data having the same number of pictures as the stream obtained by compressing and encoding the first video as the base-view video stream, and having a reduced total data amount.
- the original video is changed to a second quality higher than the first quality.
- Each picture of the second video image and generates the base view compressed coded stream pictures in the MPEG-2 format stream picture and the same time corresponding to the picture as a reference picture in the video stream.
- the video encoding apparatus has the above-described configuration, thereby maintaining the reproduction compatibility of the original video of the first quality by the reproduction apparatus based on the MPEG-2 standard, and having a data amount required than before. It is possible to suppress the increase and compress and encode the original video of the second quality higher than the first quality.
- FIG. 5 is a diagram showing video attributes to be matched and field names for indicating the video attributes when a 2D compatible video stream is an MPEG-2 video and a multi-view video stream is an MPEG-4 MVC codec.
- FIG. 1 is a diagram illustrating a data creation device according to Embodiment 1.
- FIG. 3 is a diagram showing a data creation flow of the data creation device according to the first embodiment.
- FIG. 1 is a diagram illustrating a configuration of a playback device that plays back 3D video according to Embodiment 1.
- FIG. It is a figure explaining a video decoder and a multi view video decoder.
- 6 is a diagram illustrating a flow of decoding processing and output processing relating to 3D video of the playback device according to Embodiment 1.
- FIG. 6 is a diagram for explaining management of an inter-view reference buffer according to the 3D video playback device according to Embodiment 1.
- FIG. 10 is a diagram for explaining a modification of management of an inter-view reference buffer according to the 3D video playback device according to Embodiment 1.
- FIG. 6 is a diagram for explaining a buffer sharing method according to the 3D video playback device according to Embodiment 1.
- FIG. 10 is a diagram for explaining a modification example of the video output portion according to the 3D video playback device according to the first embodiment. 10 is a modification of the method for assigning PTS and DTS related to the 3D video transport stream according to the first embodiment.
- 6 is a diagram for explaining the relationship between the structure of a transport stream and PMT packets according to Embodiment 1.
- FIG. It is a figure which shows the structure of 3D information descriptor. It is a figure explaining the reproduction
- FIG. 10 is a diagram for explaining a problem in realizing high image quality of a video according to the second embodiment. It is a figure explaining the data structure of the transport stream which concerns on Embodiment 2.
- FIG. 6 is a diagram illustrating a usage pattern according to Embodiment 2.
- FIG. 10 is a diagram illustrating a relationship between a configuration of a transport stream and PMT packets according to Embodiment 2. It is a figure which shows the structure of the image quality improvement information descriptor. It is a figure explaining a high quality reproduction system. It is a figure which shows an image quality improvement stream descriptor.
- FIG. 10 is a diagram for explaining a problem in realizing high image quality of a video according to the second embodiment. It is a figure explaining the data structure of the transport stream which concerns on Embodiment 2.
- FIG. 6 is a diagram illustrating a usage pattern according to Embodiment 2.
- FIG. 10 is a diagram illustrating a relationship between a configuration of a transport stream and PMT packet
- FIG. 6 is a diagram illustrating a data creation device according to a second embodiment.
- FIG. 10 is a diagram illustrating a data creation flow of the data creation device according to the second embodiment.
- 7 is a diagram showing a high-quality video playback device according to Embodiment 2.
- FIG. 10 is a diagram illustrating a flow of decoding processing and output processing related to high-quality video performed by the high-quality video playback device according to the second embodiment. It is a figure which shows the relationship between a high quality reproduction
- FIG. 10 is a diagram for explaining a problem in realizing a high gradation image according to the third embodiment. It is a figure explaining the production method of the extended video stream which implement
- FIG. 10 is a diagram illustrating a relationship between a configuration of a transport stream and a PMT packet according to Embodiment 3. 10 is a diagram for explaining a usage pattern according to Embodiment 3.
- FIG. FIG. 10 is a diagram illustrating a data creation device according to a third embodiment.
- FIG. 10 is a diagram illustrating a data creation flow of the data creation device according to the third embodiment.
- 6 is a diagram showing a high-gradation video reproducing apparatus according to Embodiment 3.
- FIG. 10 is a diagram for explaining a flow of decoding processing and output processing relating to high gradation video of the high gradation video reproducing device according to the third embodiment.
- FIG. 10 is a diagram for explaining a switching method according to a high gradation reproduction method in the reproduction apparatus for high gradation video according to the third embodiment. It is a figure which shows the relationship between a high gradation reproduction system, the reference switch between codecs, and a plane selector. It is a figure explaining the subject in implement
- FIG. 10 It is a figure explaining the structure of a scalable encoding system. It is a figure explaining the data structure of the transport stream which concerns on Embodiment 4.
- FIG. 10 is a diagram illustrating a relationship between a configuration of a transport stream and a PMT packet according to a fourth embodiment. It is a figure explaining the scaling method. It is a figure explaining the usage pattern which concerns on Embodiment 4.
- FIG. 10 illustrates a data creation device according to a fourth embodiment.
- FIG. 10 is a diagram illustrating a data creation flow of the data creation device according to the fourth embodiment.
- FIG. 10 is a diagram showing a high-resolution video playback apparatus according to Embodiment 4; FIG.
- FIG. 10 is a diagram for explaining a flow of decoding processing and output processing related to high-resolution video in a high-resolution video playback device according to Embodiment 4;
- FIG. 10 is a diagram for explaining a data structure according to a modification of the fourth embodiment.
- FIG. 10 is a diagram illustrating a configuration of a playback device according to a modification of the fourth embodiment.
- FIG. 10 is a diagram for explaining a data structure according to a modification of the fourth embodiment.
- FIG. 10 is a diagram illustrating a configuration of a playback device according to a modification of the fourth embodiment. It is a figure explaining the structure of a reproduction
- FIG. 10 is a diagram for explaining a data structure according to a modification of the first embodiment.
- FIG. 11 is a diagram showing a configuration of a playback device according to a modification of the first embodiment. It is a figure which shows the outline
- 10 is a diagram for explaining a data structure according to a modification of the third embodiment.
- FIG. 10 is a diagram for explaining a data structure according to a modification of the fourth embodiment.
- FIG. 10 is a diagram for explaining a data structure according to a modification of the first embodiment.
- FIG. 10 is a diagram for explaining a data structure according to a modification of the first embodiment.
- FIG. 10 is a diagram for explaining a data structure according to a modification of the first embodiment.
- FIG. 10 is a diagram for explaining a data structure according to a modification of the first embodiment. It is a figure for demonstrating the reference between views which reduces the memory amount of the reference picture which concerns on high resolution. It is a figure for demonstrating a scaling reference filter. It is a figure which shows the structure of the data creation apparatus which concerns on the modification of Embodiment 1.
- FIG. 6 is a diagram showing a configuration of a 3D video playback device according to a modification of the first embodiment.
- FIG. FIG. 10 is a diagram for explaining a data structure according to a modification of the first embodiment. It is a figure which shows the structure of the data creation apparatus which concerns on the modification of Embodiment 1.
- FIG. 11 is a diagram showing a configuration of a playback device according to a modification of the first embodiment. It is a figure which shows the structure of the data creation apparatus which concerns on the modification of Embodiment 1. FIG. It is a figure which shows the structure of the data creation apparatus which concerns on the modification of Embodiment 1.
- FIG. 11 is a diagram showing a configuration of a playback device according to a modification of the first embodiment.
- FIG. 10 is a diagram for explaining a data structure according to a modification of the first embodiment. 10 is a diagram for explaining a data structure for realizing inter-view reference between different multi-view video streams according to a modification of Embodiment 1.
- the 2D playback portion of the playback device decodes and plays back the MPEG-2 format stream using the existing decoding method, and the 3D playback portion includes the MPEG-4 MVC compliant base-view video stream and the video stream.
- the pent-view video stream is decoded and reproduced in a decoding format corresponding to the new encoding.
- FIG. 21 shows the data structure of the transport stream created by the broadcasting system in the present embodiment.
- the transport stream is composed of a 2D compatible video stream A and a multi-view video stream B.
- the latter multi-view video stream B is composed of a base-view video stream B1 and a dependent-view video stream B2.
- the 2D compatible video stream A is created by compressing and coding the left-eye image
- the base-view video stream B1 is created by compressing and coding a monochrome image such as black (hereinafter referred to as “black image”).
- the dependent-view video stream B2 is generated by compressing and encoding the difference between the left-eye image and the right-eye image.
- the base-view video stream B1 is a compression-encoded black image, it cannot be used as a reference image when generating the dependent-view video stream B2. This point is different from the existing MPEG-4 MVC format, and the reference image is set to a frame image at the same time of the 2D compatible video stream A.
- both 2D video and 3D video can be transmitted, and the base view video stream B1 compresses and encodes the black image, so the bit rate is extremely low. As a result, both 2D video and 3D video can be transmitted in the range of the existing allocated frequency band.
- the dependent-view video stream is decoded with reference to the frame image of the base-view video stream.
- the MPEG-2 compatible stream is decoded. That is, the dependent-view video stream can be decoded using the frame image of the left-eye image as a reference image.
- a descriptor for instructing the playback side to change the reference destination at the time of decoding from the base view video stream to the MPEG-2 compatible video stream is specified. ing.
- FIG. 26 is a block diagram of a functional configuration of the data creation device 2601 according to the present embodiment.
- the data creation device 2601 receives a left-eye image, a right-eye image, and a black image that constitute a 3D video as inputs, and includes a transformer including a 2D-compatible video stream, a base-view video stream, and a dependent-view video stream in a data format described later. Output the port stream.
- the data creation device 2601 includes a 2D compatible video encoder 2602, a Dec (2D compatible video decoder) 2603, an extended multi-view video encoder 2604, and a multiplexer 2610.
- the extended multi-view video encoder 2604 includes a base-view video encoder 2605, a 2D compatible video frame memory 2608, and a dependent-view video encoder 2609.
- the 2D compatible video encoder 2602 receives the left-eye image as an input, compresses and encodes it in the MPEG-2 format, and generates and outputs a 2D compatible video stream.
- Dec 2603 decodes the compression-encoded picture in the 2D compatible video stream, and outputs a decoded picture obtained as a result and 2D compatible video encoding information 2606.
- a picture is an image constituting a frame or a field, and is a unit of encoding.
- the decoded picture is stored in the 2D compatible video frame memory 2608 of the extended multiview video encoder 2604.
- the 2D compatible video encoding information 2606 is input to the base view video encoder 2605.
- the 2D compatible video encoding information 2606 includes attribute information (resolution, aspect ratio, frame rate, progressive / interlaced distinction, etc.), picture attribute information (picture type, etc.) of the corresponding picture, GOP of the decoded 2D compatible video stream. (Group of Pictures) structure and configuration including 2D compatible video frame memory management information.
- the 2D compatible video frame memory management information includes the decoded picture memory address stored in the 2D compatible video frame memory 2608, the display order information (PTS (Presentation Time Stamp) and temporal_reference) of the corresponding picture, and the code order information (code order of the file). Or DTS (Decoding Time Stamp).
- the extended multi-view video encoder 2604 receives the decoded picture output from the Dec 2603, the 2D-compatible video encoding information, the right-eye image, and the black image, performs compression encoding, and performs the base-view video stream, the dependent Dent-view video stream is output.
- the base view video encoder 2605 has a function of outputting the data compressed and encoded in accordance with the MPEG-4 MVC format as a base view video stream, and compresses and encodes a black image according to the 2D compatible video encoding information 2606. , A base-view video stream and base-view video encoding information 2607 are output.
- Base view video encoding information 2607 includes base view video stream attribute information (resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.), picture attribute information (picture type, etc.) of the corresponding picture, GOP structure, and base
- the configuration includes view video frame memory management information.
- the base-view video encoder 2605 sets the attribute information of the base-view video stream to the same value as the video attribute information included in the 2D compatible video encoding information 2606. Furthermore, the base-view video encoder 2605 determines a picture type for compression coding of pictures at the same display time in accordance with picture attribute information (picture type, etc.) and GOP structure included in the 2D compatible video coding information 2606. Thus, the black image is compression encoded. For example, if the picture type indicated by the 2D compatible video encoding information 2606 of the picture at time a is an I picture and the picture is the first picture of the GOP, the base view video encoder 2605 displays a black image indicating the same display time. The video access unit is compressed and encoded so as to be an I picture, and is used as a video access unit at the head of the GOP of the base-view video stream.
- the base-view video encoder 2605 compresses and encodes a black image indicating the same display time so as to become a B picture. To do. At this time, the DTS and PTS of the base-view video stream are matched with the DTS and PTS of the picture corresponding to the view showing the same time of the 2D compatible video stream, respectively.
- the base-view video frame memory management information includes a memory address of the frame memory 2608 that stores a decoded picture obtained by decoding a 2D-compatible video stream based on the 2D-compatible video frame memory management information, and the decoded picture
- This is information obtained by converting and associating a syntax element indicating display order information and code order information into a syntax element conforming to the rules of the compression encoding method of the base-view video stream.
- the syntax element is an element that defines attribute information necessary for encoding in the MPEG-2 or MPEG-4 MVC format compression encoding system. For example, header information such as macroblock type, motion vector, conversion coefficient, etc. It is an element to show.
- the dependent-view video encoder 2609 has a function of compressing and encoding in accordance with the MPEG-4 MVC format to generate a dependent-view video stream, and is based on information included in the base-view video encoding information 2607
- the right-eye image is compressed and encoded, and a dependent-view video stream is output.
- the dependent-view video encoder 2609 performs compression encoding using the decoded picture in the 2D compatible video frame memory as an inter-view reference.
- inter-view reference refers to referring to a picture showing a view from different viewpoints.
- the dependent-view video encoder 2609 determines a reference picture ID for performing inter-view reference based on the base-view video frame memory management information of the base-view video encoding information 2607.
- the dependent-view video encoder 2609 sets the same value as the value of the attribute information of the base-view video stream included in the base-view video encoding information 2607 as the video attribute information of the dependent-view video stream.
- the dependent-view video encoder 2609 determines the picture type of the image to be encoded based on the attribute information (for example, picture type) of the picture stored in the base-view video encoding information 2607 and the GOP structure. Then, the right-eye image is compression-encoded. For example, if the picture type indicated by the base-view video encoding information 2607 of the picture at time a is an I picture and the head of the GOP, the dependent-view video encoder 2609 sets the picture type of the picture at the same time a as an anchor picture. Then, the right-eye image is compression-encoded to obtain a video access unit at the head of the dependent GOP.
- the attribute information for example, picture type
- the dependent-view video encoder 2609 sets the picture type of the picture at the same time a as an anchor picture. Then, the right-eye image is compression-encoded to obtain a video access unit at the head of the dependent GOP.
- An anchor picture is a picture that does not refer to a picture temporally prior to the own picture and can be reproduced by jumping from the own picture. If the picture type indicated by the base-view video encoding information 2607 of the picture at time b is a B picture, the dependent-view video encoder 2609 sets the picture type of the picture at the same time b to B picture, and The image is compressed and encoded.
- the DTS and PTS of the dependent-view video stream are compression-encoded to match the DTS and PTS of the picture corresponding to the view to be displayed at the same time of the base-view video stream.
- the multiplexer 2610 divides the output 2D compatible video stream, base-view video stream, and dependent-view video stream into PES (Packetized Elementary Stream) packets, divides them in units of TS packets, and multiplexes them as a transport stream Output.
- PES Packetized Elementary Stream
- the playback device In order to allow the playback device to identify each video stream from the stream data of the multiplexed transport stream, the 2D compatible video stream, the base view video stream, and the dependent view video stream are: Separate PIDs are set.
- FIG. 22 shows video attributes to be matched in each compression encoding method in the MPEG-2 format and MPEG-4 MVC format compression encoding, and field names for indicating the video attributes.
- the resolution, aspect ratio, The frame rate and the value of the video attribute indicating progressive or interlace are configured to be the same between pictures of the respective encoding formats.
- FIG. 25 shows a GOP configuration of a 2D compatible video stream, a base view video stream, and a dependent view video stream in the present embodiment.
- the 2D compatible video stream, the base view video stream, and the dependent view video stream are all configured to have the same number of GOPs. That is, when the picture of the 2D compatible video stream is the head of the GOP, the picture of the base view video stream having the same value of the PTS and the picture of the dependent view video stream having the same value of the PTS are respectively the GOP head, Must be at the beginning of the dependent GOP.
- entry map information indicating where the GOP head picture is on the file may be stored as management information.
- the entry map information is stored as a separate information file as a management information file.
- both the base view position and the dependent view position at the same time are entered. Try to register on the map. In this way, the 3D video jump-in reproduction is facilitated by referring to the entry map.
- FIG. 36 shows the relationship between the structure of the transport stream and the PMT (Program Map Table) packet.
- a transport stream including a 3D video stream signaling information for performing 3D video decoding processing is included in a system packet such as a PMT packet.
- the descriptor includes a 3D information descriptor for performing signaling such as the relationship between each video stream and the start / end of 3D video playback of the present system, and a 3D stream descriptor set for each video stream. It is the composition which includes.
- FIG. 37 shows the structure of the 3D information descriptor.
- the 3D information descriptor includes a playback method, a left-eye video type, a 2D compatible video PID, a base view video PID, and a dependent view video PID.
- the playback method is information for signaling the playback method of the playback device.
- the playback device plays back 2D video of only the 2D compatible video stream.
- the playback apparatus When the value of the playback method is “1”, it indicates 3D video playback using 2D compatible video and dependent view video (that is, the 3D video playback method described in this embodiment).
- the playback apparatus performs 3D video playback of the 2D compatible video stream, base-view video stream, and dependent-view video stream by the playback method described in this embodiment. Note that a 3D video playback method in the present embodiment will be described later.
- the playback device plays back the 3D video of the video stream as a video stream compressed and encoded in the normal MPEG-4 MVC format.
- the playback device When the value of the playback method is “3”, this indicates that a 2D compatible video stream or a base-view video stream is double-played, and the playback device performs double playback.
- the doubling reproduction is to output one of the left and right views at a certain time a to both the L and R planes.
- Playback by this playback method is equivalent to 2D video playback as a user viewing screen, but since the frame rate does not change during 3D video playback, the playback device is connected to a display or the like via HDMI (High-Definition Multimedia Interface) or the like. In this case, re-authentication does not occur and seamless connection playback between the 2D video playback section and the 3D video playback section can be realized.
- HDMI High-Definition Multimedia Interface
- the left-eye video type is information indicating which of the multi-view video streams the left-eye image is compression-coded (the other video stream becomes the right-eye image).
- the value of the reproduction method is “0”, the value of this field has no meaning to refer to.
- the value of the reproduction method is “1”, it indicates which of the 2D compatible video and the dependent view video is the image for the left eye. That is, when the value of the playback method is “1” and the value of the left-eye video type is “0”, it indicates that the 2D compatible video stream is a video stream corresponding to the left-eye image.
- the playback method value is “2” or “3”
- the playback apparatus refers to the value of the left-eye video type to determine which video stream corresponds to the left-eye image. Judgment can be made.
- the 2D compatible video PID, the base view video PID, and the dependent view video PID indicate the PID of each video stream in the transport stream. With this information, the decoding target stream can be identified.
- FIG. 39 shows a 3D stream descriptor.
- the field name of the 3D descriptor includes a base view video type, a reference destination type, and a referenced type.
- the base view video type indicates which video is compressed and encoded in the base view video stream.
- the value of the base view video type is “0”, it indicates whether one of the left-eye image or the right-eye image data of the 3D video is compression-encoded, and when it is “1”, A black image is compressed and encoded as a dummy video that is replaced by a 2D compatible video stream and is not output to a plane.
- the reference destination type indicates the type of the video stream to which the dependent-view video stream is referred to as an inter-view reference.
- the value of the reference destination type is “0”, it indicates that the picture of the base view video stream is the inter-view reference destination, and when the value is “1”, the picture of the 2D compatible video stream is the inter-view reference destination. It shows that.
- the case where the value of the reference destination type is “1” indicates that the reference method is based on the 3D video method in the present embodiment.
- the referenced type indicates whether the corresponding video stream is referenced between views. If the reference is not made, the inter-view reference process can be skipped, so that the load of the decoding process can be reduced. Note that all or part of the information of the 3D information descriptor and the 3D stream descriptor may be inserted as supplementary data of each video stream instead of the PMT packet.
- FIG. 23 shows an example of the relationship among PTS, DTS, and picture type assigned to each video access unit of the 2D compatible video stream, base-view video stream, and dependent-view video stream in the transport stream.
- the data creation device 2601 converts the DTS and PTS values of the 2D compatible video stream picture and the dependent view video stream picture generated by compressing and encoding the left-eye image at the same time into the same DTS and PTS values, respectively. Set. Also, the same values as the PTS, DTS, and POC of the picture of the dependent-view video stream are set for the PTS, DTS, and POC of the picture of the base-view video stream to be reproduced at the same time.
- the picture of the base-view video stream having the same value of PTS, DTS, and POC is referred to.
- the picture reference ID (ref_idx_l0 or ref_idx_l1) specified from each macroblock of the picture of the dependent-view video stream has the same POC value.
- a value indicating the base view picture that is held is set.
- FIG. 27 is a diagram showing a data creation flow of the data creation device 2601. The data creation flow will be described below.
- the variable N is a variable for storing the frame number of the frame image to be compressed and encoded.
- step S2701 it is checked whether or not the Nth frame exists in the left-eye image. If it does not exist (step S2701: NO), it is determined that there is no more data to be compressed and encoded, and the process ends.
- step S2701 the number of images to be compression encoded (hereinafter referred to as “one encoded number”) is determined in one compression encoding flow (steps S2702 to S2706) (step “1 encoded number” hereinafter).
- the maximum number of video access units (the maximum number of GOPs, for example, 30 frames) set as one GOP is set as one encoded number.
- the last GOP of the video stream it is assumed that the number of frames to be encoded is less than the maximum number of GOPs depending on the length of the input video stream. In this case, the remaining number of frames is set as one encoded number. To do.
- the 2D compatible video encoder 2602 generates 2D compatible video streams for one encoded number (step S2703).
- the 2D compatible video stream is generated by the compression encoding according to the compression encoding method of the 2D compatible video stream from the Nth frame of the left-eye image by one encoded number, and output.
- the 2D compatible video decoder 2603 performs decoding processing of 2D compatible video streams for one encoded number (step S2704).
- a decoded picture and 2D compatible video encoding information obtained by decoding the compressed picture for one encoded number from the Nth frame are output.
- the base view video encoder 2605 generates base view video streams for one encoded number (step S2705). Specifically, based on 2D compatible video encoding information, base view video stream attribute information (resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.) as base view encoding information 2607, Set picture attribute information (picture type, etc.), GOP (Group of Pictures) structure, and 2D compatible video frame memory management information for each picture. Generate a video stream. Also, the set base-view encoding information 2607 is output.
- base view video stream attribute information resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.
- Set picture attribute information picture type, etc.
- GOP Group of Pictures
- the dependent view video encoder 2609 generates a dependent view video stream for one encoded number (step S2706). Specifically, based on the base-view video encoding information output in step S2705, the attribute information of the dependent-view video stream (resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.), Set picture attribute information (picture type, etc.), GOP (Group of Pictures) structure, and 2D compatible video frame memory management information for each picture.
- the attribute information of the dependent-view video stream resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.
- Set picture attribute information picture type, etc.
- GOP Group of Pictures
- the dependent-view video stream encoder 2609 does not refer to the picture of the base-view video stream, but performs the same display in the 2D-compatible video frame memory 2608 when encoding using inter-picture predictive encoding.
- the dependent-view video stream is compressed and encoded for one encoded number from the Nth frame of the right-eye image using inter-picture predictive encoding while referring to the picture obtained by decoding the 2D compatible video stream indicating the time. Is generated.
- the multiplexer 2610 converts the 2D compatible video stream, base view video stream, and dependent view video stream into PES packets. Next, the PES packet is divided into TS packets, and a multiplexed transport stream is generated. Thereafter, one encoded number is added to N (step S2707).
- step S2707 When the process of step S2707 is completed, the process returns to step S2701 and the process is repeated.
- the number of encoded sheets in one flow can be changed.
- the value of one encoded sheet in step S2702 may be set to be small. For example, when the number of reordering at the time of video encoding is two, if compression encoding is executed in units of four, it will not be affected by reordering. It is assumed that the number of reordering is a compression encoding method with two pictures and the picture types are I1, P4, B2, B3, P7, B5, and B6 (numbers are in display order). When the number of encoded images is 3, since the picture of P4 cannot be processed, the compression encoding process of B2 and B3 cannot be performed.
- the number of encoded images is set to 4, the P4 picture can be processed, so that the B2 and B3 compression encoding processes can be performed.
- the number of encoded sheets may be set to an optimal number according to the characteristics of the image within the maximum number of GOPs for each compression encoding flow.
- FIG. 28 is a block diagram showing a functional configuration of the playback device 2823.
- the playback device 2823 includes a PID filter 2801, a 2D compatible video decoder 2821, an extended multi-view video decoder 2822, a first plane 2808, and a second plane 2820.
- the PID filter 2801 filters the input transport stream.
- the PID filter 2801 determines the TS packet whose PID value matches the PID value required for reproduction, from the TS packet, to the 2D compatible video decoder 2821 or the extended multi-view video decoder 2822 according to the PID value. Forward.
- the PID of the 2D compatible video stream is 0x1011
- the PID of the base-view video stream of the multi-view video stream is 0x1012
- the PID of the dependent-view video stream of the multi-view video stream is 0x1013
- the PID value of the TS packet is When the PID value included in the TS packet matches the predetermined PID value, the TS packet is transferred to the corresponding decoder.
- the first plane 2808 is a plane memory that holds a picture decoded by the 2D compatible video decoder 2821 and output according to the value of the PTS.
- the second plane 2820 is a plane memory that holds a picture decoded by the extended multi-view video decoder 2822 and output according to the value of the PTS.
- the 2D compatible video decoder 2821 has basically the same decoding function as the MPEG-2 format decoder which is a compression encoding method of 2D video
- the extended multi-view video decoder 2822 is a 3D video that realizes inter-view reference.
- the decoder has basically the same decoding function as the MPEG-4 MVC format decoder, which is a compression encoding method.
- a general decoder of MPEG-2 format compression encoding system is a video decoder 2901
- a general decoder of MPEG-4 MVC format compression encoding system is a multi-view video decoder 2902.
- the video decoder 2901 and the multi-view video decoder 2902 will be described with reference to FIG. Thereafter, different portions between the 2D compatible video decoder 2821 and the video decoder 2901 and different portions between the extended multi-view video decoder 2822 and the multi-view video decoder 2902 will be mainly described.
- the video decoder 2901 includes TB (Transport Stream Buffer) (1) 2802, MB (Multiplexing Buffer) (1) 2803, EB (Elementary Stream Buffer) (1) 2804, D1 (2D compatible video compression video decoder) 2805. , And O (Re-ordering Buffer) 2806.
- TB (1) 2802 is a buffer that temporarily accumulates TS packets as they are when TS packets including a video stream are output from the PID filter 2801.
- MB (1) 2803 is a buffer for temporarily storing PES packets when outputting a video stream from TB (1) 2802 to EB (1) 2804.
- TB (1) 2802 to MB (1) 2803 When data is transferred from TB (1) 2802 to MB (1) 2803, the TS header and adaptation field of the TS packet are removed.
- EB (1) 2804 is a buffer for storing pictures (I picture, B picture, P picture) in a compression-encoded state.
- pictures I picture, B picture, P picture
- the PES header is removed.
- D1 (2805) creates a picture of the frame image by decoding each video access unit of the video elementary stream at a predetermined DTS time.
- the picture decoded in D1 (2805) is output to the plane 2808 or O2806.
- DTS and PTS are different, such as P picture and I picture, they are output to O2806, and when the values of DTS and PTS are the same as B picture, they are output to plane 2808 as they are.
- O2806 is a buffer for performing reordering when the values of DTS and PTS of the decoded picture are different, that is, when the decoding order and display order of the pictures are different.
- D1 (2805) performs decoding processing with reference to the picture data stored in O2806.
- Switch 2807 switches between outputting an image buffered in O2806 or outputting directly from D1 (2805) when outputting the decoded picture to plane 2808.
- the multi-view video decoder 2902 includes TB (2) 2809, MB (2) 2810, EB (2) 2811, TB (3) 2812, MB (3) 2813, EB (3) 2814, A decoding switch 2815, an inter-view buffer 2816, a D2 (multiview video compression video decoder) 2817, a DPB (decoded picture buffer) 2818, and an output plane switch 2819 are included.
- TB (2) 2809, MB (2) 2810, and EB (2) 2811 have the same functions as TB (1) 2802, MB (1) 2803, and EB (1) 2804, respectively, but the data to be buffered Is the base view video stream.
- TB (3) 2812, MB (3) 2813, and EB (3) 2814 have the same functions as TB (1) 2802, MB (1) 2803, and EB (1) 2804, respectively, but the data to be buffered Is a dependent-view video stream.
- the switch 2815 extracts the data of the video access unit to which the DTS is assigned according to the value of the DTS from the EB (2) 2811 and the EB (3) 2814, and configures the 3D video access unit. ).
- D2 (2817) performs a decoding process on the 3D video access unit transferred via the switch 2815, and creates a picture of the frame image.
- the decoded picture of the base-view video decoded by D2 (2817) is temporarily stored in the Inter-view buffer 2816.
- D2 (2817) performs a decoding process on the picture of the dependent-view video stream with reference to the decoded picture of the base-view video stream having the same PTS value stored in the Inter-view buffer 2816.
- the multi-view video decoder 2902 creates a reference picture list for designating a picture for performing inter-view reference based on the picture type and the syntax element of the picture of the base-view video stream and the picture of the dependent-view video stream.
- D2 transfers the decoded picture of the base view and the decoded picture of the dependent view stored in the inter-view buffer 2816 to the DPB 2818, and outputs it via the output plane switch 2819 according to the value of the PTS.
- the DPB 2818 is a buffer that temporarily stores decoded pictures.
- D2 (2817) is used to refer to a picture that has already been decoded when decoding a video access unit such as a P picture or a B picture using the inter picture predictive coding mode.
- the output plane switch 2819 performs a process of outputting the decoded picture to an appropriate plane. For example, when the base-view video stream indicates the left-eye image and the dependent-view video stream indicates the right-eye image, the base-view video stream picture is set to the left-eye image plane and the dependent-view video The stream picture is output to the right-eye image plane.
- the 2D compatible video decoder 2821 shown in FIG. 28 not only transfers the picture decoded in D1 (2805) to the O2806 or the switch 2807, but also in the inter-view buffer 2816 of the extended multi-view video decoder 2822 based on the DTS value. Forward.
- the extended multi-view video decoder 2822 has the same basic structure as that of the multi-view video decoder 2902. Therefore, the description of the common parts of functions will be omitted and the different parts will be described.
- the extended multi-view video decoder 2822 converts the picture transferred according to the DTS value from the 2D compatible video decoder 2821 to the same value in the PTS and DTS of the base-view video stream that has been decoded previously in the inter-view buffer 2816. Overwrite the area where the decoded picture is stored. Therefore, the extended multi-view decoder 2822 can refer to the decoded picture of the 2D-compatible video stream as if it were the decoded picture of the base-view video stream when decoding the picture of the dependent-view video stream.
- the address management on the Inter-view buffer 2816 need not be changed from the management of the decoded picture of the conventional base-view video stream.
- the extended multi-view video decoder 2822 outputs to the second plane 2820 out of the video stored in the DPB 2818, so that only the picture of the dependent-view video stream is output according to the value of the PTS. To control. The pictures of the base view video stream are not output to the plane because they are not related to display.
- the 2D-compatible video decoder 2821 outputs the picture of the 2D-compatible video stream to the first plane 2808 according to the value of the PTS
- the extended multi-view video decoder 2822 outputs the dependent view of the multi-view video stream.
- the picture of the video stream is output to the second plane 2820 according to the value of PTS.
- FIG. 30 is a diagram illustrating a flow of decoding processing and output processing in 3D video by the playback device 2823.
- the playback device 2823 determines whether there is a picture in EB (1) 2804 (step S3001). If there is no picture (step S3001: No), it is determined that the transfer of the video stream has ended, and the processing is ended.
- the playback device 2823 uses the extended multi-view video decoder 2822 to perform base-view video stream decoding processing (step S3002). Specifically, in accordance with the value of DTS, a picture attached to the corresponding DTS is extracted from EB (2), decoded, and stored in Inter-view buffer 2816.
- the management of pictures in Inter-view buffer 2816 is the same as the management in the conventional MPEG-4 MVC format and will not be described in detail. However, as the management information for creating the reference picture list, the reference destination of PTS and POC and the decoded picture
- the table information that associates the data address of Inter-view buffer 2816 indicating the internal address is managed internally.
- the playback device 2823 uses the 2D compatible video decoder 2821 to perform decoding processing of the 2D compatible video stream (step S3003). Specifically, according to the value of DTS, a picture assigned to the corresponding DTS is extracted from EB (1), and decoding processing is performed. At this time, the decoded picture is transferred to O2806 or switch 2807. Further, the decoded picture is also transferred to Inter-view buffer 2816.
- the extended multi-view video decoder overwrites the transferred picture with the base-view picture to which the same value of DTS and PTS in the Inter-view buffer 2816 is assigned.
- the picture management of the inter-view buffer 2816 is managed by the memory address of the PTS and the inter-view buffer 2816, for example.
- the decoded picture is overwritten.
- management information eg, PTS
- PTS management information
- D2 2817
- D2 can be decoded with reference to the picture obtained by decoding the 2D compatible video stream in the same process as the decoding process of the conventional MPEG-4 MVC dependent view video stream. it can.
- the extended multi-view video decoder 2822 performs a dependent-view video stream decoding process (step S3004). Specifically, the extended multi-view video decoder 2822 takes out the picture attached to the corresponding DTS from the EB (3) according to the value of the DTS, and stores the picture of the dependent-view video stream in the Inter-view buffer 2816. The decoding process is performed while referring to the picture.
- the picture referred to at this time is not a picture of the base-view video stream but a picture of the 2D compatible video stream overwritten in step S3003.
- the playback device 2823 outputs the decoded picture of the 2D-compatible video stream to the first plane 2808 according to the PTS value, and outputs the decoded picture of the dependent-view video stream to the second plane 2820 according to the PTS value (step).
- D1 decoding process performed by D1 (2805) of the playback apparatus 2823
- the LSI Large Scale Integration
- the MPEG-4 MVC format decoding process performed in D2 (2817) is the same as the existing MPEG-4 MVC format process
- the LSI or software of the existing MPEG-4 MVC format video playback apparatus Can be used.
- ⁇ Example of usage mode of playback device 2823> Referring to FIG. 5, a 3D digital television 100 that can play back 3D video of a video stream created by the data creation device 2823 and a 2D digital television 300 that can play back only conventional 2D video that does not support playback of 3D video are given as examples. I will explain.
- the user views 3D video using the 3D digital television 100 and the 3D glasses 200.
- the 3D digital television 100 can display 2D video and 3D video, and displays video by playing back a stream included in the received broadcast wave. Specifically, a 2D compatible video stream compression-encoded in the MPEG-2 format, a base-view video stream and a dependent-view video stream compression-encoded in accordance with the MPEG-4 MVC format are reproduced.
- the 3D digital television 100 alternately displays the left-eye image obtained by decoding the 2D-compatible video stream and the right-eye image obtained by decoding the dependent-view video stream.
- the user can view the image reproduced in this way as a stereoscopic image by wearing the 3D glasses 200 and viewing the image.
- FIG. 5B shows the time when the left-eye image of the 3D glasses 200 is displayed.
- the 3D glasses 200 make the liquid crystal shutter corresponding to the left eye transparent and the liquid crystal shutter corresponding to the right eye shielded.
- (C) in the figure shows when the right eye image is displayed.
- the liquid crystal shutter corresponding to the right eye is made transparent, and the liquid crystal shutter corresponding to the left eye is shielded from light.
- the 2D digital television 300 in FIG. 5D corresponds to the playback of 2D video, and 2D video obtained by decoding the 2D compatible video stream among the video streams included in the transport stream created by the data creation device 2601 is displayed. Can be played. ⁇ 1-4. Modification>
- the embodiments of the data creation device and the playback device according to the present invention have been described above. However, the illustrated data creation device and the playback device can be modified as follows, and the present invention is described in the above embodiment. Of course, the present invention is not limited to the data creation device and the playback device as shown in FIG.
- the decoded picture of the 2D compatible video stream having the same PTS value is overwritten on the decoded picture of the base-view video stream of Inter-view buffer 2816 in step S3003.
- the address reference destination may be changed without performing the overwriting process.
- the decoded picture of the base view is stored in the DPB 2818.
- the decoded picture of the base view video stream is not referred to, and may not be stored in the DPB 2818. In this way, it is possible to reduce the amount of memory for storing the pictures of the base-view video stream from the DPB 2818.
- the base-view video stream is generated by including it in the transport stream, and the decoding process of the pictures of the base-view video stream is performed in the decoding process. Processing may be omitted.
- the extended multi-view video decoder 2822 analyzes header information (eg, POC acquisition, picture type, ViewID, information acquisition with or without reference, etc.) and Inter-view buffer 2816 without decoding pictures of the base-view video stream. An area where one picture can be stored is secured.
- the extended multi-view video decoder 2822 stores the decoded picture output from the 2D compatible video decoder having the same value as the PTS and DTS obtained by header information analysis in the reserved area.
- an extended multi-view video decoder can manage an inter-view buffer 2816.
- a 2D compatible video stream including the above information may be generated.
- all or part of the syntax element of the base view video stream is recorded in the supplementary data of the 2D compatible video stream. That is, information for managing pictures in the Inter-view buffer buffer 2816 (the POC indicating the display order of the MPEG-4 MVC format, the slice_type indicating the picture type, the nal_ref_idc indicating the reference / non-reference of the picture, the base reference picture list) Ref_pic_list_mvc_modification, View ID of base-view video stream, MMCO command) and the like, which are information for creation, are included in the supplementary data of each picture of the 2D compatible video stream.
- the base-view video stream may not be multiplexed on the transport stream.
- the picture of the MPEG-4 MVC dependent view video stream directly refers to the picture of the MPEG-2 video stream.
- the data format is almost the same as the conventional one, so it is compatible with the encoding device and playback device compatible with the conventional MPEG-4 MVC format.
- the encoding device and the playback device corresponding to the video stream data of the present embodiment can be realized with good performance and a small improvement.
- the O2806 and DPB2818 memories are handled as separate areas, but the memory space may be shared as shown in FIG.
- the overwriting process can be omitted.
- Inter-view buffer 2816 and DPB 2818 are handled as separate buffers, but these may be the same buffer.
- the decoded picture of the base-view video stream having the same PTS and the same ViewID in DPB2818 may be replaced with the decoded picture of the 2D compatible video stream.
- the 2D compatible video stream and the base view video stream are P pictures (P3), but the picture of the dependent view video stream is a B picture (B3). ing.
- the pictures of the 2D compatible video stream, the base view video stream, and the dependent view video stream are all P pictures, and only the I picture and the P picture are included in all the video streams. Therefore, special reproduction processing with the I picture and P picture selected is facilitated.
- separate PIDs are set as the PID of each video stream in the multiplexing of the transport stream, but the base view video stream and the dependent view video stream are The same PID may be assigned.
- the access units of the video streams can be merged and transferred in accordance with the specifications of the compression encoding method of the multi-view video stream.
- the base-view video stream and the dependent video stream are merged according to the specifications of the compression coding system, and the playback device has a configuration as shown in FIG. 45, and the data transfer line of the extended multi-view video decoder is one. What is necessary is just to comprise.
- the base view video stream and the dependent view video stream may share header (for example, sequence header, picture header) information of each access unit storing pictures at the same time. That is, when header information is provided only in the base-view video stream and the dependent-view video stream is decoded, the header information necessary for decoding is decoded with reference to the header information of the base-view video stream. Also good. Therefore, addition of header information necessary for decoding can be omitted from the dependent-view video stream.
- header for example, sequence header, picture header
- the DTS values of the pictures of the 2D compatible video stream and the dependent view video stream at the same display time are equal to each other.
- the DTS of the picture of the pendant view video stream and the picture of the base view video stream are also set to the same value.
- the DTS values of pictures in each video stream at the same display time may not be the same.
- the DTS value of the 2D compatible video stream may be set to a value that is decoded earlier (for example, one frame before) than the base-view / dependent-view video stream.
- the inter-view buffer overwrite process and the dependent-view video stream picture decoding process can be performed with a margin. it can.
- the PTS of the picture of the 2D compatible video stream storing the parallax images at the same display time is the same as the PTS of the picture of the dependent view.
- the decoding process of the 2D compatible video stream is performed in advance. Therefore, the PTS of the pictures of the 2D compatible video stream at the same display time is set to be ahead of the base-view / dependent-view video stream (for example, one frame before).
- the PTS of the picture of the 2D compatible video stream is set to one frame before the PTS of the picture of the dependent view video stream.
- the pictures of the base-view video stream of the Inter-view buffer are exchanged, the pictures of the base-view video stream are exchanged with the pictures of the 2D compatible video stream of the PTS having a value indicating the previous frame.
- the 2D compatible video decoder 2821 outputs the decoded picture of the 2D compatible video stream to the first plane 2808 in accordance with the PTS value in step S3005.
- the extended multi-view video decoder 2822 may be configured to output both videos using the output plane switch 2819.
- the multiplexing method is the transport stream, but the present invention is not limited to this.
- the MP4 system format can be used as a multiplexing method.
- the input in FIG. 34 is separated into a 2D compatible video stream, a base-view video stream, and a dependent-view video stream as a file multiplexed by MP4 and decoded.
- the picture of the dependent-view video stream is decoded while referring to the picture in which the picture of the base-view video stream of Inter-view buffer 2816 is overwritten with the picture of the 2D-compatible video stream.
- time information in each access unit is specified based on the header information (stts, stsz, etc.) of the MP4 system format.
- the reference picture of the dependent-view video stream is a decoded picture of the 2D compatible video stream, and normal multi-view video Since it is different from the stream structure, the stream type and stream_id given to the PES packet header may be set to be different from the values in the case of the conventional multi-view video stream.
- the playback device can determine that the playback method is the 3D video playback method according to the present embodiment by referring to the stream type and stream_id, and can change the playback method.
- step S3003 When the inter-codec reference switch 2824 is ON as shown in FIG. 40, the data transfer described in step S3003 is performed from the 2D compatible video decoder to the inter-view buffer in the extended multi-view video decoder. In this case, data transfer is not performed.
- the plane selector 2825 includes a first plane 2808 in which picturer is output from the 2D compatible video decoder, a second plane 2820 in which a picture of the base-view video stream of the extended multi-view video decoder is output, and an extended multi-view video decoder dependent. Which plane is output as 2D video, 3D video left-eye image, or 3D video right-eye image to the third plane 2826 in which the view video stream picture is output select.
- the playback device 2823b changes the playback mode by switching the output from the inter-codec reference switch 2824 and the plane selector 2825 according to the playback method.
- the playback device 2823b turns off the inter-codec reference switch 2824. Then, the plane selector 2825 selects the first plane 2808 as 2D video.
- the playback device 2823b turns on the inter-codec reference switch 2824. Then, the plane selector 2825 selects the first plane 2808 or the second plane 2820 as the left-eye image, and selects the third plane 2826 as the right-eye image.
- the playback device 2823b turns off the inter-codec reference switch 2824. Then, the plane selector 2825 selects the second plane 2820 as the left-eye image and selects the third plane 2826 as the right-eye image.
- the playback device 2823b turns off the inter-codec reference switch 2824. Then, the plane selector 2825 selects the first plane 2808 as the left-eye image and selects the first plane 2808 as the right-eye image.
- the playback system when the playback system generates a transport stream that switches from 3D video playback by 2D compatible video stream and dependent-view video stream to 2D video playback of 2D compatible video stream Considering the delay of the decoding process, as shown in FIG. 42, the same video as the 2D compatible video stream may be compressed and encoded into the dependent-view video stream when the playback method changes. Good.
- a section in which the same video as the 2D compatible stream is compression-coded into a dependent-view video stream is a 2D transition period shown in the upper part of FIG. In this 2D transition section, since playback is performed as 2D video regardless of which method is used, smooth video transition is achieved for the user.
- the POC value of the picture of the dependent-view video stream indicating the same time display may be set to the value of temporal_reference indicating the display order included in each picture in MPEG-2 format compression encoding.
- the POC of the dependent-view video stream indicating the same time display may be included in the user data of each picture of the 2D compatible video stream.
- the value of temporal_reference can also be set to a unique value, so that the degree of freedom of setting in the compression encoding process increases.
- a high-quality filter 4301 may be applied to the decoding result of the 2D compatible video stream.
- the high image quality filter 4301 is a filter that reduces block noise like a deblocking filter defined by MPEG-4 AVC.
- a flag indicating whether to apply the high image quality filter 4301 is prepared. For example, the high image quality filter 430 is applied when the flag indicates ON, and is not applied when the flag is OFF.
- This flag can be included in the PMT descriptor and supplementary data of the stream.
- the playback device applies a filter to the inter-view buffer 2816 in the decoding result before transferring the data.
- This configuration realizes high quality 2D video for 2D compatible video streams.
- the dependent-view video stream can be decoded while referring to the picture that has been subjected to the high image quality processing, it is possible to improve the quality of the resulting 3D video.
- the extended multi-view video stream is configured so that a plurality of dependent-view video streams can be processed, and the PTS is the same in the replacement process with the picture of the 2D compatible video stream in the Inter-view buffer 2816.
- the base view picture having a value may be replaced.
- the ViewID to be replaced can be specified in the 2D compatible video stream, and instead of always replacing the picture of the base view, it may be selected and replaced from the pictures of the plurality of views.
- the 2D compatible video stream is described as MPEG-2 video and the multi-view video stream (base view video stream and dependent view video stream) is described as MPEG-4 MVC video. Needless to say, is not limited to this.
- this can be dealt with by appropriately changing the configuration in accordance with the characteristics of the codec. For example, when the 2D-compatible video stream is MPEG-4 AVC and the multi-view video stream is “new codec”, the O2806 and the switch 2807 shown in FIG. 34 are connected to the DPB like the playback device shown in FIG. Instead, the management of the pictures in the inter-view reference buffer 2816 may be performed according to the “new codec” configuration.
- the left-eye picture and right-eye picture are arranged alternately and alternately in the vertical direction in the screen displayed on the display, and the pixels constituting the left-eye picture are passed through the upper lens called a lenticular lens on the display surface. Only the right eye and the pixels that form the picture for the right eye may be imaged only to the right eye, so that the left and right eyes can see a picture with parallax and can be displayed as a 3D image. Further, instead of the lenticular lens, a device having the same function, for example, a liquid crystal element may be used.
- a vertically polarized filter is installed on the left-eye pixel on the display side
- a horizontally polarized filter is installed on the right-eye pixel
- viewers install a vertically polarized filter on the left eye and a horizontally polarized filter on the right eye. You may use the system called the polarization system using the installed polarized glasses.
- a depth map in which a depth value is given in pixel units to the 2D video is prepared separately. You may make it produce
- FIG. 4 schematically shows an example of generating a parallax image of a left-eye image and a right-eye image from a 2D video image and a depth map.
- the depth map has a depth value corresponding to each pixel in the 2D image.
- the circular object in the 2D image is information indicating that the circular object is nearby in the depth map (the depth value is High) is assigned, and information indicating that the other area is farther from the circular object (depth value is Low).
- This information may be represented by a bit string for each pixel, or an image image (for example, “black” indicates that the depth is Low and “white” indicates that the depth is High). Also good.
- the parallax image can be created by adjusting the parallax amount of the 2D video from the depth value of the depth map. In the example of FIG.
- the parallax amount of the pixel of the circular object is increased when creating the parallax image.
- the depth value of the region other than the circular object is Low, the parallax amount of the pixels other than the circular object is reduced, and the left-eye image and the right-eye image are created. If the left-eye image and the right-eye image are displayed using a sequential separation method or the like, stereoscopic viewing is possible.
- one dependent-view video stream is used.
- the present invention is not limited to this, and a plurality of dependent-view video streams may be used.
- high-quality 3D video reproduction can be realized, and the bit rate of each dependent-view video stream can be kept low.
- FIG. 99 is a diagram for explaining a case where two dependent-view video streams are used as an example.
- the 2D compatible video stream is created by compressing and encoding 2D video (corresponding to the video for the left eye), the base view video stream is created by compressing and encoding the black video, and the dependent view video stream 1 is the video for the left eye.
- the dependent-view video stream 2 is created by compressing and coding the right-eye video.
- the left-eye video related to the dependent-view video stream 1 and the right-eye video related to the dependent-view video stream 2 are assumed to have higher image quality than the 2D video related to the 2D-compatible video stream.
- Each picture of both dependent-view video streams is compression-coded with reference to the decoded picture of the 2D compatible video stream.
- the amount of data included in the dependent-view video stream 1 corresponds to the difference between the left-eye video with higher image quality than the 2D video included in the 2D compatible video stream and the 2D video included in the 2D compatible video stream. Compared with the case where the high-quality left-eye video is directly compressed and encoded, the data amount can be greatly reduced.
- the amount of data included in the dependent-view video stream 2 corresponds to the difference between the right-eye video with higher image quality than the 2D video included in the 2D compatible video stream and the 2D video included in the 2D compatible video stream. Compared to the case where the high-quality right-eye video is directly compressed and encoded, the data amount can be greatly reduced.
- the decoded picture of the base-view video stream is replaced with the decoded picture of the 2D-compatible video stream at the same display time (PTS), and each dependent-view video stream is decoded.
- FIG. 100 is a diagram showing a configuration of a data creation device 10001 according to this modification.
- the data creation device 10001 has basically the same configuration as the data creation device 2601 except that it includes two dependent-view video encoders (10009 and 10011). Hereinafter, the difference from the data creation device 2601 will be mainly described.
- Dependent view 1 video encoder 10009 receives the left-eye original video image, performs compression encoding according to base-view video encoding information 2607, and outputs dependent-view video stream 1.
- the dependent view 1 video encoder 10009 compresses a decoded picture of a 2D compatible video stream of the same display time (PTS) / DTS stored in the 2D compatible video frame memory 2608 as an inter-view reference image. To do.
- PTS display time
- the dependent view 2 video encoder 10011 receives the right-eye original image as input, performs compression encoding according to the base view video encoding information 2607, and outputs the dependent view video stream 2. In this compression encoding, the dependent view 2 video encoder 10011 compresses the decoded picture of the 2D compatible video stream of the same display time (PTS) / DTS stored in the 2D compatible video frame memory 2608 as an inter-view reference image. To do.
- PTS display time
- FIG. 101 is a block diagram showing a configuration of a playback apparatus 10123 according to this modification.
- the playback device 10123 basically has the same configuration as the playback device 2823, the difference will be mainly described below.
- the playback device 10123 additionally includes TB (4), TB (4), and EB (4), and two dependent-view streams are input to the multi-view video decoder.
- the picture of the base-view video stream is replaced with the picture of the 2D-compatible video stream at the same display time in the inter-view reference buffer 2816.
- decoding is performed using the replaced picture as a reference image.
- the playback device 10123 performs 3D video playback by outputting both decoded videos of the dependent-view video stream as 3D video.
- the 2D-compatible video stream is generated by encoding progressive video, but the present invention is not limited to this, and may be generated by encoding interlaced video.
- the 2D compatible video stream is generated by encoding an interlaced video
- the dependent-view video stream is generated by encoding a progressive video.
- the dependent-view video stream is configured to compress a picture obtained by enlarging the decoding result of the 2D compatible video stream twice in the vertical direction as a reference image.
- FIG. 102 is a diagram showing a stream configuration according to the present modification.
- the 2D compatible video stream is generated by compressing and encoding a full HD frame rate interlaced video (1080i / 60i) composed of a top field picture and a bottom field picture.
- the base-view video stream of the multi-view video stream is a progressive video (1080 / 60p) with a full HD frame rate of 60 Hz in which monochrome video such as black (hereinafter referred to as “black video”) is encoded at a low bit rate. ) Is compressed and encoded.
- the dependent-view video stream of the multi-view video stream is generated by compressing and encoding progressive video (1080 / 60p) with the same content as the 2D compatible video stream and full HD and a frame rate of 60 Hz.
- the decoded picture of the base-view video stream is replaced with the decoded picture of the 2D-compatible video stream at the same display time (PTS) at the timing indicated by the DTS, so that the 2D-compatible video at the same display time.
- PTS display time
- Each dependent-view video stream can be decoded with reference to the picture of the stream.
- interlaced 2D video playback can be performed by playing back a 2D compatible video stream
- progressive 2D video playback can be performed by playing back a dependent-view video stream.
- FIG. 103 is a diagram showing a configuration of a data creation device 10301 according to the present modification.
- the data creation device 10301 basically has the same configuration as the data creation device 2601, but the 2D compatible video encoder 10302, the progressive unit 10311, and the dependent view video encoder 10309 are different.
- the 2D compatible video encoder 10302 generates a 2D compatible video stream by interlacing and encoding a progressive original image with a frame rate of 60 Hz.
- the progressive unit 10311 converts the interlaced video obtained by decoding the 2D compatible video stream into a progressive video, and stores the progressive video in the 2D compatible video frame memory 2608.
- the progressive method there is a method of interpolating scanning lines between vertical scanning lines, but is not limited thereto.
- the scanning line at the same position of the previous picture may be used as it is.
- the dependent-view video encoder 10309 generates a dependent-view video stream by compressing and encoding the input progressive video based on the base-view video encoding information.
- the dependent-view video encoder 10309 views a picture at the same display time as a picture to be encoded in a progressive video among pictures stored in the 2D compatible video frame memory 2608 by the progressive unit 10311. Used as an inter-reference image.
- FIG. 104 is a diagram showing a configuration of a playback apparatus according to this modification.
- the playback device 10423 outputs the decoded video of the dependent-view video stream as a progressive video.
- the playback device 10423 has basically the same configuration as the playback device 2823, except that a progressive unit 10424 is added.
- the progressive unit 10424 progressively converts the interlaced picture decoded by the 2D compatible video compression video decoder 2805.
- the progressive processing is performed by the same method as the progressive processing unit of the data creation device. For example, as described above, two scanning lines arranged above and below an interlaced picture are used to interpolate the scanning lines between them.
- the decoding process is performed using a picture stored in the inter-view reference buffer 2816 as a reference image with a picture having the same display time as the decoding target picture.
- the progressive method by the progressive unit 10424 of the playback device 10423 is not limited to the above-described method.
- the pixels between them may not be interpolated from pixels arranged vertically, but may be interpolated from pixels arranged vertically and diagonally.
- the progressive unit 10311 of the data creation device 10301 has a function of performing progressive processing by a plurality of methods
- the optimal method is selected to perform progressive processing, and which method is applied is described as a descriptor.
- This information may be included in the supplementary data of the system packet or video stream.
- the progressive unit 10424 determines which progressive method has been used based on the descriptor information, and uses the progressive method.
- the referenced picture is progressively processed with the optimal progressive method to improve the image quality, so that the dependent-view video stream that is compressed using the image with the improved image quality for inter-view reference is also high.
- the image quality can be improved.
- interpolation may be performed using not only pictures with the same display time but also pictures with display times arranged in front and back.
- interpolation is performed using pictures stored in O2806 or DPB2818 in the 2D compatible video decoder.
- information indicating which frame is used is stored in the progressive method field of the descriptor.
- inter-view reference is performed between the 2D compatible video stream and the multi-view video stream in the present embodiment, the inter-view reference may be performed between a plurality of multi-view video streams.
- FIG. 105 is a diagram showing a stream configuration according to this modification.
- the encoding target is the left-eye original image 10511, the right-eye original image 10512, and the middle original image 10515 with the center of the left eye and right eye as the viewpoint.
- the multi-view video stream 1 (10501) includes a base-view video stream 10502 and a dependent-view stream 10503.
- the base-view video stream 10502 is generated by compressing and encoding the left-eye original image 10511
- the dependent-view stream 10503 is generated by compressing and encoding the right-eye original image 10512.
- the multi-view video stream 2 (10504) includes a base-view video stream 10505, a dependent-view stream 1 (10506), and a dependent-view stream 2 (10507).
- the base view video stream 10505 is generated by compressing and encoding a monochrome image 10513 such as black
- the base view video stream 10506 is generated by compressing and encoding a monochrome image 10514 such as black.
- the dependent view stream 10507 is generated by encoding the middle original picture 10515.
- the compression encoding of the middle original picture 10515 is performed by decoding the decoded picture of the base-view video stream 10505 corresponding to the left-eye original picture 10511 and the dependent view stream 1 (10506) corresponding to the right-eye original picture 10512.
- the decoded picture is used as a reference image.
- the base-view video stream (10505) and the dependent-view video stream 1 (10506) of the multi-view video stream 2 (10504) at the timing represented by the DTS. are replaced with the decoded pictures of the base-view video stream (10502) of the multi-view video stream 1 (10501) and the dependent-view video stream 1 (10503).
- the dependent view of the multi-view video stream 2 (10504) is then referenced with reference to the decoded pictures of the base-view video stream (10505) of the multi-view video stream 1 (10501) and the dependent-view video stream 1 (10506).
- Video stream 2 (10507) is decoded.
- FIG. 106 is a diagram showing an example of matching information indicating the correspondence between streams.
- the matching information includes information for specifying the exchange source stream and the exchange destination stream, and information indicating which View ID of the exchange-destination multi-view video corresponds to each View ID of the exchange-destination multi-view video. .
- the playback device can recognize the correspondence between the exchange source stream and the exchange destination stream by referring to the correspondence relationship between the View IDs of the multi-view video streams indicated by the matching information.
- the base-view video stream is generated by compressing and encoding a monochrome image such as a black screen.
- a differential video for improving the image quality of a 2D compatible video stream may be generated by compression encoding.
- the difference video to improve the quality of the 2D-compatible video stream and referencing the high-definition picture when encoding the dependent-view video stream, the video of the dependent-view video stream is also improved. it can.
- FIG. 86 is a diagram showing a configuration of a video stream according to the present modification.
- the 2D compatible video stream is generated by compressing and encoding the original image of the left eye.
- the multi-view video stream includes a base-view video stream and a dependent-view stream, and the base-view video stream is generated by compressing and encoding the difference between the 2D compatible video and the left-eye original video, and the dependent-view stream Is generated by compression-coding the original image of the right eye.
- FIG. 89 shows an outline of the difference video creation procedure.
- a 2D compatible video stream (8903) is generated.
- a decoded picture (8902) of the 2D compatible video is obtained.
- a difference value (difference video) between the original picture image 8901 and the decoded picture (8902) of the 2D compatible video is calculated (8913), and a value 128 is added to the calculated difference value (8914).
- the difference value is simply calculated, it becomes signed information (eg, -128 to +128 in the case of 8-bit color), and an extra bit is required to represent the sign.
- the difference between the picture of the original picture 8901 and the decoded picture 8902 of the 2D compatible video has a small dynamic range, and it is easy to obtain a large compression effect.
- FIG. 89 shows an outline of the difference video composition procedure described in the upper part of FIG.
- a 2D compatible video stream 8903 is decoded to obtain a decoded picture 8902 of the 2D compatible video.
- the difference video 8904 is obtained by decoding the base-view video stream 8905, and the information returned to the signed state by adding -128 to each picture (8033) is synthesized in units of pixels (8034).
- the quality of the 2D compatible video stream is improved using the difference video, and the picture of the dependent-view video stream can be improved by using the high-quality picture as a reference.
- FIG. 87 is a diagram showing the configuration of the data creation device.
- the data creation device 8701 has the same structure as the data creation device 2601 as a basic structure, but a base view video encoder 8705 and a dependent view video encoder 8709 are different, and a high-quality 2D compatible video frame memory 8711. Has been added.
- the base-view video encoder 8705 compresses and encodes the base-view video stream in a video codec (for example, MPEG-4 MVC) of the multi-view video stream, and is 2D compatible with the left-eye original image video according to the 2D-compatible video encoding information 2606.
- a base-view video stream is created by compressing and encoding the difference video with the decoded picture of the video stream.
- the base-view video encoder 8705 outputs base-view video encoding information 2607, and outputs a composite video of the decoded picture of the 2D-compatible video stream and the difference video to the high-quality 2D-compatible video frame memory 8711.
- the base view video encoder 8705 first generates a difference video between the left-eye original video and the decoded picture of the 2D compatible video stream.
- the method for generating the difference video has already been shown in the upper part of FIG.
- the difference video is compression-encoded by determining the encoding method of the base-view video stream with reference to the 2D compatible video encoding information 2606.
- the video compression encoding method is the same as that described in the base-view video encoder 2605.
- the base-view video encoder 8705 outputs the compression-encoded video stream as a base-view video stream, and combines the difference video obtained by decoding the base-view video stream and the video obtained by synthesizing the decoded picture of the 2D compatible video stream.
- the video is output to a high-quality 2D compatible video frame memory.
- the synthesizing method is the method described in the lower part of FIG.
- the dependent-view video encoder 8709 includes a compression encoding function of the dependent-view video stream in the video codec (eg, MPEG-4 MVC) of the multi-view video stream, and the right-eye according to the base-view video encoding information 2607. Compresses the original video and outputs a dependent-view video stream. In this compression processing, the dependent-view video encoder 2609 compresses the decoded picture in the high-quality 2D compatible video frame memory stored in the high-quality 2D compatible video frame memory 8711 using inter-view reference. Encode.
- the compression encoding method is the same as that described in the dependent view video encoder 2609.
- FIG. 88 shows the structure of the playback device.
- the playback device 8823 has the same structure as the playback device 2823 as a basic structure, but a multi-view video decoder 2817 is replaced with a multi-view video decoder 8817, and a synthesis unit 8824 is added.
- the base-view video encoder 8817 transfers the decoded picture to the synthesis unit 8824 when decoding the base-view video stream.
- the synthesizing unit 8824 performs a synthesizing process with the decoded picture of the base-view video stream and the decoded picture of the 2D compatible video stream of the same DTS / PTS.
- the synthesizing method is as described in FIG.
- the decoded picture of the base-view video stream having the same PTS in the inter-view reference buffer 2816 is replaced with the decoded picture of the 2D-compatible video stream that has been synthesized to improve the image quality.
- the decoded picture of the 2D-compatible video stream with high image quality is referred to, so that the decoded picture of the dependent-view video stream also has high image quality. Will be converted.
- the configuration of the data creation device and the playback device may be further simplified as follows.
- This modified example relates to the dependent-view video stream, and does not refer to the decoded picture of the 2D compatible video stream that has been synthesized with the difference video and improved in quality as in the above modified example (23). Reference the decoded picture of the previous 2D compatible video stream. In this case, as in the modified example (23), although the high quality of the dependent view stream cannot be achieved, the processing becomes easy.
- FIG. 97 is a diagram showing a configuration of a data creation device according to this modification.
- the data creation device 9701 has the same structure as the data creation device 2601 as the basic structure, but the base view video encoder 9705 and the dependent view video encoder 9709 are different.
- the base-view video encoder 9705 has a compression encoding function of the base-view video stream in the video codec (for example, MPEG-4 MVC) of the multi-view video stream, and is 2D compatible with the left-eye original video image according to the 2D-compatible video encoding information.
- a base-view video stream is created by compressing and encoding the difference video with the decoded picture of the video stream, and the base-view video encoding information and the decoded picture of the 2D-compatible video stream are output.
- the base-view video encoder 9705 first generates a differential video between the left-eye original video and the decoded picture of the 2D compatible video stream.
- the difference video generation method is the method shown in the upper part of FIG.
- the encoding method of the base view video stream is determined based on the 2D compatible video encoding information 2606, and the difference video is compression encoded.
- the video compression encoding method is the same as that described in the base-view video encoder 2605.
- the base view video encoder 8705 outputs the compression-encoded video stream as a base view video stream.
- the dependent-view video encoder 9709 has a compression encoding function of the dependent-view video stream in the video codec (for example, MPEG-4 MVC) of the multi-view video stream, and the right-eye according to the base-view video encoding information 2607 Compresses the original video and outputs a dependent-view video stream.
- the dependent-view video encoder 9709 performs compression encoding using a decoded picture of 2D compatible video for inter-view reference.
- the compression encoding method is the same as that described in the dependent view video encoder 2609.
- FIG. 98 is a diagram showing a configuration of a playback apparatus according to the present modification.
- the playback device 9823 has the same structure as the playback device 2823 as a basic structure, but a synthesis unit 9824 is added.
- the combining unit 9824 combines the decoded picture of the base-view video stream output to the second plane 2820 and the decoded picture of the 2D-compatible video stream of the same DTS / PTS output to the first plane 2808 at the PTS timing. And outputs a 2D compatible video with high image quality by this composition processing.
- the right-eye image is stored in the dependent-view video stream.
- the present invention is not limited to this, and a depth map image may be stored. With this configuration, 3D video can be reproduced using a depth map.
- FIG. 92 is a diagram showing a configuration of a video stream according to this modification.
- the depth map video may be compressed and encoded into the base view video stream.
- FIG. 93 is a diagram showing a configuration of a video stream when a depth map video is compression-encoded into a base-view video stream according to this modification.
- the playback apparatus that realizes 3D video playback using depth map video is configured to decode only depth map video, while playback that realizes 3D video playback using left-eye video and right-eye video.
- the apparatus can be configured to realize the video playback described in the first embodiment, and can support 3D video playback of both playback apparatuses with one data stream.
- the difference video may be stored in the dependent-view video stream.
- high-quality 2D video reproduction is performed by decoding each video stream as described in the first embodiment and synthesizing the decoded differential video and the decoded picture of the 2D compatible video stream.
- Supplementary explanation> ⁇ Description of video compression technology> ⁇ 2D video compression technology> MPEG-2 and MPEG-4 AVC (compression encoding system based on MPEG-4 MVC format), which is a standard for compressing and encoding 2D video used in the data creation device and playback device according to the present embodiment The encoding method will be briefly described.
- data amount compression encoding is performed using redundancy in the spatial direction and temporal direction of moving images.
- Inter-picture prediction coding is used as a compression coding method using redundancy.
- inter-picture predictive coding when a certain picture is coded, a picture that is forward or backward in display time order is used as a reference picture. Then, the amount of motion from the reference picture is detected, and the difference between the picture subjected to motion compensation and the picture to be coded is compression-coded.
- FIG. 1 shows a reference relationship between pictures in a video stream.
- picture P3 is compression-encoded with reference to picture I0
- picture B1 and picture B2 are compression-encoded with reference to picture I0 and picture P3, respectively.
- a right-eye image (R image) and a left-eye image (L image) are prepared, and only the image corresponding to each eye for the user is displayed for each eye.
- the stereoscopic display is realized by performing display controlled so as to be reflected on the screen.
- a moving image composed of right-eye images is called a left-view video
- a moving image composed of left-eye images is called a right-view video.
- FIG. 13 is a diagram showing an example of a stereoscopic image display.
- the example which displayed the image for left eyes of the skeleton of the dinosaur which is a target object, and the image for right eyes of the skeleton of the dinosaur which is a target object is shown. If 3D glasses are used to repeat the light and left eye light transmission and shading, the left and right scenes are overlapped by the afterimage reaction of the eyes in the user's brain, and a stereoscopic image exists on the extension line in the center of the face Then you can recognize.
- 3D video systems that compress and encode left-view video and right-view video include a frame compatible system and a multi-view encoding system.
- the frame compatible method is a method of performing compression coding by synthesizing each picture corresponding to an image showing a view at the same time of the left view video and the right view video and then synthesizing them into one picture.
- each picture corresponding to an image showing a view at the same time of the left-view video and the right-view video is compressed in half in the horizontal direction and then arranged in the left and right to form one picture.
- a video stream is generated by compressing and encoding a moving image based on a synthesized picture by a 2D video compression encoding method (for example, MPEG-2).
- the video stream is decoded based on the same compression encoding method as that for generating the video stream.
- Each decoded picture is divided into left and right images, and each of the divided images is expanded in the horizontal direction twice to obtain corresponding pictures of the left view video and the right view video.
- L image left-view video picture
- R image right-view video picture
- the multi-view encoding method is a method for compressing and encoding the left-view video and the right-view video picture as separate pictures without combining them into one picture.
- the multi-view encoding method is a method for compressing and encoding the left-view video and the right-view video picture as separate pictures without combining them into one picture.
- FIG. 2 is a diagram showing an encoding method in the MPEG-4 MVC format, which is a multi-view encoding method.
- the MPEG-4 MVC format video stream is processed simultaneously with the base-view video stream and the base-view video stream that can be played back by a playback device that plays back the conventional MPEG-4 AVC-format video stream.
- This is a configuration including a dependent-view video stream that realizes playback.
- the base-view video stream is compression-encoded using inter-picture predictive coding using only the redundancy with the video of the same viewpoint without referring to the video of another viewpoint. Is done.
- the dependent-view video stream is compression-encoded by inter-picture predictive encoding using redundancy with video from another viewpoint in addition to inter-picture predictive encoding using reference of the same viewpoint video. .
- the picture of the dependent-view video stream is compression-coded with reference to a picture showing a view at the same time in the base-view video stream.
- the picture P0 that is the first P picture of the dependent-view video stream refers to the picture I0 that is the I-picture of the base-view video stream.
- a picture B1 that is a B picture of the dependent-view video stream refers to a picture Br1 that is a Br picture of the base-view video stream.
- a picture P3 that is a second P picture of the dependent-view video stream indicates that a picture P3 that is a P-picture of the base-view video stream is referred to.
- the base-view video stream does not refer to the picture of the dependent-view video stream, it can be decoded and reproduced only by this video stream.
- the dependent-view video stream since the dependent-view video stream is decoded with reference to the base-view video stream, it cannot be played back alone.
- the dependent-view stream performs inter-picture prediction encoding using a base-view picture of another viewpoint indicating a view at the same time.
- the right-eye image and the left-eye image at the same time are similar (highly correlated), and the difference is compressed and encoded, so the dependent-view video stream has more data than the base-view video stream. Can be greatly reduced.
- MPEG-2 transport stream is a standard for multiplexing and transmitting various streams such as video and audio. It is standardized in ISO / IEC13818-1 and ITU-T recommendation H222.0.
- FIG. 6 is a diagram showing the structure of a digital stream in the MPEG-2 transport stream format.
- the transport stream 513 is obtained by multiplexing a video TS (Tranport Stream) packet 503, an audio TS packet 506, a subtitle stream TS packet 509, and the like.
- the video TS packet 503 stores the main video of the program
- the audio TS packet 506 stores the main audio portion and sub audio of the program
- the TS packet 509 of the subtitle stream stores the subtitle information of the program.
- the video frame sequence 501 is compression encoded using a video encoding method such as MPEG-2, MPEG-4 AVC.
- the audio frame sequence 504 is compression-encoded by an audio encoding method such as Dolby AC-3, MPEG-2 AAC, MPEG-4 AAC, HE-AAC.
- Each stream included in the transport stream is identified by a stream identification ID called PID.
- PID stream identification ID
- the playback apparatus can extract the stream to be processed.
- Information on the correspondence between PID and stream is stored in a descriptor of a PMT packet described below.
- a video stream 501 composed of a plurality of video frames and an audio stream 504 composed of a plurality of audio frames are converted into PES packet sequences 502 and 505, respectively.
- the PES packet sequences 502 and 505 are converted into TS packets 503 and 506, respectively.
- the data of the subtitle stream 507 is converted into a PES packet sequence 508 and further converted into a TS packet 509.
- the MPEG-2 transport stream 513 is configured by multiplexing these TS packets into one stream. The PES packet and TS packet will be described later.
- the video stream is data having a hierarchical structure as shown in FIG.
- a video stream is composed of a plurality of GOPs, and by using this as a basic unit of compression encoding processing, editing of a moving image and random access are possible.
- GOP is composed of one or more video access units.
- the video access unit is a unit for storing compression-encoded data of a picture, and stores data of one frame in the case of a frame structure and one field in the case of a field structure.
- Each video access unit includes an AU identification code, a sequence header, a picture header, supplementary data, a compressed picture, padding data, a sequence end code, and a stream end code.
- each data is stored in units called NAL units.
- AU identification code is a start code indicating the head of the access unit.
- the sequence header is a header storing common information in a playback sequence composed of a plurality of video access units, and stores information such as resolution, frame rate, aspect ratio, and bit rate.
- the picture header is a header that stores information such as the encoding method of the entire picture.
- Supplementary data is additional information that is not essential for decoding a compressed picture, and stores, for example, closed caption character information or GOP structure information that is displayed on a TV in synchronization with video.
- Compressed picture data is stored in the compressed picture.
- -Padding data stores data for formatting. For example, it is used as stuffing data for maintaining a predetermined bit rate.
- the sequence end code is data indicating the end of the playback sequence.
- the stream end code is data indicating the end of the bit stream.
- the data structure of the AU identification code, sequence header, picture header, supplemental data, compressed picture, padding data, sequence end code, and stream end code differs depending on the video encoding method.
- the AU identification code is an AU delimiter (Access Unit Delimiter)
- the sequence header is SPS (Sequence Parameter Set)
- the picture header is PPS (Picture Parameter Set)
- the compressed picture is a plurality of slices
- Supplementary data corresponds to SEI (Supplemental Enhancement Information)
- padding data corresponds to FillerData
- sequence end code corresponds to End of Sequence
- stream end code corresponds to End of Stream.
- the sequence header is sequence_Header, sequence_extension, group_of_picture_header, the picture header is picture_header, the picture_coding_extension, the compressed code is a sequence_decode, and the supplement_data is a sequence_decode.
- each attribute information is not always necessary.
- the sequence header is necessary only in the video access unit at the head of the GOP, and can be encoded without being added to other video access units.
- the picture header can be referred to that of the previous video access unit in the code order, and the picture header can be omitted in its own video access unit.
- I picture data is stored as a compressed picture, and an AU identification code, a sequence header, a picture header, and a compressed picture are always stored. Supplementary data, padding data, sequence end code, and stream end code are stored as needed.
- the AU identification code and the compressed picture are always stored in video access units other than the GOP head, and supplementary data, padding data, a sequence end code, and a stream end code are appropriately stored as necessary.
- FIG. 10 is a diagram showing how a video stream is stored in a PES packet sequence.
- the 10 shows the video frame sequence of the video stream.
- the second level shows a PES packet sequence.
- a plurality of Video Presentation Units in the video stream are divided into each picture, and stored in the payload of the PES packet. .
- Each PES packet has a PES header, and a PTS that is a picture display time and a DTS that is a picture decoding time are stored in the PES header.
- FIG. 11 is a diagram showing the data structure of TS packets constituting the transport stream.
- the TS packet is a 188-byte fixed-length packet composed of a 4-byte TS header, an adaptation field, and a TS payload.
- the TS header is configured to include transport_priority, PID, adaptation_field_control.
- the PID is an ID for identifying a stream multiplexed in the transport stream as described above.
- Transport_priority is information for identifying the type of packet in TS packets having the same PID.
- Adaptation_field_control is information for controlling the configuration of the adaptation field and the TS payload. There are cases where only one of the adaptation field and the TS payload exists or both, and adaptation_field_control is information indicating the presence / absence thereof.
- adaptation_field_control value When the adaptation_field_control value is “1”, only the TS payload is present. When the adaptation_field_control value is “2”, only the adaptation field is present. When the adaptation_field_control value is “3”, the payload is “3”. Indicates that both adaptation fields are present.
- the adaptation field is a storage area for storing information such as PCR (Program Clock Reference) and stuffing data for making the TS packet a fixed length of 188 bytes.
- PCR Program Clock Reference
- stuffing data for making the TS packet a fixed length of 188 bytes.
- a PES packet is divided and stored in the TS payload.
- TS packets included in the transport stream include packets such as PAT (Program Association Table), PMT, and PCR in addition to packets of each stream such as video, audio, and caption. These packets are called PSI (Program Specific Information).
- PAT indicates what the PID of the PMT used in the transport stream is, and the PID of the PAT itself is “0”.
- FIG. 12 is a diagram showing the data structure of the PMT.
- the PMT includes a PMT header, various descriptors related to the transport stream, and stream information related to each stream such as video / audio / subtitles included in the transport stream.
- copy control information for instructing permission / non-permission of copying of each stream such as video / audio is recorded.
- the stream information regarding each stream includes a stream descriptor in which a stream type, a stream PID, and stream attribute information (for example, a frame rate and an aspect ratio) are described in order to identify the compression encoding method of the stream.
- a stream descriptor in which a stream type, a stream PID, and stream attribute information (for example, a frame rate and an aspect ratio) are described in order to identify the compression encoding method of the stream.
- the PCR In order to synchronize the arrival time of the TS packet at the decoder and the STC (System Time Clock) which is the time axis of the PTS and DTS, the PCR has an STC time corresponding to the time when the PCR packet is transferred to the decoder. Have information.
- STC System Time Clock
- an area to be actually displayed can be designated as a cropping area from among the compression-coded frame areas.
- the frame_cropping information includes differences between the upper line, the lower line, the left line, and the right line of the cropping area and the upper line, the underline, the left line, and the right line of the compression-coded frame area. Is specified as the crop amount of the top, bottom, left and right.
- frame_cropping_flag is set to “1”
- frame_crop_top_offset, frame_crop_left_offset, frame_crop_right, and lower_crop_crop_right crop_crop
- the vertical and horizontal sizes of the cropping area (display_horizontal_size of sequence_display_extension, display_vertical_size), and the difference between the center of the compression-coded frame area and the center of the cropping area
- the cropping area is specified using information (frame_center_horizontal_offset, frame_center_vertical_offset of picture_display_extension).
- an aspect ratio is set as scaling information indicating a scaling method when the cropping area is actually displayed on a television or the like.
- the playback device uses the aspect ratio information to up-convert the cropping area for display.
- aspect ratio information (aspect_ratio_idc) is stored in the SPS as the scaling information.
- the aspect ratio is specified as 4: 3.
- sequence_header there is attribute information for storing aspect ratio information (aspect_ratio_information) called sequence_header, and the same processing as described above can be realized by appropriately setting the value of this attribute information.
- FIG. 15 is a diagram showing an example of the internal configuration of a video stream in the MPEG-4 MVC format.
- the picture of the right-view video stream is compression-coded with reference to the picture at the same display time of the left-view video stream.
- the pictures P1 and P2 of the right-view video stream refer to the pictures I1 and P2 of the left-view video stream, respectively.
- the pictures B3, B4, B6, and B7 of the right-view video stream are respectively pictures Br3 and Br4 of the left-view video stream. , Br6, Br7 with reference to compression coding.
- the second tier in the figure shows the internal structure of the left view video stream.
- This left-view video stream includes pictures I1, P2, Br3, Br4, P5, Br6, Br7, and P9. These pictures are decoded in the order of time set in the DTS.
- the first row shows the left eye image displayed on the display or the like.
- the left-eye image is the second-order decoded picture I1, P2, Br3, Br4, P5, Br6, Br7, P9 in the order of time set in the PTS, that is, I1, Br3, Br4, P2, Br6, Displayed in the order of Br7 and P5.
- the fourth level shows the internal structure of the right-view video stream.
- This right-view video stream includes pictures P1, P2, B3, B4, P5, B6, B7, and P8. These pictures are decoded in the order of time set in the DTS.
- the third row shows a right eye image displayed on a display or the like.
- the right-eye image has the decoded pictures P1, P2, B3, B4, P5, B6, B7, and P8 in the fourth stage in the order set in the PTS, that is, P1, B3, B4, P2, B6, Displayed in the order of B7 and P5.
- display of either the left-eye image or the right-eye image in which the same value is assigned to the PTS is delayed by half the interval until the next PTS time.
- the fifth row shows how the state of the 3D glasses 200 is changed. As shown in the fifth row, the right-eye shutter is closed when viewing the left-eye image, and the left-eye shutter is closed when viewing the right-eye image.
- FIG. 17 shows the configuration of the video access unit for each picture of the base-view video stream and each picture of the dependent-view video stream. As described above, as shown in the upper part of FIG. 17, in the base-view video stream, each picture is configured as one video access unit.
- the dependent-view video stream also has one video access unit in the same manner, but the data structure is different from the video access unit of the base-view video stream.
- the 3D video access unit 1701 is composed of the video access unit of the base-view video stream and the video access unit of the dependent-view video stream having the same PTS value.
- the playback device performs decoding in units of this 3D video access unit.
- FIG. 18 shows an example of the relationship between the PTS and the DTS assigned to each video access unit of the base-view video stream and the dependent-view video stream in the video stream.
- the base-view video stream picture and the dependent-view video stream picture storing the parallax images indicating the same time view are set so that the DTS and the PTS have the same value.
- the playback device that decodes the pictures of the base-view video stream and the dependent-view video stream can perform decoding and display in units of 3D video access units.
- FIG. 19 shows the GOP configuration of the base view video stream and the dependent view video stream.
- the GOP structure of the base view video stream is the same as that of the conventional video stream, and is composed of a plurality of video access units.
- the dependent view video stream is composed of a plurality of dependent GOPs.
- the leading picture of the dependent GOP is a picture displayed as a pair with the I picture at the GOP head of the base-view video stream when playing back 3D video, and has the same value as the PTS of the I picture at the GOP head of the base-view video stream This picture is assigned a PTS.
- FIG. 20 shows the data structure of the video access unit included in the dependent GOP.
- the dependent GOP head video access unit stores, as a compressed picture, picture data displayed at the same time as the GOP head I picture of the base-view video stream, and a sub AU identification code and sub sequence.
- a header, a picture header, and a compressed picture are always stored.
- Supplementary data, padding data, sequence end code, and stream end code are stored as appropriate.
- the sub AU identification code is a start code indicating the head of the access unit.
- the sub-sequence header is a header that stores common information in a playback sequence composed of a plurality of video access units, and stores information such as resolution, frame rate, aspect ratio, and bit rate.
- the frame rate, resolution, and aspect ratio values of the sub-sequence header are the same as the frame rate, resolution, and aspect ratio of the sequence header included in the video access unit at the GOP head of the corresponding base-view video stream.
- a video access unit other than the head of the GOP always stores a sub AU identification code and a compressed picture. Supplementary data, padding data, sequence end code, and stream end code are stored as appropriate.
- Second Embodiment> ⁇ 2-1. Overview>
- inter-view reference is realized between streams in which video is compressed and encoded with different codecs, thereby reducing the multi-view video stream of 3D video while maintaining playback compatibility with conventional 2D video. Configured with bit rate.
- the conventional 2D video can be improved in quality while maintaining playback compatibility with the conventional 2D video.
- a high-quality video is transmitted as a multi-view video stream configured at a low bit rate.
- “high image quality” refers to image quality closer to the original image, for example, image quality with less compression distortion such as block noise that may occur during video compression.
- an image encoded at a high bit rate is used as an example of an image with high image quality.
- the image quality depends on the encoding bit rate. The higher the bit rate, the higher the quality of the image that is closer to the original image. When the bit rate is low, the video cannot be compressed or encoded in scenes with intense motion. This is because block noise may be mixed in.
- FIG. 48 is a diagram for explaining data included in the transport stream according to the present embodiment.
- the transport stream in the present embodiment is a 2D compatible video stream, a base-view video stream of a multi-view video stream, and a dependent-view video stream, each of which is made into a PES packet, divided into TS packets, and multiplexed. .
- the 2D compatible video stream is a video stream configured in a format that can be reproduced as 2D video by a 2D video playback device as described with reference to FIG.
- the 2D compatible video stream is a video stream encoded with an MPEG-2 video codec.
- the multi-view video stream is a video stream encoded with a codec that realizes inter-view reference, and here, a high-quality original video image is compression-encoded with an MPEG-4 MVC video codec. .
- a 2D compatible video stream is generated by compressing and encoding an original image of normal quality.
- black video is generated by compression encoding at a low bit rate according to the MPEG-4 MVC video codec.
- the dependent-view video stream is generated by compressing and encoding a high-quality original picture.
- the dependent-view video stream is compressed using inter-view reference.
- the picture referred to as the inter-view reference is not a picture at the same display time (PTS is the same) of the base-view video stream, but a decoded picture at the same display time (PTS is the same) of the 2D compatible video stream.
- the dependent-view video stream compresses the difference video between the video obtained by compressing and encoding the original video with higher image quality than the MPEG-2 video codec and the video compressed and encoded according to the MPEG-2 video codec. Is generated.
- the decoded picture (black image) of the base-view video stream is replaced with the decoded picture (restored image of MPEG-2 video) of the 2D compatible video stream at the same display time, and the replaced picture (MPEG- With reference to (decoded image of two videos), a picture (high-quality image) at the same display time of the dependent-view video stream is decoded.
- the base-view video stream is a compression-encoded black image
- the dependent-view video stream is a “decoded picture of a 2D-compatible video stream” and a “picture of high-quality video”. Since only the difference is compressed, an image obtained by improving the image quality of a conventional 2D image can be configured at a low bit rate.
- FIG. 50 is a diagram for explaining a PMT packet included in the transport stream.
- signaling information for decoding high-quality video is included in system packets such as PMT packets.
- the signaling information includes a high-quality information descriptor for performing signaling such as the relationship between video streams and the start and end of high-quality video reproduction, and a high-quality stream descriptor set for each video stream.
- a high-quality information descriptor for performing signaling such as the relationship between video streams and the start and end of high-quality video reproduction
- a high-quality stream descriptor set for each video stream for each video stream.
- the high image quality information descriptor includes a high image quality playback method, a 2D compatible video PID, a base view video PID, and a dependent view video PID.
- the high-quality playback method is information for signaling the playback method of the high-quality playback device.
- the high-quality playback method is 0, 2D video playback using 2D-compatible video is shown.
- the high-quality playback system is 1, high-quality video playback using 2D-compatible video and dependent-view video (that is, the high level described in the present embodiment).
- 2 indicates high-quality video playback using base-view video stream and dependent-view video
- 3 indicates high-quality video playback using base-view video. Indicates that
- the playback device plays back 2D video only of the 2D compatible video stream.
- the value is 1
- high-quality video playback is performed on the 2D-compatible video stream, base-view video stream, and dependent-view video stream using the playback method described in this embodiment.
- the value is 2
- the 2D compatible video stream and the multi-view video stream constituting the high-quality video are not in a reference relationship because different videos are generated by compression encoding.
- the base-view video stream and the dependent-view video stream are video streams compressed with a normal multi-view codec. Therefore, the playback device plays back the multi-view video stream with high quality video.
- the value is 3
- the playback device plays back the base view video stream with high quality video.
- FIG. 53 is a diagram illustrating the structure of the high quality stream descriptor.
- the high image quality stream descriptor includes a base view video type, a reference destination type, and a referenced type.
- the base view video type indicates what video is encoded in the base view video stream. When the value is 0, it indicates that the video that is the base of the high-quality video is compression-encoded. When the value is 1, an attached video such as a black video is included and replaced with a 2D compatible video stream. This indicates that a dummy video that is not used for output to the video is compressed and encoded.
- the reference destination type indicates the type of the video stream that the dependent-view video stream is the reference destination of the inter-view reference.
- a value of 0 indicates that a picture of the base-view video stream is used as an inter-view reference, and a value of 1 indicates that a picture of a 2D compatible video stream is referenced between views.
- the referenced type indicates whether the corresponding video stream is referenced between views. If it is not referenced, the process of inter-view reference can be skipped, so the implementation can be simplified accordingly.
- PTS, DTS and GOP> The relationship between the display time (PTS), decoding time (DTS), and picture type assigned to each video access unit of the 2D compatible video stream, base-view video stream, and dependent-view video stream in the transport stream is described in the first embodiment.
- the 2D compatible video stream, the base view video stream, and the dependent view video stream have the same relationship.
- the GOP configuration of the 2D compatible video stream, the base view video stream, and the dependent view video stream is the same as the GOP configuration of the 2D compatible video stream, the base view video stream, and the dependent view video stream of the first embodiment. is there.
- the normal playback device is a device that can play back only a conventional 2D compatible video stream.
- the normal playback device can play back a stream with a bit rate of up to 10 Mbps.
- the normal playback device is already on the market, and is supposed to play back a stream distributed by a broadcast wave or the like.
- the high-quality playback device is a playback device according to the present embodiment, and has a decoding function for a multi-view video stream as well as a 2D compatible video stream. In the case where these two types of playback devices coexist, the transport stream having the configuration described in FIG. 48 is broadcast.
- the normal playback device decodes the 2D compatible video stream in the transport stream and plays back normal 2D video.
- the high-quality playback device decodes both the 2D compatible video stream and the multi-view video stream in the transport stream.
- the high-quality playback device replaces the decoded picture of the base-view video stream with the decoded picture of the 2D compatible video stream at the same display time. Then, the picture of the dependent-view video stream is decoded with reference to the decoded picture of the 2D compatible video stream at the same display time, which is the replaced picture.
- the picture obtained by decoding the dependent-view video stream is obtained by adding a difference between the picture of the high-definition video and the picture of the 2D-compatible video stream to the picture of the 2D-compatible video stream, that is, the picture of the high-quality video. . Therefore, if a dependent-view video stream is output, it is possible to reproduce high-quality video compared to the decoding result of a 2D compatible video stream.
- the base-view video stream is referred to for information included in the stream such as PTS, but the picture itself is not referred to, so the compression rate such as video composed of a black image is high.
- the bit rate can be greatly reduced.
- FIG. 54 is a diagram showing a configuration of a data creation device 5401 as a video encoding device according to the present embodiment.
- the data creation device 5401 has basically the same configuration as the data creation device 2601 of the first embodiment shown in FIG. However, since the input data is different, in FIG. 54, the “left-eye original image” in FIG. 26 is replaced with a 2D “original image image”, and the “right-eye original image” is replaced with the same 2D “original image image”. The following description will focus on differences from the data creation device 2601 of the first embodiment.
- the data creation device 5401 is different from the data creation device 2601 in the configuration of the dependent-view video encoder 5409 in the extended multi-view video encoder 5404.
- the dependent-view video encoder 5409 encodes a high-definition original picture image as an input instead of one of the videos (left-eye or right-eye original picture video) constituting the 3D video.
- the dependent-view video encoder 5409 performs the compression process by using the decoded picture in the 2D compatible video frame memory 2608 (the picture obtained by decoding the 2D compatible video stream) for inter-view reference.
- FIG. 55 is a flowchart showing data creation processing by the data creation device 5401 configured as described above.
- FIG. 55 The data creation flow shown in FIG. 55 is approximately the same as the data creation flow of the data creation device 2601 described with reference to FIG. 27 in the first embodiment.
- S2703 and S2706 in FIG. 27 are replaced with S5503 and S5506, respectively.
- the “left-eye original image” is replaced with a 2D “original image image”
- the “right-eye original image” is replaced with the same 2D “original image image”.
- FIG. 56 is a diagram showing the configuration of the playback device 5623 according to the present embodiment.
- the flowchart shown in FIG. 57 is almost the same as the 3D video decoding and output processing flow by the 3D video playback device 2823 described with reference to FIG. 30 of the first embodiment.
- the part of S3005 which is an output process is different from FIG. 30, and in FIG. 57, it is replaced with the following S5705.
- the present embodiment aims to improve the quality of 2D video images, and it is not necessary to output the 2D compatible video stream to the plane. Therefore, only the decoded pictures of the dependent-view video stream are processed at the PTS timing. 1 plane 2808 is output. ⁇ 2-5. Supplementary explanation about effects>
- the effect in the present embodiment will be supplementarily described with reference to FIG.
- the normal playback device in the figure is a playback device that can play back only a 2D compatible video stream.
- the bit rate that can be played back by the normal playback device is up to 10 Mbps.
- the normal playback device is already in the market and plays back a stream distributed by broadcast waves or the like.
- the high-quality playback device is a playback device that can play back a high-quality video stream in which a video with a higher quality than the 2D compatible video stream is encoded.
- the 2D compatible video stream is 10 Mbps
- the bit rate of the high-quality video stream obtained by compressing the same video with the same codec is 15 Mbps.
- the lower part of FIG. 47 shows an example in which the multi-view encoding method is used in order to solve the problem shown in the upper part of FIG. 47 and reduce the necessary transfer bandwidth.
- the 2D compatible video stream in the figure is encoded as a base-view video stream.
- the extended video stream is a dependent-view video stream, and is compressed and encoded using each picture of the base-view video stream as a reference image.
- the extended video stream compresses the difference between the original picture and the decoding result of the 2D compatible video stream.
- the multi-view video stream can realize a high-quality video at a low bit rate as compared with the case where the 2D compatible video stream and the high-quality video stream are combined as shown in the upper part of FIG.
- the 2D compatible video stream is 10 Mbps
- the extended video stream is compressed with the 2D compatible video stream as a reference image to 5 Mbps.
- the high-quality playback device can decode and play back high-quality video by decoding the 2D-compatible video stream constituting the multi-view video stream and decoding the extended video stream while referring to this.
- both video streams must be compressed by a multi-view video encoding method such as MPEG-4 MVC.
- the playback device can easily change the playback mode.
- the lower part of FIG. 58 shows the setting of the inter-codec reference switch and the plane selector for the example of the high image quality reproduction method of FIG.
- the high-quality playback method is “0”
- the inter-codec reference switch is turned off in the high-quality playback device, and the plane selector selects the first plane as 2D video.
- the high-quality playback method is “1”
- the inter-codec reference switch is turned on, and the plane selector selects the third plane as 2D video.
- the high-quality playback method is “2”
- the inter-codec reference switch is turned OFF, and the plane selector selects the third plane as 2D video.
- inter-view reference is realized between streams in which video is compression-coded with different codecs, thereby reducing the multi-view video stream of 3D video while maintaining playback compatibility with conventional 2D video. Configured with bit rate.
- inter-view reference is realized between streams in which video is compressed and encoded with different codecs, so that the conventional 2D video can be improved in quality while maintaining compatibility with the conventional 2D video.
- video with improved color reproducibility high gradation
- the improvement of color reproducibility means adding more accurate gradation information to color information compressed by a conventional video codec.
- color information is expressed in 8-bit color. In the case of RGB colors, display of about 16.77 million colors is possible with a total of 24 bits assigned 8 bits (0 to 255 gradations) to each of RGB.
- the user may determine that the image quality of the video expressed with this color gradation is poor.
- an image expressed with a 12-bit color gradation is used as an example of the adjusted image.
- FIG. 61 is a diagram showing an outline of an encoding procedure and a decoding procedure for increasing the gradation.
- the upper part of FIG. 61 shows an outline of a procedure for generating a 2D compatible video stream and an extended video stream related to encoding.
- the high gradation video is an original image with a color gradation of 12 bits.
- the upper 8 bits of the 12-bit original picture are taken out and compressed into a 2D compatible video stream.
- the 2D compatible video stream is decoded, and a difference value between the video shifted to the upper 4 bits (6101) and the original video of 12 bits is obtained (6102).
- the difference value is information in a signed 7-bit range (-127 to +128), and is added to this to form a positive number (6103), which is compressed and encoded into an extended video stream.
- 61 shows the outline of the decoding procedure of the 2D compatible video stream and the extended video stream described above.
- the 2D compatible video stream is decoded, and the color value of each pixel is shifted up by 4 bits (6104).
- the extended video stream is decoded and subtracted 128 from the color value of each pixel to convert it into signed information (6104). Then, these images are added (6106) to output a high gradation image.
- video can be encoded / decoded using a conventional 8-bit encoder / decoder.
- the extended video stream is a difference video between a 12-bit video and an 8-bit video (correctly, a video obtained by decoding a 2D compatible video stream compressed and encoded with 8-bit video), and is a video with little change. Therefore, compression encoding can be performed with high efficiency.
- FIG. 62 is a diagram for explaining data included in the transport stream according to the present embodiment.
- the transport stream in the present embodiment is a 2D compatible video stream, a base-view video stream of a multi-view video stream, and a dependent-view video stream, each of which is made into a PES packet, divided into TS packets, and multiplexed. .
- the 2D compatible video stream is a video stream configured in a format that can be reproduced as 2D video by a 2D video playback device as described with reference to FIG.
- the 2D compatible video stream is a video stream encoded with an MPEG-2 video codec.
- the multi-view video stream is a video stream encoded with a codec that realizes inter-view reference, and here is compressed with the MPEG-4 MVC video codec.
- the 2D compatible video stream is an original video image that is compression-encoded with 8-bit color gradation according to the MPEG-2 video codec.
- the base-view video stream is a black video that has been compression-encoded at a low bit rate in accordance with the MPEG-4 MVC video codec.
- the dependent-view video stream is obtained by compressing and encoding a differential video between a 12-bit original video and a decoded video (8-bit uncompressed video) of a 2D compatible video stream.
- the difference video is created by the procedure described with reference to FIG.
- this difference video is referred to as “differential gradation video”.
- the differential gradation video is compression-encoded according to the MPEG-4 MVC video codec.
- the dependent-view video stream is compressed using inter-view reference.
- the picture referred to as the inter-view reference is not a picture of the base view video stream at the same display time (the same PTS) but a decoded picture of a 2D compatible video stream at the same display time (the same PTS).
- the decoded picture (black image) of the base-view video stream is replaced with the decoded picture (8-bit uncompressed video) of the 2D compatible video stream at the same display time, and the replaced picture (8-bit uncompressed) With reference to (video), a picture (12-bit original video) at the same display time of the dependent-view video stream is decoded.
- the dependent-view video stream is obtained by compressing and encoding the differential gradation video, so that the bit rate of the stream can be reduced as a whole and the gradation is higher than that of the 2D compatible video stream.
- Video can be provided.
- FIG. 63 is a diagram for explaining a PMT packet included in a transport stream.
- signaling information for decoding high-gradation video is included in a system packet such as a PMT packet.
- the signaling information includes a high gradation information descriptor for performing signaling such as the relationship between each video stream and the start and end of high gradation video reproduction, and a high gradation stream descriptor set for each video stream.
- “high image quality” is set to “high image quality” in the description of the high image quality information descriptor and the high image quality stream descriptor already performed with reference to FIGS. 50, 51, and 53. Since this is replaced with “higher gradation”, further explanation is omitted.
- a field for storing the color gradation size of the high gradation video of the dependent view video stream of the stream to be highly gradation may be added to the gradation enhancement descriptor.
- the high-resolution playback device refers to the value and performs synthesis processing.
- the configuration of the part can be changed.
- the normal playback device is a device that can play back only a 2D compatible video stream.
- the normal playback device can play back a stream with a bit rate of up to 10 Mbps.
- the high gradation playback device has a decoding function of not only a 2D compatible video stream but also a multi-view video stream. In the case where these two types of playback devices exist, it is assumed that the transport stream having the configuration described in FIG. 62 is broadcast.
- the normal playback device decodes the 2D compatible video stream in the transport stream and plays back normal 2D video.
- the high gradation reproduction device is a reproduction device according to the present embodiment, and decodes both the 2D compatible video stream and the multi-view video stream.
- the high gradation playback device converts the decoded picture of the base-view video stream into the 2D compatible video stream at the same display time as described in the lower part of FIG. Replace with decoded picture. Then, the picture of the dependent-view video stream is decoded with reference to the decoded picture of the 2D compatible video stream at the same display time, which is the replaced picture.
- a picture obtained by decoding a dependent-view video stream is a picture obtained by adding a difference between a picture of a high-gradation video and a picture of a 2D-compatible video stream to a picture of a 2D-compatible video stream, that is, a picture of a high-gradation video. . Therefore, if a dependent-view video stream is output, it is possible to reproduce a high-gradation video compared to the decoding result of a 2D compatible video stream.
- the base-view video stream is referred to for information included in the stream such as PTS, but the picture itself is not referred to, so the compression rate such as video composed of a black image is high.
- the bit rate can be greatly reduced.
- FIG. 65 is a diagram showing a configuration of a data creation device 6501 as a video encoding device according to the present embodiment.
- the data creation device 6501 has basically the same configuration as the data creation device 2601 of the first embodiment shown in FIG. However, since the input data is different, in FIG. 65, the “left-eye original image” in FIG. 26 is replaced with a 2D high-gradation “12-bit original image”, and the “right-eye original image” is the same 2D “12-bit original image”. Has been replaced. The following description will focus on differences from the data creation device 2601 of the first embodiment.
- the data creation device 6501 differs from the data creation device 2601 in the configuration of the 2D compatible video encoder 6502 and the dependent-view video encoder 6509 in the extended multi-view video encoder 6504.
- the 2D compatible video encoder 6502 receives the upper 8 bits of the 12-bit original image video as input, and performs compression encoding according to the 2D compatible video codec.
- the codec is MPEG-2 video
- it is compression-encoded into the MPEG-2 video stream format and a 2D compatible video stream is output.
- the dependent-view video encoder 6509 does not input one video (left-eye or right-eye original video) constituting the 3D video, but encodes it using a 12-bit original video as input.
- the video to be compression-encoded is a difference gradation video between a 12-bit original video and an 8-bit compressed video (decoding result of a 2D compatible video stream stored in the 2D compatible video frame memory 2608).
- the difference gradation video is created by the procedure described in the upper part of FIG.
- the dependent-view video encoder 6509 uses the decoded picture in the 2D-compatible video frame memory 2608 (the picture obtained by decoding the 2D-compatible video stream) for inter-view reference and performs the compression process.
- FIG. 66 is a diagram showing a data creation flow by the data creation device 6501 configured as described above.
- the data creation flow shown in FIG. 66 is approximately the same as the data creation flow of the data creation device 2601 described with reference to FIG. 27 in the first embodiment.
- the data creation device 6501 is for increasing the gradation of 2D video.
- S2703 and S2706 in FIG. 27 are replaced with S6603 and S6606, respectively.
- the 2D compatible video encoder 6502 generates 2D compatible video streams for one encoded number.
- the upper 8 bits are compression-encoded according to the codec of the 2D compatible video stream by one encoding number from the Nth frame of the 12-bit original picture image to generate and output a 2D compatible video stream. Then, the process proceeds to S2704.
- the dependent-view video encoder 6509 generates a dependent-view video stream for one encoded number of sheets. Based on the base-view video encoding information, video attributes, picture structure, etc. are determined, and for reference between views, a 12-bit original picture and an 8-bit compressed picture are used using a picture in the 2D compatible video frame memory 2608.
- the dependent-gradation video from the (decoding result of the 2D compatible video stream stored in the 2D compatible video frame memory 2608) is compressed and encoded by the number of encoded frames from the Nth frame to generate a dependent-view video stream. To do. Then, the process proceeds to S2707. ⁇ 3-4.
- FIG. 67 is a diagram showing the configuration of the playback device 6723 according to the present embodiment.
- the playback device 6723 has basically the same configuration as the 3D video playback device 2823 of the first embodiment shown in FIG. However, since the playback device 6723 aims to increase the gradation of 2D video, in FIG. 67, the plane output from the 2D compatible video decoder and the plane output from the extended multi-view video decoder are combined in FIG. A compositing processing unit 6724 is added.
- the composition processing unit 6724 performs composition processing of the first plane and the second plane.
- the synthesis process follows the procedure described in the lower part of FIG. First, the composition processing unit 6724 shifts the color value of each pixel to the upper 4 bits for the picture of the decoded 2D compatible video stream stored in the first plane 2808. Next, the picture of the decoded dependent video stream stored in the second plane 2820 is decoded, and the color value of each pixel is subtracted by 128 to convert it into signed information. Then, these images are added to output a high gradation image. ⁇ 3-4-2. Operation>
- FIG. 68 is a diagram showing a flow of decoding processing and output processing in high-gradation video playback by the playback device 6723 configured as described above.
- the playback device 6723 outputs the decoded picture of the 2D-compatible video stream to the first plane 2808 at the PTS timing, and outputs the decoded picture of the dependent-view video stream to the second plane 2820 at the PTS timing. Then, the composition processing unit 6724 performs composition processing, and a high gradation image is output. Since the method of the synthesis process has been described above, the description is omitted. ⁇ 3-5. Supplementary explanation about effects> Hereinafter, the effect in the present embodiment will be supplementarily described with reference to FIG.
- the normal playback device in the figure is a playback device that can play back only a 2D compatible video stream.
- the bit rate that can be played back by the normal playback device is up to 10 Mbps.
- the high gradation reproduction apparatus is a reproduction apparatus capable of decoding and reproducing a video stream in which a video expressed with a higher gradation than the 2D compatible video stream is encoded.
- the 2D compatible video stream is 10 Mbps
- the bit rate of the high gradation video stream obtained by compressing the same video with the same codec is 15 Mbps.
- a 2D compatible video stream and a high gradation It is necessary to multiplex and broadcast both video streams, and a higher transfer bandwidth is required compared to the conventional case where the bit rates of both video streams are added.
- the lower part of FIG. 60 shows an example in which the multi-view encoding method is used in order to solve the problem shown in the upper part of FIG. 60 and reduce the required transfer bandwidth.
- the 2D compatible video stream in the figure is 8-bit color video and is encoded as a base-view video stream.
- the extended video stream is a dependent-view video stream, and is obtained by compressing and encoding a differential video between a 2D compatible video stream and a high-gradation original video (for example, 12-bit color).
- the dependent-view video stream is compression-coded using each picture of the base-view video stream as a reference image.
- both video streams must be compressed by a multi-view video encoding method such as MPEG-4 MVC.
- FIG. 69 is a block diagram showing a configuration of a playback apparatus according to the present modification.
- the playback device shown in FIG. 69 has the same basic structure as the playback device described in FIG. 67, but an inter-codec reference switch and a plane selector are added, and the composition processing unit controls ON / OFF of the function. A switch function has been added.
- the inter-codec reference switch When the inter-codec reference switch is ON, the 2D compatible video decoder is connected to the inter-view reference buffer in the extended multi-view video decoder, and the inter-codec reference switch is changed from the 2D compatible video decoder to the inter-view reference buffer in the extended multi-view video decoder. On the other hand, the data transfer described in S3003 is performed. On the other hand, in the case of OFF, the 2D compatible video decoder and the inter-view reference buffer in the extended multi-view video decoder are not connected and data transfer is not performed.
- the plane selector is a first plane from which a picture is output from the 2D compatible video decoder, a second plane from which a picture of the base-view video stream of the extended multi-view video decoder is output, and a dependent-view video stream of the extended multi-view video decoder. Which plane is to be output to the compositing processing unit is selected for the third plane from which the picture is output.
- FIG. 70 is a diagram showing signaling in the high gradation reproduction method.
- the high gradation reproduction apparatus turns off the inter-codec reference switch, selects the first plane for output as 8-bit video by the plane selector, and turns off the function of the synthesis processing unit. .
- the reproduction apparatus turns on the inter-codec reference switch, the plane selector selects the first plane as an output plane as 8-bit video, and outputs the differential gradation video as the plane.
- the third plane is selected, and the function of the synthesis processing unit is turned on.
- the inter-codec reference switch is turned off, and the plane selector selects the second plane as the plane to be output as 8-bit video, and the third plane as the plane to be output as the differential gradation video. Is selected, and the function of the composition processing unit is turned ON.
- the inter-codec reference switch When the high gradation reproduction method is 3, the inter-codec reference switch is turned off, and the plane selector selects the second plane as a plane to be output as 8-bit video, and turns off the function of the synthesis processing unit. In this way, by switching only the inter-codec reference switch, the plane selector, and the switch of the synthesis processing unit, the playback method in the playback device can be easily changed.
- a 12-bit video is taken up as a high-gradation video.
- the present invention is not limited to this, and the high-gradation video may be a video with a higher gradation than a normal gradation. Further, the effective bit amount in the original image with high gradation may be variably set.
- the effective bit amount may be changed for each scene such that a certain scene has 10 bits and a certain scene has 12 bits.
- the bit shift amount of the 2D compatible video stream is changed to a variable value, and the bit shift is performed so as to extract the effective gradation for the extended video stream. What is necessary is just to comprise.
- the 4-bit left shift of 6101 in FIG. 61 is changed to a 2-bit left shift.
- the 12-bit original image is shifted to the right by 2 bits to form a 10-bit image, and a difference image between the 10-bit images is created (6102).
- the 4-bit left shift of 6104 is replaced with a 2-bit shift.
- the extended video stream is shifted 2 bits to the right and added to the decoded picture of the 2D compatible video stream (6106).
- the addition process 6103 (+128) is changed to (+512)
- the subtraction process 6105 ( ⁇ 128) is changed to ( ⁇ 512).
- the base-view video stream is generated by compressing and encoding a monochrome image such as a black screen.
- a differential video (12 bits) for improving the image quality of a 2D compatible video stream may be generated by compression encoding.
- the method described with reference to FIG. 89 may be used as the difference video generation method and composition method.
- FIG. 90 shows the configuration of a video stream according to this modification.
- the base-view video stream is created by compressing and encoding the difference video between the 8-bit original picture and the 2D compatible video stream in the base-view video encoder.
- the dependent-view video encoder unit is configured to compress and encode an image using a high-quality 2D video picture generated by synthesizing the differential video with the decoded picture of the 2D compatible video stream as a reference image.
- the base-view video decoder unit transfers the decoded picture to the 2D-compatible video decoder unit.
- the picture obtained by decoding the 2D compatible video and the picture decoded by the base view video decoder unit are combined.
- the dependent-view video decoder section decodes the dependent-view video stream with reference to the picture obtained by the synthesis process.
- inter-view reference is realized between streams in which video is compressed and encoded with different codecs, thereby reducing the multi-view video stream of 3D video while maintaining playback compatibility with conventional 2D video. Configured with bit rate.
- the inter-view reference is realized between streams in which the video is compressed and encoded with different codecs, so that the conventional 2D video has reproduction compatibility.
- the high resolution corresponds to, for example, a case where the 2D video has a resolution of 1920 ⁇ 1080 in full HD and the high resolution video has a resolution of 4K2K (3840 ⁇ 2160).
- FIG. 72 is a diagram showing an outline of an encoding procedure related to high resolution.
- the 2D-compatible video stream is decoded D1 (7201), up-converted twice in the vertical and horizontal directions (7202), and stored in the picture reference buffer (7203).
- the extended video stream is decoded D2 (7204) with reference to the scaled (up-converted) picture stored in the picture reference buffer (7203).
- the 2D compatible video stream is encoded as a base video.
- full HD video is compressed and encoded.
- 4K2K high-resolution video is compression-coded.
- the extended video stream is compression-encoded using a video obtained by scaling the video of the decoded 2D compatible video stream to a high resolution (video obtained by converting full HD into 4K2K) as a reference picture.
- the bit rate of the extended video stream can be lowered by using the 2D compatible video stream as a reference picture.
- FIG. 75 is a diagram for explaining the scaling.
- the pixel set 7510 is composed of two pixels 7511 to 7514 arranged vertically and horizontally.
- a pixel set 7501 composed of four pixels is converted to 16 pixels by doubling the height and doubling the width.
- Method 1 is a method in which each pixel is simply duplicated vertically and horizontally to form 4 pixels.
- Method 2 is a method of inserting a pixel having an intermediate value of pixel values arranged between front and rear or left and right between the pixels.
- the pixel 7501 is an intermediate color between the left and right pixels 7511 and 7512
- the pixel 7502 is an intermediate color between the pixels 7511 and 7513.
- Method 3 is a method of taking the intermediate color of all pixels diagonally forward, backward, left and right.
- the pixel 7503 is an intermediate color of the surrounding pixels 7511, 7501, 7512, 7502, 7504, 7513, 7505, and 7514.
- the playback apparatus performs up-conversion using a scaling method selected from a plurality of scaling methods.
- ⁇ 4-2. Data format> First, a data structure for encoding a high-resolution video according to the present embodiment will be described.
- the transport stream according to the present embodiment is a 2D compatible video stream, a base-view video stream and a dependent-view video stream of a multi-view video stream, each of which is made into a PES packet, divided into TS packets, and multiplexed.
- the 2D compatible video stream is a video stream configured in a format that can be reproduced as 2D video by a 2D video playback device as described with reference to FIG.
- the 2D compatible video stream is a video stream encoded with an MPEG-2 video codec.
- the multi-view video stream is a video stream encoded with a codec that realizes inter-view reference, and here is compressed with the MPEG-4 MVC video codec.
- the 2D compatible video stream is a full HD original image that has been compression-encoded according to the MPEG-2 video codec.
- the base view video stream is a 4K2K black image that is compression-coded at a low bit rate in accordance with the MPEG-4 MVC video codec.
- the dependent-view video stream is a high-resolution 4K2K original image that has been compression-encoded according to the MPEG-4 MVC video codec.
- Dependent view video stream is compressed using inter-view reference.
- a picture referred to as an inter-view reference is not a picture of a base-view video stream at the same display time (same PTS) but a scaled picture of a decoded picture of a 2D compatible video stream at the same display time (same PTS).
- the decoded picture (black image) of the base-view video stream is replaced with the decoded picture (full HD) of the 2D compatible video stream at the same display time, and the replaced picture (full HD) is referred to The picture (4K2K) at the same display time of the dependent-view video stream is decoded.
- the picture of the dependent-view video stream uses the image related to the video obtained by upscaling the “decoded picture of the 2D-compatible video stream” to 4K2K as the reference image. It is possible to provide video with a lower bit rate and higher resolution than a 2D compatible video stream.
- FIG. 74 is a diagram for explaining a PMT packet included in the transport stream.
- signaling information for decoding high-resolution video is included in a system packet such as a PMT packet.
- the signaling information includes a high-resolution information descriptor for performing signaling such as the relationship between each video stream and the start and end of high-resolution video reproduction of this system, and a high-resolution stream descriptor set for each video stream. Is included.
- the high resolution information descriptor and the high resolution stream descriptor have almost the same configuration as the high quality information descriptor and high quality stream descriptor, respectively.
- image quality improvement in the description of the image quality improvement information descriptor and the image quality improvement stream descriptor already performed with reference to FIG. 50, FIG. 51 and FIG. Since this is replaced with “higher resolution”, further explanation is omitted.
- the resolution information descriptor may store resolution information of each stream.
- a playback device that plays back high-resolution video can determine a decoding method and a scaling method for each stream by referring to the descriptor information.
- the scaling method may be set in the resolution enhancement information descriptor. Since there are a plurality of scaling methods as shown in FIG. 75, information indicating which scaling method is used is stored in the descriptor. In this way, a playback device that plays back high-resolution video can determine the decoding method and scaling method of each stream with reference to the descriptor information.
- the normal playback device is a device that can play back only a 2D compatible video stream.
- the normal playback device can play back a stream with a bit rate of up to 10 Mbps.
- the normal playback device is already in the market and plays back a stream distributed by a broadcast wave or the like.
- the high-resolution playback device has a decoding function of not only a 2D compatible video stream but also a multi-view video stream. In the case where these two types of playback devices exist, it is assumed that the transport stream having the configuration described in FIG. 73 is broadcast.
- the normal playback device decodes and outputs a 2D compatible video stream, and plays back normal 2D video.
- the high-resolution playback device decodes both the 2D compatible video stream and the multi-view video stream.
- the high-resolution playback device replaces the decoded picture of the base-view video stream with a scaled picture of the decoded picture of the 2D-compatible video stream at the same display time, and thereby changes the dependent-view video stream at the same display time. Are decoded with reference to the replaced picture.
- FIG. 77 is a diagram showing a configuration of a data creation device 7701 as a video encoding device according to the present embodiment.
- the data creation device 7701 has basically the same configuration as the data creation device 2601 of the first embodiment shown in FIG. However, since the input data is different, in FIG. 77, the “left eye original image” in FIG. 26 is replaced with a 2D “high resolution original image image (here, the resolution is 4K2K)”, and the “right eye original image” is the same. It has been replaced with 2D “high resolution original image”. The following description will be focused on differences from the data creation device 2601 of the first embodiment.
- the data creation device 7701 is different from the data creation device 2601 in the configuration of the 2D compatible video encoder 7702 and the dependent-view video encoder 7709 in the extended multi-view video encoder 7704.
- the 2D compatible video encoder 7702 compresses and encodes a video (in this case, full HD) obtained by reducing the resolution of the high resolution original image as an input in accordance with a 2D compatible video codec.
- a 2D compatible video codec is MPEG-2 video
- it is compression-encoded into the MPEG-2 video stream format and a 2D compatible video stream is output.
- the dependent-view video encoder 7709 does not input one video (left-eye or right-eye original video) constituting the 3D video, but encodes the high-resolution original video as an input.
- the dependent-view video encoder 7709 uses the picture obtained by scaling the decoded picture in the 2D-compatible video frame memory 2608 (the picture obtained by decoding the 2D-compatible video stream) for inter-view reference and performs the compression process.
- FIG. 78 is a diagram showing a data creation flow by the data creation device 7701 configured as described above.
- the data creation flow shown in FIG. 78 is approximately the same as the data creation flow of the data creation device 2601 described with reference to FIG. 27 in the first embodiment. However, the data creation device 7701 increases the gradation of 2D video. In FIG. 78, S2703 and S2706 in FIG. 27 are replaced with S7803 and S7806.
- the 2D compatible video encoder 7702 generates 2D compatible video streams for one encoded number. According to the codec of the 2D compatible video stream from the Nth frame of the high-resolution original image video, according to the codec of the 2D compatible video stream, the resolution is reduced (full HD in this case) and compression encoded, and a 2D compatible video stream is generated and output. . Then, the process proceeds to S2704.
- the dependent-view video encoder 7709 generates a dependent-view video stream for one encoded number. Based on the base-view video encoding information, video attributes, picture structure, and the like are determined, and for reference between views, a picture in the 2D compatible video frame memory 2608 is used, and a full HD video (2D compatible video frame memory 2608 is used. 2D compatible video stream (decoding result of 2D-compatible video stream) stored in the video, and the high-resolution video as a reference image, the high-resolution original video is compressed and encoded by the number of encodes from the Nth frame, and the dependent view Generate a video stream. Then, the process proceeds to S2707. ⁇ 4-4.
- FIG. 79 is a diagram showing a configuration of a playback device 7923 as a high-resolution playback device that plays back high-resolution video according to the present embodiment.
- the playback device 7923 has basically the same configuration as the 3D video playback device 2823 of the first embodiment shown in FIG. However, since the playback device 7923 is intended to increase the resolution of 2D video, in FIG. 67 there is no plane output from the 2D compatible video decoder, and a scaler 7924 is added.
- FIG. 80 is a diagram showing a flow of decoding processing and output processing in high-resolution video playback by the playback device 7923 configured as described above.
- the playback device 7923 scales the decoded picture of the 2D compatible video stream by the scaler 7924, and outputs the scaled data to the inter-view reference buffer 2816.
- the playback device 7923 outputs the decoded picture of the dependent-view video stream to the first plane 2809 at the PTS timing. ⁇ 4-5. Supplementary explanation about effects> Hereinafter, the effect in the present embodiment will be supplementarily described with reference to FIG.
- the normal playback device in the figure is a playback device that can play back only a 2D compatible video stream.
- the bit rate that can be played back by the normal playback device is up to 10 Mbps.
- the normal playback device is already in the market and plays back a stream distributed by a broadcast wave or the like.
- the high-resolution playback device is a playback device that can decode and play back a high-resolution video stream encoded with a higher-resolution video than a 2D compatible video stream.
- the resolution of the normal video is full HD (1920 ⁇ 1080) and the resolution of the high resolution video is 4K2K (3820 ⁇ 2160).
- the 2D compatible video stream is 10 Mbps
- the bit rate of the high-resolution video stream obtained by compressing the same video with the same codec is 20 Mbps.
- the lower part of FIG. 71 shows an example in which a scalable video coding method is used in order to solve the above-described problems and reduce the required transfer bandwidth.
- the scalable video coding method is a method of scaling a low-resolution base video to create a reference picture and decoding a high-resolution extended video.
- both the 2D compatible video stream and the extended video stream must be configured with the same video codec.
- both video streams must be compressed and encoded by MPEG-4 MVC, which is a modified standard of MPEG-4 AVC and is a scalable video encoding method, for example.
- the multi-view video encoding method is applied to the video stream for increasing the resolution of the 2D compatible video stream.
- the scalable video encoding method can also be realized with the same configuration. In this case, if the extended multi-view video decoder of the playback device 7923 is changed to an extended scalable video decoder, the decoding result of the 2D-compatible video stream is replaced with the decoding result of the base-view video stream without scaling. Good.
- the multi-view video encoding method is applied to the video stream for increasing the resolution of the 2D compatible video stream.
- the Side-by-Side 3D video is changed to the full HD 3D video. The resolution may be increased.
- FIG. 81 is a diagram showing a stream configuration in this case.
- the 2D compatible video stream is obtained by compressing and encoding Side-by-Side 3D video by the encoding method of the 2D compatible video stream.
- the resolution of Side-by-Side 3D video is, for example, full HD resolution.
- the base-view video stream constituting the multi-view video stream is obtained by compression-coding black video at a low bit rate in accordance with the multi-view video codec.
- the dependent-view video stream is a high-resolution 4K1K Side-By-Side video that has been compression-encoded.
- the dependent-view video stream is compressed using the inter-view reference, but the picture referred to in the inter-view reference is not a picture of the base view video stream at the same display time (the same PTS),
- a decoded picture after decoding a 2D compatible video stream at the same display time (the same PTS) is a picture that has been up-converted to 4K1K.
- the video decoder that plays the video stream replaces the decoded picture of the base-view video stream with the scaled picture of the decoded picture of the 2D compatible video stream at the same display time.
- the picture of the dependent-view video stream at the same display time is decoded with reference to the picture.
- the dependent-view video stream can be reduced in bit rate by using a video obtained by up-converting “decoded picture of 2D compatible video stream” to 4K1K as a reference image.
- FIG. 82 is a diagram showing a configuration of a 3D video playback device according to the present modification.
- This configuration is almost the same as the configuration of the playback apparatus in the present embodiment, but is different in that a 3D output unit 8201 is added.
- the 3D output unit 8201 cuts out the left-eye video and the right-eye video for the first plane 2809 from which the 4K1K high-resolution Side-by-Side video is output, and outputs the video to a television or the like.
- a method different from Modification (4) may be employed.
- FIG. 83 is a diagram showing a configuration of a video stream according to this modification.
- the 2D compatible video stream is a side-by-side 3D video that has been compression-encoded by the encoding method of the 2D compatible video stream.
- the resolution of Side-by-Side 3D video is, for example, full HD resolution.
- the base-view video stream constituting the multi-view video stream is obtained by compression-coding black video at a low bit rate in accordance with the multi-view video codec.
- the dependent-view video stream is a compression-coded version of a full HD Side-By-Side video.
- This Side-by-Side video is a differential video for increasing the resolution of a 2D compatible video stream.
- the Side-by-Side video of the 2D compatible video stream is a video created by extracting pixels of the odd-numbered columns of the left-eye full HD video and the right-eye full HD video
- the Side-video of the dependent-view video stream The by-Side video is created by extracting even-numbered pixels of the left-eye full HD video and the right-eye full HD video.
- the dependent-view video stream is compressed using the inter-view reference, but the picture referred to as the inter-view reference is not a picture of the base view video stream at the same display time (the same PTS), This is a picture obtained by scaling a decoded picture after decoding of a 2D compatible video stream at the same display time (same PTS).
- the video decoder that plays back the video stream replaces the decoded picture of the base-view video stream with the decoded picture of the 2D compatible video stream at the same display time.
- the picture of the dependent-view video stream at the same display time is decoded with reference to the obtained picture.
- the dependent-view video stream can be reduced in bit rate by using “decoded picture of 2D compatible video stream” as a reference image.
- FIG. 84 is a diagram showing a configuration of a 3D video playback device according to the present modification.
- This configuration is almost the same as the configuration of the playback apparatus in the present embodiment, but is different in that a 3D output unit 8401 is added.
- the 3D output unit 8401 uses the Side-by-Side video of the 2D compatible video stream stored in the first plane 2809 and the Side-by-Side video of the dependent-view video stream stored in the second plane 2820. Then, a high-resolution full HD 3D image is created and output to a television or the like.
- (6) In the above modification (4) and modification (5), Side-by-Side 3D video is used.
- the present invention is not limited to this, and 3D of various frame compatible systems such as the Top & Bottom system and the LineByLine system.
- the Top & Bottom method is a method in which the right-eye and left-eye images are each compressed in the vertical direction (for example, compressed from 1080 pixels to 540 pixels) and transmitted in a vertically arranged manner.
- the LineByLine system is a system in which video signals for the right eye and the left eye are alternately arranged every other line. (7)
- the inter-view reference method of pictures may be changed.
- the left diagram shows the inter-view reference method described in the present embodiment
- the right diagram shows the inter-view reference method according to this modification.
- the 2D compatible video stream or the base video stream is a full HD video
- the dependent-view video stream that is a high-resolution video is a 4K2K video.
- a picture obtained by decoding (9503) a 2D-compatible video stream or base-view video stream (9501) into a 4K2K video by scaling (up-conversion) (9505) is used as an inter-view reference frame. It is stored in the buffer (9506), and when the dependent-view video stream of 4K2K video is decoded, the inter-view video is referred to by referring to the 4K2K decoded picture of this buffer.
- the decoded picture to be decoded (9503) of the 2D compatible video stream or base-view video stream (9501) and stored (9516) in the inter-view reference buffer is Store the full HD resolution without up-conversion.
- the function of the scaling reference filter (9515) is used, and the 2D-compatible video stream / base-view video stream (9501) of the macro-block unit is used.
- the decoding process is performed by enlarging the reference area of the decoded picture.
- FIG. 96 shows a specific procedure of the scaling reference filter.
- the region of the reference picture is specified from the Syntax element such as a motion vector for each macroblock, and decoding is performed with reference to the region.
- the reference method is shown as (1) to (3) in FIG.
- the target area (coordinate position (x, y) and area size (w, h)) of the high-resolution picture referenced by the macroblock is specified.
- a target region of the virtual high resolution picture is generated using a filter function.
- An example f (x, y, w, h) of the filter function is shown in a broken line frame in the figure.
- the filter function f (x, y, w, h) is applied to the method 3 in FIG. 75 for the vertical h / 2 and horizontal w / 2 regions of the picture at the (x / 2, y / 2) coordinate of the real picture.
- the up-conversion is performed twice in the vertical and horizontal directions, so that the target area of the virtual high-resolution picture can be generated by substituting the values of x, y, w, and h.
- the macroblock can be decoded by referring to the target area in the virtual high-resolution picture.
- the base-view video stream is generated by compressing and encoding a monochrome image such as a black screen.
- a differential video for improving the image quality of a 2D compatible video stream
- a difference between a 4k2k high-resolution original image and a 4K2K video obtained by up-converting a picture obtained by decoding a full HD 2D compatible video stream may be generated by compression coding.
- FIG. 91 is a diagram showing a configuration of a video stream according to this modification.
- the method for generating and synthesizing the difference video has already been described with reference to FIG. 89.
- the 2D compatible video is calculated before calculating the difference between the original image (4K2K) and the decoded picture (2K1K) of the 2D compatible video.
- the difference is that the decoded picture (2K1K) is up-converted to 4K2K to generate a differential video.
- the base-view video encoder In the data creation device, the base-view video encoder generates a base-view video stream by compressing and encoding the differential video between the 4K2K high-resolution original video and the picture obtained by up-converting the decoded picture of the 2D-compatible video stream. Then, the dependent-view video encoder unit “high-definition high-resolution 2D video” obtained by synthesizing “differential video” with “a picture obtained by up-converting a full HD decoded picture of a 2D-compatible video stream into a high-resolution 4K2K picture”. Is compressed as a reference image.
- the base-view video decoder unit transfers the decoded picture to the 2D-compatible video decoder unit, and the 2D-compatible video decoder unit up-converts the decoded picture of the 2D-compatible video stream and the difference video.
- the picture is used for decoding the dependent-view video stream.
- Embodiments 1 to 4 the basic structure of the playback device is the same in that it is composed of a 2D-compatible video stream and a multi-view video decoder. Therefore, one playback device is used in Embodiments 1 to 4. It may be performed while switching the reproduction process.
- the video streams of Embodiments 1 to 4 may exist separately for each scene in the same transport stream.
- the data creation device 7701 may record a playback method determination descriptor in supplementary data of the PMT or video stream so that it can identify which scene is the playback method in which embodiment.
- FIG. 85 is a diagram showing the structure of the playback method discrimination descriptor.
- the playback method determination descriptor stores a 3D playback flag, a high image quality flag, a high gradation level flag, and a high resolution flag.
- each flag is TRUE
- information corresponding to the flag is further stored.
- the 3D playback flag is TRUE
- the 3D playback information descriptor is stored in the playback method determination descriptor
- the image quality improvement flag is TRUE
- the image quality improvement information descriptor is stored and the gradation enhancement flag is set.
- TRUE a high gradation information descriptor is stored
- a high resolution information descriptor is stored.
- the playback device can select the playback method to be adopted from the playback methods of the above-described embodiments, and can switch the internal state to execute the processing.
- ⁇ 5. Modification> The embodiments of the data creation device and the playback device according to the present invention have been described above. However, the illustrated data creation device and the playback device can be modified as follows, and the present invention is described in the above embodiment. Of course, the present invention is not limited to the data creation device and the playback device as shown in FIG. (1) Hereinafter, the configuration and effects of a data creation device as a video encoding device and a playback device as a video playback device according to an embodiment of the present invention will be described.
- a video encoding device includes: a first encoding unit that generates an MPEG-2 stream by compressing and encoding a first video obtained by converting an original video into a first quality; Second encoding means for generating a base-view video stream and a dependent-view video stream in conformity with the MPEG-4 MVC format, and sending means for transmitting the streams generated by the first and second encoding means
- the second encoding means includes pseudo data having the same number of pictures as the stream obtained by compressing and encoding the first video as the base-view video stream, and having a reduced total data amount. While generating the stream as the dependent-view video stream, the original video is a second quality higher than the first quality.
- Each picture of the second video image is converted to generate the base view compressed coded stream pictures in the MPEG-2 format stream picture and the same time corresponding to the picture as a reference picture in the video stream.
- the second encoding means includes, in generating a stream conforming to the MPEG-4 MVC format, including information indicating that the reference image is included in the MPEG-2 format stream. Also good.
- the playback side when the dependent-view video stream is played back, the playback side can refer to pictures included in the MPEG-2 stream.
- the second encoding means as the reference image, among the pictures of the MPEG-2 format stream, a value of PTS (Presentation Time Stamp) is an object to be encoded in the second video.
- PTS Presentation Time Stamp
- a picture having the same PTS value as a picture of the base-view video stream corresponding to the picture may be selected.
- the picture to be referred to can be specified from the pictures in the MPEG-2 stream by referring to the PTS on the playback side.
- first encoding unit and the second encoding unit perform compression encoding of the first video and the second video with the same aspect ratio, and information indicating the aspect ratio in each of the generated streams May be included.
- the aspect ratio of the first video and the second video can be specified on the playback side.
- the second encoding means may use a group of pictures that do not change over time as the pseudo data.
- the second encoding means may use an image group made up of monochrome images as the picture group that does not change with time.
- the information amount of the base view video stream can be reduced as compared with the conventional art.
- the second video is a video representing the original video at a higher bit rate than the first video
- the second encoding means is a compression encoding related to the dependent-view video stream
- Information corresponding to a difference in bit rate between corresponding pictures may be compression-encoded.
- the above-mentioned predetermined bit rate can be suppressed while maintaining the reproduction compatibility of the original video of the predetermined bit rate by the reproduction apparatus based on the MPEG-2 standard, while suppressing an increase in the amount of data required than before. It is possible to compress and encode an original video having a higher bit rate.
- the second video is a video in which the original video is expressed with a higher gray level than the first video, and each of the pictures has gray level information indicating the gray level of each pixel constituting the picture.
- the second encoding means may compress and encode a difference in gradation information between corresponding pictures as compression encoding related to the dependent-view video stream.
- the second video is a video representing the original video at a higher resolution than the first video, and the second encoding means converts the reference video into the resolution of the second video.
- MPEG-2 stream pictures may be used.
- the first video is a video representing the original video in an interlaced format
- the second video is a video representing the original video in a progressive format
- the second encoding means is a top field.
- the resolution of the picture may be converted to the resolution of the second video.
- the progressive format original image is compressed and encoded while suppressing the increase in the amount of data required than before, while maintaining the compatibility of the interlace format original image by the reproduction apparatus based on the MPEG-2 standard.
- the second encoding means as the second video, a video in which the original video is expressed at a higher bit rate than the first video, and a video in which the original video is expressed with a higher gradation than the first video.
- a descriptor that indicates which video is used in at least one of the base-view video stream and the dependent-view stream. It may be included.
- the second video is a video in which the original video is expressed at a higher bit rate than the first video, the video in which the original video is expressed with a higher gradation than the first video, and the It is possible to specify which of the images represents the original image at a higher resolution than the first image.
- a video encoding method includes a first encoding step of generating an MPEG-2 format stream by compressing and encoding a first video obtained by converting an original video to a first quality, A second encoding step for generating a base-view video stream and a dependent-view video stream in conformity with the MPEG-4 MVC format; and a transmission step for transmitting the streams generated by the first and second encoding means.
- the base view video stream is made up of pseudo data having the same number of pictures as the stream obtained by compressing and encoding the first video and having a reduced total data amount.
- Each picture of the second video converted to the second quality higher than the quality is compression-coded using the picture of the MPEG-2 stream at the same time as the picture corresponding to the picture in the base-view video stream as a reference image. Generate a stream.
- a video encoding program is a video encoding program for causing a computer to function as a video encoding device, wherein the computer converts the original video into a first quality.
- First encoding means for generating an MPEG-2 stream by compressing and encoding
- second encoding for generating a base-view video stream and a dependent-view video stream in conformity with the MPEG-4 MVC format
- a sending means for sending the stream generated by the first and second encoding means, and the second encoding means compresses and encodes the first video as the base-view video stream.
- each picture of the second video obtained by converting the original video to the second quality higher than the first quality is used as the dependent-view video stream in the base-view video stream.
- a stream is generated by compression-coding a picture of the MPEG-2 stream at the same time as the picture corresponding to the picture as a reference image.
- the playback device while maintaining the playback compatibility of the original video of the first quality by the playback device based on the MPEG-2 standard, the increase in the amount of data required than before is suppressed, and higher than the first quality.
- the original video of the second quality can be compression-encoded.
- a video playback apparatus is a video playback apparatus that acquires, decodes, and plays back a stream transmitted by the above-described video encoding apparatus, and decodes the MPEG-2 format stream.
- 1 decoding means decoding the base-view video stream, and decoding each picture of the second video in the decoding of the dependent-view stream, the first video at the same time as a picture corresponding to the picture in the base-view video stream
- a second decoding unit configured to decode the picture decoded by the decoding unit as a reference image; and a reproducing unit configured to reproduce the second video obtained by the second decoding unit.
- a video playback method is a video playback method for acquiring, decoding, and playing back a stream transmitted by the above-described video coding method, wherein the MPEG-2 format stream is decoded.
- 1 decoding step decoding the base-view video stream, and decoding each picture of the second video in the decoding of the dependent-view stream, the first video at the same time as a picture corresponding to the picture in the base-view video stream
- a video playback program is a video playback program for causing a computer to function as a video playback device that acquires, decodes, and plays back a stream transmitted by the above-described video encoding device,
- a first decoding means for decoding the MPEG-2 format stream; and decoding the base-view video stream, and decoding each picture of a second video into the base-view video in the decoding of the dependent-view stream.
- a second decoding unit that decodes, as a reference image, a picture decoded by the first decoding unit at the same time as a picture corresponding to the picture in a stream, and reproduces the second video obtained by the second decoding unit It functions as a reproducing means.
- a part or all of the constituent elements constituting each of the above-described devices may be constituted by one system LSI.
- a system LSI is an ultra-multifunctional LSI manufactured by integrating a plurality of components on a single chip. Specifically, a microprocessor, a ROM (Read Only Memory), a RAM (Random Access Memory), etc. It is a computer system comprised including. A computer program is stored in the RAM. The system LSI achieves its functions by the microprocessor operating according to the computer program.
- LSI Integrated Circuit
- IC Integrated Circuit
- the method of circuit integration is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor.
- An FPGA Field Programmable Gate Array
- a reconfigurable processor that can reconfigure the connection and setting of the circuit cells inside the LSI may be used.
- Each of the above devices may be a computer system including a microprocessor, a ROM, a RAM, a hard disk unit, and the like.
- a computer program is stored in the RAM or the hard disk unit.
- Each device achieves its function by the microprocessor operating according to the computer program.
- the computer program is configured by combining a plurality of instruction codes indicating instructions for the computer in order to achieve a predetermined function.
- the present invention may be a method representing the procedure of each process described above. Further, the present invention may be a computer program that realizes these methods by a computer, or may be a digital signal composed of the computer program.
- the present invention also provides a computer-readable recording medium such as a flexible disk, hard disk, CD-ROM, MO, DVD, DVD-ROM, DVD-RAM, BD (Blu-ray Disc). ), Recorded in a semiconductor memory or the like. Further, the present invention may be the computer program or the digital signal recorded on these recording media.
- the present invention may transmit the computer program or the digital signal via an electric communication line, a wireless or wired communication line, a network represented by the Internet, a data broadcast, or the like. (5) The above embodiment and the above modifications may be combined as appropriate.
- the video encoding apparatus and the video playback apparatus according to the present invention perform encoding, transmission, and playback of higher quality video while maintaining playback compatibility with a conventional playback apparatus that plays back an MPEG-2 stream. It is suitable as a device constituting the system to be realized.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
<1-1.概要>
本発明の一実施の形態に係る放送システムは、2D映像として、既存技術であるMPEG-2形式のストリームを生成し、3D映像として、MPEG-4 MVC形式を拡張した新規形式(本明細書において、この形式をMPEG-4 MVCに準拠という。)のベースビュービデオストリームとディペンデントビュービデオストリームとを生成し、送出する。
<1-2.データ作成装置>
<1-2-1.構成>
以下、本発明に係るデータ作成装置の一実施形態について図を参照しながら説明する。
復号ピクチャは、拡張マルチビュービデオエンコーダ2604の2D互換ビデオフレームメモリ2608に格納される。また、2D互換ビデオ符号化情報2606は、ベースビュービデオエンコーダ2605に入力される。
ディペンデントビュービデオエンコーダ2609は、ビュー間参照を実行するための参照ピクチャIDを、ベースビュービデオ符号化情報2607のベースビュービデオフレームメモリ管理情報に基づいて決定する。さらに、ディペンデントビュービデオエンコーダ2609は、ディペンデントビュービデオストリームのビデオの属性情報をベースビュービデオ符号化情報2607に含まれるベースビュービデオストリームの属性情報の値と同じ値を設定する。
次に、データフォーマットについて、図面を参照しながら説明を行う。
図27はデータ作成装置2601のデータ作成フローを示す図である。以下、そのデータ作成フローを説明する。
<1-3.再生装置>
<1-3-1.構成>
次に本実施の形態に係る3D映像を再生する再生装置2823の構成について図を参照しながら説明する。
図30は再生装置2823の3D映像における復号処理と出力処理のフローを示す図である。
再生装置2823のD1(2805)で行う復号処理は、既存のMPEG-2形式のビデオストリームの復号処理と同じなので、既存のMPEG-2形式のビデオの再生装置のLSI(Large Scale Integration)やソフトウェアを利用することができる。また、D2(2817)で行うMPEG-4 MVC形式の復号処理についても、既存のMPEG-4 MVC形式の処理と同じであるので、既存のMPEG-4 MVC形式のビデオの再生装置のLSIやソフトウェアを利用できる。
<再生装置2823の使用態様の一例>
図5を用いて、データ作成装置2823で作成したビデオストリームの3D映像を再生できる3Dデジタルテレビ100と、3D映像の再生をサポートしない従来の2D映像のみを再生できる2Dデジタルテレビ300を例に挙げて説明する。
<1-4.変形例>
以上、本発明に係るデータ作成装置及び再生装置の実施の形態を説明したが、例示したデータ作成装置及び再生装置を以下のように変形することも可能であり、本発明が上述の実施の形態で示した通りのデータ作成装置及び再生装置に限らないことは勿論である。
図24は、特殊再生について説明するための図である。図24上段は、上記の制約が課されていない場合について示している。この場合、表示順で3番目のピクチャにおいて、2D互換ビデオストリームとベースビュービデオストリームは、Pピクチャ(P3)になっているが、ディペンデントビュービデオストリームのピクチャはBピクチャ(B3)となっている。
この圧縮処理においては、ディペンデントビュービデオエンコーダ2609は、高画質2D互換ビデオフレームメモリ8711に格納される高画質化された2D互換ビデオフレームメモリ内の復号ピクチャを、ビュー間参照に使って圧縮符号化する。圧縮符号化方法は、ディペンデントビュービデオエンコーダ2609の説明内容と同じである。
<1-5.補足説明>
<映像圧縮技術の説明>
<2D映像圧縮技術>
本実施の形態に係るデータ作成装置及び再生装置で使用する2D映像を圧縮符号化する規格であるMPEG-2、及びMPEG-4 AVC(MPEG-4 MVC形式のベースとなる圧縮符号化方式)の符号化方法について簡単に説明する。
次に、視差画像を用いて3D映像としてディスプレイ等で再生する方法、マルチビュー符号化方式であるMPEG-4 MVC形式の圧縮符号化について簡単に説明する。
デジタルテレビの放送波等での伝送では、MPEG-2トランスポートストリーム形式のデジタルストリームが使われている。
次に、映像を上述の符号化方式で圧縮符号化して得られるビデオストリームのデータ構成について説明する。
次に、MPEG-4 MVC形式のビデオストリームについて説明する。
<2.実施の形態2>
<2-1.概要>
実施の形態1では、異なるコーデックで映像が圧縮符号化されたストリーム間においてビュー間参照を実現することによって、従来の2D映像に再生互換性を持たせながら、3D映像のマルチビュービデオストリームを低ビットレートで構成した。本実施の形態では、異なるコーデックで映像が圧縮符号化されたストリーム間においてビュー間参照を実現することによって、従来の2D映像に再生互換性を持たせながら、従来の2D映像を高品位化の一例として高画質化した映像を、低ビットレートに構成したマルチビュービデオストリームで送信する。ここで「高画質」とは、より原画映像に近い画質、一例としては、ビデオ圧縮時に発生し得るブロックノイズなどの圧縮ひずみが少ない画質をいう。本実施の形態では、高画質化した映像の一例として、高いビットレートで符号化された映像を用いる。映像に係る画質は、符号化のビットレートに左右され、ビットレートが高いほど原画に近い高画質な映像を符号化でき、ビットレートが低くなると、動きの激しいシーンなどで圧縮、符号化しきれずにブロックノイズが混入する場合があるためである。
<2-2.使用データ>
まず、本実施の形態に係る、高画質映像を符号化するためのデータ構造について説明する。
<2-2-1.トランスポートストリーム>
図48は、本実施の形態に係るトランスポートストリームに含まれるデータについて説明するための図である。
<2-2-2.PMTパケット>
図50は、トランスポートストリームに含まれるPMTパケットについて説明するための図である。高画質映像を伝送するためのトランスポートストリームにおいては、PMTパケットなどのシステムパケット中に、高画質映像のデコード処理を行う上でのシグナリング情報が含まれる。シグナリング情報には、各ビデオストリーム相互の関係や高画質映像再生の開始、終了などのシグナリングを行うための高画質情報ディスクリプタと、ビデオストリーム毎に設定される高画質ストリームディスクリプタとが含まれる。
(1)高画質化情報ディスクリプタ
図51は、高画質化情報ディスクリプタの構造を示す図である。
(2)高画質化ストリームディスクリプタ
図53は、高画質化ストリームディスクリプタの構造を示す図である。
<2-2-3.PTS、DTS及びGOP>
トランスポートストリーム中における2D互換ビデオストリームとベースビュービデオストリームとディペンデントビュービデオストリームの各ビデオアクセスユニットに割り当てる表示時刻(PTS)、デコード時刻(DTS)およびピクチャタイプの関係は、実施の形態1の2D互換ビデオストリーム、ベースビュービデオストリームとディペンデントビュービデオストリームにおける関係と同じである。
<2-2-4.ストリームの使用形態>
上述のストリームの使用形態について、図49を用いて説明する。
<2-3.データ作成装置>
以下、本実施の形態に係るデータ作成装置について説明する。
<2-3-1.構成>
図54は、本実施の形態に係る映像符号化装置としてのデータ作成装置5401の構成を示す図である。
<2-3-2.動作>
図55は、上述のように構成したデータ作成装置5401によるデータ作成処理を示すフローチャートである。
<2-4.再生装置>
次に、本実施の形態に係る高画質映像を再生する、映像再生装置としての再生装置について説明する。
<2-4-1.構成>
図56は、本実施の形態に係る再生装置5623の構成を示す図である。
<2-4-2.動作>
図57は、上述のように構成した再生装置5623による復号処理と出力処理を示すフローチャートである。
<2-5.効果についての補足説明>
以下、本実施の形態における効果について図47を用いて補足説明する。
<2-6.本実施の形態に係る変形例>
(1)実施の形態1及び変形例の説明内容は、本実施の形態の構成等に反しない限り、本実施の形態にも適用されることはいうまでもない。
(2)本実施の形態では、図52に示す高画質化情報ディスクリプタ中の高画質再生方式フィールドの記載により、複数の高画質再生方式から1の高画質再生方式を選出する構成について説明した。複数の高画質再生方式を切り替える方法は、既に図40を用いて説明した構成を適用することで実装を容易にできる。
<3.実施の形態3>
<3-1.概要>
実施の形態1では、異なるコーデックで映像が圧縮符号化されたストリーム間においてビュー間参照を実現することによって、従来の2D映像に再生互換性を持たせながら、3D映像のマルチビュービデオストリームを低ビットレートで構成した。本実施の形態では、異なるコーデックで映像が圧縮符号化されたストリーム間においてビュー間参照を実現することによって、従来の2D映像に再生互換性を持たせながら、従来の2D映像について高品位化の一例として色再現性の向上(高階調度化)した映像を、低ビットレートに構成したマルチビュービデオストリームで送信する。ここで、色再現性の向上とは、従来のビデオコーデックによって圧縮される色情報に対し、より高精度の階調情報を付加することである。通常のビデオコーデックでは、色情報を8ビットカラーで表現している。RGBカラーの場合、RGBのそれぞれに8ビット(0~255階調)ずつ割り当てた合計24ビットで約1677万色の表示を可能としている。しかし、8ビットの色階調では表現できない色が存在するため、この色階調で表現された映像について、ユーザにより画質が悪いと判断されてしまう場合があるため、本実施の形態では、高階調度化した映像の一例として、12ビットの色階調で表現した映像を用いる。
<3-2.データフォーマット>
まず、本実施の形態に係る、高階調映像を符号化するためのデータ構造について説明する。
<3-2-1.トランスポートストリーム>
図62は、本実施の形態に係るトランスポートストリームに含まれるデータについて説明するための図である。
<3-2-2.PMTパケット>
図63は、トランスポートストリームに含まれるPMTパケットについて説明するための図である。
<3-2-4.ストリームの使用形態>
上述のストリームの使用形態について、図64を用いて説明する。
<3-3.データ作成装置>
以下、本実施の形態に係るデータ作成装置について説明する。
<3-3-1.構成>
図65は、本実施の形態に係る映像符号化装置としてのデータ作成装置6501の構成を示す図である。
<3-3-2.動作>
図66は、上述のように構成したデータ作成装置6501によるデータ作成フローを示す図である。
<3-4.再生装置>
次に、本実施の形態に係る高階調映像を再生する、映像再生装置としての再生装置について説明する。
<3-4-1.構成>
図67は、本実施の形態に係る再生装置6723の構成を示す図である。
<3-4-2.動作>
図68は、上述のように構成した再生装置6723による高階調映像再生における復号処理と出力処理のフローを示す図である。
<3-5.効果についての補足説明>
以下、本実施の形態における効果について図60を用いて補足説明する。
<3-6.本実施の形態に係る変形例>
(1)実施の形態1及び変形例の説明内容は、本実施の形態の構成等に反しない限り、本実施の形態にも適用されることはいうまでもない。
(2)本実施の形態では、高階調度化ディスクリプタ中の高階調再生方式フィールドの記載により、複数の高階調再生方式から1の高階調再生方式を選出する構成について説明した。複数の高階調再生方式を切り替える方法に関しては、下記のような構成にすることで実装処理が容易になる。
(3)本実施の形態では、高階調映像として12ビット映像を取り上げたが、これに限るものではなく、高階調映像は、通常の階調よりも高階調の映像であればよい。また、高階調の原画映像における有効ビット量は可変に設定できるようにしても良い。例えば、12ビット原画映像において、あるシーンは10ビット、あるシーンは12ビットというようにシーンごとに有効ビット量を変更しても良い。この場合には、図61で説明した合成処理において、2D互換ビデオストリームのビットシフト量を可変値にして、また、拡張ビデオストリームに対しても有効階調分を抽出するようにビットシフトを行うように構成すればよい。
(4)本実施の形態では、ベースビュービデオストリームを黒画面等の単色映像を圧縮符号化して生成したが、これに限らず、2D互換ビデオストリームを高画質化するための差分映像(12ビット原画映像の上位8ビット原画映像と2D互換ビデオとの差分など)を圧縮符号化して生成してもよい。
<4.実施の形態4>
<4-1.概要>
実施の形態1では、異なるコーデックで映像が圧縮符号化されたストリーム間においてビュー間参照を実現することによって、従来の2D映像に再生互換性を持たせながら、3D映像のマルチビュービデオストリームを低ビットレートで構成した。本実施の形態では、異なるコーデックで映像が圧縮符号化されたストリーム間においてビュー間参照を実現することによって、従来の2D映像に再生互換性を持たせながら、2D映像について高品位化の一例としての高解像度化した映像を、低ビットレートに構成したマルチビュービデオストリームで送信する。高解像度化としては、例えば、2D映像がフルHDの1920x1080の解像度で、高解像度化した映像が4K2K(3840x2160)の解像度であるような場合が該当する。
<4-2.データフォーマット>
まず、本実施の形態に係る、高解像度映像を符号化するためのデータ構造について説明する。
<4-2-1.トランスポートストリーム>
図73は、本実施の形態に係るトランスポートストリームに含まれるデータについて説明するための図である。
<4-2-2.PMTパケット>
図74は、トランスポートストリームに含まれるPMTパケットについて説明するための図である。
<4-2-4.ストリームの使用形態>
上述のストリームの使用形態について、図76を用いて説明する。
<4-3.データ作成装置>
以下、本実施の形態に係るデータ作成装置について説明する。
<4-3-1.構成>
図77は、本実施の形態に係る映像符号化装置としてのデータ作成装置7701の構成を示す図である。
<4-3-2.動作>
図78は、上述のように構成したデータ作成装置7701によるデータ作成フローを示す図である。
<4-4.再生装置>
次に、本実施の形態に係る高解像度映像を再生する、映像再生装置としての再生装置について説明する。
<4-4-1.構成>
図79は、本実施の形態に係る高解像度映像を再生する高解像度再生装置としての再生装置7923の構成を示す図である。
<4-4-2.動作>
図80は、上述のように構成した再生装置7923による高解像度映像再生におけるデコード処理と出力処理のフローを示す図である。
<4-5.効果についての補足説明>
以下、本実施の形態における効果について図71を用いて補足説明する。
<4-6.本実施の形態に係る変形例>
(1)実施の形態1及び変形例の説明内容は、本実施の形態の構成等に反しない限り、本実施の形態にも適用されることはいうまでもない。
(2)本実施の形態における再生装置の再生方式に応じたスイッチ方法については、図58を用いて説明した実施の形態2のスイッチ方法における「高画質化」を「高解像度化」に置き換えることで適用できる。
(3)本実施の形態においては、2D互換ビデオストリームを高解像度化するビデオストリームに、マルチビュービデオ符号化方式を適用したが、スケーラブルビデオ符号化方式についても同等の構成で実現できる。この場合には、再生装置7923の拡張マルチビュービデオデコーダを、拡張スケーラブルビデオデコーダに変更して、2D互換ビデオストリームのデコード結果をスケーリングせずにベースビュービデオストリームのデコード結果と入れ換える構成にすればよい。
(4)本実施の形態においては、2D互換ビデオストリームを高解像度化するビデオストリームに、マルチビュービデオ符号化方式を適用したが、Side-by-Sideの3D映像をフルHDの3D映像に高解像度化してもよい。
本実施の形態で説明したとおり、ビデオストリームを再生するビデオデコーダは、ベースビュービデオストリームのデコード済みピクチャを、同一表示時刻の2D互換ビデオストリームのデコード済みピクチャをスケーリングしたピクチャで置き換え、置き換えられたピクチャを参照して同一表示時刻のディペンデントビュービデオストリームのピクチャを復号する。この構成により、ディペンデントビュービデオストリームを、「2D互換ビデオストリームの復号ピクチャ」を4K1Kにアップコンバートした映像を参照画像として使うことによって低ビットレート化することができる。
(5)Side-by-Sideの3D映像を、フルHDの3D映像に高解像度化する方式に関し、変形例(4)とは異なる方式を採用してもよい。
(6)上述の変形例(4)及び変形例(5)では、Side-by-Sideの3D映像を用いていたが、これに限らず、Top&Bottom方式、LineByLine方式など様々なフレーム互換方式の3Dフォーマットにも同様に適用可能である。なお、Top&Bottom方式は、右目用、左目用の映像をそれぞれ垂直方向に圧縮し(例えば、1080ピクセルから540ピクセルなどに圧縮する)、上下に並べた形で伝送する方式である。また、LineByLine方式は、右目用、左目用の映像信号を1ラインおきに交互に配置する方式である。
(7)本実施の形態における高解像度化を実現するためのマルチビュー符号化方式もしくはスケーラブル符号化方式において、ピクチャのビュー間参照の方法を変更してもよい。
(8)本実施の形態では、ベースビュービデオストリームを黒画面等の単色映像を圧縮符号化して生成していたが、これに限らず、2D互換ビデオストリームを高画質化するための差分映像(4k2k高解像度原画映と、フルHDの2D互換ビデオストリームを復号したピクチャをアップコンバートした4K2K映像との差分)を圧縮符号化して生成してもよい。
(9)実施の形態1~4において、再生装置の基本構造は、2D互換ビデオストリームとマルチビュービデオデコーダから構成されている点で共通するので、1つの再生装置が、実施の形態1~4の再生処理を切り替えながら行うこととしてもよい。
<5.変形例>
以上、本発明に係るデータ作成装置及び再生装置の実施の形態を説明したが、例示したデータ作成装置及び再生装置を以下のように変形することも可能であり、本発明が上述の実施の形態で示した通りのデータ作成装置及び再生装置に限らないことは勿論である。
(1)以下、さらに、本発明の一実施の態様に係る映像符号化装置としてのデータ作成装置、及び映像再生装置としての再生装置の構成及び効果について説明する。
(2)上記の各装置を構成する構成要素の一部又は全部は、1個のシステムLSIから構成されているとしてもよい。システムLSIは、複数の構成部を1個のチップ上に集積して製造された超多機能LSIであり、具体的には、マイクロプロセッサ、ROM(Read Only Memory)、RAM(Random Access Memory )などを含んで構成されるコンピュータシステムである。前記RAMには、コンピュータプログラムが記憶されている。前記マイクロプロセッサが、前記コンピュータプログラムに従って動作することにより、システムLSIは、その機能を達成する。
(3)上記の各装置は、マイクロプロセッサ、ROM、RAM、ハードディスクユニット、などから構成されるコンピュータシステムとしてもよい。前記RAM又は前記ハードディスクユニットには、コンピュータプログラムが記憶されている。前記マイクロプロセッサが、前記コンピュータプログラムに従って動作することにより、各装置は、その機能を達成する。ここで、コンピュータプログラムは、所定の機能を達成するために、コンピュータに対する指令を示す命令コードが複数個組み合わされて構成されたものである。
(4)本発明は、上述の各処理の手順を表した方法であるとしてもよい。また、これらの方法をコンピュータにより実現するコンピュータプログラムであるとしてもよいし、前記コンピュータプログラムからなるデジタル信号であるとしてもよい。
(5)上記実施の形態及び上記変形例をそれぞれ適宜組み合わせるとしてもよい。
2603 2D互換ビデオデコーダ
2604 拡張マルチビュービデオエンコーダ
2605 ベースビュービデオエンコーダ
2606 2D互換ビデオ符号化情報
2607 ベースビュービデオ符号化情報
2608 2D互換ビデオフレームメモリ
2609 ディペンデントビュービデオエンコーダ
2610 マルチプレクサ
2801 PIDフィルタ
2802 TB(1)
2803 MB(1)
2804 EB(1)
2805 2D互換ビデオ圧縮映像デコーダ
2806 O(リオーダリングバッファ)
2807 スイッチ
2808 第1プレーン
2809 TB(2)
2810 MB(2)
2811 EB(2)
2812 TB(3)
2813 MB(3)
2814 EB(3)
2815 デコードスイッチ
2816 ビュー間参照バッファ
2817 マルチビュービデオ圧縮映像デコーダ
2818 DPB
2819 出力プレーンスイッチ
2820 第2プレーン
2821 2D互換ビデオデコーダ
2822 拡張マルチビュービデオデコーダ
5401 データ作成装置
5404 拡張マルチビュービデオエンコーダ
5409 ディペンデントビュービデオエンコーダ
5623 再生装置
Claims (16)
- 原映像を第1の品位に変換した第1映像を圧縮符号化することによりMPEG-2形式のストリームを生成する第1符号化手段と、
MPEG-4 MVC形式に準拠しベースビュービデオストリームとディペンデントビュービデオストリームとを生成する第2符号化手段と、
前記第1及び前記第2符号化手段により生成されたストリームを送出する送出手段とを備え、
前記第2符号化手段は、前記ベースビュービデオストリームとして、前記第1映像を圧縮符号化して得られるストリームとピクチャ数が同一であって、総データ量を少なくした疑似データから成るストリームを生成し、
一方、前記ディペンデントビュービデオストリームとして、前記原映像を前記第1の品位より高い第2の品位に変換した第2映像の各ピクチャを、前記ベースビュービデオストリームにおいて当該ピクチャに対応するピクチャと同時刻の前記MPEG-2形式のストリームのピクチャを参照画像として圧縮符号化したストリームを生成する
ことを特徴とする映像符号化装置。 - 前記第2符号化手段は、前記MPEG-4 MVC形式に準拠したストリームの生成において、当該ストリームに、前記参照画像が前記MPEG-2形式のストリームに含まれることを示す情報を含める
ことを特徴とする請求項1記載の映像符号化装置。 - 前記第2符号化手段は、前記参照画像として、前記MPEG-2形式のストリームのピクチャのうち、PTS(Presentation Time Stamp)の値が、前記第2映像において符号化の対象となっている当該ピクチャに対応している前記ベースビュービデオストリームのピクチャのPTSの値と同一であるピクチャを選出する
ことを特徴とする請求項2記載の映像符号化装置。 - 前記第1符号化手段及び前記第2符号化手段は、前記第1映像及び第2映像の圧縮符号化を同一のアスペクト比で行い、生成する前記各ストリームに、前記アスペクト比を示す情報を含める
ことを特徴とする請求項2記載の映像符号化装置。 - 前記第2符号化手段は、前記疑似データとして、経時的変化のないピクチャ群を用いる
ことを特徴とする請求項1記載の映像符号化装置。 - 前記第2符号化手段は、前記経時的変化のないピクチャ群として、単色の画像から成る画像群を用いる
ことを特徴とする請求項2記載の映像符号化装置。 - 前記第2映像は、前記原映像を前記第1映像よりも高いビットレートで表現した映像であり、
前記第2符号化手段は、前記ディペンデントビュービデオストリームに係る圧縮符号化として、対応するピクチャ間のビットレートの差に相当する情報を圧縮符号化する
ことを特徴とする請求項1記載の映像符号化装置。 - 前記第2映像は、前記原映像を前記第1映像よりも高階調で表現した映像であり、
前記ピクチャそれぞれは、当該ピクチャを構成する各ピクセルの階調を表す階調情報を有し、
前記第2符号化手段は、前記ディペンデントビュービデオストリームに係る圧縮符号化として、対応するピクチャ間の階調情報の差分を圧縮符号化する
ことを特徴とする請求項1記載の映像符号化装置。 - 前記第2映像は、前記原映像を前記第1映像よりも高解像度で表現した映像であり、
前記第2符号化手段は、前記参照画像として、前記第2映像の解像度に変換した前記MPEG-2形式のストリームのピクチャを用いる
ことを特徴とする請求項1記載の映像符号化装置。 - 前記第1映像は、前記原映像をインターレース形式で表現した映像であり、
前記第2映像は、前記原映像をプログレッシブ形式で表現した映像であり、
前記第2符号化手段は、トップフィールド及びボトムフィールドの各ピクチャを前記参照画像として用いる場合に、当該ピクチャの解像度を前記第2映像の解像度に変換して用いる
ことを特徴とする請求項9記載の映像符号化装置。 - 前記第2符号化手段は、前記第2映像として、前記原映像を前記第1映像よりも高いビットレートで表現した映像、前記原映像を前記第1映像よりも高階調で表現した映像及び前記原映像を前記第1映像よりも高解像度で表現した映像のいずれかを用い、いずれの映像を用いたかを示すディスクリプタを、前記ベースビュービデオストリーム及び前記ディペンデントビューストリームの少なくとも一方に含める
ことを特徴とする請求項1記載の映像符号化装置。 - 原映像を第1の品位に変換した第1映像を圧縮符号化することによりMPEG-2形式のストリームを生成する第1符号化ステップと、
MPEG-4 MVC形式に準拠しベースビュービデオストリームとディペンデントビュービデオストリームとを生成する第2符号化ステップと、
前記第1及び前記第2符号化手段により生成されたストリームを送出する送出ステップとを含み、
前記第2符号化ステップにおいて、前記ベースビュービデオストリームとして、前記第1映像を圧縮符号化して得られるストリームとピクチャ数が同一であって、総データ量を少なくした疑似データから成るストリームを生成し、
一方、前記ディペンデントビュービデオストリームとして、前記原映像を前記第1の品位より高い第2の品位に変換した第2映像の各ピクチャを、前記ベースビュービデオストリームにおいて当該ピクチャに対応するピクチャと同時刻の前記MPEG-2形式のストリームのピクチャを参照画像として圧縮符号化したストリームを生成する
ことを特徴とする映像符号化方法。 - コンピュータを映像符号化装置として機能させるための映像符号化プログラムであって、
前記コンピュータを、
原映像を第1の品位に変換した第1映像を圧縮符号化することによりMPEG-2形式のストリームを生成する第1符号化手段と、
MPEG-4 MVC形式に準拠しベースビュービデオストリームとディペンデントビュービデオストリームとを生成する第2符号化手段と、
前記第1及び前記第2符号化手段により生成されたストリームを送出する送出手段として機能させ、
前記第2符号化手段は、前記ベースビュービデオストリームとして、前記第1映像を圧縮符号化して得られるストリームとピクチャ数が同一であって、総データ量を少なくした疑似データから成るストリームを生成し、
一方、前記ディペンデントビュービデオストリームとして、前記原映像を前記第1の品位より高い第2の品位に変換した第2映像の各ピクチャを、前記ベースビュービデオストリームにおいて当該ピクチャに対応するピクチャと同時刻の前記MPEG-2形式のストリームのピクチャを参照画像として圧縮符号化したストリームを生成する
ことを特徴とする映像符号化プログラム。 - 請求項1記載の映像符号化装置により送出されるストリームを取得して復号及び再生する映像再生装置であって、
前記MPEG-2形式のストリームを復号する第1復号手段と、
前記ベースビュービデオストリームを復号し、前記ディペンデントビューストリームの復号において第2映像の各ピクチャを、前記ベースビュービデオストリームにおいて当該ピクチャに対応するピクチャと同時刻の前記第1復号手段により復号されたピクチャを参照画像として復号する第2復号手段と、
前記第2復号手段により得られた前記第2映像を再生する再生手段とを備える
ことを特徴とする映像再生装置。 - 請求項12記載の映像符号化方法により送出されるストリームを取得して復号及び再生する映像再生方法であって、
前記MPEG-2形式のストリームを復号する第1復号ステップと、
前記ベースビュービデオストリームを復号し、前記ディペンデントビューストリームの復号において第2映像の各ピクチャを、前記ベースビュービデオストリームにおいて当該ピクチャに対応するピクチャと同時刻の前記第1復号ステップにより復号されたピクチャを参照画像として復号する第2復号ステップと、
前記第2復号ステップにより得られた前記第2映像を再生する再生ステップとを含む
ことを特徴とする映像再生方法。 - コンピュータを、請求項1記載の映像符号化装置により送出されるストリームを取得して復号及び再生する映像再生装置として機能させるための映像再生プログラムであって、
前記コンピュータを、
前記MPEG-2形式のストリームを復号する第1復号手段と、
前記ベースビュービデオストリームを復号し、前記ディペンデントビューストリームの復号において第2映像の各ピクチャを、前記ベースビュービデオストリームにおいて当該ピクチャに対応するピクチャと同時刻の前記第1復号手段により復号されたピクチャを参照画像として復号する第2復号手段と、
前記第2復号手段により得られた前記第2映像を再生する再生手段として機能させる
ことを特徴とする映像再生プログラム。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
MX2013008311A MX2013008311A (es) | 2011-02-16 | 2012-02-15 | Codificador de video, metodo de codificacion de video, programa de codificacion de video, metodo de reproduccion de video y programa de reproduccion de video. |
JP2012557833A JP5906462B2 (ja) | 2011-02-16 | 2012-02-15 | 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム |
BR112013020867A BR112013020867A2 (pt) | 2011-02-16 | 2012-02-15 | codificador de vídeo, método de codificação de vídeo, programa de codificação de vídeo, dispositivo de reprodução de vídeo, método de reprodução de vídeo, e programa de reprodução de vídeo |
US13/978,458 US9277217B2 (en) | 2011-02-16 | 2012-02-15 | Video coding device for coding videos of a plurality of qualities to generate streams and video playback device for playing back streams |
CA2823288A CA2823288C (en) | 2011-02-16 | 2012-02-15 | Video encoder, video encoding method, video encoding program, video reproduction device, video reproduction method, and video reproduction program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161443353P | 2011-02-16 | 2011-02-16 | |
US61/443,353 | 2011-02-16 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012111320A1 true WO2012111320A1 (ja) | 2012-08-23 |
Family
ID=46672264
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2012/000974 WO2012111320A1 (ja) | 2011-02-16 | 2012-02-15 | 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム |
Country Status (7)
Country | Link |
---|---|
US (1) | US9277217B2 (ja) |
JP (1) | JP5906462B2 (ja) |
BR (1) | BR112013020867A2 (ja) |
CA (1) | CA2823288C (ja) |
MX (1) | MX2013008311A (ja) |
TW (1) | TW201251467A (ja) |
WO (1) | WO2012111320A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104604242A (zh) * | 2012-09-07 | 2015-05-06 | 索尼公司 | 传输设备、传输方法、接收设备和接收方法 |
CN114394100A (zh) * | 2022-01-12 | 2022-04-26 | 深圳力维智联技术有限公司 | 一种无人巡逻车控制系统及无人车 |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8081214B2 (en) | 2004-10-12 | 2011-12-20 | Enforcement Video, Llc | Method of and system for mobile surveillance and event recording |
WO2009102480A2 (en) | 2008-02-15 | 2009-08-20 | Enforcement Video, Llc | System and method for multi-resolution storage of images |
EP2645724A4 (en) * | 2011-11-11 | 2014-08-06 | Sony Corp | SENDING DEVICE, TRANSMISSION PROCEDURE, RECEPTION DEVICE AND RECEPTION PROCEDURE |
US9066082B2 (en) * | 2013-03-15 | 2015-06-23 | International Business Machines Corporation | Forensics in multi-channel media content |
US9894422B2 (en) * | 2014-02-18 | 2018-02-13 | Lg Electronics Inc. | Method and apparatus for transreceiving broadcast signal for panorama service |
US9348495B2 (en) | 2014-03-07 | 2016-05-24 | Sony Corporation | Control of large screen display using wireless portable computer and facilitating selection of audio on a headphone |
BR112017017792A2 (ja) | 2015-02-27 | 2018-04-10 | Sony Corporation | A sending set, a transmission method, a receiving set, and a receiving method |
CN106776663B (zh) * | 2015-11-25 | 2020-11-10 | 腾讯科技(深圳)有限公司 | 音频文件的压缩方法和装置 |
US10341605B1 (en) | 2016-04-07 | 2019-07-02 | WatchGuard, Inc. | Systems and methods for multiple-resolution storage of media streams |
US10769818B2 (en) * | 2017-04-09 | 2020-09-08 | Intel Corporation | Smart compression/decompression schemes for efficiency and superior results |
CN113949930B (zh) * | 2020-07-17 | 2024-03-12 | 晶晨半导体(上海)股份有限公司 | 选取参考帧的方法、电子设备和存储介质 |
US20220311984A1 (en) * | 2021-03-26 | 2022-09-29 | Lightspace Technologies, SIA | System and method for rendering three-dimensional image content |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10191394A (ja) * | 1996-12-24 | 1998-07-21 | Sharp Corp | 多視点画像符号化装置 |
JPH1185966A (ja) * | 1997-07-18 | 1999-03-30 | Sony Corp | 画像信号多重化装置および方法、画像信号逆多重化装置および方法、並びに伝送媒体 |
JP2010512096A (ja) * | 2006-12-08 | 2010-04-15 | 韓國電子通信研究院 | 非リアルタイムベースのデジタル実感放送送受信システム及びその方法 |
JP2011004411A (ja) * | 2009-02-19 | 2011-01-06 | Panasonic Corp | 半導体集積回路 |
JP2011019224A (ja) * | 2009-07-07 | 2011-01-27 | Korea Electronics Telecommun | デジタル放送システムのステレオスコピックビデオ送受信方法およびその装置 |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6055012A (en) * | 1995-12-29 | 2000-04-25 | Lucent Technologies Inc. | Digital multi-view video compression with complexity and compatibility constraints |
EP2175665B1 (en) * | 1996-12-04 | 2012-11-21 | Panasonic Corporation | Optical disk for high resolution and three-dimensional video recording, optical disk reproduction apparatus, and optical disk recording apparatus |
CN1279757C (zh) | 1997-07-18 | 2006-10-11 | 索尼公司 | 图像信号复用装置和方法、分解装置和方法 |
US7133449B2 (en) * | 2000-09-18 | 2006-11-07 | Broadcom Corporation | Apparatus and method for conserving memory in a fine granularity scalability coding system |
JP4346868B2 (ja) * | 2002-06-17 | 2009-10-21 | 株式会社日立製作所 | 動画符号化装置、及び動画記録再生装置、及び動画再生装置、及び動画符号化方法ならびに動画記録再生方法 |
US7720999B2 (en) * | 2002-11-26 | 2010-05-18 | Qualcomm Incorporated | System and method for optimizing multimedia compression using plural encoders |
JP2006237656A (ja) * | 2003-02-28 | 2006-09-07 | Secom Co Ltd | 符号化信号分離・合成装置、差分符号化信号生成装置、差分符号化信号抽出装置、符号化信号分離・合成方法、符号化信号分離・合成プログラム |
US7610603B2 (en) * | 2004-03-26 | 2009-10-27 | Broadcom Corporation | Multistream video communication with staggered access points |
US20060271990A1 (en) * | 2005-05-18 | 2006-11-30 | Rodriguez Arturo A | Higher picture rate HD encoding and transmission with legacy HD backward compatibility |
KR101366092B1 (ko) * | 2006-10-13 | 2014-02-21 | 삼성전자주식회사 | 다시점 영상의 부호화, 복호화 방법 및 장치 |
US20090320081A1 (en) * | 2008-06-24 | 2009-12-24 | Chui Charles K | Providing and Displaying Video at Multiple Resolution and Quality Levels |
JP2010041336A (ja) * | 2008-08-04 | 2010-02-18 | Toshiba Corp | 画像処理装置、および画像処理方法 |
EP2400771B1 (en) * | 2009-02-19 | 2014-11-19 | Panasonic Corporation | Recording medium, reproduction device, and integrated circuit |
JP2010263615A (ja) * | 2009-04-08 | 2010-11-18 | Sony Corp | 情報処理装置、情報処理方法、再生装置、再生方法 |
JP2010245970A (ja) * | 2009-04-08 | 2010-10-28 | Sony Corp | 再生装置、再生方法、およびプログラム |
US20110216827A1 (en) * | 2010-02-23 | 2011-09-08 | Jiancong Luo | Method and apparatus for efficient encoding of multi-view coded video data |
KR101362943B1 (ko) * | 2010-06-30 | 2014-02-17 | 한국전자통신연구원 | 통신 시스템에서 데이터 송수신 장치 및 방법 |
-
2012
- 2012-02-15 CA CA2823288A patent/CA2823288C/en not_active Expired - Fee Related
- 2012-02-15 MX MX2013008311A patent/MX2013008311A/es active IP Right Grant
- 2012-02-15 US US13/978,458 patent/US9277217B2/en not_active Expired - Fee Related
- 2012-02-15 JP JP2012557833A patent/JP5906462B2/ja active Active
- 2012-02-15 WO PCT/JP2012/000974 patent/WO2012111320A1/ja active Application Filing
- 2012-02-15 BR BR112013020867A patent/BR112013020867A2/pt not_active IP Right Cessation
- 2012-02-16 TW TW101105084A patent/TW201251467A/zh unknown
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10191394A (ja) * | 1996-12-24 | 1998-07-21 | Sharp Corp | 多視点画像符号化装置 |
JPH1185966A (ja) * | 1997-07-18 | 1999-03-30 | Sony Corp | 画像信号多重化装置および方法、画像信号逆多重化装置および方法、並びに伝送媒体 |
JP2010512096A (ja) * | 2006-12-08 | 2010-04-15 | 韓國電子通信研究院 | 非リアルタイムベースのデジタル実感放送送受信システム及びその方法 |
JP2011004411A (ja) * | 2009-02-19 | 2011-01-06 | Panasonic Corp | 半導体集積回路 |
JP2011019224A (ja) * | 2009-07-07 | 2011-01-27 | Korea Electronics Telecommun | デジタル放送システムのステレオスコピックビデオ送受信方法およびその装置 |
Non-Patent Citations (1)
Title |
---|
TETSUO NOZAWA ET AL.: "Sekai de Hajimaru 3D Hoso Gyokai Hyojun Meguru Arasoi Gekika", NIKKEI ELECTRONICS, NIKKEI BUSINESS PUBLICATIONS, INC., no. 1028, 19 April 2010 (2010-04-19), pages 55 - 62 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104604242A (zh) * | 2012-09-07 | 2015-05-06 | 索尼公司 | 传输设备、传输方法、接收设备和接收方法 |
CN104604242B (zh) * | 2012-09-07 | 2018-06-05 | 索尼公司 | 发送设备、发送方法、接收设备和接收方法 |
US10432957B2 (en) | 2012-09-07 | 2019-10-01 | Saturn Licensing Llc | Transmission device, transmitting method, reception device, and receiving method |
US10951910B2 (en) | 2012-09-07 | 2021-03-16 | Saturn Licensing Llc | Transmission device, transmitting method, reception device, and receiving method |
US11700388B2 (en) | 2012-09-07 | 2023-07-11 | Saturn Licensing Llc | Transmission device, transmitting method, reception device, and receiving method |
CN114394100A (zh) * | 2022-01-12 | 2022-04-26 | 深圳力维智联技术有限公司 | 一种无人巡逻车控制系统及无人车 |
CN114394100B (zh) * | 2022-01-12 | 2024-04-05 | 深圳力维智联技术有限公司 | 一种无人巡逻车控制系统及无人车 |
Also Published As
Publication number | Publication date |
---|---|
MX2013008311A (es) | 2013-09-06 |
JPWO2012111320A1 (ja) | 2014-07-03 |
CA2823288C (en) | 2019-05-07 |
US9277217B2 (en) | 2016-03-01 |
JP5906462B2 (ja) | 2016-04-20 |
TW201251467A (en) | 2012-12-16 |
CA2823288A1 (en) | 2012-08-23 |
BR112013020867A2 (pt) | 2016-09-27 |
US20130287090A1 (en) | 2013-10-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5906462B2 (ja) | 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム | |
WO2012111325A1 (ja) | 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム | |
JP6229962B2 (ja) | 符号化装置及び符号化方法 | |
KR101812612B1 (ko) | 3d 비디오 신호를 생성하는 방법 | |
EP3399755B1 (en) | Depth map delivery formats for stereoscopic and auto-stereoscopic displays | |
WO2012057164A1 (ja) | 符号化方法、表示装置、復号方法 | |
CN103098462A (zh) | 编码方法、显示装置以及解码方法 | |
JP6008292B2 (ja) | ビデオストリームの映像のデータ作成装置及び再生装置 | |
CN102972030A (zh) | 用于生成和重建立体兼容视频流的方法及相关的编码和解码设备 | |
US9357200B2 (en) | Video processing device and video processing method | |
WO2012169204A1 (ja) | 送信装置、受信装置、送信方法及び受信方法 | |
JP2011211605A (ja) | 画像再生装置、画像再生方法及び画像再生用プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12747400 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2823288 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13978458 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2013/008311 Country of ref document: MX |
|
ENP | Entry into the national phase |
Ref document number: 2012557833 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112013020867 Country of ref document: BR |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 12747400 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 112013020867 Country of ref document: BR Kind code of ref document: A2 Effective date: 20130815 |