US20160360248A1 - Method and apparatus for decoding segments of an audiovisual stream - Google Patents
Method and apparatus for decoding segments of an audiovisual stream Download PDFInfo
- Publication number
- US20160360248A1 US20160360248A1 US15/241,573 US201615241573A US2016360248A1 US 20160360248 A1 US20160360248 A1 US 20160360248A1 US 201615241573 A US201615241573 A US 201615241573A US 2016360248 A1 US2016360248 A1 US 2016360248A1
- Authority
- US
- United States
- Prior art keywords
- metadata
- section
- stream
- segment
- processing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 20
- BMQYVXCPAOLZOK-NJGYIYPDSA-N D-monapterin Chemical compound C1=C([C@H](O)[C@@H](O)CO)N=C2C(=O)NC(N)=NC2=N1 BMQYVXCPAOLZOK-NJGYIYPDSA-N 0.000 description 55
- 238000000605 extraction Methods 0.000 description 24
- 238000010586 diagram Methods 0.000 description 20
- 230000010365 information processing Effects 0.000 description 20
- 239000000284 extract Substances 0.000 description 14
- 238000006243 chemical reaction Methods 0.000 description 13
- 230000001360 synchronised effect Effects 0.000 description 12
- 230000005540 biological transmission Effects 0.000 description 11
- 238000004891 communication Methods 0.000 description 6
- 230000004044 response Effects 0.000 description 6
- 238000003672 processing method Methods 0.000 description 3
- 230000004927 fusion Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000003466 anti-cipated effect Effects 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000008929 regeneration Effects 0.000 description 1
- 238000011069 regeneration method Methods 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/08—Systems for the simultaneous or sequential transmission of more than one television signal, e.g. additional information signals, the signals occupying wholly or partially the same frequency band, e.g. by time division
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/242—Synchronization processes, e.g. processing of PCR [Program Clock References]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/231—Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion
- H04N21/23106—Content storage operation, e.g. caching movies for short term storage, replicating data over plural servers, prioritizing data for deletion involving caching operations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
- H04N21/2353—Processing of additional data, e.g. scrambling of additional data or processing content descriptors specifically adapted to content descriptors, e.g. coding, compressing or processing of metadata
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/238—Interfacing the downstream path of the transmission network, e.g. adapting the transmission rate of a video stream to network bandwidth; Processing of multiplex streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
- H04N21/4307—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
- H04N21/4307—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
- H04N21/43074—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of additional data with content streams on the same device, e.g. of EPG data or interactive icon with a TV program
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/458—Scheduling content for creating a personalised stream, e.g. by combining a locally stored advertisement with an incoming stream; Updating operations, e.g. for OS modules ; time-related management operations
- H04N21/4586—Content update operation triggered locally, e.g. by comparing the version of software modules in a DVB carousel to the version stored locally
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S707/00—Data processing: database and file management or data structures
- Y10S707/99941—Database schema or data structure
- Y10S707/99942—Manipulating data structure, e.g. compression, compaction, compilation
Definitions
- the present invention relates to an information provision apparatus, information receiving apparatus, and storage medium, and relates in particular to an information provision apparatus, information receiving apparatus, and storage medium for video/audio, data, etc., operating via broadcast media such as digital broadcasting and communication media such as the Internet.
- EPG information as well as audio/video information is provided by interleaving EPG (Electric Program Guide)-“Standard specification for program arrangement information used in digital broadcasting ARIB STD-B10 Version 1.1” or “prETS 300 468 Digital Broadcasting systems for television, sound and data services-Specification for Service Information (SI) in Digital Video Broadcasting (DVB) systems”) used in CS digital broadcasting, in an audio/video PES (Packetized Elementary Stream) using an MPEG-2 (Motion Picturecoding Experts Group phase 2-“ISO/IEC 13818-1 to 3”) private section.
- EPG Electronic Program Guide
- SDB10 Version 1.1 Digital Broadcasting
- MPEG-2 Motion Picturecoding Experts Group phase 2-“ISO/IEC 13818-1 to 3
- FIG. 15 is a block diagram of a conventional information processing system.
- An information provision node 1501 is provided with a storage section 1502 in which an AV stream and metadata for describing the AV stream are stored. Also provided in the information provision node 1501 is an information provision section 1504 that multiplexes the AV stream and metadata stored in the storage section 1502 and generates and outputs a multiplex stream 1503 . The information provision section 1504 transmits the multiplex stream 1503 to an information usage node 1506 via a network 1505 .
- the information usage node 1506 is provided with an information usage section 1507 that extracts an AV stream and metadata from a multiplex stream and executes processing on them in order to use them.
- the information usage node 1506 is also provided with a storage section 1508 that stores the AV stream and metadata extracted by the information usage section 1507 .
- the information usage section 1507 reads the AV stream and metadata stored in the storage section 1508 in order to use them.
- FIG. 16 is a block diagram of a conventional information provision section.
- the information provision section 1504 is provided with an access section 1601 that reads an AV stream and metadata from the storage section 1502 .
- the access section 1601 outputs an AV scream 1602 and metadata 1603 to a multiplexing section 1604 .
- the multiplexing section 1604 transmits to the information usage node 1506 a multiplex stream 1503 that multiplexes the AV stream 1602 and metadata 1603 .
- multiplex stream generation processing by the multiplexing section 1601 will be described using FIG. 17 .
- the drawing indicated by reference numeral 1503 in the drawing shows the MPEG-2 TS (Transport Stream) PES packet layer, and shows a multiplex stream.
- the drawing indicated by reference numeral 1702 shows a video PES packet
- the drawing indicated by reference numeral 1703 shows an audio PES packet
- the drawing indicated by reference numeral 1703 shows a private PES packet.
- 1603 indicates the metadata PES packet layer, in which 1704 is a first PES packet comprising metadata and 1705 is a second PES packet comprising metadata.
- the multiplexing section 1604 divides the metadata 1603 to make private PES packets, inserts the first PES packet 1704 and second PES packet 1705 in order as appropriate between AV streams consisting of video PES packets 1701 and audio PES packets 1702 , and obtains a multiplex stream 1503 that is an MPEG-2 TS.
- Metadata is AV stream ancillary data-for example, small amounts of data such as titles-processing has been performed with metadata alone. That is to say, it has not been necessary to provide time synchronization of metadata with an AV stream. Therefore, since conventional metadata does not have a configuration that provides for synchronization with an AV stream, metadata has been packetized using virtually the same size, and has been inserted as appropriate between AV streams at virtually equal intervals.
- the multiplexing section 1604 then sends this multiplex stream 1503 to the information usage node 1506 .
- FIG. 18 a block diagram of a conventional information usage section.
- the information usage section 1507 is provided with an extraction section 1803 that performs separation and extraction, and output, of an AV stream 1801 and metadata 1802 .
- the extraction section 1603 outputs the separated and extracted AV stream 1801 and metadata 1302 to an access section 1804 .
- the access section 1801 stores the AV stream 1801 and metadata 1802 input from the extraction section 1803 in a storage section 1508 . Also, the access section 1804 outputs the AV stream 1805 and metadata 1806 read from the storage section 1508 to a display section 1807 .
- the display section 1807 displays either or both of the AV stream 1805 and metadata 1806 input from the access section 1004 .
- FIG. 19 is a processing flowchart of a conventional information usage section.
- the extraction section 1803 performs metadata parsing-that is, syntax analysis (ST 1901 ). Then, execution of the processing of the access section 1804 and display section 180 is performed (ST 1902 ).
- a conventional information processing system can display a description relating to AV information, in addition to AV information, by means of the information usage node 1506 by having the information provision node 1501 transmit a multiplex stream multiplexing an AV stream and metadata to the information usage node 1506 .
- Metadata parsing cannot be carried out until all the metadata has been acquired. For example, if metadata begins with ⁇ metadata>, metadata parsing cannot be carried out until data ⁇ /metadata>indicating the end of the metadata arrives.
- the metadata processing time is closely tied to the AV stream display or processing time, and since an AV stream is processed in accordance with the metadata itself, processing cannot be started until all the metadata has been received. Therefore, in a conventional information processing system, there is a problem in that it is difficult to process an AV stream in small units.
- Metadata is distributed virtually uniformly in a multiplex stream.
- a large AV stream quantity must be read by the time all the metadata is read. Consequently, there are problems relating to inter-node response time delays and increased network traffic.
- the present invention is provided with a synchronization section which synchronizes a data stream segment with a unit of metadata corresponding to it, and a capsulization section which capsulizes a data stream packet and metadata unit packet after synchronization and generates a capsulized stream.
- the present invention is provided with an extraction section which extracts from a capsulized stream a content data stream and metadata for describing or processing that content, a synchronization section which synchronizes metadata unitized with respect to an extracted data stream segment unit by unit with a content data stream and the corresponding metadata unit, and a processing section which processes synchronized metadata unit by unit.
- the present invention uses a structured description for metadata and metadata units, and structured description re-format is performed from metadata to units and from units to metadata.
- FIG. 1 is a block diagram of an information processing system according to Embodiment 1 of the present invention.
- FIG. 2 is a block diagram of an information processing section according to Embodiment 1;
- FIG. 3A is a drawing showing an AV stream according to Embodiment 1;
- FIG. 3B is a drawing showing metadata according to Embodiment 1;
- FIG. 4A is a drawing showing DTD of XML of metadata according to Embodiment 1;
- FIG. 4B is a drawing showing DTD of XML of an MPU according to Embodiment 1;
- FIG. 5A is a drawing showing an instance of XML of metadata according to Embodiment 1;
- FIG. 5B is a drawing showing an instance of XML of an MPU according to Embodiment 1;
- FIG. 6 is a drawing showing the syntax of metadata according to Embodiment 1;
- FIG. 7 is a drawing for explaining the operation of a capsulization section according to Embodiment 1;
- FIG. 8 is a block diagram of an information usage section according to Embodiment 2 of the present invention.
- FIG. 9 is a processing flowchart showing the metadata processing operations of an information usage node according to Embodiment 2of the present invention.
- FIG. 10 is a block diagram of an information usage section according to Embodiment 3of the present invention.
- FIG. 11 is a block diagram of an information usage section according to Embodiment 4 of the present invention.
- FIG. 12 is a block diagram of an information processing system according to Embodiment 5 of the present invention.
- FIG. 13 is a block diagram of an information processing section according to Embodiment 5.
- FIG. 14 is a block diagram of an information usage section according to Embodiment 4 of the present invention according to Embodiment 6;
- FIG. 15 is a block diagram of a conventional information processing system
- FIG. 16 is a detailed drawing of a conventional information provision section
- FIG. 17 is a drawing showing the configuration of a conventional multiplex stream
- FIG. 18 is a detailed drawing of a conventional information usage section.
- FIG. 19 is a processing flowchart for a conventional extraction section.
- FIG. 1 is a block diagram of an information processing system according to Embodiment 1.
- An information provision node 101 is provided with a storage section 102 in which an AV stream and AV stream related metadata are stored.
- the metadata is data that describes the related AV stream, or data for processing the metadata itself, or the like.
- an information provision section 101 that multiplexes the AV stream and metadata stored in the storage section 102 and generates and outputs a capsulized stream 103 .
- the information provision section 104 transmits the capsulized stream 103 via a network 105 to an information usage node 106 , which is an apparatus on the information receiving side.
- the information usage node 106 is provided with an information usage section 107 that extracts an AV stream and metadata from the capsulized stream 103 and executes predetermined processing on them in order to use them.
- the information usage node 106 is also provided with a storage section 108 that stores the AV stream and metadata extracted by the information usage section 107 .
- the information usage section 107 reads the AV stream and metadata stored in the storage section 108 in order to use them.
- FIG. 2 is a block diagram of an information provision section according to Embodiment 1.
- the information provision section 104 is provided with an access section 201 that reads an AV stream and metadata from the storage section 102 .
- the access section 201 outputs an AV stream 202 and metadata 203 to a synchronization section 204 .
- the synchronization section 204 implements time synchronization for the AV stream 202 and metadata 203 read by the access section 201 , and outputs the synchronized AV stream 205 and metadata 206 to a capsulization section 207 .
- the capsulization section 207 capsulizes the synchronized AV stream 205 and metadata 206 , and transmits them to the information usage node 106 as a capsulized stream 103 .
- the present invention unitizes metadata to enable metadata to be executed in parts. Then, AV stream segments and corresponding metadata units are synchronized, synchronized data stream packets and metadata unit packets are capsulized, and a capsulized stream is generated.
- the AV stream 202 has video PES pockets 301 and audio PES packets 302 interleaved to form a stream.
- a mode 53 described whereby an AV stream 202 is scored in the storage section 102 but a mode is also possible whereby a video stream and audio stream are stored.
- the metadata 203 is configured so as to have a plurality of MPUs (Metadata Processing Units) 303 .
- the thus configured metadata 203 and AV stream 202 are read from the storage section 102 by the access section 201 . Then the access section 201 outputs the read AV stream 202 and metadata 203 to the synchronization section 204 .
- FIG. 4A and FIG. 4B are drawings showing DTD of XML.
- 101 is a drawing showing a metadata definition (metadata.dtd) that defines the metadata 203 .
- FIG. 4B the drawing indicated by reference numeral 402 shows an MPU definition (mpu.dtd) that defines an MPU 303 .
- the metadata definition 401 defines the metadata 203 as having one or more MPUs 303 . For the contents of an MPU 303 , referencing the MPU definition 402 is defined.
- the MPU definition 402 defines an MPU 303 as having one or more element_data items. For the contents of element_data, referencing user_defined.dtd is defined. Also, the MPU definition 402 defines an MPU 303 as having a serial number no assigned.
- FIG. 5A the drawing indicated by reference numeral 501 shows metadata (XML instance) whereby metadata 203 is given a structured description according to metadata definition 401
- the drawing indicated by reference numeral 502 shows an MPU (XML instance) whereby an MPU 303 is given a structured description according to MPU definition 902 .
- metadata 203 is represented by a collection of MPU definitions 102 .
- metadata (XML instance) 501 what gives a structured description of metadata 203 is metadata (XML instance) 501 .
- the metadata (XML instance) 501 instance includes a plurality of MPUs 303 .
- metadata 203 is stored in the storage section 102 as metadata (XML instance) 501 .
- an MPU 303 is represented by a collection of metadata defined by user_defined.dtd.
- MPU (XML instance) 502 what gives a structured description of MPU 303 for each MPU is MPU (XML instance) 502 .
- MPU (XML instance) 502 includes a plurality of user_defined.dtd items.
- MPU 303 is stored in the storage section 102 as MPU (XML instance) 502 .
- An MPU 303 has contents ⁇ mpu> to ⁇ /mpu>. That is to say, if there is information from ⁇ mpu> to ⁇ /mpu>, the synchronization section 204 can grasp MPU 303 contents and can perform MPU 303 processing. For this reason, when picking out an MPU 303 from metadata 203 , the synchronization section 204 extracts the contents on the inside of a tag called an MPU tag (here, ⁇ mpu>) defined by an MPU definition 402 .
- an MPU tag here, ⁇ mpu>
- the synchronization section 204 can perform metadata 203 processing for each MPU 303 , and also closely synchronize the AV data 202 and metadata 203 .
- FIG. 6 shows the syntax of metadata according to Embodiment 1 and Embodiment 2.
- metadata_type 601 is the metadata type such as position information, content information, or program.
- metadata_subtype 602 is the concrete metadata type such as GPS or structured description (MPEG-7).
- MPU_length 603 is the data length as a number of bytes from immediately after the MPU_length field co the end of the MPU.
- An MPU is composed of one or more PSS packets, and is the regeneration unit of metadata divided when a Metadata Elementary Scream is encoded.
- media_sync_flag 604 is a flag indicating the presence or absence of synchronization between the AV stream and metadata.
- overwrite_flag 605 is a flag indicating whether the previous metadata is to be overwritten.
- element_data_length 606 is the data byte length (M) of element_data 609 .
- start_time ( ) 607 is the start time of a segment that is a part of the AV stream indicated by the metadata
- duration ( ) 608 is the continuation time of a segment that is part of the AV stream indicated by the metadata.
- element_data 609 is the actual data of the metadata.
- coding uses syntax 610 from else downward even when the metadata data quantity is small and unitization is not performed.
- the synchronization section 204 capsulizes the AV stream segment for processing specified by the first packet's processing start time 607 and duration 608 . and part of the metadata 203 corresponding to the segment for processing, as a capsulized stream (private PES).
- an MPU 303 When metadata 203 is PES-packetized, an MPU 303 is packetized together with the AV stream segment first packet processing start time (start_time), duration ( ) 606 , and actual data of the metadata as an element (element_data) in the metadata syntax shown in FIG. 6 .
- an MPU 303 it is possible for an MPU 303 to have information for maintaining synchronization with the AV stream 202 .
- synchronization is maintained between the MPU 303 and AV stream 202 .
- metadata 203 operation can be determined on the information provision node 101 side.
- an MPU 303 is composed of two packets—a first PES packet 701 and a second PES packet 702 —as shown in FIG. 7 .
- the operations whereby the synchronization section 204 packetizes an MPU 303 into private PES packets and interleaves these with video PES packets 301 and audio PES packets 302 in this case will be described using FIG. 1 . How many packets an MPU 303 is made into can be determined arbitrarily according to the MPU 303 size and the packet size.
- the first PES packet 701 and second PES packet 702 are placed as private PES packets 708 earlier in time than the first packet 703 so that the first PES packet 701 and second PES packet 702 are processed before the processing start time (start_time) 705 of the first packet of the corresponding AV stream segment.
- the second PES packet 702 arrival time t 704 and the corresponding first packet 703 processing start time (start_time) 705 difference At 706 are assigned sufficient times for the information usage section 107 , which is on the information receiving side, to generate an MPU 303 from the first PES packet 701 and second PES packet 702 , and execute processing based on the contents of the generated MPU 303 .
- the AV stream 205 and metadata 205 synchronized by the synchronization section 204 in this way are input to the capsulization section 207 .
- the capsulization section 207 capsulizes the input AV stream 205 and metadata 206 , and transmits them as a capsulized stream 103 .
- Metadata can be re-formatted unit by unit and capsulized with an AV stream by providing a synchronization section 204 that maintains synchronization of the AV stream and metadata, and a capsulization section 207 that capsulizes metadata unit by unit with the AV stream.
- Embodiment 1 by using a structured description written using for metadata and metadata units, and performing structured description re-format from metadata to units and from units to metadata, it is possible to provide extensibility for metadata for processing an AV stream, and extend the degree of freedom for designing metadata.
- a structured description written in XML, etc. it is possible for a structured description written in XML, etc., to be used directly as metadata.
- FIG. 8 is a block, diagram of an information usage section 107 according to Embodiment 2.
- the information usage section 107 is provided with an extraction section 803 that performs separation and extraction, and output, of an AV stream 801 and metadata 802 .
- the extraction section 803 outputs the extracted AV stream 801 and metadata 802 to an access section 804 .
- the access section 804 records the AV stream 801 and metadata 802 in a storage section 108 . Also, the access section 804 reads an AV stream 805 and metadata 806 stored in the storage section 108 , and outputs them to a synchronization section 307 .
- the synchronization section 807 performs time synchronization every MPU 303 for the AV stream 805 and metadata 806 read by the access section 804 , and outputs them to a core processing section 808 .
- the core processing section 808 is provided with a display section 809 .
- the display section 809 performs time synchronization and display of the input synchronized AV stream 810 and metadata 811 .
- the information usage section 107 extracts an AV stream 801 and metadata 802 from the capsulized stream 103 in the extraction section 803 . Then, in the synchronization section 807 , the corresponding metadata 802 unitized in accordance with AV stream 801 segments is synchronized with the AV stream 801 unit by unit. Then the synchronized metadata 811 and AV stream 810 are displayed unit by unit by the display section 809 .
- the extraction section 803 extracts an AV stream and metadata from the received capsulized stream 103 .
- the information usage section 107 performs MPU 303 pursing (ST 901 ).
- a check is performed as to whether the MPUs 303 are to be merged and re-formatted as metadata 802 (ST 902 ).
- a check is performed as to whether MPU 303 execution is to be performed unit by unit (ST 903 ).
- processing is executed by the core processing section 808 (ST 904 ). Then MPU merging is performed in the information usage section 107 (ST 905 ).
- this processing is display processing, but it may also be conversion processing or transfer processing as in other embodiments to be described hereafter.
- rendering—that is to say, formatting—of the metadata is performed from the MPUs collected together in ST 906 .
- Metadata formatted oil the basis of this event is stored in the storage section 108 .
- the core processing section 808 reads this formatted data and performs various kinds of processing.
- an MPU merge is performed (ST 908 ). Then, in the information usage section 107 , judgment as to the presence of an MPU time or number limit—that is, an event related to completion of an MPU merge—is performed (ST 909 ), and ST 906 is repeated until the occurrence of an event. Rendering of the metadata is then performed from the MPUs collected together in processing P 107 . Then, in the information usage section 107 , rendering—that is to say, formatting—of the metadata is performed from the MPUs collected together in ST 906 (ST 910 ). Metadata formatted on the basis of this event is scored in the storage section 108 . Then the core processing section 808 reads this formatted data and performs various kinds of processing.
- processing is executed sequentially (ST 911 ). Then, in the information usage section 107 , judgment as to the presence of an MPU time or number limit—that is, an event that indicates an MPU processing unit—is performed (ST 912 ), and ST 911 is repeated until the occurrence of an event.
- the extraction method can be changed as appropriate according to the contents contained in MPUs 303 .
- the information usage section 107 extracts an AV stream 801 and metadata 802 from the capsulized stream 103 input by the extraction section 803 , and outputs them to the access section 804 .
- the access section 804 After recording the AV stream 801 and metadata 802 in the storage section 108 , the access section 804 reads an AV stream 805 and metadata 806 , and outputs them to the synchronization section 807 .
- the synchronization section 807 performs time synchronization every MPU 303 for the AV stream 805 and metadata 806 read by the access section 804 , and outputs them to the core processing section 808 .
- the display section 809 performs time synchronization and display of the input AV stream 810 and metadata 811 .
- close synchronization of the metadata and AV stream processing time can be performed by providing an extraction, section 803 for separating and extracting an AV stream and metadata, an access section 804 for reading and writing an AV stream and metadata in a storage section 108 , a synchronization section 807 for performing synchronization of the read AV stream and metadata processing, and a display section 809 , which is a core processing section 808 .
- information relating to the display method used by the display section 809 of the core processing section 808 can be provided as metadata.
- Information relating to the display method includes position information for displaying metadata related information, display size information, and display update information.
- an appropriate method for displaying metadata can be sent to the information provision node 101 by the information usage node 106 .
- metadata can be displayed appropriately by the information usage node 106 . Therefore, if metadata is an advertisement or the like, it is possible to make a specification that allows the advertisement to be displayed at the desired time, and if metadata is information related to program descriptions, it is possible to display the descriptive information so as not to interfere with images.
- Embodiment 2 by using a structured description written using XML for metadata and metadata units, and performing structured description re-format from metadata to units and from units to metadata, it is possible to extend the degree of freedom for designing metadata for processing an AV stream, and a structured description written in XML, etc., can be used directly as metadata.
- FIG. 10 is a block diagram of an information usage section 1001 according to Embodiment 3. Parts identical to those that have already been described are assigned the same reference numerals, and a description of these parts is omitted.
- the information usage section 1001 according to Embodiment 3 has the core processing section 808 of the information usage section 1001 according to Embodiment 2 replaced by a core processing section 1002 .
- the information usage section 1001 will be described centering on the core processing section 1002 .
- the core processing section 1002 is provided with a transfer section 1003 and a capsulization section 1006 .
- the transfer section 1003 performs settings, such as a destination setting, for transferring an AV scream 810 and metadata 811 input from the synchronization section 807 to another information usage node.
- the transfer section 1003 performs time synchronization every MPU 303 , and outputs an AV stream 1004 and metadata 1005 to the capsulization section 1006 .
- the capsulization section 1006 recapsulizes the input AV stream 1004 and metadata 1005 and transmits them to another node as a capsulized stream 1007 . Since the capsulization section 1006 recapsulizes the AV stream 1001 and metadata 1005 in this way, load sharing can be performed while maintaining close synchronization between the metadata and AV stream processing times.
- the operation of the capsulization section 1006 is similar to that of the capsulization section 207 according to Embodiment 1, and so a detailed description will be omitted here.
- the information usage section 1101 extracts an AV stream 801 and metadata 802 from the capsulized stream 103 input by the extraction section 803 , and outputs them to the access section 804 .
- the access section 804 After recording the AV stream 801 and metadata 802 in the storage section 108 , the access section 804 reads an AV stream 805 and metadata 806 , and outputs them to the synchronization section 807 .
- the synchronization section 807 performs time synchronization every MPU 303 for the AV stream 805 and metadata 806 read by the access section 301 , and outputs them to the core processing section 1002 .
- the core processing section 1002 per forms settings for transferring the AV stream 810 and metadata 811 input by the transfer section 1003 to another information usage node, and performs time synchronization and output to the capsulization section 1006 every MPU 303 .
- the capsulization section 1006 recapsulizes the input AV stream 1001 and metadata 1005 and transmits them to another node as a capsulized stream 1007 .
- the transfer section 1002 By configuring the information usage section 1001 as described above, it is possible for the transfer section 1002 to perform settings for transferring the AV stream 810 and metadata 811 input from the synchronization section 807 to another information usage node, perform time synchronization and output to the capsulization step 23 every MPU 303 , and for the capsulization section 1006 to recapsulize the AV stream 1004 and metadata 1005 input from the transfer section 1003 and transmit them to another node as a capsulized stream 1007 .
- Embodiment 3 it is possible for load sharing to be performed while maintaining close synchronization between the metadata and AV stream processing times, and also to make processing for a segment comprising part of a data stream variable, by providing in the information usage section 1001 an extraction section 803 for separating and extracting an AV stream and metadata, an access section 804 for reading and writing an AV stream and metadata in a storage section 108 , a synchronization section 807 for performing syncronization of the read AV stream and metadata processing, and, in the core processing section 1002 , a transfer section 1003 and a capsulization section 1006 .
- processing method refers to processing for changing the place where metadata is inserted according to the transfer destination, for instance.
- the information provision node 101 it is possible for the information provision node 101 to send appropriate information for transferring and capsulizing metadata to the information usage node 106 .
- metadata it is possible for metadata to be transferred and capsulized appropriately by the information usage node 106 .
- FIG. 11 is a block diagram of an information usage section 1101 according to Embodiment 4. Parts identical to those that have already been described are assigned the same reference numerals, and a description of these parts is omitted.
- the information usage section 1101 according to Embodiment 4 is equivalent to the information usage section 107 according to Embodiment 2 or the information usage section 1001 according to Embodiment 3 provided with a conversion section 1102 . Below, the information usage section 1101 will be described centering on the conversion section 1102 .
- the conversion section 1102 converts an AV scream 810 in accordance with metadata 811 , and outputs the result to the core processing section 1105 as a T-AV stream 1103 and T-metadata 1104 .
- the conversion referred to here is color conversion according to the transmission destination terminal or display position, graphic information format conversion according to the transmission destination terminal or display position, or conversion of the voice format to an MP3 or portable phone format according to the transmission destination terminal.
- the core processing section 1105 operates in the same way as either the core processing section 808 shown in Embodiment 2 or the core processing section 1002 shown in Embodiment 3.
- the core processing section 1105 is core processing section 808
- the core processing section 1105 is provided with a display section 809 .
- the display section 809 performs display while carrying out time synchronization of the input T-AV stream 1103 and T-metadata 1104 .
- the core processing section 1105 is core processing section 1002
- the core processing section 1105 is provided with a transfer section 1003 and capsulization section 1006 .
- the transfer section 1003 performs settings for transferring the T-AV stream 1103 and T-metadata 1104 input by the transfer section 1003 to another information usage node, and performs time synchronization and output to the capsulization section 1006 every MPU 303 .
- the operation of the capsulization section according to Embodiment 3 is similar to that of the capsulization section 207 of Embodiment 1.
- the information usage section 1101 extracts an AV stream 801 and metadata 802 from the capsulized stream 103 input by the extraction section 803 , and outputs them to the access section 804 .
- the access section 804 After recording the AV stream 801 and metadata 802 in the storage section 808 , the access section 804 reads an AV stream 805 and metadata 806 , and outputs them to the synchronization section 807 .
- the synchronization section 807 performs time synchronization every MPU 303 for the AV stream 805 and metadata 806 read by the access section 804 , and outputs them to the conversion section 1102 .
- the conversion section 1102 then converts AV stream 810 according to metadata 811 , and outputs the results to the core processing section 1105 as a T-AV stream 1103 and T-metadata 1104 .
- the display section 809 performs display while carrying out time synchronization of the input T-AV stream 1103 and T-metadata 1104 .
- the transfer section 1003 performs settings for transferring the T-AV stream 1103 and T-metadata 1301 input by the transfer section 1003 to another information usage node, and performs time synchronization and output to the capsulization section 1006 every MPU 303 .
- the capsulization section 1006 recapsulizes the input T-AV stream 1103 and T-metadata 1104 , and transmits them as a capsulized stream 1007 .
- the place where conversion processing is performed according to metadata it is possible for the place where conversion processing is performed according to metadata to be made variable by having the information usage section 1101 provided with an extraction section 303 for separating and extracting an AV stream and metadata , an access section 804 for reading and writing an AV scream and metadata in a storage section 108 , a synchronization section 30 for performing synchronization of the read AV stream and metadata processing, and, as the core processing section 1105 , a usage program composed of a display section 809 or a transfer section 1003 and capsulization section 1006 .
- the place where conversion processing is performed may be, for example, a server, terminal, network node (gateway), or the like.
- Embodiment 4 it is possible to make processing for a segment comprising part of an AV stream variable. Also, AV stream and metadata conversion can be made possible.
- Embodiment 4 by using a structured description written using XML for metadata and metadata units, and performing structured description re-format from metadata to units and from units to metadata, it is possible to extend the degree of freedom for designing metadata for processing an AV stream, and a structured description written in XML, etc., can be used directly as metadata.
- Embodiment 4 it is possible for information relating to methods for processing metadata in the core processing section 1105 —the display method, transfer method, and capsulization method—to be made metadata.
- FIG. 12 is a block diagram of an information processing system according to Embodiment 5. Parts that have already been described are assigned the same reference numerals.
- Embodiment 5 has a configuration that omits the processing for synchronizing an AV stream and metadata from the information provision section 104 according to Embodiment 1.
- processing speed can be increased by omitting synchronization processing and the configuration can be simplified.
- cases where synchronization of an AV stream and metadata need not be performed include cases where metadata is sent all together as with header information and processing need only be performed unit by unit, where it is sufficient for metadata to be synchronized implicitly with the AV stream, where it is sufficient for predetermined control to be performed by the terminal on the information usage side, and where metadata need not be processed in real time.
- An information provision node 1201 is provided with a storage section 102 in which an AV stream and AV stream related metadata are stored.
- the metadata is data that describes the related AV stream, or data for processing the metadata itself, or the like.
- an information provision section 1204 that capsulizes the AV stream and metadata stored in the storage section 102 and generates and outputs a capsulized stream 1203 .
- the information provision section 1204 transmits the capsulized stream 1203 via a network 105 to an information usage node 1206 , which is an apparatus on the information receiving side.
- the information usage node 1206 is provided with an information usage section 1207 that extracts an AV stream and metadata from the capsulized stream 1203 and executes predetermined processing on them in order to use them.
- the information usage node 1205 is also provided with a storage section 108 that stores the AV stream and metadata extracted by the information usage section 1207 .
- the information usage section 1207 reads the AV stream and metadata stored in the storage section 108 in order to use them.
- FIG. 13 is a block diagram of an information provision section according to Embodiment 5.
- the information provision section 1204 is provided with an access section 1301 that reads an AV stream and metadata from the storage section 102 .
- the access section 1301 outputs an AV stream 1302 and metadata 1303 to a unitization section 1304 .
- the unitization section 1304 reforms metadata 1306 read by the access section 1201 into MPUs 303 , and also outputs the synchronized AV stream 1305 and metadata 1306 read by the access section 1301 to a capsulization section 1307 .
- the capsulization section 130 capsulizes the input AV stream 1305 and metadata 1306 , and transmits them to the information usage node 1206 as a capsulized stream 1203 .
- Embodiment 5 as in Embodiment 1, metadata is unitized to enable it to be executed in parts. Then, the AV stream and metadata units are packetized, data stream packets and metadata unit packets are capsulized, and a capsulized stream is generated.
- the operation of the information provision section 1204 of the present invention will be described in detail below. Details of the AV stream 1302 and metadata 1303 stored in the storage section 102 are the same as for the AV stream 202 and metadata 203 according to Embodiment 1, so a description of these will be omitted here.
- metadata 1303 and an AV stream 1302 are read from the storage section 102 by the access section 1301 . Then the access section 1301 outputs the read AV stream 1302 and metadata 1303 to the unitization section 1304 .
- the unitization section 1304 On receiving the AV stream 1302 and metadata 1303 , the unitization section 1304 first proceeds to processing for unitizing the metadata 1303 .
- Metadata 1303 and MPUs 303 are the same as for the metadata 203 according to Embodiment 1 and the MPUs 303 described in Embodiment 1, so a description of these will be omitted here. Also, the process of unitization of the metadata 1303 is the same as for unitization of the metadata 203 according to Embodiment 1, so a description of this will be omitted here.
- metadata 1303 is represented by a collection of MPU definitions 402 . Therefore, metadata 1303 is given a structured description by means of metadata definition 401 , and is stored in the storage section 102 as metadata (XML instance) 501 shown in FIG. 5A .
- an MPU 303 is represented by a collection of metadata defined by user_defined.dtd. Therefore. MPUs 303 are given a structured description for each MPU by means of MPU definitions 402 , and are stored in the storage section 102 as MPU (XML instance) 502 shown in FIG. 5B .
- An MPU 303 has contents ⁇ mpu> to ⁇ /mpu>. That is to say, if there is information from ⁇ mpu> to ⁇ /mpu>, the unitization section 1301 can grasp MPU 303 contents and can perform MPU 303 processing. For this reason, when picking out an MPU 303 from metadata 1303 , the unitization section 1304 extracts the contents on the inside of a tag called an MPU tag (here, ⁇ mpu>) defined by an MPU definition 402 .
- an MPU tag here, ⁇ mpu>
- the unitization section 1304 can perform metadata 1303 processing for each MPU 303 .
- the unitization section 1304 can process AV data 1302 and metadata 1303 unit by unit.
- the capsulization section 1307 capsulizes metadata 1306 sent from the unitization section 1300 using the syntax shown in FIG. 6 .
- the capsulization section 1307 then capsulizes the AV stream segment for processing specified by the first packet's processing start time 607 and duration 608 , and part of the metadata 1303 corresponding to the segment for processing, as a capsulized stream (private PES).
- the unitization section 1304 then packetizes MPUs 303 into private PES packets and interleaves these with video PES packets and audio PES packets.
- the capsulization section 207 capsulizes the input AV stream 1305 and metadata 1306 , and transmits them as a capsulized stream 1203 .
- Metadata can be re-formatted unit by unit and capsulized with an AV stream by providing a unitization section 1304 that unitizes the AV stream and metadata, and a capsulization section 1307 that capsulizes the metadata unit by unit with the AV stream.
- Embodiment 5 unlike Embodiment 1, omits synchronization processing, when synchronization of an AV stream and metadata is not necessary, processing speed can be increased by omitting synchronization processing and the configuration can be simplified.
- FIG. 14 is a block diagram of an information usage section 1207 according to Embodiment 6.
- Embodiment 6 has a configuration that omits the processing for synchronizing an AV stream and metadata from the information usage section 107 according to Embodiment 2.
- processing speed can be increased by omitting synchronization processing and the configuration can be simplified.
- cases where synchronization of an AV stream and metadata need not be performed include cases where metadata is sent all together as with header information and processing need only be performed unit by unit, where it is sufficient for metadata to be synchronized implicitly with the AV stream, where it is sufficient for predetermined control co be performed by the terminal on the information usage side, and where metadata need not be processed in real time.
- An information usage section 120 is provided with an extraction section 1403 that extracts and outputs an AV stream 1401 and metadata 1402 from an input capsulized stream 1203 .
- the extraction section 1403 outputs the extracted AV stream 1401 and metadata 1402 to an access section 1404 .
- the access section 1404 records the AV stream 1401 and metadata 1402 in a storage section 108 . Also, the access section 1404 reads an AV stream 1405 and metadata 1406 stored in the storage section 108 , and outputs them to a core processing section 1407 .
- the core processing section 1407 operates in the same way as the core processing section 808 shown in Embodiment 2. If the core processing section 1105 is core processing section 808 , the core processing section 1407 is provided with a display section 1408 . In this case the display section 1408 displays the input AV stream 1405 and metadata 1406 .
- the information usage section 1207 extracts an AV stream 1401 and metadata 1402 from the capsulized stream 1203 in the extraction section 1403 . Then, the display section 1406 displays metadata 1406 and AV stream 1405 unit by unit.
- the information usage section 120 extracts an AV stream 1401 and metadata 1402 from the capsulized stream 1203 input by the extraction section 1403 , and outputs them to the access section 1404 .
- the access section 1404 After recording the AV stream 1401 and metadata 1402 in the storage section 108 , the access section 1404 reads an AV stream 1405 and metadata 1406 , and outputs them to the core processing section 1407 .
- the display section 1408 displays the input AV stream 1405 and metadata 1406 .
- Embodiment 6 it is possible to make processing for a segment comprising part of a data stream variable by providing an extraction section 1403 for separating and extracting an AV stream and metadata, an access section 1404 for reading and writing an AV stream and metadata in a storage section 108 , and a display section 1408 , which is a core processing section 1407 .
- Embodiment 6 unlike Embodiment 2, omits synchronization processing, when synchronization of an AV stream and metadata is not necessary, processing speed can be increased by emitting synchronization processing and the configuration can be simplified.
- Embodiment 6 has been described as having a configuration in which the synchronization section 807 is omitted from Embodiment 2, but a configuration may also be used in which the synchronization section 807 is omitted from Embodiment 3 or 4.
- each processing section is configured by having all or part of the respective operations stored as a program (software) on a computer-readable storage medium such as a CD-ROM or DVD, and having the operations of each processing section performed by the CPU of a computer, or the like, by having a computer read the program.
- a program software
- a computer-readable storage medium such as a CD-ROM or DVD
- a mode is also possible whereby all or part of the operations of each processing section are stored on a storage medium on communication means such as the Internet or the like as a program (software), the program is downloaded to an information terminal via the Internet or the like, and the operations of each processing section are performed by the information terminal.
- each processing section is configured using dedicated hardware.
- Embodiment 1 to Embodiment 6 descriptions have used an AV stream as a content data stream with timewise continuity, but the sane kind of effects as in the above-described embodiments can be obtained with not an AV stream but another stream, file, or small-volume information, as long as its use as a stream is considered useful.
- Embodiment 1 to Embodiment 6 metadata definitions and MPU definitions are performed using DTD of XML, but XML RDF or XML Schema may be used, or other definition means may also be used.
- Embodiment 1 to Embodiment 6 packetization has been described with MPEG-2 system PES packets, but an MPEG-1 system, MPBG-4, SMPTE Ancillary Data Packet, or another transmission format, streaming format, or file format may also be used.
- Embodiment 1 to Embodiment 6 private PES has been used for the description of the transmission layer for sending metadata, but metadata PES, MPEG-7 PES, MPEG-2 PSI (Program Specific Information) Section (so-called carousel) promised for the future may also be used as a transmission layer.
- metadata PES MPEG-7 PES
- MPEG-2 PSI Program Specific Information Section
- one MPU may also be inserted repeatedly to enable the necessary data to be received when starting reception midway.
- the network 105 or 1505 maybe a terrestrial broadcasting network, a satellite broadcasting network, a cable television network, a line switching network, a packet switching network, an ATM, the Internet, or another network, package medium, hard disk, memory, or the like.
- partial execution of metadata is made possible, and it is possible to carry out program distribution for processing a segment comprising part of an AV stream, speeding up of response times, reduction of the necessary storage capacity, and reduction of network traffic, by reconfiguring metadata unit by unit and capsulizing it with an AV stream, secondly, close synchronization between metadata and AV stream processing times can be performed by making processing of a segment comprising part of an AV stream variable: and thirdly, it is possible to extend the degree of freedom for designing metadata for processing an AV stream, and to use a structured description written in XML, etc., directly as metadata, by using a structured description by means of XML for metadata and metadata units, and performing structured description re-format from metadata to units and from units to metadata.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Library & Information Science (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Time-Division Multiplex Systems (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
A method and apparatus are provided for representing media content that is divided in time into a plurality of segments. A network receiver receives, from a transmitting apparatus, metadata that specifies a first segment and a second segment, first segment data that is coded data of the media content for the first segment, and second segment data that is coded data of the media content for the second segment. A processor decodes the first segment data to present the media content of the first segment according to the metadata describing the first segment data as the coded data of the media content for the first segment. The processor further decodes the second segment data to present the media content of the second segment according to the metadata describing the second segment data as the coded data of the media content for the second segment.
Description
- This is a continuing application of U.S. patent application Ser. No. 14/225,977, filed on Mar. 26, 2014, which is a continuation of U.S. patent application Ser. No. 13/609,848, filed on Sep. 11, 2012, now U.S. Pat. No. 8,832,762, issued on Sep. 9, 2014, which is a continuation of U.S. patent application Ser. No. 13/165,111, filed on Jun. 21, 2011, now U.S. Pat. No. 8,555,328, issued Oct. 8, 2013, which is a continuation of U.S. patent application Ser. No. 12/899,860, filed on Oct. 7, 2010, now U.S. Pat. No. 7,992,182, issued on Aug. 2, 2011, which is a continuation of U.S. patent application Ser. No. 12/111,021, filed on Apr. 28, 2008, now U.S. Pat. No. 7,836,479, issued on Nov. 16, 2010, which is a continuation of U.S. patent application Ser. No. 10/019,319, filed on Jan. 10, 2002, now U.S. Pat. No. 7,383,566, issued on Jun. 3, 2008, which is a U.S. National Stage of International Application No. PCT/JP00/04736, filed on Jul. 14, 2000, which claims the benefit of Japanese Application No. 11-200095, filed Jul. 14, 1999, the contents of all of which are expressly incorporated by reference herein in their entireties.
- The present invention relates to an information provision apparatus, information receiving apparatus, and storage medium, and relates in particular to an information provision apparatus, information receiving apparatus, and storage medium for video/audio, data, etc., operating via broadcast media such as digital broadcasting and communication media such as the Internet.
- In recent years, there has been an active trend of digitalization of broadcasting, and fusion with communications has also progressed. In the field of communications, satellite digital broadcasting has already been started, and it is expected that terrestrial broadcasting will also be digitalized in the future.
- As a result of digitalization of broadcast content, data broadcasting is also performed in addition to conventional video and audio. Also, in the communications field, digital content distribution via the internet has begun with music, and Internet broadcasting stations that broadcast video have also appeared.
- Henceforth, it is envisaged that continuous content media such as video and audio will enter the home via various paths (transmission media). Through such fusion and digitalization of communications and broadcasting, it has become possible to offer previously unavailable services by means of metadata chat describes content or relates to content.
- For example, EPG information as well as audio/video information is provided by interleaving EPG (Electric Program Guide)-“Standard specification for program arrangement information used in digital broadcasting ARIB STD-B10 Version 1.1” or “prETS 300 468 Digital Broadcasting systems for television, sound and data services-Specification for Service Information (SI) in Digital Video Broadcasting (DVB) systems”) used in CS digital broadcasting, in an audio/video PES (Packetized Elementary Stream) using an MPEG-2 (Motion Picturecoding Experts Group phase 2-“ISO/IEC 13818-1 to 3”) private section.
- Also, in BS digital broadcasting, data broadcasting using MPEG-2 private PBS packets is anticipated. Moreover, it is also possible to perform content management by inserting metadata that describes content in the format of user data in material transmission(“ANSI/SMPTE291M-1996 Ancillary Data Packet and Space Formatting”).
- A conventional information processing system will be described below using
FIG. 15 .FIG. 15 is a block diagram of a conventional information processing system. - An
information provision node 1501 is provided with astorage section 1502 in which an AV stream and metadata for describing the AV stream are stored. Also provided in theinformation provision node 1501 is aninformation provision section 1504 that multiplexes the AV stream and metadata stored in thestorage section 1502 and generates and outputs amultiplex stream 1503. Theinformation provision section 1504 transmits themultiplex stream 1503 to aninformation usage node 1506 via anetwork 1505. - Meanwhile, the
information usage node 1506 is provided with aninformation usage section 1507 that extracts an AV stream and metadata from a multiplex stream and executes processing on them in order to use them. Theinformation usage node 1506 is also provided with astorage section 1508 that stores the AV stream and metadata extracted by theinformation usage section 1507. Theinformation usage section 1507 reads the AV stream and metadata stored in thestorage section 1508 in order to use them. - Next, the
information provision section 1504 will be described usingFIG. 16 .FIG. 16 is a block diagram of a conventional information provision section. - The
information provision section 1504 is provided with anaccess section 1601 that reads an AV stream and metadata from thestorage section 1502. Theaccess section 1601 outputs anAV scream 1602 andmetadata 1603 to amultiplexing section 1604. - The
multiplexing section 1604 transmits to the information usage node 1506 amultiplex stream 1503 that multiplexes theAV stream 1602 andmetadata 1603. - Next, multiplex stream generation processing by the
multiplexing section 1601 will be described usingFIG. 17 . - The drawing indicated by
reference numeral 1503 in the drawing shows the MPEG-2 TS (Transport Stream) PES packet layer, and shows a multiplex stream. The drawing indicated byreference numeral 1702 shows a video PES packet, the drawing indicated byreference numeral 1703 shows an audio PES packet, and the drawing indicated byreference numeral 1703 shows a private PES packet. 1603 indicates the metadata PES packet layer, in which 1704 is a first PES packet comprising metadata and 1705 is a second PES packet comprising metadata. - The
multiplexing section 1604 divides themetadata 1603 to make private PES packets, inserts thefirst PES packet 1704 andsecond PES packet 1705 in order as appropriate between AV streams consisting ofvideo PES packets 1701 andaudio PES packets 1702, and obtains amultiplex stream 1503 that is an MPEG-2 TS. - As conventional metadata is AV stream ancillary data-for example, small amounts of data such as titles-processing has been performed with metadata alone. That is to say, it has not been necessary to provide time synchronization of metadata with an AV stream. Therefore, since conventional metadata does not have a configuration that provides for synchronization with an AV stream, metadata has been packetized using virtually the same size, and has been inserted as appropriate between AV streams at virtually equal intervals.
- The
multiplexing section 1604 then sends thismultiplex stream 1503 to theinformation usage node 1506. - Next, the information usage section 150 will be described using
FIG. 18 .FIG. 18 a block diagram of a conventional information usage section. - The
information usage section 1507 is provided with an extraction section 1803 that performs separation and extraction, and output, of anAV stream 1801 andmetadata 1802. Theextraction section 1603 outputs the separated and extractedAV stream 1801 andmetadata 1302 to anaccess section 1804. - The
access section 1801 stores theAV stream 1801 andmetadata 1802 input from the extraction section 1803 in astorage section 1508. Also, theaccess section 1804 outputs theAV stream 1805 andmetadata 1806 read from thestorage section 1508 to adisplay section 1807. Thedisplay section 1807 displays either or both of theAV stream 1805 andmetadata 1806 input from theaccess section 1004. - Next, the processing of the
information usage section 1507 will be described usingFIG. 19 .FIG. 19 is a processing flowchart of a conventional information usage section. - The extraction section 1803 performs metadata parsing-that is, syntax analysis (ST1901). Then, execution of the processing of the
access section 1804 and display section 180 is performed (ST1902). - In this way, a conventional information processing system can display a description relating to AV information, in addition to AV information, by means of the
information usage node 1506 by having theinformation provision node 1501 transmit a multiplex stream multiplexing an AV stream and metadata to theinformation usage node 1506. - In recent years, a demand has arisen for various kinds of information to be included in metadata, and for metadata to be processed coupled with an AV stream, rather than having metadata simply as ancillary data for an AV stream.
- However, in the above-described conventional information processing system, metadata parsing cannot be carried out until all the metadata has been acquired. For example, if metadata begins with <metadata>, metadata parsing cannot be carried out until data </metadata>indicating the end of the metadata arrives.
- For this reason, the metadata processing time is closely tied to the AV stream display or processing time, and since an AV stream is processed in accordance with the metadata itself, processing cannot be started until all the metadata has been received. Therefore, in a conventional information processing system, there is a problem in that it is difficult to process an AV stream in small units.
- Also, metadata is distributed virtually uniformly in a multiplex stream. As a result, especially when the data quantity of metadata is large, a large AV stream quantity must be read by the time all the metadata is read. Consequently, there are problems relating to inter-node response time delays and increased network traffic.
- It is a first objective of the present invention to carry out data and program distribution tor processing a segment comprising part of an AV stream, speeding up of response times, reduction of the necessary storage capacity, and reduction of network traffic, by making possible partial execution of metadata.
- Also, it is a second objective of the present invention to make processing of a segment comprising part of an AV stream variable, and perform close synchronization between metadata and AV stream processing times, by implementing time synchronization of metadata and an AV stream.
- Further, it is a third objective of the present invention to extend the degree of freedom for designing metadata for processing an AV stream.
- In order to meet the first objective, the present invention is provided with a synchronization section which synchronizes a data stream segment with a unit of metadata corresponding to it, and a capsulization section which capsulizes a data stream packet and metadata unit packet after synchronization and generates a capsulized stream.
- By this means, partial execution of metadata is made possible by reconfiguring metadata unit by unit and capsulizing it with the data stream. As a result, it is possible to carry out data and program distribution for processing a segment comprising part of a data scream, speeding up of response times, reduction of the necessary storage capacity, and reduction of network traffic.
- In order to meet the second objective, the present invention is provided with an extraction section which extracts from a capsulized stream a content data stream and metadata for describing or processing that content, a synchronization section which synchronizes metadata unitized with respect to an extracted data stream segment unit by unit with a content data stream and the corresponding metadata unit, and a processing section which processes synchronized metadata unit by unit.
- By this means, it is possible to make processing for a segment comprising part of a data stream variable, and perform close synchronization between metadata and AV stream processing times.
- In order to meet the third objective, the present invention uses a structured description for metadata and metadata units, and structured description re-format is performed from metadata to units and from units to metadata.
- By this means, it is possible to extend the degree of freedom for designing metadata for processing a data stream. In addition, it is possible for a structured description written in XML, etc. to be used directly as metadata.
-
FIG. 1 is a block diagram of an information processing system according toEmbodiment 1 of the present invention; -
FIG. 2 is a block diagram of an information processing section according toEmbodiment 1; -
FIG. 3A is a drawing showing an AV stream according toEmbodiment 1; -
FIG. 3B is a drawing showing metadata according toEmbodiment 1; -
FIG. 4A is a drawing showing DTD of XML of metadata according toEmbodiment 1; -
FIG. 4B is a drawing showing DTD of XML of an MPU according toEmbodiment 1; -
FIG. 5A is a drawing showing an instance of XML of metadata according toEmbodiment 1; -
FIG. 5B is a drawing showing an instance of XML of an MPU according toEmbodiment 1; -
FIG. 6 is a drawing showing the syntax of metadata according toEmbodiment 1; -
FIG. 7 is a drawing for explaining the operation of a capsulization section according toEmbodiment 1; -
FIG. 8 is a block diagram of an information usage section according toEmbodiment 2 of the present invention; -
FIG. 9 is a processing flowchart showing the metadata processing operations of an information usage node according to Embodiment 2of the present invention; -
FIG. 10 is a block diagram of an information usage section according to Embodiment 3of the present invention; -
FIG. 11 is a block diagram of an information usage section according toEmbodiment 4 of the present invention; -
FIG. 12 is a block diagram of an information processing system according to Embodiment 5 of the present invention -
FIG. 13 is a block diagram of an information processing section according to Embodiment 5; -
FIG. 14 is a block diagram of an information usage section according toEmbodiment 4 of the present invention according to Embodiment 6; -
FIG. 15 is a block diagram of a conventional information processing system; -
FIG. 16 is a detailed drawing of a conventional information provision section; -
FIG. 17 is a drawing showing the configuration of a conventional multiplex stream; -
FIG. 18 is a detailed drawing of a conventional information usage section; and -
FIG. 19 is a processing flowchart for a conventional extraction section. - With reference now to the attached drawings, embodiments of the present invention will be explained in detail below.
- (Embodiment 1)
- An information processing system according to
Embodiment 1 of the present invention will be described below.FIG. 1 is a block diagram of an information processing system according toEmbodiment 1. - An
information provision node 101 is provided with astorage section 102 in which an AV stream and AV stream related metadata are stored. The metadata is data that describes the related AV stream, or data for processing the metadata itself, or the like. Also provided in theinformation provision node 101 is aninformation provision section 101 that multiplexes the AV stream and metadata stored in thestorage section 102 and generates and outputs a capsulizedstream 103. Theinformation provision section 104 transmits the capsulizedstream 103 via anetwork 105 to aninformation usage node 106, which is an apparatus on the information receiving side. - Meanwhile, the
information usage node 106 is provided with aninformation usage section 107 that extracts an AV stream and metadata from the capsulizedstream 103 and executes predetermined processing on them in order to use them. Theinformation usage node 106 is also provided with astorage section 108 that stores the AV stream and metadata extracted by theinformation usage section 107. Theinformation usage section 107 reads the AV stream and metadata stored in thestorage section 108 in order to use them. - Next, the
information provision section 104 will be described usingFIG. 2 .FIG. 2 is a block diagram of an information provision section according toEmbodiment 1. - The
information provision section 104 is provided with anaccess section 201 that reads an AV stream and metadata from thestorage section 102. Theaccess section 201 outputs anAV stream 202 andmetadata 203 to asynchronization section 204. - The
synchronization section 204 implements time synchronization for theAV stream 202 andmetadata 203 read by theaccess section 201, and outputs thesynchronized AV stream 205 andmetadata 206 to acapsulization section 207. - The
capsulization section 207 capsulizes thesynchronized AV stream 205 andmetadata 206, and transmits them to theinformation usage node 106 as a capsulizedstream 103. - Also, the present invention unitizes metadata to enable metadata to be executed in parts. Then, AV stream segments and corresponding metadata units are synchronized, synchronized data stream packets and metadata unit packets are capsulized, and a capsulized stream is generated.
- The operation of the
information provision section 104 of the present invention will be described in detail below. - First, the
AV stream 202 andmetadata 203 stored in thestorage section 102 will be described usingFIG. 3A andFIG. 3B . - The
AV stream 202 has video PES pockets 301 andaudio PES packets 302 interleaved to form a stream. In the present embodiment, a mode 53 described whereby anAV stream 202 is scored in thestorage section 102, but a mode is also possible whereby a video stream and audio stream are stored. - The
metadata 203 is configured so as to have a plurality of MPUs (Metadata Processing Units) 303. - The thus configured
metadata 203 andAV stream 202 are read from thestorage section 102 by theaccess section 201. Then theaccess section 201 outputs theread AV stream 202 andmetadata 203 to thesynchronization section 204. - On receiving the
AV stream 202 andmetadata 203, thesynchronization section 204 first proceeds to processing for unitizing themetadata 203. Here, definitions of themetadata 203 andMPU 303 will be described usingFIG. 4A andFIG. 4B .FIG. 4A andFIG. 4B are drawings showing DTD of XML. InFIG. 4A . 101 is a drawing showing a metadata definition (metadata.dtd) that defines themetadata 203. InFIG. 4B , the drawing indicated byreference numeral 402 shows an MPU definition (mpu.dtd) that defines anMPU 303. - The
metadata definition 401 defines themetadata 203 as having one ormore MPUs 303. For the contents of anMPU 303, referencing theMPU definition 402 is defined. - The
MPU definition 402 defines anMPU 303 as having one or more element_data items. For the contents of element_data, referencing user_defined.dtd is defined. Also, theMPU definition 402 defines anMPU 303 as having a serial number no assigned. - In this way, it is possible to include in an
MPU 303 different processing contents for each of various services according to user_defined.dtd. Thus, it is possible to extend the degree of freedom for designing metadata for processing an AV stream. - Also, it is possible to include in an
MPU 303 processing contents not in accordance with a transmission specification, according to user_defined.dtd. By this means, metadata canal so be used for a different transmission specification, making it possible to provide metadata services that support a variety of transmission specifications. - Next, the unitization of
metadata 203 will be described usingFIG. 5A andFIG. 5B . InFIG. 5A , the drawing indicated byreference numeral 501 shows metadata (XML instance) wherebymetadata 203 is given a structured description according tometadata definition 401, and the drawing indicated byreference numeral 502 shows an MPU (XML instance) whereby anMPU 303 is given a structured description according to MPU definition 902. - As described above, according to
metadata definition 401,metadata 203 is represented by a collection ofMPU definitions 102. According to thismetadata definition 401, what gives a structured description ofmetadata 203 is metadata (XML instance) 501. As can be seen from the drawing, the metadata (XML instance) 501 instance includes a plurality ofMPUs 303. Also,metadata 203 is stored in thestorage section 102 as metadata (XML instance) 501. - According to
MPU definition 402, anMPU 303 is represented by a collection of metadata defined by user_defined.dtd. According to thisMPU definition 402, what gives a structured description ofMPU 303 for each MPU is MPU (XML instance) 502. As can be seen from the drawing, MPU (XML instance) 502 includes a plurality of user_defined.dtd items. Also,MPU 303 is stored in thestorage section 102 as MPU (XML instance) 502. - An
MPU 303 has contents <mpu> to </mpu>. That is to say, if there is information from <mpu> to </mpu>, thesynchronization section 204 can graspMPU 303 contents and can performMPU 303 processing. For this reason, when picking out anMPU 303 frommetadata 203, thesynchronization section 204 extracts the contents on the inside of a tag called an MPU tag (here, <mpu>) defined by anMPU definition 402. - By having
metadata 203 composed of lower-level information MPUs 303 in this way, thesynchronization section 204 can performmetadata 203 processing for eachMPU 303, and also closely synchronize theAV data 202 andmetadata 203. - Next, the
synchronization section 204 capsulizesmetadata 203 sent from theaccess section 201 using the syntax shown inFIG. 6 .FIG. 6 shows the syntax of metadata according toEmbodiment 1 andEmbodiment 2. - In
FIG. 6 ,metadata_type 601 is the metadata type such as position information, content information, or program.metadata_subtype 602 is the concrete metadata type such as GPS or structured description (MPEG-7).MPU_length 603 is the data length as a number of bytes from immediately after the MPU_length field co the end of the MPU. An MPU is composed of one or more PSS packets, and is the regeneration unit of metadata divided when a Metadata Elementary Scream is encoded.media_sync_flag 604 is a flag indicating the presence or absence of synchronization between the AV stream and metadata.overwrite_flag 605 is a flag indicating whether the previous metadata is to be overwritten.element_data_length 606 is the data byte length (M) ofelement_data 609. start_time ( ) 607 is the start time of a segment that is a part of the AV stream indicated by the metadata, duration ( ) 608 is the continuation time of a segment that is part of the AV stream indicated by the metadata.element_data 609 is the actual data of the metadata. - For the syntax shown in
FIG. 6 , coding usessyntax 610 from else downward even when the metadata data quantity is small and unitization is not performed. - The
synchronization section 204 capsulizes the AV stream segment for processing specified by the first packet'sprocessing start time 607 andduration 608. and part of themetadata 203 corresponding to the segment for processing, as a capsulized stream (private PES). - When metadata 203 is PES-packetized, an
MPU 303 is packetized together with the AV stream segment first packet processing start time (start_time), duration ( ) 606, and actual data of the metadata as an element (element_data) in the metadata syntax shown inFIG. 6 . - By this means, it is possible for an
MPU 303 to have information for maintaining synchronization with theAV stream 202. Thus, synchronization is maintained between theMPU 303 andAV stream 202. In this way,metadata 203 operation can be determined on theinformation provision node 101 side. - Also, in
Embodiment 1, anMPU 303 is composed of two packets—afirst PES packet 701 and asecond PES packet 702—as shown inFIG. 7 . The operations whereby thesynchronization section 204 packetizes anMPU 303 into private PES packets and interleaves these withvideo PES packets 301 andaudio PES packets 302 in this case will be described usingFIG. 1 . How many packets anMPU 303 is made into can be determined arbitrarily according to theMPU 303 size and the packet size. - In the case of
Embodiment 1, thefirst PES packet 701 andsecond PES packet 702 are placed asprivate PES packets 708 earlier in time than thefirst packet 703 so that thefirst PES packet 701 andsecond PES packet 702 are processed before the processing start time (start_time) 705 of the first packet of the corresponding AV stream segment. - Also, the
second PES packet 702arrival time t 704 and the correspondingfirst packet 703 processing start time (start_time) 705 difference At 706 are assigned sufficient times for theinformation usage section 107, which is on the information receiving side, to generate anMPU 303 from thefirst PES packet 701 andsecond PES packet 702, and execute processing based on the contents of the generatedMPU 303. - Then, the
AV stream 205 andmetadata 205 synchronized by thesynchronization section 204 in this way are input to thecapsulization section 207. - The
capsulization section 207 capsulizes theinput AV stream 205 andmetadata 206, and transmits them as a capsulizedstream 103. - As described above, according to
Embodiment 1, metadata can be re-formatted unit by unit and capsulized with an AV stream by providing asynchronization section 204 that maintains synchronization of the AV stream and metadata, and acapsulization section 207 that capsulizes metadata unit by unit with the AV stream. By this means, it becomes possible to perform partial execution of metadata, and to carry out program distribution for processing a segment comprising part of an AV stream, speeding up of response times, reduction of the necessary storage capacity, and reduction of network traffic. - Moreover, according to
Embodiment 1, by using a structured description written using for metadata and metadata units, and performing structured description re-format from metadata to units and from units to metadata, it is possible to provide extensibility for metadata for processing an AV stream, and extend the degree of freedom for designing metadata. In addition, it is possible for a structured description written in XML, etc., to be used directly as metadata. - (Embodiment 2)
- Next, an information processing system according to
Embodiment 2 of the present invention will be described.FIG. 8 is a block, diagram of aninformation usage section 107 according toEmbodiment 2. - The
information usage section 107 is provided with anextraction section 803 that performs separation and extraction, and output, of anAV stream 801 andmetadata 802. Theextraction section 803 outputs the extractedAV stream 801 andmetadata 802 to anaccess section 804. - The
access section 804 records theAV stream 801 andmetadata 802 in astorage section 108. Also, theaccess section 804 reads anAV stream 805 andmetadata 806 stored in thestorage section 108, and outputs them to a synchronization section 307. - The
synchronization section 807 performs time synchronization everyMPU 303 for theAV stream 805 andmetadata 806 read by theaccess section 804, and outputs them to acore processing section 808. - The
core processing section 808 is provided with adisplay section 809. Thedisplay section 809 performs time synchronization and display of the input synchronizedAV stream 810 andmetadata 811. - In this way, the
information usage section 107 extracts anAV stream 801 andmetadata 802 from the capsulizedstream 103 in theextraction section 803. Then, in thesynchronization section 807, the correspondingmetadata 802 unitized in accordance withAV stream 801 segments is synchronized with theAV stream 801 unit by unit. Then thesynchronized metadata 811 andAV stream 810 are displayed unit by unit by thedisplay section 809. - Next, the metadata processing operations of the
information usage node 106 will be described in detail using the flowchart inFIG. 9 . First, theextraction section 803 extracts an AV stream and metadata from the received capsulizedstream 103. In addition, theinformation usage section 107 performsMPU 303 pursing (ST901). Next, in theinformation usage section 107, a check is performed as to whether theMPUs 303 are to be merged and re-formatted as metadata 802 (ST902). Then, in theinformation usage section 107, a check is performed as to whetherMPU 303 execution is to be performed unit by unit (ST903). - If, in ST902 and ST903, the results confirmed by the
information usage section 107 are MPU merging and MPU execution, processing is executed by the core processing section 808 (ST904). Then MPU merging is performed in the information usage section 107 (ST905). InEmbodiment 2, this processing is display processing, but it may also be conversion processing or transfer processing as in other embodiments to be described hereafter. - Then, in the
information usage section 107, judgment as to the advent of an MPU time or number limit—that is, an event that indicates an MPU processing unit—is performed (ST906), and ST904 and ST905 are repeated until the advent of an event. Event information is given to software when providing universality, or is given to a terminal beforehand when the system is used in a fixed mode. - Then, in the
information usage section 107, rendering—that is to say, formatting—of the metadata is performed from the MPUs collected together in ST906. Metadata formatted oil the basis of this event is stored in thestorage section 108. Then thecore processing section 808 reads this formatted data and performs various kinds of processing. - In this way, it is possible not only to perform processing for each MPU, which is the minimum unit of processing, in ST904, but also to perform processing based on data obtained by merging MPUs according to an event. By this means, it is possible to set arbitrarily a unit for MPU processing according to an event, and therefore the length of AV data segments for metadata processing can be made variable. That is to say, it is possible to process metadata for small AV data and to process metadata for huge AV data. For example, it is possible to update metadata display in short cycles in a case such as a vehicle navigation system, and update metadata in long cycles in a case such as a news program.
- Also, by storing this metadata that has been formatted on the basis of an event in the
storage section 108, it is possible to read and process this information by means of user operations. - If, in ST902 and ST903, the results confirmed by the
information usage section 107 are MPU merging and MPU non-execution, an MPU merge is performed (ST908). Then, in theinformation usage section 107, judgment as to the presence of an MPU time or number limit—that is, an event related to completion of an MPU merge—is performed (ST909), and ST906 is repeated until the occurrence of an event. Rendering of the metadata is then performed from the MPUs collected together in processing P107. Then, in theinformation usage section 107, rendering—that is to say, formatting—of the metadata is performed from the MPUs collected together in ST906 (ST910). Metadata formatted on the basis of this event is scored in thestorage section 108. Then thecore processing section 808 reads this formatted data and performs various kinds of processing. - In this way, it is possible not only to perform processing for each MPU, which is the minimum unit of processing, but also to perform processing based on data obtained by merging MPUs according to an event.
- If, in ST902 and ST903, the results confirmed by the
information usage section 107 are MPU non-merging and MPU execution, processing is executed sequentially (ST911). Then, in theinformation usage section 107, judgment as to the presence of an MPU time or number limit—that is, an event that indicates an MPU processing unit—is performed (ST912), and ST911 is repeated until the occurrence of an event. - In this way, it is possible to perform processing for each MPU, which is the minimum unit of processing, and not to perform processing based on data obtained by merging MPUs according to an event.
- If, in ST902 and ST903, the results confirmed by the
information usage section 107 are MPU non-merging and MPU non-execution, no particular MPU-related processing is performed. - As described above, the extraction method can be changed as appropriate according to the contents contained in
MPUs 303. - The operation of the
information usage section 107 will now be described below. Theinformation usage section 107 extracts anAV stream 801 andmetadata 802 from the capsulizedstream 103 input by theextraction section 803, and outputs them to theaccess section 804. After recording theAV stream 801 andmetadata 802 in thestorage section 108, theaccess section 804 reads anAV stream 805 andmetadata 806, and outputs them to thesynchronization section 807. Thesynchronization section 807 performs time synchronization everyMPU 303 for theAV stream 805 andmetadata 806 read by theaccess section 804, and outputs them to thecore processing section 808. In thecore processing section 808, thedisplay section 809 performs time synchronization and display of theinput AV stream 810 andmetadata 811. - As described above, according to
Embodiment 2, close synchronization of the metadata and AV stream processing time can be performed by providing an extraction,section 803 for separating and extracting an AV stream and metadata, anaccess section 804 for reading and writing an AV stream and metadata in astorage section 108, asynchronization section 807 for performing synchronization of the read AV stream and metadata processing, and adisplay section 809, which is acore processing section 808. By this means, it is possible to vary processing for a segment, which is part of an AV stream. - Also, information relating to the display method used by the
display section 809 of thecore processing section 808 can be provided as metadata. Information relating to the display method includes position information for displaying metadata related information, display size information, and display update information. - By this means, an appropriate method for displaying metadata can be sent to the
information provision node 101 by theinformation usage node 106. As a result, metadata can be displayed appropriately by theinformation usage node 106. Therefore, if metadata is an advertisement or the like, it is possible to make a specification that allows the advertisement to be displayed at the desired time, and if metadata is information related to program descriptions, it is possible to display the descriptive information so as not to interfere with images. - Moreover, according to
Embodiment 2, by using a structured description written using XML for metadata and metadata units, and performing structured description re-format from metadata to units and from units to metadata, it is possible to extend the degree of freedom for designing metadata for processing an AV stream, and a structured description written in XML, etc., can be used directly as metadata. - (Embodiment 3)
- Next, an information processing method according to
Embodiment 3 of the present invention will be described.FIG. 10 is a block diagram of aninformation usage section 1001 according toEmbodiment 3. Parts identical to those that have already been described are assigned the same reference numerals, and a description of these parts is omitted. - The
information usage section 1001 according toEmbodiment 3 has thecore processing section 808 of theinformation usage section 1001 according toEmbodiment 2 replaced by acore processing section 1002. Below, theinformation usage section 1001 will be described centering on thecore processing section 1002. - The
core processing section 1002 is provided with atransfer section 1003 and acapsulization section 1006. - The
transfer section 1003 performs settings, such as a destination setting, for transferring anAV scream 810 andmetadata 811 input from thesynchronization section 807 to another information usage node. Thetransfer section 1003 performs time synchronization everyMPU 303, and outputs anAV stream 1004 andmetadata 1005 to thecapsulization section 1006. - The
capsulization section 1006 recapsulizes theinput AV stream 1004 andmetadata 1005 and transmits them to another node as a capsulizedstream 1007. Since thecapsulization section 1006 recapsulizes theAV stream 1001 andmetadata 1005 in this way, load sharing can be performed while maintaining close synchronization between the metadata and AV stream processing times. - The operation of the
capsulization section 1006 is similar to that of thecapsulization section 207 according toEmbodiment 1, and so a detailed description will be omitted here. - The operation of the
information usage section 1101 will now be described below. Theinformation usage section 1101 extracts anAV stream 801 andmetadata 802 from the capsulizedstream 103 input by theextraction section 803, and outputs them to theaccess section 804. After recording theAV stream 801 andmetadata 802 in thestorage section 108, theaccess section 804 reads anAV stream 805 andmetadata 806, and outputs them to thesynchronization section 807. - The
synchronization section 807 performs time synchronization everyMPU 303 for theAV stream 805 andmetadata 806 read by theaccess section 301, and outputs them to thecore processing section 1002. Thecore processing section 1002 per forms settings for transferring theAV stream 810 andmetadata 811 input by thetransfer section 1003 to another information usage node, and performs time synchronization and output to thecapsulization section 1006 everyMPU 303. Thecapsulization section 1006 recapsulizes theinput AV stream 1001 andmetadata 1005 and transmits them to another node as a capsulizedstream 1007. - By configuring the
information usage section 1001 as described above, it is possible for thetransfer section 1002 to perform settings for transferring theAV stream 810 andmetadata 811 input from thesynchronization section 807 to another information usage node, perform time synchronization and output to the capsulization step 23 everyMPU 303, and for thecapsulization section 1006 to recapsulize theAV stream 1004 andmetadata 1005 input from thetransfer section 1003 and transmit them to another node as a capsulizedstream 1007. - As described above, according to
Embodiment 3, it is possible for load sharing to be performed while maintaining close synchronization between the metadata and AV stream processing times, and also to make processing for a segment comprising part of a data stream variable, by providing in theinformation usage section 1001 anextraction section 803 for separating and extracting an AV stream and metadata, anaccess section 804 for reading and writing an AV stream and metadata in astorage section 108, asynchronization section 807 for performing syncronization of the read AV stream and metadata processing, and, in thecore processing section 1002, atransfer section 1003 and acapsulization section 1006. - Moreover, according to
Embodiment 3, it is also possible for information about the processing methods of thetransfer section 1003 andcapsulization section 1006, or a processing program itself, to be made metadata. Processing method here refers to processing for changing the place where metadata is inserted according to the transfer destination, for instance. By this means, it is possible for theinformation provision node 101 to send appropriate information for transferring and capsulizing metadata to theinformation usage node 106. As a result, it is possible for metadata to be transferred and capsulized appropriately by theinformation usage node 106. - (Embodiment 4)
- Next, an information processing system according to
Embodiment 4 of the present invention will be described.FIG. 11 is a block diagram of aninformation usage section 1101 according toEmbodiment 4. Parts identical to those that have already been described are assigned the same reference numerals, and a description of these parts is omitted. - The
information usage section 1101 according toEmbodiment 4 is equivalent to theinformation usage section 107 according toEmbodiment 2 or theinformation usage section 1001 according toEmbodiment 3 provided with aconversion section 1102. Below, theinformation usage section 1101 will be described centering on theconversion section 1102. - The
conversion section 1102 converts anAV scream 810 in accordance withmetadata 811, and outputs the result to thecore processing section 1105 as a T-AV stream 1103 and T-metadata 1104. The conversion referred to here is color conversion according to the transmission destination terminal or display position, graphic information format conversion according to the transmission destination terminal or display position, or conversion of the voice format to an MP3 or portable phone format according to the transmission destination terminal. - The
core processing section 1105 operates in the same way as either thecore processing section 808 shown inEmbodiment 2 or thecore processing section 1002 shown inEmbodiment 3. - If the
core processing section 1105 iscore processing section 808, thecore processing section 1105 is provided with adisplay section 809. In this case thedisplay section 809 performs display while carrying out time synchronization of the input T-AV stream 1103 and T-metadata 1104. - If the
core processing section 1105 iscore processing section 1002, thecore processing section 1105 is provided with atransfer section 1003 andcapsulization section 1006. In this case, thetransfer section 1003 performs settings for transferring the T-AV stream 1103 and T-metadata 1104 input by thetransfer section 1003 to another information usage node, and performs time synchronization and output to thecapsulization section 1006 everyMPU 303. The operation of the capsulization section according toEmbodiment 3 is similar to that of thecapsulization section 207 ofEmbodiment 1. - The operation of the
information usage section 1101 will now be described below. Theinformation usage section 1101 extracts anAV stream 801 andmetadata 802 from the capsulizedstream 103 input by theextraction section 803, and outputs them to theaccess section 804. After recording theAV stream 801 andmetadata 802 in thestorage section 808, theaccess section 804 reads anAV stream 805 andmetadata 806, and outputs them to thesynchronization section 807. Thesynchronization section 807 performs time synchronization everyMPU 303 for theAV stream 805 andmetadata 806 read by theaccess section 804, and outputs them to theconversion section 1102. Theconversion section 1102 then convertsAV stream 810 according tometadata 811, and outputs the results to thecore processing section 1105 as a T-AV stream 1103 and T-metadata 1104. - Then, if the
core processing section 1105 is thecore processing section 808 according toEmbodiment 2, thedisplay section 809 performs display while carrying out time synchronization of the input T-AV stream 1103 and T-metadata 1104. If thecore processing section 1105 is thecore processing section 1002 according toEmbodiment 1. thetransfer section 1003 performs settings for transferring the T-AV stream 1103 and T-metadata 1301 input by thetransfer section 1003 to another information usage node, and performs time synchronization and output to thecapsulization section 1006 everyMPU 303. Thecapsulization section 1006 recapsulizes the input T-AV stream 1103 and T-metadata 1104, and transmits them as a capsulizedstream 1007. - As described above, according to
Embodiment 4, it is possible for the place where conversion processing is performed according to metadata to be made variable by having theinformation usage section 1101 provided with anextraction section 303 for separating and extracting an AV stream and metadata , anaccess section 804 for reading and writing an AV scream and metadata in astorage section 108, a synchronization section 30 for performing synchronization of the read AV stream and metadata processing, and, as thecore processing section 1105, a usage program composed of adisplay section 809 or atransfer section 1003 andcapsulization section 1006. The place where conversion processing is performed may be, for example, a server, terminal, network node (gateway), or the like. - Moreover, according to
Embodiment 4, it is possible to make processing for a segment comprising part of an AV stream variable. Also, AV stream and metadata conversion can be made possible. - Furthermore, according to
Embodiment 4, performing further processing on a converted AV stream and metadata can be made possible. - Still further, according to
Embodiment 4, by using a structured description written using XML for metadata and metadata units, and performing structured description re-format from metadata to units and from units to metadata, it is possible to extend the degree of freedom for designing metadata for processing an AV stream, and a structured description written in XML, etc., can be used directly as metadata. - In addition, according to
Embodiment 4, it is possible for information relating to methods for processing metadata in thecore processing section 1105—the display method, transfer method, and capsulization method—to be made metadata. - (Embodiment 5)
- Next, an information processing system according to Embodiment 5 of the present invention will be described.
FIG. 12 is a block diagram of an information processing system according to Embodiment 5. Parts that have already been described are assigned the same reference numerals. - Embodiment 5 has a configuration that omits the processing for synchronizing an AV stream and metadata from the
information provision section 104 according toEmbodiment 1. By omitting synchronization processing in this way, when synchronization of an AV stream and metadata is not necessary, processing speed can be increased by omitting synchronization processing and the configuration can be simplified. Examples of cases where synchronization of an AV stream and metadata need not be performed include cases where metadata is sent all together as with header information and processing need only be performed unit by unit, where it is sufficient for metadata to be synchronized implicitly with the AV stream, where it is sufficient for predetermined control to be performed by the terminal on the information usage side, and where metadata need not be processed in real time. - The configuration of an information processing system according to Embodiment 5 will now be described below. An
information provision node 1201 is provided with astorage section 102 in which an AV stream and AV stream related metadata are stored. The metadata is data that describes the related AV stream, or data for processing the metadata itself, or the like. Also provided in theinformation provision node 1201 is aninformation provision section 1204 that capsulizes the AV stream and metadata stored in thestorage section 102 and generates and outputs a capsulizedstream 1203. theinformation provision section 1204 transmits the capsulizedstream 1203 via anetwork 105 to aninformation usage node 1206, which is an apparatus on the information receiving side. - Meanwhile, the
information usage node 1206 is provided with aninformation usage section 1207 that extracts an AV stream and metadata from the capsulizedstream 1203 and executes predetermined processing on them in order to use them. The information usage node 1205 is also provided with astorage section 108 that stores the AV stream and metadata extracted by theinformation usage section 1207. Theinformation usage section 1207 reads the AV stream and metadata stored in thestorage section 108 in order to use them. - Next, the
information provision section 1204 will be described usingFIG. 13 .FIG. 13 is a block diagram of an information provision section according to Embodiment 5. - The
information provision section 1204 is provided with anaccess section 1301 that reads an AV stream and metadata from thestorage section 102. Theaccess section 1301 outputs anAV stream 1302 andmetadata 1303 to aunitization section 1304. - The
unitization section 1304 reforms metadata 1306 read by theaccess section 1201 intoMPUs 303, and also outputs thesynchronized AV stream 1305 andmetadata 1306 read by theaccess section 1301 to acapsulization section 1307. - The capsulization section 130 capsulizes the
input AV stream 1305 andmetadata 1306, and transmits them to theinformation usage node 1206 as a capsulizedstream 1203. - In Embodiment 5, as in
Embodiment 1, metadata is unitized to enable it to be executed in parts. Then, the AV stream and metadata units are packetized, data stream packets and metadata unit packets are capsulized, and a capsulized stream is generated. - The operation of the
information provision section 1204 of the present invention will be described in detail below. Details of theAV stream 1302 andmetadata 1303 stored in thestorage section 102 are the same as for theAV stream 202 andmetadata 203 according toEmbodiment 1, so a description of these will be omitted here. - With the above-described configuration,
metadata 1303 and anAV stream 1302 are read from thestorage section 102 by theaccess section 1301. Then theaccess section 1301 outputs theread AV stream 1302 andmetadata 1303 to theunitization section 1304. - On receiving the
AV stream 1302 andmetadata 1303, theunitization section 1304 first proceeds to processing for unitizing themetadata 1303. - Definitions of the
metadata 1303 andMPUs 303 are the same as for themetadata 203 according toEmbodiment 1 and theMPUs 303 described inEmbodiment 1, so a description of these will be omitted here. Also, the process of unitization of themetadata 1303 is the same as for unitization of themetadata 203 according toEmbodiment 1, so a description of this will be omitted here. - According to metadata definitional shown in
FIG. 4A ,metadata 1303 is represented by a collection ofMPU definitions 402. Therefore,metadata 1303 is given a structured description by means ofmetadata definition 401, and is stored in thestorage section 102 as metadata (XML instance) 501 shown inFIG. 5A . - Also, according to
MPU definition 402 shown inFIG. 4B , anMPU 303 is represented by a collection of metadata defined by user_defined.dtd. Therefore.MPUs 303 are given a structured description for each MPU by means ofMPU definitions 402, and are stored in thestorage section 102 as MPU (XML instance) 502 shown inFIG. 5B . - An
MPU 303 has contents <mpu> to </mpu>. That is to say, if there is information from <mpu> to </mpu>, theunitization section 1301 can graspMPU 303 contents and can performMPU 303 processing. For this reason, when picking out anMPU 303 frommetadata 1303, theunitization section 1304 extracts the contents on the inside of a tag called an MPU tag (here, <mpu>) defined by anMPU definition 402. - By having
metadata 1303 composed of lower-level information MPUs 303 in this way, theunitization section 1304 can performmetadata 1303 processing for eachMPU 303. By this means, theunitization section 1304 can processAV data 1302 andmetadata 1303 unit by unit. - Next, as in
Embodiment 1, thecapsulization section 1307 capsulizesmetadata 1306 sent from the unitization section 1300 using the syntax shown inFIG. 6 . - The
capsulization section 1307 then capsulizes the AV stream segment for processing specified by the first packet'sprocessing start time 607 andduration 608, and part of themetadata 1303 corresponding to the segment for processing, as a capsulized stream (private PES). - The
unitization section 1304 then packetizesMPUs 303 into private PES packets and interleaves these with video PES packets and audio PES packets. - Then the
capsulization section 207 capsulizes theinput AV stream 1305 andmetadata 1306, and transmits them as a capsulizedstream 1203. - As described above, according to Embodiment 5. metadata can be re-formatted unit by unit and capsulized with an AV stream by providing a
unitization section 1304 that unitizes the AV stream and metadata, and acapsulization section 1307 that capsulizes the metadata unit by unit with the AV stream. By this means, it becomes possible to perform partial execution of metadata, and to carry out program distribution for processing a segment comprising part of an AV stream, speeding up of response times, reduction of the necessary storage capacity, and reduction of network traffic. - Moreover, since Embodiment 5, unlike
Embodiment 1, omits synchronization processing, when synchronization of an AV stream and metadata is not necessary, processing speed can be increased by omitting synchronization processing and the configuration can be simplified. - (Embodiment 6)
- Next, an information processing system according to Embodiment 6 of the present invention will be described.
FIG. 14 is a block diagram of aninformation usage section 1207 according to Embodiment 6. - Embodiment 6 has a configuration that omits the processing for synchronizing an AV stream and metadata from the
information usage section 107 according toEmbodiment 2. By omitting synchronization processing in this way, when synchronization of an AV stream and metadata is not necessary, processing speed can be increased by omitting synchronization processing and the configuration can be simplified. Examples of cases where synchronization of an AV stream and metadata need not be performed include cases where metadata is sent all together as with header information and processing need only be performed unit by unit, where it is sufficient for metadata to be synchronized implicitly with the AV stream, where it is sufficient for predetermined control co be performed by the terminal on the information usage side, and where metadata need not be processed in real time. - The configuration of an information processing system according to Embodiment 6 will now be described below.
- An information usage section 120 is provided with an
extraction section 1403 that extracts and outputs anAV stream 1401 andmetadata 1402 from an input capsulizedstream 1203. Theextraction section 1403 outputs the extractedAV stream 1401 andmetadata 1402 to anaccess section 1404. - The
access section 1404 records theAV stream 1401 andmetadata 1402 in astorage section 108. Also, theaccess section 1404 reads anAV stream 1405 andmetadata 1406 stored in thestorage section 108, and outputs them to acore processing section 1407. - The
core processing section 1407 operates in the same way as thecore processing section 808 shown inEmbodiment 2. If thecore processing section 1105 iscore processing section 808, thecore processing section 1407 is provided with adisplay section 1408. In this case thedisplay section 1408 displays theinput AV stream 1405 andmetadata 1406. - In this way, the
information usage section 1207 extracts anAV stream 1401 andmetadata 1402 from the capsulizedstream 1203 in theextraction section 1403. Then, thedisplay section 1406 displaysmetadata 1406 andAV stream 1405 unit by unit. - The operation of the
information usage section 1207 will now be described below. The information usage section 120 extracts anAV stream 1401 andmetadata 1402 from the capsulizedstream 1203 input by theextraction section 1403, and outputs them to theaccess section 1404. After recording theAV stream 1401 andmetadata 1402 in thestorage section 108, theaccess section 1404 reads anAV stream 1405 andmetadata 1406, and outputs them to thecore processing section 1407. In thecore processing section 1407, thedisplay section 1408 displays theinput AV stream 1405 andmetadata 1406. - As described above, according to Embodiment 6, it is possible to make processing for a segment comprising part of a data stream variable by providing an
extraction section 1403 for separating and extracting an AV stream and metadata, anaccess section 1404 for reading and writing an AV stream and metadata in astorage section 108, and adisplay section 1408, which is acore processing section 1407. - Moreover, since Embodiment 6, unlike
Embodiment 2, omits synchronization processing, when synchronization of an AV stream and metadata is not necessary, processing speed can be increased by emitting synchronization processing and the configuration can be simplified. - Embodiment 6 has been described as having a configuration in which the
synchronization section 807 is omitted fromEmbodiment 2, but a configuration may also be used in which thesynchronization section 807 is omitted fromEmbodiment - In
Embodiment 3 to Embodiment 6, each processing section is configured by having all or part of the respective operations stored as a program (software) on a computer-readable storage medium such as a CD-ROM or DVD, and having the operations of each processing section performed by the CPU of a computer, or the like, by having a computer read the program. - A mode is also possible whereby all or part of the operations of each processing section are stored on a storage medium on communication means such as the Internet or the like as a program (software), the program is downloaded to an information terminal via the Internet or the like, and the operations of each processing section are performed by the information terminal.
- A mode is also possible where by each processing section is configured using dedicated hardware.
- In
Embodiment 1 to Embodiment 6, descriptions have used an AV stream as a content data stream with timewise continuity, but the sane kind of effects as in the above-described embodiments can be obtained with not an AV stream but another stream, file, or small-volume information, as long as its use as a stream is considered useful. - In
Embodiment 1 to Embodiment 6, metadata definitions and MPU definitions are performed using DTD of XML, but XML RDF or XML Schema may be used, or other definition means may also be used. - In
Embodiment 1 to Embodiment 6, packetization has been described with MPEG-2 system PES packets, but an MPEG-1 system, MPBG-4, SMPTE Ancillary Data Packet, or another transmission format, streaming format, or file format may also be used. - In
Embodiment 1 to Embodiment 6, private PES has been used for the description of the transmission layer for sending metadata, but metadata PES, MPEG-7 PES, MPEG-2 PSI (Program Specific Information) Section (so-called carousel) promised for the future may also be used as a transmission layer. - In
Embodiment 1 to Embodiment 6, as a synchronization variation, one MPU may also be inserted repeatedly to enable the necessary data to be received when starting reception midway. - In
Embodiment 1 to Embodiment 6, thenetwork - This application is based on the Japanese Patent Application No. HEI11-200095 filed on Jul. 14, 1999, entire content of which is expressly incorporated by reference herein.
- As described above, according to the present invention, firstly, partial execution of metadata is made possible, and it is possible to carry out program distribution for processing a segment comprising part of an AV stream, speeding up of response times, reduction of the necessary storage capacity, and reduction of network traffic, by reconfiguring metadata unit by unit and capsulizing it with an AV stream, secondly, close synchronization between metadata and AV stream processing times can be performed by making processing of a segment comprising part of an AV stream variable: and thirdly, it is possible to extend the degree of freedom for designing metadata for processing an AV stream, and to use a structured description written in XML, etc., directly as metadata, by using a structured description by means of XML for metadata and metadata units, and performing structured description re-format from metadata to units and from units to metadata.
Claims (5)
1. A method for representing media content that is divided in time into a plurality of segments, the method comprising:
receiving, by a network receiver from a transmitting apparatus, metadata that specifies a first segment and a second segment out of the plurality of segments, first segment data that is coded data of the media content for the first segment, and second segment data that is coded data of the media content for the second segment, the first segment being different from the second segment, the metadata being described in a structured description;
decoding, by a processor, the first segment data to present the media content of the first segment according to the metadata describing the first segment data as the coded data of the media content for the first segment; and
decoding, by the processor, the second segment data to present the media content of the second segment according to the metadata describing the second segment data as the coded data of the media content for the second segment.
2. The method according to claim 1 , wherein
the metadata includes information indicating a first start time for presenting the media content of the first segment, and
the decoding of the first segment data is initiated before the first start time.
3. The method according to claim 1 , wherein
the metadata is received at least a predetermined time before the first segment data that is coded data of the media content for the first segment.
4. The method according to claim 1 , wherein
the media content of the first segment and the media content of the second segment are presented seamlessly.
5. An apparatus for representing media content that is divided in time into a plurality of segments, the apparatus comprising:
a network receiver that, in operation, receives, from a transmitting apparatus, metadata that specifies a first segment and a second segment out of the plurality of segments, first segment data that is coded data of the media content for the first segment, and second segment data that is coded data of the media content for the second segment, the first segment being different from the second segment; and
a processor that, in operation, decodes the first segment data to present the media content of the first segment according to the metadata describing the first segment data as the coded data of the media content for the first segment and decodes the second segment data to present the media content of the second segment according to the metadata describing the second segment data as the coded data of the media content for the second segment.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/241,573 US20160360248A1 (en) | 1999-07-14 | 2016-08-19 | Method and apparatus for decoding segments of an audiovisual stream |
US16/123,766 US20190007718A1 (en) | 1999-07-14 | 2018-09-06 | Method and apparatus for decoding metadata including multiple segments of an audiovisual stream |
US16/260,613 US20190182519A1 (en) | 1999-07-14 | 2019-01-29 | Method and apparatus for decoding metadata including multiple segments of an audiovisual stream |
Applications Claiming Priority (10)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JPH11-200095 | 1999-07-14 | ||
JP20009599 | 1999-07-14 | ||
PCT/JP2000/004736 WO2001006688A1 (en) | 1999-07-14 | 2000-07-14 | Apparatus for providing information, information receiver and storage medium |
US1931902A | 2002-01-10 | 2002-01-10 | |
US12/111,021 US7836479B2 (en) | 1999-07-14 | 2008-04-28 | Information provisioning apparatus and information provisioning method |
US12/899,860 US7992182B2 (en) | 1999-07-14 | 2010-10-07 | Information provisioning apparatus and information provisioning method |
US13/165,111 US8555328B2 (en) | 1999-07-14 | 2011-06-21 | Information provisioning apparatus and information provisioning method |
US13/609,848 US8832762B2 (en) | 1999-07-14 | 2012-09-11 | Information provisioning apparatus and information provisioning method |
US14/225,977 US9451293B2 (en) | 1999-07-14 | 2014-03-26 | Apparatus and method for decoding a segment of an audiovisual stream |
US15/241,573 US20160360248A1 (en) | 1999-07-14 | 2016-08-19 | Method and apparatus for decoding segments of an audiovisual stream |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/225,977 Continuation US9451293B2 (en) | 1999-07-14 | 2014-03-26 | Apparatus and method for decoding a segment of an audiovisual stream |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/123,766 Continuation US20190007718A1 (en) | 1999-07-14 | 2018-09-06 | Method and apparatus for decoding metadata including multiple segments of an audiovisual stream |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160360248A1 true US20160360248A1 (en) | 2016-12-08 |
Family
ID=16418776
Family Applications (9)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/019,319 Expired - Lifetime US7383566B1 (en) | 1999-07-14 | 2000-07-14 | Apparatus for providing information, information receiver and storage medium |
US12/111,021 Expired - Fee Related US7836479B2 (en) | 1999-07-14 | 2008-04-28 | Information provisioning apparatus and information provisioning method |
US12/899,860 Expired - Fee Related US7992182B2 (en) | 1999-07-14 | 2010-10-07 | Information provisioning apparatus and information provisioning method |
US13/165,111 Expired - Fee Related US8555328B2 (en) | 1999-07-14 | 2011-06-21 | Information provisioning apparatus and information provisioning method |
US13/609,848 Expired - Fee Related US8832762B2 (en) | 1999-07-14 | 2012-09-11 | Information provisioning apparatus and information provisioning method |
US14/225,977 Expired - Fee Related US9451293B2 (en) | 1999-07-14 | 2014-03-26 | Apparatus and method for decoding a segment of an audiovisual stream |
US15/241,573 Abandoned US20160360248A1 (en) | 1999-07-14 | 2016-08-19 | Method and apparatus for decoding segments of an audiovisual stream |
US16/123,766 Abandoned US20190007718A1 (en) | 1999-07-14 | 2018-09-06 | Method and apparatus for decoding metadata including multiple segments of an audiovisual stream |
US16/260,613 Abandoned US20190182519A1 (en) | 1999-07-14 | 2019-01-29 | Method and apparatus for decoding metadata including multiple segments of an audiovisual stream |
Family Applications Before (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/019,319 Expired - Lifetime US7383566B1 (en) | 1999-07-14 | 2000-07-14 | Apparatus for providing information, information receiver and storage medium |
US12/111,021 Expired - Fee Related US7836479B2 (en) | 1999-07-14 | 2008-04-28 | Information provisioning apparatus and information provisioning method |
US12/899,860 Expired - Fee Related US7992182B2 (en) | 1999-07-14 | 2010-10-07 | Information provisioning apparatus and information provisioning method |
US13/165,111 Expired - Fee Related US8555328B2 (en) | 1999-07-14 | 2011-06-21 | Information provisioning apparatus and information provisioning method |
US13/609,848 Expired - Fee Related US8832762B2 (en) | 1999-07-14 | 2012-09-11 | Information provisioning apparatus and information provisioning method |
US14/225,977 Expired - Fee Related US9451293B2 (en) | 1999-07-14 | 2014-03-26 | Apparatus and method for decoding a segment of an audiovisual stream |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/123,766 Abandoned US20190007718A1 (en) | 1999-07-14 | 2018-09-06 | Method and apparatus for decoding metadata including multiple segments of an audiovisual stream |
US16/260,613 Abandoned US20190182519A1 (en) | 1999-07-14 | 2019-01-29 | Method and apparatus for decoding metadata including multiple segments of an audiovisual stream |
Country Status (8)
Country | Link |
---|---|
US (9) | US7383566B1 (en) |
EP (3) | EP2288169A3 (en) |
JP (1) | JP4486291B2 (en) |
KR (3) | KR100472254B1 (en) |
CN (3) | CN1206828C (en) |
AU (1) | AU6016600A (en) |
DE (1) | DE60041259D1 (en) |
WO (1) | WO2001006688A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090006488A1 (en) * | 2007-06-28 | 2009-01-01 | Aram Lindahl | Using time-stamped event entries to facilitate synchronizing data streams |
Families Citing this family (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1206828C (en) * | 1999-07-14 | 2005-06-15 | 松下电器产业株式会社 | Apparatus for providing information, information receiver and storage medium |
AU3512301A (en) * | 2001-04-11 | 2002-10-17 | Kelvin Scott Duncan | Data management and distribution |
US7295755B2 (en) | 2001-06-22 | 2007-11-13 | Thomson Licensing | Method and apparatus for simplifying the access of metadata |
JP2003030204A (en) * | 2001-07-17 | 2003-01-31 | Takami Yasuda | Server for providing video contents, device and method for preparing file for video contents retrieval, computer program and device and method for supporting video clip preparation |
JP4491771B2 (en) * | 2001-11-29 | 2010-06-30 | 日本ビクター株式会社 | Optical transmission / reception system |
EP1387584A1 (en) | 2002-07-22 | 2004-02-04 | Deutsche Thomson-Brandt Gmbh | Method and apparatus for storing and transmiting audio-visual data |
US7991827B1 (en) * | 2002-11-13 | 2011-08-02 | Mcafee, Inc. | Network analysis system and method utilizing collected metadata |
FR2852772B1 (en) * | 2003-03-17 | 2005-07-08 | Thomson Licensing Sa | CODING METHOD APPLIED TO MULTIMEDIA DOCUMENTS AND DEVICE IMPLEMENTING THE METHOD |
KR100493896B1 (en) | 2003-04-18 | 2005-06-10 | 삼성전자주식회사 | Method and Apparatus for Transforming Digital Content Metadata, and Network System Using the Same |
KR100513290B1 (en) | 2003-06-30 | 2005-09-09 | 삼성전자주식회사 | System and Method for synchronizing between multimedia contents and segment metadata |
US20050281289A1 (en) * | 2004-06-21 | 2005-12-22 | Huang Jau H | System and method for embedding multimedia processing information in a multimedia bitstream |
US20060059510A1 (en) * | 2004-09-13 | 2006-03-16 | Huang Jau H | System and method for embedding scene change information in a video bitstream |
CN101107828B (en) * | 2004-10-05 | 2013-10-30 | 维克多曼克斯公司 | Method and system for broadcasting multimedia data |
KR100631783B1 (en) | 2004-10-06 | 2006-10-11 | 삼성전자주식회사 | System and method for synchronizing multimedia data with metadata |
US20060078288A1 (en) * | 2004-10-12 | 2006-04-13 | Huang Jau H | System and method for embedding multimedia editing information in a multimedia bitstream |
US7885921B2 (en) | 2004-11-18 | 2011-02-08 | International Business Machines Corporation | Managing atomic updates on metadata tracks in a storage system |
US8676748B2 (en) | 2004-11-18 | 2014-03-18 | International Business Machines Corporation | Clearing metadata tracks in a storage system |
US8856467B2 (en) | 2004-11-18 | 2014-10-07 | International Business Machines Corporation | Management of metadata in a storage subsystem |
KR100677545B1 (en) * | 2004-12-29 | 2007-02-02 | 삼성전자주식회사 | Method for data processing using a plurality of data processing apparatus, and recoding medium storing a program for implementing the method |
KR100808143B1 (en) * | 2005-12-01 | 2008-02-29 | 한국전자통신연구원 | XML data stream processing system and method for managing the access control using XPath |
EP2052544B1 (en) | 2006-08-14 | 2014-07-02 | NDS Limited | Controlled metadata revelation |
US7692562B1 (en) * | 2006-10-18 | 2010-04-06 | Hewlett-Packard Development Company, L.P. | System and method for representing digital media |
JP5182178B2 (en) * | 2009-03-18 | 2013-04-10 | ソニー株式会社 | Information processing apparatus and information processing method |
JP5720095B2 (en) | 2009-12-18 | 2015-05-20 | ソニー株式会社 | Reception device, reception method, transmission device, transmission method, program, and broadcasting system |
TWI574565B (en) * | 2011-03-31 | 2017-03-11 | Sony Corp | A transmitting apparatus, a receiving method, a receiving apparatus, a receiving method, a computer-readable medium, and a broadcasting system |
TW201249198A (en) | 2011-04-21 | 2012-12-01 | Sony Corp | Supplying apparatus, supplying method, receiving apparatus, receiving method, program, and broadcasting system |
WO2012174301A1 (en) | 2011-06-14 | 2012-12-20 | Related Content Database, Inc. | System and method for presenting content with time based metadata |
US9875480B2 (en) * | 2012-01-27 | 2018-01-23 | Sony Network Entertainment International Llc | System, method, and infrastructure for real-time live streaming content |
KR101947000B1 (en) * | 2012-07-17 | 2019-02-13 | 삼성전자주식회사 | Apparatus and method for delivering transport characteristics of multimedia data in broadcast system |
CN109413447B (en) * | 2012-08-22 | 2021-03-30 | 华为技术有限公司 | ISO-BMFF event box bearer in MPEG-2 transport stream |
GB2505912B (en) * | 2012-09-14 | 2015-10-07 | Canon Kk | Method and device for generating a description file, and corresponding streaming method |
JP6352931B2 (en) * | 2012-10-11 | 2018-07-04 | サムスン エレクトロニクス カンパニー リミテッド | Apparatus and method for transmitting and receiving multimedia data in a hybrid network |
GB2509178B (en) * | 2012-12-24 | 2015-10-14 | Canon Kk | Method for transmitting metadata documents associated with a video |
US9723245B2 (en) * | 2013-03-15 | 2017-08-01 | Samsung Electronics Co., Ltd. | Data transmitting apparatus, data receiving apparatus, data transceiving system, method for transmitting data, and method for receiving data |
KR102201826B1 (en) * | 2013-03-15 | 2021-01-12 | 삼성전자주식회사 | Data transmitting appratus, data receiving apparatus, data tranceiving system, method for transmitting data and method for receiving data |
US10356484B2 (en) * | 2013-03-15 | 2019-07-16 | Samsung Electronics Co., Ltd. | Data transmitting apparatus, data receiving apparatus, data transceiving system, method for transmitting data, and method for receiving data |
US9819604B2 (en) * | 2013-07-31 | 2017-11-14 | Nvidia Corporation | Real time network adaptive low latency transport stream muxing of audio/video streams for miracast |
JP6625318B2 (en) * | 2013-08-29 | 2019-12-25 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | Transmission method and reception method |
CN105491394B (en) * | 2014-09-17 | 2020-05-15 | 北京三星通信技术研究有限公司 | Method and device for sending MMT packet and method for receiving MMT packet |
EP3201923B1 (en) | 2014-10-03 | 2020-09-30 | Dolby International AB | Smart access to personalized audio |
US9838571B2 (en) | 2015-04-10 | 2017-12-05 | Gvbb Holdings S.A.R.L. | Precision timing for broadcast network |
JP6215898B2 (en) * | 2015-11-16 | 2017-10-18 | 株式会社Pfu | Video processing apparatus, video processing system, and video processing method |
TWI618405B (en) * | 2017-03-23 | 2018-03-11 | 和碩聯合科技股份有限公司 | Multimedia recording method and device thereof |
US11374992B2 (en) * | 2018-04-02 | 2022-06-28 | OVNIO Streaming Services, Inc. | Seamless social multimedia |
US11013340B2 (en) | 2018-05-23 | 2021-05-25 | L&P Property Management Company | Pocketed spring assembly having dimensionally stabilizing substrate |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010018693A1 (en) * | 1997-08-14 | 2001-08-30 | Ramesh Jain | Video cataloger system with synchronized encoders |
US20010040903A1 (en) * | 1997-03-28 | 2001-11-15 | Shinji Negishi | Multiplexing apparatus and method, transmitting apparatus and method, and recording medium |
US6341192B1 (en) * | 1997-08-08 | 2002-01-22 | Sony Corporation | Decoding method and decoding player and recording medium |
US20030128759A1 (en) * | 1999-04-17 | 2003-07-10 | Pulsent Corporation | Segment-based encoding system including segment-specific metadata |
Family Cites Families (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5119465A (en) * | 1989-06-19 | 1992-06-02 | Digital Equipment Corporation | System for selectively converting plurality of source data structures through corresponding source intermediate structures, and target intermediate structures into selected target structure |
US5175810A (en) * | 1989-06-19 | 1992-12-29 | Digital Equipment Corporation | Tabular data format |
US5253275A (en) * | 1991-01-07 | 1993-10-12 | H. Lee Browne | Audio and video transmission and receiving system |
US6112226A (en) * | 1995-07-14 | 2000-08-29 | Oracle Corporation | Method and apparatus for concurrently encoding and tagging digital information for allowing non-sequential access during playback |
US6119154A (en) * | 1995-07-14 | 2000-09-12 | Oracle Corporation | Method and apparatus for non-sequential access to an in-progress video feed |
US5966120A (en) * | 1995-11-21 | 1999-10-12 | Imedia Corporation | Method and apparatus for combining and distributing data with pre-formatted real-time video |
EP0866612B1 (en) | 1996-09-11 | 2003-11-12 | Matsushita Electric Industrial Co., Ltd. | Program reception/execution apparatus which can start execution of program even when only part of program is received, and program transmitter for it |
US20030093790A1 (en) * | 2000-03-28 | 2003-05-15 | Logan James D. | Audio and video program recording, editing and playback systems using metadata |
JPH10145755A (en) | 1996-11-06 | 1998-05-29 | Matsushita Electric Ind Co Ltd | Data output controller |
US6295380B1 (en) * | 1997-02-27 | 2001-09-25 | Matsushita Electric Industrial Co., Ltd. | Object data processing apparatus, object data recording apparatus, data storage media, data structure for transmission |
JPH10334871A (en) * | 1997-05-29 | 1998-12-18 | Matsushita Electric Ind Co Ltd | Battery-housing device |
US6263497B1 (en) * | 1997-07-31 | 2001-07-17 | Matsushita Electric Industrial Co., Ltd. | Remote maintenance method and remote maintenance apparatus |
US6567980B1 (en) * | 1997-08-14 | 2003-05-20 | Virage, Inc. | Video cataloger system with hyperlinked output |
US7295752B1 (en) * | 1997-08-14 | 2007-11-13 | Virage, Inc. | Video cataloger system with audio track extraction |
US6151602A (en) * | 1997-11-07 | 2000-11-21 | Inprise Corporation | Database system with methods providing a platform-independent self-describing data packet for transmitting information |
US7428547B2 (en) * | 1998-01-26 | 2008-09-23 | At&T Corp. | System and method of organizing data to facilitate access and streaming |
US6493720B1 (en) * | 1998-01-26 | 2002-12-10 | International Business Machines Corporation | Method and system for synchronization of metadata in an information catalog |
US6144375A (en) * | 1998-08-14 | 2000-11-07 | Praja Inc. | Multi-perspective viewer for content-based interactivity |
JP2000224257A (en) | 1999-01-29 | 2000-08-11 | Jisedai Joho Hoso System Kenkyusho:Kk | Transmitter and receiver |
JP3176893B2 (en) | 1999-03-05 | 2001-06-18 | 株式会社次世代情報放送システム研究所 | Digest creating apparatus, digest creating method, and computer-readable recording medium storing a program for causing a computer to execute each step of the method |
JP3434232B2 (en) | 1999-03-08 | 2003-08-04 | 株式会社次世代情報放送システム研究所 | Receiver and method |
JP3763992B2 (en) * | 1999-03-30 | 2006-04-05 | 富士通株式会社 | Data processing apparatus and recording medium |
US7050503B2 (en) * | 1999-04-17 | 2006-05-23 | Pts Corporation | Segment-based encoding system using residue coding by basis function coefficients |
US6574655B1 (en) * | 1999-06-29 | 2003-06-03 | Thomson Licensing Sa | Associative management of multimedia assets and associated resources using multi-domain agent-based communication between heterogeneous peers |
CN1206828C (en) * | 1999-07-14 | 2005-06-15 | 松下电器产业株式会社 | Apparatus for providing information, information receiver and storage medium |
US7137065B1 (en) * | 2000-02-24 | 2006-11-14 | International Business Machines Corporation | System and method for classifying electronically posted documents |
US6429924B1 (en) * | 2000-11-30 | 2002-08-06 | Eastman Kodak Company | Photofinishing method |
JP4099973B2 (en) * | 2001-10-30 | 2008-06-11 | 松下電器産業株式会社 | Video data transmission method, video data reception method, and video surveillance system |
JP4062230B2 (en) * | 2003-10-10 | 2008-03-19 | ソニー株式会社 | File recording / reproducing apparatus, file recording / reproducing method, file recording / reproducing method program, and recording medium recording the file recording / reproducing method program |
KR20050052790A (en) * | 2003-12-01 | 2005-06-07 | 엘지전자 주식회사 | Method for managing and reproducing a file of high density optical disc |
JP2005167742A (en) * | 2003-12-03 | 2005-06-23 | Funai Electric Co Ltd | Dvd recorder and video signal recording apparatus |
US8201073B2 (en) * | 2005-08-15 | 2012-06-12 | Disney Enterprises, Inc. | System and method for automating the creation of customized multimedia content |
JP2007173987A (en) * | 2005-12-19 | 2007-07-05 | Canon Inc | Multimedia data transmission/reception system and device, or program |
-
2000
- 2000-07-14 CN CNB008101582A patent/CN1206828C/en not_active Expired - Lifetime
- 2000-07-14 US US10/019,319 patent/US7383566B1/en not_active Expired - Lifetime
- 2000-07-14 CN CNB2004100474625A patent/CN100469139C/en not_active Expired - Lifetime
- 2000-07-14 EP EP20100173567 patent/EP2288169A3/en not_active Withdrawn
- 2000-07-14 EP EP20080165764 patent/EP2023514B1/en not_active Expired - Lifetime
- 2000-07-14 CN CNB200410047463XA patent/CN100435569C/en not_active Expired - Lifetime
- 2000-07-14 EP EP00946327A patent/EP1193899B1/en not_active Expired - Lifetime
- 2000-07-14 DE DE60041259T patent/DE60041259D1/en not_active Expired - Lifetime
- 2000-07-14 AU AU60166/00A patent/AU6016600A/en not_active Abandoned
- 2000-07-14 KR KR10-2004-7008247A patent/KR100472254B1/en active IP Right Grant
- 2000-07-14 JP JP2001511018A patent/JP4486291B2/en not_active Expired - Lifetime
- 2000-07-14 KR KR10-2002-7000514A patent/KR100462695B1/en active IP Right Grant
- 2000-07-14 KR KR10-2004-7008250A patent/KR100472255B1/en active IP Right Grant
- 2000-07-14 WO PCT/JP2000/004736 patent/WO2001006688A1/en active IP Right Grant
-
2008
- 2008-04-28 US US12/111,021 patent/US7836479B2/en not_active Expired - Fee Related
-
2010
- 2010-10-07 US US12/899,860 patent/US7992182B2/en not_active Expired - Fee Related
-
2011
- 2011-06-21 US US13/165,111 patent/US8555328B2/en not_active Expired - Fee Related
-
2012
- 2012-09-11 US US13/609,848 patent/US8832762B2/en not_active Expired - Fee Related
-
2014
- 2014-03-26 US US14/225,977 patent/US9451293B2/en not_active Expired - Fee Related
-
2016
- 2016-08-19 US US15/241,573 patent/US20160360248A1/en not_active Abandoned
-
2018
- 2018-09-06 US US16/123,766 patent/US20190007718A1/en not_active Abandoned
-
2019
- 2019-01-29 US US16/260,613 patent/US20190182519A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010040903A1 (en) * | 1997-03-28 | 2001-11-15 | Shinji Negishi | Multiplexing apparatus and method, transmitting apparatus and method, and recording medium |
US6341192B1 (en) * | 1997-08-08 | 2002-01-22 | Sony Corporation | Decoding method and decoding player and recording medium |
US20010018693A1 (en) * | 1997-08-14 | 2001-08-30 | Ramesh Jain | Video cataloger system with synchronized encoders |
US20030128759A1 (en) * | 1999-04-17 | 2003-07-10 | Pulsent Corporation | Segment-based encoding system including segment-specific metadata |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090006488A1 (en) * | 2007-06-28 | 2009-01-01 | Aram Lindahl | Using time-stamped event entries to facilitate synchronizing data streams |
US9794605B2 (en) * | 2007-06-28 | 2017-10-17 | Apple Inc. | Using time-stamped event entries to facilitate synchronizing data streams |
Also Published As
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20190182519A1 (en) | Method and apparatus for decoding metadata including multiple segments of an audiovisual stream | |
EP1705918B1 (en) | Data receiving apparatus | |
JP2002010182A (en) | Method for storing data, receiver realizing the same as well as broadcasting system | |
JP2000224257A (en) | Transmitter and receiver | |
JP4135251B2 (en) | Information processing device | |
JP7054780B2 (en) | Transmission method, reception method, transmission device and reception device | |
JP4378780B2 (en) | Receiving apparatus and receiving method | |
JP7054828B2 (en) | Transmitter, receiver, transmitter and receiver | |
JP2000295586A (en) | Information processor and information processing method for broadcast | |
JP2002171522A (en) | Division method, transmission method and integrating method for structural meta data | |
JP4378778B2 (en) | Receiving apparatus and receiving method | |
KR100461540B1 (en) | Apparatus and Method for Buffering Sychronized Non_streaming Data | |
JP2000032362A (en) | Device and method for transmitting information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: EXPRESSLY ABANDONED -- DURING PUBLICATION PROCESS |