US11622088B2 - Reception apparatus, transmission apparatus, and data processing method - Google Patents
Reception apparatus, transmission apparatus, and data processing method Download PDFInfo
- Publication number
- US11622088B2 US11622088B2 US15/564,056 US201615564056A US11622088B2 US 11622088 B2 US11622088 B2 US 11622088B2 US 201615564056 A US201615564056 A US 201615564056A US 11622088 B2 US11622088 B2 US 11622088B2
- Authority
- US
- United States
- Prior art keywords
- ttml
- file
- closed caption
- time
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/435—Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/08—Systems for the simultaneous or sequential transmission of more than one television signal, e.g. additional information signals, the signals occupying wholly or partially the same frequency band, e.g. by time division
- H04N7/087—Systems for the simultaneous or sequential transmission of more than one television signal, e.g. additional information signals, the signals occupying wholly or partially the same frequency band, e.g. by time division with signal insertion during the vertical blanking interval only
- H04N7/088—Systems for the simultaneous or sequential transmission of more than one television signal, e.g. additional information signals, the signals occupying wholly or partially the same frequency band, e.g. by time division with signal insertion during the vertical blanking interval only the inserted signal being digital
- H04N7/0884—Systems for the simultaneous or sequential transmission of more than one television signal, e.g. additional information signals, the signals occupying wholly or partially the same frequency band, e.g. by time division with signal insertion during the vertical blanking interval only the inserted signal being digital for the transmission of additional display-information, e.g. menu for programme or channel selection
- H04N7/0885—Systems for the simultaneous or sequential transmission of more than one television signal, e.g. additional information signals, the signals occupying wholly or partially the same frequency band, e.g. by time division with signal insertion during the vertical blanking interval only the inserted signal being digital for the transmission of additional display-information, e.g. menu for programme or channel selection for the transmission of subtitles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/23611—Insertion of stuffing data into a multiplex stream, e.g. to obtain a constant bitrate
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4884—Data services, e.g. news ticker for displaying subtitles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/85406—Content authoring involving a specific file format, e.g. MP4 format
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/178—Metadata, e.g. disparity information
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/183—On-screen display [OSD] information, e.g. subtitles or menus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/23614—Multiplexing of additional data and video streams
Definitions
- the present technology relates to a reception apparatus, a transmission apparatus, and a data processing method and, more particularly, to a reception apparatus, a transmission apparatus, and a data processing method which are configured to display a closed caption at desired time.
- TTML timed text markup language
- W3C World Wide Web Consortium
- the present technology has been made in view of such circumstances, and is configured to display a closed caption at desired time.
- a reception apparatus including circuitry configured to receive a digital data stream and acquire closed caption information included in the digital data stream.
- the circuitry is configured to acquire control information including selection information indicating a selection of a specific mode from a plurality of modes for specifying when closed caption text is to be displayed.
- the circuitry is further configured to output the closed caption text included in the closed caption information for display to a user, at a display time according to the specific mode, based on the selection information included in the control information.
- a data processing method including acquiring, by circuitry of a reception apparatus, closed caption information included in a digital data stream.
- the method includes acquiring, by the circuitry of the reception apparatus, control information including selection information indicating a selection of a specific mode from a plurality of modes for specifying when closed caption text is to be displayed.
- the method further includes controlling, by the circuitry of the reception apparatus, output of the closed caption text included in the closed caption information, at a display time according to the specific mode, based on the selection information included in the control information.
- a transmission apparatus including circuitry configured to generate control information including selection information indicating a selection of a specific mode from a plurality of modes for specifying when closed caption text is to be displayed.
- the circuitry is further configured to transmit a digital data stream including the control information and closed caption information including the closed caption text.
- a data processing method including generating, by circuitry of a transmission apparatus, control information including selection information indicating a selection of a specific mode from a plurality of modes for specifying when closed caption text is to be displayed.
- the method further includes transmitting, by the circuitry of the transmission apparatus, a digital data stream including the control information and closed caption information including the closed caption text.
- the closed caption can be displayed at desired time.
- FIG. 1 is a diagram illustrating a configuration of an embodiment of a transmission system to which the present technology is applied.
- FIG. 2 is a diagram illustrating a protocol stack according to the present example.
- FIG. 3 is a diagram illustrating a structure of an MPD file.
- FIG. 4 is a diagram illustrating a description example of the MPD file.
- FIG. 5 is a diagram illustrating a relationship between AdaptationSet elements and Representation elements to represent a multiplexed stream.
- FIG. 6 is a diagram illustrating a relationship between AdaptationSet elements, Representation elements, and SubRepresentation elements to represent a multiplexed stream.
- FIG. 7 is a diagram illustrating exemplary attributes and elements which can be included in an AdaptationSet element.
- FIG. 8 is a diagram illustrating exemplary attributes and elements which can be included in a Representation element.
- FIG. 9 is a diagram illustrating an MP4 file format.
- FIG. 10 is a table illustrating an exemplary TTML processing mode.
- FIG. 11 is a diagram illustrating a description example of an MPD file operated in mode 1.
- FIG. 12 is a diagram illustrating an example of closed caption display timing upon operation in mode 1.
- FIG. 13 is a diagram illustrating a description example of an MPD file operated in mode 2-1.
- FIG. 14 is a diagram illustrating an example of closed caption display timing upon operation in mode 2-1.
- FIG. 15 is a diagram illustrating a description example of an MPD file operated in mode 2-2.
- FIG. 16 is a diagram illustrating an example of closed caption display timing upon operation in mode 2-2.
- FIG. 17 is a diagram illustrating a description example of an MPD file operated in mode 3.
- FIG. 18 is a diagram illustrating an example of closed caption display timing upon operation in mode 3.
- FIG. 19 is a diagram illustrating an exemplary configuration of an ATSC server.
- FIG. 20 is a diagram illustrating a detailed exemplary configuration of the ATSC server.
- FIG. 21 is a diagram illustrating a detailed exemplary configuration of an ATSC client.
- FIG. 22 is a diagram illustrating an exemplary configuration of software of the ATSC client.
- FIG. 23 is a flowchart illustrating transmission process.
- FIG. 24 is a flowchart illustrating component signaling process.
- FIG. 25 is a flowchart illustrating reception process.
- FIG. 26 is a flowchart illustrating component signaling process.
- FIG. 27 is a diagram illustrating an exemplary configuration of a computer.
- Mode 1 TTML Time Only
- FIG. 1 is a diagram illustrating a configuration of an embodiment of a transmission system to which the present technology is applied. Note that a system represents a logical assembly of a plurality of apparatuses.
- the transmission system 1 includes an ATSC server 10 and an ATSC client 20 .
- data is transmitted in accordance with a digital broadcasting standard such as ATSC 3.0.
- ATSC 3.0 is a next generation of Advanced Television Systems Committee (ATSC) standard under development.
- ATSC Advanced Television Systems Committee
- the ATSC server 10 is a transmitter in compliance with the digital broadcasting standard such as ATSC 3.0, and includes, for example, a plurality of servers.
- the ATSC server 10 transmits, as a digital broadcasting signal, a video, audio, or closed caption (component) stream or the like constituting content such as a TV program, through a transmission path 30 .
- the ATSC client 20 is a receiver in compliance with the digital broadcasting standard such as ATSC 3.0, and includes, for example, a fixed receiver such as a TV set or a set-top box, or a mobile receiver such as a smartphone, a mobile phone, or a tablet computer. Furthermore, the ATSC client 20 may be for example a device mounted to a motor vehicle, such as a vehicle TV.
- the ATSC client 20 receives a digital broadcasting signal transmitted from the ATSC server 10 through the transmission path 30 , acquires and processes the video, audio, or closed caption (component) stream or the like, and outputs video or sound of the content such as a TV program.
- the transmission path 30 for example a satellite circuit, a cable television network (wired circuit), or the like can be used, in addition to a ground wave.
- the ATSC 3.0 is determined not to use a transport stream (TS) packet but to use an IP/UDP packet, that is, an Internet protocol (IP) packet including a user datagram protocol (UDP) packet, for data transmission.
- IP Internet protocol
- UDP user datagram protocol
- the ATSC 3.0 has parallel transport protocols, that is, Real-Time Object Delivery over Unidirectional Transport (ROUTE) and MPEG Media Transport (MMT), and any of the transport protocols is used to transmit the video, audio, or closed caption (component) stream or the like.
- ROUTE Real-Time Object Delivery over Unidirectional Transport
- MMT MPEG Media Transport
- the ROUTE is an expanded protocol of a file delivery over unidirectional transport (FLUTE) which is a protocol suitable for unidirectional multicast transfer of a binary file.
- the MMT is a transport system used over an Internet Protocol (IP), and is configured to refer to data of video, audio, or the like by setting an IP address or a uniform resource locator (URL) based on control information.
- IP Internet Protocol
- the ATSC 3.0 is configured assuming provision of link layer signaling (LLS) information and service layer signaling (SLS) information, as signaling, and the SLS information is acquired for services, according to information described in the LLS information having been acquired.
- LLS link layer signaling
- SLS service layer signaling
- the LLS information now includes, for example, metadata such as a service list table (SLT).
- SLT metadata includes information representing a configuration of the stream or the service in a broadcast network, such as information (selection information) necessary for selection of service.
- the SLS information includes metadata such as user service description (USD), LCT session instance description (LSID), or media presentation description (MPD).
- the USD metadata includes information such as an acquisition destination of other metadata.
- the LSID metadata is control information of a ROUTE protocol.
- the MPD metadata is control information for managing reproduction of a component stream.
- the metadata such as USD, LSID, or MPD is described in a markup language such as an Extensible Markup Language (XML).
- XML Extensible Markup Language
- the MPD metadata is compliant with MPEG-dynamic adaptive streaming over http (DASH) standard.
- the MPD metadata is provided in an XML format file, so that the MPD metadata will be described below as an MPD file.
- FIG. 2 is a diagram illustrating a protocol stack according to the present example.
- the lowest layer is defined as a physical layer (Physical Layer).
- An upper layer adjacent to the physical layer is defined as a layer 2 (Layer2).
- an upper layer adjacent to the layer 2 is defined as an IP layer.
- an upper layer adjacent to the IP layer is defined as a UDP layer. That is, an IP packet including a UDP packet (IP/UDP packet) is arranged and encapsulated in a payload of a generic packet in the layer 2.
- IP/UDP packet IP/UDP packet
- the physical layer has a frame (ATSC physical frame) including a preamble and a data section, and in the data section, data is mapped which is obtained by processing about the physical layer, such as interleave or mapping, after error correction parity is added to a BB frame obtained by encapsulating a plurality of generic packets.
- ASC physical frame including a preamble and a data section
- data is mapped which is obtained by processing about the physical layer, such as interleave or mapping, after error correction parity is added to a BB frame obtained by encapsulating a plurality of generic packets.
- An upper layer adjacent to the UDP layer includes the ROUTE, the MMT, and the SLT. That is, the video, audio, and closed caption streams, an SLS information stream, and an NRT content stream are stored in the IP/UDP packet, and transmitted in a ROUTE session.
- the NRT content is content distributed by non real time (NRT) broadcast, and is reproduced after once stored in a storage of the ATSC client 20 .
- a file other than the NRT content e.g., application file
- the video, audio, and closed caption streams, and the SLS information stream are stored in the IP/UDP packet, and transmitted in an MMT session.
- the SLT metadata is transmitted by being stored in the IP/UDP packet.
- the ATSC client 20 when selecting a service (channel) provided by the component stream transmitted in the ROUTE session, acquires the SLS information transmitted in the ROUTE session, according to the selection information included in the SLT metadata (S 1 - 1 , S 1 - 2 ). Then, the ATSC client 20 is connected to the component stream providing the selected service, according to the metadata such as the USD, LSID, or MPD (S 1 - 3 ). Thus, the ATSC client 20 outputs video or sound of content (e.g., TV program) according to the selected service.
- the SLT metadata S 1 - 1 , S 1 - 2
- the ATSC client 20 is connected to the component stream providing the selected service, according to the metadata such as the USD, LSID, or MPD (S 1 - 3 ).
- the ATSC client 20 outputs video or sound of content (e.g., TV program) according to the selected service.
- the ATSC client 20 when selecting a service provided by the component stream transmitted in the MMT session, acquires the SLS information transmitted in the MMT session, according to the selection information included in the SLT metadata (S 2 - 1 , S 2 - 2 ). Then, the ATSC client 20 is connected to the component stream providing the selected service, according to the metadata such as the USD, LSID, or MPD (S 2 - 3 ). Thus, the ATSC client 20 outputs video or sound of content (e.g., TV program) according to the selected service.
- the SLT metadata S 2 - 1 , S 2 - 2
- MPD MPD
- FIG. 3 is a diagram illustrating a structure of the MPD file. Furthermore, the MPD file is described in an XML hierarchical structure, as illustrated in a description example of FIG. 4 .
- the MPD file includes a Period element, an AdaptationSet element, a Representation element, and a SubRepresentation element which are described in a hierarchical structure.
- the Period element is a unit for description of a configuration of content such as a TV program.
- the AdaptationSet element, the Representation element, or the SubRepresentation element is used for each of the video, audio, and closed caption (component) streams and the like to describe attributes of each stream.
- the AdaptationSet elements represent streams which are encoded from various sources.
- the Representation elements are arranged in the AdaptationSet element to have a list of, for example, a plurality of streams to be selected having different parameters such as bit rates.
- the AdaptationSet element or the Representation element corresponds to a single stream such as the video, audio, or closed caption stream.
- the Representation elements are arranged in the AdaptationSet element to have a list of, for example, a plurality of multiplexed streams to be selected having different parameters such as bit rates. That is, as illustrated in FIG. 5 , a plurality of AdaptationSet elements representing the multiplexed stream is arranged in each of the Period elements representing a time interval, and a plurality of Representation elements are arranged in each of the AdaptationSet elements to have a list of, for example, a plurality of multiplexed streams having different bit rates.
- SubRepresentation elements can be arranged in subordination to the Representation elements to describe the attributes of the component streams constituting the multiplexed stream. That is, as illustrated in FIG. 6 , a plurality of AdaptationSet elements representing the multiplexed stream is arranged in each Period element representing a time interval, for example, a plurality of Representation elements representing the multiplexed stream having different bit rates are arranged in each AdaptationSet element, and furthermore, the SubRepresentation elements are arranged in each Representation element, and for example, an attribute of the video stream, audio stream, or closed caption stream can be described therein.
- the AdaptationSet element corresponds to a multiplexed stream of a plurality of streams, in addition to a single stream such as the video stream or the audio stream.
- the attributes and elements of such an AdaptationSet element are defined as illustrated in FIG. 7 .
- the Representation element includes, for example, a plurality of streams to be selected having different parameters such as bit rates within the range of the AdaptationSet element being a higher element (parent element).
- the attributes and elements of such a Representation element are defined as illustrated in FIG. 8 . Note that the attributes and elements of FIG. 8 can be included in the SubRepresentation element.
- an MP4 file format may be employed for a streaming file format.
- the MP4 file format is derived from an ISO base media file format defined in an ISO/IEC 14496-12.
- the ISO base media file format includes a tree structure referred to as a box (Box).
- a segment transmitted in the ROUTE session includes an initialization segment and a media segment.
- the initialization segment includes initialization information such as a data compression method.
- the media segment stores data of the video, audio, or closed caption (component) stream or the like.
- FIG. 9 is a diagram illustrating a structure of the media segment in the MP4 file format.
- the media segment includes boxes (Box) of a segment type (styp), a segment index (sidx), a subsegmentindex (ssix), a movie fragment (moot), and a media data (mdat).
- the styp box includes version information of a file format specification of a segment file.
- the sidx box includes index information in the segment.
- the ssix box includes index information for each sub-segment (level) in the segment.
- the styp, sidx, and ssix boxes are not illustrated.
- the moof box includes control information about fragmented (component) stream data.
- the moof box includes a movie fragment header (mfhd) box.
- the mfhd box includes a track fragment decode time (tfdt) box and a track fragment run (trun) box.
- the tfdt box includes Base Media Decode Time (BMDT) representing a decoding start time of a sample.
- BMDT Base Media Decode Time
- the trun box includes SampleCount representing the number of samples, SampleDuration representing duration of the samples, and information representing CompositionOffset indicating an offset value.
- the mdat box stores, as a sample, the fragmented (component) stream data (data body).
- the sample stores data of the stream (of component) in basic units to be processed.
- the samples include a sample of a file (TTML file) in TTML format for displaying a closed caption, and the sample will also be referred to as a TTML sample.
- the media segments include a media segment having the mdat box including the TTML sample, and the media segment will also be referred to as a TTML segment.
- a segment simply referred to as segment represents a media segment.
- the TTML file in TTML format may be used for display of the closed caption (CC).
- CC closed caption
- the TTML processing mode for specifying display timing of the closed caption defines a plurality of modes so that the ATSC server 10 includes, in the MPD file, selection information for selection of a specific mode from the plurality of modes.
- the ATSC client 20 can display a closed caption specified in the TTML file, at display time according to the specific mode, based on the selection information included in the MPD file.
- FIG. 10 is a table illustrating an exemplary TTML processing mode.
- Mode 1 mode 2, and mode 3 are defined as the TTML processing mode. Furthermore, mode 2 defines two types of modes, that is, mode 2-1 and mode 2-2.
- Mode 1 is a mode for display of the closed caption at time according to time information specified in the TTML file.
- “atsc:ttmlMode:ttmlTimeOnly” is specified for a value of a schemeIdUri attribute of an EssentialProperty element or a SupplementalProperty element of the AdaptationSet element.
- a body element includes a p element which specifies a character string of the closed caption, and the time information such as display start time or display finish time of the closed caption.
- Mode 2 is a mode for display of the closed caption at time according to time information defined in the MP4 file format, ignoring the time information specified in the TTML file.
- mode 2-1 of mode 2 display of the closed caption is started at time according to a BMDT stored in the moof box, and display of the closed caption is continued only during a time period according to SampleDuration stored in the moof box.
- mode 2-1 is set, in the MPD file, as the selection information, “atsc:ttmlMode:sampleTimeOnly” is specified for the value of the schemeIdUri attribute of the EssentialProperty element or the SupplementalProperty element of the AdaptationSet element.
- mode 2-2 display of the closed caption is started at time according to a BMDT stored in a moof box corresponding to a mdat box storing a target TTML sample, and display of the closed caption is continued until time according to a BMDT stored in a moof box corresponding to a mdat box storing a next TTML sample.
- mode 2-2 is set, in the MPD file, as the selection information, “atsc:ttmlMode:sampleTimeOnlyButTillNext” is specified for the value of the schemeIdUri attribute of the EssentialProperty element or the SupplementalProperty element of the AdaptationSet element.
- Mode 3 is a mode for display of the closed caption, ignoring the time information specified in the TTML file and the time information defined in the MP4 file format.
- the ATSC client 20 when the TTML file transmitted from the ATSC server 10 thorough the transmission path 30 is acquired, the closed caption specified in the TTML file is immediately displayed.
- mode 3 is set, in the MPD file, as the selection information, “atsc:ttmlMode:asap” is specified for the value of the schemeIdUri attribute of the EssentialProperty element or the SupplementalProperty element of the AdaptationSet element.
- Mode 1 TTML Time Only
- Mode 1 only uses the time information specified in the TTML file.
- FIG. 11 is a diagram illustrating a description example of an MPD file operated in mode 1.
- a Role element is arranged in the AdaptationSet element in subordination to the Period element of an MPD element being a root element, “urn:mpeg:dash:role:2011” is specified as the schemeIdUri attribute, and “closed caption” representing a closed caption is specified as a value attribute.
- AdaptationSet element “atsc:ttmlMode:ttmlTimeOnly” is specified as the schemeIdUri attribute of the EssentialProperty element. That is, mode 1 is set as the TTML processing mode by an attribute value of the schemeIdUri attribute of the EssentialProperty element.
- FIGS. 12 A to 12 C are diagrams illustrating an example of closed caption display timing upon operation in mode 1.
- FIG. 12 A schematically illustrates information about time specified in the MPD file
- FIG. 12 B illustrates a structure of the segment (TTML segment).
- FIG. 12 C illustrates a description example of the TTML file obtained from the sample (TTML sample) of the segment (TTML segment) of FIG. 12 B .
- a time direction is oriented from the left side to the right side in FIGS. 12 A to 12 C .
- these relationships are similarly applied to figures described later illustrating other modes.
- streaming start time according to coordinated universal time is specified for an availabilityStartTime attribute of the MPD element being the root element.
- Period (1), Period (2), . . . are specified as the Period element, and in the Period elements, start time is specified for each Period (period), as a start attribute.
- the sum of start time (time on WallClock time axis) specified in the availabilityStartTime attribute, and time (time on MPD time axis) specified in the Period element represents time at the head of each Period.
- time at the head of Period(2) is obtained based on the sum (MPD/@availabilityStartTime+MPD/Period(2)/@start) of the start time specified in the availabilityStartTime attribute (MPD/@availabilityStartTime), and the start time of Period(2) (MPD/Period(2)/@start).
- FIG. 12 B schematically illustrates a segment in Period (2).
- the segment is the TTML segment having the mdat box storing the TTML sample as the sample. From the TTML sample stored in the mdat box from the TTML segment, the TTML file is obtained. Note that time information (BMDT, SampleDuration, or the like) stored in the moof box is ignored, in mode 1.
- time information BMDT, SampleDuration, or the like
- FIG. 12 C illustrates the description example of the TTML file.
- a head element includes an element such as a styling element or a layout element which specifies color or font, or a display position of characters displayed as the closed caption.
- the p element in the body element specifies the character string of the closed caption, and the time information such as the display start time or the display finish time of the closed caption.
- the TTML file of FIG. 12 C has description of two p elements, and out of the two p elements, an upper p element has description of a character string “text1” between a begin tag and an end tag. Furthermore, in the upper p element, “t1” is specified as a begin attribute for setting the display start time, and “t2”is specified as an end attribute for setting the display finish time.
- a lower p element has description of a character string “text2” between the begin tag and the end tag. Furthermore, in the lower p element, “t2” is specified as the begin attribute, and “t3” is specified as the end attribute.
- “atsc:ttmlMode:ttmlTimeOnly” is specified as the schemeIdUri attribute of the EssentialProperty element of the AdaptationSet element, and mode 1 is set as the TTML processing mode.
- mode 1 is set, the closed caption is displayed at time according to the time information described in the TTML file, and thus, the closed caption is displayed at time according to values of the begin attribute and the end attribute of the p element.
- displaying a closed caption “text1” specified in the upper p element is started, after a time period t1 based on time at the head of Period(2) according to the sum (MPD/@availabilityStartTime+MPD Period(2) @start) of the start time specified in the availabilityStartTime attribute and the start time of Period(2), display of the closed caption “text1” is continued during a time period t2, and display of the closed caption “text1” is finished after the time period t2.
- displaying a closed caption “text2” specified in the lower p element is started, after the time period t2 based on time at the head of Period(2), displaying a closed caption “text2” is continued during a time period t3, and display of the closed caption “text2” is finished after the time period t3.
- the closed caption “text1” is displayed superimposed on the video, after the time period t1 and during the time period t2, based on time at the head of Period(2), and the closed caption “text2” is displayed superimposed on the video, after the time period t2 and during the time period t3.
- mode 1 This is the end of the description of the exemplary operation of the TTML processing mode in which mode 1 is set.
- the time information specified in the TTML file (values of the begin attribute and the end attribute of the p element) is used to start display of the closed caption according to the time period set for the begin attribute, and to finish display of the closed caption according to the time period set for the end attribute, and the closed caption can be displayed at desired time.
- the ATSC server 10 when a package of the video, sound, and closed caption recorded in a studio or the like is produced as content, and the content is distributed to a plurality of ATSC clients 20 through the transmission path 30 , it is considered suitable to employ operation for display of the closed caption specified in the p element of the TTML file, at time according to the time period indicated by the begin attribute and the end attribute of the p element, based on a time period relative to the head of the content.
- mode 1 is set as the TTML processing mode in the MPD file, and thus, the closed caption is displayed at time according to the time period indicated by the begin attribute and the end attribute of the p element of the TTML file, in the ATSC client 20 .
- Mode 2 uses the time information defined in the MP4 file format (time information for each TTML sample), ignoring the time information specified in the TTML file.
- Mode 2-1 and mode 2-2 will be sequentially described herein, as mode 2.
- FIG. 13 is a diagram illustrating a description example of an MPD file operated in mode 2-1.
- “atsc:ttmlMode:sampleTimeOnly” is specified as the schemeIdUri attribute of the EssentialProperty element, for the AdaptationSet element in subordination to the Period element of an MPD element being a root element. That is, mode 2-1 is set as the TTML processing mode by the attribute value of the schemeIdUri attribute of the EssentialProperty element.
- FIGS. 14 A to 14 C are diagrams illustrating an example of closed caption display timing upon operation in mode 2-1.
- the streaming start time is specified for the availabilityStartTime attribute of the MPD element. Furthermore, in the MPD file, Period (1), Period (2), . . . are specified as the Period element, and in the Period elements, the start time is specified for each Period (period), as the start attribute. Time at the head of Period(2) is obtained here, based on the sum (MPD/@availabilityStartTime+MPD/Period(2)/@start) of the start time specified in the availabilityStartTime attribute and the start time of Period(2).
- FIG. 14 B schematically illustrates the TTML segment in Period (2).
- the TTML file ( FIG. 14 C ) can be acquired from the TTML sample stored in the mdat box of the TTML segment, but, in the MPD file ( FIG. 13 ) acquired upon the operation, “atsc:ttmlMode:sampleTimeOnly” is specified in the schemeIdUri attribute of the EssentialProperty element, and mode 2-1 is set as the TTML processing mode. Therefore, the time information specified in the p element of the TTML file (“t1” of the begin attribute and “t3” of the end attribute) is ignored.
- the time information stored in the moof box of the TTML segment (time information for each TTML sample) is used, ignoring the time information specified in the TTML file.
- one moof box (time information stored therein) is assumed here to correspond to one TTML sample (TTML sample stored in the mdat box), in the TTML segment.
- display of the closed caption “text1” specified in the p element of the TTML file is started, after a time period (BMDT ⁇ ts) according to the BMDT stored in the moof box (moof/mfhd/tfdt), based on time at the head of Period(2) according to the sum (MPD/@availabilityStartTime+MPD/Period(2)/@start) of the start time specified in the availabilityStartTime attribute of the MPD file and the start time of Period(2).
- the BMDT (Base Media Decode Time) represents a time period from time at the head of Period(2) to time at which decode of the TTML sample is started, that is, offset. Furthermore, ts by which the BMDT is multiplied represents a time scale, and is used to convert a value of the BMDT to a value on the MPD time axis.
- SampleDuration represents duration of the TTML sample. Furthermore, ts by which SampleDuration is multiplied is used to convert a value of SampleDuration to a value on the MPD time axis.
- the closed caption “text1” is displayed superimposed on the video, based on time at the head of Period(2) after passage of the time period according to the BMDT stored in the moof box, and during the time period according to SampleDuration.
- mode 2-1 time information defined in the MP4 file format (BMDT and SampleDuration for each TTML sample) is used to start display of the closed caption at time according to the BMDT of the TTML sample, and continue display of the closed caption only during the time period according to SampleDuration, and the closed caption can be displayed at desired time.
- MP4 file format BMDT and SampleDuration for each TTML sample
- FIG. 15 is a diagram illustrating a description example of the MPD file operated in mode 2-2.
- FIGS. 16 A to 16 C are diagrams illustrating an example of closed caption display timing upon operation in mode 2-2.
- the streaming start time is specified for the availabilityStartTime attribute of the MPD element. Furthermore, in the MPD file, Period (1), Period (2), . . . are specified as the Period element, and in the Period elements, start time is specified for each Period (period), as a start attribute. Time at the head of Period(2) is obtained here, based on the sum (MPD/@availabilityStartTime+MPD/Period(2)/@start) of the start time specified in the availabilityStartTime attribute and the start time of Period(2).
- FIG. 16 B schematically illustrates the TTML segment in Period (2).
- the TTML file ( FIG. 16 C ) can be acquired from the TTML sample stored in the mdat box of the TTML segment, but, in the MPD file ( FIG. 15 ) acquired upon the operation, “atsc:ttmlMode:sampleTimeOnlyButTillNext” is specified in the schemeIdUri attribute of the EssentialProperty element, and mode 2-2 is set as the TTML processing mode. Therefore, the time information specified in the p element of the TTML file (“t1” of the begin attribute and “t3” of the end attribute) is ignored.
- the time information stored in the moof box of the TTML segment (time information for each TTML sample) is used, ignoring the time information specified in the TTML file.
- one moof box (time information stored therein) is assumed here to correspond to one TTML sample (TTML sample stored in the mdat box), in the TTML segment.
- FIGS. 16 A to 16 C display of the closed caption “text1” specified in the p element of the TTML file is started, after the passage of the time period (BMDT ⁇ ts) according to the BMDT stored in the moof box corresponding to the mdat box storing the target TTML sample (TTML sample of the TTML file specifying a target closed caption), based on time at the head of Period(2) according to the sum (MPD/@availabilityStartTime+MPD/Period(2)/@start) of the start time specified in the availabilityStartTime attribute of the MPD file and the start time of Period(2).
- the display of the closed caption “text1” is continued, until the passage of a time period (BMDT ⁇ ts) according to the BMDT stored in a moof box corresponding to an mdat box storing the next TTML sample (a TTML sample of a TTML file specifying a next closed caption), and display of the closed caption “text1” is finished after the passage of the time period (BMDT ⁇ ts) according to the BMDT.
- BMDT ⁇ ts a time period according to the BMDT stored in a moof box corresponding to an mdat box storing the next TTML sample (a TTML sample of a TTML file specifying a next closed caption)
- the closed caption “text1” is displayed superimposed on the video, based on time at the head of Period(2) after the time period according to the BMDT of (the moof box of) a TTML segment including the target TTML sample, and until passage of the time period according to the BMDT of (the moof box of) the TTML segment including the next TTML sample.
- next TTML sample (TTML sample of the TTML file specifying the next closed caption) is acquired, display of the closed caption “text1” is finished, and display of the closed caption (e.g., “text2”) specified in the next TTML sample (TTML file) is started.
- mode 2-2 time information defined in the MP4 file format (BMDT for each TTML sample) is used to start display of the closed caption at time according to the BMDT of the target TTML sample, and continue display of the closed caption until time according to the BMDT of the next TTML sample, and the closed caption can be displayed at desired time.
- mode 2-1 and mode 2-2 are common in specification of timing of starting closed caption display, at time according to BMDT, but are different in specification of timing of finishing the closed caption being displayed, between based on the time period according to SampleDuration and time according to the BMDT.
- the closed caption display timing seems to be operationally suitably specified in the time information defined in the MP4 file format (time information for each TTML sample), rather than in the TTML file.
- mode 2-1 and mode 2-2 are set as the TTML processing mode in the MPD file, and thus, the closed caption is displayed at time according to the time information defined in the MP4 file format (time information for each TTML sample), in the ATSC client 20 .
- Mode 3 performs processing for display of the closed caption immediately after the TTML file (TTML sample) is acquired, ignoring the time information specified in the TTML file, and the time information defined in the MP4 file format (time information for each TTML sample).
- FIG. 17 is a diagram illustrating a description example of an MPD file operated in mode 3.
- FIGS. 18 A to 18 C are diagrams illustrating an example of closed caption display timing upon operation in mode 3.
- the streaming start time is specified for the availabilityStartTime attribute of the MPD element. Furthermore, in the MPD file, Period (1), Period (2), . . . are specified as the Period element, and in the Period elements, start time is specified for each Period (period), as a start attribute.
- FIG. 18 B schematically illustrates the TTML segment in Period (2).
- the TTML file ( FIG. 18 C ) can be acquired from the TTML sample stored in the mdat box of the TTML segment, but, in the MPD file ( FIG. 17 ) acquired upon the operation, “atsc:ttmlMode:asap” is specified in the schemeIdUri attribute of the EssentialProperty element, and mode 3 is set as the TTML processing mode. Therefore, the time information specified in the p element of the TTML file (“t1” of the begin attribute and “t3” of the end attribute) is ignored. Furthermore, when mode 3 is set as the TTML processing mode, the time information (BMDT, SampleDuration, or the like) is also ignored which is stored in the moof box of the TTML segment.
- BMDT SampleDuration, or the like
- TTML file TTML sample
- the processing is performed to display the closed caption specified in the TTML file, ignoring the time information specified in the TTML file and the time information for each TTML sample.
- the TTML file is acquired from the TTML sample of the TTML segment, the TTML file is processed immediately, and display of the closed caption “text1” specified in the p element is started, ignoring the values of the begin attribute and the end attribute of the p element.
- the ATSC client 20 when the content such as a TV program is reproduced, the closed caption “text1” specified in the p attribute of the target TTML file is displayed superimposed on the video, after the TTML segment including the TTML sample of the target TTML file is acquired, and before the TTML segment including the TTML sample of the next TTML file is acquired.
- mode 3 This is the end of the description of the exemplary operation of the TTML processing mode in which mode 3 is set.
- mode 3 when the TTML file (TTML sample) is acquired, the closed caption specified in the TTML file is displayed immediately, ignoring the time information specified in the TTML file and the time information defined in the MP4 file format (time information for each TTML sample), and the closed caption can be displayed at desired time.
- the ATSC server 10 when content of live broadcast such as sports broadcast is distributed to a plurality of ATSC clients 20 through the transmission path 30 , minimum deviation between the live video and the closed caption assumed to be delayed from the live video is demanded.
- mode 3 is set as the TTML processing mode in the MPD file, and thus, the closed caption is displayed immediately after the TTML file is received, in the ATSC client 20 .
- FIG. 19 is a diagram illustrating an exemplary configuration of the ATSC server 10 of FIG. 1 .
- the ATSC server 10 includes an AV server 101 , a TTML server 102 , a DASH server 103 , and a broadcast server 104 .
- the AV server 101 , the TTML server 102 , and the DASH server 103 are provided by a service provider, such as a content provider or a broadcaster.
- the broadcast server 104 is provided by the service provider, such as the broadcaster or the like.
- the AV server 101 acquires and processes the video and audio (stream) data constituting the content, and provides the data to the DASH server 103 .
- the TTML server 102 generates and processes the TTML file as closed caption information superimposed on the video of the content, and provides the TTML file to the DASH server 103 .
- the DASH server 103 generates the MPD file. Furthermore, the DASH server 103 processes the video and audio data provided from the AV server 101 , and the TTML file supplied from the TTML server 102 , and generates the segment (segment data). The DASH server 103 provides the MPD file and the segment data to the broadcast server 104 .
- the broadcast server 104 generates signaling information such as the LLS information or the SLS information. Furthermore, the broadcast server 104 transmits the segment data (including the TTML file) or the signaling information (including the MPD file), as the digital broadcasting signal, through an antenna 105 .
- FIG. 20 is a diagram illustrating a detailed exemplary configuration of the ATSC server 10 of FIG. 19 .
- the ATSC server 10 includes a video data acquisition unit 111 , a video encoder 112 , an audio data acquisition unit 113 , an audio encoder 114 , a closed caption generation unit 115 , a closed caption encoder 116 , a signaling generation unit 117 , a signaling processing unit 118 , a segment processing unit 119 , a multiplexer 120 , and a transmission unit 121 .
- blocks achieving the function of the ATSC server 10 are included in any configuration of the AV server 101 to the broadcast server 104 of FIG. 19 , but for example, can have the following configuration. That is, in FIG. 20 , the video data acquisition unit 111 , the video encoder 112 , the audio data acquisition unit 113 , and the audio encoder 114 are included in a configuration of the AV server 101 .
- the closed caption generation unit 115 and the closed caption encoder 116 are included in a configuration of the TTML server 102 .
- the signaling generation unit 117 , the signaling processing unit 118 , and the segment processing unit 119 are included in a configuration of the DASH server 103 .
- the signaling generation unit 117 , the signaling processing unit 118 , the multiplexer 120 , and the transmission unit 121 are included in a configuration of the broadcast server 104 .
- a configuration illustrated in FIG. 20 is an example, the blocks constituting the servers can be arbitrarily combined for example to configure a server having both functions of the AV server 101 and the TTML server 102 , or a server having both functions of the DASH server 103 and the broadcast server 104 , and another configuration other than the configuration of FIG. 20 can be adopted.
- the video data acquisition unit 111 acquires video data of the content from an external server, camera, recording medium, or the like, and supplies the video data to the video encoder 112 .
- the video encoder 112 encodes the video data supplied from the video data acquisition unit 111 in accordance with a predetermined encoding method, and supplies the video data to the segment processing unit 119 .
- the audio data acquisition unit 113 acquires audio data of the content from an external server, microphone, recording medium, or the like, and supplies the audio data to the audio encoder 114 .
- the audio encoder 114 encodes the audio data supplied from the audio data acquisition unit 113 in accordance with a predetermined encoding method, and supplies the audio data to the segment processing unit 119 .
- the closed caption generation unit 115 generates, as closed caption data, the TTML file in TTML format, and supplies the TTML file to the closed caption encoder 116 .
- the closed caption encoder 116 encodes the closed caption data supplied from the closed caption generation unit 115 in accordance with a predetermined encoding method, and supplies the closed caption data to the segment processing unit 119 .
- the signaling generation unit 117 generates the signaling information, and supplies the signaling information to the signaling processing unit 118 .
- the signaling processing unit 118 processes the signaling information supplied from the signaling generation unit 117 , and supplies the signaling information to the multiplexer 120 .
- the LLS information such as the SLT metadata, or the SLS information such as the USD metadata, the LSID metadata, or the MPD metadata (MPD file) is generated and processed here.
- the MPD file includes the selection information for selection of a TTML processing mode.
- the segment processing unit 119 generates the segment (segment data) in accordance with the MP4 file format, based on the video data supplied from the video encoder 112 , the audio data supplied from the audio encoder 114 , and the closed caption data supplied from the closed caption encoder 116 , and supplies the segment to the multiplexer 120 .
- the multiplexer 120 multiplexes the segment data supplied from the segment processing unit 119 and the signaling information supplied from the signaling processing unit 118 , and supplies the thereby obtained multiplexed stream to the transmission unit 121 .
- the transmission unit 121 transmits the multiplexed stream supplied from the multiplexer 120 , as the digital broadcast wave (digital broadcasting signal), through the antenna 105 .
- the ATSC server 10 is configured as described above.
- FIG. 21 is a diagram illustrating a detailed exemplary configuration of an ATSC client 20 of FIG. 1 .
- the ATSC client 20 includes a reception unit 212 , a demultiplexer 213 , a control unit 214 , a memory 215 , an input unit 216 , a video decoder 217 , a video output unit 218 , an audio decoder 219 , an audio output unit 220 , a closed caption decoder 221 , a display unit 222 , and a speaker 223 .
- FIG. 21 illustrates a configuration including the display unit 222 and the speaker 223 , but a configuration not including the display unit 222 and the speaker 223 may be employed.
- the reception unit 212 extracts a signal according to user's selection operation, from the digital broadcast wave (digital broadcasting signal) received through an antenna 211 , and demodulates the signal, and then supplies the thereby obtained multiplexed stream by the demodulation to the demultiplexer 213 .
- the demultiplexer 213 separates the multiplexed stream supplied from the reception unit 212 into the audio, video, or closed caption stream, and the signaling information.
- the demultiplexer 213 supplies the video data to the video decoder 217 , the audio data to the audio decoder 219 , the closed caption data to the closed caption decoder 221 , and the signaling information to the control unit 214 .
- the video, audio, or closed caption data is defined as the segment (segment data) in accordance with the MP4 file format.
- the control unit 214 controls the operation of units of the ATSC client 20 . Furthermore, the control unit 214 controls the operation of the units based on the signaling information supplied from the demultiplexer 213 , in order to reproduce the content.
- the memory 215 is non-volatile memory such as non-volatile RAM (NVRAM), and records various data according to control by the control unit 214 .
- the input unit 216 supplies an operation signal to the control unit 214 according to user's operation.
- the video decoder 217 decodes the video data supplied from the demultiplexer 213 in accordance with a predetermined decoding method, and supplies the video data to the video output unit 218 .
- the video output unit 218 outputs the video data supplied from the video decoder 217 to the display unit 222 . Therefore, the video of the content is displayed on the display unit 222 , according to the user's selection operation.
- the audio decoder 219 decodes the audio data supplied from the demultiplexer 213 in accordance with a predetermined decoding method, and supplies the audio data to the audio output unit 220 .
- the audio output unit 220 outputs the audio data supplied from the audio decoder 219 to the speaker 223 . Therefore, sound of the content is output from the speaker 223 , according to the user's selection operation.
- the closed caption decoder 221 decodes the closed caption data supplied from the demultiplexer 213 in accordance with a predetermined decoding method, and supplies the closed caption data to the video output unit 218 .
- the video output unit 218 displays the closed caption corresponding to the closed caption data supplied from the closed caption decoder 221 to be superimposed on the video corresponding to the video data supplied from the video decoder 217 . Therefore, the closed caption superimposed on the video of the content is displayed on the display unit 222 , according to the user's selection operation.
- the closed caption decoder 221 includes an MP4 parser 241 and a TTML parser 242 .
- the MP4 parser 241 parses the segment data (TTML segment) from the demultiplexer 213 , and supplies the thereby obtained TTML file to the TTML parser 242 .
- the TTML parser 242 parses the TTML file supplied from the MP4 parser 241 , and supplies thereby obtained information for display of the closed caption to the video output unit 218 .
- FIG. 22 is a diagram illustrating an exemplary configuration of software of the ATSC client 20 of FIG. 21 .
- FIG. 22 illustrates the configuration of the ATSC client 20 illustrated in FIG. 21 , as a software configuration.
- the ATSC client 20 includes a broadcast client middleware 251 and a DASH client 252 .
- the broadcast client middleware 251 acquires various data such as the segment data (including TTML file) or the signaling information (including MPD file) transmitted from the ATSC server 10 , and performs processing for providing the various data to the DASH client 252 .
- the DASH client 252 processes various data such as the segment data (including TTML file) or the signaling information (including MPD file) transmitted from the broadcast client middleware 251 , and performs processing for reproducing the content. For example, the DASH client 252 controls the display of the closed caption specified in the TTML file, at display time according to the TTML processing mode, based on the MPD file.
- the ATSC client 20 is configured as described above.
- a component signaling process is performed.
- various processing is performed for the video and audio (stream) data processed by the AV server 101 , the closed caption data (TTML file) processed by the TTML server 102 , the signaling information (MPD file) processed by the DASH server 103 , or the signaling information (SLT metadata, USD metadata, LSID metadata, or the like) processed by the broadcast server 104 , and the data of a component or the signaling information can be transmitted.
- step S 101 Note that detailed contents of the component signaling process of step S 101 will be described later, with reference to a flowchart of FIG. 24 .
- step S 102 the transmission process is performed by the broadcast server 104 (transmission unit 121 or the like), and the data of a component, that is, the video, the audio, or the closed caption, and the signaling information processed in step S 101 are transmitted as the digital broadcasting signal, through the antenna 105 .
- the transmission process of FIG. 23 ends at step S 102 .
- step S 111 the TTML server 102 (closed caption generation unit 115 ) generates the TTML file.
- step S 112 the TTML server 102 (closed caption encoder 116 ) stores the TTML file generated in the process of step S 111 , in the MP4 file format.
- step S 113 the TTML server 102 makes a request to the DASH server 103 for generation of the segment (TTML segment) of the TTML file stored in the MP4 file format.
- the video and audio data are stored in the MP4 file format and the request for generation of the segment is made to the DASH server 103 also in the AV server 101 .
- step S 121 the DASH server 103 acquires the request for generation of the segment from the TTML server 102 (and the AV server 101 ).
- step S 122 the DASH server 103 (signaling generation unit 117 ) generates the MPD file.
- the selection information for selection of a TTML processing mode a character string for identifying mode 1, mode 2-1, mode 2-2, or mode 3 is specified for the value of the schemeIdUri attribute of the EssentialProperty element or the SupplementalProperty element of the AdaptationSet element.
- step S 123 the DASH server 103 (segment processing unit 119 ) uses the video and audio (stream) data from the AV server 101 and the TTML file from the TTML server 102 to generate the segment (segment data) in accordance with the MP4 file format.
- step S 124 the DASH server 103 transfers the MPD file generated in the process of step S 122 and the segment data generated in the process of step S 123 , to the broadcast server 104 .
- step S 131 the broadcast server 104 acquires the MPD file and the segment data from the DASH server 103 .
- step S 132 the broadcast server 104 (signaling generation unit 117 ) generates the signaling information such as the LLS information or the SLS information.
- step S 132 the process returns to step S 101 of FIG. 23 , and subsequent processing is performed. That is, in the broadcast server 104 (transmission unit 121 or the like), the above-mentioned process of step S 102 ( FIG. 23 ) is performed, and the segment data (including TTML file) or the signaling information (including MPD file) is transmitted as the digital broadcasting signal.
- the ATSC server 10 This is the end of the description of the component signaling process performed by the ATSC server 10 .
- various processing is performed so that the ATSC client 20 uses the data of a component or the signaling information to reproduce the content.
- the MPD file including the selection information for selection of a TTML processing mode is generated and transmitted with the TTML file.
- reception process performed by the ATSC client 20 of FIG. 1 will be described with reference to a flowchart of FIG. 25 .
- the reception process of FIG. 25 is performed, for example, when the user operates to select a desired service.
- step S 201 the reception process is performed by the reception unit 212 or the like, and the digital broadcasting signal transmitted from the ATSC server 10 through the transmission path 30 is received through the antenna 211 .
- step S 202 a component signaling process is performed.
- the data of a component that is, the video, the audio, or the closed caption, and the signaling information, obtained from the digital broadcasting signal received in the process of step S 201 are processed, and the content according to user's selection operation is reproduced.
- step S 202 The reception process of FIG. 25 ends at step S 202 .
- FIG. 26 illustrates processing performed by the broadcast client middleware 251 and the DASH client 252 of FIG. 22 .
- step S 211 the broadcast client middleware 251 acquires the MPD file. Furthermore, in step S 212 , the broadcast client middleware 251 acquires the segment data.
- step S 213 the broadcast client middleware 251 transfers the MPD file acquired in the process of step S 211 and the segment data acquired in the process of step S 212 , to the DASH client 252 .
- step S 221 the DASH client 252 acquires the MPD file and the segment data transferred from the broadcast client middleware 251 .
- step S 222 the DASH client 252 parses the MPD file acquired in the process of step S 221 .
- step S 223 the DASH client 252 performs rendering of the video and audio segment data acquired in the process of step S 221 , based on a result of the parsing of the MPD file in the process of step S 222 . Therefore, the video of the content is displayed on the display unit 222 , and the sound is output from the speaker 223 .
- step S 224 the DASH client 252 checks the value (attribute value) of the schemeIdUri attribute of the EssentialProperty element of the AdaptationSet element corresponding to the closed caption, based on a result of the parsing of the MPD file in the process of step S 222 .
- step S 224 when “ttmlTimeOnly” is determined to be specified as the value of the schemeIdUri attribute of the EssentialProperty element, mode 1 is set as the TTML processing mode. Thus, the process proceeds to step S 225 , and the processes of steps S 225 to S 227 are performed.
- step S 225 the DASH client 252 parses the segment data (TTML segment) acquired in the process of step S 221 .
- the time information defined in the MP4 file format that is, the time information such as BMDT or SampleDuration stored in the moof box is ignored.
- step S 226 the DASH client 252 parses the TTML segment in the process of step S 225 to parse the TTML file obtained from the TTML sample stored in the mdat box.
- mode 1 in consideration of the time information specified in the TTML file, that is, the time information specified in the begin attribute or the end attribute of the p element in the body element, display of the closed caption is started at time specified in the begin attribute, and the display of the closed caption is finished at time specified in the end attribute.
- step S 227 the DASH client 252 performs rendering based on a result of the parsing of the TTML file in the process of step S 226 , to display the closed caption as the character string specified in the p element in the body element, from the display start time specified in the begin attribute, to the display finish time specified in the end attribute.
- mode 1 when mode 1 is set as the TTML processing mode, the time information defined in the MP4 file format is ignored and the time information specified in the TTML file is considered, and thus, the closed caption can be displayed at desired time.
- step S 224 when “sampleTimeOnly” is determined to be specified as the value of the schemeIdUri attribute of the EssentialProperty element, mode 2-1 is set as the TTML processing mode. Thus, the process proceeds to step S 228 , and the processes of steps S 228 to S 230 are performed.
- step S 228 the DASH client 252 parses the segment data (TTML segment) acquired in the process of step S 221 .
- the DASH client 252 parses the segment data (TTML segment) acquired in the process of step S 221 .
- the time information defined in the MP4 file format that is, the time information such as BMDT or SampleDuration stored in the moof box
- display of the closed caption is started at the time according to BMDT, and the display is continued only during the time period according to SampleDuration stored in the moof box.
- step S 229 the DASH client 252 parses the TTML segment in the process of step S 228 to parse the TTML file obtained from the TTML sample stored in the mdat box.
- the time information specified in the TTML file that is, the time information specified in the begin attribute or the end attribute of the p element is ignored.
- step S 230 the DASH client 252 performs rendering based on a result of parsing in the processes of steps S 228 and S 229 to display the closed caption as the character string specified in the p element of the TTML file, from the time according to BMDT and during the time period according to SampleDuration.
- step S 224 when “sampleTimeOnlyButTillNext” is determined to be specified as the value of the schemeIdUri attribute of the EssentialProperty element, mode 2-2 is set as the TTML processing mode. Thus, the process proceeds to step S 231 , and the processes of steps S 231 to S 233 are performed.
- step S 231 the DASH client 252 parses the segment data (TTML segment) acquired in the process of step S 221 .
- the DASH client 252 parses the segment data (TTML segment) acquired in the process of step S 221 .
- mode 2-2 in consideration of the time information defined in the MP4 file format, that is, the time information of BMDT stored in the moof box, display of the closed caption is started at the time according to BMDT corresponding to the target TTML sample, and the display is continued until the time according to BMDT corresponding to the next TTML sample.
- step S 232 the DASH client 252 parses the TTML segment in the process of step S 231 to parse the TTML file obtained from the TTML sample stored in the mdat box.
- the time information specified in the TTML file that is, the time information specified in the begin attribute or the end attribute of the p element is ignored.
- step S 233 the DASH client 252 performs rendering based on a result of parsing in the processes of steps S 231 and S 232 to display the closed caption as the character string specified in the p element of the TTML file obtained from the target TTML sample, from the time according to BMDT corresponding to the target TTML sample, to the time according to BMDT corresponding to the next TTML sample.
- step S 231 for recognition of the time according to BMDT corresponding to the next TTML sample, the process returns to step S 231 to parse the TTML segment. That is, repeating the processes of steps S 231 to S 233 allows recognition of the time according to BMDT corresponding to the next TTML sample, and display of the closed caption specified in the TTML file obtained from the next TTML sample.
- mode 2 (mode 2-1, mode 2-2) is set as the TTML processing mode
- the time information defined in the MP4 file format time information for each TTML sample
- the time information specified in the TTML file is ignored, and thus, the closed caption can be displayed at desired time.
- step S 224 when “asap” is determined to be specified as the value of the schemeIdUri attribute of the EssentialProperty element, mode 3 is set as the TTML processing mode. Thus, the process proceeds to step S 234 , and the processes of steps S 234 to S 236 are performed.
- step S 234 the DASH client 252 parses the segment data (TTML segment) acquired in the process of step S 221 .
- the time information defined in the MP4 file format that is, the time information such as BMDT or SampleDuration stored in the moof box is ignored.
- step S 235 the DASH client 252 parses the TTML segment in the process of step S 234 to parse the TTML file obtained from the TTML sample stored in the mdat box.
- the time information specified in the TTML file that is, the time information specified in the begin attribute or the end attribute of the p element is ignored.
- step S 236 the DASH client 252 immediately performs rendering of the TTML file, based on a result of parsing of the TTML file in the process of step S 235 to display the closed caption as the character string specified in the p element of the TTML file. Note that, display of the closed caption thus displayed is finished, when the next TTML file (TTML sample) is acquired.
- mode 3 when mode 3 is set as the TTML processing mode, the closed caption is immediately displayed, ignoring the time information defined in the MP4 file format (time information for each TTML sample) and the time information specified in the TTML file, and thus, the closed caption can be displayed at desired time.
- the data of a component or the signaling information transmitted from the ATSC server 10 is used to reproduce the content. Furthermore, when the closed caption is displayed superimposed on the video of the content, the MPD file including the selection information for selection of a TTML processing mode is acquired, and thus, the closed caption specified in the TTML file is displayed at display time according to the TTML processing mode.
- the ATSC e.g., ATSC 3.0
- ISDB integrated services digital broadcasting
- DVD digital video broadcasting
- the name of the signaling information such as the SLT is only an example, and another name may be used. Even if another name is used as the name of the signaling information, this means the signaling information is merely formally changed in name, and does not influence the practical content of the signaling information.
- the SLT is often referred to as a fast information table (FIT).
- the TTML file or the MPD file is transmitted as the digital broadcasting signal by the ATSC server 10 , but the files may be distributed from a server on the Internet.
- the closed caption in the TTML file may be distributed through communication to be displayed superimposed on video of a content distributed through broadcasting.
- the video or audio (stream) data may be adaptively streamed from a server on the Internet.
- the streaming conforms to the MPEG-DASH standard.
- the attribute value of the schemeIdUri attribute of the EssentialProperty element or the SupplementalProperty element of the AdaptationSet element is used in the MPD file, for specification of the TTML processing mode, but the TTML processing mode may be specified using another element or attribute. Furthermore, in the Representation element or the SubRepresentation element, the TTML processing mode may be specified using the attribute value of the schemeIdUri attribute of the EssentialProperty element or the SupplementalProperty element. Furthermore, as long as the TTML processing mode is recognized upon processing of the TTML file, the TTML processing mode may be specified using signaling information or the like other than the MPD file.
- FIG. 27 is a block diagram illustrating an exemplary configuration of the hardware of the computer performing the above-mentioned series of processes by the programs.
- a central processing unit (CPU) 901 read only memory (ROM) 902 , and random access memory (RAM) 903 are connected to each other through a bus 904 . Further, the bus 904 is connected to an input/output interface 905 .
- the input/output interface 905 is connected to an input unit 906 , an output unit 907 , a recording unit 908 , a communication unit 909 , and a drive 910 .
- the input unit 906 includes a keyboard, a mouse, a microphone, or the like.
- the output unit 907 includes a display, a speaker, or the like.
- the recording unit 908 includes a hard disk, non-volatile memory, or the like.
- the communication unit 909 includes a network interface or the like.
- the drive 910 drives a removable medium 911 such as a magnetic disk, an optical disk, a magnetooptical disk, or a semiconductor memory.
- the CPU 901 loads the programs recorded in the ROM 902 or the recording unit 908 into the RAM 903 through the input/output interface 905 and the bus 904 , and executes the programs. Thereby, the above-mentioned series of processes is performed.
- the programs executed by the computer 900 can be provided by being recorded in, for example, the removable medium 911 as a package medium or the like. Additionally, the programs can be provided through a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
- the programs can be installed in the recording unit 908 through the input/output interface 905 , by mounting the removable medium 911 to the drive 910 . Additionally, the programs can be received at the communication unit 909 through the wired or wireless transmission medium, and installed in the recording unit 908 . The program can be previously installed in the ROM 902 or the recording unit 908 .
- the process performed by the computer according to the program is not necessarily performed chronologically along the flowchart sequence described above. That is, the process performed by the computer according to the program also includes processes performed separately or in parallel (e.g., parallel process or object-based process). Furthermore, the program may be executed by one computer (processor), or may be distributedly executed by a plurality of computers.
- the present technology may also include the following configuration.
- a reception apparatus including:
- circuitry configured to
- control information including selection information indicating a selection of a specific mode from a plurality of modes for specifying when closed caption text is to be displayed;
- the closed caption information is a timed text markup language (TTML) file
- the closed caption information includes data in accordance with an MPEG-4 (MP4) file format
- control information is a media presentation description (MPD) file in an extensible markup language (XML) format, and
- the TTML file and the MPD file are transmitted in a real-time object delivery over unidirectional transport (ROUTE) session.
- ROUTE unidirectional transport
- the plurality of modes includes a first mode for display of the closed caption text according to time information specified in the TTML file, and
- the circuitry when the specific mode is the first mode, the circuitry is configured to output the closed caption text specified in the TTML file for display according to the time information specified in the TTML file.
- the plurality of modes includes a second mode for display of the closed caption text according to time information defined in the MP4 file format, and
- the circuitry when the specific mode is the second mode, the circuitry is configured to output the closed caption text specified in the TTML file for display according to the time information defined in the MP4 file format.
- the circuitry is configured to
- the circuitry is configured to
- the plurality of modes includes a third mode for display of the closed caption text irrespective of time information defined in the MP4 file format, and
- the circuitry is configured to output the closed caption text specified in the TTML file for display immediately upon acquiring the TTML file.
- the selection information is specified as extensive information about the MPD file.
- the selection information is specified based on a scheme attribute of a property element, in adaptation set elements arranged in a period element of an MPD element.
- a data processing method including:
- control information including selection information indicating a selection of a specific mode from a plurality of modes for specifying when closed caption text is to be displayed;
- a transmission apparatus including:
- circuitry configured to
- control information including selection information indicating a selection of a specific mode from a plurality of modes for specifying when closed caption text is to be displayed;
- the closed caption information is a timed text markup language (TTML) file
- the closed caption information includes data in accordance with an MPEG-4 (MP4) file format
- control information is a media presentation description (MPD) file in an extensible markup language (XML) format, and
- the TTML file and the MPD file are transmitted in a real-time object delivery over unidirectional transport (ROUTE) session.
- ROUTE unidirectional transport
- the plurality of modes includes a first mode for display of the closed caption text according to time information specified in the TTML file.
- the plurality of modes includes a second mode for display of the closed caption text according to time information defined in the MP4 file format.
- the display of the closed caption text only continues during a time period according to a sample duration stored in the predetermined box.
- the display of the closed caption text to continue until a time according to a second media decode time stored in a third predetermined box corresponding to a fourth predetermined box that stores next closed caption data.
- the plurality of modes includes a third mode for display of the closed caption text irrespective of time information defined in the MP4 file format.
- the selection information is specified as extensive information about the MPD file.
- the selection information is specified based on a scheme attribute of a property element, in adaptation elements arranged in a period element of an MPD element.
- a data processing method including:
- control information including selection information indicating a selection of a specific mode from a plurality of modes for specifying when closed caption text is to be displayed;
- a reception apparatus including:
- a reception unit configured to receive digital broadcast waves
- an acquisition unit configured to acquire closed caption information about closed caption transmitted over the broadcast waves, and control information including selection information for selection of a specific mode from a plurality of modes for specifying display timing of the closed caption;
- control unit configured to control the display of the closed caption according to the closed caption information, at display time according to the specific mode, based on the selection information included in the control information.
- the closed caption information is a timed text markup language (TTML) file in TTML format
- the closed caption information having data in accordance with the MP4 file format
- control information is a media presentation description (MPD) file in extensible markup language (XML) format, and
- the TTML file and the MPD file are transmitted in a real-time object delivery over unidirectional transport (ROUTE) session.
- ROUTE unidirectional transport
- the plurality of modes includes a first mode for display of the closed caption at time according to time information specified in the TTML file
- control unit when the specific mode has the first mode, the control unit display the closed caption specified in the TTML file at time according to time information specified in the TTML file.
- the plurality of modes includes a second mode for display of the closed caption at time according to time information defined in the MP4 file format
- control unit when the specific mode has the second mode, the control unit display the closed caption specified in the TTML file at time according to time information defined in the MP4 file format.
- the control unit starts to display the closed caption, and continues the display only during a time period according to SampleDuration stored in the moof box.
- BMDT BaseMediaDecodeTime
- control unit starts to display the closed caption, and continues the display until the time according to BMDT stored in a moof box corresponding to an mdat box storing data about a next closed caption.
- the plurality of modes includes time information specified in the TTML file, and a third mode for display of the closed caption ignoring time information defined in the MP4 file format, and
- control unit displays the closed caption specified in the TTML file immediately upon acquiring the TTML file.
- the selection information is specified as extensive information about the MPD file.
- the selection information is specified based on a schemeIdUri attribute of an EssentialProperty element or a SupplementalProperty element, in AdaptationSet elements arranged in a Period element of an MPD element.
- a data processing method including:
- control information including selection information for selection of a specific mode from a plurality of modes for specifying display timing of the closed caption
- a transmission apparatus including:
- a generation unit configured to generate control information including selection information for selection of a specific mode from a plurality of modes for specifying display timing of a closed caption
- a transmission unit configured to transmit the control information over digital broadcast wave, together with closed caption information about the closed caption.
- the closed caption information is a TTML file in TTML format
- the closed caption information has data in accordance with the MP4 file format
- control information has an MPD file in XML format
- the TTML file and the MPD file are transmitted in a ROUTE session.
- the plurality of modes includes a first mode for display of the closed caption at time according to time information specified in the TTML file.
- the plurality of modes includes a second mode for display of the closed caption at time according to time information defined in the MP4 file format.
- the second mode is a mode for starting display of the closed caption at time according to BMDT defined in the MP4 file format and stored in a moof box
- the second mode is a mode for starting display of the closed caption at time according to BMDT defined in the MP4 file format and stored in a moof box corresponding to an mdat box storing data about a target closed caption, and continuing the display until time according to BMDT stored in a moof box corresponding to an mdat box storing data about a next closed caption.
- the plurality of modes includes time information specified in the TTML file, and a third mode for display of the closed caption ignoring time information defined in the MP4 file format.
- the selection information is specified as extensive information about the MPD file.
- the selection information is specified based on a schemeIdUri attribute of an EssentialProperty element or a SupplementalProperty element, in AdaptationSet elements arranged in a Period element of an MPD element.
- a data processing method including:
- control information including selection information for selection of a specific mode from a plurality of modes for specifying display timing of the closed caption, the control information being transmitted over digital broadcast wave, together with closed caption information about closed caption.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Security & Cryptography (AREA)
- Human Computer Interaction (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Circuits Of Receivers In General (AREA)
Abstract
Description
Claims (20)
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2015-121335 | 2015-06-16 | ||
JPJP2015-121335 | 2015-06-16 | ||
JP2015121335A JP6610019B2 (en) | 2015-06-16 | 2015-06-16 | Reception device, transmission device, and data processing method |
PCT/JP2016/002684 WO2016203726A1 (en) | 2015-06-16 | 2016-06-02 | Reception apparatus, transmission apparatus, and data processing method |
Publications (2)
Publication Number | Publication Date |
---|---|
US20180091769A1 US20180091769A1 (en) | 2018-03-29 |
US11622088B2 true US11622088B2 (en) | 2023-04-04 |
Family
ID=56137484
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/564,056 Active 2036-08-29 US11622088B2 (en) | 2015-06-16 | 2016-06-02 | Reception apparatus, transmission apparatus, and data processing method |
Country Status (7)
Country | Link |
---|---|
US (1) | US11622088B2 (en) |
EP (1) | EP3311583B1 (en) |
JP (1) | JP6610019B2 (en) |
KR (1) | KR20180018490A (en) |
CA (1) | CA2980851A1 (en) |
MX (1) | MX2017016011A (en) |
WO (1) | WO2016203726A1 (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106993227B (en) * | 2016-01-20 | 2020-01-21 | 腾讯科技(北京)有限公司 | Method and device for information display |
KR101863904B1 (en) * | 2016-11-17 | 2018-06-01 | 주식회사 에어코드 | System and method for broadcasting integrated signaling of av encoder in ground radio wave uhd service |
KR102153717B1 (en) * | 2017-12-08 | 2020-09-08 | 주식회사 에어코드 | System and method for transmitting closed caption data in terrestrial UHD broadcasting service |
KR20200135324A (en) * | 2018-03-26 | 2020-12-02 | 소니 주식회사 | Information processing devices, information processing devices and programs |
US12003825B1 (en) * | 2022-09-21 | 2024-06-04 | Amazon Technologies, Inc. | Enhanced control of video subtitles |
Citations (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS5713141B2 (en) | 1973-09-10 | 1982-03-15 | ||
JP2012169885A (en) | 2011-02-15 | 2012-09-06 | Sony Corp | Display control method, recording medium and display control unit |
WO2013088986A1 (en) | 2011-12-16 | 2013-06-20 | ソニー株式会社 | Reception device, method for controlling same, distribution device, distribution method, program, and distribution system |
US20130185760A1 (en) * | 2011-07-29 | 2013-07-18 | Sony Corporation | Streaming distributing device and method, streaming receiving device and method, streaming system, program, and recording medium |
US20130188922A1 (en) * | 2012-01-23 | 2013-07-25 | Research In Motion Limited | Multimedia File Support for Media Capture Device Position and Location Timed Metadata |
US20130291001A1 (en) * | 2012-04-25 | 2013-10-31 | Jan Besehanic | Methods and apparatus to measure exposure to streaming media |
US8695048B1 (en) * | 2012-10-15 | 2014-04-08 | Wowza Media Systems, LLC | Systems and methods of processing closed captioning for video on demand content |
US8782721B1 (en) * | 2013-04-05 | 2014-07-15 | Wowza Media Systems, LLC | Closed captions for live streams |
WO2014109321A1 (en) | 2013-01-09 | 2014-07-17 | ソニー株式会社 | Transmission device, transmission method, receiving device, and receiving method |
US20140229970A1 (en) * | 2013-02-14 | 2014-08-14 | Jan Besehanic | Methods and apparatus to measure exposure to streaming media |
US20140280888A1 (en) * | 2013-03-15 | 2014-09-18 | Francis Gavin McMillan | Methods, Apparatus and Articles of Manufacture to Monitor Media Devices |
US20140280874A1 (en) * | 2013-03-14 | 2014-09-18 | Alan N. Bosworth | Methods and apparatus to monitor media presentations |
US20150128162A1 (en) * | 2013-11-01 | 2015-05-07 | Adobe Systems Incorporated | Real-time tracking collection for video experiences |
WO2015064383A1 (en) | 2013-10-30 | 2015-05-07 | ソニー株式会社 | Transmission device, transmission method, reception device, and reception method |
JP5713141B1 (en) | 2014-12-05 | 2015-05-07 | ソニー株式会社 | Receiving apparatus and data processing method |
WO2015065104A1 (en) | 2013-11-01 | 2015-05-07 | Lg Electronics Inc. | Apparatus for transmitting broadcast signals, apparatus for receiving broadcast signals, method for transmitting broadcast signals and method for receiving broadcast signals |
US20150208102A1 (en) * | 2013-07-19 | 2015-07-23 | Sony Corporation | File generation apparatus, file generating method, file reproduction apparatus, and file reproducing method |
US20160300596A1 (en) * | 2015-04-09 | 2016-10-13 | Avid Technology, Inc. | Methods and systems for processing synchronous data tracks in a media editing system |
US20160345078A1 (en) * | 2015-05-22 | 2016-11-24 | Echostar Technologies L.L.C. | Carrier-based active text enhancement |
US20170048485A1 (en) * | 2014-05-15 | 2017-02-16 | Sony Corporation | Reception apparatus, transmission apparatus, and data processing method |
US20170223432A1 (en) * | 2014-10-14 | 2017-08-03 | Lg Electronics Inc. | Broadcasting signal transmission device, broadcasting signal reception device, broadcasting signal transmission method, and broadcasting signal reception method |
US9748504B2 (en) * | 2014-03-25 | 2017-08-29 | Universal Display Corporation | Organic electroluminescent materials and devices |
US20180035153A1 (en) * | 2015-03-01 | 2018-02-01 | Lg Electronics Inc. | Broadcasting signal transmission apparatus, broadcasting signal reception apparatus, broadcasting signal transmission method, and broadcasting signal reception method |
US20180054660A1 (en) * | 2015-04-05 | 2018-02-22 | Lg Electronics Inc. | Method and device for transmitting and receiving broadcast signal for broadcast service on basis of xml subtitle |
US20180077456A1 (en) * | 2015-06-22 | 2018-03-15 | Sony Corporation | Receiving device, transmitting device, and data processing method |
US20180205975A1 (en) * | 2015-07-06 | 2018-07-19 | Lg Electronics Inc. | Broadcast signal transmission device, broadcast signal reception device, broadcast signal transmission method, and broadcast signal reception method |
US20180249167A1 (en) * | 2015-09-04 | 2018-08-30 | Sharp Kabushiki Kaisha | Systems and methods for signaling of video parameters and information associated with caption services |
-
2015
- 2015-06-16 JP JP2015121335A patent/JP6610019B2/en not_active Expired - Fee Related
-
2016
- 2016-06-02 WO PCT/JP2016/002684 patent/WO2016203726A1/en active Application Filing
- 2016-06-02 CA CA2980851A patent/CA2980851A1/en not_active Abandoned
- 2016-06-02 US US15/564,056 patent/US11622088B2/en active Active
- 2016-06-02 EP EP16730509.3A patent/EP3311583B1/en active Active
- 2016-06-02 KR KR1020177032108A patent/KR20180018490A/en unknown
- 2016-06-02 MX MX2017016011A patent/MX2017016011A/en active IP Right Grant
Patent Citations (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS5713141B2 (en) | 1973-09-10 | 1982-03-15 | ||
JP2012169885A (en) | 2011-02-15 | 2012-09-06 | Sony Corp | Display control method, recording medium and display control unit |
US20130185760A1 (en) * | 2011-07-29 | 2013-07-18 | Sony Corporation | Streaming distributing device and method, streaming receiving device and method, streaming system, program, and recording medium |
WO2013088986A1 (en) | 2011-12-16 | 2013-06-20 | ソニー株式会社 | Reception device, method for controlling same, distribution device, distribution method, program, and distribution system |
US20130188922A1 (en) * | 2012-01-23 | 2013-07-25 | Research In Motion Limited | Multimedia File Support for Media Capture Device Position and Location Timed Metadata |
US20130291001A1 (en) * | 2012-04-25 | 2013-10-31 | Jan Besehanic | Methods and apparatus to measure exposure to streaming media |
US8695048B1 (en) * | 2012-10-15 | 2014-04-08 | Wowza Media Systems, LLC | Systems and methods of processing closed captioning for video on demand content |
WO2014109321A1 (en) | 2013-01-09 | 2014-07-17 | ソニー株式会社 | Transmission device, transmission method, receiving device, and receiving method |
US20140229970A1 (en) * | 2013-02-14 | 2014-08-14 | Jan Besehanic | Methods and apparatus to measure exposure to streaming media |
US20140280874A1 (en) * | 2013-03-14 | 2014-09-18 | Alan N. Bosworth | Methods and apparatus to monitor media presentations |
US20140280888A1 (en) * | 2013-03-15 | 2014-09-18 | Francis Gavin McMillan | Methods, Apparatus and Articles of Manufacture to Monitor Media Devices |
US8782721B1 (en) * | 2013-04-05 | 2014-07-15 | Wowza Media Systems, LLC | Closed captions for live streams |
US20150208102A1 (en) * | 2013-07-19 | 2015-07-23 | Sony Corporation | File generation apparatus, file generating method, file reproduction apparatus, and file reproducing method |
WO2015064383A1 (en) | 2013-10-30 | 2015-05-07 | ソニー株式会社 | Transmission device, transmission method, reception device, and reception method |
US20150128162A1 (en) * | 2013-11-01 | 2015-05-07 | Adobe Systems Incorporated | Real-time tracking collection for video experiences |
WO2015065104A1 (en) | 2013-11-01 | 2015-05-07 | Lg Electronics Inc. | Apparatus for transmitting broadcast signals, apparatus for receiving broadcast signals, method for transmitting broadcast signals and method for receiving broadcast signals |
US20160227263A1 (en) | 2013-11-01 | 2016-08-04 | Lg Electronics Inc. | Apparatus for transmitting broadcast signals, apparatus for receiving broadcast signals, method for transmitting broadcast signals and method for receiving broadcast signals |
US9748504B2 (en) * | 2014-03-25 | 2017-08-29 | Universal Display Corporation | Organic electroluminescent materials and devices |
US20170048485A1 (en) * | 2014-05-15 | 2017-02-16 | Sony Corporation | Reception apparatus, transmission apparatus, and data processing method |
US20170223432A1 (en) * | 2014-10-14 | 2017-08-03 | Lg Electronics Inc. | Broadcasting signal transmission device, broadcasting signal reception device, broadcasting signal transmission method, and broadcasting signal reception method |
JP5713141B1 (en) | 2014-12-05 | 2015-05-07 | ソニー株式会社 | Receiving apparatus and data processing method |
US20180035153A1 (en) * | 2015-03-01 | 2018-02-01 | Lg Electronics Inc. | Broadcasting signal transmission apparatus, broadcasting signal reception apparatus, broadcasting signal transmission method, and broadcasting signal reception method |
US20180054660A1 (en) * | 2015-04-05 | 2018-02-22 | Lg Electronics Inc. | Method and device for transmitting and receiving broadcast signal for broadcast service on basis of xml subtitle |
US20160300596A1 (en) * | 2015-04-09 | 2016-10-13 | Avid Technology, Inc. | Methods and systems for processing synchronous data tracks in a media editing system |
US20160345078A1 (en) * | 2015-05-22 | 2016-11-24 | Echostar Technologies L.L.C. | Carrier-based active text enhancement |
US20180077456A1 (en) * | 2015-06-22 | 2018-03-15 | Sony Corporation | Receiving device, transmitting device, and data processing method |
US20180205975A1 (en) * | 2015-07-06 | 2018-07-19 | Lg Electronics Inc. | Broadcast signal transmission device, broadcast signal reception device, broadcast signal transmission method, and broadcast signal reception method |
US20180249167A1 (en) * | 2015-09-04 | 2018-08-30 | Sharp Kabushiki Kaisha | Systems and methods for signaling of video parameters and information associated with caption services |
Non-Patent Citations (5)
Title |
---|
DAVID SINGER (APPLE): "Editor's study of 14496-30 and 14496-12 DAM (Timed Text in MP4)", 105. MPEG MEETING; 20130729 - 20130802; VIENNA; (MOTION PICTURE EXPERT GROUP OR ISO/IEC JTC1/SC29/WG11), 19 July 2013 (2013-07-19), XP030058631 |
David Singer, "Editor's study of 14496-30 and 14496-12 DAM (Timed Text in MP4)", 105. MPEG Meeting; Jul. 29, 2013-Feb. 8, 2013; Vienna; (Motion Picture Expert Group or ISO/IEC JTC1/SC29/WG11), No. m30103, XP030058631, sections 1, 4.1,4.2, 5.1; p. vii; Jul. 19, 2013, 23 pages. |
International Search Report dated Aug. 19, 2016 in PCT/JP2016/002684. |
Notification of Reasons(s) for Refusal dated Feb. 5, 2019 in corresponding to Japanese Patent Application No. 2015-121335 (with English translation) (8 pages). |
Office Action dated Jun. 11, 2019 in corresponding Japanese Patent Application No. 2015-121335 (with English Translation), 12 pages. |
Also Published As
Publication number | Publication date |
---|---|
MX2017016011A (en) | 2018-04-20 |
CA2980851A1 (en) | 2016-12-22 |
WO2016203726A1 (en) | 2016-12-22 |
KR20180018490A (en) | 2018-02-21 |
EP3311583B1 (en) | 2022-03-30 |
JP6610019B2 (en) | 2019-11-27 |
EP3311583A1 (en) | 2018-04-25 |
US20180091769A1 (en) | 2018-03-29 |
JP2017011325A (en) | 2017-01-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9854375B2 (en) | Selection of coded next generation audio data for transport | |
US10693936B2 (en) | Transporting coded audio data | |
JP6310111B2 (en) | Control message composition apparatus in broadcasting system | |
CN107113460B (en) | Session description information for over-the-air broadcast media data | |
JP6258856B2 (en) | Control message composition apparatus and method in broadcast system | |
US20160337424A1 (en) | Transferring media data using a websocket subprotocol | |
US11622088B2 (en) | Reception apparatus, transmission apparatus, and data processing method | |
EP2754301A1 (en) | Streaming of multimedia data from multiple sources | |
US10834153B2 (en) | System level signaling of SEI tracks for media data streaming | |
KR102499231B1 (en) | Receiving device, sending device and data processing method | |
US11388485B2 (en) | Receiving device, receiving method, transmission device, and transmission method | |
WO2015115253A1 (en) | Receiving device, reception method, transmitting device, and transmission method | |
JP2020025325A (en) | Receiver, transmitter, and data processing method | |
KR102600762B1 (en) | Apparatus and method for transmitting broadcasting content based on atsc 3.0, and apparatus and method for receiving broadcasting content based on atsc 3.0 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KITAZATO, NAOHISA;YAMAGISHI, YASUAKI;YAMANE, TAKETOSHI;SIGNING DATES FROM 20170831 TO 20170909;REEL/FRAME:044113/0690 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
AS | Assignment |
Owner name: SATURN LICENSING LLC, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONY CORPORATION;REEL/FRAME:052256/0443 Effective date: 20190911 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCV | Information on status: appeal procedure |
Free format text: NOTICE OF APPEAL FILED |
|
STCV | Information on status: appeal procedure |
Free format text: APPEAL BRIEF (OR SUPPLEMENTAL BRIEF) ENTERED AND FORWARDED TO EXAMINER |
|
STCV | Information on status: appeal procedure |
Free format text: EXAMINER'S ANSWER TO APPEAL BRIEF MAILED |
|
STCV | Information on status: appeal procedure |
Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS |
|
STCV | Information on status: appeal procedure |
Free format text: BOARD OF APPEALS DECISION RENDERED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction |