Nothing Special   »   [go: up one dir, main page]

US20210014472A1 - Methods and apparatus of viewport adaptive 360 degree video delivery - Google Patents

Methods and apparatus of viewport adaptive 360 degree video delivery Download PDF

Info

Publication number
US20210014472A1
US20210014472A1 US17/036,578 US202017036578A US2021014472A1 US 20210014472 A1 US20210014472 A1 US 20210014472A1 US 202017036578 A US202017036578 A US 202017036578A US 2021014472 A1 US2021014472 A1 US 2021014472A1
Authority
US
United States
Prior art keywords
video
viewport
representation
layer
quality
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/036,578
Inventor
Yong He
Yan Ye
Srinivas Gudumasu
Eduardo Asbun
Ahmed Hamza
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
InterDigital VC Holdings Inc
Original Assignee
Vid Scale Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vid Scale Inc filed Critical Vid Scale Inc
Priority to US17/036,578 priority Critical patent/US20210014472A1/en
Publication of US20210014472A1 publication Critical patent/US20210014472A1/en
Assigned to INTERDIGITAL VC HOLDINGS, INC. reassignment INTERDIGITAL VC HOLDINGS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: VID SCALE, INC.
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/282Image signal generators for generating image signals corresponding to three or more geometrical viewpoints, e.g. multi-view systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/111Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
    • H04N13/117Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/349Multi-view displays for displaying three or more geometrical viewpoints without viewer tracking
    • H04N13/354Multi-view displays for displaying three or more geometrical viewpoints without viewer tracking for displaying sequentially
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/39Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability involving multiple description coding [MDC], i.e. with separate layers being structured as independently decodable descriptions of input picture data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234327Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by decomposing into layers, e.g. base layer and one or more enhancement layers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234345Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements the reformatting operation being performed only on part of the stream, e.g. a region of the image or a time segment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/23439Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements for generating different versions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/4728End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for selecting a Region Of Interest [ROI], e.g. for requesting a higher resolution version of a selected region
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8543Content authoring using a description language, e.g. Multimedia and Hypermedia information coding Expert Group [MHEG], eXtensible Markup Language [XML]

Definitions

  • 360° video is a rapidly growing format emerging in the media industry.
  • 360° video is enabled by the growing availability of virtual reality (VR) devices.
  • VR virtual reality
  • 360° video may provide the viewer a new sense of presence.
  • 360° video may pose difficult engineering challenges on video processing and/or delivery. Enabling comfort and/or an immersive user experience may require high video quality and/or very low latency.
  • the large video size of 360° video may be an impediment to delivering the 360° video in a quality manner at scale.
  • 360° video applications and/or services may encode the entire 360° video into a standard compliant stream for progressive downloading and/or adaptive streaming.
  • Delivering the entire 360° video to the clients may enable low-latency rendering (e.g., the client may have access to the entirety of the 360° video content and/or can choose to render the portion it desires to see without further constraint).
  • the same streams can support multiple users with possibly different viewports.
  • the video size may be tremendously high, incurring high transmission bandwidth when the video is delivered (e.g., because the entire 360° video has to be encoded in high quality, such as 4K@60 fps or 6K@90 fps per eye).
  • the high bandwidth consumption during delivery may be wasted, for example, because the user may only view a small portion (e.g., a viewport) of the entire picture.
  • Viewport enhancement based 360 degree video may be delivered and/or signaled.
  • 360 degree video may be delivered using a layer-based viewport overlay. Signaling of 360 degree video mapping may be provided.
  • a first viewport of a 360 degree video may be determined.
  • the 360 degree video may comprise one or more of an equirectangular, a cube-map, a cylindrical, a pyramidal, and/or a spherical projection mapping.
  • the first viewport may be associated with a spatial region of the 360 degree video. An adjacent area that extends around the spatial region may be determined.
  • a second viewport of the 360 degree video may be determined.
  • a bitstream associated with the 360 degree video may be received.
  • the bitstream may comprise one or more enhanced regions.
  • the one or more enhanced regions may correspond to the first and/or second viewport.
  • a high coding bitrate may be associated with the first viewport and/or the second viewport. Signaling, that indicates one or more viewport properties associated with the 360 degree video delivery, may be received.
  • a WTRU, for processing a 360 degree video may include a processor configured, (e.g., with executable instructions saved in memory), for one or more of the following: (i) receiving a media presentation description (MPD) associated with a 360 degree video comprising an essential property element that indicates a face-packing layout for a multi-face geometric projection format of a media segment; (ii) receiving the media segment; (iv) determining, based on the essential property element, at least one face-packing layout, from a set of face-packing layouts, for the received media segment; and (iv) constructing the received media segment based on the determined at least one face-packing layout.
  • MPD media presentation description
  • the set of face-packing layouts comprise plate carree, poles on the side half height, poles on the side full height, single row, two by three, and one hundred and eighty degrees.
  • the essential property element may be at one of an adaptation level and a representation level.
  • a WTRU processor may be configured, (e.g., with executable instructions saved in memory), to determine a media representation, that is associated with the MPD, to request for future media segments and to send a request for the determined media representation.
  • the MPD may include a video type, selected from a set of video types, for the media segment.
  • the set of video types may include comprise rectilinear, panorama, sphere, and lightfield formats.
  • the WTRU processor may be configured, (e.g., with executable instructions saved in memory), to determine the video type for the received media segment and/or to construct the received media segment using the determined video type.
  • the MPD may include at least one projection format used for projecting the 360 degree video from an omnidirectional format to a rectilinear video frame.
  • the projection format may include one of a equirectangular, cube, an offset cube, a squished sphere, a pyramid, and a cylinder.
  • the WTRU processor may be configured, (e.g., with executable instructions saved in memory), to determine a projection format for receiving a video file and/or to send a request for the determined projection format.
  • the 360 degree video may include an Omnidirectional Media Application Format (OMAF) file.
  • OMAF Omnidirectional Media Application Format
  • a method of using a WTRU for processing a 360 degree video may include one or more of: (i) receiving a media presentation description (MPD) associated with a 360 degree video comprising an essential property element that indicates a face-packing layout for a multi-face geometric projection format of a media segment; (ii) receiving the media segment; (iii) determining, based on the essential property element at least one face-packing layout, from a set of face-packing layouts, for the received media segment; and (iv) constructing the received media segment based on the determined at least one face-packing layout.
  • MPD media presentation description
  • the method of using a WTRU may include determining a media representation, that is associated with the MPD, to request for future received vide files and/or sending a request for the determined media representation.
  • the method may include determining the video type for the received media segment and/or constructing the received media segment using the determined video type.
  • the method may include determining a projection format for a video file and/or sending a request for the determined projection format.
  • a WTRU for processing a 360 degree video file my include a processor configured (e.g., with executable instructions saved in memory) for one or more of: receiving a media presentation description (MPD), associated with a 360 degree video, that includes a first essential property element that, indicates a first face-packing layout, for a multi-face geometric projection format of a media segment, and a second essential property element that, indicates a first second-packing layout, for a multi-face geometric projection format of a media segment; determine whether to use the first or the second face-packing layout for the media segment; request the determined at least first or the second face-packing layout; receive the media segment; and reconstruct the 360 degree video associated with the received media segment based on the requested face-packing layout.
  • MPD media presentation description
  • FIG. 1 depicts an example portion of 360° video displayed on a head mounted device (HMD).
  • HMD head mounted device
  • FIG. 2 depicts an example equirectangular projection for a 360° video.
  • FIG. 3 depicts example 360° video mappings.
  • FIG. 4 depicts an example media presentation description (MPD) hierarchical data model.
  • MPD media presentation description
  • FIG. 5 depicts an example Dynamic Adaptive Streaming over HTTP (DASH) spatial relationship description (SRD) for a video.
  • DASH Dynamic Adaptive Streaming over HTTP
  • SRD spatial relationship description
  • FIG. 6 depicts an example tile based video partitioning.
  • FIG. 7 depicts an example temporal motion constrained tile set.
  • FIG. 8 depicts an example 360° video streaming quality degradation.
  • FIG. 9 depicts an example viewport area with an associated adjacent area.
  • FIGS. 10A-10D depict example cube-map layouts.
  • FIG. 11A depicts an example equirectangular coordinate viewport mapping.
  • FIG. 11B depicts an example cube-map coordinate viewport mapping.
  • FIG. 12 depicts an example spherical coordinate viewport mapping.
  • FIGS. 13A and 13B depict an example viewport enhancement representation.
  • FIG. 14 depicts an example layer-based 360° video overlay.
  • FIG. 15 depicts an example layer-based 360° video representation.
  • FIG. 16 depicts an example layer-based 360° video overlay.
  • FIG. 17 depicts a flowchart of an example layer-based 360° video overlay.
  • FIG. 18 depicts an example layer-based 360° video overlay with multiple viewports.
  • FIG. 19 depicts an example equirectangular representation with half height poles on the side.
  • FIG. 20 depicts an example equirectangular representation with full height poles on the side.
  • FIG. 21 depicts an example single row layout cube representation.
  • FIG. 22 depicts an example 2 ⁇ 3 layout cube representation.
  • FIG. 23 depicts an example 180° cube-map layout.
  • FIG. 24A is a system diagram of an example communications system in which one or more disclosed embodiments may be implemented.
  • FIG. 24B is a system diagram of an example wireless transmit/receive unit (WTRU) that may be used within the communications system illustrated in FIG. 24A .
  • WTRU wireless transmit/receive unit
  • FIG. 24C is a system diagram of an example radio access network and an example core network that may be used within the communications system illustrated in FIG. 24A .
  • FIG. 24D is a system diagram of another example radio access network and an example core network that may be used within the communications system illustrated in FIG. 24A .
  • FIG. 24E is a system diagram of another example radio access network and an example core network that may be used within the communications system illustrated in FIG. 24A .
  • FIG. 1 depicts example portion of 360° video displayed on a head mounted device (HMD).
  • HMD head mounted device
  • a user may be presented with a part of the video, for example, as shown in FIG. 1 .
  • the part of the video may be changed when the user looks around and/or zooms the image.
  • the part of the video may be changed based on feedback provided by the HMD and/or other types of user interface (e.g., a wireless transmit/receive unit (WTRU)).
  • WTRU wireless transmit/receive unit
  • a viewport may be or may include a spatial region of the entire 360° video.
  • the viewport may be fully or partially presented to the user.
  • the viewport may have one or more different qualities than other parts of the 360° video.
  • a 360° video may be captured and/or rendered on a sphere (e.g., to give the user the ability to choose an arbitrary viewport).
  • a spherical video format may not be delivered directly using conventional video codecs.
  • a 360° video (e.g., such as a spherical video) may be compressed by projecting spherical video onto a 2D plane using a projection method.
  • the projected 2D video may be coded (e.g., using conventional video codecs).
  • An example of the projection method may include an equirectangular projection.
  • FIG. 2 depicts an example equirectangular projection for a 360° video.
  • an equirectangular projection method may use one or more of the following equations to map a first point P with coordinate ( ⁇ , ⁇ ) on a sphere to a second point P with coordinate (u, v) on a 2D plane,
  • FIG. 3 depicts example 360° video mappings.
  • one or more other projection methods e.g., mappings
  • the one or more other projection methods may include a pyramid-map, a cube-map, and/or an offset cube-map.
  • the one or more other projection methods may be used to represent the spherical video with less data.
  • a viewport-specific representation may be used.
  • One or more of the projection methods shown in FIG. 3 may provide uneven quality representations for different viewports (e.g., some viewports may be represented in a higher quality than other view/ports).
  • Multiple versions of the same video with different target viewports may be generated and/or stored at the server side (e.g., to support all viewports of the spherical video).
  • the offset cube-map format shown in FIG. 3 may be used.
  • the offset cube-map may provide a highest resolution (e.g., highest quality) to the front viewport, a lowest resolution (e.g., lowest quality) to the back view, and an intermediate resolution (e.g., intermediate quality) to one or more side views.
  • a server may store multiple versions of the same content (e.g., to accommodate a client request for different viewports of the same content). For example, a total of 150 different versions (e.g., 30 viewports times 5 resolutions for each viewport) of the same content.
  • delivery for example, streaming
  • the client may request a specific version that corresponds to its current viewport.
  • the specific version may be delivered by the server.
  • ISO/IEC/MPEG may define an Omnidirectional Media Application Format (OMAF).
  • OMAF file format metadata for the projection methods described herein may include support for projection metadata for video onto sphere, cube, cylinder, pyramid, and/or the like.
  • Table 1 may show a syntax example of OMAF to support projection methods such as squished sphere, cylinder, and pyramid.
  • HTTP streaming has become a dominant approach in commercial deployments.
  • streaming platforms such as Apple's HTTP Live Streaming (HLS), Microsoft's Smooth Streaming (SS), and/or Adobe's HTTP Dynamic Streaming (HDS) may use HTTP streaming as an underlying delivery method.
  • a standard for HTTP streaming of multimedia content may enable a standard-based client to stream content from any standard-based server (e.g., thereby enabling interoperability between servers and clients of different vendors).
  • MPEG Dynamic Adaptive Streaming over HTTP MPEG-DASH
  • MPEG-DASH may be a universal delivery format that provides end users with the best possible video experience by dynamically adapting to changing network conditions.
  • DASH may be built on top of the HTTP/TCP/IP stack.
  • DASH may define a manifest format, Media Presentation Description (MPD), and segment formats for ISO Base Media File Format and MPEG-2 Transport Streams.
  • Dynamic HTTP streaming may be associated with various bitrate alternatives of the multimedia content to be available at the server.
  • the multimedia content may include several media components (e.g., audio, video, text), each of which may have different characteristics.
  • the characteristics may be described by a Media Presentation Description (MPD).
  • MPD Media Presentation Description
  • the MPD may be an XML document that includes metadata necessary for a DASH Client to construct appropriate HTTP-URLs to access video segments (e.g., as described herein) in an adaptive manner during streaming sessions.
  • FIG. 4 depicts an example media presentation description (MPD) hierarchical data model.
  • the MPD may describe a sequence of Periods, where a consistent set of encoded versions of the media content components does not change during a Period.
  • a Period may have a starting time and duration.
  • a Period may be composed of one or more adaptation sets (e.g., AdaptationSet).
  • An AdaptationSet may represent a set of encoded versions of one or several media content components sharing one or more identical properties (e.g., such as the language, the media type, the picture aspect ratio, the role, the accessibility, the viewpoint, and/or the rating property).
  • a first AdaptationSet may include different bitrates of the video component of the same multimedia content.
  • a second AdaptationSet may include different bitrates of the audio component (e.g., lower quality stereo and/or higher quality surround sound) of the same multimedia content.
  • An AdaptationSet may include multiple Representations.
  • a Representation may describe a deliverable encoded version of one or several media components, varying from other representations by bitrate, resolution, number of channels, and/or other characteristics.
  • a representation may include one or more segments.
  • One or more attributes of a Representation element e.g., such as @id, @bandwidth, @qualityRanking, and @dependencyId
  • a Segment may be the largest unit of data that can be retrieved with a single HTTP request.
  • a segment may have a URL (e.g., an addressable location on a server).
  • a segment may be downloaded using HTTP GET or HTTP GET with byte ranges.
  • a DASH client may parse a MPD XML document.
  • the DASH client may select a collection of AdaptationSets suitable for its environment, for example, based on information provided in the AdaptationSet elements.
  • the client may select a Representation.
  • the client may select the Representation based on the value of @bandwidth attribute, client decoding capabilities, and/or client rendering capabilities.
  • the client may download an initialization segment of the selected Representation.
  • the client may access content (e.g., by requesting entire Segments or byte ranges of Segments).
  • the client may continue consuming the media content. For example, the client may request (e.g., continuously request) Media Segments and/or parts of Media Segments during the presentation.
  • the client may play content according to a media presentation timeline.
  • the client may switch from a first Representation to a second representation, based on updated information from the client's environment.
  • the client may play the content continuously across two or more Periods.
  • the Media Presentation may be terminated, a Period may be started, and/or the MPD may be re-fetched.
  • the MPD descriptor element, Descriptor may be provided to the application (e.g., to instantiate one or more description elements with the appropriate scheme information).
  • One or more Descriptors e.g., such as content protection, role, accessibility, rating, viewpoint, frame packing, and/or UTC timing descriptor
  • a supplemental property descriptor may include metadata that may be used by the DASH client for optimizing processing.
  • An essential property descriptor may include metadata for processing the containing element.
  • a Role MPD element may use a @schemeIdUri attribute to identify the role scheme employed to identify the role of the media content component.
  • One or more Roles may define and/or describe one or more characteristics and/or structural functions of media content components.
  • An Adaptation Set and/or a media content component may have multiple assigned roles (e.g., even within the same scheme).
  • MPEG-DASH may provide a spatial relationship description (SRD) scheme.
  • the SRD scheme may express a spatial relationship of a video representing a spatial part of another full-frame video in two MPD elements (e.g., AdaptationSet and SubRepresentation).
  • the SupplementalProperty and/or EssentialProperty descriptors with @schemeIdURI equal to “urn:mpeg:dash:srd:2014” may be used to provide spatial relationship information associated with the AdaptationSet and/or SubRepresentation.
  • An attribute @value of the SupplementalProperty and/or EssentialProperty elements may provide one or more values for SRD parameters such as source_id, object_x, object_y, object_width, object_height, total_width, total_height, and/or spatial_set_id.
  • SRD parameters such as source_id, object_x, object_y, object_width, object_height, total_width, total_height, and/or spatial_set_id.
  • the value and the semantics of the SRD parameters may be defined as shown in Table 2.
  • source_id a non-negative integer in decimal representation that may provide the identifier for the source of the content object_x a non-negative integer in decimal representation that may express the horizontal position of the top-left corner of the Spatial Object in arbitrary units object_y a non-negative integer in decimal representation that may express the vertical position of the top-left corner of the Spatial Object in arbitrary units object_width a non-negative integer in decimal representation that may express the width of the Spatial Object in arbitrary units object_height a non-negative integer in decimal representation that may express the height of the Spatial Object in arbitrary units total_width a non-negative integer in decimal representation that may express the width of the reference space in arbitrary units.
  • At least one descriptor may provide a value for the total_width parameter. If two or more descriptors provide different total_width values, all other descriptors may explicitly provide the value of total_width. If the total width value is provided in only one descriptor, all other descriptors may be assumed to use the total_width value.
  • the value of total_width may be defined such that, for a descriptor using the value of total_width, the sum of object_x and object_width is smaller or equal to the total_width. When the value total_width is present, the value total_height may be present.
  • total_height a non-negative integer in decimal representation that may express the height of the reference space in arbitrary units.
  • At least one descriptor may provide a value for the total_height parameter. If two or more descriptors provide different total_height values, all other descriptors may explicitly provide the value of total_height. If the total_height value is provided in only one descriptor, all other descriptors may be assumed to use the total_height value.
  • the value of total height may be defined such that, for each descriptor using the value of total_height, the sum of object_y and object_height is smaller or equal to the total_height.
  • spatial_set_id a non-negative integer in decimal representation that may provide an identifier for a group of Spatial Objects.
  • the Spatial Object associated to this descriptor may not belong to any spatial set and/or no spatial set information may be given.
  • the value of spatial set_id the value of total_width and/or total_height may be present.
  • FIG. 5 depicts an example DASH SRD video.
  • the SRD may express that a video stream represents a spatial part of a full-frame video.
  • the spatial part may be a tile and/or a region of interest (ROI) of the full-frame video.
  • An SRD may describe the video stream in terms of the spatial part's position (object_x, object_y) and/or size (object_width, object_height) relative to the full-frame video (total_width, total_height).
  • the SRD description may provide flexibility for the client in terms of adaptation.
  • An SRD-aware DASH client may select a full-frame representation and/or a spatial part of the full-frame representation using one or more SRD annotations.
  • Using the one or more SRD annotations to select the full-frame representation or a spatial part may save bandwidth and/or client-side computations, e.g., avoiding full-frame fetching, decoding, and/or cropping.
  • Using the one or more SRD annotations to determine which representation to select may increase the quality of a given spatial part (e.g., the Region of Interest or ROI) of the full-frame video, for example, after a zoom.
  • the client may request a first video stream that corresponds to the ROI spatial part with higher quality and the client may request a second video stream that does not correspond to the ROI with lower quality, without increasing the overall bitrate.
  • Table 3 is an MPD example supporting SRD for the scenario as shown in FIG. 5 , where each tile has a resolution of 1920 ⁇ 1080 and the whole frame has a resolution of 5760 ⁇ 3240 with 9 tiles.
  • FIG. 6 depicts an example tile based video partitioning.
  • a 2D frame may be partitioned into one or more tiles, as shown in FIG. 6 .
  • tile based adaptive streaming TAS
  • An SRD aware DASH client may use one or more SRD annotations to select a full-frame representation or a tile representation. Using the one or more SRD annotations to determine whether to select the full-frame representation of the tile representation may save bandwidth and/or client-side computations (e.g., avoiding full-frame fetching, decoding, and/or cropping).
  • FIG. 7 depicts an example temporal motion-constrained tile set.
  • a frame may be encoded into a bitstream that includes a number of temporal motion-constrained tile sets as specified in HEVC.
  • Each of the temporal motion-constrained tile sets may be decoded independently.
  • two or more left tiles may form a motion-constrained tile set that can be decoded independently (e.g., without decoding the full picture).
  • 360° video content may be delivered via progressive download or DASH adaptive streaming, which are primarily HTTP based streaming solutions.
  • Dynamic streaming for 360° video based on UDP instead of HTTP has been proposed (e.g., by Facebook) to reduce latency.
  • mappings and/or projections may be used to reduce the amount of data by allowing different portions of the 360 video to be represented in different qualities. For example, a front view (e.g., viewport) may be represented in a high quality and a back (e.g., opposite) view may be represented in a low quality. One or more other views may be represented in one or more intermediate qualities.
  • the offset cube-map and the pyramidal map, shown in FIG. 3 may be examples of mappings and/or projections that represent different portions of the 360 video in different qualities.
  • the pyramidal map may reduce the number of pixels and/or save bitrate for each viewport, multiple viewport versions may handle different viewing positions that the client may request.
  • the viewport-specific representation and/or delivery may have higher latency adapting when the user changes viewing positions. For example, using the offset cube-map representation, the video quality may degrade when the user's head position rotates 180 degrees.
  • FIG. 8 depicts an example 360 video streaming quality degradation.
  • DASH segment length and/or client buffer size may affect viewing quality. Longer segment lengths may result in higher coding efficiency. The longer segment lengths may not adapt to viewport changes as quickly.
  • a user may have three possible viewports of a 360 video (e.g., A, B and C).
  • One or more (e.g., 3) segment types, S A , S B and S C may be associated with the 360 video.
  • Each of the one or more segment types may carry a higher quality of the corresponding viewport and a lower quality of the other viewports.
  • the user may pan from viewport A to viewport B at time t 1 during the playback of segment S A that carries a higher quality video for viewport A and a lower quality video for viewports B and C.
  • the user may have to watch lower quality viewport B before switching to the next segment (S B ), which carries higher quality video for viewport B and lower quality video for viewports A and C.
  • Such negative user experience may be resolved with shorter segment lengths. Shorter segment lengths may reduce coding efficiency.
  • the user may have to watch lower quality video when the user's streaming client logic pre-downloads too many segments based on a previous viewport. To prevent the user's streaming client from pre-downloading too many segments based on the previous viewport, the streaming buffer size may be reduced. A smaller streaming buffer size may affect streaming quality adaptation, for example, by causing more frequent buffer underflow.
  • Viewport adaptive 360° video streaming may include viewport enhancement based delivery and/or layer based delivery.
  • Viewport enhancement based 360° video delivery may include encoding one or more identified viewports of a 360 video frame in a high quality. For example, a viewport-based bit allocation may be performed during encoding. The viewport-based bit allocation may assign a bigger portion of the bits to the one or more viewports and/or assign a correspondingly reduced amount of bits to the other areas.
  • FIG. 9 depicts an example viewport area with an associated adjacent area. A viewport area, an adjacent area, and other area may be determined for a 360° video frame.
  • a bitrate weight for the viewport area may be defined as a.
  • the bitrate weight for the adjacent area may be defined as ⁇ .
  • the bitrate weight for the other area may be defined as ⁇ .
  • One or more of the following equations may be used to determine the target bitrate for each area.
  • R may represent a constant coding bitrate for the entire 360° video
  • BR HQ may represent a coding bitrate for the target viewports area
  • BR MQ may represent a coding bitrate for the viewport adjacent area
  • BR LQ may represent the coding bitrate for the other area.
  • the values of ⁇ , ⁇ , and ⁇ may add up to 1, which may mean that the overall bitrate is kept the same. For example, bits may only be redistributed among the different areas (e.g., viewport, adjacent areas, and other areas).
  • the same video may be encoded into different versions. Each of the different versions may be associated with a different viewport quality level. Each of the different viewport quality levels may correspond to a different value of ⁇ .
  • the overall bitrate R may not be kept the same.
  • the viewport area, the adjacent area, and the other area may be encoded to a target quality level.
  • the representation bitrate for each of the areas may be different.
  • the projection methods described herein may enhance the quality of one or more target viewports and/or reduce the quality of other areas of the video.
  • an AdaptationSet may include multiple Representation elements.
  • a Representation of a 360 video stream may be coded in a specific resolution and/or at a specific bitrate.
  • a Representation may be associated with one or more specific quality enhanced viewport(s).
  • a Viewport Relationship Description may specify one or more corresponding viewport spatial coordinate relationships.
  • the SupplementalProperty and/or EssentialProperty descriptors with @schemeIdUri equal to “urn:mpeg:dash:viewport:2d:2016” may be used to provide a VRD associated with the AdaptationSet, Representation, and/or Sub-Representation elements.
  • the @value of the SupplementalProperty and/or EssentialProperty elements using the VRD scheme may be a comma separated list of values for viewport description parameters.
  • Each AdaptationSet, Representation, and/or Sub-Representation may include one or more VRDs to represent one or more enhanced viewports.
  • the properties of an enhanced viewport may be described by the VRD parameters as shown in Table 4.
  • the value may be inferred to be the same as the value of viewport_width.
  • full_height A non-negative integer in decimal representation that may express the height of the entire 360 video. If not present, the value may be inferred to be the same as the value of viewport_height.
  • Table 5 provides a MPD example of a 4 k 360 video.
  • the AdaptationSet may be annotated with two SupplementalProperty descriptors with VRD scheme identifier “urn:mpeg:dash:viewport:2d:2016”.
  • a first descriptor may specify an enhanced 320 ⁇ 640 viewport # 1 at (150,150) with quality level 3.
  • a second descriptor may specify 640 ⁇ 960 viewport # 2 at (1000, 1000) with quality level 5. Both viewports may represent spatial parts of a 4096 ⁇ 2048 full-frame 360° video.
  • the viewport position and/or size of a Representation may be identified based on the values of VRD attributes @full_width and/or @full_height and/or the value of Representation attributes @width and/or @height.
  • the half resolution video may be scaled up to the full resolution.
  • a projection may be used to map the surface of a sphere video to a flat image for processing.
  • One or more layouts may be available for a particular projection. For example, for the equirectangular and/or cube-map projection format, different layouts may be used.
  • FIGS. 10A-10D depict example cube-map layouts.
  • the cube-map layouts may include a cube layout, a 2 ⁇ 3 layout, a poles on the side layout, and/or a single row layout.
  • the VRD in Table 4 may be specified for a particular projection format, for example, equirectangular projection.
  • the VRD in Table 4 may be specified for a particular projection and layout combination, for example, cube-map project and a 2 ⁇ 3 layout (e.g., layout B shown in FIG. 10B ).
  • the VRD in Table 4 may be extended to support a variety of projection formats and/or projection and layout combination formats at the same time, as shown in Table 6.
  • the server may support a set of popular projection and/or projection and layout formats, for example, using the signalling syntax shown in Table 6.
  • FIG. 11 depicts example viewport coordinates in equirectangular and cube-map.
  • One or more (e.g., two) viewports of the 360 video may be identified. As shown, the coordinate values of a viewport may be different in equirectangular and cube-map projection formats.
  • the VRD may specify the one or more viewport(s) associated with the Representation element.
  • Table 7 shows an MPD example where both equirectangular and cube-map Representations are provided and the corresponding VRD is signalled accordingly.
  • may specify one or more properties of a viewport in one or more first Representations (e.g., Representation@id 0,1) that are in an equirectangular projection format.
  • the VRD may specify one or more viewport(s) coordinates in one or more (e.g., all) common projection/layout formats (e.g., even though the associated Representation is in one projection/layout format).
  • Table 8 shows an MPD example where the Representations of equirectangular projection are provided and the VRD of corresponding viewports for both equirectangular and cube-map are provided.
  • the projection and layout formats may be signalled at a Representation level as described herein.
  • the server may specify the viewports in different formats and/or may give the client the flexibility to choose an appropriate viewport (e.g., depending on the client's capability and/or technical specification). If the client cannot find a preferred format (e.g., either from the method specified in Table 4 or from the set of methods specified in Table), the client may convert one or more viewports from one of the specified formats into a format that it wants to use. For example, the one or more viewports may specified in the cube-map format but the client wants to use the equirectangular format. Based on the projection and/or layout description available in MPD, the client may derive a user orientation position from its gyroscope, accelerometer, and/or magnetometer tracking information.
  • a preferred format e.g., either from the method specified in Table 4 or from the set of methods specified in Table
  • the client may convert one or more viewports from one of the specified formats into a format that it wants to use. For example, the one or more viewports may specified in the cu
  • the client may convert the orientation position into a corresponding 2D position on a specific projection layout.
  • the client may request the Representation with the identified viewport based on the value of one or more VRD parameters such as @viewport_x, @viewport_y, @viewport_width, and/or @viewport_height.
  • viewport_id A non-negative integer in decimal representation that may provide an identifier for an identified viewport viewport_x A non-negative integer in decimal representation that may express the horizontal position of the top-left corner of the identified viewport viewport_y A non-negative integer in decimal representation that may express the vertical position of the top-left corner of the identified viewport viewport_width A non-negative integer in decimal representation that may express the width of the identified viewport viewport_height A non-negative integer in decimal representation that may express the height of the identified viewport full_width A non-negative integer in decimal representation that may express the width of the entire 360 video.
  • the value may be inferred to be the same as the value of viewport_width.
  • full_height A non-negative integer in decimal representation that may express the height of the entire 360 video. If not present, the value may be inferred to be the same as the value of viewport_height.
  • layout May specify the layout format in the form of a string.
  • a general viewport descriptor as shown in Table 9, may be provided.
  • the general viewport descriptor may describe the viewport positions using spherical coordinates ( ⁇ , ⁇ ) as shown in FIG. 2 , where ⁇ may represent the inclination or polar angle, ⁇ may represent the azimuthal angle, and/or a normalized radius may be 1.
  • SupplementalProperty and/or EssentialProperty descriptors with @schemeIdUri equal to “urn:mpeg:dash:viewport:sphere:2016” may be used to provide spherical coordinate based VRD associated with the AdaptationSet, Representation, and/or Sub-Representation elements.
  • a region specified with the spherical coordinates may not correspond to a rectangular area on the 2D plane after projection. If the region does not correspond to a rectangular area on the 2D plane after projection, a bounding rectangle of the signalled region may be derived and/or used to specify the viewport.
  • the @value of the SupplementalProperty and/or EssentialProperty elements using the VRD may be a comma separated list of values for one or more viewport description parameters.
  • Each AdaptationSet, Representation and/or Sub-Representation may include one or several VRD to represent one or more enhanced viewports.
  • the properties of an enhanced viewport may be described by the parameters as shown in Table 9.
  • FIG. 12 depicts an example spherical coordinate viewport.
  • viewport_inc may specify a polar angle ⁇
  • viewport_az may specify the azimuthal angle ( ⁇ )
  • viewport_delta_inc may specify d ⁇
  • viewport_delta_az may specify d ⁇ .
  • @schemeIdUri values may be used in each case.
  • the @schemeIdUri value may be urn:mpeg:dash:viewport:2d:2016.
  • the @schemeIdUri value may be “urn:mpeg:dash:viewport:sphere:2016”.
  • Using a viewport descriptor that is based on the spherical coordinates may reduce signaling cost since viewports may only be specified for one coordinate system (e.g., the spherical coordinate system).
  • Using a viewport descriptor that is based on the spherical coordinates may result in a simplified conversion process on the client side since each client may only need to implement a predefined conversion process. For example, each client may convert between the projection format the client chooses to use (e.g., equirectangular) and the spherical representation. When the client aligns the viewport coordinates, the client may use similar logic to decide which representation to request.
  • the VRD may be signaled in one or more Period.SupplementalProperty elements.
  • the VRD may list available (e.g., all available) enhanced viewports respectively.
  • a Representation may use an attribute, @viewportId, to signal one or more viewport indices to identify which enhanced viewport(s) are associated with the current Representation or Sub-Representation. With such a descriptor reference approach, redundant signaling of VRDs within a Representation may be avoided.
  • One or more Representations with different associated viewport(s), projection formats, and/or layout formats may be allocated within a single AdaptationSet. Table 10 shows example semantics of Representation element attribute @viewportId and @viewport_quality.
  • the attribute @viewport_quality may be signaled at a Representation level (e.g., instead of at the Period level) as part of the VRD.
  • the @viewport_quality attribute may allow a client to select an appropriate quality level for one or more viewports of interest. For example, if the user is navigating through the 360 view frequently (e.g., turning head to look around all the time), the client may select a Representation with balanced quality between viewport and non-viewport. If the user is focusing on the viewport, the client may select a Representation with a high viewport quality (e.g., but relatively reduced quality in non-viewport areas).
  • the @viewport_quality attribute may be signaled at the Representation level.
  • @viewportId May specify one or more (e.g., all) identified viewports associated with the element as a whitespace-separated list of values of ViewportDescriptor@viewport_id attributes.
  • @viewport_quality May specify the quality level of one or more corresponding viewports specified by @viewportId as a whitespace-separated list of values.
  • Table 11 is an MPD example using the Representation attributes, @viewportId and @viewport_quality, shown in Table 10 to specify the associated viewports.
  • Two VRDs may be specified in Period.SupplementalProperty.
  • viewport # 0 quality level may be 2.
  • An enhanced viewport area may be selected to be larger than an actual viewing resolution, which may reduce quality degradation when the user changes (e.g., slightly changes) viewports.
  • the enhanced viewport area may be selected to cover an area that a target object of interest moves around during a segment period (e.g., such that the user can watch the target object as the target moves around at the same high quality).
  • One or more most watched viewports may be enhanced in a single bitstream, for example, to reduce the total number of Representations and/or corresponding media streams at the origin server or CDN.
  • the viewport descriptors specified in Table 4, Table 6, Table 9, and/or Table 10 may support multiple quality-enhanced viewports within a single bitstream.
  • FIG. 13 depicts an example viewport enhancement representation.
  • a 360° video may have two identified most viewed viewports (e.g., viewport # 1 and viewport # 2 ).
  • One or more DASH Representations may correspond to the two identified most viewed viewports.
  • both viewports may be enhanced at high quality.
  • one Representation may be provided with both enhanced viewports (e.g., to facilitate fast viewport changes and/or save the storage cost).
  • three Representations may be provided with the quality enhancement on both viewports or on an individual viewport.
  • the client may request different representations based on a preference for fast viewport changes and/or high quality of one of the viewports.
  • two Representations may be provided with enhancement of each viewport separately such that the client may request decent viewport quality based on a viewing direction.
  • Representation selection may enable different trade-offs between low-latency viewport switching, lower storage cost, and/or decent viewing quality.
  • a client may request a specific representation based on the available bandwidth, the viewport the user is watching, and/or the viewing direction change (for example, how fast and/or how often the viewport changes).
  • the client WTRU may analyze one or more user habits locally using one or more gyroscope, accelerometer, and/or magnetometer tracking parameters to determine which Representations to request. For example, the client WTRU may request a Representation with an individual enhanced viewport if it detects that the user has not and/or does not change viewing direction frequently. To ensure low-latency rendering and/or sufficient viewport quality, the client WTRU may request a Representation with multiple enhanced viewports if it detects that the user keeps changing and/or has a tendency to change viewing directions.
  • Layer based 360 video delivery may be a viewport adaptive approach for 360° video streaming. Layer based 360 video delivery may decouple the viewport area from the entire frame. Layer based 360 video deliver may allow more flexible and/or efficient composition of various virtual and/or real objects onto the sphere.
  • An entire frame may be encoded as a full-frame video layer in lower quality, lower frame rate, and/or lower resolution.
  • One or more viewports may be encoded into one or more quality Representations as the viewport layer.
  • the viewport layer may be coded independently from the full-frame video layer.
  • the viewport layer may be coded more efficiently using scalable coding, e.g., using scalable extensions of HEVC (SHVC), where the full-frame video layer is used as a reference layer to code the one or more viewport layer representations with inter-layer prediction.
  • SHVC scalable extensions of HEVC
  • the user may always request the full-frame video layer (e.g., as a fallback layer).
  • the user WTRU has sufficient additional resource (e.g., bandwidth and/or computing resource)
  • one or more high quality enhancement viewport(s) may be requested to overlay onto the full-frame video layer.
  • the pixels outside the viewport may be subsampled.
  • the frame rate of the areas outside the viewport may not be directly reduced.
  • the layer-based 360 video delivery may decouple the viewport(s) from the entire 360° video frame.
  • the 360 video deliver may allow the viewport(s) to be coded at a higher bitrate, higher resolution, and/or higher frame rate, while the full-frame video layer 360° video may be coded at a lower bitrate, lower resolution, and/or lower frame rate.
  • the full-frame video layer may be coded in lower resolution and up-sampled for overlay at the client side. Upsampling the full-frame video layer at the client side may reduce the storage cost and/or provide acceptable quality at a lower bitrate. For the viewport layer, multiple quality Representations of the viewport(s) may be generated for a fine-granularity quality adaptation.
  • FIG. 14 depicts an example layer-based 360 video overlay.
  • a high quality viewport may be overlayed onto a 360 full frame video layer.
  • the full-frame video layer stream may include the entire 360° video at a lower quality.
  • the enhancement layer stream may include an enhanced viewport at a high quality.
  • One or more directional low pass filters may be applied across the overlay boundaries (e.g., horizontal and/or vertical boundaries) to smooth the abrupt quality changes.
  • a 1D or 2D low pass filter may be applied on a vertical boundary to smooth out one or more horizontal neighboring pixels along the vertical boundary and/or a similar low pass filter may be applied on a horizontal boundary to smooth out one or more vertical neighboring pixels along the horizontal boundary.
  • the full-frame video layer may include full 360° video in different projection formats such as equirectangular, cube-map, offset cube-map, pyramid, etc.
  • a projection format may have multiple full-frame video layer Representations supporting different quality levels such as resolution, bitrate, and/or frame rate.
  • the viewport layer may include multiple viewports.
  • a viewport may have a number of quality Representations with different resolutions, bitrate, and/or frame rate for adaptive streaming. Multiple Representations may be provided to support fine-granularity quality adaptation, without incurring high storage and/or transmission cost (e.g., since each viewport's size is relatively small compared to the size of the entire 360° video).
  • FIG. 15 depicts an example layer based 360 video representation.
  • two or more Representations may be available at the full-frame video layer for full 360° video with different resolutions (2048 ⁇ 1024@30 fps and 4096 ⁇ 2048@30 fps).
  • Two or more target viewports e.g., viewport # 1 and viewport # 2
  • Viewport # 1 and # 2 may have Representations with different resolution(s) and/or different bitrate(s).
  • a viewport layer Representation may use @dependencyId to identify a particular full-frame video layer Representation as the dependent Representation.
  • the user may request the target viewport Representation and/or its dependent full-frame video layer full frame Representation to compose the final 360° video for rendering.
  • a MPEG-DASH SRD element may be used to support layer-based 360 video streaming.
  • the MPD author may use one or more SRD values to describe a spatial relationship between a full-frame video layer 360° full video and a viewport layer video.
  • the SRD element may specify the spatial relationship of spatial objects.
  • the SRD element may not specify how to overlay one or more viewport(s) video onto the full-frame video layer. Viewport overlay may be specified to improve streaming quality for layer-based 360° video streaming.
  • Viewpoint overlay may be used with the Role descriptor applied to an AdaptationSet element.
  • a Role element with @schemeIdURI equal to “urn:mpeg:dash:viewport:overlay:2016” may signal which Representation is associated with the viewport layer video and/or which Representation is associated with the full-frame video layer.
  • the @value of the Role element may include one or more overlay indicators.
  • the one or more overlay indicators may include ‘f’ and/or ‘v,’ where “f” indicates the associated video Representation is a full-frame video layer and “v” indicates the associated video Representation is a viewport video to be overlaid on the full-frame video layer.
  • One or more viewport videos may be overlaid onto a full-frame video layer.
  • the @dependencyId may specify a particular full-frame video layer Representation for an associated viewport video overlay composition.
  • An original full-frame video layer resolution may be indicated by one or more SRD parameters (e.g., total_width and total_height of associated AdaptationSet).
  • the viewport original resolution may be indicated by the one or more SRD parameters object_width and object_height of associated AdaptationSet.
  • the reconstructed video may be up-sampled to align the full-frame video layer and the enhancement layer video for the overlay.
  • Table 12 is an example MPD SRD annotation for the layer based 360° video Representation example shown in FIG. 15 .
  • the AdaptationSet may include a SRD describing that the AdaptationSet element spans the entire reference space since the object_width and object_height parameters are equal to the total_width and total_height parameters.
  • the AdaptationSet may include two or more Representations that represent the same spatial part of the source but with different resolutions (e.g., a first Representation may have 4096 ⁇ 2048 resolution and a second Representation may have 2048 ⁇ 1024 resolution).
  • the Role element of the AdaptationSet with @schemeIdUri equal to “urn:mpeg:dash:viewport:overlay:2016” and @value equal to “b1” may indicate that one or more AdaptationSet elements are full-frame video layer full video.
  • the full-frame video layer Representation resolution specified by value of attribute @width and @height is not equal to the value of SRD parameters @total_width and @total_height
  • the full-frame video layer video may be up-scaled to match the full-frame video layer Representation resolution.
  • viewport # 1 Representations may be included in a first AdaptationSet
  • viewport # 2 Representations may be included in a second AdaptationSet.
  • Each viewport may have different resolution and/or bandwidth.
  • the AdaptationSet may include a SRD describing that the video in the AdaptationSet element represents only a part of the full 360° video (e.g., because its object_width and object_height parameters are less than its total_width and total_height respectively).
  • the Role element of the AdaptationSet with @schemeIdUri equal to “urn:mpeg:dash:viewport:overlay:2016” and @value equal to “f” may indicate the AdaptationSet elements are viewport video.
  • the viewport Representation resolution specified by value of attribute @width and @height is not equal to the value of SRD parameters @object_width and @object_height
  • the viewport video may be up-scaled to match the original full-frame resolution.
  • One or more (e.g., all) AdaptationSets may use the same first parameter source_id to indicate that the videos in AdaptationSets spatially relate to each other within a reference space at 4096 ⁇ 2048.
  • a non SRD-aware client may only see the full-frame video layer full video due to the use of SupplementalProperty rather than EssentialProperty.
  • FIG. 16 depicts an example layer-based 360 video overlay.
  • the example layer-based 360 video overlay may be associated with the Representations listed in Table 12.
  • the high quality viewport may be overlaid on the 360 video at the position signaled by object_x (e.g., 100) and object_y (e.g., 100) as signaled in the associated SRD parameters.
  • the video conversion shown in FIG. 16 may perform up-scaling, projection conversion, and/or layout conversion (e.g., when the Representation resolution, projection and/or layout are not matched between the full-frame video layer and the viewport video).
  • FIG. 17 depicts an example layer-based 360 video overlay flowchart.
  • the full-frame video layer and the enhancement layer representation may have different frame rates. Because the attribute @frameRate, along with @width, @height, and @bandwidth, are specified as common attributes and elements for AdaptationSet, Representation, and Sub-Representation elements, @width, @heightthe and @framerate may be signaled for AdaptationSet, Representation, and/or Sub-Representation. When the @frameRate is signaled at AdaptationSet level, one or more Representations allocated to the AdaptationSet may share the same value of @frameRate. Table 13 is an MPD example for the Representations with different frame rate at the AdaptationSet level.
  • Multiple target viewport videos may be requested and/or overlaid on a 360° full video sphere. Based on a user orientation position, the user may request multiple high quality viewports at the same time (e.g., to facilitate fast viewport changes and/or reduce overall system latency).
  • One or more (e.g., all) viewport layers may share the same full-frame video layer 360° video as a reference space.
  • the client may overlay the multiple viewports it received onto the same reference space for rendering.
  • the full-frame video layer video may be used as an anchor format and/or the enhancement layer viewport may be converted into the anchor format (e.g., so that the composition position can be aligned between full-frame video layer and the viewport video).
  • FIG. 18 depicts an example multiple viewports overlay.
  • a layer based overlay with two enhanced viewports may be signaled, as shown in Table 13.
  • a composite video with two enhanced viewports may be generated based on a determination that both viewports may be watched in the near-term timeframe (e.g., during the same segment duration). Additional project and/or layout conversion may be performed if the video type, projection, and/or layout are different between the full-frame video layer video and the enhancement layer viewport.
  • a layer-based 360 video overlay may be able to adapt to the viewport changes more efficiently.
  • the segment length of the full-frame video layer may be increased to improve the coding efficiency (e.g., because the full-frame video layer video is always delivered).
  • the segment length of the viewport video layer may be kept shorter to accommodate for fast switching between viewports.
  • the layer-based 360 video overlay described herein may use SRD to describe the spatial relationship of viewport video and/or overlay to describe the overlay procedure to form the composite video.
  • One or more spatial parameters such as source_id, object_x, object_y, object_width, object_height, total_width, and/or total_height specified in SRD may be merged into an overlay @value to simplify the MPD structure.
  • the @value of the Role element may include a comma-delimited list of viewport indicator, source_id, object_x, object_y, object_width, object_height, total_width, and/or total_height. Table 14 shows an MPD example with a merged overlay.
  • the layer-based 360 video overlay may be provided by one or more descriptors such as SupplementalProperty and/or EssentialProperty with @schemeIdUri equal to “urn:mpeg:dash:viewport:overlay:2016”.
  • Different video Representations may be in different projection formats (e.g., one Representation is equirectangular whereas another representation is cube-map or one video Representation is spherical video while another video Representation is rectilinear).
  • the video property may be specified from the OMAF file format.
  • One or more Common attributes specified for AdaptationSet, Representation, and/or Sub-Representation, @videoType may indicate the video type such as spherical video, light field video, and/or rectilinear video.
  • different projection formats and/or projection+layout combinations such as equirectangular, cube-map (e.g., in combination with different layouts in FIG.
  • AdaptationSet may be signaled as Common attribute @projection and/or @layout for AdaptationSet, Representation, and/or Sub-Representation elements.
  • the element and attributes may be provided using SupplementalProperty and/or EssentialProperty elements.
  • Table 15 is a semantic example of a video type and/or a projection attribute.
  • the projection attribute may assist the user to request the appropriate video based on the client's capability. For example, a client that does not support spherical video and/or light field video may request a rectilinear video only. In an example, a client that supports spherical video may select an appropriate format (e.g., equirectangular instead of cube-map) from the set of available formats.
  • an appropriate format e.g., equirectangular instead of cube-map
  • FIG. 19 depicts an example equirectangular representation with half height poles on the side.
  • FIG. 20 depicts an example equirectangular representation with full height poles on the side.
  • FIG. 21 depicts an example single row layout for the cube representation format (e.g., with region_id).
  • FIG. 22 depicts an example 2 ⁇ 3 layout for the cube representation format (e.g., with region_id).
  • FIG. 23 depicts an example 180° layout for cube-map.
  • the 180° layout for the cube map projection may reduce the resolution of the back half of the cube by 25% (e.g., by reducing the width and height of the back half areas by half) and/or may use the layout shown in FIG. 23 for the result, where regions a, b, c, d, and e belong to the front half and regions f, g, h, i, and j belong to the back half.
  • Table 19 is an MPD example using such common attributes and elements to support both equirectangular and cube-map projections.
  • FIG. 24A is a diagram of an example communications system 100 in which one or more disclosed embodiments may be implemented.
  • the communications system 100 may be a multiple access system that provides content, such as voice, data, video, messaging, broadcast, etc., to multiple wireless users.
  • the communications system 100 may enable multiple wireless users to access such content through the sharing of system resources, including wireless bandwidth.
  • the communications systems 100 may employ one or more channel access methods, such as code division multiple access (CDMA), time division multiple access (TDMA), frequency division multiple access (FDMA), orthogonal FDMA (OFDMA), single-carrier FDMA (SC-FDMA), and the like.
  • CDMA code division multiple access
  • TDMA time division multiple access
  • FDMA frequency division multiple access
  • OFDMA orthogonal FDMA
  • SC-FDMA single-carrier FDMA
  • the communications system 100 may include wireless transmit/receive units (WTRUs) 102 a, 102 b, 102 c, and/or 102 d (which generally or collectively may be referred to as WTRU 102 ), a radio access network (RAN) 103 / 104 / 105 , a core network 106 / 107 / 109 , a public switched telephone network (PSTN) 108 , the Internet 110 , and other networks 112 , though it will be appreciated that the disclosed embodiments contemplate any number of WTRUs, base stations, networks, and/or network elements.
  • WTRUs wireless transmit/receive units
  • RAN radio access network
  • PSTN public switched telephone network
  • Each of the WTRUs 102 a, 102 b, 102 c, 102 d may be any type of device configured to operate and/or communicate in a wireless environment.
  • the WTRUs 102 a, 102 b, 102 c, 102 d may be configured to transmit and/or receive wireless signals and may include user equipment (UE), a mobile station, a fixed or mobile subscriber unit, a pager, a cellular telephone, a personal digital assistant (PDA), a smartphone, a laptop, a netbook, a personal computer, a wireless sensor, consumer electronics, and the like.
  • UE user equipment
  • PDA personal digital assistant
  • the communications systems 100 may also include a base station 114 a and a base station 114 b.
  • Each of the base stations 114 a, 114 b may be any type of device configured to wirelessly interface with at least one of the WTRUs 102 a, 102 b, 102 c, 102 d to facilitate access to one or more communication networks, such as the core network 106 / 107 / 109 , the Internet 110 , and/or the networks 112 .
  • the base stations 114 a, 114 b may be a base transceiver station (BTS), a Node-B, an eNode B, a Home Node B, a Home eNode B, a site controller, an access point (AP), a wireless router, and the like. While the base stations 114 a, 114 b are each depicted as a single element, it will be appreciated that the base stations 114 a, 114 b may include any number of interconnected base stations and/or network elements.
  • BTS base transceiver station
  • AP access point
  • the base station 114 a may be part of the RAN 103 / 104 / 105 , which may also include other base stations and/or network elements (not shown), such as a base station controller (BSC), a radio network controller (RNC), relay nodes, etc.
  • BSC base station controller
  • RNC radio network controller
  • the base station 114 a and/or the base station 114 b may be configured to transmit and/or receive wireless signals within a particular geographic region, which may be referred to as a cell (not shown).
  • the cell may further be divided into cell sectors.
  • the cell associated with the base station 114 a may be divided into three sectors.
  • the base station 114 a may include three transceivers, e.g., one for each sector of the cell.
  • the base station 114 a may employ multiple-input multiple output (MIMO) technology and, therefore, may utilize multiple transceivers for each sector of the cell.
  • MIMO multiple-input multiple output
  • the base stations 114 a, 114 b may communicate with one or more of the WTRUs 102 a, 102 b, 102 c, 102 d over an air interface 115 / 116 / 117 , which may be any suitable wireless communication link (e.g., radio frequency (RF), microwave, infrared (IR), ultraviolet (UV), visible light, etc.).
  • the air interface 115 / 116 / 117 may be established using any suitable radio access technology (RAT).
  • RAT radio access technology
  • the communications system 100 may be a multiple access system and may employ one or more channel access schemes, such as CDMA, TDMA, FDMA, OFDMA, SC-FDMA, and the like.
  • the base station 114 a in the RAN 103 / 104 / 105 and the WTRUs 102 a, 102 b, 102 c may implement a radio technology such as Universal Mobile Telecommunications System (UMTS) Terrestrial Radio Access (UTRA), which may establish the air interface 115 / 116 / 117 using wideband CDMA (WCDMA).
  • WCDMA may include communication protocols such as High-Speed Packet Access (HSPA) and/or Evolved HSPA (HSPA+).
  • HSPA may include High-Speed Downlink Packet Access (HSDPA) and/or High-Speed Uplink Packet Access (HSUPA).
  • the base station 114 a and the WTRUs 102 a, 102 b, 102 c may implement a radio technology such as Evolved UMTS Terrestrial Radio Access (E-UTRA), which may establish the air interface 115 / 116 / 117 using Long Term Evolution (LTE) and/or LTE-Advanced (LTE-A).
  • E-UTRA Evolved UMTS Terrestrial Radio Access
  • LTE Long Term Evolution
  • LTE-A LTE-Advanced
  • the base station 114 a and the WTRUs 102 a, 102 b, 102 c may implement radio technologies such as IEEE 802.16 (e.g., Worldwide Interoperability for Microwave Access (WiMAX)), CDMA2000, CDMA2000 1X, CDMA2000 EV-DO, Interim Standard 2000 (IS-2000), Interim Standard 95 (IS-95), Interim Standard 856 (IS-856), Global System for Mobile communications (GSM), Enhanced Data rates for GSM Evolution (EDGE), GSM EDGE (GERAN), and the like.
  • IEEE 802.16 e.g., Worldwide Interoperability for Microwave Access (WiMAX)
  • CDMA2000, CDMA2000 1X, CDMA2000 EV-DO Code Division Multiple Access 2000
  • IS-95 Interim Standard 95
  • IS-856 Interim Standard 856
  • GSM Global System for Mobile communications
  • GSM Global System for Mobile communications
  • EDGE Enhanced Data rates for GSM Evolution
  • GERAN GSM EDGERAN
  • the base station 114 b in FIG. 24A may be a wireless router, Home Node B, Home eNode B, or access point, for example, and may utilize any suitable RAT for facilitating wireless connectivity in a localized area, such as a place of business, a home, a vehicle, a campus, and the like.
  • the base station 114 b and the WTRUs 102 c, 102 d may implement a radio technology such as IEEE 802.11 to establish a wireless local area network (WLAN).
  • the base station 114 b and the WTRUs 102 c, 102 d may implement a radio technology such as IEEE 802.15 to establish a wireless personal area network (WPAN).
  • WPAN wireless personal area network
  • the base station 114 b and the WTRUs 102 c, 102 d may utilize a cellular-based RAT (e.g., WCDMA, CDMA2000, GSM, LTE, LTE-A, etc.) to establish a picocell or femtocell.
  • a cellular-based RAT e.g., WCDMA, CDMA2000, GSM, LTE, LTE-A, etc.
  • the base station 114 b may have a direct connection to the Internet 110 .
  • the base station 114 b may not be required to access the Internet 110 via the core network 106 / 107 / 109 .
  • the RAN 103 / 104 / 105 may be in communication with the core network 106 / 107 / 109 , which may be any type of network configured to provide voice, data, applications, and/or voice over internet protocol (VoIP) services to one or more of the WTRUs 102 a, 102 b, 102 c, 102 d.
  • the core network 106 / 107 / 109 may provide call control, billing services, mobile location-based services, pre-paid calling, Internet connectivity, video distribution, etc., and/or perform high-level security functions, such as user authentication.
  • the RAN 103 / 104 / 105 and/or the core network 106 / 107 / 109 may be in direct or indirect communication with other RANs that employ the same RAT as the RAN 103 / 104 / 105 or a different RAT.
  • the core network 106 / 107 / 109 may also be in communication with another RAN (not shown) employing a GSM radio technology.
  • the core network 106 / 107 / 109 may also serve as a gateway for the WTRUs 102 a, 102 b, 102 c, 102 d to access the PSTN 108 , the Internet 110 , and/or other networks 112 .
  • the PSTN 108 may include circuit-switched telephone networks that provide plain old telephone service (POTS).
  • POTS plain old telephone service
  • the Internet 110 may include a global system of interconnected computer networks and devices that use common communication protocols, such as the transmission control protocol (TCP), user datagram protocol (UDP) and the internet protocol (IP) in the TCP/IP internet protocol suite.
  • the networks 112 may include wired or wireless communications networks owned and/or operated by other service providers.
  • the networks 112 may include another core network connected to one or more RANs, which may employ the same RAT as the RAN 103 / 104 / 105 or a different RAT.
  • the WTRUs 102 a, 102 b, 102 c, 102 d in the communications system 100 may include multi-mode capabilities, e.g., the WTRUs 102 a, 102 b, 102 c, 102 d may include multiple transceivers for communicating with different wireless networks over different wireless links.
  • the WTRU 102 c shown in FIG. 24A may be configured to communicate with the base station 114 a, which may employ a cellular-based radio technology, and with the base station 114 b, which may employ an IEEE 802 radio technology.
  • FIG. 24B is a system diagram of an example WTRU 102 .
  • the WTRU 102 may include a processor 118 , a transceiver 120 , a transmit/receive element 122 , a speaker/microphone 124 , a keypad 126 , a display/touchpad 128 , non-removable memory 130 , removable memory 132 , a power source 134 , a global positioning system (GPS) chipset 136 , and other peripherals 138 .
  • GPS global positioning system
  • the base stations 114 a and 114 b, and/or the nodes that base stations 114 a and 114 b may represent, such as but not limited to transceiver station (BTS), a Node-B, a site controller, an access point (AP), a home node-B, an evolved home node-B (eNodeB), a home evolved node-B (HeNB), a home evolved node-B gateway, and proxy nodes, among others, may include some or all of the elements depicted in FIG. 24B and described herein.
  • BTS transceiver station
  • Node-B a Node-B
  • AP access point
  • eNodeB evolved home node-B
  • HeNB home evolved node-B gateway
  • proxy nodes among others, may include some or all of the elements depicted in FIG. 24B and described herein.
  • the processor 118 may be a general purpose processor, a special purpose processor, a conventional processor, a digital signal processor (DSP), a plurality of microprocessors, one or more microprocessors in association with a DSP core, a controller, a microcontroller, Application Specific Integrated Circuits (ASICs), Field Programmable Gate Array (FPGAs) circuits, any other type of integrated circuit (IC), a state machine, and the like.
  • the processor 118 may perform signal coding, data processing, power control, input/output processing, and/or any other functionality that enables the WTRU 102 to operate in a wireless environment.
  • the processor 118 may be coupled to the transceiver 120 , which may be coupled to the transmit/receive element 122 . While FIG. 24B depicts the processor 118 and the transceiver 120 as separate components, it will be appreciated that the processor 118 and the transceiver 120 may be integrated together in an electronic package or chip.
  • the transmit/receive element 122 may be configured to transmit signals to, or receive signals from, a base station (e.g., the base station 114 a ) over the air interface 115 / 116 / 117 .
  • a base station e.g., the base station 114 a
  • the transmit/receive element 122 may be an antenna configured to transmit and/or receive RF signals.
  • the transmit/receive element 122 may be an emitter/detector configured to transmit and/or receive IR, UV, or visible light signals, for example.
  • the transmit/receive element 122 may be configured to transmit and receive both RF and light signals. It will be appreciated that the transmit/receive element 122 may be configured to transmit and/or receive any combination of wireless signals.
  • the WTRU 102 may include any number of transmit/receive elements 122 . More specifically, the WTRU 102 may employ MIMO technology. Thus, in one embodiment, the WTRU 102 may include two or more transmit/receive elements 122 (e.g., multiple antennas) for transmitting and receiving wireless signals over the air interface 115 / 116 / 117 .
  • the transceiver 120 may be configured to modulate the signals that are to be transmitted by the transmit/receive element 122 and to demodulate the signals that are received by the transmit/receive element 122 .
  • the WTRU 102 may have multi-mode capabilities.
  • the transceiver 120 may include multiple transceivers for enabling the WTRU 102 to communicate via multiple RATs, such as UTRA and IEEE 802.11, for example.
  • the processor 118 of the WTRU 102 may be coupled to, and may receive user input data from, the speaker/microphone 124 , the keypad 126 , and/or the display/touchpad 128 (e.g., a liquid crystal display (LCD) display unit or organic light-emitting diode (OLED) display unit).
  • the processor 118 may also output user data to the speaker/microphone 124 , the keypad 126 , and/or the display/touchpad 128 .
  • the processor 118 may access information from, and store data in, any type of suitable memory, such as the non-removable memory 130 and/or the removable memory 132 .
  • the non-removable memory 130 may include random-access memory (RAM), read-only memory (ROM), a hard disk, or any other type of memory storage device.
  • the removable memory 132 may include a subscriber identity module (SIM) card, a memory stick, a secure digital (SD) memory card, and the like.
  • SIM subscriber identity module
  • SD secure digital
  • the processor 118 may access information from, and store data in, memory that is not physically located on the WTRU 102 , such as on a server or a home computer (not shown).
  • the processor 118 may receive power from the power source 134 , and may be configured to distribute and/or control the power to the other components in the WTRU 102 .
  • the power source 134 may be any suitable device for powering the WTRU 102 .
  • the power source 134 may include one or more dry cell batteries (e.g., nickel-cadmium (NiCd), nickel-zinc (NiZn), nickel metal hydride (NiMH), lithium-ion (Li-ion), etc.), solar cells, fuel cells, and the like.
  • the processor 118 may also be coupled to the GPS chipset 136 , which may be configured to provide location information (e.g., longitude and latitude) regarding the current location of the WTRU 102 .
  • location information e.g., longitude and latitude
  • the WTRU 102 may receive location information over the air interface 115 / 116 / 117 from a base station (e.g., base stations 114 a, 114 b ) and/or determine its location based on the timing of the signals being received from two or more nearby base stations. It will be appreciated that the WTRU 102 may acquire location information by way of any suitable location-determination method while remaining consistent with an embodiment.
  • the processor 118 may further be coupled to other peripherals 138 , which may include one or more software and/or hardware modules that provide additional features, functionality and/or wired or wireless connectivity.
  • the peripherals 138 may include an accelerometer, an e-compass, a satellite transceiver, a digital camera (for photographs or video), a universal serial bus (USB) port, a vibration device, a television transceiver, a hands-free headset, a Bluetooth® module, a frequency modulated (FM) radio unit, a digital music player, a media player, a video game player module, an Internet browser, and the like.
  • the peripherals 138 may include an accelerometer, an e-compass, a satellite transceiver, a digital camera (for photographs or video), a universal serial bus (USB) port, a vibration device, a television transceiver, a hands-free headset, a Bluetooth® module, a frequency modulated (FM) radio unit, a digital music player, a media player, a
  • FIG. 24C is a system diagram of the RAN 103 and the core network 106 according to an embodiment.
  • the RAN 103 may employ a UTRA radio technology to communicate with the WTRUs 102 a, 102 b, 102 c over the air interface 115 .
  • the RAN 103 may also be in communication with the core network 106 .
  • the RAN 103 may include Node-Bs 140 a, 140 b, 140 c, which may each include one or more transceivers for communicating with the WTRUs 102 a, 102 b, 102 c over the air interface 115 .
  • the Node-Bs 140 a, 140 b, 140 c may each be associated with a particular cell (not shown) within the RAN 103 .
  • the RAN 103 may also include RNCs 142 a, 142 b. It will be appreciated that the RAN 103 may include any number of Node-Bs and RNCs while remaining consistent with an embodiment.
  • the Node-Bs 140 a, 140 b may be in communication with the RNC 142 a. Additionally, the Node-B 140 c may be in communication with the RNC 142 b.
  • the Node-Bs 140 a, 140 b, 140 c may communicate with the respective RNCs 142 a, 142 b via an Iub interface.
  • the RNCs 142 a, 142 b may be in communication with one another via an Iur interface.
  • Each of the RNCs 142 a, 142 b may be configured to control the respective Node-Bs 140 a, 140 b, 140 c to which it is connected.
  • each of the RNCs 142 a, 142 b may be configured to carry out or support other functionality, such as outer loop power control, load control, admission control, packet scheduling, handover control, macrodiversity, security functions, data encryption, and the like.
  • the core network 106 shown in FIG. 24C may include a media gateway (MGW) 144 , a mobile switching center (MSC) 146 , a serving GPRS support node (SGSN) 148 , and/or a gateway GPRS support node (GGSN) 150 . While each of the foregoing elements are depicted as part of the core network 106 , it will be appreciated that any one of these elements may be owned and/or operated by an entity other than the core network operator.
  • MGW media gateway
  • MSC mobile switching center
  • SGSN serving GPRS support node
  • GGSN gateway GPRS support node
  • the RNC 142 a in the RAN 103 may be connected to the MSC 146 in the core network 106 via an IuCS interface.
  • the MSC 146 may be connected to the MGW 144 .
  • the MSC 146 and the MGW 144 may provide the WTRUs 102 a, 102 b, 102 c with access to circuit-switched networks, such as the PSTN 108 , to facilitate communications between the WTRUs 102 a, 102 b, 102 c and traditional land-line communications devices.
  • the RNC 142 a in the RAN 103 may also be connected to the SGSN 148 in the core network 106 via an IuPS interface.
  • the SGSN 148 may be connected to the GGSN 150 .
  • the SGSN 148 and the GGSN 150 may provide the WTRUs 102 a, 102 b, 102 c with access to packet-switched networks, such as the Internet 110 , to facilitate communications between and the WTRUs 102 a, 102 b, 102 c and IP-enabled devices.
  • the core network 106 may also be connected to the networks 112 , which may include other wired or wireless networks that are owned and/or operated by other service providers.
  • FIG. 24D is a system diagram of the RAN 104 and the core network 107 according to an embodiment.
  • the RAN 104 may employ an E-UTRA radio technology to communicate with the WTRUs 102 a, 102 b, 102 c over the air interface 116 .
  • the RAN 104 may also be in communication with the core network 107 .
  • the RAN 104 may include eNode-Bs 160 a, 160 b, 160 c, though it will be appreciated that the RAN 104 may include any number of eNode-Bs while remaining consistent with an embodiment.
  • the eNode-Bs 160 a, 160 b, 160 c may each include one or more transceivers for communicating with the WTRUs 102 a, 102 b, 102 c over the air interface 116 .
  • the eNode-Bs 160 a, 160 b, 160 c may implement MIMO technology.
  • the eNode-B 160 a for example, may use multiple antennas to transmit wireless signals to, and receive wireless signals from, the WTRU 102 a.
  • Each of the eNode-Bs 160 a, 160 b, 160 c may be associated with a cell (not shown) and may be configured to handle radio resource management decisions, handover decisions, scheduling of users in the uplink and/or downlink, and the like. As shown in FIG. 24D , the eNode-Bs 160 a, 160 b, 160 c may communicate with one another over an X2 interface.
  • the core network 107 shown in FIG. 24D may include a mobility management gateway (MME) 162 , a serving gateway 164 , and a packet data network (PDN) gateway 166 . While each of the foregoing elements are depicted as part of the core network 107 , it will be appreciated that any one of these elements may be owned and/or operated by an entity other than the core network operator.
  • MME mobility management gateway
  • PDN packet data network
  • the MME 162 may be connected to each of the eNode-Bs 160 a, 160 b, 160 c in the RAN 104 via an S1 interface and may serve as a control node.
  • the MME 162 may be responsible for authenticating users of the WTRUs 102 a, 102 b, 102 c, bearer activation/deactivation, selecting a serving gateway during an initial attach of the WTRUs 102 a, 102 b, 102 c, and the like.
  • the MME 162 may also provide a control plane function for switching between the RAN 104 and other RANs (not shown) that employ other radio technologies, such as GSM or WCDMA.
  • the serving gateway 164 may be connected to each of the eNode-Bs 160 a, 160 b, 160 c in the RAN 104 via the S1 interface.
  • the serving gateway 164 may generally route and forward user data packets to/from the WTRUs 102 a, 102 b, 102 c.
  • the serving gateway 164 may also perform other functions, such as anchoring user planes during inter-eNode B handovers, triggering paging when downlink data is available for the WTRUs 102 a, 102 b, 102 c, managing and storing contexts of the WTRUs 102 a, 102 b, 102 c, and the like.
  • the serving gateway 164 may also be connected to the PDN gateway 166 , which may provide the WTRUs 102 a, 102 b, 102 c with access to packet-switched networks, such as the Internet 110 , to facilitate communications between the WTRUs 102 a, 102 b, 102 c and IP-enabled devices.
  • the PDN gateway 166 may provide the WTRUs 102 a, 102 b, 102 c with access to packet-switched networks, such as the Internet 110 , to facilitate communications between the WTRUs 102 a, 102 b, 102 c and IP-enabled devices.
  • the core network 107 may facilitate communications with other networks.
  • the core network 107 may provide the WTRUs 102 a, 102 b, 102 c with access to circuit-switched networks, such as the PSTN 108 , to facilitate communications between the WTRUs 102 a, 102 b, 102 c and traditional land-line communications devices.
  • the core network 107 may include, or may communicate with, an IP gateway (e.g., an IP multimedia subsystem (IMS) server) that serves as an interface between the core network 107 and the PSTN 108 .
  • the core network 107 may provide the WTRUs 102 a, 102 b, 102 c with access to the networks 112 , which may include other wired or wireless networks that are owned and/or operated by other service providers.
  • IMS IP multimedia subsystem
  • FIG. 24E is a system diagram of the RAN 105 and the core network 109 according to an embodiment.
  • the RAN 105 may be an access service network (ASN) that employs IEEE 802.16 radio technology to communicate with the WTRUs 102 a, 102 b, 102 c over the air interface 117 .
  • ASN access service network
  • the communication links between the different functional entities of the WTRUs 102 a, 102 b, 102 c, the RAN 105 , and the core network 109 may be defined as reference points.
  • the RAN 105 may include base stations 180 a, 180 b, 180 c, and an ASN gateway 182 , though it will be appreciated that the RAN 105 may include any number of base stations and ASN gateways while remaining consistent with an embodiment.
  • the base stations 180 a, 180 b, 180 c may each be associated with a cell (not shown) in the RAN 105 and may each include one or more transceivers for communicating with the WTRUs 102 a, 102 b, 102 c over the air interface 117 .
  • the base stations 180 a, 180 b, 180 c may implement MIMO technology.
  • the base station 180 a may use multiple antennas to transmit wireless signals to, and receive wireless signals from, the WTRU 102 a.
  • the base stations 180 a, 180 b, 180 c may also provide mobility management functions, such as handoff triggering, tunnel establishment, radio resource management, traffic classification, quality of service (QoS) policy enforcement, and the like.
  • the ASN gateway 182 may serve as a traffic aggregation point and may be responsible for paging, caching of subscriber profiles, routing to the core network 109 , and the like.
  • the air interface 117 between the WTRUs 102 a, 102 b, 102 c and the RAN 105 may be defined as an R1 reference point that implements the IEEE 802.16 specification.
  • each of the WTRUs 102 a, 102 b, 102 c may establish a logical interface (not shown) with the core network 109 .
  • the logical interface between the WTRUs 102 a, 102 b, 102 c and the core network 109 may be defined as an R2 reference point, which may be used for authentication, authorization, IP host configuration management, and/or mobility management.
  • the communication link between each of the base stations 180 a, 180 b, 180 c may be defined as an R8 reference point that includes protocols for facilitating WTRU handovers and the transfer of data between base stations.
  • the communication link between the base stations 180 a, 180 b, 180 c and the ASN gateway 182 may be defined as an R6 reference point.
  • the R6 reference point may include protocols for facilitating mobility management based on mobility events associated with each of the WTRUs 102 a, 102 b, 102 c.
  • the RAN 105 may be connected to the core network 109 .
  • the communication link between the RAN 105 and the core network 109 may defined as an R3 reference point that includes protocols for facilitating data transfer and mobility management capabilities, for example.
  • the core network 109 may include a mobile IP home agent (MIP-HA) 184 , an authentication, authorization, accounting (AAA) server 186 , and a gateway 188 . While each of the foregoing elements are depicted as part of the core network 109 , it will be appreciated that any one of these elements may be owned and/or operated by an entity other than the core network operator.
  • MIP-HA mobile IP home agent
  • AAA authentication, authorization, accounting
  • the MIP-HA may be responsible for IP address management, and may enable the WTRUs 102 a, 102 b, 102 c to roam between different ASNs and/or different core networks.
  • the MIP-HA 184 may provide the WTRUs 102 a, 102 b, 102 c with access to packet-switched networks, such as the Internet 110 , to facilitate communications between the WTRUs 102 a, 102 b, 102 c and IP-enabled devices.
  • the AAA server 186 may be responsible for user authentication and for supporting user services.
  • the gateway 188 may facilitate interworking with other networks.
  • the gateway 188 may provide the WTRUs 102 a, 102 b, 102 c with access to circuit-switched networks, such as the PSTN 108 , to facilitate communications between the WTRUs 102 a, 102 b, 102 c and traditional land-line communications devices.
  • the gateway 188 may provide the WTRUs 102 a, 102 b, 102 c with access to the networks 112 , which may include other wired or wireless networks that are owned and/or operated by other service providers.
  • Each of the computing systems described herein may have one or more computer processors having memory that are configured with executable instructions or hardware for accomplishing the functions described herein including determining the parameters described herein and sending and receiving messages between entities (e.g., WTRU and network or client and server) to accomplish the described functions.
  • entities e.g., WTRU and network or client and server
  • the processes described above may be implemented in a computer program, software, and/or firmware incorporated in a computer-readable medium for execution by a computer and/or processor.
  • the RAN 105 may be connected to other ASNs and the core network 109 may be connected to other core networks.
  • the communication link between the RAN 105 the other ASNs may be defined as an R4 reference point, which may include protocols for coordinating the mobility of the WTRUs 102 a, 102 b, 102 c between the RAN 105 and the other ASNs.
  • the communication link between the core network 109 and the other core networks may be defined as an R5 reference, which may include protocols for facilitating interworking between home core networks and visited core networks.
  • ROM read only memory
  • RAM random access memory
  • register cache memory
  • semiconductor memory devices magnetic media such as internal hard disks and removable disks, magneto-optical media, and optical media such as CD-ROM disks, and digital versatile disks (DVDs).
  • a processor in association with software may be used to implement a radio frequency transceiver for use in a WTRU, WTRU, terminal, base station, RNC, or any host computer.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Computer Security & Cryptography (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Information Transfer Between Computers (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Systems, methods, and instrumentalities are disclosed for client centric service quality control. A first viewport of a 360 degree video may be determined. The 360 degree video may comprise one or more of an equirectangular, a cube-map, a cylindrical, a pyramidal, and/or a spherical projection mapping. The first viewport may be associated with a spatial region of the 360 degree video. An adjacent area that extends around the spatial region may be determined. A second viewport of the 360 degree video may be determined. A bitstream associated with the 360 degree video may be received. One or more enhanced regions may be included in the bitstream. The one or more enhanced regions may correspond to the first and/or second viewport. A high coding bitrate may be associated with the first viewport and/or the second viewport.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims priority to and the benefit of U.S. Provisional Application Ser. No. 62/342,158, filed May 26, 2016, which is hereby incorporated by reference herein.
  • BACKGROUND
  • 360° video is a rapidly growing format emerging in the media industry. 360° video is enabled by the growing availability of virtual reality (VR) devices. 360° video may provide the viewer a new sense of presence. When compared to rectilinear video (e.g., 2D or 3D), 360° video may pose difficult engineering challenges on video processing and/or delivery. Enabling comfort and/or an immersive user experience may require high video quality and/or very low latency. The large video size of 360° video may be an impediment to delivering the 360° video in a quality manner at scale.
  • 360° video applications and/or services may encode the entire 360° video into a standard compliant stream for progressive downloading and/or adaptive streaming. Delivering the entire 360° video to the clients may enable low-latency rendering (e.g., the client may have access to the entirety of the 360° video content and/or can choose to render the portion it desires to see without further constraint). From the server's perspective, the same streams can support multiple users with possibly different viewports. The video size may be tremendously high, incurring high transmission bandwidth when the video is delivered (e.g., because the entire 360° video has to be encoded in high quality, such as 4K@60 fps or 6K@90 fps per eye). The high bandwidth consumption during delivery may be wasted, for example, because the user may only view a small portion (e.g., a viewport) of the entire picture.
  • SUMMARY
  • Systems, methods, and instrumentalities are disclosed for viewport adaptive 360° video delivery. Viewport enhancement based 360 degree video may be delivered and/or signaled. 360 degree video may be delivered using a layer-based viewport overlay. Signaling of 360 degree video mapping may be provided.
  • A first viewport of a 360 degree video may be determined. The 360 degree video may comprise one or more of an equirectangular, a cube-map, a cylindrical, a pyramidal, and/or a spherical projection mapping. The first viewport may be associated with a spatial region of the 360 degree video. An adjacent area that extends around the spatial region may be determined. A second viewport of the 360 degree video may be determined. A bitstream associated with the 360 degree video may be received. The bitstream may comprise one or more enhanced regions. The one or more enhanced regions may correspond to the first and/or second viewport. A high coding bitrate may be associated with the first viewport and/or the second viewport. Signaling, that indicates one or more viewport properties associated with the 360 degree video delivery, may be received.
  • A WTRU, for processing a 360 degree video may include a processor configured, (e.g., with executable instructions saved in memory), for one or more of the following: (i) receiving a media presentation description (MPD) associated with a 360 degree video comprising an essential property element that indicates a face-packing layout for a multi-face geometric projection format of a media segment; (ii) receiving the media segment; (iv) determining, based on the essential property element, at least one face-packing layout, from a set of face-packing layouts, for the received media segment; and (iv) constructing the received media segment based on the determined at least one face-packing layout.
  • The set of face-packing layouts comprise plate carree, poles on the side half height, poles on the side full height, single row, two by three, and one hundred and eighty degrees. The essential property element may be at one of an adaptation level and a representation level.
  • A WTRU processor may be configured, (e.g., with executable instructions saved in memory), to determine a media representation, that is associated with the MPD, to request for future media segments and to send a request for the determined media representation.
  • The MPD may include a video type, selected from a set of video types, for the media segment. The set of video types may include comprise rectilinear, panorama, sphere, and lightfield formats. The WTRU processor may be configured, (e.g., with executable instructions saved in memory), to determine the video type for the received media segment and/or to construct the received media segment using the determined video type.
  • The MPD may include at least one projection format used for projecting the 360 degree video from an omnidirectional format to a rectilinear video frame. The projection format may include one of a equirectangular, cube, an offset cube, a squished sphere, a pyramid, and a cylinder. The WTRU processor may be configured, (e.g., with executable instructions saved in memory), to determine a projection format for receiving a video file and/or to send a request for the determined projection format. The 360 degree video may include an Omnidirectional Media Application Format (OMAF) file.
  • A method of using a WTRU for processing a 360 degree video may include one or more of: (i) receiving a media presentation description (MPD) associated with a 360 degree video comprising an essential property element that indicates a face-packing layout for a multi-face geometric projection format of a media segment; (ii) receiving the media segment; (iii) determining, based on the essential property element at least one face-packing layout, from a set of face-packing layouts, for the received media segment; and (iv) constructing the received media segment based on the determined at least one face-packing layout.
  • The method of using a WTRU may include determining a media representation, that is associated with the MPD, to request for future received vide files and/or sending a request for the determined media representation. The method may include determining the video type for the received media segment and/or constructing the received media segment using the determined video type. The method may include determining a projection format for a video file and/or sending a request for the determined projection format.
  • A WTRU for processing a 360 degree video file my include a processor configured (e.g., with executable instructions saved in memory) for one or more of: receiving a media presentation description (MPD), associated with a 360 degree video, that includes a first essential property element that, indicates a first face-packing layout, for a multi-face geometric projection format of a media segment, and a second essential property element that, indicates a first second-packing layout, for a multi-face geometric projection format of a media segment; determine whether to use the first or the second face-packing layout for the media segment; request the determined at least first or the second face-packing layout; receive the media segment; and reconstruct the 360 degree video associated with the received media segment based on the requested face-packing layout.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 depicts an example portion of 360° video displayed on a head mounted device (HMD).
  • FIG. 2 depicts an example equirectangular projection for a 360° video.
  • FIG. 3 depicts example 360° video mappings.
  • FIG. 4 depicts an example media presentation description (MPD) hierarchical data model.
  • FIG. 5 depicts an example Dynamic Adaptive Streaming over HTTP (DASH) spatial relationship description (SRD) for a video.
  • FIG. 6 depicts an example tile based video partitioning.
  • FIG. 7 depicts an example temporal motion constrained tile set.
  • FIG. 8 depicts an example 360° video streaming quality degradation.
  • FIG. 9 depicts an example viewport area with an associated adjacent area.
  • FIGS. 10A-10D depict example cube-map layouts.
  • FIG. 11A depicts an example equirectangular coordinate viewport mapping.
  • FIG. 11B depicts an example cube-map coordinate viewport mapping.
  • FIG. 12 depicts an example spherical coordinate viewport mapping.
  • FIGS. 13A and 13B depict an example viewport enhancement representation.
  • FIG. 14 depicts an example layer-based 360° video overlay.
  • FIG. 15 depicts an example layer-based 360° video representation.
  • FIG. 16 depicts an example layer-based 360° video overlay.
  • FIG. 17 depicts a flowchart of an example layer-based 360° video overlay.
  • FIG. 18 depicts an example layer-based 360° video overlay with multiple viewports.
  • FIG. 19 depicts an example equirectangular representation with half height poles on the side.
  • FIG. 20 depicts an example equirectangular representation with full height poles on the side.
  • FIG. 21 depicts an example single row layout cube representation.
  • FIG. 22 depicts an example 2×3 layout cube representation.
  • FIG. 23 depicts an example 180° cube-map layout.
  • FIG. 24A is a system diagram of an example communications system in which one or more disclosed embodiments may be implemented.
  • FIG. 24B is a system diagram of an example wireless transmit/receive unit (WTRU) that may be used within the communications system illustrated in FIG. 24A.
  • FIG. 24C is a system diagram of an example radio access network and an example core network that may be used within the communications system illustrated in FIG. 24A.
  • FIG. 24D is a system diagram of another example radio access network and an example core network that may be used within the communications system illustrated in FIG. 24A.
  • FIG. 24E is a system diagram of another example radio access network and an example core network that may be used within the communications system illustrated in FIG. 24A.
  • DETAILED DESCRIPTION
  • A detailed description of illustrative embodiments will now be described with reference to the various Figures. Although this description provides a detailed example of possible implementations, it should be noted that the details are intended to be exemplary and in no way limit the scope of the application.
  • FIG. 1 depicts example portion of 360° video displayed on a head mounted device (HMD). When viewing a 360° video, a user may be presented with a part of the video, for example, as shown in FIG. 1. The part of the video may be changed when the user looks around and/or zooms the image. The part of the video may be changed based on feedback provided by the HMD and/or other types of user interface (e.g., a wireless transmit/receive unit (WTRU)). A viewport may be or may include a spatial region of the entire 360° video. The viewport may be fully or partially presented to the user. The viewport may have one or more different qualities than other parts of the 360° video.
  • A 360° video may be captured and/or rendered on a sphere (e.g., to give the user the ability to choose an arbitrary viewport). A spherical video format may not be delivered directly using conventional video codecs. A 360° video (e.g., such as a spherical video) may be compressed by projecting spherical video onto a 2D plane using a projection method. The projected 2D video may be coded (e.g., using conventional video codecs). An example of the projection method may include an equirectangular projection. FIG. 2 depicts an example equirectangular projection for a 360° video. For example, an equirectangular projection method may use one or more of the following equations to map a first point P with coordinate (θ, ϕ) on a sphere to a second point P with coordinate (u, v) on a 2D plane,

  • u=ϕ/(2*pi)+0.5   (1)

  • v=0.5−θ/(pi)   (2)
  • FIG. 3 depicts example 360° video mappings. For example, one or more other projection methods (e.g., mappings) may be used to convert 360 video to 2D planar video (e.g., in order to reduce the bandwidth requirement, alternative). For example, the one or more other projection methods may include a pyramid-map, a cube-map, and/or an offset cube-map. The one or more other projection methods may be used to represent the spherical video with less data.
  • A viewport-specific representation may be used. One or more of the projection methods shown in FIG. 3, for example, cube-map and/or pyramidal projection, may provide uneven quality representations for different viewports (e.g., some viewports may be represented in a higher quality than other view/ports). Multiple versions of the same video with different target viewports may be generated and/or stored at the server side (e.g., to support all viewports of the spherical video). For example, in Facebook's implementation of VR video delivery, the offset cube-map format shown in FIG. 3 may be used. The offset cube-map may provide a highest resolution (e.g., highest quality) to the front viewport, a lowest resolution (e.g., lowest quality) to the back view, and an intermediate resolution (e.g., intermediate quality) to one or more side views. A server may store multiple versions of the same content (e.g., to accommodate a client request for different viewports of the same content). For example, a total of 150 different versions (e.g., 30 viewports times 5 resolutions for each viewport) of the same content. During delivery (for example, streaming), the client may request a specific version that corresponds to its current viewport. The specific version may be delivered by the server.
  • In order to describe the various projection methods that may be used to represent 360 video and/or other non-conventional video formats (e.g., cylinder video that may be used in panoramic video representation), ISO/IEC/MPEG may define an Omnidirectional Media Application Format (OMAF). The OMAF file format metadata for the projection methods described herein may include support for projection metadata for video onto sphere, cube, cylinder, pyramid, and/or the like. Table 1 may show a syntax example of OMAF to support projection methods such as squished sphere, cylinder, and pyramid.
  • TABLE 1
    Omnidirectional media texture mapping metadata
    aligned(8) class OmnidirectionalMediaTextureMappingSampleEntry
    extends MetadatasampleEntry (‘omtm’){
    unsigned int(1) is_stereoscopic;
    unsigned int(1) is_default_front;
    unsigned int(6) reserved;
    if ( is_sterescopic )
    unsigned int(8) stereoscopic_type;
    unsigned int(8) geometry_type;
    if ( geometry_type == Squished Sphere ){
    unsigned int (16) squish_start_pitch_top;
    unsigned int (16) squish_start_pitch_bottom;
    unsigned int (8) squish_ratio_top;
    unsigned int (8) squish_ratio_bottom;
    }
    if ( geometry_type == cylinder )
    unsigned int(8) radius;
    if ( geometry_type == pyramid )
    unsigned int(8) pyramid_height;
    if ( !is_default_front ) {
    unsigned int(16)center_pitch;
    unsigned int(16)center_yaw;
    }
    }
  • HTTP streaming has become a dominant approach in commercial deployments. For instance, streaming platforms such as Apple's HTTP Live Streaming (HLS), Microsoft's Smooth Streaming (SS), and/or Adobe's HTTP Dynamic Streaming (HDS) may use HTTP streaming as an underlying delivery method. A standard for HTTP streaming of multimedia content may enable a standard-based client to stream content from any standard-based server (e.g., thereby enabling interoperability between servers and clients of different vendors). MPEG Dynamic Adaptive Streaming over HTTP (MPEG-DASH) may be a universal delivery format that provides end users with the best possible video experience by dynamically adapting to changing network conditions. DASH may be built on top of the HTTP/TCP/IP stack. DASH may define a manifest format, Media Presentation Description (MPD), and segment formats for ISO Base Media File Format and MPEG-2 Transport Streams.
  • Dynamic HTTP streaming may be associated with various bitrate alternatives of the multimedia content to be available at the server. The multimedia content may include several media components (e.g., audio, video, text), each of which may have different characteristics. In MPEG-DASH, the characteristics may be described by a Media Presentation Description (MPD).
  • The MPD may be an XML document that includes metadata necessary for a DASH Client to construct appropriate HTTP-URLs to access video segments (e.g., as described herein) in an adaptive manner during streaming sessions. FIG. 4 depicts an example media presentation description (MPD) hierarchical data model. The MPD may describe a sequence of Periods, where a consistent set of encoded versions of the media content components does not change during a Period. A Period may have a starting time and duration. A Period may be composed of one or more adaptation sets (e.g., AdaptationSet).
  • An AdaptationSet may represent a set of encoded versions of one or several media content components sharing one or more identical properties (e.g., such as the language, the media type, the picture aspect ratio, the role, the accessibility, the viewpoint, and/or the rating property). For instance, a first AdaptationSet may include different bitrates of the video component of the same multimedia content. A second AdaptationSet may include different bitrates of the audio component (e.g., lower quality stereo and/or higher quality surround sound) of the same multimedia content. An AdaptationSet may include multiple Representations.
  • A Representation may describe a deliverable encoded version of one or several media components, varying from other representations by bitrate, resolution, number of channels, and/or other characteristics. A representation may include one or more segments. One or more attributes of a Representation element (e.g., such as @id, @bandwidth, @qualityRanking, and @dependencyId) may be used to specify one or more properties of the associated Representation.
  • A Segment may be the largest unit of data that can be retrieved with a single HTTP request. A segment may have a URL (e.g., an addressable location on a server). A segment may be downloaded using HTTP GET or HTTP GET with byte ranges.
  • A DASH client may parse a MPD XML document. The DASH client may select a collection of AdaptationSets suitable for its environment, for example, based on information provided in the AdaptationSet elements. Within an AdaptationSet, the client may select a Representation. The client may select the Representation based on the value of @bandwidth attribute, client decoding capabilities, and/or client rendering capabilities. The client may download an initialization segment of the selected Representation. The client may access content (e.g., by requesting entire Segments or byte ranges of Segments). When the presentation has started, the client may continue consuming the media content. For example, the client may request (e.g., continuously request) Media Segments and/or parts of Media Segments during the presentation. The client may play content according to a media presentation timeline. The client may switch from a first Representation to a second representation, based on updated information from the client's environment. The client may play the content continuously across two or more Periods. When the client is consuming media contained in the Segments towards the end of the announced media in the Representation, the Media Presentation may be terminated, a Period may be started, and/or the MPD may be re-fetched.
  • The MPD descriptor element, Descriptor, may be provided to the application (e.g., to instantiate one or more description elements with the appropriate scheme information). One or more Descriptors (e.g., such as content protection, role, accessibility, rating, viewpoint, frame packing, and/or UTC timing descriptor) may include a @schemeIdUri attribute to identify the relative scheme.
  • A supplemental property descriptor (SupplementalProperty) may include metadata that may be used by the DASH client for optimizing processing.
  • An essential property descriptor (EssentialProperty) may include metadata for processing the containing element.
  • A Role MPD element may use a @schemeIdUri attribute to identify the role scheme employed to identify the role of the media content component. One or more Roles may define and/or describe one or more characteristics and/or structural functions of media content components. An Adaptation Set and/or a media content component may have multiple assigned roles (e.g., even within the same scheme).
  • MPEG-DASH may provide a spatial relationship description (SRD) scheme. The SRD scheme may express a spatial relationship of a video representing a spatial part of another full-frame video in two MPD elements (e.g., AdaptationSet and SubRepresentation). The SupplementalProperty and/or EssentialProperty descriptors with @schemeIdURI equal to “urn:mpeg:dash:srd:2014” may be used to provide spatial relationship information associated with the AdaptationSet and/or SubRepresentation. An attribute @value of the SupplementalProperty and/or EssentialProperty elements may provide one or more values for SRD parameters such as source_id, object_x, object_y, object_width, object_height, total_width, total_height, and/or spatial_set_id. The value and the semantics of the SRD parameters may be defined as shown in Table 2.
  • TABLE 2
    EssentialProperty and/or SupplementalProperty attributes for the SRD scheme
    EssentialProperty
    @value and/or
    SupplementalProperty
    @value parameter Description
    source_id a non-negative integer in decimal representation that may
    provide the identifier for the source of the content
    object_x a non-negative integer in decimal representation that may
    express the horizontal position of the top-left corner of
    the Spatial Object in arbitrary units
    object_y a non-negative integer in decimal representation that may
    express the vertical position of the top-left corner of the
    Spatial Object in arbitrary units
    object_width a non-negative integer in decimal representation that may
    express the width of the Spatial Object in arbitrary units
    object_height a non-negative integer in decimal representation that may
    express the height of the Spatial Object in arbitrary units
    total_width a non-negative integer in decimal representation that may
    express the width of the reference space in arbitrary units.
    At a Period and/or for a given source_id value, one or more
    of the following rules may apply:
    At least one descriptor may provide a value for the total_width
    parameter.
    If two or more descriptors provide different total_width
    values, all other descriptors may explicitly provide the value
    of total_width.
    If the total width value is provided in only one descriptor,
    all other descriptors may be assumed to use the total_width
    value.
    The value of total_width may be defined such that, for a
    descriptor using the value of total_width, the sum of object_x
    and object_width is smaller or equal to the total_width.
    When the value total_width is present, the value total_height
    may be present.
    total_height a non-negative integer in decimal representation that may
    express the height of the reference space in arbitrary units.
    At each Period and/or for a given source_id value, one or more
    of the following rules may apply:
    At least one descriptor may provide a value for the total_height
    parameter.
    If two or more descriptors provide different total_height
    values, all other descriptors may explicitly provide the value
    of total_height.
    If the total_height value is provided in only one descriptor,
    all other descriptors may be assumed to use the total_height
    value.
    The value of total height may be defined such that, for each
    descriptor using the value of total_height, the sum of object_y
    and object_height is smaller or equal to the total_height.
    When the value total_height is present, the value total_width
    may be present.
    spatial_set_id a non-negative integer in decimal representation that may
    provide an identifier for a group of Spatial Objects.
    When not present, the Spatial Object associated to this
    descriptor may not belong to any spatial set and/or no spatial
    set information may be given.
    When the value of spatial set_id is present, the value of
    total_width and/or total_height may be present.
  • FIG. 5 depicts an example DASH SRD video. The SRD may express that a video stream represents a spatial part of a full-frame video. The spatial part may be a tile and/or a region of interest (ROI) of the full-frame video. An SRD may describe the video stream in terms of the spatial part's position (object_x, object_y) and/or size (object_width, object_height) relative to the full-frame video (total_width, total_height). The SRD description may provide flexibility for the client in terms of adaptation. An SRD-aware DASH client may select a full-frame representation and/or a spatial part of the full-frame representation using one or more SRD annotations. Using the one or more SRD annotations to select the full-frame representation or a spatial part may save bandwidth and/or client-side computations, e.g., avoiding full-frame fetching, decoding, and/or cropping. Using the one or more SRD annotations to determine which representation to select may increase the quality of a given spatial part (e.g., the Region of Interest or ROI) of the full-frame video, for example, after a zoom. For example, the client may request a first video stream that corresponds to the ROI spatial part with higher quality and the client may request a second video stream that does not correspond to the ROI with lower quality, without increasing the overall bitrate.
  • Table 3 is an MPD example supporting SRD for the scenario as shown in FIG. 5, where each tile has a resolution of 1920×1080 and the whole frame has a resolution of 5760×3240 with 9 tiles.
  • TABLE 3
    MPD example with SRD elements
    <?xml version=“1.0” encoding=“UTF-8”?>
    <MPD
    xmlns:xsi=“http://www.w3.org/2001/XMLSchema-instance”
    xmlns=“urn:mpeg:dash:schema:mpd:2011”
    xsi:schemaLocation=“urn:mpeg:dash:schema:mpd:2011 DASH-MPD.xsd”
    [...]>
    <Period>
    <!-9 tiles -->
    <AdaptationSet [...]>
    <EssentialProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 0, 0, 1920, 1080, 5760, 3240”/>
    <Representation id=“1” bandwidth=“5000000”>
    <BaseURL>tile1.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet [...]>
    <EssentialProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 1920, 0, 1920, 1080”/>
    <Representation id=“2” bandwidth=“5000000”>
    <BaseURL>tile2.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet [...]>
    <EssentialProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 3840, 0, 1920, 1080”/>
    <Representation id=“3” bandwidth=“5000000”>
    <BaseURL>tile3.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet [...]>
    <EssentialProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 0, 1080, 1920, 1080”/>
    <Representation id=“4” bandwidth=“5000000”>
    <BaseURL>tile4.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
     <!-Legacy clients will end up with a valid MPD after processing unknown
    SupplementalProperty or EssentialProperty SRD descriptors. It will contain only this
    AdaptationSet since SupplementalProperty is used here. -->
    <AdaptationSet [...]>
    <SupplementalProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 1920, 1080, 1920, 1080”/>
    <Representation id=“5” bandwidth=“5000000”>
    <BaseURL>tile5.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet [...]>
    <EssentialProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 3840, 1080, 1920, 1080”/>
    <Representation id=“6” bandwidth=“5000000”>
    <BaseURL>tile6.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet [...]>
    <EssentialProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 0, 2160, 1920,1080”/>
    <Representation id=“7” bandwidth=“5000000”>
    <BaseURL>tile7.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet [...]>
    <EssentialProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 1920, 2160, 1920, 1080”>
    <Representation id=“8” bandwidth=“5000000”>
    <BaseURL>tile8.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet [...]>
    <EssentialProperty schemeIdUri=“urn:mpeg:dash:srd:2014”
    value=“1, 3840, 2160, 1920, 1080”/>
    <Representation id=“9” bandwidth=“5000000”>
    <BaseURL>tile9.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • FIG. 6 depicts an example tile based video partitioning. A 2D frame may be partitioned into one or more tiles, as shown in FIG. 6. Given MPEG-DASH SRD support, tile based adaptive streaming (TAS) may be used to support features like zooming and panning in a large panorama, spatial resolution enhancement, and/or server-based mosaic service. An SRD aware DASH client may use one or more SRD annotations to select a full-frame representation or a tile representation. Using the one or more SRD annotations to determine whether to select the full-frame representation of the tile representation may save bandwidth and/or client-side computations (e.g., avoiding full-frame fetching, decoding, and/or cropping).
  • FIG. 7 depicts an example temporal motion-constrained tile set. A frame may be encoded into a bitstream that includes a number of temporal motion-constrained tile sets as specified in HEVC. Each of the temporal motion-constrained tile sets may be decoded independently. As shown in FIG. 7, two or more left tiles may form a motion-constrained tile set that can be decoded independently (e.g., without decoding the full picture).
  • 360° video content may be delivered via progressive download or DASH adaptive streaming, which are primarily HTTP based streaming solutions. Dynamic streaming for 360° video based on UDP instead of HTTP has been proposed (e.g., by Facebook) to reduce latency.
  • Representing 360 video with sufficient quality may require the 2D projection to have a high resolution. When a full 2D layout is coded in high quality, the resulting bandwidth may be too high for effective delivery. Some mappings and/or projections may be used to reduce the amount of data by allowing different portions of the 360 video to be represented in different qualities. For example, a front view (e.g., viewport) may be represented in a high quality and a back (e.g., opposite) view may be represented in a low quality. One or more other views may be represented in one or more intermediate qualities. The offset cube-map and the pyramidal map, shown in FIG. 3, may be examples of mappings and/or projections that represent different portions of the 360 video in different qualities. Though the pyramidal map may reduce the number of pixels and/or save bitrate for each viewport, multiple viewport versions may handle different viewing positions that the client may request. When compared to delivering the entire 360 video, the viewport-specific representation and/or delivery may have higher latency adapting when the user changes viewing positions. For example, using the offset cube-map representation, the video quality may degrade when the user's head position rotates 180 degrees.
  • FIG. 8 depicts an example 360 video streaming quality degradation. DASH segment length and/or client buffer size may affect viewing quality. Longer segment lengths may result in higher coding efficiency. The longer segment lengths may not adapt to viewport changes as quickly. As shown in FIG. 8, a user may have three possible viewports of a 360 video (e.g., A, B and C). One or more (e.g., 3) segment types, SA, SB and SC, may be associated with the 360 video. Each of the one or more segment types may carry a higher quality of the corresponding viewport and a lower quality of the other viewports. The user may pan from viewport A to viewport B at time t1 during the playback of segment SA that carries a higher quality video for viewport A and a lower quality video for viewports B and C. The user may have to watch lower quality viewport B before switching to the next segment (SB), which carries higher quality video for viewport B and lower quality video for viewports A and C. Such negative user experience may be resolved with shorter segment lengths. Shorter segment lengths may reduce coding efficiency. The user may have to watch lower quality video when the user's streaming client logic pre-downloads too many segments based on a previous viewport. To prevent the user's streaming client from pre-downloading too many segments based on the previous viewport, the streaming buffer size may be reduced. A smaller streaming buffer size may affect streaming quality adaptation, for example, by causing more frequent buffer underflow.
  • Viewport adaptive 360° video streaming may include viewport enhancement based delivery and/or layer based delivery.
  • Efficient 360 video streaming may consider both bitrate adaptation and viewport adaptation. Viewport enhancement based 360° video delivery may include encoding one or more identified viewports of a 360 video frame in a high quality. For example, a viewport-based bit allocation may be performed during encoding. The viewport-based bit allocation may assign a bigger portion of the bits to the one or more viewports and/or assign a correspondingly reduced amount of bits to the other areas. FIG. 9 depicts an example viewport area with an associated adjacent area. A viewport area, an adjacent area, and other area may be determined for a 360° video frame.
  • A bitrate weight for the viewport area may be defined as a. The bitrate weight for the adjacent area may be defined as β. The bitrate weight for the other area may be defined as γ. One or more of the following equations may be used to determine the target bitrate for each area.

  • α+βγ=1   (3)

  • BR HQ =R×α  (4)

  • BR MQ =R×β  (5)

  • BR LQ =R×γ  (6)
  • Where R may represent a constant coding bitrate for the entire 360° video, BRHQ may represent a coding bitrate for the target viewports area, BRMQ may represent a coding bitrate for the viewport adjacent area and/or BRLQ may represent the coding bitrate for the other area.
  • In equation (3), the values of α, β, and γ may add up to 1, which may mean that the overall bitrate is kept the same. For example, bits may only be redistributed among the different areas (e.g., viewport, adjacent areas, and other areas). At an overall bitrate of R, the same video may be encoded into different versions. Each of the different versions may be associated with a different viewport quality level. Each of the different viewport quality levels may correspond to a different value of α.
  • The overall bitrate R may not be kept the same. For example, the viewport area, the adjacent area, and the other area may be encoded to a target quality level. In this case, the representation bitrate for each of the areas may be different.
  • The projection methods described herein, (e.g., such as offset cube-map, pyramid map) may enhance the quality of one or more target viewports and/or reduce the quality of other areas of the video.
  • At the server side, an AdaptationSet may include multiple Representation elements. A Representation of a 360 video stream may be coded in a specific resolution and/or at a specific bitrate. A Representation may be associated with one or more specific quality enhanced viewport(s). A Viewport Relationship Description (VRD) may specify one or more corresponding viewport spatial coordinate relationships. The SupplementalProperty and/or EssentialProperty descriptors with @schemeIdUri equal to “urn:mpeg:dash:viewport:2d:2016” may be used to provide a VRD associated with the AdaptationSet, Representation, and/or Sub-Representation elements.
  • The @value of the SupplementalProperty and/or EssentialProperty elements using the VRD scheme may be a comma separated list of values for viewport description parameters. Each AdaptationSet, Representation, and/or Sub-Representation may include one or more VRDs to represent one or more enhanced viewports. The properties of an enhanced viewport may be described by the VRD parameters as shown in Table 4.
  • TABLE 4
    Viewport relationship description scheme
    VRD scheme Description
    @schemeIdUri urn:mpeg:dash:viewport:2d:2016
    @value A comma-separate list of values for one or
    more of the following fields: viewport_id,
    viewport_x, viewport_y, viewport_width,
    viewport_height, full_width, full_height
    and/or viewport_quality
    viewport_id A non-negative integer in decimal
    representation that may provide an identifier
    for an identified viewport
    viewport_x A non-negative integer in decimal
    representation that may express the horizontal
    position of the top-left corner of the
    identified viewport
    viewport_y A non-negative integer in decimal
    representation that may express the vertical
    position of the top-left corner of the
    identified viewport
    viewport_width A non-negative integer in decimal
    representation that may express the width
    of the identified viewport
    viewport_height A non-negative integer in decimal
    representation that may express the height
    of the identified viewport
    full_width A non-negative integer in decimal
    representation that may express the width
    of the entire 360 video. If not present,
    the value may be inferred to be the same as
    the value of viewport_width.
    full_height A non-negative integer in decimal
    representation that may express the height
    of the entire 360 video. If not present,
    the value may be inferred to be the same as
    the value of viewport_height.
    viewport_quality A non-negative integer that may express a
    relative quality of the identified viewport,
    for example, on a scale of 0 to N − 1 (e.g.
    N = 5 or 10)
  • Table 5 provides a MPD example of a 4 k 360 video. The AdaptationSet may be annotated with two SupplementalProperty descriptors with VRD scheme identifier “urn:mpeg:dash:viewport:2d:2016”. A first descriptor may specify an enhanced 320×640 viewport # 1 at (150,150) with quality level 3. A second descriptor may specify 640×960 viewport # 2 at (1000, 1000) with quality level 5. Both viewports may represent spatial parts of a 4096×2048 full-frame 360° video. There may be two Representations, a first Representation may be full resolution and a second Representation may be half resolution. The viewport position and/or size of a Representation may be identified based on the values of VRD attributes @full_width and/or @full_height and/or the value of Representation attributes @width and/or @height. For example, viewport # 1 in full resolution Representation (e.g., @width=4096 and @height=2048) may be at (150,150) with the size 320×640, while viewport # 1 in half resolution Representation (e.g., @width=2048 and @height=1024) may be at (75,75) with the size 160×320. Depending on one or more capabilities of the user's WTRU, the half resolution video may be scaled up to the full resolution.
  • TABLE 5
    MPD example for the viewport description scheme
    <?xml version=“1.0” encoding=“UTF-8”?>
    <MPD
    xmlns=“urn:mpeg:dash:schema:mpd:2011”
    type=“static”
    mediaPresentationDuration=“PT10S”
    minBufferTime=“PT1S”
    profiles=“urn:mpeg:dash:profile:isoff-on-demand:2011”>
    <ProgramInformation>
    <Title>Example of a DASH Media Presentation Description using Viewport
    Relationship Description to indicate viewports of a 360 video</Title>
    </ProgramInformation>
    <Period>
    <!-- Main Video -->
    <AdaptationSet segmentAlignment=“true” subsegmentAlignment=“true”
    subsegmentStartsWithSAP=“1”>
    <Role schemeIdUri=“urn:mpeg:dash:role:2011” value=“main”/>
    <SupplementalProperty schemeIdUri=“urn:mpeg:dash:viewport:2d:2016”
    value=“0,150,150,320,640,4096,2048,3”/>
    <SupplementalProperty schemeIdUri=“urn:mpeg:dash:viewport:2d:2016”
    value=“1,1000,1000,640,960,4096,2048,5”/>
    <Representation mimeType=“video/mp4” width=“2048” height=“1024”
    bandwidth=“226597” startWithSAP=“1”>
    <BaseURL> half_360 video.mp4</BaseURL>
    </Representation>
    <Representation mimeType=“video/mp4” width=“4096” height=“2048”
    bandwidth=“1055223” startWithSAP=“1”>
    <BaseURL> full_360 video.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • A projection (e.g., equirectangular, cube-map, cylinder, and pyramid) may be used to map the surface of a sphere video to a flat image for processing. One or more layouts may be available for a particular projection. For example, for the equirectangular and/or cube-map projection format, different layouts may be used. FIGS. 10A-10D depict example cube-map layouts. The cube-map layouts may include a cube layout, a 2×3 layout, a poles on the side layout, and/or a single row layout.
  • The VRD in Table 4 may be specified for a particular projection format, for example, equirectangular projection. The VRD in Table 4 may be specified for a particular projection and layout combination, for example, cube-map project and a 2×3 layout (e.g., layout B shown in FIG. 10B). The VRD in Table 4 may be extended to support a variety of projection formats and/or projection and layout combination formats at the same time, as shown in Table 6. The server may support a set of popular projection and/or projection and layout formats, for example, using the signalling syntax shown in Table 6. FIG. 11 depicts example viewport coordinates in equirectangular and cube-map. One or more (e.g., two) viewports of the 360 video may be identified. As shown, the coordinate values of a viewport may be different in equirectangular and cube-map projection formats.
  • The VRD may specify the one or more viewport(s) associated with the Representation element. Table 7 shows an MPD example where both equirectangular and cube-map Representations are provided and the corresponding VRD is signalled accordingly. One or more first VRDs (e.g., VRD@viewport_id=0, 1) may specify one or more properties of a viewport in one or more first Representations (e.g., Representation@id=0,1) that are in an equirectangular projection format. One or more second VRDs (VRD@viewport_id=2, 3) may specify one or more properties of the viewport in one or more second Representations (e.g., Representation@id=2,3) that are in a cube-map projection format.
  • The VRD may specify one or more viewport(s) coordinates in one or more (e.g., all) common projection/layout formats (e.g., even though the associated Representation is in one projection/layout format). Table 8 shows an MPD example where the Representations of equirectangular projection are provided and the VRD of corresponding viewports for both equirectangular and cube-map are provided. For example, the projection and layout formats may be signalled at a Representation level as described herein.
  • The server may specify the viewports in different formats and/or may give the client the flexibility to choose an appropriate viewport (e.g., depending on the client's capability and/or technical specification). If the client cannot find a preferred format (e.g., either from the method specified in Table 4 or from the set of methods specified in Table), the client may convert one or more viewports from one of the specified formats into a format that it wants to use. For example, the one or more viewports may specified in the cube-map format but the client wants to use the equirectangular format. Based on the projection and/or layout description available in MPD, the client may derive a user orientation position from its gyroscope, accelerometer, and/or magnetometer tracking information. The client may convert the orientation position into a corresponding 2D position on a specific projection layout. The client may request the Representation with the identified viewport based on the value of one or more VRD parameters such as @viewport_x, @viewport_y, @viewport_width, and/or @viewport_height.
  • TABLE 6
    Extended EssentialProperty@value and/or SupplementalProperty@value
    attributes for VRD
    EssentialProperty
    @value and/or
    SupplementalProperty
    @value parameter Description
    viewport_id A non-negative integer in decimal representation
    that may provide an identifier for an identified
    viewport
    viewport_x A non-negative integer in decimal representation
    that may express the horizontal position of the
    top-left corner of the identified viewport
    viewport_y A non-negative integer in decimal representation
    that may express the vertical position of the
    top-left corner of the identified viewport
    viewport_width A non-negative integer in decimal representation
    that may express the width of the identified
    viewport
    viewport_height A non-negative integer in decimal representation
    that may express the height of the identified
    viewport
    full_width A non-negative integer in decimal representation
    that may express the width of the entire 360
    video. If not present, the value may be inferred
    to be the same as the value of viewport_width.
    full_height A non-negative integer in decimal representation
    that may express the height of the entire 360
    video. If not present, the value may be inferred to
    be the same as the value of viewport_height.
    viewport_quality A non-negative integer that may express a relative
    quality of the identified viewport, for example, on
    a scale of 0 to N − 1 (e.g. N = 5 or 10)
    projection May specify the projection method used to project
    omnidirectional video into flat video frame in
    the form of a string.
    layout May specify the layout format in the form of a
    string.
  • TABLE 7
    MPD example #1 for the VRD
    <?xml version=″1.0″ encoding=″UTF-8″?>
    <MPD
    xmlns=″urn:mpeg:dash:schema:mpd:2011″
    type=″static″
    mediaPresentationDuration=″PT10S″
    minBufferTime=″PT1S″
    profiles=″urn:mpeg:dash:profile:isoff-on-demand:2011″>
    <ProgramInformation>
    <Title>Example of a DASH Media Presentation Description using Viewport
    Relationship Description to indicate viewports of a 360 video</Title>
    </ProgramInformation>
    <Period>
    <!-- Equirectangular Video -->
    <AdaptationSet segmentAlignment=″true″ subsegmentAlignment=″true″
    subsegmentStartsWithSAP=″1″>
    <Role schemeIdUri=″urn:mpeg:dash:role:2011″ value=″main″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″0,10,10,720,1280,4096,2048,3,’equirectangular’,’regular’″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″1,2000,100,540,960,4096,2048,5,’equirectangul ar’,’regular’″/>
    <Representation @id=0 mimeType=″video/mp4″ codecs=″avc1.42c01e″
    width=″1920″ height=″1080″ bandwidth=″226597″ startWithSAP=″1″>
    <BaseURL> full_video_HD.mp4</BaseURL>
    </Representation>
    <Representation @id=1 mimeType=″video/mp4″ codecs=″avc1.42c033″
    width=″4096″ height=″2048″ bandwidth=″ 1055223″ startWithSAP=″1″>
    <BaseURL> full_360 video_4k.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <!-Cubemap Video -->
    <AdaptationSet segmentAlignment=″true″ subsegmentAlignment=″true″
    subsegmentStartsWithSAP=″1″>
    <Role schemeIdUri=″urn:mpeg:dash:role:2011″ value=″main″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″2,150,50,720,1280,4096,2048,3,’cubemap’,’2x3’″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″3,1000,1200,540,960,4096,2048,5,’cubemap’, ’2x3’″/>
    <Representation @id=2 mimeType=″video/mp4″ codecs=″avc1.42c01e″
    width=″1920″ height=″1080″ bandwidth=″226597″ startWithSAP=″1″>
    <BaseURL> Cubemap_video_HD.mp4</BaseURL>
    </Representation>
    <Representation @id=3 mimeType=″video/mp4″ codecs=″avc1.42c033″
    width=″4096″ height=″2048″ bandwidth=″1055223″ startWithSAP=″1″>
    <BaseURL> Cubemap_360 video_4k.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • TABLE 8
    MPD example #2 for the VRD
    <?xml version=″1.0″ encoding=″UTF-8″?>
    <MPD
    xmlns=″urn:mpeg:dash:schema:mpd:2011″
    type=″static″
    mediaPresentationDuration=″PT10S″
    minBufferTime=″PT1S″
    profiles=″urn:mpeg:dash:profile:isoff-on-demand:2011″>
    <ProgramInformation>
    <Title>Example of a DASH Media Presentation Description using Viewport
    Relationship Description to indicate viewports of a 360 video</Title>
    </ProgramInformation>
    <Period>
    <!-- Equirectangular Video -->
    <AdaptationSet segmentAlignment=″true″ subsegmentAlignment=″true″
    subsegmentStartsWithSAP=″1″>
    <Role schemeIdUri=″urn:mpeg:dash:role:2011″ value=″main″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″0,10,10,720,1280,4096,2048,3,’Equirectangular’,’regular’″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″1,2000,100,540,960,4096,2048,5,’Equirectangular’,’regular’″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″0,150,50,720,1280,4096,2048,3,’cubemap’,’2x3’″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″1,1000,1200,540,960,4096,2048,5,’cubemap’,’2x3’″/>
    <Representation mimeType=″video/mp4″ codecs=″avc1.42c01e″ width=″1920″
    height=″1080″ bandwidth=″226597″ startWithSAP=″1″ projection=”equirectangular”>
    <BaseURL> full_video_HD.mp4</BaseURL>
    </Representation>
    <Representation mimeType=″video/mp4″ codecs=″avc1.42c033″ width=″4096″
    height=″2048″ bandwidth=″1055223″ startWithSAP=″1″ projection=”equirectangular”>
    <BaseURL> full_360 video_4k.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • A general viewport descriptor, as shown in Table 9, may be provided. The general viewport descriptor may describe the viewport positions using spherical coordinates (θ, ϕ) as shown in FIG. 2, where θ may represent the inclination or polar angle, ϕ may represent the azimuthal angle, and/or a normalized radius may be 1.
  • SupplementalProperty and/or EssentialProperty descriptors with @schemeIdUri equal to “urn:mpeg:dash:viewport:sphere:2016” may be used to provide spherical coordinate based VRD associated with the AdaptationSet, Representation, and/or Sub-Representation elements. A region specified with the spherical coordinates may not correspond to a rectangular area on the 2D plane after projection. If the region does not correspond to a rectangular area on the 2D plane after projection, a bounding rectangle of the signalled region may be derived and/or used to specify the viewport.
  • The @value of the SupplementalProperty and/or EssentialProperty elements using the VRD may be a comma separated list of values for one or more viewport description parameters. Each AdaptationSet, Representation and/or Sub-Representation may include one or several VRD to represent one or more enhanced viewports. The properties of an enhanced viewport may be described by the parameters as shown in Table 9.
  • TABLE 9
    Viewport relationship description in spherical coordinates
    EssentialProperty and/or
    SupplementalProperty
    parameter Description
    @schemeIdUri urn:mpeg:dash:viewport:sphere:2016
    @value A comma-separate list of values for one or
    more of the following fields: viewport_id,
    viewport_inc, viewport_az,
    viewport_delta_inc, viewport_delta_az,
    and/or viewport_quality
    viewport_id A non-negative integer in decimal
    representation that may provide an identifier
    for one or more viewports
    viewport_inc An integer in decimal representation measured
    in degrees or radian that may express the
    inclination of an identified viewport top-left
    position
    viewport_az An integer in decimal representation measured
    in degrees or radian that may express the
    azimuthal of the identified viewport top-left
    position
    viewport_delta_inc A non-negative integer in decimal
    representation measured in degrees or radian
    that may express the inclination range of the
    identified viewport
    viewport_delta_az A non-negative integer in decimal
    representation measured in degrees or radian
    that may express the azimuthal range of the
    identified viewport
    viewport_quality A non-negative integer that may express a
    relative quality of the identified viewport,
    for example, on a scale of 0 to N − 1 (e.g.
    N = 5 or 10)
  • FIG. 12 depicts an example spherical coordinate viewport. For a viewport (e.g., noted as V in FIG. 12), viewport_inc may specify a polar angle θ, viewport_az may specify the azimuthal angle (ϕ), viewport_delta_inc may specify dθ, and/or viewport_delta_az may specify dϕ.
  • To distinguish between a VRD using 2D coordinates (e.g., Table 4 or Table 6) and a VRD using spherical coordinates (e.g., Table 9), different @schemeIdUri values may be used in each case. For example, for 2D viewport descriptors, the @schemeIdUri value may be urn:mpeg:dash:viewport:2d:2016. For viewports based on spherical coordinates, the @schemeIdUri value may be “urn:mpeg:dash:viewport:sphere:2016”.
  • Using a viewport descriptor that is based on the spherical coordinates may reduce signaling cost since viewports may only be specified for one coordinate system (e.g., the spherical coordinate system). Using a viewport descriptor that is based on the spherical coordinates may result in a simplified conversion process on the client side since each client may only need to implement a predefined conversion process. For example, each client may convert between the projection format the client chooses to use (e.g., equirectangular) and the spherical representation. When the client aligns the viewport coordinates, the client may use similar logic to decide which representation to request.
  • The VRD may be signaled in one or more Period.SupplementalProperty elements. The VRD may list available (e.g., all available) enhanced viewports respectively. A Representation may use an attribute, @viewportId, to signal one or more viewport indices to identify which enhanced viewport(s) are associated with the current Representation or Sub-Representation. With such a descriptor reference approach, redundant signaling of VRDs within a Representation may be avoided. One or more Representations with different associated viewport(s), projection formats, and/or layout formats may be allocated within a single AdaptationSet. Table 10 shows example semantics of Representation element attribute @viewportId and @viewport_quality.
  • As shown in Table 10, the attribute @viewport_quality may be signaled at a Representation level (e.g., instead of at the Period level) as part of the VRD. The @viewport_quality attribute may allow a client to select an appropriate quality level for one or more viewports of interest. For example, if the user is navigating through the 360 view frequently (e.g., turning head to look around all the time), the client may select a Representation with balanced quality between viewport and non-viewport. If the user is focusing on the viewport, the client may select a Representation with a high viewport quality (e.g., but relatively reduced quality in non-viewport areas). The @viewport_quality attribute may be signaled at the Representation level.
  • TABLE 10
    Semantic of Representation attributes
    Element and/or
    Attribute Name
    Representation Description
    @viewportId May specify one or more (e.g., all) identified
    viewports associated with the element as a
    whitespace-separated list of values of
    ViewportDescriptor@viewport_id attributes.
    @viewport_quality May specify the quality level of one or more
    corresponding viewports specified by
    @viewportId as a whitespace-separated
    list of values.
  • Table 11 is an MPD example using the Representation attributes, @viewportId and @viewport_quality, shown in Table 10 to specify the associated viewports. Two VRDs may be specified in Period.SupplementalProperty. A first Representation (e.g., @id=0) may include one or more viewports (e.g., @viewportId=0), viewport # 0 quality level may be 2. A second Representation (e.g., @id=1) may include the same one or more viewports as the first Representation (e.g., @viewportId=0), but the quality level of viewport # 0 may be 4. A third Representation (e.g., @id=2) may include one or more enhanced viewports (e.g., @viewportId=1) with quality level 5. A fourth Representation (e.g., @id=3) may include two or more enhanced viewports (e.g., @viewportId=0, 1) with highest quality level (e.g., @viewport_quality=5,5).
  • TABLE 11
    MPD example with representation @viewportId attribute
    <?xml version=″1.0″ encoding=″UTF-8″?>
    <MPD
    xmlns=″urn:mpeg:dash:schema:mpd:2011″
    type=″static″
    mediaPresentationDuration=″PT10S″
    minBufferTime=″PT1S″
    profiles=″urn:mpeg:dash:profile:isoff-on-demand:2011″>
    <ProgramInformation>
    <Title>Example of a DASH Media Presentation Description using Viewport
    Relationship Description to indicate viewports of a 360 video</Title>
    </ProgramInformation>
    <Period>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″0,0,0,360,640,4096,2048″/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:viewport:2d:2016″
    value=″1,1000,1000,540,960,4096,2048″/>
    <AdaptationSet segmentAlignment=″true″ subsegmentAlignment=″true″
    subsegmentStartsWithSAP=″1″>
    <Representation @id=”0” mimeType=″video/mp4″ codecs=″avc1.42c01e″
    width=″1920″ height=″1080″ bandwidth=″226597″ startWithSAP=″1″ @viewportId=”0”
    @viewport_quality=”2”>
    <BaseURL> e0.mp4</BaseURL>
    </Representation>
    <Representation @id=”1” mimeType=″video/mp4″ codecs=″avc1.42c033″
    width=″4096″ height=″2048″ bandwidth=″ 1055223″ startWithSAP=″1″
    @viewportId=”0” @viewport_quality=”4”>
    <BaseURL> e1.mp4</BaseURL>
    </Representation>
    <Representation @id=”2” mimeType=″video/mp4″ codecs=″avc1.42c01e″
    width=″1920″ height=″1080″ bandwidth=″226597″ startWithSAP=″1″ @viewportId=”1”
    @viewport_quality=”5”>
    <BaseURL> e2.mp4</BaseURL>
    </Representation>
    <Representation @id=”3” mimeType=″video/mp4″ codecs=″avc1.42c033″
    width=″4096″ height=″2048″ bandwidth=″1055223″ startWithSAP=″1″ @viewportId=”0
    1” @viewport_quahty=”5 5”>
    <BaseURL> e3.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • An enhanced viewport area may be selected to be larger than an actual viewing resolution, which may reduce quality degradation when the user changes (e.g., slightly changes) viewports. The enhanced viewport area may be selected to cover an area that a target object of interest moves around during a segment period (e.g., such that the user can watch the target object as the target moves around at the same high quality). One or more most watched viewports may be enhanced in a single bitstream, for example, to reduce the total number of Representations and/or corresponding media streams at the origin server or CDN. The viewport descriptors specified in Table 4, Table 6, Table 9, and/or Table 10 may support multiple quality-enhanced viewports within a single bitstream.
  • FIG. 13 depicts an example viewport enhancement representation. A 360° video may have two identified most viewed viewports (e.g., viewport # 1 and viewport #2). One or more DASH Representations may correspond to the two identified most viewed viewports. At a high coding bitrate (e.g., 20 mbps), both viewports may be enhanced at high quality. When both viewports are enhanced at high quality, one Representation may be provided with both enhanced viewports (e.g., to facilitate fast viewport changes and/or save the storage cost). At a medium bitrate, three Representations may be provided with the quality enhancement on both viewports or on an individual viewport. The client may request different representations based on a preference for fast viewport changes and/or high quality of one of the viewports. At a lower bitrate, two Representations may be provided with enhancement of each viewport separately such that the client may request decent viewport quality based on a viewing direction. Representation selection may enable different trade-offs between low-latency viewport switching, lower storage cost, and/or decent viewing quality.
  • During an adaptive 360° video streaming session, a client may request a specific representation based on the available bandwidth, the viewport the user is watching, and/or the viewing direction change (for example, how fast and/or how often the viewport changes). The client WTRU may analyze one or more user habits locally using one or more gyroscope, accelerometer, and/or magnetometer tracking parameters to determine which Representations to request. For example, the client WTRU may request a Representation with an individual enhanced viewport if it detects that the user has not and/or does not change viewing direction frequently. To ensure low-latency rendering and/or sufficient viewport quality, the client WTRU may request a Representation with multiple enhanced viewports if it detects that the user keeps changing and/or has a tendency to change viewing directions.
  • Layer based 360 video delivery may be a viewport adaptive approach for 360° video streaming. Layer based 360 video delivery may decouple the viewport area from the entire frame. Layer based 360 video deliver may allow more flexible and/or efficient composition of various virtual and/or real objects onto the sphere.
  • An entire frame may be encoded as a full-frame video layer in lower quality, lower frame rate, and/or lower resolution. One or more viewports may be encoded into one or more quality Representations as the viewport layer. The viewport layer may be coded independently from the full-frame video layer. The viewport layer may be coded more efficiently using scalable coding, e.g., using scalable extensions of HEVC (SHVC), where the full-frame video layer is used as a reference layer to code the one or more viewport layer representations with inter-layer prediction. The user may always request the full-frame video layer (e.g., as a fallback layer). When the user WTRU has sufficient additional resource (e.g., bandwidth and/or computing resource), one or more high quality enhancement viewport(s) may be requested to overlay onto the full-frame video layer.
  • The pixels outside the viewport may be subsampled. The frame rate of the areas outside the viewport may not be directly reduced. The layer-based 360 video delivery may decouple the viewport(s) from the entire 360° video frame. The 360 video deliver may allow the viewport(s) to be coded at a higher bitrate, higher resolution, and/or higher frame rate, while the full-frame video layer 360° video may be coded at a lower bitrate, lower resolution, and/or lower frame rate.
  • The full-frame video layer may be coded in lower resolution and up-sampled for overlay at the client side. Upsampling the full-frame video layer at the client side may reduce the storage cost and/or provide acceptable quality at a lower bitrate. For the viewport layer, multiple quality Representations of the viewport(s) may be generated for a fine-granularity quality adaptation.
  • FIG. 14 depicts an example layer-based 360 video overlay. For example, a high quality viewport may be overlayed onto a 360 full frame video layer. The full-frame video layer stream may include the entire 360° video at a lower quality. The enhancement layer stream may include an enhanced viewport at a high quality.
  • One or more directional low pass filters may be applied across the overlay boundaries (e.g., horizontal and/or vertical boundaries) to smooth the abrupt quality changes. For example, a 1D or 2D low pass filter may be applied on a vertical boundary to smooth out one or more horizontal neighboring pixels along the vertical boundary and/or a similar low pass filter may be applied on a horizontal boundary to smooth out one or more vertical neighboring pixels along the horizontal boundary.
  • The full-frame video layer may include full 360° video in different projection formats such as equirectangular, cube-map, offset cube-map, pyramid, etc. A projection format may have multiple full-frame video layer Representations supporting different quality levels such as resolution, bitrate, and/or frame rate.
  • The viewport layer may include multiple viewports. A viewport may have a number of quality Representations with different resolutions, bitrate, and/or frame rate for adaptive streaming. Multiple Representations may be provided to support fine-granularity quality adaptation, without incurring high storage and/or transmission cost (e.g., since each viewport's size is relatively small compared to the size of the entire 360° video).
  • FIG. 15 depicts an example layer based 360 video representation. For example, two or more Representations may be available at the full-frame video layer for full 360° video with different resolutions (2048×1024@30 fps and 4096×2048@30 fps). Two or more target viewports (e.g., viewport # 1 and viewport #2) may be available at the viewport layer. Viewport # 1 and #2 may have Representations with different resolution(s) and/or different bitrate(s). A viewport layer Representation may use @dependencyId to identify a particular full-frame video layer Representation as the dependent Representation. The user may request the target viewport Representation and/or its dependent full-frame video layer full frame Representation to compose the final 360° video for rendering.
  • A MPEG-DASH SRD element may be used to support layer-based 360 video streaming. Depending on the application, the MPD author may use one or more SRD values to describe a spatial relationship between a full-frame video layer 360° full video and a viewport layer video. The SRD element may specify the spatial relationship of spatial objects. The SRD element may not specify how to overlay one or more viewport(s) video onto the full-frame video layer. Viewport overlay may be specified to improve streaming quality for layer-based 360° video streaming.
  • Viewpoint overlay may be used with the Role descriptor applied to an AdaptationSet element. A Role element with @schemeIdURI equal to “urn:mpeg:dash:viewport:overlay:2016” may signal which Representation is associated with the viewport layer video and/or which Representation is associated with the full-frame video layer. The @value of the Role element may include one or more overlay indicators. For example, the one or more overlay indicators may include ‘f’ and/or ‘v,’ where “f” indicates the associated video Representation is a full-frame video layer and “v” indicates the associated video Representation is a viewport video to be overlaid on the full-frame video layer. One or more viewport videos may be overlaid onto a full-frame video layer. The @dependencyId may specify a particular full-frame video layer Representation for an associated viewport video overlay composition. An original full-frame video layer resolution may be indicated by one or more SRD parameters (e.g., total_width and total_height of associated AdaptationSet). The viewport original resolution may be indicated by the one or more SRD parameters object_width and object_height of associated AdaptationSet. When the resolution of the Representation indicated by @width and @height is less than a corresponding resolution specified by SRD, the reconstructed video may be up-sampled to align the full-frame video layer and the enhancement layer video for the overlay. Table 12 is an example MPD SRD annotation for the layer based 360° video Representation example shown in FIG. 15.
  • In Table 12, two or more full-frame Representations identified with @id may be included in the same AdaptationSet. A first Representation (@id=2) may include full resolution 4096×2048, and a second Representation (@id=1) may include half resolution. The AdaptationSet may include a SRD describing that the AdaptationSet element spans the entire reference space since the object_width and object_height parameters are equal to the total_width and total_height parameters. The AdaptationSet may include two or more Representations that represent the same spatial part of the source but with different resolutions (e.g., a first Representation may have 4096×2048 resolution and a second Representation may have 2048×1024 resolution). The Role element of the AdaptationSet with @schemeIdUri equal to “urn:mpeg:dash:viewport:overlay:2016” and @value equal to “b1” may indicate that one or more AdaptationSet elements are full-frame video layer full video. When the full-frame video layer Representation resolution specified by value of attribute @width and @height is not equal to the value of SRD parameters @total_width and @total_height, the full-frame video layer video may be up-scaled to match the full-frame video layer Representation resolution.
  • For the enhancement layer, viewport # 1 Representations may be included in a first AdaptationSet, and viewport # 2 Representations may be included in a second AdaptationSet. Each viewport may have different resolution and/or bandwidth. The AdaptationSet may include a SRD describing that the video in the AdaptationSet element represents only a part of the full 360° video (e.g., because its object_width and object_height parameters are less than its total_width and total_height respectively). The Role element of the AdaptationSet with @schemeIdUri equal to “urn:mpeg:dash:viewport:overlay:2016” and @value equal to “f” may indicate the AdaptationSet elements are viewport video. When the viewport Representation resolution specified by value of attribute @width and @height is not equal to the value of SRD parameters @object_width and @object_height, the viewport video may be up-scaled to match the original full-frame resolution.
  • One or more (e.g., all) AdaptationSets may use the same first parameter source_id to indicate that the videos in AdaptationSets spatially relate to each other within a reference space at 4096×2048. A non SRD-aware client may only see the full-frame video layer full video due to the use of SupplementalProperty rather than EssentialProperty. An SRD-aware client may form a 360° video by overlaying viewport #1 (@id=3/4/5) and/or viewport #2 (@id=6/7/8) video (e.g., to be up-sampled if necessary) on either full-frame video layer (@id=1/2).
  • TABLE 12
    MPD SRD example for layer-based 360 video
    <?xml version=″1.0″ encoding=″UTF-8″?>
    <MPD
    xmlns:xsi=″http://www.w3.org/2001/XMLSchema-instance″
    xmlns=″urn:mpeg:dash:schema:mpd:2011″
    xsi:schemaLocation=″urn:mpeg:dash:schema:mpd:2011 DASH-MPD.xsd″
    [...]>
    <Period>
    <!-- Full 360 video full-frame video layer -->
    <AdaptationSet [...]>
    <Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”f”/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 0, 0, 4096, 2048, 4096, 2048″/>
    <Representation id=″1″ bandwidth=″1000000″ width=″2048″ height=″1024″ ...>
    <BaseURL>b1.mp4</BaseURL>
    </Representation>
    <Representation id=″2″ bandwidth=″4000000″ width=″4096″ height=″2048″ ...>
    <BaseURL>b2.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <!-- viewport 1 video -->
    <AdaptationSet [...]>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”v”/>
    < EssentialProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 100, 100, 720, 1280, 4096, 2048″/>
    <Representation id=″3″ bandwidth=″1000000″ width=″360″ height=″640″...>
    <BaseURL>e3.mp4</BaseURL>
    </Representation>
    <Representation id=″4″ bandwidth=″2000000″ width=″360″ height=″640″...>
    <BaseURL>e4.mp4</BaseURL>
    </Representation>
    <Representation id=″5″ bandwidth=″3000000″ width=″720″ height=″1280″...>
    <BaseURL>e5.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <!-- viewport 2 video -->
    <AdaptationSet [...]>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”v”/>
    < EssentialProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 2000, 100, 1080, 1920, 4096, 2048″/>
    <Representation id=″6″ bandwidth=″2000000″ width=″720″ height=″1280″...>
    <BaseURL>e6.mp4</BaseURL>
    </Representation>
    <Representation id=″7″ bandwidth=″4000000″ width=″720″ height=″1280″...>
    <BaseURL>e7.mp4</BaseURL>
    </Representation>
    <Representation id=″8″ bandwidth=″6000000″ width=″1080″ height=″1920″...>
    <BaseURL>e8.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • FIG. 16 depicts an example layer-based 360 video overlay. For example, the example layer-based 360 video overlay may be associated with the Representations listed in Table 12. The client may request a half resolution full-frame video layer Representation (@id=1) and the enhancement layer viewport # 1 Representation (@id=5). The half resolution full-frame video layer may be up-sampled to a full resolution 4096×2048 since its Representation resolution, @width=2048 and @height=1024, is less than the associated SRD parameters, @total_width=4096 and @total_height=2048. The high quality viewport may be overlaid on the 360 video at the position signaled by object_x (e.g., 100) and object_y (e.g., 100) as signaled in the associated SRD parameters.
  • The video conversion shown in FIG. 16 may perform up-scaling, projection conversion, and/or layout conversion (e.g., when the Representation resolution, projection and/or layout are not matched between the full-frame video layer and the viewport video).
  • FIG. 17 depicts an example layer-based 360 video overlay flowchart.
  • The full-frame video layer and the enhancement layer representation may have different frame rates. Because the attribute @frameRate, along with @width, @height, and @bandwidth, are specified as common attributes and elements for AdaptationSet, Representation, and Sub-Representation elements, @width, @heightthe and @framerate may be signaled for AdaptationSet, Representation, and/or Sub-Representation. When the @frameRate is signaled at AdaptationSet level, one or more Representations allocated to the AdaptationSet may share the same value of @frameRate. Table 13 is an MPD example for the Representations with different frame rate at the AdaptationSet level.
  • TABLE 13
    MPD SRD example for layer based 360 video with different frame rate representations
    <?xml version=″1.0″ encoding=″UTF-8″?>
    <MPD
    xmlns:xsi=″http://www.w3.org/2001/XMLSchema-instance″
    xmlns=″urn:mpeg:dash:schema:mpd:2011″
    xsi:schemaLocation=″urn:mpeg:dash:schema:mpd:2011 DASH-MPD.xsd″
    [...]>
    <Period>
    <!-Full 360 video full-frame video layer -->
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=“30”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”f”/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 0, 0, 4096, 2048, 4096, 2048″/>
    <Representation id=″1″ bandwidth=″1000000″ width=″2048″ height=″1024″ ...>
    <BaseURL>b1.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”60”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”f”/>
    <SupplementalProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 0, 0, 4096, 2048, 4096, 2048″/>
    <Representation id=″2″ bandwidth=″4000000″ width=″4096″ height=″2048″ ...>
    <BaseURL>b2.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <!-viewport 1 video -->
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”30”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”v”/>
    < EssentialProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 100, 100, 720, 1280, 4096, 2048″/>
    <Representation id=″3″ bandwidth=″1000000″ width=″360″ height=″640″
    dependencyId=”1”...>
    <BaseURL>e3.mp4</BaseURL>
    </Representation>
    <Representation id=″4″ bandwidth=″2000000″ width=″360″ height=″640″
    dependencyId=”1”...>
    <BaseURL>e4.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”60”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”v”/>
    < EssentialProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 100, 100, 720, 1280, 4096, 2048″/>
    <Representation id=″5″ bandwidth=″3000000″ width=″720″ height=″1280″
    dependencyId=”1”...>
    <BaseURL>e5.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <!-viewport 2 video -->
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”30”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”v”/>
    < EssentialProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 2000, 100, 1080, 1920, 4096, 2048″/>
    <Representation id=″6″ bandwidth=″2000000″ width=″720″ height=″1280″
    dependencyId=”2”...>
    <BaseURL>e6.mp4</BaseURL>
    </Representation>
    <Representation id=″7″ bandwidth=″4000000″ width=″720″ height=″1280″
    dependencyId=”2”...>
    <BaseURL>e7.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”60”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”v”/>
    < EssentialProperty schemeIdUri=″urn:mpeg:dash:srd:2014″
    value=″1, 2000, 100, 1080, 1920, 4096, 2048″/>
    <Representation id=″8″ bandwidth=″6000000″ width=″1080″ height=″1920″
    dependencyId=”2”...>
    <BaseURL>e8.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • Multiple target viewport videos may be requested and/or overlaid on a 360° full video sphere. Based on a user orientation position, the user may request multiple high quality viewports at the same time (e.g., to facilitate fast viewport changes and/or reduce overall system latency). One or more (e.g., all) viewport layers may share the same full-frame video layer 360° video as a reference space. The client may overlay the multiple viewports it received onto the same reference space for rendering. When the projection and layout are different among full-frame video layer and enhancement layer viewport, the full-frame video layer video may be used as an anchor format and/or the enhancement layer viewport may be converted into the anchor format (e.g., so that the composition position can be aligned between full-frame video layer and the viewport video).
  • FIG. 18 depicts an example multiple viewports overlay. For example, a layer based overlay with two enhanced viewports may be signaled, as shown in Table 13. The client may request a full-frame video layer 360° video (e.g., Representation @id=1) as a reference space and/or two high quality enhancement layer viewports (e.g., Representations @id=5 and @id=8). A composite video with two enhanced viewports may be generated based on a determination that both viewports may be watched in the near-term timeframe (e.g., during the same segment duration). Additional project and/or layout conversion may be performed if the video type, projection, and/or layout are different between the full-frame video layer video and the enhancement layer viewport. A layer-based 360 video overlay may be able to adapt to the viewport changes more efficiently. The segment length of the full-frame video layer may be increased to improve the coding efficiency (e.g., because the full-frame video layer video is always delivered). The segment length of the viewport video layer may be kept shorter to accommodate for fast switching between viewports.
  • The layer-based 360 video overlay described herein may use SRD to describe the spatial relationship of viewport video and/or overlay to describe the overlay procedure to form the composite video. One or more spatial parameters such as source_id, object_x, object_y, object_width, object_height, total_width, and/or total_height specified in SRD may be merged into an overlay @value to simplify the MPD structure. The @value of the Role element may include a comma-delimited list of viewport indicator, source_id, object_x, object_y, object_width, object_height, total_width, and/or total_height. Table 14 shows an MPD example with a merged overlay.
  • TABLE 14
    MPD example of overlay w/o SRD
    <?xml version=″1.0″ encoding=″UTF-8″?>
    <MPD
    xmlns:xsi=″http://www.w3.org/2001/XMLSchema-instance″
    xmlns=″urn:mpeg:dash:schema:mpd:2011″
    xsi:schemaLocation=″urn:mpeg:dash:schema:mpd:2011 DASH-MPD.xsd″
    [...]>
    <Period>
    <!-Full 360 video full-frame video layer -->
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”30”>
    <Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”’f’, 1, 0, 0,
    4096, 2048, 4096, 2048”/>
    <Representation id=″1″ bandwidth=″1000000″ width=″2048″ height=″1024″ ...>
    <BaseURL>b1.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”60”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”’f’, 1, 0, 0,
    4096, 2048, 4096, 2048”/>
    <Representation id=″2″ bandwidth=″4000000″ width=″4096″ height=″2048″ ...>
    <BaseURL>b2.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <!-viewport 1 video -->
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”30”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”’v’, 1, 100,
    100, 720, 1280, 4096, 2048”/>
    <Representation id=″3″ bandwidth=″1000000″ width=″360″ height=″640″
    <BaseURL>e3.mp4</BaseURL>
    </Representation>
    <Representation id=″4″ bandwidth=″2000000″ width=″360″ height=″640″
    <BaseURL>e4.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”60”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”’v’, 1, 100,
    100, 720, 1280, 4096, 2048”/>
    <Representation id=″5″ bandwidth=″3000000″ width=″720″ height=″1280″ ...>
    <BaseURL>e5.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <!-viewport 2 video -->
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”30”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”’v’, 1, 2000,
    100, 1080, 1920, 4096, 2048”/>
    <Representation id=″6″ bandwidth=″2000000″ width=″720″ height=″1280″...>
    <BaseURL>e6.mp4</BaseURL>
    </Representation>
    <Representation id=″7″ bandwidth=″4000000″ width=″720″ height=″1280″...>
    <BaseURL>e7.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <AdaptationSet mimeType=”video/mp4” width=″4096″ height=″2048″ framerate=”60”>
    < Role schemeIdUri=”urn:mpeg:dash:viewport:overlay:2016” value=”’v’, 1, 2000,
    100, 1080, 1920, 4096, 2048”/>
    <Representation id=″8″ bandwidth=″6000000″ width=″1080″ height=″1920″...>
    <BaseURL>e8.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • The layer-based 360 video overlay may be provided by one or more descriptors such as SupplementalProperty and/or EssentialProperty with @schemeIdUri equal to “urn:mpeg:dash:viewport:overlay:2016”.
  • Different video Representations may be in different projection formats (e.g., one Representation is equirectangular whereas another representation is cube-map or one video Representation is spherical video while another video Representation is rectilinear). The video property may be specified from the OMAF file format. One or more Common attributes specified for AdaptationSet, Representation, and/or Sub-Representation, @videoType, may indicate the video type such as spherical video, light field video, and/or rectilinear video. For Spherical video, different projection formats and/or projection+layout combinations, such as equirectangular, cube-map (e.g., in combination with different layouts in FIG. 10) and/or Pyramid map may be signaled as Common attribute @projection and/or @layout for AdaptationSet, Representation, and/or Sub-Representation elements. The element and attributes may be provided using SupplementalProperty and/or EssentialProperty elements.
  • Table 15 is a semantic example of a video type and/or a projection attribute. The projection attribute may assist the user to request the appropriate video based on the client's capability. For example, a client that does not support spherical video and/or light field video may request a rectilinear video only. In an example, a client that supports spherical video may select an appropriate format (e.g., equirectangular instead of cube-map) from the set of available formats.
  • TABLE 15
    Common Adaptation Set, Representation, and/or
    Sub-Representation attributes and elements
    Element and/or
    Attribute Name
    Common attributes
    and elements Description
    @videoType May specify the video type in the form of
    an integer as specified in Table 16. If not
    present, the value may be inferred to be
    “rectilinear”
    @projection May specify the projection method used to
    project omnidirectional video into rectilinear
    video frame as specified in Table 17. If
    @videoType is equal to 2, the value of
    @projection may be specified in Table 17
    @layout May specify the layout of projected
    omnidirectional video in the form of an
    integer. The value of @layout may be
    specified in Table 18
  • TABLE 16
    Value of @videoType
    Value Geometry Type
    0 Rectilinear video
    1 Panorama video
    2 Sphere video
    3 Lightfield video
    . . . . . .
  • TABLE 17
    Value of @projection
    Value Geometry Type
    0 Equirectangular
    1 Cube
    2 Offset cube
    3 Squished sphere
    4 Pyramid
    5 Cylinder
    . . . . . .
  • TABLE 18
    Value of @layout
    @projection @layout Layout format
    0 0 Plate carree
    1 Poles on the side
    (half height)
    2 Poles on the side
    (full height)
    1 3 Single row
    4 2 × 3
    5 180°
    . . . . . . . . .
  • The layout formats in Table 18 may be illustrated in the following figures.
  • FIG. 19 depicts an example equirectangular representation with half height poles on the side.
  • FIG. 20 depicts an example equirectangular representation with full height poles on the side.
  • FIG. 21 depicts an example single row layout for the cube representation format (e.g., with region_id).
  • FIG. 22 depicts an example 2×3 layout for the cube representation format (e.g., with region_id).
  • FIG. 23 depicts an example 180° layout for cube-map.
  • The 180° layout for the cube map projection may reduce the resolution of the back half of the cube by 25% (e.g., by reducing the width and height of the back half areas by half) and/or may use the layout shown in FIG. 23 for the result, where regions a, b, c, d, and e belong to the front half and regions f, g, h, i, and j belong to the back half.
  • Table 19 is an MPD example using such common attributes and elements to support both equirectangular and cube-map projections.
  • TABLE 19
    MPD example with Common attribute
    <?xml version=“1.0” encoding=“UTF-8”?>
    <MPD
    xmlns=“urn:mpeg:dash:schema:mpd:2011”
    type=“static”
    mediaPresentationDuration=“PT10S”
    minBufferTime=“PT1S”
    profiles=“urn:mpeg:dash:profile:isoff-on-demand:2011”>
    <ProgramInformation>
    <Title>Example of a DASH Media Presentation Description using @projection and
    @layout attributes to indicate viewports of a 360 video</Title>
    </ProgramInformation>
    <Period>
    <!-- Equirectangular Video -->
    <AdaptationSet segmentAlignment=“true” subsegmentAlignment=“true”
    subsegmentStartsWithSAP=“1” videoType=“2” projection=“0” layout=“0”>
    <Representation mimeType=“video/mp4” codecs=“avc1.42c01e” width=“1920”
    height=“1080” bandwidth=“226597” startWithSAP=“1”>
    <BaseURL> e1.mp4</BaseURL>
    </Representation>
    <Representation mimeType=“video/mp4” codecs=“avc1.42c033” width=“4096”
    height=“2048” bandwidth=“1055223” startWithSAP=“1”>
    <BaseURL> e2.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    <!-- Cubemap 2x3 Video -->
    <AdaptationSet segmentAlignment=“true” subsegmentAlignment=“true”
    subsegmentStartsWithSAP=“1” videoType=“2” projection=“1” layout=“1”>
    <Representation mimeType=“video/mp4” codecs=“avc1.42c01e” width=“1920”
    height=“1080” bandwidth=“226597” startWithSAP=“1”>
    <BaseURL> c1.mp4</BaseURL>
    </Representation>
    <Representation mimeType=“video/mp4” codecs=“avc1.42c033” width=“4096”
    height=“2048” bandwidth=“1055223” startWithSAP=“1”>
    <BaseURL> c2.mp4</BaseURL>
    </Representation>
    </AdaptationSet>
    </Period>
    </MPD>
  • FIG. 24A is a diagram of an example communications system 100 in which one or more disclosed embodiments may be implemented. The communications system 100 may be a multiple access system that provides content, such as voice, data, video, messaging, broadcast, etc., to multiple wireless users. The communications system 100 may enable multiple wireless users to access such content through the sharing of system resources, including wireless bandwidth. For example, the communications systems 100 may employ one or more channel access methods, such as code division multiple access (CDMA), time division multiple access (TDMA), frequency division multiple access (FDMA), orthogonal FDMA (OFDMA), single-carrier FDMA (SC-FDMA), and the like.
  • As shown in FIG. 24A, the communications system 100 may include wireless transmit/receive units (WTRUs) 102 a, 102 b, 102 c, and/or 102 d (which generally or collectively may be referred to as WTRU 102), a radio access network (RAN) 103/104/105, a core network 106/107/109, a public switched telephone network (PSTN) 108, the Internet 110, and other networks 112, though it will be appreciated that the disclosed embodiments contemplate any number of WTRUs, base stations, networks, and/or network elements. Each of the WTRUs 102 a, 102 b, 102 c, 102 d may be any type of device configured to operate and/or communicate in a wireless environment. By way of example, the WTRUs 102 a, 102 b, 102 c, 102 d may be configured to transmit and/or receive wireless signals and may include user equipment (UE), a mobile station, a fixed or mobile subscriber unit, a pager, a cellular telephone, a personal digital assistant (PDA), a smartphone, a laptop, a netbook, a personal computer, a wireless sensor, consumer electronics, and the like.
  • The communications systems 100 may also include a base station 114 a and a base station 114 b. Each of the base stations 114 a, 114 b may be any type of device configured to wirelessly interface with at least one of the WTRUs 102 a, 102 b, 102 c, 102 d to facilitate access to one or more communication networks, such as the core network 106/107/109, the Internet 110, and/or the networks 112. By way of example, the base stations 114 a, 114 b may be a base transceiver station (BTS), a Node-B, an eNode B, a Home Node B, a Home eNode B, a site controller, an access point (AP), a wireless router, and the like. While the base stations 114 a, 114 b are each depicted as a single element, it will be appreciated that the base stations 114 a, 114 b may include any number of interconnected base stations and/or network elements.
  • The base station 114 a may be part of the RAN 103/104/105, which may also include other base stations and/or network elements (not shown), such as a base station controller (BSC), a radio network controller (RNC), relay nodes, etc. The base station 114 a and/or the base station 114 b may be configured to transmit and/or receive wireless signals within a particular geographic region, which may be referred to as a cell (not shown). The cell may further be divided into cell sectors. For example, the cell associated with the base station 114 a may be divided into three sectors. Thus, in one embodiment, the base station 114 a may include three transceivers, e.g., one for each sector of the cell. In another embodiment, the base station 114 a may employ multiple-input multiple output (MIMO) technology and, therefore, may utilize multiple transceivers for each sector of the cell.
  • The base stations 114 a, 114 b may communicate with one or more of the WTRUs 102 a, 102 b, 102 c, 102 d over an air interface 115/116/117, which may be any suitable wireless communication link (e.g., radio frequency (RF), microwave, infrared (IR), ultraviolet (UV), visible light, etc.). The air interface 115/116/117 may be established using any suitable radio access technology (RAT).
  • More specifically, as noted above, the communications system 100 may be a multiple access system and may employ one or more channel access schemes, such as CDMA, TDMA, FDMA, OFDMA, SC-FDMA, and the like. For example, the base station 114 a in the RAN 103/104/105 and the WTRUs 102 a, 102 b, 102 c may implement a radio technology such as Universal Mobile Telecommunications System (UMTS) Terrestrial Radio Access (UTRA), which may establish the air interface 115/116/117 using wideband CDMA (WCDMA). WCDMA may include communication protocols such as High-Speed Packet Access (HSPA) and/or Evolved HSPA (HSPA+). HSPA may include High-Speed Downlink Packet Access (HSDPA) and/or High-Speed Uplink Packet Access (HSUPA).
  • In another embodiment, the base station 114 a and the WTRUs 102 a, 102 b, 102 c may implement a radio technology such as Evolved UMTS Terrestrial Radio Access (E-UTRA), which may establish the air interface 115/116/117 using Long Term Evolution (LTE) and/or LTE-Advanced (LTE-A).
  • In other embodiments, the base station 114 a and the WTRUs 102 a, 102 b, 102 c may implement radio technologies such as IEEE 802.16 (e.g., Worldwide Interoperability for Microwave Access (WiMAX)), CDMA2000, CDMA2000 1X, CDMA2000 EV-DO, Interim Standard 2000 (IS-2000), Interim Standard 95 (IS-95), Interim Standard 856 (IS-856), Global System for Mobile communications (GSM), Enhanced Data rates for GSM Evolution (EDGE), GSM EDGE (GERAN), and the like.
  • The base station 114 b in FIG. 24A may be a wireless router, Home Node B, Home eNode B, or access point, for example, and may utilize any suitable RAT for facilitating wireless connectivity in a localized area, such as a place of business, a home, a vehicle, a campus, and the like. In one embodiment, the base station 114 b and the WTRUs 102 c, 102 d may implement a radio technology such as IEEE 802.11 to establish a wireless local area network (WLAN). In another embodiment, the base station 114 b and the WTRUs 102 c, 102 d may implement a radio technology such as IEEE 802.15 to establish a wireless personal area network (WPAN). In yet another embodiment, the base station 114 b and the WTRUs 102 c, 102 d may utilize a cellular-based RAT (e.g., WCDMA, CDMA2000, GSM, LTE, LTE-A, etc.) to establish a picocell or femtocell. As shown in FIG. 24A, the base station 114 b may have a direct connection to the Internet 110. Thus, the base station 114 b may not be required to access the Internet 110 via the core network 106/107/109.
  • The RAN 103/104/105 may be in communication with the core network 106/107/109, which may be any type of network configured to provide voice, data, applications, and/or voice over internet protocol (VoIP) services to one or more of the WTRUs 102 a, 102 b, 102 c, 102 d. For example, the core network 106/107/109 may provide call control, billing services, mobile location-based services, pre-paid calling, Internet connectivity, video distribution, etc., and/or perform high-level security functions, such as user authentication. Although not shown in FIG. 24A, it will be appreciated that the RAN 103/104/105 and/or the core network 106/107/109 may be in direct or indirect communication with other RANs that employ the same RAT as the RAN 103/104/105 or a different RAT. For example, in addition to being connected to the RAN 103/104/105, which may be utilizing an E-UTRA radio technology, the core network 106/107/109 may also be in communication with another RAN (not shown) employing a GSM radio technology.
  • The core network 106/107/109 may also serve as a gateway for the WTRUs 102 a, 102 b, 102 c, 102 d to access the PSTN 108, the Internet 110, and/or other networks 112. The PSTN 108 may include circuit-switched telephone networks that provide plain old telephone service (POTS). The Internet 110 may include a global system of interconnected computer networks and devices that use common communication protocols, such as the transmission control protocol (TCP), user datagram protocol (UDP) and the internet protocol (IP) in the TCP/IP internet protocol suite. The networks 112 may include wired or wireless communications networks owned and/or operated by other service providers. For example, the networks 112 may include another core network connected to one or more RANs, which may employ the same RAT as the RAN 103/104/105 or a different RAT.
  • Some or all of the WTRUs 102 a, 102 b, 102 c, 102 d in the communications system 100 may include multi-mode capabilities, e.g., the WTRUs 102 a, 102 b, 102 c, 102 d may include multiple transceivers for communicating with different wireless networks over different wireless links. For example, the WTRU 102 c shown in FIG. 24A may be configured to communicate with the base station 114 a, which may employ a cellular-based radio technology, and with the base station 114 b, which may employ an IEEE 802 radio technology.
  • FIG. 24B is a system diagram of an example WTRU 102. As shown in FIG. 24B, the WTRU 102 may include a processor 118, a transceiver 120, a transmit/receive element 122, a speaker/microphone 124, a keypad 126, a display/touchpad 128, non-removable memory 130, removable memory 132, a power source 134, a global positioning system (GPS) chipset 136, and other peripherals 138. It will be appreciated that the WTRU 102 may include any sub-combination of the foregoing elements while remaining consistent with an embodiment. Also, embodiments contemplate that the base stations 114 a and 114 b, and/or the nodes that base stations 114 a and 114 b may represent, such as but not limited to transceiver station (BTS), a Node-B, a site controller, an access point (AP), a home node-B, an evolved home node-B (eNodeB), a home evolved node-B (HeNB), a home evolved node-B gateway, and proxy nodes, among others, may include some or all of the elements depicted in FIG. 24B and described herein.
  • The processor 118 may be a general purpose processor, a special purpose processor, a conventional processor, a digital signal processor (DSP), a plurality of microprocessors, one or more microprocessors in association with a DSP core, a controller, a microcontroller, Application Specific Integrated Circuits (ASICs), Field Programmable Gate Array (FPGAs) circuits, any other type of integrated circuit (IC), a state machine, and the like. The processor 118 may perform signal coding, data processing, power control, input/output processing, and/or any other functionality that enables the WTRU 102 to operate in a wireless environment. The processor 118 may be coupled to the transceiver 120, which may be coupled to the transmit/receive element 122. While FIG. 24B depicts the processor 118 and the transceiver 120 as separate components, it will be appreciated that the processor 118 and the transceiver 120 may be integrated together in an electronic package or chip.
  • The transmit/receive element 122 may be configured to transmit signals to, or receive signals from, a base station (e.g., the base station 114 a) over the air interface 115/116/117. For example, in one embodiment, the transmit/receive element 122 may be an antenna configured to transmit and/or receive RF signals. In another embodiment, the transmit/receive element 122 may be an emitter/detector configured to transmit and/or receive IR, UV, or visible light signals, for example. In yet another embodiment, the transmit/receive element 122 may be configured to transmit and receive both RF and light signals. It will be appreciated that the transmit/receive element 122 may be configured to transmit and/or receive any combination of wireless signals.
  • In addition, although the transmit/receive element 122 is depicted in FIG. 24B as a single element, the WTRU 102 may include any number of transmit/receive elements 122. More specifically, the WTRU 102 may employ MIMO technology. Thus, in one embodiment, the WTRU 102 may include two or more transmit/receive elements 122 (e.g., multiple antennas) for transmitting and receiving wireless signals over the air interface 115/116/117.
  • The transceiver 120 may be configured to modulate the signals that are to be transmitted by the transmit/receive element 122 and to demodulate the signals that are received by the transmit/receive element 122. As noted above, the WTRU 102 may have multi-mode capabilities. Thus, the transceiver 120 may include multiple transceivers for enabling the WTRU 102 to communicate via multiple RATs, such as UTRA and IEEE 802.11, for example.
  • The processor 118 of the WTRU 102 may be coupled to, and may receive user input data from, the speaker/microphone 124, the keypad 126, and/or the display/touchpad 128 (e.g., a liquid crystal display (LCD) display unit or organic light-emitting diode (OLED) display unit). The processor 118 may also output user data to the speaker/microphone 124, the keypad 126, and/or the display/touchpad 128. In addition, the processor 118 may access information from, and store data in, any type of suitable memory, such as the non-removable memory 130 and/or the removable memory 132. The non-removable memory 130 may include random-access memory (RAM), read-only memory (ROM), a hard disk, or any other type of memory storage device. The removable memory 132 may include a subscriber identity module (SIM) card, a memory stick, a secure digital (SD) memory card, and the like. In other embodiments, the processor 118 may access information from, and store data in, memory that is not physically located on the WTRU 102, such as on a server or a home computer (not shown).
  • The processor 118 may receive power from the power source 134, and may be configured to distribute and/or control the power to the other components in the WTRU 102. The power source 134 may be any suitable device for powering the WTRU 102. For example, the power source 134 may include one or more dry cell batteries (e.g., nickel-cadmium (NiCd), nickel-zinc (NiZn), nickel metal hydride (NiMH), lithium-ion (Li-ion), etc.), solar cells, fuel cells, and the like.
  • The processor 118 may also be coupled to the GPS chipset 136, which may be configured to provide location information (e.g., longitude and latitude) regarding the current location of the WTRU 102. In addition to, or in lieu of, the information from the GPS chipset 136, the WTRU 102 may receive location information over the air interface 115/116/117 from a base station (e.g., base stations 114 a, 114 b) and/or determine its location based on the timing of the signals being received from two or more nearby base stations. It will be appreciated that the WTRU 102 may acquire location information by way of any suitable location-determination method while remaining consistent with an embodiment.
  • The processor 118 may further be coupled to other peripherals 138, which may include one or more software and/or hardware modules that provide additional features, functionality and/or wired or wireless connectivity. For example, the peripherals 138 may include an accelerometer, an e-compass, a satellite transceiver, a digital camera (for photographs or video), a universal serial bus (USB) port, a vibration device, a television transceiver, a hands-free headset, a Bluetooth® module, a frequency modulated (FM) radio unit, a digital music player, a media player, a video game player module, an Internet browser, and the like.
  • FIG. 24C is a system diagram of the RAN 103 and the core network 106 according to an embodiment. As noted above, the RAN 103 may employ a UTRA radio technology to communicate with the WTRUs 102 a, 102 b, 102 c over the air interface 115. The RAN 103 may also be in communication with the core network 106. As shown in FIG. 24C, the RAN 103 may include Node- Bs 140 a, 140 b, 140 c, which may each include one or more transceivers for communicating with the WTRUs 102 a, 102 b, 102 c over the air interface 115. The Node- Bs 140 a, 140 b, 140 c may each be associated with a particular cell (not shown) within the RAN 103. The RAN 103 may also include RNCs 142 a, 142 b. It will be appreciated that the RAN 103 may include any number of Node-Bs and RNCs while remaining consistent with an embodiment.
  • As shown in FIG. 24C, the Node- Bs 140 a, 140 b may be in communication with the RNC 142 a. Additionally, the Node-B 140 c may be in communication with the RNC 142 b. The Node- Bs 140 a, 140 b, 140 c may communicate with the respective RNCs 142 a, 142 b via an Iub interface. The RNCs 142 a, 142 b may be in communication with one another via an Iur interface. Each of the RNCs 142 a, 142 b may be configured to control the respective Node- Bs 140 a, 140 b, 140 c to which it is connected. In addition, each of the RNCs 142 a, 142 b may be configured to carry out or support other functionality, such as outer loop power control, load control, admission control, packet scheduling, handover control, macrodiversity, security functions, data encryption, and the like.
  • The core network 106 shown in FIG. 24C may include a media gateway (MGW) 144, a mobile switching center (MSC) 146, a serving GPRS support node (SGSN) 148, and/or a gateway GPRS support node (GGSN) 150. While each of the foregoing elements are depicted as part of the core network 106, it will be appreciated that any one of these elements may be owned and/or operated by an entity other than the core network operator.
  • The RNC 142 a in the RAN 103 may be connected to the MSC 146 in the core network 106 via an IuCS interface. The MSC 146 may be connected to the MGW 144. The MSC 146 and the MGW 144 may provide the WTRUs 102 a, 102 b, 102 c with access to circuit-switched networks, such as the PSTN 108, to facilitate communications between the WTRUs 102 a, 102 b, 102 c and traditional land-line communications devices.
  • The RNC 142 a in the RAN 103 may also be connected to the SGSN 148 in the core network 106 via an IuPS interface. The SGSN 148 may be connected to the GGSN 150. The SGSN 148 and the GGSN 150 may provide the WTRUs 102 a, 102 b, 102 c with access to packet-switched networks, such as the Internet 110, to facilitate communications between and the WTRUs 102 a, 102 b, 102 c and IP-enabled devices.
  • As noted above, the core network 106 may also be connected to the networks 112, which may include other wired or wireless networks that are owned and/or operated by other service providers.
  • FIG. 24D is a system diagram of the RAN 104 and the core network 107 according to an embodiment. As noted above, the RAN 104 may employ an E-UTRA radio technology to communicate with the WTRUs 102 a, 102 b, 102 c over the air interface 116. The RAN 104 may also be in communication with the core network 107.
  • The RAN 104 may include eNode- Bs 160 a, 160 b, 160 c, though it will be appreciated that the RAN 104 may include any number of eNode-Bs while remaining consistent with an embodiment. The eNode- Bs 160 a, 160 b, 160 c may each include one or more transceivers for communicating with the WTRUs 102 a, 102 b, 102 c over the air interface 116. In one embodiment, the eNode- Bs 160 a, 160 b, 160 c may implement MIMO technology. Thus, the eNode-B 160 a, for example, may use multiple antennas to transmit wireless signals to, and receive wireless signals from, the WTRU 102 a.
  • Each of the eNode- Bs 160 a, 160 b, 160 c may be associated with a cell (not shown) and may be configured to handle radio resource management decisions, handover decisions, scheduling of users in the uplink and/or downlink, and the like. As shown in FIG. 24D, the eNode- Bs 160 a, 160 b, 160 c may communicate with one another over an X2 interface.
  • The core network 107 shown in FIG. 24D may include a mobility management gateway (MME) 162, a serving gateway 164, and a packet data network (PDN) gateway 166. While each of the foregoing elements are depicted as part of the core network 107, it will be appreciated that any one of these elements may be owned and/or operated by an entity other than the core network operator.
  • The MME 162 may be connected to each of the eNode- Bs 160 a, 160 b, 160 c in the RAN 104 via an S1 interface and may serve as a control node. For example, the MME 162 may be responsible for authenticating users of the WTRUs 102 a, 102 b, 102 c, bearer activation/deactivation, selecting a serving gateway during an initial attach of the WTRUs 102 a, 102 b, 102 c, and the like. The MME 162 may also provide a control plane function for switching between the RAN 104 and other RANs (not shown) that employ other radio technologies, such as GSM or WCDMA.
  • The serving gateway 164 may be connected to each of the eNode- Bs 160 a, 160 b, 160 c in the RAN 104 via the S1 interface. The serving gateway 164 may generally route and forward user data packets to/from the WTRUs 102 a, 102 b, 102 c. The serving gateway 164 may also perform other functions, such as anchoring user planes during inter-eNode B handovers, triggering paging when downlink data is available for the WTRUs 102 a, 102 b, 102 c, managing and storing contexts of the WTRUs 102 a, 102 b, 102 c, and the like.
  • The serving gateway 164 may also be connected to the PDN gateway 166, which may provide the WTRUs 102 a, 102 b, 102 c with access to packet-switched networks, such as the Internet 110, to facilitate communications between the WTRUs 102 a, 102 b, 102 c and IP-enabled devices.
  • The core network 107 may facilitate communications with other networks. For example, the core network 107 may provide the WTRUs 102 a, 102 b, 102 c with access to circuit-switched networks, such as the PSTN 108, to facilitate communications between the WTRUs 102 a, 102 b, 102 c and traditional land-line communications devices. For example, the core network 107 may include, or may communicate with, an IP gateway (e.g., an IP multimedia subsystem (IMS) server) that serves as an interface between the core network 107 and the PSTN 108. In addition, the core network 107 may provide the WTRUs 102 a, 102 b, 102 c with access to the networks 112, which may include other wired or wireless networks that are owned and/or operated by other service providers.
  • FIG. 24E is a system diagram of the RAN 105 and the core network 109 according to an embodiment. The RAN 105 may be an access service network (ASN) that employs IEEE 802.16 radio technology to communicate with the WTRUs 102 a, 102 b, 102 c over the air interface 117. As will be further discussed below, the communication links between the different functional entities of the WTRUs 102 a, 102 b, 102 c, the RAN 105, and the core network 109 may be defined as reference points.
  • As shown in FIG. 24E, the RAN 105 may include base stations 180 a, 180 b, 180 c, and an ASN gateway 182, though it will be appreciated that the RAN 105 may include any number of base stations and ASN gateways while remaining consistent with an embodiment. The base stations 180 a, 180 b, 180 c may each be associated with a cell (not shown) in the RAN 105 and may each include one or more transceivers for communicating with the WTRUs 102 a, 102 b, 102 c over the air interface 117. In one embodiment, the base stations 180 a, 180 b, 180 c may implement MIMO technology. Thus, the base station 180 a, for example, may use multiple antennas to transmit wireless signals to, and receive wireless signals from, the WTRU 102 a. The base stations 180 a, 180 b, 180 c may also provide mobility management functions, such as handoff triggering, tunnel establishment, radio resource management, traffic classification, quality of service (QoS) policy enforcement, and the like. The ASN gateway 182 may serve as a traffic aggregation point and may be responsible for paging, caching of subscriber profiles, routing to the core network 109, and the like.
  • The air interface 117 between the WTRUs 102 a, 102 b, 102 c and the RAN 105 may be defined as an R1 reference point that implements the IEEE 802.16 specification. In addition, each of the WTRUs 102 a, 102 b, 102 c may establish a logical interface (not shown) with the core network 109. The logical interface between the WTRUs 102 a, 102 b, 102 c and the core network 109 may be defined as an R2 reference point, which may be used for authentication, authorization, IP host configuration management, and/or mobility management.
  • The communication link between each of the base stations 180 a, 180 b, 180 c may be defined as an R8 reference point that includes protocols for facilitating WTRU handovers and the transfer of data between base stations. The communication link between the base stations 180 a, 180 b, 180 c and the ASN gateway 182 may be defined as an R6 reference point. The R6 reference point may include protocols for facilitating mobility management based on mobility events associated with each of the WTRUs 102 a, 102 b, 102 c.
  • As shown in FIG. 24E, the RAN 105 may be connected to the core network 109. The communication link between the RAN 105 and the core network 109 may defined as an R3 reference point that includes protocols for facilitating data transfer and mobility management capabilities, for example. The core network 109 may include a mobile IP home agent (MIP-HA) 184, an authentication, authorization, accounting (AAA) server 186, and a gateway 188. While each of the foregoing elements are depicted as part of the core network 109, it will be appreciated that any one of these elements may be owned and/or operated by an entity other than the core network operator.
  • The MIP-HA may be responsible for IP address management, and may enable the WTRUs 102 a, 102 b, 102 c to roam between different ASNs and/or different core networks. The MIP-HA 184 may provide the WTRUs 102 a, 102 b, 102 c with access to packet-switched networks, such as the Internet 110, to facilitate communications between the WTRUs 102 a, 102 b, 102 c and IP-enabled devices. The AAA server 186 may be responsible for user authentication and for supporting user services. The gateway 188 may facilitate interworking with other networks. For example, the gateway 188 may provide the WTRUs 102 a, 102 b, 102 c with access to circuit-switched networks, such as the PSTN 108, to facilitate communications between the WTRUs 102 a, 102 b, 102 c and traditional land-line communications devices. In addition, the gateway 188 may provide the WTRUs 102 a, 102 b, 102 c with access to the networks 112, which may include other wired or wireless networks that are owned and/or operated by other service providers.
  • Each of the computing systems described herein may have one or more computer processors having memory that are configured with executable instructions or hardware for accomplishing the functions described herein including determining the parameters described herein and sending and receiving messages between entities (e.g., WTRU and network or client and server) to accomplish the described functions. The processes described above may be implemented in a computer program, software, and/or firmware incorporated in a computer-readable medium for execution by a computer and/or processor.
  • Although not shown in FIG. 24E, it will be appreciated that the RAN 105 may be connected to other ASNs and the core network 109 may be connected to other core networks. The communication link between the RAN 105 the other ASNs may be defined as an R4 reference point, which may include protocols for coordinating the mobility of the WTRUs 102 a, 102 b, 102 c between the RAN 105 and the other ASNs. The communication link between the core network 109 and the other core networks may be defined as an R5 reference, which may include protocols for facilitating interworking between home core networks and visited core networks.
  • Although features and elements are described above in particular combinations, one of ordinary skill in the art will appreciate that each feature or element can be used alone or in any combination with the other features and elements. In addition, the methods described herein may be implemented in a computer program, software, or firmware incorporated in a computer-readable medium for execution by a computer or processor. Examples of computer-readable media include electronic signals (transmitted over wired or wireless connections) and computer-readable storage media. Examples of computer-readable storage media include, but are not limited to, a read only memory (ROM), a random access memory (RAM), a register, cache memory, semiconductor memory devices, magnetic media such as internal hard disks and removable disks, magneto-optical media, and optical media such as CD-ROM disks, and digital versatile disks (DVDs). A processor in association with software may be used to implement a radio frequency transceiver for use in a WTRU, WTRU, terminal, base station, RNC, or any host computer.

Claims (13)

1-24. (canceled)
25. An apparatus comprising:
a processor configured to:
receive a media presentation description (MPD) file comprising a plurality of representations associated with a 360 degree video, the plurality of representations comprising a first representation and a second representation, wherein the first representation is associated with a first video layer for the 360 degree video at a first quality and wherein the second representation is associated with a second video layer for one or more spatial subsets of the 360 degree video at a second quality;
request the first video layer in the first representation based on the received MPD file;
identify a viewport of a user that is expected to be viewed;
request the second video layer that corresponds to a spatial subset of the 360 degree video in the second representation based on the identified viewport of the user and the received MPD file; and
render a current viewport of the user, the current viewport comprising the second video layer in the second representation overlaid onto the first video layer in the first representation, wherein the second quality has higher quality than the first quality.
26. The apparatus of claim 25, wherein the first video layer is a full-frame video layer associated with the 360 degree video, and the second video layer is the one or more spatial subsets of the 360 degree video.
27. The apparatus of claim 25, wherein the 360 degree video comprises a scalable 360 degree video.
28. The apparatus of claim 25, wherein the first video layer is a scalable base layer associated with the 360 degree video, and the second video layer is a scalable enhancement layer associated with the 360 degree video.
29. The apparatus of claim 25, wherein the second video layer is coded with one or more of a higher bitrate, a higher resolution, or a higher frame rate than the first video layer.
30. The apparatus of claim 25, wherein the second video layer is coded independently from the first video layer.
31. A method comprising:
receiving a media presentation description (MPD) file comprising a plurality of representations associated with a 360 degree video, the plurality of representations comprising a first representation and a second representation, wherein the first representation is associated with a first video layer for the 360 degree video at a first quality and wherein the second representation is associated with a second video layer for one or more spatial subsets of the 360 degree video at a second quality;
requesting the first video layer in the first representation based on the received MPD file;
identifying a viewport of a user that is expected to be viewed;
requesting the second video layer that corresponds to a spatial subset of the 360 degree video in the second representation based on the identified viewport of the user and the received MPD file; and
rendering a current viewport of the user, the current viewport comprising the second video layer in the second representation at the second quality overlaid onto the first video layer in the first representation at the first quality, wherein the second quality has higher quality than the first quality.
32. The method of claim 31, wherein the first video layer is a full-frame video layer associated with the 360 degree video, and the second video layer is the one or more spatial subsets of the 360 degree video.
33. The method of claim 31, wherein the 360 degree video comprises a scalable 360 degree video.
34. The method of claim 31, wherein the first video layer is a scalable base layer associated with the 360 degree video, and the second video layer is a scalable enhancement layer associated with the 360 degree video.
35. The method of claim 31, wherein the second video layer is coded with one or more of a higher bitrate, a higher resolution, or a higher frame rate than the first video layer.
36. The method of claim 31, wherein the second video layer is coded independently from the first video layer.
US17/036,578 2016-05-26 2020-09-29 Methods and apparatus of viewport adaptive 360 degree video delivery Pending US20210014472A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/036,578 US20210014472A1 (en) 2016-05-26 2020-09-29 Methods and apparatus of viewport adaptive 360 degree video delivery

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201662342158P 2016-05-26 2016-05-26
PCT/US2017/034770 WO2017205794A1 (en) 2016-05-26 2017-05-26 Methods and apparatus of viewport adaptive 360 degree video delivery
US201816304443A 2018-11-26 2018-11-26
US17/036,578 US20210014472A1 (en) 2016-05-26 2020-09-29 Methods and apparatus of viewport adaptive 360 degree video delivery

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US16/304,443 Continuation US10841566B2 (en) 2016-05-26 2017-05-26 Methods and apparatus of viewport adaptive 360 degree video delivery
PCT/US2017/034770 Continuation WO2017205794A1 (en) 2016-05-26 2017-05-26 Methods and apparatus of viewport adaptive 360 degree video delivery

Publications (1)

Publication Number Publication Date
US20210014472A1 true US20210014472A1 (en) 2021-01-14

Family

ID=59071081

Family Applications (2)

Application Number Title Priority Date Filing Date
US16/304,443 Active US10841566B2 (en) 2016-05-26 2017-05-26 Methods and apparatus of viewport adaptive 360 degree video delivery
US17/036,578 Pending US20210014472A1 (en) 2016-05-26 2020-09-29 Methods and apparatus of viewport adaptive 360 degree video delivery

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US16/304,443 Active US10841566B2 (en) 2016-05-26 2017-05-26 Methods and apparatus of viewport adaptive 360 degree video delivery

Country Status (6)

Country Link
US (2) US10841566B2 (en)
EP (2) EP3466076A1 (en)
JP (4) JP2019521587A (en)
KR (4) KR20190021229A (en)
CN (2) CN117336499A (en)
WO (1) WO2017205794A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11190801B2 (en) 2019-07-02 2021-11-30 Mediatek Inc. Video encoding method with syntax element signaling of mapping function employed by cube-based projection and associated video decoding method
US11190768B2 (en) * 2019-07-02 2021-11-30 Mediatek Inc. Video encoding method with syntax element signaling of packing of projection faces derived from cube-based projection and associated video decoding method and apparatus
US11659206B2 (en) 2019-07-02 2023-05-23 Mediatek Inc. Video encoding method with syntax element signaling of guard band configuration of projection-based frame and associated video decoding method and apparatus

Families Citing this family (84)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016056411A1 (en) * 2014-10-10 2016-04-14 ソニー株式会社 Encoding device and method, playback device and method, and program
WO2018011054A1 (en) 2016-07-15 2018-01-18 Koninklijke Kpn N.V. Streaming virtual reality video
EP3501014A1 (en) 2016-08-17 2019-06-26 VID SCALE, Inc. Secondary content insertion in 360-degree video
EP4084480A1 (en) * 2016-09-09 2022-11-02 Vid Scale, Inc. Methods and apparatus to reduce latency for 360-degree viewport adaptive streaming
US20190362151A1 (en) * 2016-09-14 2019-11-28 Koninklijke Kpn N.V. Streaming virtual reality video
CN109644296A (en) * 2016-10-10 2019-04-16 华为技术有限公司 A kind of video stream transmission method, relevant device and system
GB2554877B (en) * 2016-10-10 2021-03-31 Canon Kk Methods, devices, and computer programs for improving rendering display during streaming of timed media data
EP3535644B1 (en) 2016-11-04 2023-02-22 Koninklijke KPN N.V. Streaming virtual reality video
WO2018124820A1 (en) * 2017-01-02 2018-07-05 주식회사 케이티 Method and apparatus for processing video signals
WO2018147329A1 (en) * 2017-02-10 2018-08-16 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ Free-viewpoint image generation method and free-viewpoint image generation system
KR102277267B1 (en) * 2017-03-29 2021-07-14 엘지전자 주식회사 How to send 360 video, how to receive 360 video, 360 video sending device, 360 video receiving device
KR20200014281A (en) * 2017-05-30 2020-02-10 소니 주식회사 Image processing apparatus and method, file generating apparatus and method, and program
WO2018222996A1 (en) 2017-06-02 2018-12-06 Vid Scale, Inc. 360-degree video delivery over next generation network
GB2563275B (en) * 2017-06-09 2019-10-09 Sony Interactive Entertainment Inc Image processing device and system
GB2563387B (en) * 2017-06-09 2020-04-15 Sony Interactive Entertainment Inc Image processing device and system
DE112018002432T5 (en) * 2017-08-18 2020-01-23 Mediatek Inc. Method and device for reducing artifacts in a projection-based frame
WO2019037558A1 (en) * 2017-08-22 2019-02-28 优酷网络技术(北京)有限公司 Image processing method and apparatus
US10827159B2 (en) * 2017-08-23 2020-11-03 Mediatek Inc. Method and apparatus of signalling syntax for immersive video coding
US10818087B2 (en) 2017-10-02 2020-10-27 At&T Intellectual Property I, L.P. Selective streaming of immersive video based on field-of-view prediction
US10217488B1 (en) 2017-12-15 2019-02-26 Snap Inc. Spherical video editing
EP3503559A1 (en) * 2017-12-20 2019-06-26 Thomson Licensing Method and terminal for managing a streaming session of an immersive video spatially tiled with a set of tiles and stored on a network equipment
JP2019118026A (en) * 2017-12-27 2019-07-18 キヤノン株式会社 Information processing device, information processing method, and program
CN108111899B (en) * 2017-12-29 2021-05-11 中兴通讯股份有限公司 Video transmission method and device, client, server and storage medium
JPWO2019138927A1 (en) * 2018-01-12 2020-12-24 ソニー株式会社 Information processing equipment and methods
WO2019141907A1 (en) * 2018-01-22 2019-07-25 Nokia Technologies Oy An apparatus, a method and a computer program for omnidirectional video
US10659815B2 (en) 2018-03-08 2020-05-19 At&T Intellectual Property I, L.P. Method of dynamic adaptive streaming for 360-degree videos
JP6859286B2 (en) * 2018-03-26 2021-04-14 Kddi株式会社 VR video distribution device and method, VR video playback device and method, and VR video system
CN110351607B (en) * 2018-04-04 2022-01-14 阿里巴巴(中国)有限公司 Method for switching panoramic video scenes, computer storage medium and client
US11140373B2 (en) 2018-04-05 2021-10-05 Lg Electronics Inc. Method for transmitting 360-degree video, method for receiving 360-degree video, apparatus for transmitting 360-degree video, and apparatus for receiving 360-degree video
CN111869221B (en) * 2018-04-05 2021-07-20 华为技术有限公司 Efficient association between DASH objects
US10812828B2 (en) 2018-04-10 2020-10-20 At&T Intellectual Property I, L.P. System and method for segmenting immersive video
GB2573593B (en) * 2018-05-10 2021-11-03 Samsung Electronics Co Ltd Augmented reality rendering method and apparatus
US11917127B2 (en) 2018-05-25 2024-02-27 Interdigital Madison Patent Holdings, Sas Monitoring of video streaming events
US10623791B2 (en) 2018-06-01 2020-04-14 At&T Intellectual Property I, L.P. Field of view prediction in live panoramic video streaming
KR102154530B1 (en) * 2018-06-06 2020-09-21 엘지전자 주식회사 Method and apparatus for processing overaly meaid in 360 degree video system
US10812774B2 (en) 2018-06-06 2020-10-20 At&T Intellectual Property I, L.P. Methods and devices for adapting the rate of video content streaming
US10573060B1 (en) * 2018-06-14 2020-02-25 Kilburn Live, Llc Controller binding in virtual domes
US10740957B1 (en) * 2018-06-14 2020-08-11 Kilburn Live, Llc Dynamic split screen
US10419738B1 (en) 2018-06-14 2019-09-17 Telefonaktiebolaget Lm Ericsson (Publ) System and method for providing 360° immersive video based on gaze vector information
US10616621B2 (en) 2018-06-29 2020-04-07 At&T Intellectual Property I, L.P. Methods and devices for determining multipath routing for panoramic video content
US10841662B2 (en) 2018-07-27 2020-11-17 Telefonaktiebolaget Lm Ericsson (Publ) System and method for inserting advertisement content in 360° immersive video
BR112021001958A2 (en) * 2018-08-03 2021-04-27 V-Nova International Limited transformations for signal enhancement encoding
FR3084980A1 (en) 2018-08-10 2020-02-14 Orange 360 DEGREE VIDEO BROADCASTING METHOD AND DEVICE
US10708494B2 (en) 2018-08-13 2020-07-07 At&T Intellectual Property I, L.P. Methods, systems and devices for adjusting panoramic video content
US11019361B2 (en) 2018-08-13 2021-05-25 At&T Intellectual Property I, L.P. Methods, systems and devices for adjusting panoramic view of a camera for capturing video content
US11032590B2 (en) 2018-08-31 2021-06-08 At&T Intellectual Property I, L.P. Methods, devices, and systems for providing panoramic video content to a mobile device from an edge server
US10999583B2 (en) 2018-09-14 2021-05-04 Apple Inc. Scalability of multi-directional video streaming
JP2022500931A (en) * 2018-09-14 2022-01-04 華為技術有限公司Huawei Technologies Co., Ltd. Improved attribute layers and signaling in point cloud coding
CN112771878B (en) * 2018-09-27 2023-01-13 华为技术有限公司 Method, client and server for processing media data
CN116980967A (en) 2018-09-28 2023-10-31 苹果公司 Electronic device content provision adjustment based on wireless communication channel bandwidth conditions
US11818419B2 (en) 2018-09-28 2023-11-14 Apple Inc. Mobile device content provisioning adjustments based on wireless communication channel bandwidth condition
CN113206826B (en) * 2018-09-28 2022-10-04 华为技术有限公司 Method, client and server for transmitting media data
US10757389B2 (en) 2018-10-01 2020-08-25 Telefonaktiebolaget Lm Ericsson (Publ) Client optimization for providing quality control in 360° immersive video during pause
EP3644619A1 (en) * 2018-10-23 2020-04-29 InterDigital CE Patent Holdings Method and apparatus for receiving a tile-based immersive video
US10939139B2 (en) 2018-11-29 2021-03-02 Apple Inc. Adaptive coding and streaming of multi-directional video
KR20210094646A (en) * 2018-12-14 2021-07-29 지티이 코포레이션 Immersive video bitstream processing
EP3905532A4 (en) * 2018-12-28 2022-11-30 Sony Group Corporation Information processing device and information processing method
CN114979750A (en) * 2019-01-08 2022-08-30 诺基亚技术有限公司 Method, apparatus and computer readable medium for virtual reality
US20220150464A1 (en) * 2019-03-08 2022-05-12 Sony Group Corporation Image processing apparatus, image processing method, and image processing program
US11991402B2 (en) 2019-03-26 2024-05-21 Interdigital Vc Holdings, Inc. System and method for multiplexed rendering of light fields
KR102178111B1 (en) * 2019-04-25 2020-11-12 주식회사 엘지유플러스 Apparatus and method for providing video streaming service
US20220217314A1 (en) * 2019-05-24 2022-07-07 Lg Electronics Inc. Method for transmitting 360 video, method for receiving 360 video, 360 video transmitting device, and 360 video receiving device
US11523185B2 (en) 2019-06-19 2022-12-06 Koninklijke Kpn N.V. Rendering video stream in sub-area of visible display area
US11159823B2 (en) * 2019-06-20 2021-10-26 At&T Intellectual Property I, L.P. Multi-viewport transcoding for volumetric video streaming
EP3777223A4 (en) * 2019-06-25 2021-08-25 Beijing Xiaomi Mobile Software Co., Ltd. Omnidirectional media playback method and device and computer readable storage medium thereof
WO2021000276A1 (en) * 2019-07-03 2021-01-07 Beijing Xiaomi Mobile Software Co., Ltd. Method, system and apparatus for building virtual reality envioronment
US11343567B1 (en) * 2019-08-07 2022-05-24 Meta Platforms, Inc. Systems and methods for providing a quality metric for media content
US11831861B2 (en) * 2019-08-12 2023-11-28 Intel Corporation Methods for viewport-dependent adaptive streaming of point cloud content
US11956295B2 (en) 2019-09-27 2024-04-09 Apple Inc. Client-end enhanced view prediction for multi-view video streaming exploiting pre-fetched data and side information
WO2021151761A1 (en) * 2020-01-29 2021-08-05 Nokia Technologies Oy A method, an apparatus and a computer program product for video streaming
JP7439556B2 (en) * 2020-02-14 2024-02-28 オムロン株式会社 Image processing device, image sensor, and method for controlling the image processing device
KR20210107409A (en) 2020-02-24 2021-09-01 삼성전자주식회사 Method and apparatus for transmitting video content using edge computing service
CN115225937B (en) * 2020-03-24 2023-12-01 腾讯科技(深圳)有限公司 Immersive media providing method, acquisition method, device, equipment and storage medium
US11503289B2 (en) * 2020-05-27 2022-11-15 Tencent America LLC Bitstream structure for viewport-based streaming with a fallback bitstream
KR102655908B1 (en) * 2020-07-07 2024-04-11 인하대학교 산학협력단 Method and device for building a stackable light field-based virtual space
KR20220037617A (en) * 2020-09-18 2022-03-25 삼성전자주식회사 Method and apparatus for video processing using edge computing service
US20230300309A1 (en) * 2020-09-23 2023-09-21 Sony Group Corporation Information processing device, information processing method, and information processing system
CN112752158B (en) * 2020-12-29 2023-06-20 北京达佳互联信息技术有限公司 Video display method and device, electronic equipment and storage medium
US11800184B2 (en) 2021-01-06 2023-10-24 Tencent America LLC Method and apparatus for media scene description
WO2022234699A1 (en) * 2021-05-07 2022-11-10 キヤノン株式会社 Moving image delivery device, display control device, moving image delivery system, moving image delivery method, and program
US11570417B2 (en) 2021-05-20 2023-01-31 Apple Inc. Immersive video streaming using view-adaptive prefetching and buffer control
US20230224557A1 (en) * 2022-01-12 2023-07-13 Tencent America LLC Auxiliary mpds for mpeg dash to support prerolls, midrolls and endrolls with stacking properties
US11893699B2 (en) 2022-03-15 2024-02-06 Zeality Inc Method and processing unit for providing content in a bandwidth constrained environment
WO2024123915A1 (en) * 2022-12-08 2024-06-13 Dolby Laboratories Licensing Corporation Offset low discrepancy spherical sampling for image rendering

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160150212A1 (en) * 2014-11-26 2016-05-26 Sony Corporation Live selective adaptive bandwidth
US20160260196A1 (en) * 2015-03-05 2016-09-08 Nokia Technologies Oy Video streaming method
US20170347026A1 (en) * 2016-05-24 2017-11-30 Nokia Technologies Oy Method and an apparatus and a computer program for encoding media content

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005142654A (en) 2003-11-04 2005-06-02 Matsushita Electric Ind Co Ltd Video transmitting apparatus and video receiving apparatus
JP2010212811A (en) 2009-03-06 2010-09-24 Panasonic Corp Moving image encoding device and moving image decoding device
US9232257B2 (en) * 2010-09-22 2016-01-05 Thomson Licensing Method for navigation in a panoramic scene
CN102834849B (en) * 2011-03-31 2016-08-31 松下知识产权经营株式会社 Carry out the image displaying device of the description of three-dimensional view picture, image drawing method, image depiction program
CN104067628B (en) 2012-01-19 2018-12-04 Vid拓展公司 For supporting the method and system of the transmission of video adaptive to viewing condition
KR102283241B1 (en) * 2012-07-10 2021-07-29 브이아이디 스케일, 인크. Quality-driven streaming
GB2509954B (en) 2013-01-18 2016-03-23 Canon Kk Method of displaying a region of interest in a video stream
WO2015009676A1 (en) 2013-07-15 2015-01-22 Sony Corporation Extensions of motion-constrained tile sets sei message for interactivity
US10306273B2 (en) 2013-07-19 2019-05-28 Sony Corporation Information processing device and method for generating partial image information including group identification information
CN106233745B (en) 2013-07-29 2021-01-15 皇家Kpn公司 Providing tile video streams to clients
GB2558086B (en) * 2014-03-25 2019-02-20 Canon Kk Methods, devices, and computer programs for improving streaming of partitioned timed media data
EP3162074A1 (en) * 2014-06-27 2017-05-03 Koninklijke KPN N.V. Determining a region of interest on the basis of a hevc-tiled video stream
US10979691B2 (en) * 2016-05-20 2021-04-13 Qualcomm Incorporated Circular fisheye video in virtual reality
EP3249929A1 (en) * 2016-05-25 2017-11-29 Thomson Licensing Method and network equipment for establishing a manifest

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160150212A1 (en) * 2014-11-26 2016-05-26 Sony Corporation Live selective adaptive bandwidth
US20160260196A1 (en) * 2015-03-05 2016-09-08 Nokia Technologies Oy Video streaming method
US20170347026A1 (en) * 2016-05-24 2017-11-30 Nokia Technologies Oy Method and an apparatus and a computer program for encoding media content

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CORBILLON, XAVIER et al., "Viewport-Adaptive Navigable 360-Degree Video Delivery", Telecom Bretagne, France, September 26, 2016, 7 pages *
Martens, Geoffrey "Bandwidth management for ODV tiled streaming with MPEG-DASH" Masters thesis, Universiteit Hasselt, 01 January 2015 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11190801B2 (en) 2019-07-02 2021-11-30 Mediatek Inc. Video encoding method with syntax element signaling of mapping function employed by cube-based projection and associated video decoding method
US11190768B2 (en) * 2019-07-02 2021-11-30 Mediatek Inc. Video encoding method with syntax element signaling of packing of projection faces derived from cube-based projection and associated video decoding method and apparatus
US11659206B2 (en) 2019-07-02 2023-05-23 Mediatek Inc. Video encoding method with syntax element signaling of guard band configuration of projection-based frame and associated video decoding method and apparatus

Also Published As

Publication number Publication date
US10841566B2 (en) 2020-11-17
WO2017205794A1 (en) 2017-11-30
JP7288875B2 (en) 2023-06-08
KR20210027543A (en) 2021-03-10
EP3761645A1 (en) 2021-01-06
KR102173635B1 (en) 2020-11-03
CN109511284A (en) 2019-03-22
KR20190021229A (en) 2019-03-05
JP2019521587A (en) 2019-07-25
CN109511284B (en) 2023-09-01
JP2020110005A (en) 2020-07-16
JP2023116547A (en) 2023-08-22
KR20200084904A (en) 2020-07-13
US20190158815A1 (en) 2019-05-23
WO2017205794A9 (en) 2019-03-07
JP7519503B2 (en) 2024-07-19
KR102277287B1 (en) 2021-07-15
CN117336499A (en) 2024-01-02
JP2024138375A (en) 2024-10-08
EP3466076A1 (en) 2019-04-10
KR20200127044A (en) 2020-11-09
KR102224826B1 (en) 2021-03-09

Similar Documents

Publication Publication Date Title
US20210014472A1 (en) Methods and apparatus of viewport adaptive 360 degree video delivery
US12022117B2 (en) Apparatus, a method and a computer program for video coding and decoding
US10893256B2 (en) Apparatus, a method and a computer program for omnidirectional video
RU2741507C1 (en) Device and method for video encoding and decoding
US12108125B2 (en) 360-degree video delivery over next generation network
US20220159267A1 (en) An apparatus, a method and a computer program for video coding and decoding
KR20190050817A (en) Information processing device and method of controlling same, and computer program
US20240022773A1 (en) Mmt signaling for streaming of visual volumetric video-based and geometry-based point cloud media

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

AS Assignment

Owner name: INTERDIGITAL VC HOLDINGS, INC., DELAWARE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VID SCALE, INC.;REEL/FRAME:068284/0031

Effective date: 20240627

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION