US20200177956A1 - Method and apparatus for content adaptation based on audience monitoring - Google Patents
Method and apparatus for content adaptation based on audience monitoring Download PDFInfo
- Publication number
- US20200177956A1 US20200177956A1 US16/733,399 US202016733399A US2020177956A1 US 20200177956 A1 US20200177956 A1 US 20200177956A1 US 202016733399 A US202016733399 A US 202016733399A US 2020177956 A1 US2020177956 A1 US 2020177956A1
- Authority
- US
- United States
- Prior art keywords
- content
- segment
- audience
- context
- bandwidth
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims description 46
- 230000006978 adaptation Effects 0.000 title description 8
- 238000012544 monitoring process Methods 0.000 title description 7
- 238000006243 chemical reaction Methods 0.000 claims abstract description 142
- 238000012545 processing Methods 0.000 claims description 28
- 238000003909 pattern recognition Methods 0.000 claims description 21
- 230000008859 change Effects 0.000 claims description 15
- 230000015654 memory Effects 0.000 claims description 13
- 230000003247 decreasing effect Effects 0.000 claims description 10
- 230000007423 decrease Effects 0.000 claims description 5
- 230000006854 communication Effects 0.000 description 57
- 238000004891 communication Methods 0.000 description 57
- 238000005516 engineering process Methods 0.000 description 26
- 230000006870 function Effects 0.000 description 25
- 230000033001 locomotion Effects 0.000 description 14
- 238000001514 detection method Methods 0.000 description 4
- 230000002452 interceptive effect Effects 0.000 description 4
- 238000003780 insertion Methods 0.000 description 3
- 230000037431 insertion Effects 0.000 description 3
- 230000004044 response Effects 0.000 description 3
- 238000003491 array Methods 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 206010011469 Crying Diseases 0.000 description 1
- 206010039740 Screaming Diseases 0.000 description 1
- 230000007175 bidirectional communication Effects 0.000 description 1
- 230000033228 biological regulation Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 230000007123 defense Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000036651 mood Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/442—Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
- H04N21/44213—Monitoring of end-user related data
- H04N21/44218—Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/234363—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the spatial resolution, e.g. for clients with a lower screen resolution
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/251—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/252—Processing of multiple end-users' preferences to derive collaborative data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/4104—Peripherals receiving signals from specially adapted client devices
- H04N21/4126—The peripheral being portable, e.g. PDAs or mobile phones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/4104—Peripherals receiving signals from specially adapted client devices
- H04N21/4126—The peripheral being portable, e.g. PDAs or mobile phones
- H04N21/41265—The peripheral being portable, e.g. PDAs or mobile phones having a remote control device for bidirectional communication between the remote control device and client device
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42203—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/466—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/4667—Processing of monitored end-user data, e.g. trend analysis based on the log file of viewer selections
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/812—Monomedia components thereof involving advertisement data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8456—Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/57—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for processing of video signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/442—Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
- H04N21/44204—Monitoring of content usage, e.g. the number of times a movie has been viewed, copied or the amount which has been watched
Definitions
- the subject disclosure relates to a method and apparatus for content adaptation based on audience monitoring
- Selection of content for presentation to the users including movies, programming, advertising and the like is often performed based on generalizations associated with demographics, such as age and gender. However, not all users fit those generalizations which can lead to selection of content which is not of interest to a particular user.
- FIGS. 1 and 2 depict illustrative embodiments of content adjustment systems that can adjust content according to monitored audience reaction
- FIG. 3 depicts an illustrative embodiment of a method used in portions of the system described in FIGS. 1 and 2 ;
- FIG. 4 depicts an illustrative embodiment of a communication system that provides media services including content adjustment based on audience reactions;
- FIG. 5 depicts an illustrative embodiment of a communication device
- FIG. 6 is a diagrammatic representation of a machine in the form of a computer system within which a set of instructions, when executed, may cause the machine to perform any one or more of the methods described herein.
- the subject disclosure describes, among other things, illustrative embodiments for adjusting content based on the reaction of the audience.
- the audience reaction can be compared to an expected audience reaction that is determined from a context of the particular segment of content that the audience is viewing.
- the audience reaction can be utilized to determine a level of interest or a change in the level of interest which can trigger an adjustment of the content.
- Various criteria can be utilized for comparing the actual and expected audience reaction, including monitoring sound volume or determining a context of the audience reaction, such as whether there is excitement directed towards the content or whether the captured sounds are part of a conversation indicating disinterest in the content, or even different environments, such as mobility state or the vibration pattern of the mobile device (via which content is delivered) in a car or a train.
- Other embodiments are described in the subject disclosure.
- One or more aspects of the subject disclosure are a method including determining, by a system including a processor, a content context of a first segment of content being presented at a display to an audience.
- the method can include determining, by the system, an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected volume and an expected audio reaction context.
- the method can include obtaining, by the system, sensor data captured from a sensor device in proximity to the audience, where the sensor data is indicative of a sensed audience reaction to the first segment of the content, and where the sensed audience reaction comprises a sensed volume and a sensed audio reaction context.
- the method can include comparing, by the system, the sensed volume and the sensed audio reaction context with the expected volume and the expected audio reaction context to determine a level of interest in the first segment.
- the method can include adjusting, by the system, a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display.
- One or more aspects of the subject disclosure include a machine-readable storage medium, comprising executable instructions that, when executed by a media processor, facilitate performance of operations, including determining a content context of a first segment of content.
- the media processor can present the first segment of the content at a display to an audience; and can determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context.
- the media processor can obtain sensor data captured from a sensor device in proximity to the display, where the sensor data is indicative of a sensed audience reaction to the first segment of the content.
- the media processor can apply speech pattern recognition to the sensor data to determine a sensed audio reaction context of the sensed audience reaction.
- the media processor can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment.
- the media processor can increase a resolution of a second segment of the content according to a determination that the level of interest has increased, where the increasing of the resolution of the second segment generates an adjusted second segment.
- the media processor can present the adjusted second segment at the display.
- the media processor can provide a notice to a second media processor that causes the second media processor to decrease another resolution of other content being presented by the second media processor.
- One or more aspects of the subject disclosure include a network server having a processor; and a memory that stores executable instructions that, when executed by the processor, facilitate performance of operations, including determining a content context of a first segment of content being presented by a media processor at a display to an audience.
- the network server can determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context.
- the network server can receive sensor data captured from a sensor device in proximity to the audience, where the sensor data is indicative of a sensed audience reaction to the first segment of the content.
- the network server can apply speech recognition to the sensor data to identify words spoken by the audience.
- the network server can determine a sensed audio reaction context according to the words.
- the network server can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment.
- the network server can adjust a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display.
- FIG. 1 depicts an illustrative embodiment of a system 100 in which content is presented by a media processor 106 (e.g., a set top box) at a display 108 (e.g., a television).
- An audience 110 can be viewing or otherwise consuming the content.
- Three users are illustrated for audience 110 , although the audience could be a single user or any other number of users.
- the media processor 106 , the display 108 and the audience 110 are illustrated at a single location (e.g., a single premises), however, the media processor, the display and the audience can be groups of media processors, displays and audiences that are located at different locations, such as different premises.
- the content can be provided to the media processor 106 by a network server 130 .
- one or more sensor devices 150 can capture sensor data of a sensed audience reaction for the audience 110 .
- the sensor data can be of various types including audio and/or image information.
- the sensor device 150 can be integrated with the media processor 106 or can be otherwise in communication with the media processor to provide the sensor data to the media processor.
- the sensor device 150 can be a separate device from the media processor 106 , where the sensor device provides the sensor data to the network server 130 with or without sharing the sensor data with the media processor.
- Combinations of sensor devices 150 can also be utilized, such as a first sensor device that is an audio recorder of the media processor 106 and a second sensor device that is a camera for capturing images of the audience 110 .
- the camera can be integrated with the media processor 106 or can be otherwise in communication with the media processor to provide the sensor data to the media processor, or the camera can be a separate device from the media processor, where the camera provides the sensor data to the network server 130 with or without sharing the sensor data with the media processor.
- the sensor device 150 can be part of an end user device of one or more of the users in the audience 110 .
- a mobile phone of a user of the audience 110 can record audio and provide that audio to the media processor 106 and/or to the network server 130 .
- the sensor device 150 can capture other sensor data, such as motion detection, lighting and so forth.
- System 100 enables determining a level of interest or a change in the level of interest for a particular segment of the content being presented so that a content adaptation can be performed. For example, a determination can be made for a content context of a first segment of the content being presented at the display 108 for the audience 110 .
- the content context can be performed based on various factors and utilizing various techniques.
- segments of the content can include or otherwise be associated with metadata that can be accessed where the metadata describes the particular content context of the particular segment.
- a movie can include metadata that describes various scenes in the movie.
- the content context can be categorized in such a way as to facilitate determining an expected audience reaction to the scene. For instance, the metadata can describe that the scene is funny from which it can be determined that audience laughter is expected.
- the metadata can describe that the scene is scary from which it can be determined that an audience scream and/or an audience movement is expected.
- the metadata can describe that the scene is somber from which it can be determined that audience quiet is expected.
- the metadata can also be utilized for determining that which is not an expected reaction of the audience, such as it is not expected for there to be laughter in a scene that is described by the metadata as tragic.
- pattern recognition can be applied to the segment of the content to determine the content context and thus the expected audience reaction.
- audio and/or image pattern recognition can be applied to the segment to determine that a movie scene is scary or tragic.
- Detected words and/or sounds e.g., laughter, scream, sighing, crying
- detected actions, or other characteristics of the scene detected from the pattern recognition can be utilized in determining the particular content context.
- the pattern recognition can also be performed on live events that are being presented as the content. For instance, pattern recognition can be utilized to determine that a football team is close to the end zone from which it can be determined that the expected audience reaction can be excited (for an audience member favoring the team on offense) or discouraged (for an audience member favoring the team on defense).
- an expected volume of audience reaction can be determined, such as determining that an audio reaction to a touchdown in a football game should be above a threshold level if the audience member is interested in the particular game.
- the determination of the content context can be performed by various devices, such as by the media processor 106 and/or the network server 130 .
- System 100 enables comparing an expected audience reaction with a sensed audience reaction to determine a level of interest or a change in the level of interest for a particular segment of the content being presented so that a content adaptation can be performed.
- the comparison can be performed by various devices, such as by the media processor 106 and/or the network server 130 .
- the expected and sensed audience reactions can be characterized in various ways, including audio volume, spoken words, sounds, movement and so forth.
- a content adaptation can be performed.
- the content adaptation can be of various types. For example, characteristics of the content can be adjusted, such as a resolution or quality, in response to a determination of an increased level of interest. For instance, a determination can be made that the audience has become more interested in a football game because the offensive team is near the end zone and may score a touchdown. As described above, this determination can be performed according to a comparison of the expected audience reaction and the sensed audience reaction.
- the content i.e., the football game
- the content can be adjusted to a high resolution so that the upcoming plays near the end zone are of a higher quality.
- the system 100 can continue to monitor for content context of subsequent segments, expected audience reactions for those subsequent segments, and sensed audience reactions for those subsequent segments, from which changes in level of interest can be determined. In this example, if it is determined that the level of interest decreases then the content can be adjusted again, such as lowering the resolution (for plays that are deemed to be of less interest to the audience).
- the adjustment of the resolution can be performed in conjunction with adjustment to other content being provided to the location. For instance, an increase in resolution at the media processor 106 resulting in an increase in bandwidth can be offset by a decrease in resolution at another media processor so that the bandwidth usage for the location are maintained relatively constant.
- the location can have multiple audiences and multiple media processors where the monitoring for content context, expected audience reactions, and sensed audience reactions are performed for those multiple audiences, from which changes in level of interest can be determined for those multiple audiences.
- the changes in level of interest for each of the multiple audiences (at different displays watching different content) can be compared and the content adjustment can be made according to a higher level of change in interest. For instance, a first audience determined to have a higher level of interest can be presented with a high definition version of first content while a second audience determined to have a lower level of interest can be presented with a standard definition version of second content.
- the adjustment of the content can include inserting a different segment into the content, such as a targeted advertisement or a different version of a scene.
- the adjustment of the content can include replacing the subsequent segments with completely different content, such as switching to a different episode of a show or a different movie.
- the monitoring of the audience reaction can be performed on an individual basis. For example, a captured audio reaction can be identified with a particular user of the audience and compared with an expected audio reaction for that user. User profiles and/or historical reactions for that particular user can be analyzed in conjunction with the content context to determine the expected audience reaction for that particular user. As an example, it can be determined that the content context for a scene in a movie is humor. It can be further determined that a particular user in the past has reacted with laughter at a particular volume. The sensed laughter for that particular user can be compared to the particular volume of past laughter to determine a level of interest in the scene for the particular user.
- a notification can be presented at the display 108 indicating that the content adjustment is going to occur (e.g., resolution of a subsequent segment of the content will be increased).
- a characteristic of the audience can be determined (e.g., by the media processor 106 and/or the network server 130 ) based on a monitored consumption history associated with the audience.
- the adjusting of a subsequent segment of the content can include selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement for presentation at the display 108 , where the selecting of the targeted advertisement is according to the determined level of interest and according to the characteristic of the audience.
- system 100 can teach a video broadcast system how to insert and/or adapt content according to an audience signal-to-noise ratio.
- the audience signal-to-noise ratio can be a measure of expected audio to actual audio and can filter out background noise.
- System 100 can capture the mood, emotion, need and/or requirement of the audience at that particular moment. As a result, the content inserted and/or adapted to the video broadcast system can instantaneously capture the need and/or requirement of the audience at that moment.
- System 100 can enable content insertion to be adapted and personalized according to the audience signal-to-noise which is distinct from blindly or randomly inserting content.
- FIG. 2 depicts an illustrative embodiment of system 200 that enables content adaptation according to monitored audience reactions to segments of the content. By determining an expected audience reaction according to the context of the particular segment of content being presented, system 200 can detect a level of interest or a change in the level of interest via a comparison to a sensed audience reaction.
- a device 205 can determine a content context of a first segment of content being presented by the media processor to the audience 110 .
- the content context determination can be performed utilizing various techniques including analyzing metadata that describes the segment content and/or performing pattern recognition (e.g., audio and/or image) to determine the segment context.
- An expected audience reaction can be determined according to the content context of the first segment, such as determining an expected audio reaction context (e.g., laughter, screaming, enthusiasm, anger, and so forth).
- the device 205 can obtain sensor data captured from one or more sensor devices in proximity to the audience, such as an audio recorder, a camera, a motion detector and so forth.
- the sensor data can be indicative of a sensed audience reaction to the first segment of the content.
- the sensor data can be filtered, such as removing background noise, to isolate the sensed audience reaction.
- speech recognition can be applied to the sensor data to identify words spoken by the audience. Based on the speech recognition, a sensed audio reaction context can be determined according to the words. For example, words indicative of anger or words indicative of grief can be identified.
- Device 205 can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment. For instance, device 205 can determine that a sad scene in a movie should invoke a griefful audience reaction. If the comparison of the expected and sensed audience reactions is not a match then device 205 can determine that there is a low level of interest in the scene. In one embodiment, the device 205 can detect a sensed audience reaction which is contrary to what is expected, such as sensing laughter during a sad scene. Based on this detection, device 205 can determine there is little interest in the scene.
- device 205 can adjust a second segment of the content according to the level of interest to generate an adjusted second segment for presentation at the display.
- the adjusting of the second segment of the content can include adjusting a resolution of the second segment according to the level of interest.
- the adjusting of the resolution of the second segment can be performed in conjunction with decreasing of another resolution of second content being presented by another media processor, where the media processor and the other media processor are located at a same premises.
- the display can be a group of displays located at different premises, where the audience 110 is a group of users located at the different premises.
- the sensor device can be a group of sensor devices that capture sensor data for the multiple audiences at the different premises.
- the content adjustment can be performed for content being presented at the different premises.
- the device 205 can determine a characteristic of the audience based on a monitored consumption history associated with the audience.
- the adjusting of the second segment of the content can include selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement to the media processor for presentation at the display. The selecting of the targeted advertisement can be according to the level of interest and according to the characteristic of the audience.
- the determining of the content context of the first segment can include applying image pattern recognition and speech pattern recognition to the first segment.
- device 205 can determine an expected volume of the expected audience reaction according to the content context of the first segment; determine a sensed volume of the sensed audience reaction according to the sensor data; and compare the sensed volume with the expected volume, where the level of interest in the first segment is determined based in part on the comparing of the sensed volume with the expected volume.
- FIG. 3 depicts an illustrative embodiment of a method 300 used by systems 100 and 200 for adjusting content according to a level of audience interest determined from a comparison of an expected audience reaction to an actual audience reaction.
- Method 300 can commence at 302 where a content context is determined by a system for a first segment of content being presented at a display for an audience.
- the system can be various devices or combinations of devices, including a media processor presenting the content or a network server.
- an expected audience reaction can be determined by the system according to the content context of the first segment.
- the expected audience reaction can include an expected volume and/or an expected audio reaction context.
- the system can obtain sensor data that was captured from a sensor device in proximity to the audience.
- the sensor data can be indicative of a sensed audience reaction to the first segment of the content.
- the sensed audience reaction can include a sensed volume and/or a sensed audio reaction context.
- the system can compare the sensed audience reaction (e.g., the sensed volume and/or the sensed audio reaction context) with the expected audience reaction (e.g., the expected volume and/or the expected audio reaction context) to determine change in level of interest for the first segment. If there is no change, then method 300 can return to 302 to continue monitoring for changes in interest to subsequent segments. If on the other hand there is a change in level of interest then at 310 the system can adjust a second segment of the content according to the changed level of interest to generate an adjusted second segment for presentation at the display. For example, if it is determined that the level of interest has increased then the second segment can be adjusted according to that increased level of interest.
- the sensed audience reaction e.g., the sensed volume and/or the sensed audio reaction context
- the expected audience reaction e.g., the expected volume and/or the expected audio reaction context
- the sensor device can be an audio recorder of a media processor presenting the content at the display, and method 300 can include applying speech recognition to the sensor data to identify words spoken by the audience; and determining the sensed audio reaction context according to the words.
- the adjusting of the second segment of the content can include adjusting a resolution of the second segment according to the level of interest.
- the adjusting of the second segment of the content can include increasing a resolution of the second segment being presented by a first media processor at the display, where the adjusting of the resolution of the second segment is performed in conjunction with decreasing of another resolution of second content being presented by a second media processor, and where the first and second media processors are located at a same premises.
- method 300 can include transmitting, by the system over a network, the content to a media processor presenting the content at the display, where the obtaining of the sensor data by the system comprises receiving the sensor data captured from the sensor device housed in the media processor.
- the display can be a group of displays located at different premises, where the audience is a group of users located at the different premises, and where the sensor device is a group of sensor devices that are each housed by a corresponding media processor presenting the content at one of the group of displays.
- the method 300 can include determining a characteristic of the audience based on a monitored consumption history associated with the audience, where the adjusting of the second segment of the content comprises selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement for presentation at the display, and where the selecting of the targeted advertisement is according to the level of interest and according to the characteristic of the audience.
- the determining of the content context of the first segment can be based on metadata associated with the content. In one embodiment, the determining of the content context of the first segment can include applying image pattern recognition and speech pattern recognition to the first segment.
- FIG. 4 depicts an illustrative embodiment of a first communication system 400 for delivering media content.
- the communication system 400 can represent an interactive television network, such as an Internet Protocol Television (IPTV) media system.
- IPTV Internet Protocol Television
- Communication system 400 can be overlaid or operably coupled with systems 100 and 200 of FIGS. 1 and 2 as another representative embodiment of communication system 400 .
- one or more devices illustrated in the communication system 400 of FIG. 4 can adjust content based on the reaction of the audience.
- the audience reaction can be compared to an expected audience reaction that is determined from a context of the particular segment of content that the audience is viewing.
- the audience reaction can be utilized to determine a level of interest or a change in the level of interest which can trigger an adjustment of the content.
- Various criteria can be utilized for comparing the actual and expected audience reaction, such as monitoring sound volume, determining a context of the audience reaction, such as whether there is excitement directed towards the content or whether the captured sounds are part of a conversation indicating disinterest in the content.
- the IPTV media system can include a super head-end office (SHO) 410 with at least one super headend office server (SHS) 411 which receives media content from satellite and/or terrestrial communication systems.
- media content can represent, for example, audio content, moving image content such as 2D or 3D videos, video games, virtual reality content, still image content, and combinations thereof.
- the SHS server 411 can forward packets associated with the media content to one or more video head-end servers (VHS) 414 via a network of video head-end offices (VHO) 412 according to a multicast communication protocol.
- the VHS 414 can distribute multimedia broadcast content via an access network 418 to commercial and/or residential buildings 402 housing a gateway 404 (such as a residential or commercial gateway).
- the access network 418 can represent a group of digital subscriber line access multiplexers (DSLAMs) located in a central office or a service area interface that provide broadband services over fiber optical links or copper twisted pairs 419 to buildings 402 .
- DSLAMs digital subscriber line access multiplexers
- the gateway 404 can use communication technology to distribute broadcast signals to media processors 406 such as Set-Top Boxes (STBs) which in turn present broadcast channels to media devices 408 such as computers or television sets managed in some instances by a media controller 407 (such as an infrared or RF remote controller).
- STBs Set-Top Boxes
- media devices 408 such as computers or television sets managed in some instances by a media controller 407 (such as an infrared or RF remote controller).
- the gateway 404 , the media processors 406 , and media devices 408 can utilize tethered communication technologies (such as coaxial, powerline or phone line wiring) or can operate over a wireless access protocol such as Wireless Fidelity (WiFi), Bluetooth®, Zigbee or other present or next generation local or personal area wireless network technologies.
- WiFi Wireless Fidelity
- Bluetooth® Bluetooth®
- Zigbee Zigbee
- unicast communications can also be invoked between the media processors 406 and subsystems of the IPTV media system for services such as video-on-demand (VoD), browsing an electronic programming guide (EPG), or other infrastructure services.
- VoD video-on-demand
- EPG electronic programming guide
- a satellite broadcast television system 429 can be used in the media system of FIG. 4 .
- the satellite broadcast television system can be overlaid, operably coupled with, or replace the IPTV system as another representative embodiment of communication system 400 .
- signals transmitted by a satellite 415 that include media content can be received by a satellite dish receiver 431 coupled to the building 402 .
- Modulated signals received by the satellite dish receiver 431 can be transferred to the media processors 406 for demodulating, decoding, encoding, and/or distributing broadcast channels to the media devices 408 .
- the media processors 406 can be equipped with a broadband port to an Internet Service Provider (ISP) network 432 to enable interactive services such as VoD and EPG as described above.
- ISP Internet Service Provider
- an analog or digital cable broadcast distribution system such as cable TV system 433 can be overlaid, operably coupled with, or replace the IPTV system and/or the satellite TV system as another representative embodiment of communication system 400 .
- the cable TV system 433 can also provide Internet, telephony, and interactive media services.
- System 400 enables various types of interactive television and/or services including IPTV, cable and/or satellite.
- the subject disclosure can apply to other present or next generation over-the-air and/or landline media content services system.
- Some of the network elements of the IPTV media system can be coupled to one or more computing devices 430 , a portion of which can operate as a web server for providing web portal services over the ISP network 432 to wireline media devices 408 or wireless communication devices 416 .
- Communication system 400 can also provide for all or a portion of the computing devices 430 to function as a content adjustment server (herein referred to as server 430 ).
- the server 430 can use computing and communication technology to perform function 462 , which can include among other things, determining a content context of a first segment of content being presented by a media processor at a display for an audience; determining an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context; receiving sensor data captured from a sensor device in proximity to the audience, the sensor data indicative of a sensed audience reaction to the first segment of the content; applying speech recognition to the sensor data to identify words spoken by the audience; determining a sensed audio reaction context according to the words; comparing the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment; and/or adjusting a second segment of the content according to the level of interest to generate an adjusted second segment for presentation at the display.
- function 462 of server 430 can be similar to the functions described network server 130 of system 100 .
- the media processors 406 and wireless communication devices 416 can be provisioned with software functions 464 and 466 , respectively, to utilize the services of server 430 .
- functions 46 Y and 46 Z of media processors 406 and wireless communication devices 416 can be similar to the functions described for the communication devices 106 of FIG. 1 in accordance with method 300 .
- media services can be offered to media devices over landline technologies such as those described above. Additionally, media services can be offered to media devices by way of a wireless access base station 417 operating according to common wireless access protocols such as Global System for Mobile or GSM, Code Division Multiple Access or CDMA, Time Division Multiple Access or TDMA, Universal Mobile Telecommunications or UMTS, World interoperability for Microwave or WiMAX, Software Defined Radio or SDR, Long Term Evolution or LTE, and so on.
- GSM Global System for Mobile or GSM
- CDMA Code Division Multiple Access
- TDMA Time Division Multiple Access or TDMA
- Universal Mobile Telecommunications or UMTS Universal Mobile Telecommunications or UMTS
- World interoperability for Microwave or WiMAX Software Defined Radio or SDR, Long Term Evolution or LTE, and so on.
- Other present and next generation wide area wireless access network technologies can be used in one or more embodiments of the subject disclosure.
- FIG. 5 depicts an illustrative embodiment of a communication device 500 .
- Communication device 500 can serve in whole or in part as an illustrative embodiment of the devices depicted in systems 100 , 200 and 400 and can be configured to perform portions of method 300 .
- communication device 500 can determine a content context of a first segment of content; present the first segment of the content at a display for an audience; determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context; obtain sensor data captured from a sensor device in proximity to the display, the sensor data indicative of a sensed audience reaction to the first segment of the content; apply speech pattern recognition to the sensor data to determine a sensed audio reaction context of the sensed audience reaction; compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment; increase a resolution of a second segment of the content according to a determination that the level of interest has increased, where the increasing of the resolution of the second segment generates an adjusted second segment; present the adjusted second segment at the display; and/or provide a notice to a second media processor that causes the second media processor to decrease another resolution of other content being presented by the second media processor.
- communication device 500 can determine an expected volume of the expected audience reaction according to the content context of the first segment; determine a sensed volume of the sensed audience reaction according to the sensor data; and compare the sensed volume with the expected volume, where the level of interest in the first segment is determined based in part on the comparing of the sensed volume with the expected volume.
- communication device 500 can present a notification at the display indicating that the resolution of the second segment of the content will be increased prior to the presenting of the adjusted second segment at the display.
- Communication device 500 can comprise a wireline and/or wireless transceiver 502 (herein transceiver 502 ), a user interface (UI) 504 , a power supply 514 , a location receiver 516 , a motion sensor 518 , an orientation sensor 520 , and a controller 506 for managing operations thereof.
- the transceiver 502 can support short-range or long-range wireless access technologies such as Bluetooth®, ZigBee®, WiFi, DECT, or cellular communication technologies, just to mention a few (Bluetooth® and ZigBee® are trademarks registered by the Bluetooth® Special Interest Group and the ZigBee® Alliance, respectively).
- Cellular technologies can include, for example, CDMA-1X, UMTS/HSDPA, GSM/GPRS, TDMA/EDGE, EV/DO, WiMAX, SDR, LTE, as well as other next generation wireless communication technologies as they arise.
- the transceiver 502 can also be adapted to support circuit-switched wireline access technologies (such as PSTN), packet-switched wireline access technologies (such as TCP/IP, VoIP, etc.), and combinations thereof.
- the UI 504 can include a depressible or touch-sensitive keypad 508 with a navigation mechanism such as a roller ball, a joystick, a mouse, or a navigation disk for manipulating operations of the communication device 500 .
- the keypad 508 can be an integral part of a housing assembly of the communication device 500 or an independent device operably coupled thereto by a tethered wireline interface (such as a USB cable) or a wireless interface supporting for example Bluetooth®.
- the keypad 508 can represent a numeric keypad commonly used by phones, and/or a QWERTY keypad with alphanumeric keys.
- the UI 504 can further include a display 510 such as monochrome or color LCD (Liquid Crystal Display), OLED (Organic Light Emitting Diode) or other suitable display technology for conveying images to an end user of the communication device 500 .
- a display 510 such as monochrome or color LCD (Liquid Crystal Display), OLED (Organic Light Emitting Diode) or other suitable display technology for conveying images to an end user of the communication device 500 .
- a display 510 is touch-sensitive, a portion or all of the keypad 508 can be presented by way of the display 510 with navigation features.
- the display 510 can use touch screen technology to also serve as a user interface for detecting user input.
- the communication device 500 can be adapted to present a user interface with graphical user interface (GUI) elements that can be selected by a user with a touch of a finger.
- GUI graphical user interface
- the touch screen display 510 can be equipped with capacitive, resistive or other forms of sensing technology to detect how much surface area of a user's finger has been placed on a portion of the touch screen display. This sensing information can be used to control the manipulation of the GUI elements or other functions of the user interface.
- the display 510 can be an integral part of the housing assembly of the communication device 500 or an independent device communicatively coupled thereto by a tethered wireline interface (such as a cable) or a wireless interface.
- the UI 504 can also include an audio system 512 that utilizes audio technology for conveying low volume audio (such as audio heard in proximity of a human ear) and high volume audio (such as speakerphone for hands free operation).
- the audio system 512 can further include a microphone for receiving audible signals of an end user.
- the audio system 512 can also be used for voice recognition applications.
- the UI 504 can further include an image sensor 513 such as a charged coupled device (CCD) camera for capturing still or moving images.
- CCD charged coupled device
- the power supply 514 can utilize common power management technologies such as replaceable and rechargeable batteries, supply regulation technologies, and/or charging system technologies for supplying energy to the components of the communication device 500 to facilitate long-range or short-range portable applications.
- the charging system can utilize external power sources such as DC power supplied over a physical interface such as a USB port or other suitable tethering technologies.
- the location receiver 516 can utilize location technology such as a global positioning system (GPS) receiver capable of assisted GPS for identifying a location of the communication device 500 based on signals generated by a constellation of GPS satellites, which can be used for facilitating location services such as navigation.
- GPS global positioning system
- the motion sensor 518 can utilize motion sensing technology such as an accelerometer, a gyroscope, or other suitable motion sensing technology to detect motion of the communication device 500 in three-dimensional space.
- the orientation sensor 520 can utilize orientation sensing technology such as a magnetometer to detect the orientation of the communication device 500 (north, south, west, and east, as well as combined orientations in degrees, minutes, or other suitable orientation metrics).
- the communication device 500 can use the transceiver 502 to also determine a proximity to a cellular, WiFi, Bluetooth®, or other wireless access points by sensing techniques such as utilizing a received signal strength indicator (RSSI) and/or signal time of arrival (TOA) or time of flight (TOF) measurements.
- the controller 506 can utilize computing technologies such as a microprocessor, a digital signal processor (DSP), programmable gate arrays, application specific integrated circuits, and/or a video processor with associated storage memory such as Flash, ROM, RAM, SRAM, DRAM or other storage technologies for executing computer instructions, controlling, and processing data supplied by the aforementioned components of the communication device 500 .
- computing technologies such as a microprocessor, a digital signal processor (DSP), programmable gate arrays, application specific integrated circuits, and/or a video processor with associated storage memory such as Flash, ROM, RAM, SRAM, DRAM or other storage technologies for executing computer instructions, controlling, and processing data supplied by the aforementioned components of the communication device
- the communication device 500 can include a reset button (not shown).
- the reset button can be used to reset the controller 506 of the communication device 500 .
- the communication device 500 can also include a factory default setting button positioned, for example, below a small hole in a housing assembly of the communication device 500 to force the communication device 500 to re-establish factory settings.
- a user can use a protruding object such as a pen or paper clip tip to reach into the hole and depress the default setting button.
- the communication device 500 can also include a slot for adding or removing an identity module such as a Subscriber Identity Module (SIM) card. SIM cards can be used for identifying subscriber services, executing programs, storing subscriber data, and so forth.
- SIM Subscriber Identity Module
- the communication device 500 as described herein can operate with more or less of the circuit components shown in FIG. 5 . These variant embodiments can be used in one or more embodiments of the subject disclosure.
- the communication device 500 can be adapted to perform the functions of devices 106 , 130 , 205 of FIGS. 1 and 2 , the media processor 406 , the media devices 408 , or the portable communication devices 416 of FIG. 4 . It will be appreciated that the communication device 500 can also represent other devices that can operate in systems 100 and 200 of FIGS. 1 and 2 ], communication system 400 of FIG. 4 such as a gaming console and a media player. In addition, the controller 506 can be adapted in various embodiments to perform the functions 462 - 466 .
- Other actions can be triggered by the determination of the level of interest or a change in the level of interest, such as proposing different content to the audience or to a subset of the audience.
- the determination of the level of interest can be multi-tiered based on different thresholds. For example, a scene in a movie can be determined to have a content context of frightening.
- the expected audience reaction can be determined to be a scream or movement of the audience. If the sensed audience reaction is both a scream and movement then a determination of very high interest can be made. If the sensed audience reaction is either the scream or the movement then a determination of high interest can be made. If the sensed audience reaction is neither the scream nor the movement, but rather silence, then a determination of medium interest can be made. If the sensed audience reaction is neither the scream nor the movement, but rather laughter, then a determination of low interest can be made.
- the content adjustments can then be made according to the group of different interest levels. For instance, the highest level of interest can result in an increase to the highest resolution and insertion of a targeted advertisement for another movie in the same genre while the lowest level of interest can result in a notification that a different movie of a different genre is currently available.
- sensed audience responses can be distinguished amongst users of the audience and a content adjustment determination can be made according to individual responses.
- sensed audio can be analyzed via speech pattern recognition to determine that first and second users are engaged in a conversation that is not related to the content (e.g., a football game) being presented.
- the sensed audio can be further analyzed to determine that third and fourth users are cheering.
- the determination of which users are speaking can be performed in a number of different ways, such as based on accessible voice samples.
- a content adjustment can be made (e.g., a change in resolution of the presented football game or an insertion of a commercial for another football game that is scheduled in the near future) based on the interest of the third and fourth users.
- the content can be broadcast content which is then adjusted by a network server responsive to detecting the change in level of interest.
- a network server responsive to detecting the change in level of interest.
- devices described in the exemplary embodiments can be in communication with each other via various wireless and/or wired methodologies.
- the methodologies can be links that are described as coupled, connected and so forth, which can include unidirectional and/or bidirectional communication over wireless paths and/or wired paths that utilize one or more of various protocols or methodologies, where the coupling and/or connection can be direct (e.g., no intervening processing device) and/or indirect (e.g., an intermediary processing device such as a router).
- FIG. 6 depicts an exemplary diagrammatic representation of a machine in the form of a computer system 600 within which a set of instructions, when executed, may cause the machine to perform any one or more of the methods described above.
- One or more instances of the machine can operate, for example, as the media processor 106 , the network server 130 , the server 430 , the media processor 406 and other devices of FIGS. 1-2 and 4 in order to adjust content according to a comparison of an expected audience reaction with a sensed audience reaction.
- the machine may be connected (e.g., using a network 626 ) to other machines.
- the machine may operate in the capacity of a server or a client user machine in a server-client user network environment, or as a peer machine in a peer-to-peer (or distributed) network environment.
- the machine may comprise a server computer, a client user computer, a personal computer (PC), a tablet, a smart phone, a laptop computer, a desktop computer, a control system, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine.
- a communication device of the subject disclosure includes broadly any electronic device that provides voice, video or data communication.
- the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methods discussed herein.
- the computer system 600 may include a processor (or controller) 602 (e.g., a central processing unit (CPU)), a graphics processing unit (GPU, or both), a main memory 604 and a static memory 606 , which communicate with each other via a bus 608 .
- the computer system 600 may further include a display unit 610 (e.g., a liquid crystal display (LCD), a flat panel, or a solid state display).
- the computer system 600 may include an input device 612 (e.g., a keyboard), a cursor control device 614 (e.g., a mouse), a disk drive unit 616 , a signal generation device 618 (e.g., a speaker or remote control) and a network interface device 620 .
- the embodiments described in the subject disclosure can be adapted to utilize multiple display units 610 controlled by two or more computer systems 600 .
- presentations described by the subject disclosure may in part be shown in a first of the display units 610 , while the remaining portion is presented in a second of the display units 610 .
- the disk drive unit 616 may include a tangible computer-readable storage medium 622 on which is stored one or more sets of instructions (e.g., software 624 ) embodying any one or more of the methods or functions described herein, including those methods illustrated above.
- the instructions 624 may also reside, completely or at least partially, within the main memory 604 , the static memory 606 , and/or within the processor 602 during execution thereof by the computer system 600 .
- the main memory 604 and the processor 602 also may constitute tangible computer-readable storage media.
- Dedicated hardware implementations including, but not limited to, application specific integrated circuits, programmable logic arrays and other hardware devices can likewise be constructed to implement the methods described herein.
- Application specific integrated circuits and programmable logic array can use downloadable instructions for executing state machines and/or circuit configurations to implement embodiments of the subject disclosure.
- Applications that may include the apparatus and systems of various embodiments broadly include a variety of electronic and computer systems. Some embodiments implement functions in two or more specific interconnected hardware modules or devices with related control and data signals communicated between and through the modules, or as portions of an application-specific integrated circuit.
- the example system is applicable to software, firmware, and hardware implementations.
- the operations or methods described herein are intended for operation as software programs or instructions running on or executed by a computer processor or other computing device, and which may include other forms of instructions manifested as a state machine implemented with logic components in an application specific integrated circuit or field programmable gate array.
- software implementations e.g., software programs, instructions, etc.
- a computing device such as a processor, a controller, a state machine or other suitable device for executing instructions to perform operations or methods may perform such operations directly or indirectly by way of one or more intermediate devices directed by the computing device.
- tangible computer-readable storage medium 622 is shown in an example embodiment to be a single medium, the term “tangible computer-readable storage medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions.
- tangible computer-readable storage medium shall also be taken to include any non-transitory medium that is capable of storing or encoding a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methods of the subject disclosure.
- non-transitory as in a non-transitory computer-readable storage includes without limitation memories, drives, devices and anything tangible but not a signal per se.
- tangible computer-readable storage medium shall accordingly be taken to include, but not be limited to: solid-state memories such as a memory card or other package that houses one or more read-only (non-volatile) memories, random access memories, or other re-writable (volatile) memories, a magneto-optical or optical medium such as a disk or tape, or other tangible media which can be used to store information. Accordingly, the disclosure is considered to include any one or more of a tangible computer-readable storage medium, as listed herein and including art-recognized equivalents and successor media, in which the software implementations herein are stored.
- Each of the standards for Internet and other packet switched network transmission (e.g., TCP/IP, UDP/IP, HTML, HTTP) represent examples of the state of the art. Such standards are from time-to-time superseded by faster or more efficient equivalents having essentially the same functions.
- Wireless standards for device detection e.g., RFID
- short-range communications e.g., Bluetooth®, WiFi, Zigbee®
- long-range communications e.g., WiMAX, GSM, CDMA, LTE
- facilitating e.g., facilitating access or facilitating establishing a connection
- the facilitating can include less than every step needed to perform the function or can include all of the steps needed to perform the function.
- a processor (which can include a controller or circuit) has been described that performs various functions. It should be understood that the processor can be multiple processors, which can include distributed processors or parallel processors in a single machine or multiple machines.
- the processor can be used in supporting a virtual processing environment.
- the virtual processing environment may support one or more virtual machines representing computers, servers, or other computing devices. In such virtual machines, components such as microprocessors and storage devices may be virtualized or logically represented.
- the processor can include a state machine, application specific integrated circuit, and/or programmable gate array including a Field PGA.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Social Psychology (AREA)
- Computing Systems (AREA)
- Marketing (AREA)
- Business, Economics & Management (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
- This application is a continuation of U.S. patent application Ser. No. 14/938,421 filed Nov. 11, 2015, pending. All sections of the aforementioned application are incorporated herein by reference in their entirety.
- The subject disclosure relates to a method and apparatus for content adaptation based on audience monitoring
- Users have various options to select from in communication services. Selection of content for presentation to the users including movies, programming, advertising and the like is often performed based on generalizations associated with demographics, such as age and gender. However, not all users fit those generalizations which can lead to selection of content which is not of interest to a particular user.
- Reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
-
FIGS. 1 and 2 depict illustrative embodiments of content adjustment systems that can adjust content according to monitored audience reaction; -
FIG. 3 depicts an illustrative embodiment of a method used in portions of the system described inFIGS. 1 and 2 ; -
FIG. 4 depicts an illustrative embodiment of a communication system that provides media services including content adjustment based on audience reactions; -
FIG. 5 depicts an illustrative embodiment of a communication device; and -
FIG. 6 is a diagrammatic representation of a machine in the form of a computer system within which a set of instructions, when executed, may cause the machine to perform any one or more of the methods described herein. - The subject disclosure describes, among other things, illustrative embodiments for adjusting content based on the reaction of the audience. The audience reaction can be compared to an expected audience reaction that is determined from a context of the particular segment of content that the audience is viewing. The audience reaction can be utilized to determine a level of interest or a change in the level of interest which can trigger an adjustment of the content. Various criteria can be utilized for comparing the actual and expected audience reaction, including monitoring sound volume or determining a context of the audience reaction, such as whether there is excitement directed towards the content or whether the captured sounds are part of a conversation indicating disinterest in the content, or even different environments, such as mobility state or the vibration pattern of the mobile device (via which content is delivered) in a car or a train. Other embodiments are described in the subject disclosure.
- One or more aspects of the subject disclosure are a method including determining, by a system including a processor, a content context of a first segment of content being presented at a display to an audience. The method can include determining, by the system, an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected volume and an expected audio reaction context. The method can include obtaining, by the system, sensor data captured from a sensor device in proximity to the audience, where the sensor data is indicative of a sensed audience reaction to the first segment of the content, and where the sensed audience reaction comprises a sensed volume and a sensed audio reaction context. The method can include comparing, by the system, the sensed volume and the sensed audio reaction context with the expected volume and the expected audio reaction context to determine a level of interest in the first segment. The method can include adjusting, by the system, a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display.
- One or more aspects of the subject disclosure include a machine-readable storage medium, comprising executable instructions that, when executed by a media processor, facilitate performance of operations, including determining a content context of a first segment of content. The media processor can present the first segment of the content at a display to an audience; and can determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context. The media processor can obtain sensor data captured from a sensor device in proximity to the display, where the sensor data is indicative of a sensed audience reaction to the first segment of the content. The media processor can apply speech pattern recognition to the sensor data to determine a sensed audio reaction context of the sensed audience reaction. The media processor can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment. The media processor can increase a resolution of a second segment of the content according to a determination that the level of interest has increased, where the increasing of the resolution of the second segment generates an adjusted second segment. The media processor can present the adjusted second segment at the display. The media processor can provide a notice to a second media processor that causes the second media processor to decrease another resolution of other content being presented by the second media processor.
- One or more aspects of the subject disclosure include a network server having a processor; and a memory that stores executable instructions that, when executed by the processor, facilitate performance of operations, including determining a content context of a first segment of content being presented by a media processor at a display to an audience. The network server can determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context. The network server can receive sensor data captured from a sensor device in proximity to the audience, where the sensor data is indicative of a sensed audience reaction to the first segment of the content. The network server can apply speech recognition to the sensor data to identify words spoken by the audience. The network server can determine a sensed audio reaction context according to the words. The network server can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment. The network server can adjust a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display.
-
FIG. 1 depicts an illustrative embodiment of asystem 100 in which content is presented by a media processor 106 (e.g., a set top box) at a display 108 (e.g., a television). Anaudience 110 can be viewing or otherwise consuming the content. Three users are illustrated foraudience 110, although the audience could be a single user or any other number of users. Also, themedia processor 106, thedisplay 108 and theaudience 110 are illustrated at a single location (e.g., a single premises), however, the media processor, the display and the audience can be groups of media processors, displays and audiences that are located at different locations, such as different premises. The content can be provided to themedia processor 106 by anetwork server 130. - In one embodiment, one or more sensor devices 150 (only one of which is shown) can capture sensor data of a sensed audience reaction for the
audience 110. The sensor data can be of various types including audio and/or image information. In one embodiment, thesensor device 150 can be integrated with themedia processor 106 or can be otherwise in communication with the media processor to provide the sensor data to the media processor. In another embodiment, thesensor device 150 can be a separate device from themedia processor 106, where the sensor device provides the sensor data to thenetwork server 130 with or without sharing the sensor data with the media processor. - Combinations of
sensor devices 150 can also be utilized, such as a first sensor device that is an audio recorder of themedia processor 106 and a second sensor device that is a camera for capturing images of theaudience 110. For instance, the camera can be integrated with themedia processor 106 or can be otherwise in communication with the media processor to provide the sensor data to the media processor, or the camera can be a separate device from the media processor, where the camera provides the sensor data to thenetwork server 130 with or without sharing the sensor data with the media processor. - In another embodiment, the
sensor device 150 can be part of an end user device of one or more of the users in theaudience 110. For example, a mobile phone of a user of theaudience 110 can record audio and provide that audio to themedia processor 106 and/or to thenetwork server 130. In addition to audio and image recordings, thesensor device 150 can capture other sensor data, such as motion detection, lighting and so forth. -
System 100 enables determining a level of interest or a change in the level of interest for a particular segment of the content being presented so that a content adaptation can be performed. For example, a determination can be made for a content context of a first segment of the content being presented at thedisplay 108 for theaudience 110. The content context can be performed based on various factors and utilizing various techniques. In one embodiment, segments of the content can include or otherwise be associated with metadata that can be accessed where the metadata describes the particular content context of the particular segment. For example, a movie can include metadata that describes various scenes in the movie. The content context can be categorized in such a way as to facilitate determining an expected audience reaction to the scene. For instance, the metadata can describe that the scene is funny from which it can be determined that audience laughter is expected. As another example, the metadata can describe that the scene is scary from which it can be determined that an audience scream and/or an audience movement is expected. As another example, the metadata can describe that the scene is somber from which it can be determined that audience quiet is expected. The metadata can also be utilized for determining that which is not an expected reaction of the audience, such as it is not expected for there to be laughter in a scene that is described by the metadata as tragic. - In another embodiment, pattern recognition can be applied to the segment of the content to determine the content context and thus the expected audience reaction. For example, audio and/or image pattern recognition can be applied to the segment to determine that a movie scene is scary or tragic. Detected words and/or sounds (e.g., laughter, scream, sighing, crying), detected actions, or other characteristics of the scene detected from the pattern recognition can be utilized in determining the particular content context. The pattern recognition can also be performed on live events that are being presented as the content. For instance, pattern recognition can be utilized to determine that a football team is close to the end zone from which it can be determined that the expected audience reaction can be excited (for an audience member favoring the team on offense) or discouraged (for an audience member favoring the team on defense). In one embodiment, an expected volume of audience reaction can be determined, such as determining that an audio reaction to a touchdown in a football game should be above a threshold level if the audience member is interested in the particular game. The determination of the content context can be performed by various devices, such as by the
media processor 106 and/or thenetwork server 130. -
System 100 enables comparing an expected audience reaction with a sensed audience reaction to determine a level of interest or a change in the level of interest for a particular segment of the content being presented so that a content adaptation can be performed. The comparison can be performed by various devices, such as by themedia processor 106 and/or thenetwork server 130. As explained above, the expected and sensed audience reactions can be characterized in various ways, including audio volume, spoken words, sounds, movement and so forth. - Once a level of interest or a change in the level of interest for a particular segment of the content being presented is determined, then a content adaptation can be performed. The content adaptation can be of various types. For example, characteristics of the content can be adjusted, such as a resolution or quality, in response to a determination of an increased level of interest. For instance, a determination can be made that the audience has become more interested in a football game because the offensive team is near the end zone and may score a touchdown. As described above, this determination can be performed according to a comparison of the expected audience reaction and the sensed audience reaction. The content (i.e., the football game) can be adjusted to a high resolution so that the upcoming plays near the end zone are of a higher quality. In one embodiment, the
system 100 can continue to monitor for content context of subsequent segments, expected audience reactions for those subsequent segments, and sensed audience reactions for those subsequent segments, from which changes in level of interest can be determined. In this example, if it is determined that the level of interest decreases then the content can be adjusted again, such as lowering the resolution (for plays that are deemed to be of less interest to the audience). - In one embodiment, the adjustment of the resolution can be performed in conjunction with adjustment to other content being provided to the location. For instance, an increase in resolution at the
media processor 106 resulting in an increase in bandwidth can be offset by a decrease in resolution at another media processor so that the bandwidth usage for the location are maintained relatively constant. In one embodiment, the location can have multiple audiences and multiple media processors where the monitoring for content context, expected audience reactions, and sensed audience reactions are performed for those multiple audiences, from which changes in level of interest can be determined for those multiple audiences. In this example, the changes in level of interest for each of the multiple audiences (at different displays watching different content) can be compared and the content adjustment can be made according to a higher level of change in interest. For instance, a first audience determined to have a higher level of interest can be presented with a high definition version of first content while a second audience determined to have a lower level of interest can be presented with a standard definition version of second content. - In one embodiment, the adjustment of the content can include inserting a different segment into the content, such as a targeted advertisement or a different version of a scene. In another embodiment, the adjustment of the content can include replacing the subsequent segments with completely different content, such as switching to a different episode of a show or a different movie.
- In one embodiment, the monitoring of the audience reaction can be performed on an individual basis. For example, a captured audio reaction can be identified with a particular user of the audience and compared with an expected audio reaction for that user. User profiles and/or historical reactions for that particular user can be analyzed in conjunction with the content context to determine the expected audience reaction for that particular user. As an example, it can be determined that the content context for a scene in a movie is humor. It can be further determined that a particular user in the past has reacted with laughter at a particular volume. The sensed laughter for that particular user can be compared to the particular volume of past laughter to determine a level of interest in the scene for the particular user.
- In one embodiment, a notification can be presented at the
display 108 indicating that the content adjustment is going to occur (e.g., resolution of a subsequent segment of the content will be increased). In another embodiment, a characteristic of the audience can be determined (e.g., by themedia processor 106 and/or the network server 130) based on a monitored consumption history associated with the audience. In this example, the adjusting of a subsequent segment of the content can include selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement for presentation at thedisplay 108, where the selecting of the targeted advertisement is according to the determined level of interest and according to the characteristic of the audience. - In one or more embodiments,
system 100 can teach a video broadcast system how to insert and/or adapt content according to an audience signal-to-noise ratio. The audience signal-to-noise ratio can be a measure of expected audio to actual audio and can filter out background noise.System 100 can capture the mood, emotion, need and/or requirement of the audience at that particular moment. As a result, the content inserted and/or adapted to the video broadcast system can instantaneously capture the need and/or requirement of the audience at that moment.System 100 can enable content insertion to be adapted and personalized according to the audience signal-to-noise which is distinct from blindly or randomly inserting content. -
FIG. 2 depicts an illustrative embodiment ofsystem 200 that enables content adaptation according to monitored audience reactions to segments of the content. By determining an expected audience reaction according to the context of the particular segment of content being presented,system 200 can detect a level of interest or a change in the level of interest via a comparison to a sensed audience reaction. - In one embodiment, a device 205 (e.g., a media processor presenting the content, a network server, or some other device) can determine a content context of a first segment of content being presented by the media processor to the
audience 110. The content context determination can be performed utilizing various techniques including analyzing metadata that describes the segment content and/or performing pattern recognition (e.g., audio and/or image) to determine the segment context. An expected audience reaction can be determined according to the content context of the first segment, such as determining an expected audio reaction context (e.g., laughter, screaming, enthusiasm, anger, and so forth). Thedevice 205 can obtain sensor data captured from one or more sensor devices in proximity to the audience, such as an audio recorder, a camera, a motion detector and so forth. In one embodiment, the sensor data can be indicative of a sensed audience reaction to the first segment of the content. In another embodiment, the sensor data can be filtered, such as removing background noise, to isolate the sensed audience reaction. - In one embodiment, speech recognition can be applied to the sensor data to identify words spoken by the audience. Based on the speech recognition, a sensed audio reaction context can be determined according to the words. For example, words indicative of anger or words indicative of sorrow can be identified.
Device 205 can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment. For instance,device 205 can determine that a sad scene in a movie should invoke a sorrowful audience reaction. If the comparison of the expected and sensed audience reactions is not a match thendevice 205 can determine that there is a low level of interest in the scene. In one embodiment, thedevice 205 can detect a sensed audience reaction which is contrary to what is expected, such as sensing laughter during a sad scene. Based on this detection,device 205 can determine there is little interest in the scene. - Based on a detection of a particular level of interest (e.g., satisfying a threshold),
device 205 can adjust a second segment of the content according to the level of interest to generate an adjusted second segment for presentation at the display. In one embodiment, the adjusting of the second segment of the content can include adjusting a resolution of the second segment according to the level of interest. In another embodiment, the adjusting of the resolution of the second segment can be performed in conjunction with decreasing of another resolution of second content being presented by another media processor, where the media processor and the other media processor are located at a same premises. - In one embodiment, the display can be a group of displays located at different premises, where the
audience 110 is a group of users located at the different premises. In this example, the sensor device can be a group of sensor devices that capture sensor data for the multiple audiences at the different premises. The content adjustment can be performed for content being presented at the different premises. - In one embodiment, the
device 205 can determine a characteristic of the audience based on a monitored consumption history associated with the audience. In this example, the adjusting of the second segment of the content can include selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement to the media processor for presentation at the display. The selecting of the targeted advertisement can be according to the level of interest and according to the characteristic of the audience. - In one embodiment, the determining of the content context of the first segment can include applying image pattern recognition and speech pattern recognition to the first segment. In another embodiment,
device 205 can determine an expected volume of the expected audience reaction according to the content context of the first segment; determine a sensed volume of the sensed audience reaction according to the sensor data; and compare the sensed volume with the expected volume, where the level of interest in the first segment is determined based in part on the comparing of the sensed volume with the expected volume. -
FIG. 3 depicts an illustrative embodiment of amethod 300 used bysystems -
Method 300 can commence at 302 where a content context is determined by a system for a first segment of content being presented at a display for an audience. The system can be various devices or combinations of devices, including a media processor presenting the content or a network server. At 304, an expected audience reaction can be determined by the system according to the content context of the first segment. In one embodiment, the expected audience reaction can include an expected volume and/or an expected audio reaction context. - At 306, the system can obtain sensor data that was captured from a sensor device in proximity to the audience. The sensor data can be indicative of a sensed audience reaction to the first segment of the content. In one embodiment, the sensed audience reaction can include a sensed volume and/or a sensed audio reaction context.
- At 308 the system can compare the sensed audience reaction (e.g., the sensed volume and/or the sensed audio reaction context) with the expected audience reaction (e.g., the expected volume and/or the expected audio reaction context) to determine change in level of interest for the first segment. If there is no change, then
method 300 can return to 302 to continue monitoring for changes in interest to subsequent segments. If on the other hand there is a change in level of interest then at 310 the system can adjust a second segment of the content according to the changed level of interest to generate an adjusted second segment for presentation at the display. For example, if it is determined that the level of interest has increased then the second segment can be adjusted according to that increased level of interest. - In one embodiment, the sensor device can be an audio recorder of a media processor presenting the content at the display, and
method 300 can include applying speech recognition to the sensor data to identify words spoken by the audience; and determining the sensed audio reaction context according to the words. In one embodiment, the adjusting of the second segment of the content can include adjusting a resolution of the second segment according to the level of interest. - In one embodiment, the adjusting of the second segment of the content can include increasing a resolution of the second segment being presented by a first media processor at the display, where the adjusting of the resolution of the second segment is performed in conjunction with decreasing of another resolution of second content being presented by a second media processor, and where the first and second media processors are located at a same premises. In one embodiment,
method 300 can include transmitting, by the system over a network, the content to a media processor presenting the content at the display, where the obtaining of the sensor data by the system comprises receiving the sensor data captured from the sensor device housed in the media processor. - In one embodiment, the display can be a group of displays located at different premises, where the audience is a group of users located at the different premises, and where the sensor device is a group of sensor devices that are each housed by a corresponding media processor presenting the content at one of the group of displays. In one embodiment, the
method 300 can include determining a characteristic of the audience based on a monitored consumption history associated with the audience, where the adjusting of the second segment of the content comprises selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement for presentation at the display, and where the selecting of the targeted advertisement is according to the level of interest and according to the characteristic of the audience. - In one embodiment, the determining of the content context of the first segment can be based on metadata associated with the content. In one embodiment, the determining of the content context of the first segment can include applying image pattern recognition and speech pattern recognition to the first segment.
- While for purposes of simplicity of explanation, the respective processes are shown and described as a series of blocks in
FIG. 3 , it is to be understood and appreciated that the claimed subject matter is not limited by the order of the blocks, as some blocks may occur in different orders and/or concurrently with other blocks from what is depicted and described herein. Moreover, not all illustrated blocks may be required to implement the methods described herein. -
FIG. 4 depicts an illustrative embodiment of a first communication system 400 for delivering media content. The communication system 400 can represent an interactive television network, such as an Internet Protocol Television (IPTV) media system. Communication system 400 can be overlaid or operably coupled withsystems FIGS. 1 and 2 as another representative embodiment of communication system 400. For instance, one or more devices illustrated in the communication system 400 ofFIG. 4 can adjust content based on the reaction of the audience. The audience reaction can be compared to an expected audience reaction that is determined from a context of the particular segment of content that the audience is viewing. The audience reaction can be utilized to determine a level of interest or a change in the level of interest which can trigger an adjustment of the content. Various criteria can be utilized for comparing the actual and expected audience reaction, such as monitoring sound volume, determining a context of the audience reaction, such as whether there is excitement directed towards the content or whether the captured sounds are part of a conversation indicating disinterest in the content. - The IPTV media system can include a super head-end office (SHO) 410 with at least one super headend office server (SHS) 411 which receives media content from satellite and/or terrestrial communication systems. In the present context, media content can represent, for example, audio content, moving image content such as 2D or 3D videos, video games, virtual reality content, still image content, and combinations thereof. The
SHS server 411 can forward packets associated with the media content to one or more video head-end servers (VHS) 414 via a network of video head-end offices (VHO) 412 according to a multicast communication protocol. - The
VHS 414 can distribute multimedia broadcast content via anaccess network 418 to commercial and/orresidential buildings 402 housing a gateway 404 (such as a residential or commercial gateway). Theaccess network 418 can represent a group of digital subscriber line access multiplexers (DSLAMs) located in a central office or a service area interface that provide broadband services over fiber optical links or copper twistedpairs 419 tobuildings 402. Thegateway 404 can use communication technology to distribute broadcast signals tomedia processors 406 such as Set-Top Boxes (STBs) which in turn present broadcast channels tomedia devices 408 such as computers or television sets managed in some instances by a media controller 407 (such as an infrared or RF remote controller). - The
gateway 404, themedia processors 406, andmedia devices 408 can utilize tethered communication technologies (such as coaxial, powerline or phone line wiring) or can operate over a wireless access protocol such as Wireless Fidelity (WiFi), Bluetooth®, Zigbee or other present or next generation local or personal area wireless network technologies. By way of these interfaces, unicast communications can also be invoked between themedia processors 406 and subsystems of the IPTV media system for services such as video-on-demand (VoD), browsing an electronic programming guide (EPG), or other infrastructure services. - A satellite
broadcast television system 429 can be used in the media system ofFIG. 4 . The satellite broadcast television system can be overlaid, operably coupled with, or replace the IPTV system as another representative embodiment of communication system 400. In this embodiment, signals transmitted by asatellite 415 that include media content can be received by asatellite dish receiver 431 coupled to thebuilding 402. Modulated signals received by thesatellite dish receiver 431 can be transferred to themedia processors 406 for demodulating, decoding, encoding, and/or distributing broadcast channels to themedia devices 408. Themedia processors 406 can be equipped with a broadband port to an Internet Service Provider (ISP)network 432 to enable interactive services such as VoD and EPG as described above. - In yet another embodiment, an analog or digital cable broadcast distribution system such as
cable TV system 433 can be overlaid, operably coupled with, or replace the IPTV system and/or the satellite TV system as another representative embodiment of communication system 400. In this embodiment, thecable TV system 433 can also provide Internet, telephony, and interactive media services. System 400 enables various types of interactive television and/or services including IPTV, cable and/or satellite. - The subject disclosure can apply to other present or next generation over-the-air and/or landline media content services system.
- Some of the network elements of the IPTV media system can be coupled to one or
more computing devices 430, a portion of which can operate as a web server for providing web portal services over theISP network 432 towireline media devices 408 orwireless communication devices 416. - Communication system 400 can also provide for all or a portion of the
computing devices 430 to function as a content adjustment server (herein referred to as server 430). Theserver 430 can use computing and communication technology to performfunction 462, which can include among other things, determining a content context of a first segment of content being presented by a media processor at a display for an audience; determining an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context; receiving sensor data captured from a sensor device in proximity to the audience, the sensor data indicative of a sensed audience reaction to the first segment of the content; applying speech recognition to the sensor data to identify words spoken by the audience; determining a sensed audio reaction context according to the words; comparing the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment; and/or adjusting a second segment of the content according to the level of interest to generate an adjusted second segment for presentation at the display. - For instance, function 462 of
server 430 can be similar to the functions describednetwork server 130 ofsystem 100. Themedia processors 406 andwireless communication devices 416 can be provisioned withsoftware functions server 430. For instance, functions 46Y and 46Z ofmedia processors 406 andwireless communication devices 416 can be similar to the functions described for thecommunication devices 106 ofFIG. 1 in accordance withmethod 300. - Multiple forms of media services can be offered to media devices over landline technologies such as those described above. Additionally, media services can be offered to media devices by way of a wireless
access base station 417 operating according to common wireless access protocols such as Global System for Mobile or GSM, Code Division Multiple Access or CDMA, Time Division Multiple Access or TDMA, Universal Mobile Telecommunications or UMTS, World interoperability for Microwave or WiMAX, Software Defined Radio or SDR, Long Term Evolution or LTE, and so on. Other present and next generation wide area wireless access network technologies can be used in one or more embodiments of the subject disclosure. -
FIG. 5 depicts an illustrative embodiment of acommunication device 500.Communication device 500 can serve in whole or in part as an illustrative embodiment of the devices depicted insystems method 300. As an example,communication device 500 can determine a content context of a first segment of content; present the first segment of the content at a display for an audience; determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context; obtain sensor data captured from a sensor device in proximity to the display, the sensor data indicative of a sensed audience reaction to the first segment of the content; apply speech pattern recognition to the sensor data to determine a sensed audio reaction context of the sensed audience reaction; compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment; increase a resolution of a second segment of the content according to a determination that the level of interest has increased, where the increasing of the resolution of the second segment generates an adjusted second segment; present the adjusted second segment at the display; and/or provide a notice to a second media processor that causes the second media processor to decrease another resolution of other content being presented by the second media processor. - In one embodiment,
communication device 500 can determine an expected volume of the expected audience reaction according to the content context of the first segment; determine a sensed volume of the sensed audience reaction according to the sensor data; and compare the sensed volume with the expected volume, where the level of interest in the first segment is determined based in part on the comparing of the sensed volume with the expected volume. In one embodiment,communication device 500 can present a notification at the display indicating that the resolution of the second segment of the content will be increased prior to the presenting of the adjusted second segment at the display. -
Communication device 500 can comprise a wireline and/or wireless transceiver 502 (herein transceiver 502), a user interface (UI) 504, apower supply 514, alocation receiver 516, amotion sensor 518, anorientation sensor 520, and acontroller 506 for managing operations thereof. Thetransceiver 502 can support short-range or long-range wireless access technologies such as Bluetooth®, ZigBee®, WiFi, DECT, or cellular communication technologies, just to mention a few (Bluetooth® and ZigBee® are trademarks registered by the Bluetooth® Special Interest Group and the ZigBee® Alliance, respectively). Cellular technologies can include, for example, CDMA-1X, UMTS/HSDPA, GSM/GPRS, TDMA/EDGE, EV/DO, WiMAX, SDR, LTE, as well as other next generation wireless communication technologies as they arise. Thetransceiver 502 can also be adapted to support circuit-switched wireline access technologies (such as PSTN), packet-switched wireline access technologies (such as TCP/IP, VoIP, etc.), and combinations thereof. - The
UI 504 can include a depressible or touch-sensitive keypad 508 with a navigation mechanism such as a roller ball, a joystick, a mouse, or a navigation disk for manipulating operations of thecommunication device 500. Thekeypad 508 can be an integral part of a housing assembly of thecommunication device 500 or an independent device operably coupled thereto by a tethered wireline interface (such as a USB cable) or a wireless interface supporting for example Bluetooth®. Thekeypad 508 can represent a numeric keypad commonly used by phones, and/or a QWERTY keypad with alphanumeric keys. TheUI 504 can further include adisplay 510 such as monochrome or color LCD (Liquid Crystal Display), OLED (Organic Light Emitting Diode) or other suitable display technology for conveying images to an end user of thecommunication device 500. In an embodiment where thedisplay 510 is touch-sensitive, a portion or all of thekeypad 508 can be presented by way of thedisplay 510 with navigation features. - The
display 510 can use touch screen technology to also serve as a user interface for detecting user input. As a touch screen display, thecommunication device 500 can be adapted to present a user interface with graphical user interface (GUI) elements that can be selected by a user with a touch of a finger. Thetouch screen display 510 can be equipped with capacitive, resistive or other forms of sensing technology to detect how much surface area of a user's finger has been placed on a portion of the touch screen display. This sensing information can be used to control the manipulation of the GUI elements or other functions of the user interface. Thedisplay 510 can be an integral part of the housing assembly of thecommunication device 500 or an independent device communicatively coupled thereto by a tethered wireline interface (such as a cable) or a wireless interface. - The
UI 504 can also include anaudio system 512 that utilizes audio technology for conveying low volume audio (such as audio heard in proximity of a human ear) and high volume audio (such as speakerphone for hands free operation). Theaudio system 512 can further include a microphone for receiving audible signals of an end user. Theaudio system 512 can also be used for voice recognition applications. TheUI 504 can further include animage sensor 513 such as a charged coupled device (CCD) camera for capturing still or moving images. - The
power supply 514 can utilize common power management technologies such as replaceable and rechargeable batteries, supply regulation technologies, and/or charging system technologies for supplying energy to the components of thecommunication device 500 to facilitate long-range or short-range portable applications. Alternatively, or in combination, the charging system can utilize external power sources such as DC power supplied over a physical interface such as a USB port or other suitable tethering technologies. - The
location receiver 516 can utilize location technology such as a global positioning system (GPS) receiver capable of assisted GPS for identifying a location of thecommunication device 500 based on signals generated by a constellation of GPS satellites, which can be used for facilitating location services such as navigation. Themotion sensor 518 can utilize motion sensing technology such as an accelerometer, a gyroscope, or other suitable motion sensing technology to detect motion of thecommunication device 500 in three-dimensional space. Theorientation sensor 520 can utilize orientation sensing technology such as a magnetometer to detect the orientation of the communication device 500 (north, south, west, and east, as well as combined orientations in degrees, minutes, or other suitable orientation metrics). - The
communication device 500 can use thetransceiver 502 to also determine a proximity to a cellular, WiFi, Bluetooth®, or other wireless access points by sensing techniques such as utilizing a received signal strength indicator (RSSI) and/or signal time of arrival (TOA) or time of flight (TOF) measurements. Thecontroller 506 can utilize computing technologies such as a microprocessor, a digital signal processor (DSP), programmable gate arrays, application specific integrated circuits, and/or a video processor with associated storage memory such as Flash, ROM, RAM, SRAM, DRAM or other storage technologies for executing computer instructions, controlling, and processing data supplied by the aforementioned components of thecommunication device 500. - Other components not shown in
FIG. 5 can be used in one or more embodiments of the subject disclosure. For instance, thecommunication device 500 can include a reset button (not shown). The reset button can be used to reset thecontroller 506 of thecommunication device 500. In yet another embodiment, thecommunication device 500 can also include a factory default setting button positioned, for example, below a small hole in a housing assembly of thecommunication device 500 to force thecommunication device 500 to re-establish factory settings. In this embodiment, a user can use a protruding object such as a pen or paper clip tip to reach into the hole and depress the default setting button. Thecommunication device 500 can also include a slot for adding or removing an identity module such as a Subscriber Identity Module (SIM) card. SIM cards can be used for identifying subscriber services, executing programs, storing subscriber data, and so forth. - The
communication device 500 as described herein can operate with more or less of the circuit components shown inFIG. 5 . These variant embodiments can be used in one or more embodiments of the subject disclosure. - The
communication device 500 can be adapted to perform the functions ofdevices FIGS. 1 and 2 , themedia processor 406, themedia devices 408, or theportable communication devices 416 ofFIG. 4 . It will be appreciated that thecommunication device 500 can also represent other devices that can operate insystems FIGS. 1 and 2 ], communication system 400 ofFIG. 4 such as a gaming console and a media player. In addition, thecontroller 506 can be adapted in various embodiments to perform the functions 462-466. - Upon reviewing the aforementioned embodiments, it would be evident to an artisan with ordinary skill in the art that said embodiments can be modified, reduced, or enhanced without departing from the scope of the claims described below. For example, other user activity can be monitored to detect a level of interest of a user of the audience, such as detecting a text message, a voice call, or surfing the web on a mobile device of a user of the audience during the presentation of the content which may be indicative of a lack of interest.
- Other actions can be triggered by the determination of the level of interest or a change in the level of interest, such as proposing different content to the audience or to a subset of the audience.
- In one embodiment, the determination of the level of interest can be multi-tiered based on different thresholds. For example, a scene in a movie can be determined to have a content context of frightening. The expected audience reaction can be determined to be a scream or movement of the audience. If the sensed audience reaction is both a scream and movement then a determination of very high interest can be made. If the sensed audience reaction is either the scream or the movement then a determination of high interest can be made. If the sensed audience reaction is neither the scream nor the movement, but rather silence, then a determination of medium interest can be made. If the sensed audience reaction is neither the scream nor the movement, but rather laughter, then a determination of low interest can be made. The content adjustments can then be made according to the group of different interest levels. For instance, the highest level of interest can result in an increase to the highest resolution and insertion of a targeted advertisement for another movie in the same genre while the lowest level of interest can result in a notification that a different movie of a different genre is currently available.
- In one embodiment, sensed audience responses can be distinguished amongst users of the audience and a content adjustment determination can be made according to individual responses. For example, sensed audio can be analyzed via speech pattern recognition to determine that first and second users are engaged in a conversation that is not related to the content (e.g., a football game) being presented. The sensed audio can be further analyzed to determine that third and fourth users are cheering. The determination of which users are speaking can be performed in a number of different ways, such as based on accessible voice samples. In this example, a content adjustment can be made (e.g., a change in resolution of the presented football game or an insertion of a commercial for another football game that is scheduled in the near future) based on the interest of the third and fourth users.
- In one embodiment, the content can be broadcast content which is then adjusted by a network server responsive to detecting the change in level of interest. Other embodiments can be used in the subject disclosure.
- It should be understood that devices described in the exemplary embodiments can be in communication with each other via various wireless and/or wired methodologies. The methodologies can be links that are described as coupled, connected and so forth, which can include unidirectional and/or bidirectional communication over wireless paths and/or wired paths that utilize one or more of various protocols or methodologies, where the coupling and/or connection can be direct (e.g., no intervening processing device) and/or indirect (e.g., an intermediary processing device such as a router).
-
FIG. 6 depicts an exemplary diagrammatic representation of a machine in the form of acomputer system 600 within which a set of instructions, when executed, may cause the machine to perform any one or more of the methods described above. One or more instances of the machine can operate, for example, as themedia processor 106, thenetwork server 130, theserver 430, themedia processor 406 and other devices ofFIGS. 1-2 and 4 in order to adjust content according to a comparison of an expected audience reaction with a sensed audience reaction. In some embodiments, the machine may be connected (e.g., using a network 626) to other machines. In a networked deployment, the machine may operate in the capacity of a server or a client user machine in a server-client user network environment, or as a peer machine in a peer-to-peer (or distributed) network environment. - The machine may comprise a server computer, a client user computer, a personal computer (PC), a tablet, a smart phone, a laptop computer, a desktop computer, a control system, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine. It will be understood that a communication device of the subject disclosure includes broadly any electronic device that provides voice, video or data communication. Further, while a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methods discussed herein.
- The
computer system 600 may include a processor (or controller) 602 (e.g., a central processing unit (CPU)), a graphics processing unit (GPU, or both), amain memory 604 and astatic memory 606, which communicate with each other via abus 608. Thecomputer system 600 may further include a display unit 610 (e.g., a liquid crystal display (LCD), a flat panel, or a solid state display). Thecomputer system 600 may include an input device 612 (e.g., a keyboard), a cursor control device 614 (e.g., a mouse), adisk drive unit 616, a signal generation device 618 (e.g., a speaker or remote control) and anetwork interface device 620. In distributed environments, the embodiments described in the subject disclosure can be adapted to utilizemultiple display units 610 controlled by two ormore computer systems 600. In this configuration, presentations described by the subject disclosure may in part be shown in a first of thedisplay units 610, while the remaining portion is presented in a second of thedisplay units 610. - The
disk drive unit 616 may include a tangible computer-readable storage medium 622 on which is stored one or more sets of instructions (e.g., software 624) embodying any one or more of the methods or functions described herein, including those methods illustrated above. Theinstructions 624 may also reside, completely or at least partially, within themain memory 604, thestatic memory 606, and/or within theprocessor 602 during execution thereof by thecomputer system 600. Themain memory 604 and theprocessor 602 also may constitute tangible computer-readable storage media. - Dedicated hardware implementations including, but not limited to, application specific integrated circuits, programmable logic arrays and other hardware devices can likewise be constructed to implement the methods described herein. Application specific integrated circuits and programmable logic array can use downloadable instructions for executing state machines and/or circuit configurations to implement embodiments of the subject disclosure. Applications that may include the apparatus and systems of various embodiments broadly include a variety of electronic and computer systems. Some embodiments implement functions in two or more specific interconnected hardware modules or devices with related control and data signals communicated between and through the modules, or as portions of an application-specific integrated circuit. Thus, the example system is applicable to software, firmware, and hardware implementations.
- In accordance with various embodiments of the subject disclosure, the operations or methods described herein are intended for operation as software programs or instructions running on or executed by a computer processor or other computing device, and which may include other forms of instructions manifested as a state machine implemented with logic components in an application specific integrated circuit or field programmable gate array. Furthermore, software implementations (e.g., software programs, instructions, etc.) including, but not limited to, distributed processing or component/object distributed processing, parallel processing, or virtual machine processing can also be constructed to implement the methods described herein. It is further noted that a computing device such as a processor, a controller, a state machine or other suitable device for executing instructions to perform operations or methods may perform such operations directly or indirectly by way of one or more intermediate devices directed by the computing device.
- While the tangible computer-
readable storage medium 622 is shown in an example embodiment to be a single medium, the term “tangible computer-readable storage medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “tangible computer-readable storage medium” shall also be taken to include any non-transitory medium that is capable of storing or encoding a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methods of the subject disclosure. The term “non-transitory” as in a non-transitory computer-readable storage includes without limitation memories, drives, devices and anything tangible but not a signal per se. - The term “tangible computer-readable storage medium” shall accordingly be taken to include, but not be limited to: solid-state memories such as a memory card or other package that houses one or more read-only (non-volatile) memories, random access memories, or other re-writable (volatile) memories, a magneto-optical or optical medium such as a disk or tape, or other tangible media which can be used to store information. Accordingly, the disclosure is considered to include any one or more of a tangible computer-readable storage medium, as listed herein and including art-recognized equivalents and successor media, in which the software implementations herein are stored.
- Although the present specification describes components and functions implemented in the embodiments with reference to particular standards and protocols, the disclosure is not limited to such standards and protocols. Each of the standards for Internet and other packet switched network transmission (e.g., TCP/IP, UDP/IP, HTML, HTTP) represent examples of the state of the art. Such standards are from time-to-time superseded by faster or more efficient equivalents having essentially the same functions. Wireless standards for device detection (e.g., RFID), short-range communications (e.g., Bluetooth®, WiFi, Zigbee®), and long-range communications (e.g., WiMAX, GSM, CDMA, LTE) can be used by
computer system 600. - The illustrations of embodiments described herein are intended to provide a general understanding of the structure of various embodiments, and they are not intended to serve as a complete description of all the elements and features of apparatus and systems that might make use of the structures described herein. Many other embodiments will be apparent to those of skill in the art upon reviewing the above description. The exemplary embodiments can include combinations of features and/or steps from multiple embodiments. Other embodiments may be utilized and derived therefrom, such that structural and logical substitutions and changes may be made without departing from the scope of this disclosure. Figures are also merely representational and may not be drawn to scale. Certain proportions thereof may be exaggerated, while others may be minimized. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
- Although specific embodiments have been illustrated and described herein, it should be appreciated that any arrangement which achieves the same or similar purpose may be substituted for the embodiments described or shown by the subject disclosure. The subject disclosure is intended to cover any and all adaptations or variations of various embodiments. Combinations of the above embodiments, and other embodiments not specifically described herein, can be used in the subject disclosure. For instance, one or more features from one or more embodiments can be combined with one or more features of one or more other embodiments. In one or more embodiments, features that are positively recited can also be negatively recited and excluded from the embodiment with or without replacement by another structural and/or functional feature. The steps or functions described with respect to the embodiments of the subject disclosure can be performed in any order. The steps or functions described with respect to the embodiments of the subject disclosure can be performed alone or in combination with other steps or functions of the subject disclosure, as well as from other embodiments or from other steps that have not been described in the subject disclosure. Further, more than or less than all of the features described with respect to an embodiment can also be utilized.
- Less than all of the steps or functions described with respect to the exemplary processes or methods can also be performed in one or more of the exemplary embodiments. Further, the use of numerical terms to describe a device, component, step or function, such as first, second, third, and so forth, is not intended to describe an order or function unless expressly stated so. The use of the terms first, second, third and so forth, is generally to distinguish between devices, components, steps or functions unless expressly stated otherwise. Additionally, one or more devices or components described with respect to the exemplary embodiments can facilitate one or more functions, where the facilitating (e.g., facilitating access or facilitating establishing a connection) can include less than every step needed to perform the function or can include all of the steps needed to perform the function.
- In one or more embodiments, a processor (which can include a controller or circuit) has been described that performs various functions. It should be understood that the processor can be multiple processors, which can include distributed processors or parallel processors in a single machine or multiple machines. The processor can be used in supporting a virtual processing environment. The virtual processing environment may support one or more virtual machines representing computers, servers, or other computing devices. In such virtual machines, components such as microprocessors and storage devices may be virtualized or logically represented. The processor can include a state machine, application specific integrated circuit, and/or programmable gate array including a Field PGA. In one or more embodiments, when a processor executes instructions to perform “operations”, this can include the processor performing the operations directly and/or facilitating, directing, or cooperating with another device or component to perform the operations.
- The Abstract of the Disclosure is provided with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, it can be seen that various features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separately claimed subject matter.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/733,399 US20200177956A1 (en) | 2015-11-11 | 2020-01-03 | Method and apparatus for content adaptation based on audience monitoring |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/938,421 US10542315B2 (en) | 2015-11-11 | 2015-11-11 | Method and apparatus for content adaptation based on audience monitoring |
US16/733,399 US20200177956A1 (en) | 2015-11-11 | 2020-01-03 | Method and apparatus for content adaptation based on audience monitoring |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/938,421 Continuation US10542315B2 (en) | 2015-11-11 | 2015-11-11 | Method and apparatus for content adaptation based on audience monitoring |
Publications (1)
Publication Number | Publication Date |
---|---|
US20200177956A1 true US20200177956A1 (en) | 2020-06-04 |
Family
ID=58664388
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/938,421 Expired - Fee Related US10542315B2 (en) | 2015-11-11 | 2015-11-11 | Method and apparatus for content adaptation based on audience monitoring |
US16/733,399 Abandoned US20200177956A1 (en) | 2015-11-11 | 2020-01-03 | Method and apparatus for content adaptation based on audience monitoring |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/938,421 Expired - Fee Related US10542315B2 (en) | 2015-11-11 | 2015-11-11 | Method and apparatus for content adaptation based on audience monitoring |
Country Status (1)
Country | Link |
---|---|
US (2) | US10542315B2 (en) |
Families Citing this family (47)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10437288B2 (en) | 2014-10-06 | 2019-10-08 | Fasetto, Inc. | Portable storage device with modular power and housing system |
US9467718B1 (en) | 2015-05-06 | 2016-10-11 | Echostar Broadcasting Corporation | Apparatus, systems and methods for a content commentary community |
EP4080794A1 (en) | 2016-01-06 | 2022-10-26 | TVision Insights, Inc. | Systems and methods for assessing viewer engagement |
US11540009B2 (en) | 2016-01-06 | 2022-12-27 | Tvision Insights, Inc. | Systems and methods for assessing viewer engagement |
WO2017119604A1 (en) * | 2016-01-08 | 2017-07-13 | 주식회사 아이플래테아 | Audience rating calculation server, audience rating calculation method, and audience rating calculation remote device |
US10268689B2 (en) | 2016-01-28 | 2019-04-23 | DISH Technologies L.L.C. | Providing media content based on user state detection |
US11146865B2 (en) | 2016-03-03 | 2021-10-12 | Comcast Cable Communications, Llc | Determining points of interest in a content item |
US11012719B2 (en) * | 2016-03-08 | 2021-05-18 | DISH Technologies L.L.C. | Apparatus, systems and methods for control of sporting event presentation based on viewer engagement |
US10984036B2 (en) | 2016-05-03 | 2021-04-20 | DISH Technologies L.L.C. | Providing media content based on media element preferences |
US10057746B1 (en) | 2016-11-16 | 2018-08-21 | Wideorbit, Inc. | Method and system for detecting a user device in an environment associated with a content presentation system presenting content |
US10764381B2 (en) | 2016-12-23 | 2020-09-01 | Echostar Technologies L.L.C. | Communications channels in media systems |
US11196826B2 (en) | 2016-12-23 | 2021-12-07 | DISH Technologies L.L.C. | Communications channels in media systems |
US10390084B2 (en) | 2016-12-23 | 2019-08-20 | DISH Technologies L.L.C. | Communications channels in media systems |
US11708051B2 (en) | 2017-02-03 | 2023-07-25 | Fasetto, Inc. | Systems and methods for data storage in keyed devices |
JP6565084B2 (en) * | 2017-03-29 | 2019-08-28 | 本田技研工業株式会社 | Object authentication apparatus and object authentication method |
US11770574B2 (en) * | 2017-04-20 | 2023-09-26 | Tvision Insights, Inc. | Methods and apparatus for multi-television measurements |
US10511888B2 (en) | 2017-09-19 | 2019-12-17 | Sony Corporation | Calibration system for audience response capture and analysis of media content |
US10763630B2 (en) | 2017-10-19 | 2020-09-01 | Fasetto, Inc. | Portable electronic device connection systems |
US11869039B1 (en) | 2017-11-13 | 2024-01-09 | Wideorbit Llc | Detecting gestures associated with content displayed in a physical environment |
WO2019109033A1 (en) | 2017-12-01 | 2019-06-06 | Fasetto, Inc. | Systems and methods for improved data encryption |
US20190205469A1 (en) * | 2018-01-04 | 2019-07-04 | International Business Machines Corporation | Cognitive system and method to select best suited audio content based on individual's past reactions |
US11043230B1 (en) * | 2018-01-25 | 2021-06-22 | Wideorbit Inc. | Targeted content based on user reactions |
US10462422B1 (en) * | 2018-04-09 | 2019-10-29 | Facebook, Inc. | Audio selection based on user engagement |
CN112292708B (en) * | 2018-04-17 | 2022-06-17 | 法斯埃托股份有限公司 | Presentation system and method with real-time feedback |
US10616649B2 (en) * | 2018-07-19 | 2020-04-07 | Rovi Guides, Inc. | Providing recommendations based on passive microphone detections |
US10897647B1 (en) * | 2018-07-25 | 2021-01-19 | Imdb.Com, Inc. | Ascertaining audience reactions for a media item |
US11412295B2 (en) * | 2018-10-02 | 2022-08-09 | Comcast Cable Communications, Llc | Systems and methods for determining usage information |
US11817005B2 (en) * | 2018-10-31 | 2023-11-14 | International Business Machines Corporation | Internet of things public speaking coach |
US11037550B2 (en) | 2018-11-30 | 2021-06-15 | Dish Network L.L.C. | Audio-based link generation |
US20200186862A1 (en) | 2018-12-05 | 2020-06-11 | Roku, Inc. | Downloadable avod content |
US11240555B2 (en) | 2018-12-05 | 2022-02-01 | Roku, Inc. | Providing over-the-air content to any device |
US10834452B2 (en) * | 2019-01-02 | 2020-11-10 | International Business Machines Corporation | Dynamic live feed recommendation on the basis of user real time reaction to a live feed |
US11089356B2 (en) * | 2019-03-26 | 2021-08-10 | Rovi Guides, Inc. | Systems and methods for media content hand-off based on type of buffered data |
JP2021135803A (en) * | 2020-02-27 | 2021-09-13 | 富士フイルムビジネスイノベーション株式会社 | Document preparation device and program |
US11128925B1 (en) * | 2020-02-28 | 2021-09-21 | Nxp Usa, Inc. | Media presentation system using audience and audio feedback for playback level control |
US11310553B2 (en) | 2020-06-19 | 2022-04-19 | Apple Inc. | Changing resource utilization associated with a media object based on an engagement score |
US20220174357A1 (en) * | 2020-11-30 | 2022-06-02 | At&T Intellectual Property I, L.P. | Simulating audience feedback in remote broadcast events |
US20220377413A1 (en) * | 2021-05-21 | 2022-11-24 | Rovi Guides, Inc. | Methods and systems for personalized content based on captured gestures |
US20230019723A1 (en) * | 2021-07-14 | 2023-01-19 | Rovi Guides, Inc. | Interactive supplemental content system |
US11849179B2 (en) * | 2021-12-21 | 2023-12-19 | Disney Enterprises, Inc. | Characterizing audience engagement based on emotional alignment with characters |
US20230217086A1 (en) * | 2021-12-30 | 2023-07-06 | Interwise Ltd. | Providing and Using a Branching Narrative Content Service |
US11910061B2 (en) * | 2022-05-23 | 2024-02-20 | Rovi Guides, Inc. | Leveraging emotional transitions in media to modulate emotional impact of secondary content |
US11871081B2 (en) * | 2022-05-23 | 2024-01-09 | Rovi Guides, Inc. | Leveraging emotional transitions in media to modulate emotional impact of secondary content |
US20240022791A1 (en) * | 2022-07-15 | 2024-01-18 | Rovi Guides, Inc. | Systems and methods to adapt a schedule to be played by a media player |
US11632413B1 (en) * | 2022-07-18 | 2023-04-18 | Rovi Guides, Inc. | Methods and systems for streaming media content |
US12063260B2 (en) | 2022-08-31 | 2024-08-13 | Rovi Guides, Inc. | Intelligent delivery and playout to prevent stalling in video streaming |
US11936948B1 (en) * | 2023-01-24 | 2024-03-19 | Roku, Inc. | Method and system for generating a visual composition of user reactions in a shared content viewing session |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120324493A1 (en) * | 2011-06-17 | 2012-12-20 | Microsoft Corporation | Interest-based video streams |
US20130145385A1 (en) * | 2011-12-02 | 2013-06-06 | Microsoft Corporation | Context-based ratings and recommendations for media |
US20140168277A1 (en) * | 2011-05-10 | 2014-06-19 | Cisco Technology Inc. | Adaptive Presentation of Content |
US8898687B2 (en) * | 2012-04-04 | 2014-11-25 | Microsoft Corporation | Controlling a media program based on a media reaction |
US20140366049A1 (en) * | 2013-06-11 | 2014-12-11 | Nokia Corporation | Method, apparatus and computer program product for gathering and presenting emotional response to an event |
US20150189376A1 (en) * | 2013-12-31 | 2015-07-02 | Echostar Technologies L.L.C. | Transponder loss and join-in-progress channel monitoring |
Family Cites Families (55)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6078349A (en) * | 1995-06-07 | 2000-06-20 | Compaq Computer Corporation | Process and system for increasing the display resolution of a point-to-point video transmission relative to the actual amount of video data sent |
US6708335B1 (en) | 1999-08-18 | 2004-03-16 | Webtv Networks, Inc. | Tracking viewing behavior of advertisements on a home entertainment system |
US8910199B2 (en) | 2000-02-25 | 2014-12-09 | Interval Licensing Llc | Targeted television content display |
US6904408B1 (en) | 2000-10-19 | 2005-06-07 | Mccarthy John | Bionet method, system and personalized web content manager responsive to browser viewers' psychological preferences, behavioral responses and physiological stress indicators |
US7284201B2 (en) * | 2001-09-20 | 2007-10-16 | Koninklijke Philips Electronics N.V. | User attention-based adaptation of quality level to improve the management of real-time multi-media content delivery and distribution |
US8561095B2 (en) | 2001-11-13 | 2013-10-15 | Koninklijke Philips N.V. | Affective television monitoring and control in response to physiological data |
US7409639B2 (en) | 2003-06-19 | 2008-08-05 | Accenture Global Services Gmbh | Intelligent collaborative media |
US8156444B1 (en) | 2003-12-31 | 2012-04-10 | Google Inc. | Systems and methods for determining a user interface attribute |
US20050289582A1 (en) | 2004-06-24 | 2005-12-29 | Hitachi, Ltd. | System and method for capturing and using biometrics to review a product, service, creative work or thing |
US20070005752A1 (en) | 2005-06-29 | 2007-01-04 | Jitendra Chawla | Methods and apparatuses for monitoring attention of a user during a collaboration session |
WO2006043925A1 (en) | 2004-10-14 | 2006-04-27 | John Mccarthy | Bionet system for personalizing web content manager |
US7492821B2 (en) * | 2005-02-08 | 2009-02-17 | International Business Machines Corporation | System and method for selective image capture, transmission and reconstruction |
US20070214471A1 (en) * | 2005-03-23 | 2007-09-13 | Outland Research, L.L.C. | System, method and computer program product for providing collective interactive television experiences |
US20070038516A1 (en) | 2005-08-13 | 2007-02-15 | Jeff Apple | Systems, methods, and computer program products for enabling an advertiser to measure user viewing of and response to an advertisement |
US20080169930A1 (en) | 2007-01-17 | 2008-07-17 | Sony Computer Entertainment Inc. | Method and system for measuring a user's level of attention to content |
JP2008205861A (en) * | 2007-02-20 | 2008-09-04 | Matsushita Electric Ind Co Ltd | Viewing and listening quality determining apparatus, viewing and listening quality determining method, viewing and listening quality determining program, and storage medium |
US8764652B2 (en) | 2007-03-08 | 2014-07-01 | The Nielson Company (US), LLC. | Method and system for measuring and ranking an “engagement” response to audiovisual or interactive media, products, or activities using physiological signals |
US8392503B2 (en) | 2007-06-19 | 2013-03-05 | Cisco Technology, Inc. | Reporting participant attention level to presenter during a web-based rich-media conference |
US7673088B2 (en) * | 2007-06-29 | 2010-03-02 | Microsoft Corporation | Multi-tasking interference model |
US7889073B2 (en) | 2008-01-31 | 2011-02-15 | Sony Computer Entertainment America Llc | Laugh detector and system and method for tracking an emotional response to a media presentation |
US8630192B2 (en) | 2009-01-28 | 2014-01-14 | Headwater Partners I Llc | Verifiable and accurate service usage monitoring for intermediate networking devices |
US20100070987A1 (en) | 2008-09-12 | 2010-03-18 | At&T Intellectual Property I, L.P. | Mining viewer responses to multimedia content |
US8156054B2 (en) | 2008-12-04 | 2012-04-10 | At&T Intellectual Property I, L.P. | Systems and methods for managing interactions between an individual and an entity |
US20120017231A1 (en) | 2009-09-15 | 2012-01-19 | Jackson Chao | Behavior monitoring system |
US8875167B2 (en) | 2009-09-21 | 2014-10-28 | Mobitv, Inc. | Implicit mechanism for determining user response to media |
US20110321073A1 (en) | 2010-06-23 | 2011-12-29 | Yarvis Mark D | Techniques for customization |
US8949871B2 (en) | 2010-09-08 | 2015-02-03 | Opentv, Inc. | Smart media selection based on viewer user presence |
US8493390B2 (en) * | 2010-12-08 | 2013-07-23 | Sony Computer Entertainment America, Inc. | Adaptive displays using gaze tracking |
US9009298B2 (en) | 2010-12-10 | 2015-04-14 | The Nielsen Company (Us), Llc | Methods and apparatus to determine audience engagement indices associated with media presentations |
CN103688245A (en) | 2010-12-30 | 2014-03-26 | 安比恩特兹公司 | Information processing using a population of data acquisition devices |
US20120210383A1 (en) * | 2011-02-11 | 2012-08-16 | Sayers Craig P | Presenting streaming media for an event |
US9049259B2 (en) | 2011-05-03 | 2015-06-02 | Onepatont Software Limited | System and method for dynamically providing visual action or activity news feed |
GB2494945A (en) * | 2011-09-22 | 2013-03-27 | Ibm | Targeted digital media content delivery based on emotional change |
US8635637B2 (en) | 2011-12-02 | 2014-01-21 | Microsoft Corporation | User interface presenting an animated avatar performing a media reaction |
US8943526B2 (en) * | 2011-12-02 | 2015-01-27 | Microsoft Corporation | Estimating engagement of consumers of presented content |
US9100685B2 (en) | 2011-12-09 | 2015-08-04 | Microsoft Technology Licensing, Llc | Determining audience state or interest using passive sensor data |
US20130205314A1 (en) * | 2012-02-07 | 2013-08-08 | Arun Ramaswamy | Methods and apparatus to select media based on engagement levels |
WO2012103820A2 (en) * | 2012-03-08 | 2012-08-09 | 华为技术有限公司 | Method, device, and system for highlighting party of interest |
WO2013184488A1 (en) | 2012-06-05 | 2013-12-12 | Almondnet, Inc. | Targeted television advertising based on a profile linked to an online device associated with a content-selecting device |
KR20150021039A (en) * | 2012-06-14 | 2015-02-27 | 톰슨 라이센싱 | Method, apparatus and system for determining viewer reaction to content elements |
US8931064B2 (en) | 2012-12-18 | 2015-01-06 | Bank Of America Corporation | Identity attribute exchange and validation ecosystem |
WO2014108194A1 (en) * | 2013-01-10 | 2014-07-17 | Telefonaktiebolaget L M Ericsson (Publ) | Apparatus and method for controlling adaptive streaming of media |
US20140298364A1 (en) * | 2013-03-26 | 2014-10-02 | Rawllin International Inc. | Recommendations for media content based on emotion |
US20140337880A1 (en) | 2013-05-08 | 2014-11-13 | Shayna Sorbel | Peer-to-Peer Communication Advertising Method |
US20140337868A1 (en) * | 2013-05-13 | 2014-11-13 | Microsoft Corporation | Audience-aware advertising |
CN105264906A (en) * | 2013-06-05 | 2016-01-20 | 汤姆逊许可公司 | Method and apparatus for content distribution for multiscreen viewing |
WO2015031671A1 (en) | 2013-08-30 | 2015-03-05 | Biscotti Inc. | Physical presence and advertising |
CN105830108A (en) | 2013-09-20 | 2016-08-03 | 交互数字专利控股公司 | Verification Of Ad Impressions In User-Adptive Multimedia Delivery Framework |
EP3058873A4 (en) * | 2013-10-17 | 2017-06-28 | Natsume Research Institute, Co., Ltd. | Device for measuring visual efficacy |
GB201402536D0 (en) * | 2014-02-13 | 2014-04-02 | Piksel Inc | Sensed content delivery |
US10264211B2 (en) * | 2014-03-14 | 2019-04-16 | Comcast Cable Communications, Llc | Adaptive resolution in software applications based on dynamic eye tracking |
US20150281783A1 (en) * | 2014-03-18 | 2015-10-01 | Vixs Systems, Inc. | Audio/video system with viewer-state based recommendations and methods for use therewith |
US10129312B2 (en) * | 2014-09-11 | 2018-11-13 | Microsoft Technology Licensing, Llc | Dynamic video streaming based on viewer activity |
US20160182955A1 (en) * | 2014-12-23 | 2016-06-23 | Rovi Guides, Inc. | Methods and systems for recommending media assets |
US10200725B2 (en) * | 2015-06-15 | 2019-02-05 | Intel Corporation | Adaptive data streaming based on virtual screen size |
-
2015
- 2015-11-11 US US14/938,421 patent/US10542315B2/en not_active Expired - Fee Related
-
2020
- 2020-01-03 US US16/733,399 patent/US20200177956A1/en not_active Abandoned
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140168277A1 (en) * | 2011-05-10 | 2014-06-19 | Cisco Technology Inc. | Adaptive Presentation of Content |
US20120324493A1 (en) * | 2011-06-17 | 2012-12-20 | Microsoft Corporation | Interest-based video streams |
US20130145385A1 (en) * | 2011-12-02 | 2013-06-06 | Microsoft Corporation | Context-based ratings and recommendations for media |
US8898687B2 (en) * | 2012-04-04 | 2014-11-25 | Microsoft Corporation | Controlling a media program based on a media reaction |
US20140366049A1 (en) * | 2013-06-11 | 2014-12-11 | Nokia Corporation | Method, apparatus and computer program product for gathering and presenting emotional response to an event |
US20150189376A1 (en) * | 2013-12-31 | 2015-07-02 | Echostar Technologies L.L.C. | Transponder loss and join-in-progress channel monitoring |
Also Published As
Publication number | Publication date |
---|---|
US10542315B2 (en) | 2020-01-21 |
US20170134803A1 (en) | 2017-05-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20200177956A1 (en) | Method and apparatus for content adaptation based on audience monitoring | |
US10491946B2 (en) | Methods and systems for providing multiple video content streams | |
US10812752B2 (en) | Method and apparatus to present multiple audio content | |
US10567810B2 (en) | Method and apparatus for managing bandwidth in providing communication services | |
US11509958B2 (en) | Automated content selection for groups | |
US10652622B2 (en) | Method and apparatus for providing content based upon a selected language | |
US10433011B2 (en) | Apparatus and method for providing programming information for media content to a wearable device | |
US10581779B2 (en) | Method and apparatus for managing content distribution according to social networks | |
US10257585B2 (en) | Method and apparatus for directed advertisement | |
US20210125231A1 (en) | Targeted user digital embedded advertising | |
US11457278B2 (en) | Method and apparatus for recording advertised media content |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |