Nothing Special   »   [go: up one dir, main page]

US20200177956A1 - Method and apparatus for content adaptation based on audience monitoring - Google Patents

Method and apparatus for content adaptation based on audience monitoring Download PDF

Info

Publication number
US20200177956A1
US20200177956A1 US16/733,399 US202016733399A US2020177956A1 US 20200177956 A1 US20200177956 A1 US 20200177956A1 US 202016733399 A US202016733399 A US 202016733399A US 2020177956 A1 US2020177956 A1 US 2020177956A1
Authority
US
United States
Prior art keywords
content
segment
audience
context
bandwidth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/733,399
Inventor
Venson Shaw
Sangar Dowlatkhah
Zhi Cui
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
AT&T Intellectual Property I LP
Original Assignee
AT&T Intellectual Property I LP
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by AT&T Intellectual Property I LP filed Critical AT&T Intellectual Property I LP
Priority to US16/733,399 priority Critical patent/US20200177956A1/en
Publication of US20200177956A1 publication Critical patent/US20200177956A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44218Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234363Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the spatial resolution, e.g. for clients with a lower screen resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/251Learning process for intelligent management, e.g. learning user preferences for recommending movies
    • H04N21/252Processing of multiple end-users' preferences to derive collaborative data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4126The peripheral being portable, e.g. PDAs or mobile phones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4126The peripheral being portable, e.g. PDAs or mobile phones
    • H04N21/41265The peripheral being portable, e.g. PDAs or mobile phones having a remote control device for bidirectional communication between the remote control device and client device
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42203Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/466Learning process for intelligent management, e.g. learning user preferences for recommending movies
    • H04N21/4667Processing of monitored end-user data, e.g. trend analysis based on the log file of viewer selections
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/812Monomedia components thereof involving advertisement data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/57Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for processing of video signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44204Monitoring of content usage, e.g. the number of times a movie has been viewed, copied or the amount which has been watched

Definitions

  • the subject disclosure relates to a method and apparatus for content adaptation based on audience monitoring
  • Selection of content for presentation to the users including movies, programming, advertising and the like is often performed based on generalizations associated with demographics, such as age and gender. However, not all users fit those generalizations which can lead to selection of content which is not of interest to a particular user.
  • FIGS. 1 and 2 depict illustrative embodiments of content adjustment systems that can adjust content according to monitored audience reaction
  • FIG. 3 depicts an illustrative embodiment of a method used in portions of the system described in FIGS. 1 and 2 ;
  • FIG. 4 depicts an illustrative embodiment of a communication system that provides media services including content adjustment based on audience reactions;
  • FIG. 5 depicts an illustrative embodiment of a communication device
  • FIG. 6 is a diagrammatic representation of a machine in the form of a computer system within which a set of instructions, when executed, may cause the machine to perform any one or more of the methods described herein.
  • the subject disclosure describes, among other things, illustrative embodiments for adjusting content based on the reaction of the audience.
  • the audience reaction can be compared to an expected audience reaction that is determined from a context of the particular segment of content that the audience is viewing.
  • the audience reaction can be utilized to determine a level of interest or a change in the level of interest which can trigger an adjustment of the content.
  • Various criteria can be utilized for comparing the actual and expected audience reaction, including monitoring sound volume or determining a context of the audience reaction, such as whether there is excitement directed towards the content or whether the captured sounds are part of a conversation indicating disinterest in the content, or even different environments, such as mobility state or the vibration pattern of the mobile device (via which content is delivered) in a car or a train.
  • Other embodiments are described in the subject disclosure.
  • One or more aspects of the subject disclosure are a method including determining, by a system including a processor, a content context of a first segment of content being presented at a display to an audience.
  • the method can include determining, by the system, an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected volume and an expected audio reaction context.
  • the method can include obtaining, by the system, sensor data captured from a sensor device in proximity to the audience, where the sensor data is indicative of a sensed audience reaction to the first segment of the content, and where the sensed audience reaction comprises a sensed volume and a sensed audio reaction context.
  • the method can include comparing, by the system, the sensed volume and the sensed audio reaction context with the expected volume and the expected audio reaction context to determine a level of interest in the first segment.
  • the method can include adjusting, by the system, a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display.
  • One or more aspects of the subject disclosure include a machine-readable storage medium, comprising executable instructions that, when executed by a media processor, facilitate performance of operations, including determining a content context of a first segment of content.
  • the media processor can present the first segment of the content at a display to an audience; and can determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context.
  • the media processor can obtain sensor data captured from a sensor device in proximity to the display, where the sensor data is indicative of a sensed audience reaction to the first segment of the content.
  • the media processor can apply speech pattern recognition to the sensor data to determine a sensed audio reaction context of the sensed audience reaction.
  • the media processor can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment.
  • the media processor can increase a resolution of a second segment of the content according to a determination that the level of interest has increased, where the increasing of the resolution of the second segment generates an adjusted second segment.
  • the media processor can present the adjusted second segment at the display.
  • the media processor can provide a notice to a second media processor that causes the second media processor to decrease another resolution of other content being presented by the second media processor.
  • One or more aspects of the subject disclosure include a network server having a processor; and a memory that stores executable instructions that, when executed by the processor, facilitate performance of operations, including determining a content context of a first segment of content being presented by a media processor at a display to an audience.
  • the network server can determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context.
  • the network server can receive sensor data captured from a sensor device in proximity to the audience, where the sensor data is indicative of a sensed audience reaction to the first segment of the content.
  • the network server can apply speech recognition to the sensor data to identify words spoken by the audience.
  • the network server can determine a sensed audio reaction context according to the words.
  • the network server can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment.
  • the network server can adjust a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display.
  • FIG. 1 depicts an illustrative embodiment of a system 100 in which content is presented by a media processor 106 (e.g., a set top box) at a display 108 (e.g., a television).
  • An audience 110 can be viewing or otherwise consuming the content.
  • Three users are illustrated for audience 110 , although the audience could be a single user or any other number of users.
  • the media processor 106 , the display 108 and the audience 110 are illustrated at a single location (e.g., a single premises), however, the media processor, the display and the audience can be groups of media processors, displays and audiences that are located at different locations, such as different premises.
  • the content can be provided to the media processor 106 by a network server 130 .
  • one or more sensor devices 150 can capture sensor data of a sensed audience reaction for the audience 110 .
  • the sensor data can be of various types including audio and/or image information.
  • the sensor device 150 can be integrated with the media processor 106 or can be otherwise in communication with the media processor to provide the sensor data to the media processor.
  • the sensor device 150 can be a separate device from the media processor 106 , where the sensor device provides the sensor data to the network server 130 with or without sharing the sensor data with the media processor.
  • Combinations of sensor devices 150 can also be utilized, such as a first sensor device that is an audio recorder of the media processor 106 and a second sensor device that is a camera for capturing images of the audience 110 .
  • the camera can be integrated with the media processor 106 or can be otherwise in communication with the media processor to provide the sensor data to the media processor, or the camera can be a separate device from the media processor, where the camera provides the sensor data to the network server 130 with or without sharing the sensor data with the media processor.
  • the sensor device 150 can be part of an end user device of one or more of the users in the audience 110 .
  • a mobile phone of a user of the audience 110 can record audio and provide that audio to the media processor 106 and/or to the network server 130 .
  • the sensor device 150 can capture other sensor data, such as motion detection, lighting and so forth.
  • System 100 enables determining a level of interest or a change in the level of interest for a particular segment of the content being presented so that a content adaptation can be performed. For example, a determination can be made for a content context of a first segment of the content being presented at the display 108 for the audience 110 .
  • the content context can be performed based on various factors and utilizing various techniques.
  • segments of the content can include or otherwise be associated with metadata that can be accessed where the metadata describes the particular content context of the particular segment.
  • a movie can include metadata that describes various scenes in the movie.
  • the content context can be categorized in such a way as to facilitate determining an expected audience reaction to the scene. For instance, the metadata can describe that the scene is funny from which it can be determined that audience laughter is expected.
  • the metadata can describe that the scene is scary from which it can be determined that an audience scream and/or an audience movement is expected.
  • the metadata can describe that the scene is somber from which it can be determined that audience quiet is expected.
  • the metadata can also be utilized for determining that which is not an expected reaction of the audience, such as it is not expected for there to be laughter in a scene that is described by the metadata as tragic.
  • pattern recognition can be applied to the segment of the content to determine the content context and thus the expected audience reaction.
  • audio and/or image pattern recognition can be applied to the segment to determine that a movie scene is scary or tragic.
  • Detected words and/or sounds e.g., laughter, scream, sighing, crying
  • detected actions, or other characteristics of the scene detected from the pattern recognition can be utilized in determining the particular content context.
  • the pattern recognition can also be performed on live events that are being presented as the content. For instance, pattern recognition can be utilized to determine that a football team is close to the end zone from which it can be determined that the expected audience reaction can be excited (for an audience member favoring the team on offense) or discouraged (for an audience member favoring the team on defense).
  • an expected volume of audience reaction can be determined, such as determining that an audio reaction to a touchdown in a football game should be above a threshold level if the audience member is interested in the particular game.
  • the determination of the content context can be performed by various devices, such as by the media processor 106 and/or the network server 130 .
  • System 100 enables comparing an expected audience reaction with a sensed audience reaction to determine a level of interest or a change in the level of interest for a particular segment of the content being presented so that a content adaptation can be performed.
  • the comparison can be performed by various devices, such as by the media processor 106 and/or the network server 130 .
  • the expected and sensed audience reactions can be characterized in various ways, including audio volume, spoken words, sounds, movement and so forth.
  • a content adaptation can be performed.
  • the content adaptation can be of various types. For example, characteristics of the content can be adjusted, such as a resolution or quality, in response to a determination of an increased level of interest. For instance, a determination can be made that the audience has become more interested in a football game because the offensive team is near the end zone and may score a touchdown. As described above, this determination can be performed according to a comparison of the expected audience reaction and the sensed audience reaction.
  • the content i.e., the football game
  • the content can be adjusted to a high resolution so that the upcoming plays near the end zone are of a higher quality.
  • the system 100 can continue to monitor for content context of subsequent segments, expected audience reactions for those subsequent segments, and sensed audience reactions for those subsequent segments, from which changes in level of interest can be determined. In this example, if it is determined that the level of interest decreases then the content can be adjusted again, such as lowering the resolution (for plays that are deemed to be of less interest to the audience).
  • the adjustment of the resolution can be performed in conjunction with adjustment to other content being provided to the location. For instance, an increase in resolution at the media processor 106 resulting in an increase in bandwidth can be offset by a decrease in resolution at another media processor so that the bandwidth usage for the location are maintained relatively constant.
  • the location can have multiple audiences and multiple media processors where the monitoring for content context, expected audience reactions, and sensed audience reactions are performed for those multiple audiences, from which changes in level of interest can be determined for those multiple audiences.
  • the changes in level of interest for each of the multiple audiences (at different displays watching different content) can be compared and the content adjustment can be made according to a higher level of change in interest. For instance, a first audience determined to have a higher level of interest can be presented with a high definition version of first content while a second audience determined to have a lower level of interest can be presented with a standard definition version of second content.
  • the adjustment of the content can include inserting a different segment into the content, such as a targeted advertisement or a different version of a scene.
  • the adjustment of the content can include replacing the subsequent segments with completely different content, such as switching to a different episode of a show or a different movie.
  • the monitoring of the audience reaction can be performed on an individual basis. For example, a captured audio reaction can be identified with a particular user of the audience and compared with an expected audio reaction for that user. User profiles and/or historical reactions for that particular user can be analyzed in conjunction with the content context to determine the expected audience reaction for that particular user. As an example, it can be determined that the content context for a scene in a movie is humor. It can be further determined that a particular user in the past has reacted with laughter at a particular volume. The sensed laughter for that particular user can be compared to the particular volume of past laughter to determine a level of interest in the scene for the particular user.
  • a notification can be presented at the display 108 indicating that the content adjustment is going to occur (e.g., resolution of a subsequent segment of the content will be increased).
  • a characteristic of the audience can be determined (e.g., by the media processor 106 and/or the network server 130 ) based on a monitored consumption history associated with the audience.
  • the adjusting of a subsequent segment of the content can include selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement for presentation at the display 108 , where the selecting of the targeted advertisement is according to the determined level of interest and according to the characteristic of the audience.
  • system 100 can teach a video broadcast system how to insert and/or adapt content according to an audience signal-to-noise ratio.
  • the audience signal-to-noise ratio can be a measure of expected audio to actual audio and can filter out background noise.
  • System 100 can capture the mood, emotion, need and/or requirement of the audience at that particular moment. As a result, the content inserted and/or adapted to the video broadcast system can instantaneously capture the need and/or requirement of the audience at that moment.
  • System 100 can enable content insertion to be adapted and personalized according to the audience signal-to-noise which is distinct from blindly or randomly inserting content.
  • FIG. 2 depicts an illustrative embodiment of system 200 that enables content adaptation according to monitored audience reactions to segments of the content. By determining an expected audience reaction according to the context of the particular segment of content being presented, system 200 can detect a level of interest or a change in the level of interest via a comparison to a sensed audience reaction.
  • a device 205 can determine a content context of a first segment of content being presented by the media processor to the audience 110 .
  • the content context determination can be performed utilizing various techniques including analyzing metadata that describes the segment content and/or performing pattern recognition (e.g., audio and/or image) to determine the segment context.
  • An expected audience reaction can be determined according to the content context of the first segment, such as determining an expected audio reaction context (e.g., laughter, screaming, enthusiasm, anger, and so forth).
  • the device 205 can obtain sensor data captured from one or more sensor devices in proximity to the audience, such as an audio recorder, a camera, a motion detector and so forth.
  • the sensor data can be indicative of a sensed audience reaction to the first segment of the content.
  • the sensor data can be filtered, such as removing background noise, to isolate the sensed audience reaction.
  • speech recognition can be applied to the sensor data to identify words spoken by the audience. Based on the speech recognition, a sensed audio reaction context can be determined according to the words. For example, words indicative of anger or words indicative of grief can be identified.
  • Device 205 can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment. For instance, device 205 can determine that a sad scene in a movie should invoke a griefful audience reaction. If the comparison of the expected and sensed audience reactions is not a match then device 205 can determine that there is a low level of interest in the scene. In one embodiment, the device 205 can detect a sensed audience reaction which is contrary to what is expected, such as sensing laughter during a sad scene. Based on this detection, device 205 can determine there is little interest in the scene.
  • device 205 can adjust a second segment of the content according to the level of interest to generate an adjusted second segment for presentation at the display.
  • the adjusting of the second segment of the content can include adjusting a resolution of the second segment according to the level of interest.
  • the adjusting of the resolution of the second segment can be performed in conjunction with decreasing of another resolution of second content being presented by another media processor, where the media processor and the other media processor are located at a same premises.
  • the display can be a group of displays located at different premises, where the audience 110 is a group of users located at the different premises.
  • the sensor device can be a group of sensor devices that capture sensor data for the multiple audiences at the different premises.
  • the content adjustment can be performed for content being presented at the different premises.
  • the device 205 can determine a characteristic of the audience based on a monitored consumption history associated with the audience.
  • the adjusting of the second segment of the content can include selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement to the media processor for presentation at the display. The selecting of the targeted advertisement can be according to the level of interest and according to the characteristic of the audience.
  • the determining of the content context of the first segment can include applying image pattern recognition and speech pattern recognition to the first segment.
  • device 205 can determine an expected volume of the expected audience reaction according to the content context of the first segment; determine a sensed volume of the sensed audience reaction according to the sensor data; and compare the sensed volume with the expected volume, where the level of interest in the first segment is determined based in part on the comparing of the sensed volume with the expected volume.
  • FIG. 3 depicts an illustrative embodiment of a method 300 used by systems 100 and 200 for adjusting content according to a level of audience interest determined from a comparison of an expected audience reaction to an actual audience reaction.
  • Method 300 can commence at 302 where a content context is determined by a system for a first segment of content being presented at a display for an audience.
  • the system can be various devices or combinations of devices, including a media processor presenting the content or a network server.
  • an expected audience reaction can be determined by the system according to the content context of the first segment.
  • the expected audience reaction can include an expected volume and/or an expected audio reaction context.
  • the system can obtain sensor data that was captured from a sensor device in proximity to the audience.
  • the sensor data can be indicative of a sensed audience reaction to the first segment of the content.
  • the sensed audience reaction can include a sensed volume and/or a sensed audio reaction context.
  • the system can compare the sensed audience reaction (e.g., the sensed volume and/or the sensed audio reaction context) with the expected audience reaction (e.g., the expected volume and/or the expected audio reaction context) to determine change in level of interest for the first segment. If there is no change, then method 300 can return to 302 to continue monitoring for changes in interest to subsequent segments. If on the other hand there is a change in level of interest then at 310 the system can adjust a second segment of the content according to the changed level of interest to generate an adjusted second segment for presentation at the display. For example, if it is determined that the level of interest has increased then the second segment can be adjusted according to that increased level of interest.
  • the sensed audience reaction e.g., the sensed volume and/or the sensed audio reaction context
  • the expected audience reaction e.g., the expected volume and/or the expected audio reaction context
  • the sensor device can be an audio recorder of a media processor presenting the content at the display, and method 300 can include applying speech recognition to the sensor data to identify words spoken by the audience; and determining the sensed audio reaction context according to the words.
  • the adjusting of the second segment of the content can include adjusting a resolution of the second segment according to the level of interest.
  • the adjusting of the second segment of the content can include increasing a resolution of the second segment being presented by a first media processor at the display, where the adjusting of the resolution of the second segment is performed in conjunction with decreasing of another resolution of second content being presented by a second media processor, and where the first and second media processors are located at a same premises.
  • method 300 can include transmitting, by the system over a network, the content to a media processor presenting the content at the display, where the obtaining of the sensor data by the system comprises receiving the sensor data captured from the sensor device housed in the media processor.
  • the display can be a group of displays located at different premises, where the audience is a group of users located at the different premises, and where the sensor device is a group of sensor devices that are each housed by a corresponding media processor presenting the content at one of the group of displays.
  • the method 300 can include determining a characteristic of the audience based on a monitored consumption history associated with the audience, where the adjusting of the second segment of the content comprises selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement for presentation at the display, and where the selecting of the targeted advertisement is according to the level of interest and according to the characteristic of the audience.
  • the determining of the content context of the first segment can be based on metadata associated with the content. In one embodiment, the determining of the content context of the first segment can include applying image pattern recognition and speech pattern recognition to the first segment.
  • FIG. 4 depicts an illustrative embodiment of a first communication system 400 for delivering media content.
  • the communication system 400 can represent an interactive television network, such as an Internet Protocol Television (IPTV) media system.
  • IPTV Internet Protocol Television
  • Communication system 400 can be overlaid or operably coupled with systems 100 and 200 of FIGS. 1 and 2 as another representative embodiment of communication system 400 .
  • one or more devices illustrated in the communication system 400 of FIG. 4 can adjust content based on the reaction of the audience.
  • the audience reaction can be compared to an expected audience reaction that is determined from a context of the particular segment of content that the audience is viewing.
  • the audience reaction can be utilized to determine a level of interest or a change in the level of interest which can trigger an adjustment of the content.
  • Various criteria can be utilized for comparing the actual and expected audience reaction, such as monitoring sound volume, determining a context of the audience reaction, such as whether there is excitement directed towards the content or whether the captured sounds are part of a conversation indicating disinterest in the content.
  • the IPTV media system can include a super head-end office (SHO) 410 with at least one super headend office server (SHS) 411 which receives media content from satellite and/or terrestrial communication systems.
  • media content can represent, for example, audio content, moving image content such as 2D or 3D videos, video games, virtual reality content, still image content, and combinations thereof.
  • the SHS server 411 can forward packets associated with the media content to one or more video head-end servers (VHS) 414 via a network of video head-end offices (VHO) 412 according to a multicast communication protocol.
  • the VHS 414 can distribute multimedia broadcast content via an access network 418 to commercial and/or residential buildings 402 housing a gateway 404 (such as a residential or commercial gateway).
  • the access network 418 can represent a group of digital subscriber line access multiplexers (DSLAMs) located in a central office or a service area interface that provide broadband services over fiber optical links or copper twisted pairs 419 to buildings 402 .
  • DSLAMs digital subscriber line access multiplexers
  • the gateway 404 can use communication technology to distribute broadcast signals to media processors 406 such as Set-Top Boxes (STBs) which in turn present broadcast channels to media devices 408 such as computers or television sets managed in some instances by a media controller 407 (such as an infrared or RF remote controller).
  • STBs Set-Top Boxes
  • media devices 408 such as computers or television sets managed in some instances by a media controller 407 (such as an infrared or RF remote controller).
  • the gateway 404 , the media processors 406 , and media devices 408 can utilize tethered communication technologies (such as coaxial, powerline or phone line wiring) or can operate over a wireless access protocol such as Wireless Fidelity (WiFi), Bluetooth®, Zigbee or other present or next generation local or personal area wireless network technologies.
  • WiFi Wireless Fidelity
  • Bluetooth® Bluetooth®
  • Zigbee Zigbee
  • unicast communications can also be invoked between the media processors 406 and subsystems of the IPTV media system for services such as video-on-demand (VoD), browsing an electronic programming guide (EPG), or other infrastructure services.
  • VoD video-on-demand
  • EPG electronic programming guide
  • a satellite broadcast television system 429 can be used in the media system of FIG. 4 .
  • the satellite broadcast television system can be overlaid, operably coupled with, or replace the IPTV system as another representative embodiment of communication system 400 .
  • signals transmitted by a satellite 415 that include media content can be received by a satellite dish receiver 431 coupled to the building 402 .
  • Modulated signals received by the satellite dish receiver 431 can be transferred to the media processors 406 for demodulating, decoding, encoding, and/or distributing broadcast channels to the media devices 408 .
  • the media processors 406 can be equipped with a broadband port to an Internet Service Provider (ISP) network 432 to enable interactive services such as VoD and EPG as described above.
  • ISP Internet Service Provider
  • an analog or digital cable broadcast distribution system such as cable TV system 433 can be overlaid, operably coupled with, or replace the IPTV system and/or the satellite TV system as another representative embodiment of communication system 400 .
  • the cable TV system 433 can also provide Internet, telephony, and interactive media services.
  • System 400 enables various types of interactive television and/or services including IPTV, cable and/or satellite.
  • the subject disclosure can apply to other present or next generation over-the-air and/or landline media content services system.
  • Some of the network elements of the IPTV media system can be coupled to one or more computing devices 430 , a portion of which can operate as a web server for providing web portal services over the ISP network 432 to wireline media devices 408 or wireless communication devices 416 .
  • Communication system 400 can also provide for all or a portion of the computing devices 430 to function as a content adjustment server (herein referred to as server 430 ).
  • the server 430 can use computing and communication technology to perform function 462 , which can include among other things, determining a content context of a first segment of content being presented by a media processor at a display for an audience; determining an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context; receiving sensor data captured from a sensor device in proximity to the audience, the sensor data indicative of a sensed audience reaction to the first segment of the content; applying speech recognition to the sensor data to identify words spoken by the audience; determining a sensed audio reaction context according to the words; comparing the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment; and/or adjusting a second segment of the content according to the level of interest to generate an adjusted second segment for presentation at the display.
  • function 462 of server 430 can be similar to the functions described network server 130 of system 100 .
  • the media processors 406 and wireless communication devices 416 can be provisioned with software functions 464 and 466 , respectively, to utilize the services of server 430 .
  • functions 46 Y and 46 Z of media processors 406 and wireless communication devices 416 can be similar to the functions described for the communication devices 106 of FIG. 1 in accordance with method 300 .
  • media services can be offered to media devices over landline technologies such as those described above. Additionally, media services can be offered to media devices by way of a wireless access base station 417 operating according to common wireless access protocols such as Global System for Mobile or GSM, Code Division Multiple Access or CDMA, Time Division Multiple Access or TDMA, Universal Mobile Telecommunications or UMTS, World interoperability for Microwave or WiMAX, Software Defined Radio or SDR, Long Term Evolution or LTE, and so on.
  • GSM Global System for Mobile or GSM
  • CDMA Code Division Multiple Access
  • TDMA Time Division Multiple Access or TDMA
  • Universal Mobile Telecommunications or UMTS Universal Mobile Telecommunications or UMTS
  • World interoperability for Microwave or WiMAX Software Defined Radio or SDR, Long Term Evolution or LTE, and so on.
  • Other present and next generation wide area wireless access network technologies can be used in one or more embodiments of the subject disclosure.
  • FIG. 5 depicts an illustrative embodiment of a communication device 500 .
  • Communication device 500 can serve in whole or in part as an illustrative embodiment of the devices depicted in systems 100 , 200 and 400 and can be configured to perform portions of method 300 .
  • communication device 500 can determine a content context of a first segment of content; present the first segment of the content at a display for an audience; determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context; obtain sensor data captured from a sensor device in proximity to the display, the sensor data indicative of a sensed audience reaction to the first segment of the content; apply speech pattern recognition to the sensor data to determine a sensed audio reaction context of the sensed audience reaction; compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment; increase a resolution of a second segment of the content according to a determination that the level of interest has increased, where the increasing of the resolution of the second segment generates an adjusted second segment; present the adjusted second segment at the display; and/or provide a notice to a second media processor that causes the second media processor to decrease another resolution of other content being presented by the second media processor.
  • communication device 500 can determine an expected volume of the expected audience reaction according to the content context of the first segment; determine a sensed volume of the sensed audience reaction according to the sensor data; and compare the sensed volume with the expected volume, where the level of interest in the first segment is determined based in part on the comparing of the sensed volume with the expected volume.
  • communication device 500 can present a notification at the display indicating that the resolution of the second segment of the content will be increased prior to the presenting of the adjusted second segment at the display.
  • Communication device 500 can comprise a wireline and/or wireless transceiver 502 (herein transceiver 502 ), a user interface (UI) 504 , a power supply 514 , a location receiver 516 , a motion sensor 518 , an orientation sensor 520 , and a controller 506 for managing operations thereof.
  • the transceiver 502 can support short-range or long-range wireless access technologies such as Bluetooth®, ZigBee®, WiFi, DECT, or cellular communication technologies, just to mention a few (Bluetooth® and ZigBee® are trademarks registered by the Bluetooth® Special Interest Group and the ZigBee® Alliance, respectively).
  • Cellular technologies can include, for example, CDMA-1X, UMTS/HSDPA, GSM/GPRS, TDMA/EDGE, EV/DO, WiMAX, SDR, LTE, as well as other next generation wireless communication technologies as they arise.
  • the transceiver 502 can also be adapted to support circuit-switched wireline access technologies (such as PSTN), packet-switched wireline access technologies (such as TCP/IP, VoIP, etc.), and combinations thereof.
  • the UI 504 can include a depressible or touch-sensitive keypad 508 with a navigation mechanism such as a roller ball, a joystick, a mouse, or a navigation disk for manipulating operations of the communication device 500 .
  • the keypad 508 can be an integral part of a housing assembly of the communication device 500 or an independent device operably coupled thereto by a tethered wireline interface (such as a USB cable) or a wireless interface supporting for example Bluetooth®.
  • the keypad 508 can represent a numeric keypad commonly used by phones, and/or a QWERTY keypad with alphanumeric keys.
  • the UI 504 can further include a display 510 such as monochrome or color LCD (Liquid Crystal Display), OLED (Organic Light Emitting Diode) or other suitable display technology for conveying images to an end user of the communication device 500 .
  • a display 510 such as monochrome or color LCD (Liquid Crystal Display), OLED (Organic Light Emitting Diode) or other suitable display technology for conveying images to an end user of the communication device 500 .
  • a display 510 is touch-sensitive, a portion or all of the keypad 508 can be presented by way of the display 510 with navigation features.
  • the display 510 can use touch screen technology to also serve as a user interface for detecting user input.
  • the communication device 500 can be adapted to present a user interface with graphical user interface (GUI) elements that can be selected by a user with a touch of a finger.
  • GUI graphical user interface
  • the touch screen display 510 can be equipped with capacitive, resistive or other forms of sensing technology to detect how much surface area of a user's finger has been placed on a portion of the touch screen display. This sensing information can be used to control the manipulation of the GUI elements or other functions of the user interface.
  • the display 510 can be an integral part of the housing assembly of the communication device 500 or an independent device communicatively coupled thereto by a tethered wireline interface (such as a cable) or a wireless interface.
  • the UI 504 can also include an audio system 512 that utilizes audio technology for conveying low volume audio (such as audio heard in proximity of a human ear) and high volume audio (such as speakerphone for hands free operation).
  • the audio system 512 can further include a microphone for receiving audible signals of an end user.
  • the audio system 512 can also be used for voice recognition applications.
  • the UI 504 can further include an image sensor 513 such as a charged coupled device (CCD) camera for capturing still or moving images.
  • CCD charged coupled device
  • the power supply 514 can utilize common power management technologies such as replaceable and rechargeable batteries, supply regulation technologies, and/or charging system technologies for supplying energy to the components of the communication device 500 to facilitate long-range or short-range portable applications.
  • the charging system can utilize external power sources such as DC power supplied over a physical interface such as a USB port or other suitable tethering technologies.
  • the location receiver 516 can utilize location technology such as a global positioning system (GPS) receiver capable of assisted GPS for identifying a location of the communication device 500 based on signals generated by a constellation of GPS satellites, which can be used for facilitating location services such as navigation.
  • GPS global positioning system
  • the motion sensor 518 can utilize motion sensing technology such as an accelerometer, a gyroscope, or other suitable motion sensing technology to detect motion of the communication device 500 in three-dimensional space.
  • the orientation sensor 520 can utilize orientation sensing technology such as a magnetometer to detect the orientation of the communication device 500 (north, south, west, and east, as well as combined orientations in degrees, minutes, or other suitable orientation metrics).
  • the communication device 500 can use the transceiver 502 to also determine a proximity to a cellular, WiFi, Bluetooth®, or other wireless access points by sensing techniques such as utilizing a received signal strength indicator (RSSI) and/or signal time of arrival (TOA) or time of flight (TOF) measurements.
  • the controller 506 can utilize computing technologies such as a microprocessor, a digital signal processor (DSP), programmable gate arrays, application specific integrated circuits, and/or a video processor with associated storage memory such as Flash, ROM, RAM, SRAM, DRAM or other storage technologies for executing computer instructions, controlling, and processing data supplied by the aforementioned components of the communication device 500 .
  • computing technologies such as a microprocessor, a digital signal processor (DSP), programmable gate arrays, application specific integrated circuits, and/or a video processor with associated storage memory such as Flash, ROM, RAM, SRAM, DRAM or other storage technologies for executing computer instructions, controlling, and processing data supplied by the aforementioned components of the communication device
  • the communication device 500 can include a reset button (not shown).
  • the reset button can be used to reset the controller 506 of the communication device 500 .
  • the communication device 500 can also include a factory default setting button positioned, for example, below a small hole in a housing assembly of the communication device 500 to force the communication device 500 to re-establish factory settings.
  • a user can use a protruding object such as a pen or paper clip tip to reach into the hole and depress the default setting button.
  • the communication device 500 can also include a slot for adding or removing an identity module such as a Subscriber Identity Module (SIM) card. SIM cards can be used for identifying subscriber services, executing programs, storing subscriber data, and so forth.
  • SIM Subscriber Identity Module
  • the communication device 500 as described herein can operate with more or less of the circuit components shown in FIG. 5 . These variant embodiments can be used in one or more embodiments of the subject disclosure.
  • the communication device 500 can be adapted to perform the functions of devices 106 , 130 , 205 of FIGS. 1 and 2 , the media processor 406 , the media devices 408 , or the portable communication devices 416 of FIG. 4 . It will be appreciated that the communication device 500 can also represent other devices that can operate in systems 100 and 200 of FIGS. 1 and 2 ], communication system 400 of FIG. 4 such as a gaming console and a media player. In addition, the controller 506 can be adapted in various embodiments to perform the functions 462 - 466 .
  • Other actions can be triggered by the determination of the level of interest or a change in the level of interest, such as proposing different content to the audience or to a subset of the audience.
  • the determination of the level of interest can be multi-tiered based on different thresholds. For example, a scene in a movie can be determined to have a content context of frightening.
  • the expected audience reaction can be determined to be a scream or movement of the audience. If the sensed audience reaction is both a scream and movement then a determination of very high interest can be made. If the sensed audience reaction is either the scream or the movement then a determination of high interest can be made. If the sensed audience reaction is neither the scream nor the movement, but rather silence, then a determination of medium interest can be made. If the sensed audience reaction is neither the scream nor the movement, but rather laughter, then a determination of low interest can be made.
  • the content adjustments can then be made according to the group of different interest levels. For instance, the highest level of interest can result in an increase to the highest resolution and insertion of a targeted advertisement for another movie in the same genre while the lowest level of interest can result in a notification that a different movie of a different genre is currently available.
  • sensed audience responses can be distinguished amongst users of the audience and a content adjustment determination can be made according to individual responses.
  • sensed audio can be analyzed via speech pattern recognition to determine that first and second users are engaged in a conversation that is not related to the content (e.g., a football game) being presented.
  • the sensed audio can be further analyzed to determine that third and fourth users are cheering.
  • the determination of which users are speaking can be performed in a number of different ways, such as based on accessible voice samples.
  • a content adjustment can be made (e.g., a change in resolution of the presented football game or an insertion of a commercial for another football game that is scheduled in the near future) based on the interest of the third and fourth users.
  • the content can be broadcast content which is then adjusted by a network server responsive to detecting the change in level of interest.
  • a network server responsive to detecting the change in level of interest.
  • devices described in the exemplary embodiments can be in communication with each other via various wireless and/or wired methodologies.
  • the methodologies can be links that are described as coupled, connected and so forth, which can include unidirectional and/or bidirectional communication over wireless paths and/or wired paths that utilize one or more of various protocols or methodologies, where the coupling and/or connection can be direct (e.g., no intervening processing device) and/or indirect (e.g., an intermediary processing device such as a router).
  • FIG. 6 depicts an exemplary diagrammatic representation of a machine in the form of a computer system 600 within which a set of instructions, when executed, may cause the machine to perform any one or more of the methods described above.
  • One or more instances of the machine can operate, for example, as the media processor 106 , the network server 130 , the server 430 , the media processor 406 and other devices of FIGS. 1-2 and 4 in order to adjust content according to a comparison of an expected audience reaction with a sensed audience reaction.
  • the machine may be connected (e.g., using a network 626 ) to other machines.
  • the machine may operate in the capacity of a server or a client user machine in a server-client user network environment, or as a peer machine in a peer-to-peer (or distributed) network environment.
  • the machine may comprise a server computer, a client user computer, a personal computer (PC), a tablet, a smart phone, a laptop computer, a desktop computer, a control system, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine.
  • a communication device of the subject disclosure includes broadly any electronic device that provides voice, video or data communication.
  • the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methods discussed herein.
  • the computer system 600 may include a processor (or controller) 602 (e.g., a central processing unit (CPU)), a graphics processing unit (GPU, or both), a main memory 604 and a static memory 606 , which communicate with each other via a bus 608 .
  • the computer system 600 may further include a display unit 610 (e.g., a liquid crystal display (LCD), a flat panel, or a solid state display).
  • the computer system 600 may include an input device 612 (e.g., a keyboard), a cursor control device 614 (e.g., a mouse), a disk drive unit 616 , a signal generation device 618 (e.g., a speaker or remote control) and a network interface device 620 .
  • the embodiments described in the subject disclosure can be adapted to utilize multiple display units 610 controlled by two or more computer systems 600 .
  • presentations described by the subject disclosure may in part be shown in a first of the display units 610 , while the remaining portion is presented in a second of the display units 610 .
  • the disk drive unit 616 may include a tangible computer-readable storage medium 622 on which is stored one or more sets of instructions (e.g., software 624 ) embodying any one or more of the methods or functions described herein, including those methods illustrated above.
  • the instructions 624 may also reside, completely or at least partially, within the main memory 604 , the static memory 606 , and/or within the processor 602 during execution thereof by the computer system 600 .
  • the main memory 604 and the processor 602 also may constitute tangible computer-readable storage media.
  • Dedicated hardware implementations including, but not limited to, application specific integrated circuits, programmable logic arrays and other hardware devices can likewise be constructed to implement the methods described herein.
  • Application specific integrated circuits and programmable logic array can use downloadable instructions for executing state machines and/or circuit configurations to implement embodiments of the subject disclosure.
  • Applications that may include the apparatus and systems of various embodiments broadly include a variety of electronic and computer systems. Some embodiments implement functions in two or more specific interconnected hardware modules or devices with related control and data signals communicated between and through the modules, or as portions of an application-specific integrated circuit.
  • the example system is applicable to software, firmware, and hardware implementations.
  • the operations or methods described herein are intended for operation as software programs or instructions running on or executed by a computer processor or other computing device, and which may include other forms of instructions manifested as a state machine implemented with logic components in an application specific integrated circuit or field programmable gate array.
  • software implementations e.g., software programs, instructions, etc.
  • a computing device such as a processor, a controller, a state machine or other suitable device for executing instructions to perform operations or methods may perform such operations directly or indirectly by way of one or more intermediate devices directed by the computing device.
  • tangible computer-readable storage medium 622 is shown in an example embodiment to be a single medium, the term “tangible computer-readable storage medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions.
  • tangible computer-readable storage medium shall also be taken to include any non-transitory medium that is capable of storing or encoding a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methods of the subject disclosure.
  • non-transitory as in a non-transitory computer-readable storage includes without limitation memories, drives, devices and anything tangible but not a signal per se.
  • tangible computer-readable storage medium shall accordingly be taken to include, but not be limited to: solid-state memories such as a memory card or other package that houses one or more read-only (non-volatile) memories, random access memories, or other re-writable (volatile) memories, a magneto-optical or optical medium such as a disk or tape, or other tangible media which can be used to store information. Accordingly, the disclosure is considered to include any one or more of a tangible computer-readable storage medium, as listed herein and including art-recognized equivalents and successor media, in which the software implementations herein are stored.
  • Each of the standards for Internet and other packet switched network transmission (e.g., TCP/IP, UDP/IP, HTML, HTTP) represent examples of the state of the art. Such standards are from time-to-time superseded by faster or more efficient equivalents having essentially the same functions.
  • Wireless standards for device detection e.g., RFID
  • short-range communications e.g., Bluetooth®, WiFi, Zigbee®
  • long-range communications e.g., WiMAX, GSM, CDMA, LTE
  • facilitating e.g., facilitating access or facilitating establishing a connection
  • the facilitating can include less than every step needed to perform the function or can include all of the steps needed to perform the function.
  • a processor (which can include a controller or circuit) has been described that performs various functions. It should be understood that the processor can be multiple processors, which can include distributed processors or parallel processors in a single machine or multiple machines.
  • the processor can be used in supporting a virtual processing environment.
  • the virtual processing environment may support one or more virtual machines representing computers, servers, or other computing devices. In such virtual machines, components such as microprocessors and storage devices may be virtualized or logically represented.
  • the processor can include a state machine, application specific integrated circuit, and/or programmable gate array including a Field PGA.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Computing Systems (AREA)
  • Marketing (AREA)
  • Business, Economics & Management (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

Aspects of the subject disclosure may include, for example, determining a content context of a first segment of content being presented by a media processor at a display to an audience, determining an expected audience reaction according to the content context of the first segment, receiving sensor data captured from a sensor device in proximity to the audience where the sensor data is indicative of a sensed audience reaction to the first segment of the content, comparing the sensed audience reaction with the expected audience reaction to determine a level of interest in the first segment, and adjusting a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display. Other embodiments are disclosed.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 14/938,421 filed Nov. 11, 2015, pending. All sections of the aforementioned application are incorporated herein by reference in their entirety.
  • FIELD OF THE DISCLOSURE
  • The subject disclosure relates to a method and apparatus for content adaptation based on audience monitoring
  • BACKGROUND
  • Users have various options to select from in communication services. Selection of content for presentation to the users including movies, programming, advertising and the like is often performed based on generalizations associated with demographics, such as age and gender. However, not all users fit those generalizations which can lead to selection of content which is not of interest to a particular user.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
  • FIGS. 1 and 2 depict illustrative embodiments of content adjustment systems that can adjust content according to monitored audience reaction;
  • FIG. 3 depicts an illustrative embodiment of a method used in portions of the system described in FIGS. 1 and 2;
  • FIG. 4 depicts an illustrative embodiment of a communication system that provides media services including content adjustment based on audience reactions;
  • FIG. 5 depicts an illustrative embodiment of a communication device; and
  • FIG. 6 is a diagrammatic representation of a machine in the form of a computer system within which a set of instructions, when executed, may cause the machine to perform any one or more of the methods described herein.
  • DETAILED DESCRIPTION
  • The subject disclosure describes, among other things, illustrative embodiments for adjusting content based on the reaction of the audience. The audience reaction can be compared to an expected audience reaction that is determined from a context of the particular segment of content that the audience is viewing. The audience reaction can be utilized to determine a level of interest or a change in the level of interest which can trigger an adjustment of the content. Various criteria can be utilized for comparing the actual and expected audience reaction, including monitoring sound volume or determining a context of the audience reaction, such as whether there is excitement directed towards the content or whether the captured sounds are part of a conversation indicating disinterest in the content, or even different environments, such as mobility state or the vibration pattern of the mobile device (via which content is delivered) in a car or a train. Other embodiments are described in the subject disclosure.
  • One or more aspects of the subject disclosure are a method including determining, by a system including a processor, a content context of a first segment of content being presented at a display to an audience. The method can include determining, by the system, an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected volume and an expected audio reaction context. The method can include obtaining, by the system, sensor data captured from a sensor device in proximity to the audience, where the sensor data is indicative of a sensed audience reaction to the first segment of the content, and where the sensed audience reaction comprises a sensed volume and a sensed audio reaction context. The method can include comparing, by the system, the sensed volume and the sensed audio reaction context with the expected volume and the expected audio reaction context to determine a level of interest in the first segment. The method can include adjusting, by the system, a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display.
  • One or more aspects of the subject disclosure include a machine-readable storage medium, comprising executable instructions that, when executed by a media processor, facilitate performance of operations, including determining a content context of a first segment of content. The media processor can present the first segment of the content at a display to an audience; and can determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context. The media processor can obtain sensor data captured from a sensor device in proximity to the display, where the sensor data is indicative of a sensed audience reaction to the first segment of the content. The media processor can apply speech pattern recognition to the sensor data to determine a sensed audio reaction context of the sensed audience reaction. The media processor can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment. The media processor can increase a resolution of a second segment of the content according to a determination that the level of interest has increased, where the increasing of the resolution of the second segment generates an adjusted second segment. The media processor can present the adjusted second segment at the display. The media processor can provide a notice to a second media processor that causes the second media processor to decrease another resolution of other content being presented by the second media processor.
  • One or more aspects of the subject disclosure include a network server having a processor; and a memory that stores executable instructions that, when executed by the processor, facilitate performance of operations, including determining a content context of a first segment of content being presented by a media processor at a display to an audience. The network server can determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context. The network server can receive sensor data captured from a sensor device in proximity to the audience, where the sensor data is indicative of a sensed audience reaction to the first segment of the content. The network server can apply speech recognition to the sensor data to identify words spoken by the audience. The network server can determine a sensed audio reaction context according to the words. The network server can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment. The network server can adjust a second segment of the content according to the level of interest to generate an adjusted second segment displayable at the display.
  • FIG. 1 depicts an illustrative embodiment of a system 100 in which content is presented by a media processor 106 (e.g., a set top box) at a display 108 (e.g., a television). An audience 110 can be viewing or otherwise consuming the content. Three users are illustrated for audience 110, although the audience could be a single user or any other number of users. Also, the media processor 106, the display 108 and the audience 110 are illustrated at a single location (e.g., a single premises), however, the media processor, the display and the audience can be groups of media processors, displays and audiences that are located at different locations, such as different premises. The content can be provided to the media processor 106 by a network server 130.
  • In one embodiment, one or more sensor devices 150 (only one of which is shown) can capture sensor data of a sensed audience reaction for the audience 110. The sensor data can be of various types including audio and/or image information. In one embodiment, the sensor device 150 can be integrated with the media processor 106 or can be otherwise in communication with the media processor to provide the sensor data to the media processor. In another embodiment, the sensor device 150 can be a separate device from the media processor 106, where the sensor device provides the sensor data to the network server 130 with or without sharing the sensor data with the media processor.
  • Combinations of sensor devices 150 can also be utilized, such as a first sensor device that is an audio recorder of the media processor 106 and a second sensor device that is a camera for capturing images of the audience 110. For instance, the camera can be integrated with the media processor 106 or can be otherwise in communication with the media processor to provide the sensor data to the media processor, or the camera can be a separate device from the media processor, where the camera provides the sensor data to the network server 130 with or without sharing the sensor data with the media processor.
  • In another embodiment, the sensor device 150 can be part of an end user device of one or more of the users in the audience 110. For example, a mobile phone of a user of the audience 110 can record audio and provide that audio to the media processor 106 and/or to the network server 130. In addition to audio and image recordings, the sensor device 150 can capture other sensor data, such as motion detection, lighting and so forth.
  • System 100 enables determining a level of interest or a change in the level of interest for a particular segment of the content being presented so that a content adaptation can be performed. For example, a determination can be made for a content context of a first segment of the content being presented at the display 108 for the audience 110. The content context can be performed based on various factors and utilizing various techniques. In one embodiment, segments of the content can include or otherwise be associated with metadata that can be accessed where the metadata describes the particular content context of the particular segment. For example, a movie can include metadata that describes various scenes in the movie. The content context can be categorized in such a way as to facilitate determining an expected audience reaction to the scene. For instance, the metadata can describe that the scene is funny from which it can be determined that audience laughter is expected. As another example, the metadata can describe that the scene is scary from which it can be determined that an audience scream and/or an audience movement is expected. As another example, the metadata can describe that the scene is somber from which it can be determined that audience quiet is expected. The metadata can also be utilized for determining that which is not an expected reaction of the audience, such as it is not expected for there to be laughter in a scene that is described by the metadata as tragic.
  • In another embodiment, pattern recognition can be applied to the segment of the content to determine the content context and thus the expected audience reaction. For example, audio and/or image pattern recognition can be applied to the segment to determine that a movie scene is scary or tragic. Detected words and/or sounds (e.g., laughter, scream, sighing, crying), detected actions, or other characteristics of the scene detected from the pattern recognition can be utilized in determining the particular content context. The pattern recognition can also be performed on live events that are being presented as the content. For instance, pattern recognition can be utilized to determine that a football team is close to the end zone from which it can be determined that the expected audience reaction can be excited (for an audience member favoring the team on offense) or discouraged (for an audience member favoring the team on defense). In one embodiment, an expected volume of audience reaction can be determined, such as determining that an audio reaction to a touchdown in a football game should be above a threshold level if the audience member is interested in the particular game. The determination of the content context can be performed by various devices, such as by the media processor 106 and/or the network server 130.
  • System 100 enables comparing an expected audience reaction with a sensed audience reaction to determine a level of interest or a change in the level of interest for a particular segment of the content being presented so that a content adaptation can be performed. The comparison can be performed by various devices, such as by the media processor 106 and/or the network server 130. As explained above, the expected and sensed audience reactions can be characterized in various ways, including audio volume, spoken words, sounds, movement and so forth.
  • Once a level of interest or a change in the level of interest for a particular segment of the content being presented is determined, then a content adaptation can be performed. The content adaptation can be of various types. For example, characteristics of the content can be adjusted, such as a resolution or quality, in response to a determination of an increased level of interest. For instance, a determination can be made that the audience has become more interested in a football game because the offensive team is near the end zone and may score a touchdown. As described above, this determination can be performed according to a comparison of the expected audience reaction and the sensed audience reaction. The content (i.e., the football game) can be adjusted to a high resolution so that the upcoming plays near the end zone are of a higher quality. In one embodiment, the system 100 can continue to monitor for content context of subsequent segments, expected audience reactions for those subsequent segments, and sensed audience reactions for those subsequent segments, from which changes in level of interest can be determined. In this example, if it is determined that the level of interest decreases then the content can be adjusted again, such as lowering the resolution (for plays that are deemed to be of less interest to the audience).
  • In one embodiment, the adjustment of the resolution can be performed in conjunction with adjustment to other content being provided to the location. For instance, an increase in resolution at the media processor 106 resulting in an increase in bandwidth can be offset by a decrease in resolution at another media processor so that the bandwidth usage for the location are maintained relatively constant. In one embodiment, the location can have multiple audiences and multiple media processors where the monitoring for content context, expected audience reactions, and sensed audience reactions are performed for those multiple audiences, from which changes in level of interest can be determined for those multiple audiences. In this example, the changes in level of interest for each of the multiple audiences (at different displays watching different content) can be compared and the content adjustment can be made according to a higher level of change in interest. For instance, a first audience determined to have a higher level of interest can be presented with a high definition version of first content while a second audience determined to have a lower level of interest can be presented with a standard definition version of second content.
  • In one embodiment, the adjustment of the content can include inserting a different segment into the content, such as a targeted advertisement or a different version of a scene. In another embodiment, the adjustment of the content can include replacing the subsequent segments with completely different content, such as switching to a different episode of a show or a different movie.
  • In one embodiment, the monitoring of the audience reaction can be performed on an individual basis. For example, a captured audio reaction can be identified with a particular user of the audience and compared with an expected audio reaction for that user. User profiles and/or historical reactions for that particular user can be analyzed in conjunction with the content context to determine the expected audience reaction for that particular user. As an example, it can be determined that the content context for a scene in a movie is humor. It can be further determined that a particular user in the past has reacted with laughter at a particular volume. The sensed laughter for that particular user can be compared to the particular volume of past laughter to determine a level of interest in the scene for the particular user.
  • In one embodiment, a notification can be presented at the display 108 indicating that the content adjustment is going to occur (e.g., resolution of a subsequent segment of the content will be increased). In another embodiment, a characteristic of the audience can be determined (e.g., by the media processor 106 and/or the network server 130) based on a monitored consumption history associated with the audience. In this example, the adjusting of a subsequent segment of the content can include selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement for presentation at the display 108, where the selecting of the targeted advertisement is according to the determined level of interest and according to the characteristic of the audience.
  • In one or more embodiments, system 100 can teach a video broadcast system how to insert and/or adapt content according to an audience signal-to-noise ratio. The audience signal-to-noise ratio can be a measure of expected audio to actual audio and can filter out background noise. System 100 can capture the mood, emotion, need and/or requirement of the audience at that particular moment. As a result, the content inserted and/or adapted to the video broadcast system can instantaneously capture the need and/or requirement of the audience at that moment. System 100 can enable content insertion to be adapted and personalized according to the audience signal-to-noise which is distinct from blindly or randomly inserting content.
  • FIG. 2 depicts an illustrative embodiment of system 200 that enables content adaptation according to monitored audience reactions to segments of the content. By determining an expected audience reaction according to the context of the particular segment of content being presented, system 200 can detect a level of interest or a change in the level of interest via a comparison to a sensed audience reaction.
  • In one embodiment, a device 205 (e.g., a media processor presenting the content, a network server, or some other device) can determine a content context of a first segment of content being presented by the media processor to the audience 110. The content context determination can be performed utilizing various techniques including analyzing metadata that describes the segment content and/or performing pattern recognition (e.g., audio and/or image) to determine the segment context. An expected audience reaction can be determined according to the content context of the first segment, such as determining an expected audio reaction context (e.g., laughter, screaming, enthusiasm, anger, and so forth). The device 205 can obtain sensor data captured from one or more sensor devices in proximity to the audience, such as an audio recorder, a camera, a motion detector and so forth. In one embodiment, the sensor data can be indicative of a sensed audience reaction to the first segment of the content. In another embodiment, the sensor data can be filtered, such as removing background noise, to isolate the sensed audience reaction.
  • In one embodiment, speech recognition can be applied to the sensor data to identify words spoken by the audience. Based on the speech recognition, a sensed audio reaction context can be determined according to the words. For example, words indicative of anger or words indicative of sorrow can be identified. Device 205 can compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment. For instance, device 205 can determine that a sad scene in a movie should invoke a sorrowful audience reaction. If the comparison of the expected and sensed audience reactions is not a match then device 205 can determine that there is a low level of interest in the scene. In one embodiment, the device 205 can detect a sensed audience reaction which is contrary to what is expected, such as sensing laughter during a sad scene. Based on this detection, device 205 can determine there is little interest in the scene.
  • Based on a detection of a particular level of interest (e.g., satisfying a threshold), device 205 can adjust a second segment of the content according to the level of interest to generate an adjusted second segment for presentation at the display. In one embodiment, the adjusting of the second segment of the content can include adjusting a resolution of the second segment according to the level of interest. In another embodiment, the adjusting of the resolution of the second segment can be performed in conjunction with decreasing of another resolution of second content being presented by another media processor, where the media processor and the other media processor are located at a same premises.
  • In one embodiment, the display can be a group of displays located at different premises, where the audience 110 is a group of users located at the different premises. In this example, the sensor device can be a group of sensor devices that capture sensor data for the multiple audiences at the different premises. The content adjustment can be performed for content being presented at the different premises.
  • In one embodiment, the device 205 can determine a characteristic of the audience based on a monitored consumption history associated with the audience. In this example, the adjusting of the second segment of the content can include selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement to the media processor for presentation at the display. The selecting of the targeted advertisement can be according to the level of interest and according to the characteristic of the audience.
  • In one embodiment, the determining of the content context of the first segment can include applying image pattern recognition and speech pattern recognition to the first segment. In another embodiment, device 205 can determine an expected volume of the expected audience reaction according to the content context of the first segment; determine a sensed volume of the sensed audience reaction according to the sensor data; and compare the sensed volume with the expected volume, where the level of interest in the first segment is determined based in part on the comparing of the sensed volume with the expected volume.
  • FIG. 3 depicts an illustrative embodiment of a method 300 used by systems 100 and 200 for adjusting content according to a level of audience interest determined from a comparison of an expected audience reaction to an actual audience reaction.
  • Method 300 can commence at 302 where a content context is determined by a system for a first segment of content being presented at a display for an audience. The system can be various devices or combinations of devices, including a media processor presenting the content or a network server. At 304, an expected audience reaction can be determined by the system according to the content context of the first segment. In one embodiment, the expected audience reaction can include an expected volume and/or an expected audio reaction context.
  • At 306, the system can obtain sensor data that was captured from a sensor device in proximity to the audience. The sensor data can be indicative of a sensed audience reaction to the first segment of the content. In one embodiment, the sensed audience reaction can include a sensed volume and/or a sensed audio reaction context.
  • At 308 the system can compare the sensed audience reaction (e.g., the sensed volume and/or the sensed audio reaction context) with the expected audience reaction (e.g., the expected volume and/or the expected audio reaction context) to determine change in level of interest for the first segment. If there is no change, then method 300 can return to 302 to continue monitoring for changes in interest to subsequent segments. If on the other hand there is a change in level of interest then at 310 the system can adjust a second segment of the content according to the changed level of interest to generate an adjusted second segment for presentation at the display. For example, if it is determined that the level of interest has increased then the second segment can be adjusted according to that increased level of interest.
  • In one embodiment, the sensor device can be an audio recorder of a media processor presenting the content at the display, and method 300 can include applying speech recognition to the sensor data to identify words spoken by the audience; and determining the sensed audio reaction context according to the words. In one embodiment, the adjusting of the second segment of the content can include adjusting a resolution of the second segment according to the level of interest.
  • In one embodiment, the adjusting of the second segment of the content can include increasing a resolution of the second segment being presented by a first media processor at the display, where the adjusting of the resolution of the second segment is performed in conjunction with decreasing of another resolution of second content being presented by a second media processor, and where the first and second media processors are located at a same premises. In one embodiment, method 300 can include transmitting, by the system over a network, the content to a media processor presenting the content at the display, where the obtaining of the sensor data by the system comprises receiving the sensor data captured from the sensor device housed in the media processor.
  • In one embodiment, the display can be a group of displays located at different premises, where the audience is a group of users located at the different premises, and where the sensor device is a group of sensor devices that are each housed by a corresponding media processor presenting the content at one of the group of displays. In one embodiment, the method 300 can include determining a characteristic of the audience based on a monitored consumption history associated with the audience, where the adjusting of the second segment of the content comprises selecting a targeted advertisement from among a group of advertisements and providing the targeted advertisement for presentation at the display, and where the selecting of the targeted advertisement is according to the level of interest and according to the characteristic of the audience.
  • In one embodiment, the determining of the content context of the first segment can be based on metadata associated with the content. In one embodiment, the determining of the content context of the first segment can include applying image pattern recognition and speech pattern recognition to the first segment.
  • While for purposes of simplicity of explanation, the respective processes are shown and described as a series of blocks in FIG. 3, it is to be understood and appreciated that the claimed subject matter is not limited by the order of the blocks, as some blocks may occur in different orders and/or concurrently with other blocks from what is depicted and described herein. Moreover, not all illustrated blocks may be required to implement the methods described herein.
  • FIG. 4 depicts an illustrative embodiment of a first communication system 400 for delivering media content. The communication system 400 can represent an interactive television network, such as an Internet Protocol Television (IPTV) media system. Communication system 400 can be overlaid or operably coupled with systems 100 and 200 of FIGS. 1 and 2 as another representative embodiment of communication system 400. For instance, one or more devices illustrated in the communication system 400 of FIG. 4 can adjust content based on the reaction of the audience. The audience reaction can be compared to an expected audience reaction that is determined from a context of the particular segment of content that the audience is viewing. The audience reaction can be utilized to determine a level of interest or a change in the level of interest which can trigger an adjustment of the content. Various criteria can be utilized for comparing the actual and expected audience reaction, such as monitoring sound volume, determining a context of the audience reaction, such as whether there is excitement directed towards the content or whether the captured sounds are part of a conversation indicating disinterest in the content.
  • The IPTV media system can include a super head-end office (SHO) 410 with at least one super headend office server (SHS) 411 which receives media content from satellite and/or terrestrial communication systems. In the present context, media content can represent, for example, audio content, moving image content such as 2D or 3D videos, video games, virtual reality content, still image content, and combinations thereof. The SHS server 411 can forward packets associated with the media content to one or more video head-end servers (VHS) 414 via a network of video head-end offices (VHO) 412 according to a multicast communication protocol.
  • The VHS 414 can distribute multimedia broadcast content via an access network 418 to commercial and/or residential buildings 402 housing a gateway 404 (such as a residential or commercial gateway). The access network 418 can represent a group of digital subscriber line access multiplexers (DSLAMs) located in a central office or a service area interface that provide broadband services over fiber optical links or copper twisted pairs 419 to buildings 402. The gateway 404 can use communication technology to distribute broadcast signals to media processors 406 such as Set-Top Boxes (STBs) which in turn present broadcast channels to media devices 408 such as computers or television sets managed in some instances by a media controller 407 (such as an infrared or RF remote controller).
  • The gateway 404, the media processors 406, and media devices 408 can utilize tethered communication technologies (such as coaxial, powerline or phone line wiring) or can operate over a wireless access protocol such as Wireless Fidelity (WiFi), Bluetooth®, Zigbee or other present or next generation local or personal area wireless network technologies. By way of these interfaces, unicast communications can also be invoked between the media processors 406 and subsystems of the IPTV media system for services such as video-on-demand (VoD), browsing an electronic programming guide (EPG), or other infrastructure services.
  • A satellite broadcast television system 429 can be used in the media system of FIG. 4. The satellite broadcast television system can be overlaid, operably coupled with, or replace the IPTV system as another representative embodiment of communication system 400. In this embodiment, signals transmitted by a satellite 415 that include media content can be received by a satellite dish receiver 431 coupled to the building 402. Modulated signals received by the satellite dish receiver 431 can be transferred to the media processors 406 for demodulating, decoding, encoding, and/or distributing broadcast channels to the media devices 408. The media processors 406 can be equipped with a broadband port to an Internet Service Provider (ISP) network 432 to enable interactive services such as VoD and EPG as described above.
  • In yet another embodiment, an analog or digital cable broadcast distribution system such as cable TV system 433 can be overlaid, operably coupled with, or replace the IPTV system and/or the satellite TV system as another representative embodiment of communication system 400. In this embodiment, the cable TV system 433 can also provide Internet, telephony, and interactive media services. System 400 enables various types of interactive television and/or services including IPTV, cable and/or satellite.
  • The subject disclosure can apply to other present or next generation over-the-air and/or landline media content services system.
  • Some of the network elements of the IPTV media system can be coupled to one or more computing devices 430, a portion of which can operate as a web server for providing web portal services over the ISP network 432 to wireline media devices 408 or wireless communication devices 416.
  • Communication system 400 can also provide for all or a portion of the computing devices 430 to function as a content adjustment server (herein referred to as server 430). The server 430 can use computing and communication technology to perform function 462, which can include among other things, determining a content context of a first segment of content being presented by a media processor at a display for an audience; determining an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context; receiving sensor data captured from a sensor device in proximity to the audience, the sensor data indicative of a sensed audience reaction to the first segment of the content; applying speech recognition to the sensor data to identify words spoken by the audience; determining a sensed audio reaction context according to the words; comparing the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment; and/or adjusting a second segment of the content according to the level of interest to generate an adjusted second segment for presentation at the display.
  • For instance, function 462 of server 430 can be similar to the functions described network server 130 of system 100. The media processors 406 and wireless communication devices 416 can be provisioned with software functions 464 and 466, respectively, to utilize the services of server 430. For instance, functions 46Y and 46Z of media processors 406 and wireless communication devices 416 can be similar to the functions described for the communication devices 106 of FIG. 1 in accordance with method 300.
  • Multiple forms of media services can be offered to media devices over landline technologies such as those described above. Additionally, media services can be offered to media devices by way of a wireless access base station 417 operating according to common wireless access protocols such as Global System for Mobile or GSM, Code Division Multiple Access or CDMA, Time Division Multiple Access or TDMA, Universal Mobile Telecommunications or UMTS, World interoperability for Microwave or WiMAX, Software Defined Radio or SDR, Long Term Evolution or LTE, and so on. Other present and next generation wide area wireless access network technologies can be used in one or more embodiments of the subject disclosure.
  • FIG. 5 depicts an illustrative embodiment of a communication device 500. Communication device 500 can serve in whole or in part as an illustrative embodiment of the devices depicted in systems 100, 200 and 400 and can be configured to perform portions of method 300. As an example, communication device 500 can determine a content context of a first segment of content; present the first segment of the content at a display for an audience; determine an expected audience reaction according to the content context of the first segment, where the expected audience reaction comprises an expected audio reaction context; obtain sensor data captured from a sensor device in proximity to the display, the sensor data indicative of a sensed audience reaction to the first segment of the content; apply speech pattern recognition to the sensor data to determine a sensed audio reaction context of the sensed audience reaction; compare the sensed audio reaction context with the expected audio reaction context to determine a level of interest in the first segment; increase a resolution of a second segment of the content according to a determination that the level of interest has increased, where the increasing of the resolution of the second segment generates an adjusted second segment; present the adjusted second segment at the display; and/or provide a notice to a second media processor that causes the second media processor to decrease another resolution of other content being presented by the second media processor.
  • In one embodiment, communication device 500 can determine an expected volume of the expected audience reaction according to the content context of the first segment; determine a sensed volume of the sensed audience reaction according to the sensor data; and compare the sensed volume with the expected volume, where the level of interest in the first segment is determined based in part on the comparing of the sensed volume with the expected volume. In one embodiment, communication device 500 can present a notification at the display indicating that the resolution of the second segment of the content will be increased prior to the presenting of the adjusted second segment at the display.
  • Communication device 500 can comprise a wireline and/or wireless transceiver 502 (herein transceiver 502), a user interface (UI) 504, a power supply 514, a location receiver 516, a motion sensor 518, an orientation sensor 520, and a controller 506 for managing operations thereof. The transceiver 502 can support short-range or long-range wireless access technologies such as Bluetooth®, ZigBee®, WiFi, DECT, or cellular communication technologies, just to mention a few (Bluetooth® and ZigBee® are trademarks registered by the Bluetooth® Special Interest Group and the ZigBee® Alliance, respectively). Cellular technologies can include, for example, CDMA-1X, UMTS/HSDPA, GSM/GPRS, TDMA/EDGE, EV/DO, WiMAX, SDR, LTE, as well as other next generation wireless communication technologies as they arise. The transceiver 502 can also be adapted to support circuit-switched wireline access technologies (such as PSTN), packet-switched wireline access technologies (such as TCP/IP, VoIP, etc.), and combinations thereof.
  • The UI 504 can include a depressible or touch-sensitive keypad 508 with a navigation mechanism such as a roller ball, a joystick, a mouse, or a navigation disk for manipulating operations of the communication device 500. The keypad 508 can be an integral part of a housing assembly of the communication device 500 or an independent device operably coupled thereto by a tethered wireline interface (such as a USB cable) or a wireless interface supporting for example Bluetooth®. The keypad 508 can represent a numeric keypad commonly used by phones, and/or a QWERTY keypad with alphanumeric keys. The UI 504 can further include a display 510 such as monochrome or color LCD (Liquid Crystal Display), OLED (Organic Light Emitting Diode) or other suitable display technology for conveying images to an end user of the communication device 500. In an embodiment where the display 510 is touch-sensitive, a portion or all of the keypad 508 can be presented by way of the display 510 with navigation features.
  • The display 510 can use touch screen technology to also serve as a user interface for detecting user input. As a touch screen display, the communication device 500 can be adapted to present a user interface with graphical user interface (GUI) elements that can be selected by a user with a touch of a finger. The touch screen display 510 can be equipped with capacitive, resistive or other forms of sensing technology to detect how much surface area of a user's finger has been placed on a portion of the touch screen display. This sensing information can be used to control the manipulation of the GUI elements or other functions of the user interface. The display 510 can be an integral part of the housing assembly of the communication device 500 or an independent device communicatively coupled thereto by a tethered wireline interface (such as a cable) or a wireless interface.
  • The UI 504 can also include an audio system 512 that utilizes audio technology for conveying low volume audio (such as audio heard in proximity of a human ear) and high volume audio (such as speakerphone for hands free operation). The audio system 512 can further include a microphone for receiving audible signals of an end user. The audio system 512 can also be used for voice recognition applications. The UI 504 can further include an image sensor 513 such as a charged coupled device (CCD) camera for capturing still or moving images.
  • The power supply 514 can utilize common power management technologies such as replaceable and rechargeable batteries, supply regulation technologies, and/or charging system technologies for supplying energy to the components of the communication device 500 to facilitate long-range or short-range portable applications. Alternatively, or in combination, the charging system can utilize external power sources such as DC power supplied over a physical interface such as a USB port or other suitable tethering technologies.
  • The location receiver 516 can utilize location technology such as a global positioning system (GPS) receiver capable of assisted GPS for identifying a location of the communication device 500 based on signals generated by a constellation of GPS satellites, which can be used for facilitating location services such as navigation. The motion sensor 518 can utilize motion sensing technology such as an accelerometer, a gyroscope, or other suitable motion sensing technology to detect motion of the communication device 500 in three-dimensional space. The orientation sensor 520 can utilize orientation sensing technology such as a magnetometer to detect the orientation of the communication device 500 (north, south, west, and east, as well as combined orientations in degrees, minutes, or other suitable orientation metrics).
  • The communication device 500 can use the transceiver 502 to also determine a proximity to a cellular, WiFi, Bluetooth®, or other wireless access points by sensing techniques such as utilizing a received signal strength indicator (RSSI) and/or signal time of arrival (TOA) or time of flight (TOF) measurements. The controller 506 can utilize computing technologies such as a microprocessor, a digital signal processor (DSP), programmable gate arrays, application specific integrated circuits, and/or a video processor with associated storage memory such as Flash, ROM, RAM, SRAM, DRAM or other storage technologies for executing computer instructions, controlling, and processing data supplied by the aforementioned components of the communication device 500.
  • Other components not shown in FIG. 5 can be used in one or more embodiments of the subject disclosure. For instance, the communication device 500 can include a reset button (not shown). The reset button can be used to reset the controller 506 of the communication device 500. In yet another embodiment, the communication device 500 can also include a factory default setting button positioned, for example, below a small hole in a housing assembly of the communication device 500 to force the communication device 500 to re-establish factory settings. In this embodiment, a user can use a protruding object such as a pen or paper clip tip to reach into the hole and depress the default setting button. The communication device 500 can also include a slot for adding or removing an identity module such as a Subscriber Identity Module (SIM) card. SIM cards can be used for identifying subscriber services, executing programs, storing subscriber data, and so forth.
  • The communication device 500 as described herein can operate with more or less of the circuit components shown in FIG. 5. These variant embodiments can be used in one or more embodiments of the subject disclosure.
  • The communication device 500 can be adapted to perform the functions of devices 106, 130, 205 of FIGS. 1 and 2, the media processor 406, the media devices 408, or the portable communication devices 416 of FIG. 4. It will be appreciated that the communication device 500 can also represent other devices that can operate in systems 100 and 200 of FIGS. 1 and 2], communication system 400 of FIG. 4 such as a gaming console and a media player. In addition, the controller 506 can be adapted in various embodiments to perform the functions 462-466.
  • Upon reviewing the aforementioned embodiments, it would be evident to an artisan with ordinary skill in the art that said embodiments can be modified, reduced, or enhanced without departing from the scope of the claims described below. For example, other user activity can be monitored to detect a level of interest of a user of the audience, such as detecting a text message, a voice call, or surfing the web on a mobile device of a user of the audience during the presentation of the content which may be indicative of a lack of interest.
  • Other actions can be triggered by the determination of the level of interest or a change in the level of interest, such as proposing different content to the audience or to a subset of the audience.
  • In one embodiment, the determination of the level of interest can be multi-tiered based on different thresholds. For example, a scene in a movie can be determined to have a content context of frightening. The expected audience reaction can be determined to be a scream or movement of the audience. If the sensed audience reaction is both a scream and movement then a determination of very high interest can be made. If the sensed audience reaction is either the scream or the movement then a determination of high interest can be made. If the sensed audience reaction is neither the scream nor the movement, but rather silence, then a determination of medium interest can be made. If the sensed audience reaction is neither the scream nor the movement, but rather laughter, then a determination of low interest can be made. The content adjustments can then be made according to the group of different interest levels. For instance, the highest level of interest can result in an increase to the highest resolution and insertion of a targeted advertisement for another movie in the same genre while the lowest level of interest can result in a notification that a different movie of a different genre is currently available.
  • In one embodiment, sensed audience responses can be distinguished amongst users of the audience and a content adjustment determination can be made according to individual responses. For example, sensed audio can be analyzed via speech pattern recognition to determine that first and second users are engaged in a conversation that is not related to the content (e.g., a football game) being presented. The sensed audio can be further analyzed to determine that third and fourth users are cheering. The determination of which users are speaking can be performed in a number of different ways, such as based on accessible voice samples. In this example, a content adjustment can be made (e.g., a change in resolution of the presented football game or an insertion of a commercial for another football game that is scheduled in the near future) based on the interest of the third and fourth users.
  • In one embodiment, the content can be broadcast content which is then adjusted by a network server responsive to detecting the change in level of interest. Other embodiments can be used in the subject disclosure.
  • It should be understood that devices described in the exemplary embodiments can be in communication with each other via various wireless and/or wired methodologies. The methodologies can be links that are described as coupled, connected and so forth, which can include unidirectional and/or bidirectional communication over wireless paths and/or wired paths that utilize one or more of various protocols or methodologies, where the coupling and/or connection can be direct (e.g., no intervening processing device) and/or indirect (e.g., an intermediary processing device such as a router).
  • FIG. 6 depicts an exemplary diagrammatic representation of a machine in the form of a computer system 600 within which a set of instructions, when executed, may cause the machine to perform any one or more of the methods described above. One or more instances of the machine can operate, for example, as the media processor 106, the network server 130, the server 430, the media processor 406 and other devices of FIGS. 1-2 and 4 in order to adjust content according to a comparison of an expected audience reaction with a sensed audience reaction. In some embodiments, the machine may be connected (e.g., using a network 626) to other machines. In a networked deployment, the machine may operate in the capacity of a server or a client user machine in a server-client user network environment, or as a peer machine in a peer-to-peer (or distributed) network environment.
  • The machine may comprise a server computer, a client user computer, a personal computer (PC), a tablet, a smart phone, a laptop computer, a desktop computer, a control system, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine. It will be understood that a communication device of the subject disclosure includes broadly any electronic device that provides voice, video or data communication. Further, while a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methods discussed herein.
  • The computer system 600 may include a processor (or controller) 602 (e.g., a central processing unit (CPU)), a graphics processing unit (GPU, or both), a main memory 604 and a static memory 606, which communicate with each other via a bus 608. The computer system 600 may further include a display unit 610 (e.g., a liquid crystal display (LCD), a flat panel, or a solid state display). The computer system 600 may include an input device 612 (e.g., a keyboard), a cursor control device 614 (e.g., a mouse), a disk drive unit 616, a signal generation device 618 (e.g., a speaker or remote control) and a network interface device 620. In distributed environments, the embodiments described in the subject disclosure can be adapted to utilize multiple display units 610 controlled by two or more computer systems 600. In this configuration, presentations described by the subject disclosure may in part be shown in a first of the display units 610, while the remaining portion is presented in a second of the display units 610.
  • The disk drive unit 616 may include a tangible computer-readable storage medium 622 on which is stored one or more sets of instructions (e.g., software 624) embodying any one or more of the methods or functions described herein, including those methods illustrated above. The instructions 624 may also reside, completely or at least partially, within the main memory 604, the static memory 606, and/or within the processor 602 during execution thereof by the computer system 600. The main memory 604 and the processor 602 also may constitute tangible computer-readable storage media.
  • Dedicated hardware implementations including, but not limited to, application specific integrated circuits, programmable logic arrays and other hardware devices can likewise be constructed to implement the methods described herein. Application specific integrated circuits and programmable logic array can use downloadable instructions for executing state machines and/or circuit configurations to implement embodiments of the subject disclosure. Applications that may include the apparatus and systems of various embodiments broadly include a variety of electronic and computer systems. Some embodiments implement functions in two or more specific interconnected hardware modules or devices with related control and data signals communicated between and through the modules, or as portions of an application-specific integrated circuit. Thus, the example system is applicable to software, firmware, and hardware implementations.
  • In accordance with various embodiments of the subject disclosure, the operations or methods described herein are intended for operation as software programs or instructions running on or executed by a computer processor or other computing device, and which may include other forms of instructions manifested as a state machine implemented with logic components in an application specific integrated circuit or field programmable gate array. Furthermore, software implementations (e.g., software programs, instructions, etc.) including, but not limited to, distributed processing or component/object distributed processing, parallel processing, or virtual machine processing can also be constructed to implement the methods described herein. It is further noted that a computing device such as a processor, a controller, a state machine or other suitable device for executing instructions to perform operations or methods may perform such operations directly or indirectly by way of one or more intermediate devices directed by the computing device.
  • While the tangible computer-readable storage medium 622 is shown in an example embodiment to be a single medium, the term “tangible computer-readable storage medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “tangible computer-readable storage medium” shall also be taken to include any non-transitory medium that is capable of storing or encoding a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methods of the subject disclosure. The term “non-transitory” as in a non-transitory computer-readable storage includes without limitation memories, drives, devices and anything tangible but not a signal per se.
  • The term “tangible computer-readable storage medium” shall accordingly be taken to include, but not be limited to: solid-state memories such as a memory card or other package that houses one or more read-only (non-volatile) memories, random access memories, or other re-writable (volatile) memories, a magneto-optical or optical medium such as a disk or tape, or other tangible media which can be used to store information. Accordingly, the disclosure is considered to include any one or more of a tangible computer-readable storage medium, as listed herein and including art-recognized equivalents and successor media, in which the software implementations herein are stored.
  • Although the present specification describes components and functions implemented in the embodiments with reference to particular standards and protocols, the disclosure is not limited to such standards and protocols. Each of the standards for Internet and other packet switched network transmission (e.g., TCP/IP, UDP/IP, HTML, HTTP) represent examples of the state of the art. Such standards are from time-to-time superseded by faster or more efficient equivalents having essentially the same functions. Wireless standards for device detection (e.g., RFID), short-range communications (e.g., Bluetooth®, WiFi, Zigbee®), and long-range communications (e.g., WiMAX, GSM, CDMA, LTE) can be used by computer system 600.
  • The illustrations of embodiments described herein are intended to provide a general understanding of the structure of various embodiments, and they are not intended to serve as a complete description of all the elements and features of apparatus and systems that might make use of the structures described herein. Many other embodiments will be apparent to those of skill in the art upon reviewing the above description. The exemplary embodiments can include combinations of features and/or steps from multiple embodiments. Other embodiments may be utilized and derived therefrom, such that structural and logical substitutions and changes may be made without departing from the scope of this disclosure. Figures are also merely representational and may not be drawn to scale. Certain proportions thereof may be exaggerated, while others may be minimized. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
  • Although specific embodiments have been illustrated and described herein, it should be appreciated that any arrangement which achieves the same or similar purpose may be substituted for the embodiments described or shown by the subject disclosure. The subject disclosure is intended to cover any and all adaptations or variations of various embodiments. Combinations of the above embodiments, and other embodiments not specifically described herein, can be used in the subject disclosure. For instance, one or more features from one or more embodiments can be combined with one or more features of one or more other embodiments. In one or more embodiments, features that are positively recited can also be negatively recited and excluded from the embodiment with or without replacement by another structural and/or functional feature. The steps or functions described with respect to the embodiments of the subject disclosure can be performed in any order. The steps or functions described with respect to the embodiments of the subject disclosure can be performed alone or in combination with other steps or functions of the subject disclosure, as well as from other embodiments or from other steps that have not been described in the subject disclosure. Further, more than or less than all of the features described with respect to an embodiment can also be utilized.
  • Less than all of the steps or functions described with respect to the exemplary processes or methods can also be performed in one or more of the exemplary embodiments. Further, the use of numerical terms to describe a device, component, step or function, such as first, second, third, and so forth, is not intended to describe an order or function unless expressly stated so. The use of the terms first, second, third and so forth, is generally to distinguish between devices, components, steps or functions unless expressly stated otherwise. Additionally, one or more devices or components described with respect to the exemplary embodiments can facilitate one or more functions, where the facilitating (e.g., facilitating access or facilitating establishing a connection) can include less than every step needed to perform the function or can include all of the steps needed to perform the function.
  • In one or more embodiments, a processor (which can include a controller or circuit) has been described that performs various functions. It should be understood that the processor can be multiple processors, which can include distributed processors or parallel processors in a single machine or multiple machines. The processor can be used in supporting a virtual processing environment. The virtual processing environment may support one or more virtual machines representing computers, servers, or other computing devices. In such virtual machines, components such as microprocessors and storage devices may be virtualized or logically represented. The processor can include a state machine, application specific integrated circuit, and/or programmable gate array including a Field PGA. In one or more embodiments, when a processor executes instructions to perform “operations”, this can include the processor performing the operations directly and/or facilitating, directing, or cooperating with another device or component to perform the operations.
  • The Abstract of the Disclosure is provided with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, it can be seen that various features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separately claimed subject matter.

Claims (20)

What is claimed is:
1. A method, comprising:
transmitting, by a processing system including a processor, a first content to a first media processor and a second content to a second media processor;
determining, by the processing system, a content context of a first segment of the first content presented by the first media processor to an audience;
determining, by the processing system, an expected audience reaction according to the content context of the first segment;
obtaining, by the processing system, sensor data from a sensor device in proximity to an audience, the sensor data indicative of a sensed audience reaction to the first segment of the first content;
comparing, by the processing system, the sensed audience reaction to the expected audience reaction to determine a level of interest in the first segment;
determining that the level of interest is above a predetermined threshold;
responsive to the determining that the level of interest is above the predetermined threshold, increasing, by the processing system, a resolution of a second segment of the first content being presented by the first media processor, wherein the increasing the resolution of the second segment results in a first adjusted bandwidth; and
responsive to the determining that the level of interest is above the predetermined threshold, replacing, by the processing system, the second content being presented by the second media processor with third content presented by the second media processor, wherein the replacing comprises decreasing a resolution of the third content relative to an original resolution of the second content, wherein the decreasing the resolution of the third content results in a second adjusted bandwidth, where a change to the first adjusted bandwidth is offset by a change to the second adjusted bandwidth.
2. The method of claim 1, wherein the sensor device comprises an audio recorder of the first media processor presenting the first content, wherein the method further comprises:
applying, by the processing system, speech recognition to the sensor data to identify words spoken by the audience; and
determining, by the processing system, the sensed audience reaction according to the words spoken by the audience.
3. The method of claim 1, wherein obtaining sensor data from the sensor device in proximity to the audience comprises, obtaining, by the processing system, a sensed volume of the sensed audience reaction to the first segment of the first content, and wherein determining the sensed audience reaction comprises determining, by the processing system, the sensed audience reaction according to the sensed volume.
4. The method of claim 3, wherein comparing the sensed audience reaction to the expected audience reaction to determine a level of interest in the first segment comprises comparing, by the processing system, the sensed volume with an expected volume of the expected audience reaction.
5. The method of claim 1, wherein the determining the content context of the first segment of the first content presented by the first media processor to the audience comprises applying, by the processing system, image pattern recognition and speech pattern recognition to the first segment.
6. The method of claim 5, wherein the determining the content context of the first segment of the first content presented by the first media processor to the audience comprises applying, by the processing system, image pattern recognition and speech pattern recognition to a live event that is being presented as the first content.
7. The method of claim 1, wherein the determining the content context of the first segment of the first content presented by the first media processor to the audience comprises determining, by the processing system, the content context of the first segment based on metadata associated with the first content.
8. The method of claim 7, further comprising:
accessing, by the processing system, metadata that describes the content context of the first segment of the first content; and
determining, by the processing system, the expected audience reaction according to the metadata.
9. A network server, comprising:
a processing system including a processor; and
a memory that stores executable instructions that, when executed by the processing system, facilitate performance of operations, comprising:
transmitting a first content at a first bandwidth to a first media processor and a second content at a second bandwidth to a second media processor, wherein the first bandwidth and the second bandwidth form a first total bandwidth;
determining a content context of a first segment of the first content being presented by the first media processor to an audience;
determining an expected audience reaction context according to the content context of the first segment;
receiving sensor data captured from a sensor device in proximity to the audience;
determining a sensed audience reaction context to the first segment of the first content based on the sensor data;
comparing the sensed audience reaction context with the expected audience reaction context to determine a level of interest in the first segment;
determining that the level of interest is above a threshold; and
adjusting a second segment of the first content according to a determination that the level of interest is above the threshold to generate an adjusted second segment of the first content,
wherein adjusting the second segment of the first content comprises increasing a resolution of the second segment being presented by the first media processor in conjunction with a replacement of the second content being presented by the second media processor with third content at a third resolution that is decreased relative to an original resolution of the second content,
wherein the increasing of the resolution of the second segment of the first content results in an adjusted first bandwidth, and the decreasing of the other resolution of the second content results in an adjusted second bandwidth,
wherein a second total bandwidth comprises the first adjusted bandwidth and the second adjusted bandwidth, and
wherein the increasing of the resolution of the second segment of the first content resulting in the adjusted first bandwidth is offset by the decreasing of the other resolution of the second content resulting in the adjusted second bandwidth to maintain the first total bandwidth and the second total bandwidth equal.
10. The network server of claim 9, wherein the sensor device comprises an audio recorder of the first media processor presenting the first content.
11. The network server of claim 10, wherein the operations further comprise:
applying speech recognition to the sensor data to identify words spoken by the audience; and
determining a sensed audio reaction context according to the words.
12. The network server of claim 11, wherein the operations further comprise:
determining an expected audio reaction according to the content context of the first segment; and
comparing the sensed audio reaction context with an expected audio reaction context to determine the level of interest in the first segment.
13. The network server of claim 10, wherein the operations further comprise:
obtaining sensor data from the sensor device in proximity to the audience;
obtaining a sensed volume of the sensed audience reaction to the first segment of the first content;
comparing the sensed volume with an expected and an expected audience volume; and
determining the level of interest in the first segment according to the comparing.
14. The network server of claim 9, wherein the operations further comprise:
applying one of image pattern recognition and speech pattern recognition to the first segment to determine the content context of the first segment of the first content.
15. The network server of claim 9, wherein the operations further comprise:
identifying a first genre of the first segment of the first content;
determining a target advertisement according to the level of interest and the first genre of the first segment, wherein the target advertisement is associated with the first genre; and
wherein the adjusting of the second segment of the first content comprises presenting the target advertisement.
16. A non-transitory machine-readable storage medium, comprising executable instructions that, when executed by a first media processor, facilitate performance of operations, comprising:
receiving a first content at a first bandwidth, wherein a second media processor receives a second content at a second bandwidth, wherein the first bandwidth and the second bandwidth form a first total bandwidth;
determining a content context of a first segment of the first content;
presenting the first segment of the first content at a first display to an audience;
determining an expected audience reaction context according to the content context of the first segment;
obtaining sensor data captured from a sensor device in proximity to the first display, the sensor data indicative of a sensed audience reaction context to the first segment of the first content;
comparing the sensed audience reaction context with the expected audience reaction context to determine a level of interest in the first segment;
determining that the level of interest is above a threshold;
identifying a first genre of the first segment of the first segment;
increasing a second resolution of a second segment of the first content being presented by the first media processor at the first display in conjunction with a replacement of the second content being presented by a second media processor at a second display with third content at another resolution that is decreased relative to an original resolution of the second content according to a determination that the level of interest is above the threshold, wherein the increasing of the second resolution of the second segment generates an adjusted second segment, wherein a third genre of the third content is different from a second genre of the second content;
presenting the adjusted second segment at the first display; and
wherein the increasing of the resolution of the second segment of the first content results in an adjusted first bandwidth, and decreasing of the other resolution of the second content results in an adjusted second bandwidth,
wherein the first adjusted bandwidth and the second adjusted bandwidth form a second total bandwidth, and
wherein the increasing of the resolution of the second segment of the first content resulting in the adjusted first bandwidth is offset by the decreasing of the other resolution of the second content resulting in the adjusted second bandwidth to maintain the first total bandwidth and the second total bandwidth equal.
17. The non-transitory machine-readable storage medium of claim 16 wherein the operations further comprise:
identifying a first genre of the first segment of the first segment;
determining a target advertisement according to the level of interest and the first genre of the first segment, wherein the target advertisement is associated with the first genre; and
wherein the adjusting of the second segment of the first content comprises presenting the target advertisement.
18. The non-transitory machine-readable storage medium of claim 16 wherein the operations further comprise:
providing a notice to the second media processor that causes the second media processor to decrease another resolution of the second content being presented by the second media processor on a second display.
19. The non-transitory machine-readable storage medium of claim 16 wherein the operations further comprise:
applying speech recognition to the sensor data to identify words spoken by the audience;
determining a sensed audio reaction context according to the words; and
comparing the sensed audio reaction context with an expected audio reaction context to determine the level of interest in the first segment.
20. The non-transitory machine-readable storage medium of claim 16 wherein the operations further comprise:
applying image pattern recognition and speech pattern recognition to a live event that is being presented as the first content to determine the content context of the first segment of the first content.
US16/733,399 2015-11-11 2020-01-03 Method and apparatus for content adaptation based on audience monitoring Abandoned US20200177956A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/733,399 US20200177956A1 (en) 2015-11-11 2020-01-03 Method and apparatus for content adaptation based on audience monitoring

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US14/938,421 US10542315B2 (en) 2015-11-11 2015-11-11 Method and apparatus for content adaptation based on audience monitoring
US16/733,399 US20200177956A1 (en) 2015-11-11 2020-01-03 Method and apparatus for content adaptation based on audience monitoring

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/938,421 Continuation US10542315B2 (en) 2015-11-11 2015-11-11 Method and apparatus for content adaptation based on audience monitoring

Publications (1)

Publication Number Publication Date
US20200177956A1 true US20200177956A1 (en) 2020-06-04

Family

ID=58664388

Family Applications (2)

Application Number Title Priority Date Filing Date
US14/938,421 Expired - Fee Related US10542315B2 (en) 2015-11-11 2015-11-11 Method and apparatus for content adaptation based on audience monitoring
US16/733,399 Abandoned US20200177956A1 (en) 2015-11-11 2020-01-03 Method and apparatus for content adaptation based on audience monitoring

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US14/938,421 Expired - Fee Related US10542315B2 (en) 2015-11-11 2015-11-11 Method and apparatus for content adaptation based on audience monitoring

Country Status (1)

Country Link
US (2) US10542315B2 (en)

Families Citing this family (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10437288B2 (en) 2014-10-06 2019-10-08 Fasetto, Inc. Portable storage device with modular power and housing system
US9467718B1 (en) 2015-05-06 2016-10-11 Echostar Broadcasting Corporation Apparatus, systems and methods for a content commentary community
EP4080794A1 (en) 2016-01-06 2022-10-26 TVision Insights, Inc. Systems and methods for assessing viewer engagement
US11540009B2 (en) 2016-01-06 2022-12-27 Tvision Insights, Inc. Systems and methods for assessing viewer engagement
WO2017119604A1 (en) * 2016-01-08 2017-07-13 주식회사 아이플래테아 Audience rating calculation server, audience rating calculation method, and audience rating calculation remote device
US10268689B2 (en) 2016-01-28 2019-04-23 DISH Technologies L.L.C. Providing media content based on user state detection
US11146865B2 (en) 2016-03-03 2021-10-12 Comcast Cable Communications, Llc Determining points of interest in a content item
US11012719B2 (en) * 2016-03-08 2021-05-18 DISH Technologies L.L.C. Apparatus, systems and methods for control of sporting event presentation based on viewer engagement
US10984036B2 (en) 2016-05-03 2021-04-20 DISH Technologies L.L.C. Providing media content based on media element preferences
US10057746B1 (en) 2016-11-16 2018-08-21 Wideorbit, Inc. Method and system for detecting a user device in an environment associated with a content presentation system presenting content
US10764381B2 (en) 2016-12-23 2020-09-01 Echostar Technologies L.L.C. Communications channels in media systems
US11196826B2 (en) 2016-12-23 2021-12-07 DISH Technologies L.L.C. Communications channels in media systems
US10390084B2 (en) 2016-12-23 2019-08-20 DISH Technologies L.L.C. Communications channels in media systems
US11708051B2 (en) 2017-02-03 2023-07-25 Fasetto, Inc. Systems and methods for data storage in keyed devices
JP6565084B2 (en) * 2017-03-29 2019-08-28 本田技研工業株式会社 Object authentication apparatus and object authentication method
US11770574B2 (en) * 2017-04-20 2023-09-26 Tvision Insights, Inc. Methods and apparatus for multi-television measurements
US10511888B2 (en) 2017-09-19 2019-12-17 Sony Corporation Calibration system for audience response capture and analysis of media content
US10763630B2 (en) 2017-10-19 2020-09-01 Fasetto, Inc. Portable electronic device connection systems
US11869039B1 (en) 2017-11-13 2024-01-09 Wideorbit Llc Detecting gestures associated with content displayed in a physical environment
WO2019109033A1 (en) 2017-12-01 2019-06-06 Fasetto, Inc. Systems and methods for improved data encryption
US20190205469A1 (en) * 2018-01-04 2019-07-04 International Business Machines Corporation Cognitive system and method to select best suited audio content based on individual's past reactions
US11043230B1 (en) * 2018-01-25 2021-06-22 Wideorbit Inc. Targeted content based on user reactions
US10462422B1 (en) * 2018-04-09 2019-10-29 Facebook, Inc. Audio selection based on user engagement
CN112292708B (en) * 2018-04-17 2022-06-17 法斯埃托股份有限公司 Presentation system and method with real-time feedback
US10616649B2 (en) * 2018-07-19 2020-04-07 Rovi Guides, Inc. Providing recommendations based on passive microphone detections
US10897647B1 (en) * 2018-07-25 2021-01-19 Imdb.Com, Inc. Ascertaining audience reactions for a media item
US11412295B2 (en) * 2018-10-02 2022-08-09 Comcast Cable Communications, Llc Systems and methods for determining usage information
US11817005B2 (en) * 2018-10-31 2023-11-14 International Business Machines Corporation Internet of things public speaking coach
US11037550B2 (en) 2018-11-30 2021-06-15 Dish Network L.L.C. Audio-based link generation
US20200186862A1 (en) 2018-12-05 2020-06-11 Roku, Inc. Downloadable avod content
US11240555B2 (en) 2018-12-05 2022-02-01 Roku, Inc. Providing over-the-air content to any device
US10834452B2 (en) * 2019-01-02 2020-11-10 International Business Machines Corporation Dynamic live feed recommendation on the basis of user real time reaction to a live feed
US11089356B2 (en) * 2019-03-26 2021-08-10 Rovi Guides, Inc. Systems and methods for media content hand-off based on type of buffered data
JP2021135803A (en) * 2020-02-27 2021-09-13 富士フイルムビジネスイノベーション株式会社 Document preparation device and program
US11128925B1 (en) * 2020-02-28 2021-09-21 Nxp Usa, Inc. Media presentation system using audience and audio feedback for playback level control
US11310553B2 (en) 2020-06-19 2022-04-19 Apple Inc. Changing resource utilization associated with a media object based on an engagement score
US20220174357A1 (en) * 2020-11-30 2022-06-02 At&T Intellectual Property I, L.P. Simulating audience feedback in remote broadcast events
US20220377413A1 (en) * 2021-05-21 2022-11-24 Rovi Guides, Inc. Methods and systems for personalized content based on captured gestures
US20230019723A1 (en) * 2021-07-14 2023-01-19 Rovi Guides, Inc. Interactive supplemental content system
US11849179B2 (en) * 2021-12-21 2023-12-19 Disney Enterprises, Inc. Characterizing audience engagement based on emotional alignment with characters
US20230217086A1 (en) * 2021-12-30 2023-07-06 Interwise Ltd. Providing and Using a Branching Narrative Content Service
US11910061B2 (en) * 2022-05-23 2024-02-20 Rovi Guides, Inc. Leveraging emotional transitions in media to modulate emotional impact of secondary content
US11871081B2 (en) * 2022-05-23 2024-01-09 Rovi Guides, Inc. Leveraging emotional transitions in media to modulate emotional impact of secondary content
US20240022791A1 (en) * 2022-07-15 2024-01-18 Rovi Guides, Inc. Systems and methods to adapt a schedule to be played by a media player
US11632413B1 (en) * 2022-07-18 2023-04-18 Rovi Guides, Inc. Methods and systems for streaming media content
US12063260B2 (en) 2022-08-31 2024-08-13 Rovi Guides, Inc. Intelligent delivery and playout to prevent stalling in video streaming
US11936948B1 (en) * 2023-01-24 2024-03-19 Roku, Inc. Method and system for generating a visual composition of user reactions in a shared content viewing session

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120324493A1 (en) * 2011-06-17 2012-12-20 Microsoft Corporation Interest-based video streams
US20130145385A1 (en) * 2011-12-02 2013-06-06 Microsoft Corporation Context-based ratings and recommendations for media
US20140168277A1 (en) * 2011-05-10 2014-06-19 Cisco Technology Inc. Adaptive Presentation of Content
US8898687B2 (en) * 2012-04-04 2014-11-25 Microsoft Corporation Controlling a media program based on a media reaction
US20140366049A1 (en) * 2013-06-11 2014-12-11 Nokia Corporation Method, apparatus and computer program product for gathering and presenting emotional response to an event
US20150189376A1 (en) * 2013-12-31 2015-07-02 Echostar Technologies L.L.C. Transponder loss and join-in-progress channel monitoring

Family Cites Families (55)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6078349A (en) * 1995-06-07 2000-06-20 Compaq Computer Corporation Process and system for increasing the display resolution of a point-to-point video transmission relative to the actual amount of video data sent
US6708335B1 (en) 1999-08-18 2004-03-16 Webtv Networks, Inc. Tracking viewing behavior of advertisements on a home entertainment system
US8910199B2 (en) 2000-02-25 2014-12-09 Interval Licensing Llc Targeted television content display
US6904408B1 (en) 2000-10-19 2005-06-07 Mccarthy John Bionet method, system and personalized web content manager responsive to browser viewers' psychological preferences, behavioral responses and physiological stress indicators
US7284201B2 (en) * 2001-09-20 2007-10-16 Koninklijke Philips Electronics N.V. User attention-based adaptation of quality level to improve the management of real-time multi-media content delivery and distribution
US8561095B2 (en) 2001-11-13 2013-10-15 Koninklijke Philips N.V. Affective television monitoring and control in response to physiological data
US7409639B2 (en) 2003-06-19 2008-08-05 Accenture Global Services Gmbh Intelligent collaborative media
US8156444B1 (en) 2003-12-31 2012-04-10 Google Inc. Systems and methods for determining a user interface attribute
US20050289582A1 (en) 2004-06-24 2005-12-29 Hitachi, Ltd. System and method for capturing and using biometrics to review a product, service, creative work or thing
US20070005752A1 (en) 2005-06-29 2007-01-04 Jitendra Chawla Methods and apparatuses for monitoring attention of a user during a collaboration session
WO2006043925A1 (en) 2004-10-14 2006-04-27 John Mccarthy Bionet system for personalizing web content manager
US7492821B2 (en) * 2005-02-08 2009-02-17 International Business Machines Corporation System and method for selective image capture, transmission and reconstruction
US20070214471A1 (en) * 2005-03-23 2007-09-13 Outland Research, L.L.C. System, method and computer program product for providing collective interactive television experiences
US20070038516A1 (en) 2005-08-13 2007-02-15 Jeff Apple Systems, methods, and computer program products for enabling an advertiser to measure user viewing of and response to an advertisement
US20080169930A1 (en) 2007-01-17 2008-07-17 Sony Computer Entertainment Inc. Method and system for measuring a user's level of attention to content
JP2008205861A (en) * 2007-02-20 2008-09-04 Matsushita Electric Ind Co Ltd Viewing and listening quality determining apparatus, viewing and listening quality determining method, viewing and listening quality determining program, and storage medium
US8764652B2 (en) 2007-03-08 2014-07-01 The Nielson Company (US), LLC. Method and system for measuring and ranking an “engagement” response to audiovisual or interactive media, products, or activities using physiological signals
US8392503B2 (en) 2007-06-19 2013-03-05 Cisco Technology, Inc. Reporting participant attention level to presenter during a web-based rich-media conference
US7673088B2 (en) * 2007-06-29 2010-03-02 Microsoft Corporation Multi-tasking interference model
US7889073B2 (en) 2008-01-31 2011-02-15 Sony Computer Entertainment America Llc Laugh detector and system and method for tracking an emotional response to a media presentation
US8630192B2 (en) 2009-01-28 2014-01-14 Headwater Partners I Llc Verifiable and accurate service usage monitoring for intermediate networking devices
US20100070987A1 (en) 2008-09-12 2010-03-18 At&T Intellectual Property I, L.P. Mining viewer responses to multimedia content
US8156054B2 (en) 2008-12-04 2012-04-10 At&T Intellectual Property I, L.P. Systems and methods for managing interactions between an individual and an entity
US20120017231A1 (en) 2009-09-15 2012-01-19 Jackson Chao Behavior monitoring system
US8875167B2 (en) 2009-09-21 2014-10-28 Mobitv, Inc. Implicit mechanism for determining user response to media
US20110321073A1 (en) 2010-06-23 2011-12-29 Yarvis Mark D Techniques for customization
US8949871B2 (en) 2010-09-08 2015-02-03 Opentv, Inc. Smart media selection based on viewer user presence
US8493390B2 (en) * 2010-12-08 2013-07-23 Sony Computer Entertainment America, Inc. Adaptive displays using gaze tracking
US9009298B2 (en) 2010-12-10 2015-04-14 The Nielsen Company (Us), Llc Methods and apparatus to determine audience engagement indices associated with media presentations
CN103688245A (en) 2010-12-30 2014-03-26 安比恩特兹公司 Information processing using a population of data acquisition devices
US20120210383A1 (en) * 2011-02-11 2012-08-16 Sayers Craig P Presenting streaming media for an event
US9049259B2 (en) 2011-05-03 2015-06-02 Onepatont Software Limited System and method for dynamically providing visual action or activity news feed
GB2494945A (en) * 2011-09-22 2013-03-27 Ibm Targeted digital media content delivery based on emotional change
US8635637B2 (en) 2011-12-02 2014-01-21 Microsoft Corporation User interface presenting an animated avatar performing a media reaction
US8943526B2 (en) * 2011-12-02 2015-01-27 Microsoft Corporation Estimating engagement of consumers of presented content
US9100685B2 (en) 2011-12-09 2015-08-04 Microsoft Technology Licensing, Llc Determining audience state or interest using passive sensor data
US20130205314A1 (en) * 2012-02-07 2013-08-08 Arun Ramaswamy Methods and apparatus to select media based on engagement levels
WO2012103820A2 (en) * 2012-03-08 2012-08-09 华为技术有限公司 Method, device, and system for highlighting party of interest
WO2013184488A1 (en) 2012-06-05 2013-12-12 Almondnet, Inc. Targeted television advertising based on a profile linked to an online device associated with a content-selecting device
KR20150021039A (en) * 2012-06-14 2015-02-27 톰슨 라이센싱 Method, apparatus and system for determining viewer reaction to content elements
US8931064B2 (en) 2012-12-18 2015-01-06 Bank Of America Corporation Identity attribute exchange and validation ecosystem
WO2014108194A1 (en) * 2013-01-10 2014-07-17 Telefonaktiebolaget L M Ericsson (Publ) Apparatus and method for controlling adaptive streaming of media
US20140298364A1 (en) * 2013-03-26 2014-10-02 Rawllin International Inc. Recommendations for media content based on emotion
US20140337880A1 (en) 2013-05-08 2014-11-13 Shayna Sorbel Peer-to-Peer Communication Advertising Method
US20140337868A1 (en) * 2013-05-13 2014-11-13 Microsoft Corporation Audience-aware advertising
CN105264906A (en) * 2013-06-05 2016-01-20 汤姆逊许可公司 Method and apparatus for content distribution for multiscreen viewing
WO2015031671A1 (en) 2013-08-30 2015-03-05 Biscotti Inc. Physical presence and advertising
CN105830108A (en) 2013-09-20 2016-08-03 交互数字专利控股公司 Verification Of Ad Impressions In User-Adptive Multimedia Delivery Framework
EP3058873A4 (en) * 2013-10-17 2017-06-28 Natsume Research Institute, Co., Ltd. Device for measuring visual efficacy
GB201402536D0 (en) * 2014-02-13 2014-04-02 Piksel Inc Sensed content delivery
US10264211B2 (en) * 2014-03-14 2019-04-16 Comcast Cable Communications, Llc Adaptive resolution in software applications based on dynamic eye tracking
US20150281783A1 (en) * 2014-03-18 2015-10-01 Vixs Systems, Inc. Audio/video system with viewer-state based recommendations and methods for use therewith
US10129312B2 (en) * 2014-09-11 2018-11-13 Microsoft Technology Licensing, Llc Dynamic video streaming based on viewer activity
US20160182955A1 (en) * 2014-12-23 2016-06-23 Rovi Guides, Inc. Methods and systems for recommending media assets
US10200725B2 (en) * 2015-06-15 2019-02-05 Intel Corporation Adaptive data streaming based on virtual screen size

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140168277A1 (en) * 2011-05-10 2014-06-19 Cisco Technology Inc. Adaptive Presentation of Content
US20120324493A1 (en) * 2011-06-17 2012-12-20 Microsoft Corporation Interest-based video streams
US20130145385A1 (en) * 2011-12-02 2013-06-06 Microsoft Corporation Context-based ratings and recommendations for media
US8898687B2 (en) * 2012-04-04 2014-11-25 Microsoft Corporation Controlling a media program based on a media reaction
US20140366049A1 (en) * 2013-06-11 2014-12-11 Nokia Corporation Method, apparatus and computer program product for gathering and presenting emotional response to an event
US20150189376A1 (en) * 2013-12-31 2015-07-02 Echostar Technologies L.L.C. Transponder loss and join-in-progress channel monitoring

Also Published As

Publication number Publication date
US10542315B2 (en) 2020-01-21
US20170134803A1 (en) 2017-05-11

Similar Documents

Publication Publication Date Title
US20200177956A1 (en) Method and apparatus for content adaptation based on audience monitoring
US10491946B2 (en) Methods and systems for providing multiple video content streams
US10812752B2 (en) Method and apparatus to present multiple audio content
US10567810B2 (en) Method and apparatus for managing bandwidth in providing communication services
US11509958B2 (en) Automated content selection for groups
US10652622B2 (en) Method and apparatus for providing content based upon a selected language
US10433011B2 (en) Apparatus and method for providing programming information for media content to a wearable device
US10581779B2 (en) Method and apparatus for managing content distribution according to social networks
US10257585B2 (en) Method and apparatus for directed advertisement
US20210125231A1 (en) Targeted user digital embedded advertising
US11457278B2 (en) Method and apparatus for recording advertised media content

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION