US20140111693A9 - Method and system for video and film recommendation - Google Patents
Method and system for video and film recommendation Download PDFInfo
- Publication number
- US20140111693A9 US20140111693A9 US13/726,874 US201213726874A US2014111693A9 US 20140111693 A9 US20140111693 A9 US 20140111693A9 US 201213726874 A US201213726874 A US 201213726874A US 2014111693 A9 US2014111693 A9 US 2014111693A9
- Authority
- US
- United States
- Prior art keywords
- video
- scene
- digital video
- descriptors
- videos
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
- G06F16/735—Filtering based on additional data, e.g. user or group profiles
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9535—Search customisation based on user profiles and personalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/02—Marketing; Price estimation or determination; Fundraising
- G06Q30/0201—Market modelling; Market analysis; Collecting market data
- G06Q30/0203—Market surveys; Market polls
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/06—Buying, selling or leasing transactions
- G06Q30/0601—Electronic shopping [e-shopping]
- G06Q30/0603—Catalogue ordering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9538—Presentation of query results
Definitions
- the invention disclosed herein relates generally to methods and systems for analyzing and using digital video compositions, and more particularly to a method and system for determining the characteristics of a video composition using measurable criteria.
- video as a generic term to include all video formats including films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs, among others.
- videos films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs
- an artificial intelligence recommendation system and method accomplish this task by analyzing for each video format a database of films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips and television programs in order to identify key similarities between different videos, and then recommending videos to a user depending upon their preferences.
- the system uses a series of complex artificial intelligence algorithms to analyze a plurality of characteristics in a video composition.
- the characteristics analyzed are relevant variables for characterizing films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips and television programs, among others, including: image analysis (both static and video dynamics), sound analysis (including video, voice, sounds, and noise), script analysis, structural analysis (length, scene structure, scene changes etc), and metadata analysis (name of studio, director, actors, etc. . . . ).
- This approach enables the creation of “constellations” of videos with similar characteristics, even from different genres and styles, enabling fast, accurate, and less costly comparison of films, TV programs, and videos for potential market success, video classification, and recommendation.
- the video recommendation includes video discovery and personalization through all digital channels including internet, mobile, personal computers and hand set devices.
- FIG. 1 is a basic overview of a system according to the present invention.
- FIG. 2 is a flowchart of the process steps according to the present invention.
- FIG. 3 is a frame/photogram sequence showing an image when translating channels according to the present invention.
- FIG. 4 is a tone and saturation histogram according to the present invention.
- FIG. 5 is a frame/photogram histogram with a marker according to the present invention.
- FIG. 6 illustrates a processed frame/photogram sequence according to the present invention.
- FIG. 7 is a frame/photogram sequence for detecting abrupt changes of shots and scene according to the present invention.
- FIG. 8 represents the calculations made in a record/computation process of the optical flow parameter according to the present invention.
- FIG. 9 is a frame/photogram sequence displaying frequency analysis according to the present invention.
- FIG. 10 is a frame/photogram sequence of FIG. 9 having corresponding spectrums relocated into a polar system according to the present invention.
- FIGS. 11 and 12 illustrate examples of complex scene analysis according to the present invention.
- FIGS. 13 and 14 show monograms of selected sound tracks according to the present invention.
- FIG. 15 illustrates two paths of video analysis according to the present invention.
- FIG. 16 shows some examples of perceptually processed sequences of frames according to the present invention.
- FIG. 17 shows the ratings for a TV show over a period of time according to the present invention.
- FIG. 18 shows media analyzed parameters of FIG. 17 according to the present invention.
- FIG. 19 shows a graph of a growing function according to the present invention.
- FIG. 20 is a schematic overview of the technology and process according to the present invention.
- FIG. 21 is a schematic of the global system architecture of the present invention.
- FIG. 22 is a view of a portion of the universe of analyzed videos illustrating some success clusters.
- FIG. 23 is a view of a portion of the universe of analyzed videos shown in FIG. 22 , with some new videos superimposed thereon.
- FIG. 1 depicts a schematic view of a system architecture for enabling the transfer and processing of digital video files to an automated digital video file analysis tool in order to generate an output file that, as described in greater detail below, serves as a descriptor of the measurable characteristics of the particular video composition that was analyzed.
- an “analyst wrapper” script receives digital video files from the video database or a video file archive, preferably copies the particular digital video file to a location on a computer network on which the analysis system is stored, decompresses the digital video file when necessary, and passes the file to the analysis engine for processing.
- the output of such analysis engine is directed to an output text file that, in turn, may be used by the recommendation utilities described below.
- the raw materials for the system are films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs. These are stored in a digital file, which is the main starting point for all embodiments of the invention.
- the first step performed by the system is to analyze an existing digital video file in order to create a descriptive profile for the video characteristics of the file.
- the analysis portion reads a “video” file and extracts some data.
- This data can be represented as a series of numbers, which are the main input for future processing.
- Such processing depends on the final application, and can use algorithms such as Principal Components Analysis (PCA), KNearest Neighbors (kNN), etc.
- PCA Principal Components Analysis
- kNN KNearest Neighbors
- the processes start by analyzing a large and representative sample of video files.
- a database is created for each country consisting of movies that have performed well, within the best window time frame.
- historical data for movies would include: national and worldwide sales and distribution figures, revenues, DVD/video sales and rental figures, audience ratings, awards garnered, release date, season, permanence of the movie, re-releases of the movie, etc. Additional data may include data associated with the Title, director, actors, studio, music production, associated books for the script (if any) etc.
- the database is similar, but may also include historical information on minute-to-minute audience ratings for each program. The minute-to-minute information will be aggregated and disaggregated according to the learning systems described herein.
- the process analyzes several characteristics of the video, such as sound and camera movements, and measures how the characteristics change over time. Parameter analysis is described in U.S. Pat. No. 7,081,579 to Alcalde et al., the specification of which is included herein by reference, in its entirety.
- the purpose of the initial analysis performed by the system is to analyze a variety of physical parameters of the videos stored in the database.
- Such physical parameters describe quantifiable characteristics of the film or TV program that may be mathematically modeled to create a descriptive, electronic “descriptors' vector” for each video.
- the analyzed parameters are based on cognitive and perceptual analysis, and the system is referred to as a Media Intelligence Universe System.
- the characteristics have been identified to be the ones that are measurable and quantifiable. Often the characteristics are detected unconsciously. In general, the mix of parameters may be more important than any individual parameter.
- the system particularly analyzes one or more of the following characteristics for each video composition. Not all of the characteristics necessarily provide distinctions in the video program. Combinations of some or all of these characteristics may be employed without departing from the spirit and scope of the instant invention.
- the parameters that are mathematically analyzed can be divided into four primary components with accompanying sub-components, which include:
- FIG. 2 The main processes of the Media Intelligence Technology are shown in FIG. 2 .
- Process #1 organizes data in a Database and allows detecting and correcting errors in the input data, both in the files and structure.
- the data coming from the audience and from the programs are processed together.
- the videos are transformed (consisting of several recordings in QuickTime format, divided by advertisements in some cases) into sequences of frames/photograms and audio files.
- the frame/photogram extraction method has proven to be very efficient to analyze a great amount of videos quickly. While projecting on a high resolution, up to 13 ⁇ 10 frames/photograms (grabbed with a rate of two per second), it is possible to immediately display about 30 seconds per chapter. This display, together with the characters' identification, is carried out in parallel.
- Process #3 comprises the following main modules:
- the first processing unit calculates simple statistics (such as the amount of red, green, blue, saturation, light level/luminosity, tone, and other complex parameters such as the light level period (until the second order) and the statistic parameter called Hu's Momentum which is invariable with rotation and scale/range.
- FIG. 3 shows a frame/photogram sequence that shows an image when translating the channels red, green, and blue to tone, saturation, and light level/luminosity.
- the following processing unit calculates the tone and saturation histogram in order to distinguish the most present tonalities. As it can be observed, at the left top section of each photogram there is a black box of 4 ⁇ 4 strips corresponding to the tone-saturation two-dimensional histogram. This system is very useful to automatically locate scene and shot changes because the histogram should not vary with mouth movements and face gestures during a conversation among several characters of the video.
- the advertisements have different histograms compared with the rest of frames/photograms in the case of TV programs, with the objective of attracting more attention.
- the third processing unit looks for structures made up by lines, aiming to help with the determination of the visual complexity of an image.
- FIG. 5 shows the original frames/photograms with a marker (it appears in red color at the bottom of each frame/photogram). For this specific example, the marker provides information about the quantity of linear structures in the image.
- FIG. 6 illustrates a processed sequence.
- the next processing unit of an image is one of the most complex in the whole system. It deals with the calculation of the amount of movement between two consecutive frames/photograms (optical flow). From other point of view, the objective is to calculate if a frame/photogram sequence does not show abrupt changes of shots and scene (see FIG. 7 ).
- FIG. 8 represents the calculations made in the record/computation process of the optical flow parameter.
- the basic idea is to find an equivalent for every pixel in the frame/photogram below. If there is a short distance between both positions for all image pixels, the frame/photogram sequence is static.
- the image is divided into quadrants, which are analyzed separately in order to obtain their corresponding equivalents in the frame/photogram below and, like this, calculate the optical flux.
- the fifth image-processing unit analyzes data from a frequency point of view (the same way can be done with the sound).
- the frame/photogram sequence shown in FIG. 9 displays the frequency analysis where it can be extracted where the ‘image energy’ concentrates.
- FIG. 10 shows, for the same frames/photograms of the sequence in FIG. 9 , their corresponding spectrums relocated into a polar system where an average for all the components is calculated by using the same distance as the zero frequency.
- the last processing unit supplements/complements the complexity measurement by means of a color segmentation algorithm.
- An example of this analysis is presented in FIGS. 11 and 12 .
- the segmentation algorithm output sequence is shown in FIG. 12 .
- FIGS. 13 and 14 show a monogram for 5 minutes of audio as well as an example of two of the extracted parameters (percentile 70% of the frequency energy and the level of the most powerful harmonic). Sound analysis techniques are described in U.S. Pat. No. 7,081,579, which is incorporated herein by reference in its entirety.
- Process #4 establishes correlations among descriptors. Having obtained all necessary descriptors, a preliminary analysis that relates those parameters with audience data is processed. Some simple examples related to the underway study of cognitive data and its connection with the audience rating are presented below.
- the average rating for every scene type, as well as the associated average rating, the maximum rating and the standard drift/variation/deviation are analyzed. See below a data sample ordered according to the maximum rating.
- the associated average rating (average rating per minute of the scene group per defined intensity), the maximum rating, and the Standard variation/deviation are presented for every scene intensity (1-5) of video 1.
- the ELO rating is a means of comparing the relative strengths of chess players, devised by Professor Arpad Elo. Players gain or lose rating points depending on the ELO rating of their opponents. If a player wins a game of chess in a rated tournament, they gain a number of rating points that increases in proportion to the difference between their rating and their opponent's rating.
- the Elo rating does not supply any information on the individual aspects of a chess player's capabilities; it does not rate the individual style as a player, or how well his defense and game endings are. Ratings provide merely a comparison of performances, no more and no less. Nevertheless, the Elo rating system has proved to be a relatively accurate measure for predicting the outcome of chess matches, based on a quantified figure of the strength of individual chess players.
- the GELO Measures Output Module of the Media Intelligence Universe is based on an innovative measure system that generalizes the application of ELO rating systems that are commonly applied in chess. Hereinafter we will denominate this new Generalized ELO system the GELO system.
- the GELO measures allow in general when comparing new “Videos” (Films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs, among others) with previously successful videos and with the predefined measure of success stated in each case, to predict the success potential of the new video in relation to the pre-defined measure.
- the first module is the cognitive analysis module where scenes, commercials, and introductory materials are tagged with a time stamp in order to relate them to minute-to-minute rating data. Scenes are classified as interior, exterior, and also the type of scene (as described in process #3).
- Characters are also identified for each scene to determine which are the characters that produce the highest or lowest ratings. Other parameters are considered as for example the intensity of the scene, where each scene is measured with a number between 0 and 5.
- the perceptual parameters applied in this example include (but are not restricted to):
- Audio analysis music, voices, and sounds (sound power, statistical model, low and high frequencies power ratio, harmonic detector, power steps, and frequency centroid).
- FIG. 16 shows some examples of perceptually processed sequences of frames.
- hue measurement is shown in the upper left
- line structure detection in the upper right optical flow in the lower left
- image complexity measurement in the lower right is shown.
- the audience rating information is only given in a minute-by-minute basis, all this information is condensed by minutes using the minimum, the maximum, the mean and the standard deviation of all the parameters obtained during the same minute.
- the GELO number can be used to transform all this information into knowledge that reflects the causes underneath audience rating variation.
- Chapters and their ratings are displayed in a matrix with fast-time (minute-by-minute rating) in the columns and slow-time (chapter to chapter) in the rows. This can be seen in FIG. 17 , where the ratings for a TV show are displayed over a period of two months approximately. It is important to show that the start and the end time of the TV show vary from chapter to chapter.
- FIG. 17 shows the rating over two months for the same TV show (between 0% and 30%).
- the media analyzed parameters are also transformed into a similar matrix such as the one seen in FIG. 18 , which shows an example of image complexity mean (red) and image complexity standard deviation in a minute-by-minute basis.
- the GELO system establishes a different rating (from the ELO system) gain/loss step depending on the number of times an identity plays a game.
- the same approach can be extended to characters, scene intensity, and media parameters. This technique is executed in the following steps:
- the GELO number update step is given by:
- ⁇ s ⁇ square root over ( ⁇ a 2 + ⁇ b 2 ) ⁇
- the processes described herein measure innovation/prediction cycles in video structure by using spectrum variables for power law detection. They also measure deviation analysis from the universality trend through detection of cycles from the universality trend and the detection of innovation and prediction wavelets.
- software modules learn a user's video preferences.
- the software uses two modules: one called ADAM, a cluster recognition engine, and another called EVE, a recommendation engine.
- ADAM is a conceptual clustering engine that is based on physical pattern recognition models. This non-supervised learning system generates a hierarchical tree structure that is based on topological metrics, which automatically determines the final number of clusters while allowing for automated related variable detection.
- the methodology for detecting social trends is completely scalable, and has been successfully applied in many other areas. It is also used for the preliminary visualization engine described in more detail below.
- EVE is a non-linear kernel learner, which had been successfully used in many other commercial applications.
- the application of EVE in video is similar to the referred application “Music Taste Test” in U.S. Pat. No. 7,081,579.
- This supervised learning system uses technology that has been proven to outperform statistical and neural network systems.
- a mathematically elegant solution which is relatively easy to customize and refine, the algorithm uses a direct strategy to capture personal Von Neumann Morgenstern utility functions. Due to their elegant and parsimonious mathematical architecture, both ADAM and EVE have been easily ported to new operating system environments, such as Symbian 60.
- the system After the system has learned a user's video preferences, it can connect the user with video selections based on his or her likes and dislikes. User preference techniques are described in U.S. Pat. No. 7,081,579.
- FIG. 21 shows the server architecture that has been developed to interact with a user though an active web page.
- the user interacts locally with a user visual interface.
- the visual interface may be a downloaded Macromedia Flash Player that runs locally in the user terminal and communicates with the Media Intelligence Universe (MIU) Environment.
- MIU Media Intelligence Universe
- the starting point of the Media Intelligence Universe is the ability to extract quantitative information from a video stored in digital format.
- the different types of mathematical procedures used to extract characteristic parameters are described in detail in U.S. Pat. No. 7,081,579.
- the analysis module is designed to be extremely portable and self-constituent, which means that it contains all the information it requires.
- the input of the MIU server is a list of videos with their respective descriptors (set of real numbers that could define different type of signal analysis, for example, the mean frequency, the level of noise, the mean power, image noise, rhythm, camera movements, etc).
- a unique identifier is assigned to each video, which is used to retrieve metadata from the database, such as title, name of studio, director, actors, clip location, etc.
- auxiliary non-mathematical information is also sent (such as, the year and the genre, among other relevant information related to the video). All this information is stored in an ASCII file that the MIU Server can read, parse, and analyze.
- the format of the MIU Server input ASCII file may be as follows:
- PCA_ refers to specific parameter values and VAR_refers to the video descriptors.
- the PHP module has the possibility to select whether to use the PCAs or the descriptors. With PCAs, the calculations are done faster, but with the descriptors, it will be possible to also send weights to disable some descriptors and find, for example, videos with similar rhythm but with different moods.
- An important aspect of the present invention is to analyze historical data related to the commercial success of films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs in order to create a “success” database.
- Such parameters may include additional data such as total sales, date of release, awards, critical acclaim, and other common indicators of commercial success.
- the success database will enable comparisons between new or unreleased content and previously released content regarding the intrinsic “video” parameters (as the ones described above), in order to predict the market success of new projects as well as gain other meaningful data to inform other marketing decisions.
- Each analyzed video is mapped onto a grid and positioned according to its mathematical characteristics, as illustrated in FIG. 22 . Videos with mathematical similarities are positioned very close to one another. The videos that had been recent successes were grouped into a limited number of small “success” clusters all over the grid but with vast spaces between them.
- the techniques taught by the present invention can be used throughout the creation and lifetime of the project. Below is listed some of the utility in the Pre-production, Post production, and Marketing phases. However, the potential of the product can have an impact in all the Industry value chain.
- the system generally does not have a global application. However it can analyze separately the main ingredients of the video when ready (before the production) such as script, initial music tracks selection, actors voices, dailies, etc.
- the system of the present invention can aid in pacing, scene length, music placement, Foley recording, scene exclusion, etc.
- a new film in production if the project is close to a “success cluster” (of similar films), editing can be optimized in order to move the final production toward a “success cluster”.
- FIG. 23 is an illustration of three new videos as analyzed and graphed on the grid of “success clusters”. In this illustration, there is one video that falls squarely within a success cluster, one video that is on the edge of a success cluster and one video that is clearly outside the success clusters.
- the final product will be analyzed and compared to other releases enabling market comparisons, success potential objective measures, and clues to better position the product.
- a film recommendation system for TV programs or videos can be developed based on the similarities that determine the film clusters/TV program clusters or video clusters.
- the similarities are determined trough the EVE and ADAM non-supervised and supervised learning systems using affinity values.
- the affinity value is a rating that shows how closely related the mathematical patterns in one video are to another. The lower the affinity value between two videos the more closely related they are.
- the learning systems use artificial intelligence applications as well as other methods to analyze the underlying mathematical patterns in the videos.
- the technology can isolate and separate many distinct events that occur in films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs, among others. By doing this, combined with other mathematical calculations, the system can develop a highly accurate and scientific tool. By revealing some before-unseen scientific information about films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips and television programs, we can better understand the art and man's desire to be engulfed in compelling entertainment.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Business, Economics & Management (AREA)
- Finance (AREA)
- Accounting & Taxation (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Strategic Management (AREA)
- Development Economics (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- General Business, Economics & Management (AREA)
- Entrepreneurship & Innovation (AREA)
- Economics (AREA)
- Marketing (AREA)
- Game Theory and Decision Science (AREA)
- Library & Information Science (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Television Signal Processing For Recording (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
- This application is a division of U.S. patent application Ser. No. 11/881,810, filed with the U.S. Patent and Trademark Office on Jul. 27, 2007 entitled “Method and System for Video and Film Recommendation”, which is a continuation-in-part of copending and co-owned U.S. patent application Ser. No. 11/492,355, filed with the U.S. Patent and Trademark Office on Jul. 25, 2006 entitled “Method and System for Music Recommendation”, which is a continuation of co-owned U.S. patent application Ser. No. 10/678,505, filed with the U.S. Patent and Trademark Office on Oct. 3, 2003 entitled “Method and System for Music Recommendation”, now U.S. Pat. No. 7,081,579, which is based upon and claims benefit of copending and co-owned U.S. Provisional Patent Application Ser. No. 60/415,868 entitled “Method and System for Music Recommendation”, filed with the U.S. Patent and Trademark Office on Oct. 3, 2002 by the inventors herein, the specifications of which are incorporated herein by reference.
- This application also claims benefit of copending and co-owned U.S. Provisional Patent Application Ser. No. 60/857,627 entitled “Music Intelligence Universe Server”, filed with the U.S. Patent and Trademark Office on Nov. 8, 2006 by the inventors herein, the specification of which is incorporated herein by reference.
- 1. Field of the Invention
- The invention disclosed herein relates generally to methods and systems for analyzing and using digital video compositions, and more particularly to a method and system for determining the characteristics of a video composition using measurable criteria.
- 2. Description of the Background
- Historically, much of what attracts us to a particular song is found in the basic structure of the music. Particular rhythms, changes in key and certain melodic patterns define the psychological and very human response we all have to music. In a similar fashion, particular characteristic variables of films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs affect how viewers respond to the production.
- While the number of possible plot lines combined with all of the other variables in recorded films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips and television programs allow for a seemingly infinite number of combinations, the patterns that we find pleasing have not changed. By analyzing images, sound, script, and structure similar mathematical patterns can be discerned.
- This document will refer to the term “video” as a generic term to include all video formats including films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs, among others.
- It is an object of the present invention to provide a method and system for measuring the characteristics of a video composition, and establishing a collection of digital video compositions that may be sorted upon such characteristics.
- It is another object of the present invention to enable a method and system to compare digital video files to discover video similarities.
- It is another object of the present invention to provide a method and system for determining a preferred video characteristic profile for a video viewer.
- It is another object of the present invention to provide a method and system for analyzing “videos” (films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs) to make video recommendations, video classification, video discovery, personalized video games.
- It is another object of the present invention to enable a method and system to compare new digital video files with historical commercially successful videos.
- It is another object of the present invention to provide a method and system for analyzing “videos (films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips and television programs, among others) to predict market success.
- In accordance with the above objects, an artificial intelligence recommendation system and method is provided. The system and method accomplish this task by analyzing for each video format a database of films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips and television programs in order to identify key similarities between different videos, and then recommending videos to a user depending upon their preferences.
- The system uses a series of complex artificial intelligence algorithms to analyze a plurality of characteristics in a video composition. The characteristics analyzed are relevant variables for characterizing films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips and television programs, among others, including: image analysis (both static and video dynamics), sound analysis (including video, voice, sounds, and noise), script analysis, structural analysis (length, scene structure, scene changes etc), and metadata analysis (name of studio, director, actors, etc. . . . ). This approach enables the creation of “constellations” of videos with similar characteristics, even from different genres and styles, enabling fast, accurate, and less costly comparison of films, TV programs, and videos for potential market success, video classification, and recommendation. The video recommendation includes video discovery and personalization through all digital channels including internet, mobile, personal computers and hand set devices.
- The various features of novelty that characterize the invention will be pointed out with particularity in the claims of this application.
- Other objects, features, and advantages of the present invention will become more apparent from the following detailed description of the preferred embodiments and certain modifications thereof when taken together with the accompanying drawings, in which:
-
FIG. 1 is a basic overview of a system according to the present invention. -
FIG. 2 is a flowchart of the process steps according to the present invention. -
FIG. 3 is a frame/photogram sequence showing an image when translating channels according to the present invention. -
FIG. 4 is a tone and saturation histogram according to the present invention. -
FIG. 5 is a frame/photogram histogram with a marker according to the present invention. -
FIG. 6 illustrates a processed frame/photogram sequence according to the present invention. -
FIG. 7 is a frame/photogram sequence for detecting abrupt changes of shots and scene according to the present invention. -
FIG. 8 represents the calculations made in a record/computation process of the optical flow parameter according to the present invention. -
FIG. 9 is a frame/photogram sequence displaying frequency analysis according to the present invention. -
FIG. 10 is a frame/photogram sequence ofFIG. 9 having corresponding spectrums relocated into a polar system according to the present invention. -
FIGS. 11 and 12 illustrate examples of complex scene analysis according to the present invention. -
FIGS. 13 and 14 show monograms of selected sound tracks according to the present invention. -
FIG. 15 illustrates two paths of video analysis according to the present invention. -
FIG. 16 shows some examples of perceptually processed sequences of frames according to the present invention. -
FIG. 17 shows the ratings for a TV show over a period of time according to the present invention. -
FIG. 18 shows media analyzed parameters ofFIG. 17 according to the present invention. -
FIG. 19 shows a graph of a growing function according to the present invention. -
FIG. 20 is a schematic overview of the technology and process according to the present invention. -
FIG. 21 is a schematic of the global system architecture of the present invention. -
FIG. 22 is a view of a portion of the universe of analyzed videos illustrating some success clusters. -
FIG. 23 is a view of a portion of the universe of analyzed videos shown inFIG. 22 , with some new videos superimposed thereon. - The invention summarized above and defined by the enumerated claims may be better understood by referring to the following description, which should be read in conjunction with the accompanying drawings. This description of an embodiment, set out below to enable one to build and use an implementation of the invention, is not intended to limit the enumerated claims, but to serve as particular examples thereof. Those skilled in the art should appreciate that they may readily use the conception and specific embodiments disclosed as a basis for modifying or designing other methods and systems for carrying out the same purposes of the present invention. Those skilled in the art should also realize that such equivalent assemblies do not depart from the spirit and scope of the invention in its broadest form.
-
FIG. 1 depicts a schematic view of a system architecture for enabling the transfer and processing of digital video files to an automated digital video file analysis tool in order to generate an output file that, as described in greater detail below, serves as a descriptor of the measurable characteristics of the particular video composition that was analyzed. As shown inFIG. 1 , an “analyst wrapper” script receives digital video files from the video database or a video file archive, preferably copies the particular digital video file to a location on a computer network on which the analysis system is stored, decompresses the digital video file when necessary, and passes the file to the analysis engine for processing. The output of such analysis engine is directed to an output text file that, in turn, may be used by the recommendation utilities described below. - The raw materials for the system are films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs. These are stored in a digital file, which is the main starting point for all embodiments of the invention. The first step performed by the system is to analyze an existing digital video file in order to create a descriptive profile for the video characteristics of the file. In a first stage, the analysis portion reads a “video” file and extracts some data. This data can be represented as a series of numbers, which are the main input for future processing. Such processing depends on the final application, and can use algorithms such as Principal Components Analysis (PCA), KNearest Neighbors (kNN), etc.
- The processes, according to the present invention, start by analyzing a large and representative sample of video files. A database is created for each country consisting of movies that have performed well, within the best window time frame. For example, historical data for movies would include: national and worldwide sales and distribution figures, revenues, DVD/video sales and rental figures, audience ratings, awards garnered, release date, season, permanence of the movie, re-releases of the movie, etc. Additional data may include data associated with the Title, director, actors, studio, music production, associated books for the script (if any) etc. For TV programs, the database is similar, but may also include historical information on minute-to-minute audience ratings for each program. The minute-to-minute information will be aggregated and disaggregated according to the learning systems described herein. The process analyzes several characteristics of the video, such as sound and camera movements, and measures how the characteristics change over time. Parameter analysis is described in U.S. Pat. No. 7,081,579 to Alcalde et al., the specification of which is included herein by reference, in its entirety.
- The purpose of the initial analysis performed by the system is to analyze a variety of physical parameters of the videos stored in the database. Such physical parameters describe quantifiable characteristics of the film or TV program that may be mathematically modeled to create a descriptive, electronic “descriptors' vector” for each video. Moreover, the analyzed parameters are based on cognitive and perceptual analysis, and the system is referred to as a Media Intelligence Universe System. The characteristics have been identified to be the ones that are measurable and quantifiable. Often the characteristics are detected unconsciously. In general, the mix of parameters may be more important than any individual parameter. To implement the methods described herein, the system particularly analyzes one or more of the following characteristics for each video composition. Not all of the characteristics necessarily provide distinctions in the video program. Combinations of some or all of these characteristics may be employed without departing from the spirit and scope of the instant invention.
- The parameters that are mathematically analyzed can be divided into four primary components with accompanying sub-components, which include:
- 1. Static Image Analysis:
-
- Brightness Scale
- Gamma Scale
- Color Spectrum
- Object Presence and Focus
- Picture's Spatial Correlation
- Image textures
- Symmetries
- Image Noise
- 2. Video and Image Movements:
-
- Cut Frequency
- Camera Movements: Static Camera, zoom, panning
- Rhythm
- Scene objects' movement
- Scene Length
- Object presence: Faces, stenography elements, etc
- 3. Script Analysis:
-
- Key words, stop words, stemming techniques
- Nominal sintagmas, expressions, combination of key words
- Structural information
- Mood related parameters
- 4. Music and Sound
-
- Sound
- Voices
- Noise
- Music Variables including
- Sonic Brilliance
- Timbre
- Rhythm and Tempo
- Placement of Music
- Volume and Amplitude
- Octave
- Mood related parameters of the music
- The main processes of the Media Intelligence Technology are shown in
FIG. 2 . -
- PROCESS 1: To order the input data by means of a database.
- PROCESS 2: To transform the videos into QuickTime format with image sequences, and to isolate the audio.
- PROCESS 3: To extract video descriptors and metadata from the scripts and cognitive scene structure and classification.
- PROCESS 4: To establish correlations among descriptors.
- PROCESS 5: To establish output measures.
-
Process # 1 organizes data in a Database and allows detecting and correcting errors in the input data, both in the files and structure. In the case of TV programs, the data coming from the audience and from the programs are processed together. - For
Process # 2, the videos are transformed (consisting of several recordings in QuickTime format, divided by advertisements in some cases) into sequences of frames/photograms and audio files. - As an example, we will describe a chapter of a soap opera. It can be observed that the advertisements are located immediately after the soap opera. However, most of the advertisements are also divided into video files respectively. In the case of TV programs, frames/photograms provide a lot of information concerning the introduction of advertisements.
- The frame/photogram extraction method has proven to be very efficient to analyze a great amount of videos quickly. While projecting on a high resolution, up to 13×10 frames/photograms (grabbed with a rate of two per second), it is possible to immediately display about 30 seconds per chapter. This display, together with the characters' identification, is carried out in parallel.
-
Process # 3 comprises the following main modules: - Module 1: Supervised Perceptual Analysis.
- Extraction of the amount of characters and identification of the characters in a scene without using the script information.
- Module 2: Supervised Cognitive Analysis.
- Three parameters are evaluated for each scene.
- 2.1 Scene Character:
- The first of them is the scene typology. Some of the possible values considered include:
-
- a. COMIC RELIEF
- b. FRIENDSHIP
- c. LOVE
- d. FAMILY LOVE
- e. ATTRACTION
- f. JEALOUSY
- g. DISAPPOINTMENT/LACK OF LOVE
- h. DISCUSSION/PSYCHOLOGICAL VIOLENCE
- i. UNFAITHFULNESS/DECEPTION
- j. DAYDREAM
- k. END OF PLOT
- l. TURN-PLOT
- m. DRAMATIC IMPACT
- n. START OF PLOT
- o. INTEREST
- p. INTRIGUE
- q. DEATH
- r. CHARACTERS' PRESENTATION
- s. NO APPARENT CONTRIBUTION
- t. SUSPENSE
- u. LOCATION
- v. TRANSITION
- w. CHARACTERS
- 2.2 Scene Intensity:
- Level of intensity inside the contextual thread followed by the chapter. It corresponds to a number between 1 and 5. This supervised analysis has been carried out for both soap operas.
- 2.3 Type of Scene:
- Advertisement, credits, or scene with characters from the series.
- Module 3: Automatic Perceptual Analysis.
- For this analysis, several techniques can be used. The results of applying a frame/photogram input rate to the system of two frames/photograms per second and an audio sample frequency of 24 kHz with only one channel is presented as an example.
- The first processing unit calculates simple statistics (such as the amount of red, green, blue, saturation, light level/luminosity, tone, and other complex parameters such as the light level period (until the second order) and the statistic parameter called Hu's Momentum which is invariable with rotation and scale/range.
FIG. 3 shows a frame/photogram sequence that shows an image when translating the channels red, green, and blue to tone, saturation, and light level/luminosity. - The following processing unit calculates the tone and saturation histogram in order to distinguish the most present tonalities. As it can be observed, at the left top section of each photogram there is a black box of 4×4 strips corresponding to the tone-saturation two-dimensional histogram. This system is very useful to automatically locate scene and shot changes because the histogram should not vary with mouth movements and face gestures during a conversation among several characters of the video.
- It is possible to observe this characteristic in the photogram sequence. As seen in
FIG. 4 , the advertisements have different histograms compared with the rest of frames/photograms in the case of TV programs, with the objective of attracting more attention. - The third processing unit looks for structures made up by lines, aiming to help with the determination of the visual complexity of an image.
FIG. 5 shows the original frames/photograms with a marker (it appears in red color at the bottom of each frame/photogram). For this specific example, the marker provides information about the quantity of linear structures in the image.FIG. 6 illustrates a processed sequence. - The next processing unit of an image is one of the most complex in the whole system. It deals with the calculation of the amount of movement between two consecutive frames/photograms (optical flow). From other point of view, the objective is to calculate if a frame/photogram sequence does not show abrupt changes of shots and scene (see
FIG. 7 ). -
FIG. 8 represents the calculations made in the record/computation process of the optical flow parameter. The basic idea is to find an equivalent for every pixel in the frame/photogram below. If there is a short distance between both positions for all image pixels, the frame/photogram sequence is static. - As we can observe, the image is divided into quadrants, which are analyzed separately in order to obtain their corresponding equivalents in the frame/photogram below and, like this, calculate the optical flux. The fifth image-processing unit analyzes data from a frequency point of view (the same way can be done with the sound). The frame/photogram sequence shown in
FIG. 9 displays the frequency analysis where it can be extracted where the ‘image energy’ concentrates. - This processing is repeated for the image in a grey scale, and for the components in red, green, and blue.
FIG. 10 shows, for the same frames/photograms of the sequence inFIG. 9 , their corresponding spectrums relocated into a polar system where an average for all the components is calculated by using the same distance as the zero frequency. - The last processing unit supplements/complements the complexity measurement by means of a color segmentation algorithm. The more complex is a scene, the more objects it contains. An example of this analysis is presented in
FIGS. 11 and 12 . The segmentation algorithm output sequence is shown inFIG. 12 . - It is possible to observe how the upper frames/photograms are less complex than the last frames/photograms from the sequence.
- Concerning the automatic analysis of the audio parameters, the following list of Basic parameters is processed at a first level:
- 1. Sound power
- 2. Spectrum fractiles (50% to 90%) from low frequencies
- 3. Maximum frequency and frequency centroid
- 4. Power of low, medium and high frequencies
- 5. Power break (low, medium, and high totals).
- 6. Harmonic detector (frequency and power)
- With these basic parameters, it is possible to develop more complex parameters such as music detectors, music and voice, or voice alone.
FIGS. 13 and 14 show a monogram for 5 minutes of audio as well as an example of two of the extracted parameters (percentile 70% of the frequency energy and the level of the most powerful harmonic). Sound analysis techniques are described in U.S. Pat. No. 7,081,579, which is incorporated herein by reference in its entirety. -
Process # 4 establishes correlations among descriptors. Having obtained all necessary descriptors, a preliminary analysis that relates those parameters with audience data is processed. Some simple examples related to the underway study of cognitive data and its connection with the audience rating are presented below. - 1. Rating/
Scene Character Video 1. - The average rating for every scene type, as well as the associated average rating, the maximum rating and the standard drift/variation/deviation are analyzed. See below a data sample ordered according to the maximum rating.
-
SCENE TYPE Rating Average Std dev Maximum Intrigue 8.10% 1.29% 12.60% Dramatic impact 8.40% 1.25% 12.40% No apparent contribution 7.80% 1.36% 12.40% Violence 8.20% 1.10% 12.30% Interest 8.30% 1.25% 12.10% Suspense 8.30% 1.18% 12.10% Friendship 7.80% 1.28% 11.70% Love 8.60% 1.42% 11.70% End of plot 10.60% 0.70% 11.50% Family love 8.30% 1.32% 11.40% Daydream 8.00% 1.31% 11.40% Transition 8.30% 1.12% 11.20% Disappointment/lack of 7.60% 0.94% 11.10% love Unfaithfulness/Deception 8.00% 1.00% 10.50% Comic relief 7.80% 0.86% 9.80% Attraction 7.30% 0.96% 9.70% Turn-Plot 7.30% 1.32% 9.70% Stara of plot 7.80% 1.23% 9.60% Jealousy 7.30% 0.95% 8.90% Death 7.60% 0.54% 8.70% Characters' presentation 7.70% 0.31% 8.00% - 2. Rating/
Scene Character Video 2 -
SCENE TYPE Average Std dev Max Comic relief 30.40% 3.78% 38.90% Characters 32.30% 3.58% 38.90% Interest 30.60% 3.84% 38.80% No apparent contribution 27.80% 5.74% 38.50% Friendship 31.30% 3.56% 38.20% Intrigue 30.50% 3.50% 38.10% Unfaithfulness 30.70% 3.78% 37.90% Transition 32.30% 4.48% 37.90% Family love 30.60% 3.24% 37.50% Suspense 28.80% 4.37% 37.30% Dramatic impact 31.20% 3.36% 37.10% Disappointment 30.10% 3.39% 36.90% Discussion 31.10% 3.56% 36.50% Jealousy 30.20% 2.81% 36.20% Daydream 29.50% 4.14% 35.90% Love 30.60% 3.20% 35.50% Attraction 29.90% 3.25% 35.50% Characters' presentation 30.20% 3.62% 34.10% Death 27.60% 4.11% 32.40% Turn-plot 29.40% 1.55% 31.90% Location 25.70% 0.50% 26.20% Stara plot 24.90% 0.00% 24.90% - 3. Rating/
Scene Intensity Video 1 - The associated average rating (average rating per minute of the scene group per defined intensity), the maximum rating, and the Standard variation/deviation are presented for every scene intensity (1-5) of
video 1. -
Level Average Std dev Max 5 8.4% 1.23% 12.60% 2 7.6% 1.18% 12.40% 4 8.0% 1.21% 12.10% 3 7.4% 1.11% 11.30% 1 7.7% 0.67% 8.50% -
Process # 5 Measures the Output. Generalized ELO (“GELO”) Module - The ELO rating is a means of comparing the relative strengths of chess players, devised by Professor Arpad Elo. Players gain or lose rating points depending on the ELO rating of their opponents. If a player wins a game of chess in a rated tournament, they gain a number of rating points that increases in proportion to the difference between their rating and their opponent's rating.
- Of course, the Elo rating does not supply any information on the individual aspects of a chess player's capabilities; it does not rate the individual style as a player, or how well his defense and game endings are. Ratings provide merely a comparison of performances, no more and no less. Nevertheless, the Elo rating system has proved to be a relatively accurate measure for predicting the outcome of chess matches, based on a quantified figure of the strength of individual chess players.
- The GELO Measures Output Module of the Media Intelligence Universe is based on an innovative measure system that generalizes the application of ELO rating systems that are commonly applied in chess. Hereinafter we will denominate this new Generalized ELO system the GELO system.
- The GELO measures allow in general when comparing new “Videos” (Films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs, among others) with previously successful videos and with the predefined measure of success stated in each case, to predict the success potential of the new video in relation to the pre-defined measure.
- As an example (but by no means restricting the scope of application), and in order to fix ideas the method is described as applied to TV shows (exemplifying the concept of “video”) and the correlation between GELO measures and the audience ratings (exemplifying the concept of pre-defined measure of success). We will also present the way knowledge of the causes of TV rating gain or loss can be extracted using the GELO system to cognitive and image processing parameters such as image complexity, scene type (violence, love, family love, or action), presence of certain characters, audio spectrum, etc. The same process can be used for any video type.
- To understand the behavior of GELO measures for audience ratings for a TV show (and also the competitors' evolution) it is necessary to process the historical successful videos in two separate ways as illustrated in
FIG. 15 . - The first module is the cognitive analysis module where scenes, commercials, and introductory materials are tagged with a time stamp in order to relate them to minute-to-minute rating data. Scenes are classified as interior, exterior, and also the type of scene (as described in process #3).
- Characters are also identified for each scene to determine which are the characters that produce the highest or lowest ratings. Other parameters are considered as for example the intensity of the scene, where each scene is measured with a number between 0 and 5.
- The second module as described in process #3: the perceptual analysis module would be applied in this case to gather visual and audio parameters from frames and sequences of frames in order to look for relationships with the minute-to-minute ratings. The perceptual parameters applied in this example include (but are not restricted to):
- 1. Simple red, green, blue, saturation, hue, and luminosity statistical measurements per frame.
- 2. Hue and saturation two-dimensional histogram
- 3. Line structures
- 4. Optical flow to detect movement and camera displacement over the scene
- 5. Fourier Spectral histogram of the luminosity map
- 6. Image complexity
- 7. Audio analysis: music, voices, and sounds (sound power, statistical model, low and high frequencies power ratio, harmonic detector, power steps, and frequency centroid).
-
FIG. 16 shows some examples of perceptually processed sequences of frames. InFIG. 16 hue measurement is shown in the upper left, line structure detection in the upper right, optical flow in the lower left, and image complexity measurement in the lower right. - All the chapters are pre-processed with both modules (cognitive and perceptual modules) systematically creates a large database of media descriptors, scene types, characters identification, etc.
- Due to the fact that, in this example, the audience rating information is only given in a minute-by-minute basis, all this information is condensed by minutes using the minimum, the maximum, the mean and the standard deviation of all the parameters obtained during the same minute.
- Once all the information has been processed and included into a database, then the GELO number can be used to transform all this information into knowledge that reflects the causes underneath audience rating variation. Chapters and their ratings are displayed in a matrix with fast-time (minute-by-minute rating) in the columns and slow-time (chapter to chapter) in the rows. This can be seen in
FIG. 17 , where the ratings for a TV show are displayed over a period of two months approximately. It is important to show that the start and the end time of the TV show vary from chapter to chapter.FIG. 17 shows the rating over two months for the same TV show (between 0% and 30%). - The media analyzed parameters are also transformed into a similar matrix such as the one seen in
FIG. 18 , which shows an example of image complexity mean (red) and image complexity standard deviation in a minute-by-minute basis. - This allows the application of the GELO number for the immediate term (that means, comparing ratings inside the same chapter) and for the long term (comparing between chapters separated in time).
- The GELO system establishes a different rating (from the ELO system) gain/loss step depending on the number of times an identity plays a game. The same approach can be extended to characters, scene intensity, and media parameters. This technique is executed in the following steps:
- 1) For each minute of each chapter, the information considered includes (but is not restricted to) the following:
-
- a) type and intensity of all scene present in this minute
- b) characters that appear in this minute
- c) number of seconds of commercials
- d) number of seconds of introductory material from the TV show
- e) evaluation of the minimum, the maximum, the mean and the standard deviation of all the media parameters of the frames belonging to the minute and separate them by the clustered classes
- f) Audience rating values for this minute.
- 2) Meaningful pairs of elements that will compete in groups within a modified GELO rating system; for example, pairs of characters, groups of characters, noise powers, image complexity classes, etc.
- 3) Selected pairs of elements belonging to the same chapter are evaluated using the GELO system and the procedure is repeated for the rest of chapters. In the same way, selected elements from different chapters are evaluated (again with GELO) in order to gather knowledge about the chapter-to-chapter rating changes.
- 4) The GELO number update step is given by:
-
- a) Sa is the TV show audience rating for player A
- b) Sb is the TV show audience rating for player B
- c) Ra is the applied GELO rating for player A
- d) Rb is the applied GELO rating for player B
- e) The new rating Wa is updated with the equation: R′a=Ra+Cab Ka (Sa−Sb)
- f) Where Ka is the inverse of the number of games in which player A participates, and
- g) Cab determines the rating update related to the entropy of the result of the game. This means that if the result is not the expected, then GELO number rating will increase or decrease more than if the results are likely the expected. For the first games of every player, Cab does not take effect in order to achieve a good statistical base. In other words, Cab=0 if the result is the expected one and Cab=1 if the result is an absolute surprise.
- h) Being ma (mean) and ∫a (standard deviation) the parameters that estimate the a posteriori TV rating difference performance of player A against its opponents (variable A) and mb (mean) and ∫b (standard deviation) an opponent posteriori same parameters (variable B), then the subtraction of these two random variables is also a Gaussian random variable with mean (ms=ma−mb) and standard deviation:
-
σs=√{square root over (σa 2+σb 2)} -
- i) The probability to win for player A against player B is given by:
-
-
- j) To finally evaluate Cab, given the result of the game (A−B, that is, the difference between the audience ratings of both players), the normalized distance is evaluated as a measure of the entropy:
-
-
- k) The last step is to apply a growing function which transforms the u value into Cab, taking into consideration its Gaussian nature (i.e. the
region 0<|u|<0.5 corresponds to the 68% of the distribution, the region 0.5<|u|<1 to the 95%, etc). The selection of the transforming function contemplates heuristics analysis and is always a growing function, such as the one shown inFIG. 19 .
- k) The last step is to apply a growing function which transforms the u value into Cab, taking into consideration its Gaussian nature (i.e. the
- The procedure followed to apply the new GELO methodology to TV show audience ratings has been described in detail as an example on how this is applied for any kind of video with any pre-defined success measure.
- Some non-trivial and key generalizations of the ELO concept have been invented in order to design the GELO rating system that is applicable to “video competitions”. Furthermore, the ability to seek the videos in sets of perceptual and cognitive temporal parameters and apply the GELO rating model to them, represents the next step beyond the state of art of video analysis. The direct application to TV programs presented shows the next generation—beyond the state of the art—for TV rating analysis.
- In a preferred embodiment, the processes described herein measure innovation/prediction cycles in video structure by using spectrum variables for power law detection. They also measure deviation analysis from the universality trend through detection of cycles from the universality trend and the detection of innovation and prediction wavelets.
- As shown in
FIG. 20 , following analysis of the video characteristics, software modules according to a preferred embodiment of the present invention learn a user's video preferences. The software uses two modules: one called ADAM, a cluster recognition engine, and another called EVE, a recommendation engine. - ADAM is a conceptual clustering engine that is based on physical pattern recognition models. This non-supervised learning system generates a hierarchical tree structure that is based on topological metrics, which automatically determines the final number of clusters while allowing for automated related variable detection. The methodology for detecting social trends is completely scalable, and has been successfully applied in many other areas. It is also used for the preliminary visualization engine described in more detail below.
- EVE is a non-linear kernel learner, which had been successfully used in many other commercial applications. The application of EVE in video is similar to the referred application “Music Taste Test” in U.S. Pat. No. 7,081,579. This supervised learning system uses technology that has been proven to outperform statistical and neural network systems. A mathematically elegant solution which is relatively easy to customize and refine, the algorithm uses a direct strategy to capture personal Von Neumann Morgenstern utility functions. Due to their elegant and parsimonious mathematical architecture, both ADAM and EVE have been easily ported to new operating system environments, such as Symbian 60.
- After the system has learned a user's video preferences, it can connect the user with video selections based on his or her likes and dislikes. User preference techniques are described in U.S. Pat. No. 7,081,579.
-
FIG. 21 shows the server architecture that has been developed to interact with a user though an active web page. The user interacts locally with a user visual interface. In a preferred embodiment, the visual interface may be a downloaded Macromedia Flash Player that runs locally in the user terminal and communicates with the Media Intelligence Universe (MIU) Environment. The MIU environment is made of several modules: -
- The database module consists of a database server, such as a MySQL Database, that stores metadata information for videos (name of studio, director, actors, clip location, etc).
- The video clip storage module stores representative clips of all the videos that are in the database so the visual interface can download them.
- The MIU server module makes all the computations necessary to provide the services. In a preferred embodiment, it is an ANSI C++ portable application that communicates via sockets.
- The Web server module manages all the communications with the rest of modules and with the user visual interface. In a preferred embodiment, the web server module may be an active web page coded in PHP that runs under an Apache Tomcat Web server.
- The starting point of the Media Intelligence Universe is the ability to extract quantitative information from a video stored in digital format. The different types of mathematical procedures used to extract characteristic parameters are described in detail in U.S. Pat. No. 7,081,579. The analysis module is designed to be extremely portable and self-constituent, which means that it contains all the information it requires. Accordingly, the input of the MIU server is a list of videos with their respective descriptors (set of real numbers that could define different type of signal analysis, for example, the mean frequency, the level of noise, the mean power, image noise, rhythm, camera movements, etc). A unique identifier is assigned to each video, which is used to retrieve metadata from the database, such as title, name of studio, director, actors, clip location, etc.
- Sometimes it is useful to apply a technique called Eigen Value Decomposition to find a reduced set of useful descriptors such as based on Principle Component Analysis (PCA) condensing the information because descriptors are not totally independent. So as to be able to filter the recommended list, some auxiliary non-mathematical information is also sent (such as, the year and the genre, among other relevant information related to the video). All this information is stored in an ASCII file that the MIU Server can read, parse, and analyze. The format of the MIU Server input ASCII file may be as follows:
- id;0;1;1;YEAR;GENRE;PCA1;PCA2;PCA3;VAR1;VAR2;VAR3;VAR4;VAR5 . . . .
- where PCA_refers to specific parameter values and VAR_refers to the video descriptors.
- The PHP module has the possibility to select whether to use the PCAs or the descriptors. With PCAs, the calculations are done faster, but with the descriptors, it will be possible to also send weights to disable some descriptors and find, for example, videos with similar rhythm but with different moods.
- An important aspect of the present invention is to analyze historical data related to the commercial success of films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs in order to create a “success” database. Such parameters may include additional data such as total sales, date of release, awards, critical acclaim, and other common indicators of commercial success. The success database will enable comparisons between new or unreleased content and previously released content regarding the intrinsic “video” parameters (as the ones described above), in order to predict the market success of new projects as well as gain other meaningful data to inform other marketing decisions. Each analyzed video is mapped onto a grid and positioned according to its mathematical characteristics, as illustrated in
FIG. 22 . Videos with mathematical similarities are positioned very close to one another. The videos that had been recent successes were grouped into a limited number of small “success” clusters all over the grid but with vast spaces between them. - The techniques taught by the present invention can be used throughout the creation and lifetime of the project. Below is listed some of the utility in the Pre-production, Post production, and Marketing phases. However, the potential of the product can have an impact in all the Industry value chain.
- 1) Pre-Production
- At the pre-production level, the system generally does not have a global application. However it can analyze separately the main ingredients of the video when ready (before the production) such as script, initial music tracks selection, actors voices, dailies, etc.
- 2) Post Production
- During postproduction, the system of the present invention can aid in pacing, scene length, music placement, Foley recording, scene exclusion, etc. Consider, for example, a new film in production, if the project is close to a “success cluster” (of similar films), editing can be optimized in order to move the final production toward a “success cluster”.
FIG. 23 is an illustration of three new videos as analyzed and graphed on the grid of “success clusters”. In this illustration, there is one video that falls squarely within a success cluster, one video that is on the edge of a success cluster and one video that is clearly outside the success clusters. - 3) Marketing
- The final product will be analyzed and compared to other releases enabling market comparisons, success potential objective measures, and clues to better position the product.
- 4) Film Recommendation for Final Consumers
- In the retail space, either physical or digital, for TV, films, or videos, a film recommendation system for TV programs or videos can be developed based on the similarities that determine the film clusters/TV program clusters or video clusters. The similarities are determined trough the EVE and ADAM non-supervised and supervised learning systems using affinity values. The affinity value is a rating that shows how closely related the mathematical patterns in one video are to another. The lower the affinity value between two videos the more closely related they are.
- The learning systems use artificial intelligence applications as well as other methods to analyze the underlying mathematical patterns in the videos. The technology can isolate and separate many distinct events that occur in films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips, and television programs, among others. By doing this, combined with other mathematical calculations, the system can develop a highly accurate and scientific tool. By revealing some before-unseen scientific information about films, video-shorts, mobile videos, music videos, videogames, advertisings, video clips and television programs, we can better understand the art and man's desire to be engulfed in compelling entertainment.
- The invention has been described with references to a preferred embodiment. While specific values, relationships, materials and steps have been set forth for purposes of describing concepts of the invention, it will be appreciated by persons skilled in the art that numerous variations and/or modifications may be made to the invention as shown in the specific embodiments without departing from the spirit or scope of the basic concepts and operating principles of the invention as broadly described. It should be recognized that, in the light of the above teachings, those skilled in the art could modify those specifics without departing from the invention taught herein. Having now fully set forth the preferred embodiments and certain modifications of the concept underlying the present invention, various other embodiments as well as certain variations and modifications of the embodiments herein shown and described will obviously occur to those skilled in the art upon becoming familiar with such underlying concept. It is intended to include all such modifications, alternatives and other embodiments insofar as they come within the scope of the appended claims or equivalents thereof. It should be understood, therefore, that the invention might be practiced otherwise than as specifically set forth herein. Consequently, the present embodiments are to be considered as illustrative and not restrictive.
Claims (14)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/726,874 US8841534B2 (en) | 2002-10-03 | 2012-12-26 | Method and system for video and film recommendation |
Applications Claiming Priority (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US41586802P | 2002-10-03 | 2002-10-03 | |
US10/678,505 US7081579B2 (en) | 2002-10-03 | 2003-10-03 | Method and system for music recommendation |
US11/492,355 US20060254411A1 (en) | 2002-10-03 | 2006-07-25 | Method and system for music recommendation |
US85762706P | 2006-11-08 | 2006-11-08 | |
US11/881,810 US8063295B2 (en) | 2002-10-03 | 2007-07-27 | Method and system for video and film recommendation |
US13/174,904 US8338685B2 (en) | 2002-10-03 | 2011-07-01 | Method and system for video and film recommendation |
US13/726,874 US8841534B2 (en) | 2002-10-03 | 2012-12-26 | Method and system for video and film recommendation |
Related Parent Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/881,810 Division US8063295B2 (en) | 2002-10-03 | 2007-07-27 | Method and system for video and film recommendation |
US13/174,904 Continuation US8338685B2 (en) | 2002-10-03 | 2011-07-01 | Method and system for video and film recommendation |
Publications (3)
Publication Number | Publication Date |
---|---|
US20130215330A1 US20130215330A1 (en) | 2013-08-22 |
US20140111693A9 true US20140111693A9 (en) | 2014-04-24 |
US8841534B2 US8841534B2 (en) | 2014-09-23 |
Family
ID=44993268
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/881,810 Active 2026-05-25 US8063295B2 (en) | 2002-10-03 | 2007-07-27 | Method and system for video and film recommendation |
US13/174,904 Expired - Lifetime US8338685B2 (en) | 2002-10-03 | 2011-07-01 | Method and system for video and film recommendation |
US13/726,874 Expired - Lifetime US8841534B2 (en) | 2002-10-03 | 2012-12-26 | Method and system for video and film recommendation |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/881,810 Active 2026-05-25 US8063295B2 (en) | 2002-10-03 | 2007-07-27 | Method and system for video and film recommendation |
US13/174,904 Expired - Lifetime US8338685B2 (en) | 2002-10-03 | 2011-07-01 | Method and system for video and film recommendation |
Country Status (1)
Country | Link |
---|---|
US (3) | US8063295B2 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10015548B1 (en) | 2016-12-29 | 2018-07-03 | Arris Enterprises Llc | Recommendation of segmented content |
WO2021242527A1 (en) * | 2020-05-26 | 2021-12-02 | Koenig Errol Avery | Automatic movie performance predictor |
Families Citing this family (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9172915B2 (en) * | 2004-08-04 | 2015-10-27 | Dizpersion Corporation | Method of operating a channel recommendation system |
US8053659B2 (en) * | 2002-10-03 | 2011-11-08 | Polyphonic Human Media Interface, S.L. | Music intelligence universe server |
US8063295B2 (en) * | 2002-10-03 | 2011-11-22 | Polyphonic Human Media Interface, S.L. | Method and system for video and film recommendation |
JP3969656B2 (en) * | 2003-05-12 | 2007-09-05 | ソニー株式会社 | Information processing apparatus and method, program recording medium, and program |
US7490775B2 (en) | 2004-12-30 | 2009-02-17 | Aol Llc, A Deleware Limited Liability Company | Intelligent identification of multimedia content for synchronization |
US7937422B1 (en) * | 2005-11-03 | 2011-05-03 | Aol Inc. | Digital asset hosting and distribution |
US8869037B2 (en) | 2006-06-22 | 2014-10-21 | Linkedin Corporation | Event visualization |
US7831928B1 (en) | 2006-06-22 | 2010-11-09 | Digg, Inc. | Content visualization |
US20090006368A1 (en) * | 2007-06-29 | 2009-01-01 | Microsoft Corporation | Automatic Video Recommendation |
US20090070198A1 (en) * | 2007-09-12 | 2009-03-12 | Sony Corporation | Studio farm |
US7921071B2 (en) * | 2007-11-16 | 2011-04-05 | Amazon Technologies, Inc. | Processes for improving the utility of personalized recommendations generated by a recommendation engine |
US8260778B2 (en) * | 2008-01-16 | 2012-09-04 | Kausik Ghatak | Mood based music recommendation method and system |
US20090216625A1 (en) * | 2008-02-27 | 2009-08-27 | Adam Jeffrey Erlebacher | Systems and Methods for Automated Identification and Evaluation of Brand Integration Opportunities in Scripted Entertainment |
US8079054B1 (en) * | 2008-04-14 | 2011-12-13 | Adobe Systems Incorporated | Location for secondary content based on data differential |
US20090326970A1 (en) * | 2008-06-30 | 2009-12-31 | Microsoft Corporation | Awarding users for discoveries of content based on future popularity in a social network |
US20100198685A1 (en) * | 2009-01-30 | 2010-08-05 | Microsoft Corporation | Predicting web advertisement click success by using head-to-head ratings |
CN102655607B (en) * | 2011-03-04 | 2016-03-30 | 富士通株式会社 | For adding up the method and apparatus of the audience ratings of target audience |
JP5732994B2 (en) * | 2011-04-19 | 2015-06-10 | ソニー株式会社 | Music searching apparatus and method, program, and recording medium |
TWI449410B (en) * | 2011-07-29 | 2014-08-11 | Nat Univ Chung Cheng | Personalized Sorting Method of Internet Audio and Video Data |
US9741061B2 (en) * | 2012-03-01 | 2017-08-22 | Sarah Nash Brechner | System and method for personal customization of digital content |
US20140201778A1 (en) * | 2013-01-15 | 2014-07-17 | Sap Ag | Method and system of interactive advertisement |
US9197926B2 (en) | 2013-06-05 | 2015-11-24 | International Business Machines Corporation | Location based determination of related content |
CN104156472B (en) * | 2014-08-25 | 2018-05-08 | 北京四达时代软件技术股份有限公司 | A kind of video recommendation method and system |
CN106375781B (en) * | 2015-07-23 | 2019-09-13 | 无锡天脉聚源传媒科技有限公司 | A kind of judgment method and device repeating video |
CN106375850B (en) * | 2015-07-23 | 2019-09-13 | 无锡天脉聚源传媒科技有限公司 | A kind of judgment method and device matching video |
CN105550334B (en) * | 2015-12-22 | 2019-01-22 | 海信集团有限公司 | A kind of video recommendation method and device |
CN105843876B (en) * | 2016-03-18 | 2020-07-14 | 阿里巴巴(中国)有限公司 | Quality evaluation method and device for multimedia resources |
CN106354867A (en) * | 2016-09-12 | 2017-01-25 | 传线网络科技(上海)有限公司 | Multimedia resource recommendation method and device |
US20180165696A1 (en) * | 2016-12-09 | 2018-06-14 | Authors, Inc. | Predictive Analytics Diagnostic System and Results on Market Viability and Audience Metrics for Scripted Media |
EP3340069A1 (en) * | 2016-12-23 | 2018-06-27 | ScriptBook NV | Automated characterization of scripted narratives |
US10963798B2 (en) | 2017-01-31 | 2021-03-30 | Accenture Global Solutions Limited | Multimedia content distribution and recommendation system |
CN107748759A (en) * | 2017-09-21 | 2018-03-02 | 广州华多网络科技有限公司 | Video pushing method and its equipment |
US10628486B2 (en) * | 2017-11-15 | 2020-04-21 | Google Llc | Partitioning videos |
CN109640176B (en) * | 2018-12-18 | 2021-01-22 | 北京字节跳动网络技术有限公司 | Method and apparatus for generating information |
US11704495B2 (en) * | 2019-05-20 | 2023-07-18 | Sony Group Corporation | Prediction of film success-quotient |
CN111339326A (en) * | 2020-02-19 | 2020-06-26 | 北京达佳互联信息技术有限公司 | Multimedia resource display method, multimedia resource providing method and multimedia resource providing device |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5616876A (en) * | 1995-04-19 | 1997-04-01 | Microsoft Corporation | System and methods for selecting music on the basis of subjective content |
US5792971A (en) * | 1995-09-29 | 1998-08-11 | Opcode Systems, Inc. | Method and system for editing digital audio information with music-like parameters |
GB2308262B (en) * | 1995-12-16 | 1999-08-04 | Paul Gordon Wilkins | Method for analysing the content of a video signal |
US5918223A (en) * | 1996-07-22 | 1999-06-29 | Muscle Fish | Method and article of manufacture for content-based analysis, storage, retrieval, and segmentation of audio information |
US6735249B1 (en) * | 1999-08-11 | 2004-05-11 | Nokia Corporation | Apparatus, and associated method, for forming a compressed motion vector field utilizing predictive motion coding |
JP3601371B2 (en) * | 1999-09-27 | 2004-12-15 | ヤマハ株式会社 | Waveform generation method and apparatus |
US20020002899A1 (en) * | 2000-03-22 | 2002-01-10 | Gjerdingen Robert O. | System for content based music searching |
US7532943B2 (en) * | 2001-08-21 | 2009-05-12 | Microsoft Corporation | System and methods for providing automatic classification of media entities according to sonic properties |
US7031980B2 (en) * | 2000-11-02 | 2006-04-18 | Hewlett-Packard Development Company, L.P. | Music similarity function based on signal analysis |
US20020147628A1 (en) * | 2001-02-16 | 2002-10-10 | Jeffrey Specter | Method and apparatus for generating recommendations for consumer preference items |
US20040039913A1 (en) * | 2002-04-16 | 2004-02-26 | Sky Kruse | Method and system for watermarking digital content and for introducing failure points into digital content |
US8063295B2 (en) * | 2002-10-03 | 2011-11-22 | Polyphonic Human Media Interface, S.L. | Method and system for video and film recommendation |
US7081579B2 (en) * | 2002-10-03 | 2006-07-25 | Polyphonic Human Media Interface, S.L. | Method and system for music recommendation |
US7460730B2 (en) * | 2005-08-04 | 2008-12-02 | Microsoft Corporation | Video registration and image sequence stitching |
US7966632B1 (en) * | 2007-12-12 | 2011-06-21 | Google Inc. | Visual presentation of video recommendations |
US20100191689A1 (en) * | 2009-01-27 | 2010-07-29 | Google Inc. | Video content analysis for automatic demographics recognition of users and videos |
US20110320380A1 (en) * | 2010-06-23 | 2011-12-29 | Microsoft Corporation | Video content recommendations |
-
2007
- 2007-07-27 US US11/881,810 patent/US8063295B2/en active Active
-
2011
- 2011-07-01 US US13/174,904 patent/US8338685B2/en not_active Expired - Lifetime
-
2012
- 2012-12-26 US US13/726,874 patent/US8841534B2/en not_active Expired - Lifetime
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10015548B1 (en) | 2016-12-29 | 2018-07-03 | Arris Enterprises Llc | Recommendation of segmented content |
WO2021242527A1 (en) * | 2020-05-26 | 2021-12-02 | Koenig Errol Avery | Automatic movie performance predictor |
US11769162B2 (en) * | 2020-05-26 | 2023-09-26 | Errol Avery Koenig | Automatic movie performance predictor |
Also Published As
Publication number | Publication date |
---|---|
US20080059287A1 (en) | 2008-03-06 |
US8063295B2 (en) | 2011-11-22 |
US8841534B2 (en) | 2014-09-23 |
US8338685B2 (en) | 2012-12-25 |
US20130215330A1 (en) | 2013-08-22 |
US20110270848A1 (en) | 2011-11-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8841534B2 (en) | Method and system for video and film recommendation | |
US20200058088A1 (en) | Method and system for determining content treatment | |
Hanjalic | Adaptive extraction of highlights from a sport video based on excitement modeling | |
JP6807389B2 (en) | Methods and equipment for immediate prediction of media content performance | |
US8369686B2 (en) | Intelligent overlay for video advertising | |
US20180082127A1 (en) | Video segmentation techniques | |
KR102112973B1 (en) | Estimating and displaying social interest in time-based media | |
CN111683209A (en) | Mixed-cut video generation method and device, electronic equipment and computer-readable storage medium | |
EP1067800A1 (en) | Signal processing method and video/voice processing device | |
CN107368965A (en) | A kind of script data processing method, device and apply its computer equipment | |
TW201340690A (en) | Video recommendation system and method thereof | |
CN108632640A (en) | The following insert region metadata of prediction | |
CN109640112B (en) | Video processing method, device, equipment and storage medium | |
KR20040054773A (en) | Identification and evaluation of audience exposure to logos in a broadcast event | |
KR102161080B1 (en) | Device, method and program of generating background music of video | |
CN114005077B (en) | Audience evaluation data driven silent product video creation auxiliary method and device | |
CN109902823A (en) | A kind of model training method and equipment based on generation confrontation network | |
Chu et al. | On broadcasted game video analysis: event detection, highlight detection, and highlight forecast | |
JP2000285243A (en) | Signal processing method and video sound processing device | |
Shin et al. | On-chart success dynamics of popular songs | |
CN111581435A (en) | Video cover image generation method and device, electronic equipment and storage medium | |
del Río et al. | Profiling players with engagement predictions | |
US7974440B2 (en) | Use of statistical data in estimating an appearing-object | |
JP7158902B2 (en) | Information processing device, information processing method, and information processing program | |
US20210225408A1 (en) | Content Pushing Method for Display Device, Pushing Device and Display Device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: POLYPHONIC HUMAN MEDIA INTERFACE, S.L., SPAIN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BONET, ANTONIO TRIAS;MARCOS, JESUS SANZ;LLOPIS, ANTONIO TRIAS;AND OTHERS;REEL/FRAME:035696/0387 Effective date: 20120409 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.) |
|
FEPP | Fee payment procedure |
Free format text: SURCHARGE FOR LATE PAYMENT, SMALL ENTITY (ORIGINAL EVENT CODE: M2554) |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551) Year of fee payment: 4 |
|
AS | Assignment |
Owner name: MUSIC INTELLIGENCE SOLUTIONS, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:POLYPHONIC HUMAN MEDIA INTERFACE, S.L.;REEL/FRAME:054481/0806 Effective date: 20100917 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 8 |