Methods, apparatus, systems and articles of manufacture to adjust content presented to an individual are disclosed. An example system includes a first modality sensor to measure a first response of an individual to first content during a first time frame and a second modality sensor to measure a second response of the individual to the first content during the first time frame. The first modality sensor is to measure a third response of the individual to first content during a second time frame, and the second modality sensor is to measure a fourth response of the individual to the first content during the second time frame. The example system also includes a mental classifier executing instructions to determine a first mental classification of the individual based on a first comparison of the first response to a first threshold and a second comparison of the second response to a second threshold. The mental classifier also is to determine a second mental classification of the individual based on a third comparison of the third response to a third threshold and a fourth comparison of the fourth response to a fourth threshold. In addition, the mental classified is to determine a mental state of the individual based on a degree of similarity between the first mental classification and the second mental classification. The example system also includes a content modifier to at least one of modify the first content to include second content or replace the first content with second content based on the mental state.
|
1. A system comprising:
a first sensor to measure a first response of an individual to first content during a first time frame and a second response of the individual to the first content during a second time frame, the first sensor including a pupil dilation sensor;
a second sensor to measure a third response of the individual to the first content during the first time frame and a fourth response of the individual to the first content during the second time frame; and
a processor to:
generate a cognitive load index based on data from the pupil dilation sensor, the cognitive load index representative of how much of an information processing capacity of the individual is being used;
determine a first mental classification of the individual based on (1) a first comparison of the first response to a first threshold and (2) a second comparison of the third response to a second threshold;
determine a second mental classification of the individual based on (1) a third comparison of the second response to a third threshold and (2) a fourth comparison of the fourth response to a fourth threshold;
determine a mental state of the individual based on a degree of similarity between the first mental classification and the second mental classification; and
at least one of modify the first content to include second content or replace the first content with the second content based on the mental state.
18. An apparatus comprising:
means for measuring a first response of an individual to first content during a first time frame and a second response of the individual to the first content during a second time frame, the means for measuring the first response and the second response including sensing pupil dilation;
means for measuring a third response of the individual to the first content during the first time frame and a fourth response of the individual to the first content during the second time frame;
means for analyzing responses, the response analyzing means to:
generate a cognitive load index based on pupil dilation data, the cognitive load index representative of how much of an information processing capacity of the individual is being used;
determine a first mental classification of the individual based on (1) a first comparison of the first response to a first threshold and (2) a second comparison of the third response to a second threshold; and
determine a second mental classification of the individual based on (1) a third comparison of the second response to a third threshold and (2) a fourth comparison of the fourth response to a fourth threshold;
means for determining a mental state of the individual based on a degree of similarity between the first mental classification and the second mental classification; and
means for at least one of modifying the first content to include second content or replacing the first content with the second content based on the mental state.
13. A tangible computer readable storage medium comprising instructions that, when executed, cause a machine to at least:
determine a first mental classification of an individual based on (1) a first comparison of a first response of the individual to first content to a first threshold and (2) a second comparison of a second response of the individual to the first content to a second threshold, the first response measured via a first modality during a first time frame, and the second response measured via a second modality during the first time frame, the second modality including sensing pupil dilation;
generate a cognitive load index based on pupil dilation data, the cognitive load index representative of how much of an information processing capacity of the individual is being used;
determine a second mental classification of the individual based on (1) a third comparison of a third response of the individual to the first content to a third threshold and (2) a fourth comparison of a fourth response of the individual to the first content to a fourth threshold, the third response measured via the first modality during a second time frame, and the fourth response measured during the second time frame;
determine a mental state of the individual based on a degree of similarity between the first mental classification and the second mental classification; and
at least one of modify the first content to include second content or replace the first content with the second content based on the mental state.
3. The system of
5. The system of
measure a fifth response of the individual to the first content during the first time frame; and
measure a sixth response of the individual to the first content during the second time frame;
the processor to:
determine the first mental classification of the individual based, in part, on a fifth comparison of the fifth response with a fifth threshold; and
determine the second mental classification of the individual based, in part, on a sixth comparison of the sixth response with a sixth threshold.
6. The system of
7. The system of
8. The system of
9. The system of
10. The system of
11. The system of
12. The system of
14. The tangible computer readable storage medium of
15. The tangible computer readable storage medium of
16. The tangible computer readable storage medium of
17. The tangible computer readable storage medium of
|
This patent arises from a continuation of U.S. patent application Ser. No. 15/155,543, titled “METHODS AND APPARATUS TO ADJUST CONTENT PRESENTED TO AN INDIVIDUAL,” and filed on May 16, 2016, which claims priority to U.S. Provisional Application No. 62/163,874, titled “MULTI-PHASIC EMOTION AND COGNITION CLASSIFIERS,” filed on May 19, 2015, and to U.S. Provisional Application No. 62/272,423, titled “METHODS AND APPARATUS TO ADJUST CONTENT PRESENTED TO AN INDIVIDUAL,” filed on Dec. 29, 2015. U.S. patent application Ser. No. 15/155,543; U.S. Provisional Application No. 62/163,874; and U.S. Provisional Application No. 62/272,423 are hereby incorporated herein by reference in their entireties. Priority to U.S. patent application Ser. No. 15/155,543; U.S. Provisional Application No. 62/163,874; and U.S. Provisional Application No. 62/272,423 is hereby claimed.
This disclosure relates generally to presenting content to an individual and, more particularly, to methods and apparatus to adjust content presented to an individual.
Individuals are exposed to multiple passive and interactive audio, visual, and audio-visual media content every day. The media content produces biologically based responses in the user that can be measured by one or more sensors. An individual's biological and/or physical response to an image can indicate emotional and cognitive responses. Personal logs and self-reporting of responses are often inaccurate and include biases due to human input. Additionally, personal logs and self-reporting rely on an accurate account by the individual of their own emotional and cognitive reaction.
The figures are not to scale. Wherever possible, the same reference numbers will be used throughout the drawing(s) and accompanying written description to refer to the same or like parts.
Many different kinds of media content, such as audio, visual, and audio-visual content are presented to individuals every day. The presentation of media content to the individual can result in a biologically based response in the individual, which can be used to determine a mental state of the user. For example, a user may be frustrated or confused by media content displayed by a website, including the website interface, a game, etc. Data related to a particular determined mental state of a user may be used, for example, in marketing applications. For example, if a user exhibits a biological response indicative of frustration caused by a website design, the user may be less likely spend time on the website and purchase products from the business owner of the website. Thus, market researchers could use the information about determined mental states of users to adapt the website to provide a more enjoyable experience to a user and potential consumer as indicated by positive mental or emotional states detected through the biological responses. Users having enjoyable and pleasant experiences are more likely to spend more time on a website and may be more likely to purchase products and/or services. In some examples, the media content (e.g., the website) may adapt automatically based on the determined mental state of the individual.
Example apparatus and methods described herein adjust content (e.g., media content such as commercial advertisements, websites, videos, Internet content, etc.) presented to an individual based on a determined mental state (e.g., frustration, concentration, boredom, etc.) of the individual. The mental state of the individual is determined based on measured responses (e.g., biometric responses such as heart rate, pupil dilation, etc.) to presented content. In some examples, the responses of the individual are measured using modality sensors (e.g., sensors to measure biometric responses such as heart rate sensors, pressure sensors, facial expression detectors or facial action coding (FAC), etc.) and compared to respective thresholds to determine response classifications (e.g., high or low heart rate; high, medium, or low pupil dilation; etc.).
In some examples, the thresholds may correspond to a baseline (e.g., a threshold amount above a baseline) associated with the response of the individual over a period of time during presentation of the content. For example, the measured response of an individual over time may be used to develop a baseline of biometric response activity. In such examples, an individual may have biometric data recorded while exposed to neutral or background media (as opposed to a targeted media or stimulus). The baseline determines the level of biometric activity of the person in, for example, an inactive or uninvolved state. This determination of the baseline may be different from individual to individual because, for example, some individuals have higher resting heart rates than others or different rates of respiration, etc. A threshold is set to indicate when, for example, an individual response indicative of excitement is high enough relative to the baseline to register as a positive response. Thus, the biometric data may be compared to the individualized threshold(s) to determine a high or low classification on a moment-by-moment basis for each subject, panelist, or user. Furthermore, in this example, the baseline is used as a reference for the threshold. If the comparison of the signal (the biometric data) to the threshold is constantly resulting in a “low” or “high” classification, for example, the baseline and, in some instance also the threshold, are adjusted because the response of the individual has changed over time and is not being accurately represented or detected by the constantly “low” or “high” classification. In other words, the fluctuations or changes in the individual's response may not be detected because the measured responses consistently remain below the threshold and, therefore, the threshold and baseline are to be changed to capture the fluctuations or changes.
In another example in which the baseline is adjusted based on the measured responses, a baseline is increased when a measured response is consistently higher than the threshold because this indicates that the response of the individual has changed. In this example, the baseline is increased after a period of time (e.g., thirty seconds) to account for the shift in the response of the individual (i.e., the fluctuations of the response of the individual are better measured if the baseline and threshold are increased). For example, if an individual has a resting heart rate of 70 beats per minute (bpm) (e.g., a baseline measurement), the threshold for a “high” heart rate classification may be 75 bpm. If the individual is frustrated, the heart rate may rise by, for example, approximately 10 bpm and, thus, be consistently classified as “high.” In this example, when the baseline is held at 70 bpm and the threshold is held at 75 bpm, the fluctuations of the heart rate between 77 bpm and 85 bpm are not detected because the entire signal between 75 bpm and 85 bpm is registered as “high.” Thus, changing the baseline and the threshold to more accurately represent the new baseline heart rate of the individual enables the fluctuations at the higher heart rate range to be classified as high or low with respect to the new baseline and threshold.
Additionally or alternatively, the baseline and/or threshold may be changed based on the task being performed by the individual. For example, a simple task (e.g., shopping for a toothbrush) may result in little or no change to the baseline and/or threshold, but a more complex task (e.g., configuring a car) may result in a larger change to the baseline and/or threshold. In some examples, the length of time the task is estimated to take may affect the adjustment of the baseline and/or threshold.
The response classifications measured during a first time frame may be combined to determine a mental classification (e.g., concentration, frustration, confusion, etc.). For example, the response classifications of high GSR, high pupil dilation, high negative FAC, and low positive FAC may indicate, when combined, that the individual is actively engaged and has a mental classification of frustration. In some examples, the mental classification is determined by, for example, combining response classifications corresponding to responses measured by different modality sensors (e.g., two or more different sensors). For example, a low GSR response, a low pupil dilation, a low negative FAC, and a low positive FAC may be combined to indicate a mental classification of low and/or no engagement (boredom). In some examples, a second mental classification is determined based on additional responses measured during a second time frame. A mental state is designated if consecutive mental classifications are similar (e.g., the mental state is designated based on a degree of similarity between consecutive mental classifications). In some examples, the mental state indicates a reaction of the individual to the presented content (e.g., frustration, confusion, etc.). In such examples, the content is adjusted or new content (e.g., second content), different from the content originally presented (e.g., first content), is presented to the individual to increase the positivity level of the reaction.
In the examples disclosed herein, the responses are measured during overlapping time frames to ensure that no peak measurements (e.g., a peak galvanic skin response (GSR) measurement) are missed. For example, the first time frame begins when the content is presented to the individual and the second time frame begins prior to an end of the first time frame. In some such examples, each time frame has a duration of four seconds and the second time frame begins one second after the first time frame begins. In conventional methods in which discrete windows are used, a peak that occurs at the end of one window and into the next may be lost.
In some examples, the response classifications are combined in time segments that include multiple overlapping time frames. For example, if each time frame has a duration of four seconds, the time segments may have a duration of two seconds. In such examples, the time segments include responses from up to four different time frames. In some examples, a mental classification is determined for each time segment based on the response classification corresponding to the respective time segments. In some such examples, the mental state is designated if the mental classifications for consecutive time segments (e.g., two or more consecutive segments) are similar. Alternatively, no mental state is designated if no consecutive mental classifications are similar.
Disclosed in some examples herein, are methods to adjust content presented to an individual. The example method includes measuring, via a first modality sensor, a first response of the individual to first content during a first time frame and determining a first response classification based on a first comparison of the first response and a first threshold. The example method also includes measuring, via a second modality sensor, a second response of the individual to the first content during the first time frame and determining a second response classification based on a second comparison of the second response to a second threshold. In addition, the example method includes determining a first mental classification of the individual based on combining the first response classification and the second response classification and determining a first baseline during the first time frame, at least one of the first threshold or second threshold based on the first baseline. The example method includes measuring, via the first modality sensor, a third response of the individual to first content during a second time frame and measuring, via the second modality sensor, a fourth response of the individual to the first content during the second time frame. In addition, the method includes adjusting the first baseline to a second baseline based on at least one of the third response or the fourth response in the second time frame, adjusting at least one of the first threshold to a third threshold or second threshold to a fourth threshold based on the second baseline, determining a third response classification based on a third comparison of the third response and the third threshold, and determining a fourth response classification based on a fourth comparison of the fourth response and the fourth threshold; determining a second mental classification of the individual based on combining the third response classification and the fourth response classification. Other aspects of the example method include determining a mental state of a user based on a degree of similarity between the first mental classification and the second mental classification, and at least one of modifying the first content to include second content or replacing the first content with second content based on the mental state.
In some examples, the first modality sensor includes a galvanic skin response sensor. Also, in some examples, the second modality sensor includes a pupil dilation sensor.
In some examples, the method also includes generating a cognitive load index based on data from the pupil dilation sensor. The cognitive load index is representative of how much of a maximum information processing capacity of the individual is being used.
In some example methods, the second time frame partially overlaps the first time frame.
In some examples, the method includes measuring, via a third modality sensor, a fifth response of the individual to the first content during the first time frame and determining a fifth response classification based on a fifth comparison of the fifth response and a fifth threshold, the first mental classification of the individual based on combining the first response classification and the second response classification further with the fifth response classification. The method also includes measuring, via the third modality sensor, a sixth response of the individual to the first content during the second time frame, and determining a sixth response classification based on a sixth comparison of the sixth response and the fifth threshold, the second mental classification of the individual based on combining the third response classification and the fourth response classification further with the sixth response classification.
In some examples, the third modality sensor includes a facial action coding sensor. Also, in some examples, the third modality sensor includes an eye tracking sensor.
Also, in some examples disclosed herein, the second content is to increase a positivity level of the mental state. In addition, in some examples, the second content is to at least one of induce a purchase or increase a total spend amount on a purchase.
Also disclosed herein are example systems including, a system that includes a first modality sensor, a second modality sensor, and a processor. In the example system, the processor is to measure, via the first modality sensor, a first response of an individual to first content during a first time frame and determine a first response classification based on a first comparison of the first response and a first threshold. The example processor also is to measure, via the second modality sensor, a second response of the individual to the first content during the first time frame, and determine a second response classification based on a second comparison of the second response to a second threshold. The example system also uses the processor to determine a first baseline during the first time frame, at least one of the first threshold or second threshold based on the first baseline and determine a first mental classification of the individual based on combining the first response classification and the second response classification. In addition, the processor is to measure, via the first modality sensor, a third response of the individual to first content during a second time frame, measure, via the second modality sensor, a fourth response of the individual to the first content during the second time frame, adjust the first baseline to a second baseline based on at least one of the third response or the fourth response in the second time frame, and adjust at least one of the first threshold to a third threshold or the second threshold to a fourth threshold based on the second baseline. Other determinations are also made by the example processor including, for examples, determining a third response classification based on a third comparison of the third response and the third threshold, determining a fourth response classification based on a fourth comparison of the fourth response to the fourth threshold, determining a second mental classification of the individual based on combining the third response classification and the fourth response classification, and determining a mental state of a user based on a degree of similarity between the first mental classification and the second mental classification. In addition, the example system uses the processor to at least one of modify the first content to include second content or replace the first content with second content based on the mental state.
Also disclosed herein are tangible computer readable storage media comprising instructions that, when executed, causes a processor of a content presentation device to at least measure, via a first modality sensor, a first response of an individual to first content during a first time frame, determine a first response classification based on a first comparison of the first response and a first threshold, measure, via a second modality sensor, a second response of the individual to the first content during the first time frame, and determine a second response classification based on a second comparison of the second response to a second threshold. In these examples, the instructions further cause the machine to determine a first baseline during the first time frame, at least one of the first threshold or second threshold based on the first baseline, and determine a first mental classification of the individual based on combining the first response classification and the second response classification. In addition, executing the instructions also causes the machine to measure, via the first modality sensor, a third response of the individual to first content during a second time frame, measure, via the second modality sensor, a fourth response of the individual to the first content during the second time frame, adjust the first baseline to a second baseline based on the third response or the fourth response in the second time frame, and adjust at least one of the first threshold to a third threshold or the second threshold to a fourth threshold based on the second baselines. Furthermore, in this example, the machine is caused by the executed instructions to determine a third response classification based on a third comparison of the third response and the third threshold, determine a fourth response classification based on a fourth comparison of the fourth response to the fourth threshold, determine a second mental classification of the individual based on combining the third response classification and the fourth response classification, and determine a mental state of a user based on a degree of similarity between the first mental classification and the second mental classification. Also, in this example, the machine is to at least one of modify the first content to include second content or replace the first content with second content based on the mental state.
Further disclosed herein are systems such as an example system that includes a first modality sensor to measure a first response of an individual to first content during a first time frame. The example system also includes a second modality sensor to measure a second response of the individual to the first content during the first time frame. The first modality sensor is to measure a third response of the individual to first content during a second time frame, and the second modality sensor is to measure a fourth response of the individual to the first content during the second time frame. The example system includes a response classifier to determine a first response classification based on a first comparison of the first response and a first threshold and determine a second response classification based on a second comparison of the second response to a second threshold. In addition, the example system includes a baseline generator to determine a first baseline during the first time frame, at least one of the first threshold or second threshold based on the first baseline and adjust the first baseline to a second baseline based on at least one of the third response or the fourth response in the second time frame. The baseline generator also is to adjust at least one of the first threshold to a third threshold or the second threshold to a fourth threshold based on the second baseline. In addition, the response classifier is to further determine a third response classification based on a third comparison of the third response and the third threshold and determine a fourth response classification based on a fourth comparison of the fourth response to the fourth threshold. The example system also includes a mental classifier to determine a first mental classification of the individual based on combining the first response classification and the second response classification and determine a second mental classification of the individual based on combining the third response classification and the fourth response classification. The mental classifier also is to determine a mental state of a user based on a degree of similarity between the first mental classification and the second mental classification. The example system also includes a content modifier to at least one of modify the first content to include second content or replace the first content with second content based on the mental state.
In some examples, the system further includes a third modality sensor to measure a fifth response of the individual to the first content during the first time frame and measure a sixth response of the individual to the first content during the second time frame. Also, in such example systems, the response classifier is to determine a fifth response classification based on a fifth comparison of the fifth response and a fifth threshold, and the mental classifier is to base the first mental classification of the individual on combining the first response classification and the second response classification further with the fifth response classification. In addition, the response classifier is to determine a sixth response classification based on a sixth comparison of the sixth response and the fifth threshold, and the mental classifier is to base the second mental classification of the individual on combining the third response classification and the fourth response classification further with the sixth response classification.
Turning now to the figures,
In the illustrated example system 100, the content presentation device 102 is a desktop computer. In other examples, the content presentation device 102 may be any device suitable to present media content to an individual 104, such as a television, a radio, an Internet-streamed audio source, a workstation, a kiosk, a laptop computer, a tablet computer, an e-reader, a smartphone, etc. The example content presentation device 102 presents media content to the individual 104 that includes audio, visual, and/or audio-visual content. In some examples, the content is advertisement(s) and/or entertainment. Also, in some examples, the content is interactive, such as a video game, live interaction, or an Internet experience (e.g., a website). The example content presentation device 102 includes a display 106 and/or an audio output 108 (e.g., speakers, a headset) to present the media content to the individual 104. In some examples, the display 106 and/or the audio output 108 enables the individual 104 to interact with the content presentation device 102. The content presentation device 102 includes one or more of a keyboard 110, a mouse 112, a touchscreen, a microphone, a remote control, etc. to facilitate an interaction between the individual 104 and the content presentation device 102.
In some examples, the content presentation device 102 is used to measure and/or record self-reported responses, such as responses to computer generated surveys, text input, and/or audio responses. Self-reported measurements include, but are not limited to, survey responses to items such as perception of the experience, perception of the usability or likeability of the experience, level of personal relevance to user, attitude toward content or advertising embedded in the content, intent to purchase a product, game, or service, and changes in responses from before and after testing.
In some examples, the input devices (e.g. the mouse 112 and/or keyboard 110 and/or other input devices) include sensors (e.g., biometric sensors, pressure sensors) to measure a response of the individual 104. For example, interactive content is presented to the individual 104 according to a predefined program or sequence biometric response data is recorded and synchronized or mapped to the content presentation to indicate what biological response the individual 104 had to what portion of the presentation.
As shown in
In some examples, the measured response data is linked and/or synchronized with the content presentation using time stamps and/or event windows. For example, the presentation is divided into event windows based on specific tasks or activities that are included in the interactive content presented to the individual 104, and the measured response data is associated with the event windows based on the tasks or activities. In some examples, each task or activity has one or more event windows associated with the task or activity. Additionally, each event window can be the same or a different duration of time as the other event windows.
The one or more modality sensors 114, 116, 118 and/or the content presentation device 102 are in communication with a server 120 via a wired or wireless network 122. In some examples, the sensors 114, 116, 118 are coupled to the network 122 via the content presentation device 102. In some examples, the network 122 uses communication technologies such as RS-232, Ethernet, Wi-Fi, Bluetooth or ZigBee. The server 120 additionally is in communication with a results analyzing device 124, which is illustrated as a desktop computer but other devices may be used as noted herein. Additionally or alternatively, more than one communication technology is used at the same time, including wired components (e.g., Ethernet, digital cable, etc.) and wireless components (Wi-Fi, WiMAX, Bluetooth, etc.) to connect the sensors 114, 116, 118 and/or other computer system components to the server 120.
Alternatively or additionally, the results analyzing device 124 includes any device suitable to analyze data collected by the content presentation device 102 and/or the modality sensors 114, 116, 118, including a workstation, a kiosk, a laptop computer, a tablet computer, and a smartphone. In some examples, the results analyzing device 124 receives input from a reviewer 126 related to the results corresponding to the individual 104. The results are transmitted to, for example, the server 120, a second server, and/or an additional computing device. Alternatively, the results include a generated report 128 (e.g., a hard or a soft copy) distributed to, for example, a client. In some examples, the results analyzing device 124 is integrated with the content presentation device 102 to determine moment-to-moment, event-to-event or total level of emotion and cognition classifiers and provides the results to the server 120. Analyzing the results using the results analyzing device 124 prior to transmitting the results the server 120 decreases the amount of data transferred, resulting in faster data processing and lower transmission bandwidth requirements to increase the operating efficiency of the system.
As used herein, the phrase “in communication,” including variances thereof, encompasses direct communication and/or indirect communication through one or more intermediary components and does not require direct physical (e.g., wired) communication and/or constant communication, but rather additionally includes selective communication at periodic or aperiodic intervals, as well as one-time events.
The example mental state determination module 202 is communicatively coupled to a plurality of modality sensors including, for example, a first modality sensor 204, a second modality sensor 206, and an Nth modality sensor 208. In some examples, the first modality sensor 204, the second modality sensor 206, and the Nth modality sensor 208 correspond to any of the camera 114, the biometric sensing clothing 116, and the biometric bracelet 118 of
The example mental state determination module 202 includes an input/output interface (I/O interface) 210. The example I/O interface 210 is operatively coupled to the first, second, and Nth modality sensors 204, 206, and 208 to communicate a response of the individual 104 to the example mental state determination module 202. Additionally or alternatively, the I/O interface 210 is operatively coupled to any of the display 106, the audio output 108, the keyboard 110, the mouse 112, a touchscreen, a microphone, or any other device capable of providing an output to the individual 104 and/or providing an input to the content presentation device 102. Additionally, the I/O interface 210 is in communication with the server 120 of
In the illustrated example, the mental state determination module 202 includes storage 212 (e.g., a mass storage device) to store response data corresponding to the individual 104, content to be presented to the individual 104, and/or instructions for processing the response data. The example storage 212 is in communication with the I/O interface 210 to send and receive response data and/or media content to and/or from, for example, the first modality sensor 204, the second modality sensor 206, the Nth modality sensor 208, and/or the server 120. Alternatively or additionally, in some examples, the example storage 212 is in direct communication with one or more of the first modality sensor 204, the second modality sensor 206, the Nth modality sensor 208, and the server 120.
The example mental state determination module 202 includes a response classifier 214 to determine a response classification of measured response data received from one or more of the first, second, and Nth modality sensors 204, 206, and 208. The example response classifier 214 determines the response classification (e.g., high or low heart rate; high or low GSR; high, medium, or low pupil dilation; etc.) by, for example, comparing the measured response to a threshold. In some examples, each measured response corresponding to one of the modality sensors 204, 20, 208 is compared to a different threshold (e.g., a respective threshold) based on, for example, different biological characteristics of the signals and responses for the respective modality. The threshold is determined based on, for example, an average value of the measured response during an initial time period (e.g., a baseline). In some examples, the response classifier 214 determines which responses are most likely to be relevant to the mental state of the individual 104 from the available response measurements. In some examples, the selection of responses relevant to the mental state is confirmed using a research methodology. For example, a hypothesis is generated, a study is created, participants are recruited, data is collected and analyzed, and a conclusion is drawn. Additionally or alternatively, in some examples, a statistical model of the contributions of each of the responses is created to select the responses with the greatest relevance to the mental state of the individual 104. The example statistical model may be used to classify responses and/or determine a range for characterization of responses using an assumed statistical probability density. In some examples, the statistical model may form the bases for classification barriers and determine if one mental state is more likely than another.
The example mental state determination module 202 includes an example bin classifier 216. In some examples, the bin classifier 216 creates one or more bins in which to place each response based on the respective thresholds and/or a baseline. For example, each response (e.g., the measurement from each modality sensor 204, 206, 208) is sorted into a bin (e.g., a high bin, a low bin) based on the comparison to the threshold, and the one or more bins may be created based on the baseline. In some examples, the bin classifier 216 determines binning criteria based on the response measurement and/or sensor measuring the response being binned. For example, for GSR binning, the baseline (e.g., the binning criteria) is determined by calculating a mean GSR for a portion of the response measurement. Example GSR bins include a high bin (e.g., 50% increase above the mean GSR) and a low bin (e.g., 25% increase above the mean GSR). Additionally or alternatively, binning criteria for an HR response measurement is determined by a change in absolute beats-per-minute (bpm) within a two-second window. Example HR bins include a high bin (e.g., increase of 12-15 bpm) and a low bin (e.g., decrease of 12-15 bpm).
Additionally or alternatively, the bin classifier 216 divides some responses, such as facial responses, into positive and negative categories prior to sorting the response into a high or low bin. In such examples, a database of facial responses is created from participants during testing to determine relative baseline(s) for positive and negative expressions. Example facial response bins include a high bin (e.g., one standard deviation in probability of coding an expression as positive/negative above the standard deviation and the opposite response is low (i.e., to code high positive FAC, negative FAC response must be low)) and a low bin (e.g., one standard deviation decrease in probability, below data baseline, of coding an expression as positive/negative).
In some examples, the bin classifier 216 creates an intermediate bin for response measurements, such as pupil dilation. For example, pupil dilation binning includes determining the baseline based on a function of change from mean pupil dilation during some portion of the measured response. In some examples, pupil dilation binning includes a medium bin to capture times when the individual 104 is cognitively engaged, but not necessarily heavily concentrating, frustrated, or bored. Example pupil dilation bins include a high bin (e.g., mean dilation plus at least one half standard deviation), a medium bin (e.g., within one half standard deviation of the mean), and a low bin (e.g., mean dilation minus at least one half standard deviation). In some examples, the response classifier 214 uses the bins to determine the response classification of a response and/or places the responses in bins based on the comparison of the response to the threshold. In some such examples, the response classifier 214 and the bin classifier 216 work cooperatively to place responses in an appropriate bin based on the response classification and/or the comparison of the response to the threshold or baseline.
Typically, all responses are weighted equally when determining the mental state of the individual 104 (e.g., if three responses were measured, each response contributes to the mental state 33%). In some examples, the responses are weighted by adjusting the contribution of one or more responses to the overall results to be more or less than the contribution of other responses. For example, if one of the modality sensors 204, 206, 208 is not measuring data for all or part of the content presentation, the weighting of the contribution of each response is adjusted (e.g., if only two responses are measured at a given time, each response contributes to the mental state 50%). As more response data is collected and/or as the reaction to the presented content changes, the weights of the responses contributions can be adjusted to improve accuracy of the results (e.g., the response classification, a mental classification, the determined mental state).
The example mental state determination module 202 includes an example baseline generator 218. In some examples, the baseline generator 218 determines an initial baseline using a baselining procedure. For example, response measurements are not binned (e.g., classified) for an initial time period such as, for example, thirty seconds. The length of the initial time period may vary based on a task being performed by the individual. In some examples, neutral content is presented to the individual 104 during the initial time period. During the initial time period, a representative value (e.g., a mean value, a standard deviation, etc.) is determined by the baseline generator 218, for example, for the responses related to each sensor and are used as the initial baseline. In such examples, responses are compared to the initial baseline and/or a respective one of the thresholds. Additionally or alternatively, after the initial time period, the baseline is periodically adjusted based on the measured responses. For example, the baseline generator 218 re-evaluates the baseline for each baseline time period (e.g., thirty seconds) and adjusts the baseline based on the response. Alternatively or additionally, the baseline time period is the same as the first time frame. For example, if a user's GSR is above the mean (e.g., sorted in the high bin) for a period of time and then drops below the mean, the baseline determiner 218 adjusts the baseline in response to the drop in the GSR measurement to establish a new baseline. Thus, the baseline determiner 218 automatically adjusts the baseline corresponding to each response measurement in response to the occurrence of relevant events.
Automatically adjusting the baseline as the response of the individual changes and/or develops increases the accuracy of the determined mental state. For example, determining the mental state based on a single and/or constant baseline may not detect fluctuations or changes in the response of the individual (e.g., drop in heart rate after a period of higher heart rate) because the response (e.g., heart rate) may still be higher than the initial baseline and, thus, classified as high. The failure to detect these fluctuations or changes may result in an incorrect classification of the response.
In addition, there are many advantages to adjusting the baseline. For example, if an individual is experiencing frustration with a website, which is detected based on the individual's GSR being above a threshold relative to the baseline, and there may be a modification of the content to alter the mental state of the user to a more enjoyable experience. The modification of the content may begin to work to bring the individual to a less frustrated state. However, at the initial stages of the change, the individual's GSR may remain above the threshold relative to the baseline, though the individual's mental state is changing in accordance with the goals of the modified content. However, these changes may go undetected based on the level of GSR compared to the threshold relative to the baseline. Whereas, an adjusted baseline would change the threshold trigger, and enable detection of the GSR (in this example) moving across the threshold and provide indication that the content modification is effective. Therefore, the content modification can continue to bring the individual into the desired mental state. In addition, where the baseline is moved and the threshold has not been triggered though content has been modified, the operator or website owner would know that the content modification did not work (or did not work fast enough) or that a secondary baseline adjustment may be needed for a finer detection of biometric responses and/or mental classification and state changes.
In some examples, a running window implementation is used. In some such examples, the running windows include overlapping time windows (e.g., four-second windows). In some examples, the responses are measured using overlapping windows to avoid inaccuracies and/or missed events in the collected data. For example, GSR measurements typically peak at approximately four to five seconds, which can be missed or misinterpreted using non-overlapping time windows to measure GSR responses. In the example disclosed herein, the time windows each have a duration of four seconds and begin in one-second increments. In other examples, any other suitable or desired time duration(s) and/or increment(s) may be used. In some examples, the response is binned and/or a response classification is determined for each of the time windows.
The example mental state determination module 202 also includes an example mental classifier 220. The example mental classifier 220 determines mental classifications related to the measured responses (e.g., raw data from the modality sensors) and/or the response classifications (e.g., response data that is classified based on a threshold). In some examples, the mental classifications are determined by combining response classifications (e.g., high heart rate and low GSR) corresponding to one or more modality sensors 204, 206, and 208. In the illustrated example, the response classifications are combined in time segments shorter than the time windows (e.g., two seconds) and include the response classifications determined for each time window related to the time segment.
In some examples, the mental classifier 220 uses a mental classification grid, such as the example mental classification grid 300 in
In some examples, one of the axes used to determine a mental classification of an individual is cognitive load. The cognitive load axis 302 refines the classifications and/or the emotional valence 304 and the emotional arousal 306. The cognitive load is determined based on biological measures, such as measurements of pupil dilation. A cognitive load index represents the maximum amount of information the individual 104 can process at a given time. Cognitive load 302 is quantified based on the index to represent how much information the individual 104 is processing at a given time. Including cognitive load 302 as an axis in the example mental classification grid 300 provides significant functionality. Each individual 104 is determined to have a maximum information processing capacity. Comparing the cognitive load of the individual 104 during a period of time to the cognitive load index provides information related to the mental state of the individual 104. For example, if the individual 104 exhibits a high cognitive load index and a low emotional index, the determined mental state is concentration. In some examples, the emotional index is based on the emotional valence 304 and/or the emotional arousal 306. Thus, the use of cognitive load 302 allows the example mental state determination module 202 and/or the mental classifier 220 to distinguish between mental states such as frustration, confusion, and concentration.
Additionally or alternatively, the mental classifier 220 uses a mental classification matrix, such as the example mental classification matrix 400 of
In the illustrated example, response classifications 402 that can be combined by the mental classifier 220 to create other mental classifications 404 (e.g., active engagement (frustration), active concentration (flow state), passive concentration, low/no engagement (boredom), etc.) include one or more response classifications 402 different than the response classifications 402 combined to provide an active engagement (positive) mental classification 404. The example response classifications 402 in the example mental classifications matrix 400 include response classifications 402 corresponding to measurements (e.g., GSR, pupil dilation, FAC, etc.) using the sensors (e.g., the first modality sensor 204, the second modality sensor 206, the Nth modality sensor 208, the camera 114, the biometric sensing clothing 116, the biometric sensing bracelet 118, etc.). In other examples, the example response classifications 402 correspond to additional and/or alternative sensor measurements (e.g., HR, EEG, pupil tracking, etc.). The example mental classification matrix 400 illustrated in
In some examples, the example response classifications corresponding to time windows 502-510 are combined in two-second time segments 514-520 to determine a mental classification for each of the time segments 514-520. In some examples, the response measurements from all modality sensors 204, 206, 208 are combined during the same time segment (e.g., time segment 514) to determine a mental classification corresponding to the time segment 514. Additionally or alternatively, the time windows 502-510 falling within each of the time segments 514-520 are combined to determine the mental classification for a time segment (e.g., time segment 516). For example, the mental classification corresponding to time segment 516 is determined by combining response classifications from all time windows (e.g., the first four time windows 502-508) that at least partially overlap and/or fall within the time segment 516. In the illustrated example of
In some examples, the mental classifier 220 (
After the mental state determination module 202 determines the mental state of the individual 104, an example content modifier 222 determines whether to modify the content. In some examples, the content modifier 222 edits the content presented to the individual 104 based on the determined mental state of the individual 104. For example, if the mental state indicates that the individual 104 is frustrated, the content modifier 222 presents new content (e.g., second content) to the individual and/or adjusts the content to increase a positivity of the mental state. In some examples, the new content includes a coupon and/or a video (e.g., a tutorial video). Alternatively, the new content is a coupon, a free gift, a suggestion, etc. In some such examples, the new content induces a purchase of a product. In other examples, the new content increases a total amount spent on a purchase.
In some examples, the new content is provided as an output 224, such as content displayed via the display 106, printed content, audio content, or any other type of media content presentable to the individual 104. In some examples, the output 224 includes response data (e.g., response classifications, mental classifications, and the determined mental state) transmitted to the server 120. In some examples, the output 224 is in communication with the server 120 and/or the content presentation device 102 via the I/O interface 210 of the mental state determination module 202.
While an example manner of implementing the system 100 of
A flowchart representative of example machine readable instructions for implementing the systems 100, 200 of
As mentioned above, the example processes of
The example instructions 600 include measuring biometric and/or neurophysiological responses to content (block 604). For example, one or more of the example sensors (e.g., the camera 114, the biometric sensing clothing 116, the biometric sensing bracelet 118, and/or the first, second, and Nth modality sensors 204, 206, 208) measures the response of the individual 104 to the content, which may include a biometric response, a neurophysiological response, and/or a behavioral response. In some examples, a first response to first content is measured by the first modality sensor 204 during a first time frame and a second response to first content is measured by the second modality sensor 206 during the first time frame. Additionally, in some examples, the first modality sensor 204 measures a third response of the individual to first content during a second time frame and the second modality sensor 206 measures a fourth response of the individual to the first content during a second time frame. In the example implementation, sensors collected data related to GSR, FAC, eye tracking and pupil dilation for the individuals exposed to the Adagio tea configurator. The responses are averaged over a time period (block 606) using, for example, the response classifier 214 to average the responses (e.g., heart rate, pupil dilation, GSR, etc.) of each sensor 204, 206, 208 over a first time frame.
For example, the responses of each of the individuals were monitored for the duration of the exposure to various images, text, displays, and/or other options presented via the Adagio tea configurator. The duration included multiple time frames. The responses of the individual were averaged for each of the time frames. For example, a sensor detected FAC may detect a furrowed brow and a tight lip during a time window. These detected features could change over the duration of the time window, and an average across the window in determined.
The example instructions 600 include comparing the average of each response to a threshold (block 608). For example, the response classifier 214 compares the average response values to respective thresholds corresponding to the sensors 204, 206, 208 to classify each response (e.g., high heart rate, low heart rate, low GSR, etc.). In some examples, a first response is compared to a first threshold and a second response is compared to a second threshold. In some examples, the first threshold or the second threshold is based on a first baseline determined for a first time frame. In some such examples, the first baseline is adjusted to a second baseline based on the third response and/or the fourth response. In some examples, the first threshold is adjusted to a third threshold and/or the second threshold is adjusted to a fourth threshold. Additionally, in some examples, a third response is compared to a third threshold and a fourth response is compared to a fourth threshold.
For example, in the Adagio tea implementation described above, the individuals using the configurator may be presented with 30 images and 4 video clips to establish a baseline and/or threshold to which the responses measured during the exposure to the Adagio tea configurator are compared. The individuals' responses from one or more sensors (including for example GSR, FAC, and/or pupillary dilation) are compared to the thresholds over time. For example, the average FAC response (based on the furrowed brow and tight lip mentioned above) is compared to a threshold value related to features detected via FACs sensors to determine a relative level of, for example, furrowed brows and tight lips. In addition, the system operating the configurator may, at times, determine that a baseline and/or threshold may have to be adjusted, as described above and below with respect to
Each response is placed in a bin based on the comparison (block 610). In addition, the example bin classifier 216 places each response into a respective bin (e.g., high, medium, low) based on the comparison to the threshold. The response classifier 214 determines a response classification (e.g., high GSR, low heart rate, low pupil dilation, etc.) for each response based on the comparison of the average response value(s) to the threshold(s) and/or the bin(s) in which each response is placed.
For example, the responses of the individuals using the Adagio tea configurator are placed in high or low GSR bin; high, medium, or low FAC bins; and high, medium, or low pupillary dilation bins and/or other bins relative to the biometric responses detected from the individual while presented with the configurator. For example, a low FAC bin may include negative responses such as, for example, those identifiable by furrowed brows and tight lips.
The example instructions 600 include assigning a weight to each response (block 612). For example, the response classifier 214 assigns a weight to each response corresponding to the amount each response contributes to the determined mental state of the individual 104. For example, if three response classifications are available, each response classification may be weighted as 33%. In another example implementation, each of the responses are given a weight corresponding to the contribution of each response to the determined mental state. For example, pupil dilation data can be adversely affected due to changing lighting when viewing dynamic media and, thus, the weighting for pupil dilation may be less than the weighting for each of GSR, FAC, and/or eye tracking, etc.
The example instructions 600 also determine first response classification for each response over a first time period (block 614) and second response classification of each response over a second time period (block 616). For example, the response classifier 214 of
In the Adagio tea example implementation, responses are measured over numerous time periods, and response classifications are determined for each time period by comparing the responses to relevant thresholds to detect fluctuations or changes in the response of the individual. For example, based on the comparison of the FACs data (e.g., the furrowed brows and tight lips) to the thresholds and/or bin data, it may be determined that the responses during the measured time periods if low or negative.
In this example, the instructions 600 combine classifications of each response to create a mental classification (block 618). For example, the mental classifier 220 determines a mental classification (e.g., frustration, confusion, boredom, etc.) for a time segment by combining response classifications from the first time period and/or the second time period. In some examples, a first mental classification of the individual is determined based on combining the first response classification and the second response classification. Additionally, in some examples, a second mental classification is determined based on combining the third response classification and the fourth response classification. In the example implementation, an individual with response classifications including high GSR, high pupil dilation, high negative FAC, and low positive FAC was determined to have a mental classification of frustrated.
In addition, the example instructions 600 are used to identify consecutive similar mental classifications (block 620). For example, the mental state determination module 202 of
Based on the determined mental state, the instructions are further executed to determine whether the mental state should be adjusted (block 624) to, for example, increase the positivity, decrease negativity, increase intensity, heighten a concentration and/or otherwise make a change to the mental state. For example, any individual operating the Adagio tea configurator who is experiencing frustration would be identified as a candidate in need of a mental state adjustment.
If it is determined that the mental state is to be adjusted, the instructions 600 include modifying the content (block 626). For example, the content modifier 222 of
After content has been modified (block 626), the control returns to block 602 and the modified content is presented to the individual and the example instructions 600 continue with the data gathering and analysis disclosed above. However, if it is determined that the mental state does not need to be adjusted (block 624), the example instructions are also used to determine whether or not to continue monitoring the individual (block 628). For example, if the mental state determination module 202 decides to continue monitoring the individual 104, control returns to block 604 and monitoring continues. However, if it is determined that monitoring is not to continue (block 628), monitoring ceases and the process 600 is complete (block 630). In the example implementation, the system continued to monitor the individual through the entire experience with the tea configurator and the process was designated as complete after the individual completed the checkout process. The results of the particular example implementation showed a positive impact on the amount of money spent by individuals receiving new or modified content based on the determined mental state (i.e., the individuals in the minimal and maximal adaptation groups spent more than the individuals in the random and control groups), with the maximal adaptation spending a slightly higher amount.
The example instructions 700 include monitoring biometric responses (block 704). For example, one or more of the example sensors (e.g., the camera 114, the biometric sensing clothing 116, the biometric sensing bracelet 118, and/or the first, second, and Nth modality sensors 204, 206, 208) measures the response of the individual 104 to the content, which may include a biometric response, a neurophysiological response, and/or a behavioral response.
A baseline response is determined (block 706), using for example the baseline generator 218 if
The example instructions 700 also include presenting the individual with content that includes stimulus and/or target material (block 710). For example, the presentation device 102 of
The example instructions 700 are executed to determine if a period of time elapses in which the threshold has not been triggered (block 716). Triggering the threshold may mean, for example, meeting a threshold, crossing or exceeding a threshold, falling without or outside of a threshold range, etc. For example, the example system 200 analyzes the monitored biometric responses over time and continually compares the responses to the threshold, which was established in relation to the individual's baseline. In some examples, the individual's heart rate may be monitored to determine if the heart rate moves higher than 10 bmp over a baseline heart rate, or if the heart exceeds an absolute value change, or if the heart rate passes 80 bmp, and/or any other suitable or desired metrics. If the threshold has been triggered within the established time period, the individual's responses are continued to be monitored (block 712).
However, the threshold has not been triggered within the set time period (block 716), the example instructions 700 are executed to determine if the content has been modified (block 718) and, if not, content is modified (block 720), presented to the individual (block 710), and monitoring continues (block 712). The content may be modified, for example, in accordance with the example systems 100, 200 of
If the content has been modified (block 718), the example instructions 700 adjust the individual's baseline and/or reestablishes the threshold relative to the baseline (block 722) using, for example, the example system 200 including the baseline generator 218 as disclosed above. For example, if the individual is experiencing frustration and adjustments are made the content to change the individual's biometric responses (and, thus, mental state), but the system continues to read the individual's response as frustrated, there may be an indication that the adjusted content is not sufficient to change the individual's response to a more positive response.
Additionally or alternatively, this may be an indication that the threshold set, with respect to the baseline, is insufficient to detect a change in response. For example, if an individual has a baseline heart rate of 75 bpm and a threshold is set to a 5 bpm change (plus or minus), the change may determine when a person is experiencing boredom or frustration. If the person has a change of 10 bpm, the threshold has been crossed. There may be a desire to present modified content to the individual to change the response back to a positive response as indicated by the heart rate crossing the threshold back toward the 75 bpm baseline. New content may be provided, which alters the individual's 10 bmp change to 8 bmp but does not cross the 5 bmp threshold. This change indicates that the altered content is effective in changing the responses to the desired response. However, the change is not enough to trigger threshold and, therefore, goes undetected. This may cause the content provider to abandon the content modification, may ultimately be effective for changing the response to the desired response, or cause the content provider to overcompensate resulting in further and unnecessary modification of the content. With the present example systems and processes, the threshold may be adjusted to, for example a change of 2 bmp, for a more fine detection of responses changes. This advancement provides the content with enhanced detection capabilities and advanced knowledge of the effectiveness in content modification in causes an individual to have a desired response.
In addition, there are examples in which the baseline itself is to be changed. For example, if a content provider would like to know when an individual has a change in heart rate when the measurements already exceed the threshold and/or continuously exceed the threshold for a period of time, the baseline is adjusted to reflect the changes in heart rate at levels higher than the previous baseline. In some examples, the baseline may be changed based on task. A low stress task (e.g., buying a toothbrush) may have a lower baseline for heart rate than a high stress task (e.g., configuring a car) because the individual is more likely have a higher heart rate while performing the higher stress task and would likely continuously exceed a baseline for a lower stress task unless the baseline is adjusted according to the task. In some examples, the length of time for which the task is performed affects the change in the baseline.
The processor platform 800 of the illustrated example includes a processor 812. The processor 812 of the illustrated example is hardware. For example, the processor 812 can be implemented by one or more integrated circuits, logic circuits, microprocessors or controllers from any desired family or manufacturer.
The processor 812 of the illustrated example includes a local memory 813 (e.g., a cache). The processor 812 of the illustrated example is in communication with a main memory including a volatile memory 814 and a non-volatile memory 816 via a bus 818. The volatile memory 814 may be implemented by Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS Dynamic Random Access Memory (RDRAM) and/or any other type of random access memory device. The non-volatile memory 816 may be implemented by flash memory and/or any other desired type of memory device. Access to the main memory 814, 816 is controlled by a memory controller.
The processor platform 800 of the illustrated example also includes an interface circuit 820. The interface circuit 820 may be implemented by any type of interface standard, such as an Ethernet interface, a universal serial bus (USB), and/or a PCI express interface.
In the illustrated example, one or more input devices 822 are connected to the interface circuit 820. The input device(s) 822 permit(s) a user to enter data and commands into the processor 812. The input device(s) can be implemented by, for example, an audio sensor, a microphone, a camera (still or video), a keyboard, a button, a mouse, a touchscreen, a track-pad, a trackball, isopoint and/or a voice recognition system.
One or more output devices 824 are also connected to the interface circuit 820 of the illustrated example. The output devices 824 can be implemented, for example, by display devices (e.g., a light emitting diode (LED), an organic light emitting diode (OLED), a liquid crystal display, a cathode ray tube display (CRT), a touchscreen, a tactile output device, a light emitting diode (LED), a printer and/or speakers). The interface circuit 820 of the illustrated example, thus, typically includes a graphics driver card, a graphics driver chip or a graphics driver processor.
The interface circuit 820 of the illustrated example also includes a communication device such as a transmitter, a receiver, a transceiver, a modem and/or network interface card to facilitate exchange of data with external machines (e.g., computing devices of any kind) via a network 826 (e.g., an Ethernet connection, a digital subscriber line (DSL), a telephone line, coaxial cable, a cellular telephone system, etc.).
The processor platform 800 of the illustrated example also includes one or more mass storage devices 828 for storing software and/or data. Examples of such mass storage devices 828 include floppy disk drives, hard drive disks, compact disk drives, Blu-ray disk drives, RAID systems, and digital versatile disk (DVD) drives.
The coded instructions 832 of
From the foregoing, it will appreciate that the above disclosed methods, apparatus and articles of manufacture are operative to provide the individual with a better experience when interacting with media content, including websites, by altering content and/or presenting new content based on a mental state of the individual while viewing and/or interacting with the content.
Although certain example methods, apparatus and articles of manufacture have been disclosed herein, the scope of coverage of this patent is not limited thereto. On the contrary, this patent covers all methods, apparatus and articles of manufacture fairly falling within the scope of the claims of this patent.
Marci, Carl D., Levine, Brian, Murray, Brendan
Patent | Priority | Assignee | Title |
11290779, | May 19 2015 | Nielsen Consumer LLC | Methods and apparatus to adjust content presented to an individual |
11700421, | Dec 27 2012 | The Nielsen Company (US), LLC | Methods and apparatus to determine engagement levels of audience members |
11924509, | Dec 27 2012 | The Nielsen Company (US), LLC | Methods and apparatus to determine engagement levels of audience members |
11956502, | Dec 27 2012 | The Nielsen Company (US), LLC | Methods and apparatus to determine engagement levels of audience members |
12142068, | Dec 15 2011 | The Nielsen Company (US), LLC | Methods and apparatus to capture images |
ER6959, |
Patent | Priority | Assignee | Title |
2549836, | |||
3490439, | |||
3572322, | |||
3735753, | |||
3880144, | |||
3901215, | |||
3998213, | Apr 08 1975 | Bio-Volt Corporation | Self-adjustable holder for automatically positioning electroencephalographic electrodes |
4075657, | Mar 03 1977 | MICROMEASUREMENTS, INC | Eye movement monitoring apparatus |
4145122, | May 31 1977 | Colorado Seminary | Method and apparatus for monitoring the position of the eye |
4149716, | Jun 24 1977 | ATARI, INC A CORP OF DE | Bionic apparatus for controlling television games |
4201224, | Dec 29 1978 | NEW YORK UNIVERSITY, A NY CORP | Electroencephalographic method and system for the quantitative description of patient brain states |
4279258, | Mar 26 1980 | NEW YORK UNIVERSITY, A NY CORP | Rapid automatic electroencephalographic evaluation |
4411273, | Jan 30 1978 | New York Medical College | System and method for electrode pair derivations in electroencephalography |
4417592, | May 11 1981 | NEW YORK UNIVERSITY, A NY CORP | Digital electroencephalographic instrument and method |
4537198, | May 03 1983 | Electrode cap | |
4557270, | Aug 23 1983 | NEW YORK UNIVERSITY, A NY CORP | Electroencephalographic system for intra-operative open-heart surgery |
4610259, | Aug 31 1983 | AEQUITRON MEDICAL, INC | EEG signal analysis system |
4613951, | Oct 11 1984 | Hewlett-Packard Company | Time interval measuring apparatus and method |
4626904, | Nov 12 1985 | ARBITRON INC | Meter for passively logging the presence and identity of TV viewers |
4632122, | Apr 24 1985 | HZI RESEARCH CENTER, INC | Method and apparatus for conducting brain function diagnostic test |
4683891, | Apr 26 1982 | CORNELLIER, VINCENT | Biomonitoring stress management method and device |
4683892, | Apr 24 1985 | HZI RESEARCH CENTER, INC | Method and apparatus for conducting brain function diagnostic test |
4686999, | Apr 10 1985 | INTERNATIONAL ADAPTIVE MONITORS, INC | Multi-channel ventilation monitor and method |
4695879, | Feb 07 1986 | Television viewer meter | |
4736751, | Dec 16 1986 | EEG Systems Laboratory | Brain wave source network location scanning method and system |
4800888, | Aug 17 1987 | HZI RESEARCH CENTER, INC | Enhanced electrode headset |
4802484, | Jun 13 1983 | Ernest H., Friedman | Method and apparatus to monitor asymmetric and interhemispheric brain functions |
4846190, | Aug 23 1983 | UNIVERSITY, NEW YORK | Electroencephalographic system data display |
4870579, | Oct 01 1987 | Adobe Systems Incorporated | System and method of predicting subjective reactions |
4885687, | May 08 1986 | Regents of the University of Minnesota | Trackig instrumentation for measuring human motor control |
4894777, | Jul 28 1986 | CANON KABUSHIKI KAISHA, A CORP OF JAPAN | Operator mental condition detector |
4913160, | Sep 30 1987 | NEW YORK UNIVERSITY, A CORP OF NY | Electroencephalographic system and method using factor structure of the evoked potentials |
4955388, | Jul 30 1985 | INSTITUTE FOR INNOVATION AND ENTERPRISE LTD AS TRUSTEE FOR BRAIN SCIENCES INSTITUTE TRUST | Electroencephalographic attention monitor |
4967038, | Dec 16 1986 | Sam Techology Inc. | Dry electrode brain wave recording system |
4973149, | Aug 19 1987 | University of Virginia | Eye movement detector |
4987903, | Nov 14 1988 | KEPPEL, WILLIAM | Method and apparatus for identifying and alleviating semantic memory deficiencies |
5003986, | Nov 17 1988 | POOL, KENNETH D , JR ; ELECTROPHYSIOLOGY LABORATORIES OF DALLAS, INC | Hierarchial analysis for processing brain stem signals to define a prominent wave |
5010891, | Oct 09 1987 | ASPECT MEDICAL SYSTEMS, INC | Cerebral biopotential analysis system and method |
5038782, | Dec 16 1986 | Sam Technology, Inc. | Electrode system for brain wave detection |
5052401, | Aug 06 1986 | Northrop Grumman Corporation | Product detector for a steady visual evoked potential stimulator and product detector |
5083571, | Apr 18 1988 | NEW YORK UNIVERSITY, A CORP OF NEW YORK | Use of brain electrophysiological quantitative data to classify and subtype an individual into diagnostic categories by discriminant and cluster analysis |
5137027, | May 01 1987 | Method for the analysis and utilization of P300 brain waves | |
5213338, | Sep 30 1991 | Brain wave-directed amusement device | |
5226177, | Mar 27 1990 | VIEWFACTS, INC | Real-time wireless audience response system |
5243517, | Aug 03 1988 | Northrop Grumman Corporation | Method and apparatus for physiological evaluation of short films and entertainment materials |
5273037, | Aug 01 1991 | HZI RESEARCH CENTER, INC | Electrode assembly for EEG headset |
5291888, | Aug 26 1991 | Electrical Geodesics, Inc. | Head sensor positioning network |
5293867, | Sep 24 1992 | Method and apparatus for marking electrode locations for electroencephalographic procedure | |
5295491, | Sep 26 1991 | Sam Technology, Inc. | Non-invasive human neurocognitive performance capability testing method and system |
5331544, | Apr 23 1992 | A C NIELSEN US , INC | Market research method and system for collecting retail store and shopper market research data |
5339826, | Dec 09 1991 | WESTINGHOUSE ELECTRIC CO LLC | Method for training material evaluation with method of EEG spectral estimation |
5345281, | Dec 17 1992 | Eye tracking system and method | |
5357957, | Aug 01 1991 | HZI RESEARCH CENTER, INC | Electrode assembly for EEG headset |
5363858, | Feb 11 1993 | BRAINWAVE SCIENCE INC | Method and apparatus for multifaceted electroencephalographic response analysis (MERA) |
5392788, | Feb 03 1993 | Method and device for interpreting concepts and conceptual thought from brainwave data and for assisting for diagnosis of brainwave disfunction | |
5406956, | Feb 11 1993 | BRAINWAVE SCIENCE INC | Method and apparatus for truth detection |
5410609, | Aug 09 1991 | MATSUSHITA ELECTRIC INDUSTRIAL CO , LTD | Apparatus for identification of individuals |
5436830, | Feb 01 1993 | OLSON ZALTMAN & ASSOCIATES, LLC | Metaphor elicitation method and apparatus |
5447166, | Sep 26 1991 | Neurocognitive adaptive computer interface method and system based on on-line measurement of the user's mental effort | |
5474082, | Jan 06 1993 | BRAIN ACTUATED TECHNOLOGIES, INC | Brain-body actuated system |
5479934, | Nov 08 1991 | Masimo Corporation | EEG headpiece with disposable electrodes and apparatus and system and method for use therewith |
5513649, | Mar 22 1994 | Sam Technology, Inc. | Adaptive interference canceler for EEG movement and eye artifacts |
5518007, | Dec 01 1994 | Electrode locator | |
5537618, | Dec 23 1993 | TIMDOLIR LTD , LIMITED LIABILITY COMPANY | Method and apparatus for implementing user feedback |
5540285, | Dec 06 1988 | Fuel containment medium | |
5550928, | Dec 15 1992 | NIELSEN COMPANY US , LLC, THE, A DELAWARE LIMITED LIABILITY COMPANY | Audience measurement system and method |
5617855, | Sep 01 1994 | Medical testing device and associated method | |
5655534, | Nov 30 1992 | Nexstim Oy | Method and apparatus for separating the different components of evoked response and spontaneous activity brain signals as well as of signals measured from the heart |
5676138, | Mar 15 1996 | Emotional response analyzer system with multimedia display | |
5676148, | Mar 31 1995 | Siemens Medical Solutions USA, Inc | Method and system for doppler ultrasound audio dealiasing |
5687322, | May 01 1989 | Catalina Marketing Corporation | Method and system for selective incentive point-of-sale marketing in response to customer shopping histories |
5720619, | Apr 24 1995 | Interactive computer assisted multi-media biofeedback system | |
5724987, | Sep 26 1991 | Sam Technology, Inc. | Neurocognitive adaptive computer-aided training method and system |
5726701, | Apr 20 1995 | Intel Corporation | Method and apparatus for stimulating the responses of a physically-distributed audience |
5729205, | Mar 07 1997 | Hyundai Motor Company | Automatic transmission system of an emergency signal and a method thereof using a driver's brain wave |
5736986, | Jul 14 1995 | Virtual reality mental conditioning medium | |
5740035, | Jul 23 1991 | THE NIELSEN COMPANY US , LLC | Self-administered survey systems, methods and devices |
5762611, | Nov 12 1996 | The United States of America as represented by the Secretary of the Navy; NAVY, UNITED STATES OF AMERICA, THE, AS REPRESENTED BY THE SECRETARY | Evaluation of a subject's interest in education, training and other materials using brain activity patterns |
5771897, | Apr 08 1996 | SITNIKOV, L | Method of and apparatus for quantitative evaluation of current changes in a functional state of human organism |
5774591, | Dec 15 1995 | University of Maryland | Apparatus and method for recognizing facial expressions and facial gestures in a sequence of images |
5787187, | Apr 01 1996 | Sandia Corporation | Systems and methods for biometric identification using the acoustic properties of the ear canal |
5800351, | Oct 04 1996 | Rest Technologies, Inc. | Electrode supporting head set |
5802208, | May 06 1996 | Alcatel-Lucent USA Inc | Face recognition using DCT-based feature vectors |
5802220, | Dec 15 1995 | University of Maryland | Apparatus and method for tracking facial motion through a sequence of images |
5812642, | Jul 12 1995 | Audience response monitor and analysis system and method | |
5817029, | Jun 26 1997 | Sam Technology, Inc. | Spatial measurement of EEG electrodes |
5842199, | Oct 18 1996 | Regents of the University of Minnesota | System, method and article of manufacture for using receiver operating curves to evaluate predictive utility |
5848399, | Nov 30 1993 | DASSAULT SYSTEMES, S A | Computer system for allowing a consumer to purchase packaged goods at home |
5892566, | Jan 20 1998 | RIVERSIDE NATIONAL BANK OF FLORIDA | Fiber optic eye-tracking system |
5945863, | Jun 18 1997 | MACOM CONNECTIVITY SOLUTIONS, LLC | Analog delay circuit |
5961332, | Sep 08 1992 | KICKAPOO RUN, LLC | Apparatus for processing psychological data and method of use thereof |
5974262, | Aug 15 1997 | Fuller Research Corporation | System for generating output based on involuntary and voluntary user input without providing output information to induce user to alter involuntary input |
5983129, | Feb 19 1998 | COWAN, JONATHAN DANIEL | Method for determining an individual's intensity of focused attention and integrating same into computer program |
5995868, | Jan 23 1996 | University of Kansas | System for the prediction, rapid detection, warning, prevention, or control of changes in activity states in the brain of a subject |
6001065, | Aug 02 1995 | IBVA Technologies, Inc. | Method and apparatus for measuring and analyzing physiological signals for active or passive control of physical and virtual spaces and the contents therein |
6016475, | Oct 08 1996 | MINNESOTA UNIVERSITY OF, REGENTS OF, THE | System, method, and article of manufacture for generating implicit ratings based on receiver operating curves |
6021346, | Nov 13 1997 | Electronics and Telecommunications Research Institute | Method for determining positive and negative emotional states by electroencephalogram (EEG) |
6032129, | Sep 06 1997 | International Business Machines Corporation | Customer centric virtual shopping experience with actors agents and persona |
6052619, | Aug 07 1997 | New York University | Brain function scan system |
6088040, | Sep 17 1996 | ATR Human Information Processing Research Laboratories | Method and apparatus of facial image conversion by interpolation/extrapolation for plurality of facial expression components representing facial image |
6099319, | Feb 24 1998 | THE NIELSEN COMPANY US , LLC , A DELAWARE LIMITED LIABILITY COMPANY | Neuroimaging as a marketing tool |
6120440, | Sep 11 1990 | Diagnostic method | |
6128521, | Jul 10 1998 | Masimo Corporation | Self adjusting headgear appliance using reservoir electrodes |
6154669, | Nov 06 1998 | CAPITA SYSTEMS, INC | Headset for EEG measurements |
6155927, | Apr 06 1995 | Electronic competition system and method for using same | |
6161030, | Feb 05 1999 | ADVANCED BRAIN MONITORING, INC | Portable EEG electrode locator headgear |
6170018, | Nov 27 1995 | Oracle America, Inc | Remote procedure calling using an existing descriptor mechanism |
6171239, | Aug 17 1998 | Emory University | Systems, methods, and devices for controlling external devices by signals derived directly from the nervous system |
6173260, | Oct 29 1997 | Vulcan Patents LLC | System and method for automatic classification of speech based upon affective content |
6175753, | Jul 02 1998 | BALTIMORE BIOMEDICAL, INC | Methods and mechanisms for quick-placement electroencephalogram (EEG) electrodes |
6182113, | Sep 16 1997 | SNAP INC | Dynamic multiplexing of hyperlinks and bookmarks |
6190314, | Jul 15 1998 | UNILOC 2017 LLC | Computer input device with biosensors for sensing user emotions |
6212502, | Mar 23 1998 | Microsoft Technology Licensing, LLC | Modeling and projecting emotion and personality from a computer user interface |
6228038, | Apr 14 1997 | OBJECT SOFTWARE MODULES N V | Measuring and processing data in reaction to stimuli |
6236885, | Jun 30 1999 | CAPITA RESEARCH GROUP INC | System for correlating in a display stimuli and a test subject's response to the stimuli |
6236975, | Sep 29 1998 | Ignite Sales, Inc.; IGNITE SALES, INC | System and method for profiling customers for targeted marketing |
6254536, | Aug 02 1995 | IBVA Technologies, Inc. | Method and apparatus for measuring and analyzing physiological signals for active or passive control of physical and virtual spaces and the contents therein |
6280198, | Jan 29 1999 | LUMOS LABS INC | Remote computer implemented methods for cognitive testing |
6286005, | Mar 11 1998 | NIELSEN COMPANY US , LLC , THE | Method and apparatus for analyzing data and advertising optimization |
6289234, | Dec 02 1998 | Siemens Healthcare GmbH | Method for time-resolved and location-resolved presentation of functional brain activities with magnetic resonance and apparatus for the implementation of the method |
6292688, | Feb 28 1996 | Advanced Neurotechnologies, Inc.; ADVANCED NEUROTECHNOLOGIES, INC | Method and apparatus for analyzing neurological response to emotion-inducing stimuli |
6299308, | Apr 02 1999 | JOLLY SEVEN, SERIES 70 OF ALLIED SECURITY TRUST I | Low-cost non-imaging eye tracker system for computer control |
6301493, | Jul 10 1999 | Masimo Corporation | Reservoir electrodes for electroencephalograph headgear appliance |
6315569, | Feb 24 1998 | OLSON ZALTMAN & ASSOCIATES, LLC | Metaphor elicitation technique with physiological function monitoring |
6330470, | Dec 17 1999 | Philips North America LLC | Method for localizing electrical activity in the body |
6334778, | Apr 26 1994 | HEALTH HERO NETWORK, INC | Remote psychological diagnosis and monitoring system |
6358201, | Mar 02 1999 | QUANTUM INTECH, INC | Method and apparatus for facilitating physiological coherence and autonomic balance |
6370513, | Aug 08 1997 | DIETFOOD CORP | Method and apparatus for automated selection, organization, and recommendation of items |
6374143, | Aug 18 1999 | MED-EL ELEKTRO-MEDIZINISCHE GERATE GESELLSCHAFT M B H | Modiolar hugging electrode array |
6381481, | Feb 05 1999 | ADVANCED BRAIN MONITORING, INC | Portable EEG electrode locator headgear |
6398643, | Sep 30 1999 | Promotional gaming device | |
6422999, | May 13 1999 | SENSORY LOGIC, INC | Method of measuring consumer reaction |
6434419, | Jun 26 2000 | SAM TECHNOLOGY, INC | Neurocognitive ability EEG measurement method and system |
6435878, | Feb 27 1997 | BCI, LLC; COGNITIVE DIAGNOSTICS, INC ; BRAIN COM, INC ; @BRAIN COM, INC | Interactive computer program for measuring and analyzing mental ability |
6453194, | Mar 29 2000 | Method of measuring consumer reaction while participating in a consumer activity | |
6453241, | Dec 23 1998 | Microsoft Technology Licensing, LLC | Method and system for analyzing biological response signal data |
6487444, | Mar 28 2000 | Design evaluation method, equipment thereof, and goods design method | |
6488617, | Oct 13 2000 | Universal Hedonics | Method and device for producing a desired brain state |
6510340, | Jan 10 2000 | JORDAN NEUROSCIENCE, INC | Method and apparatus for electroencephalography |
6520905, | Feb 26 1998 | CARESTREAM HEALTH, INC | Management of physiological and psychological state of an individual using images portable biosensor device |
6545685, | Jan 14 1999 | RPX Corporation | Method and system for efficient edge blending in high fidelity multichannel computer graphics displays |
6575902, | Jan 27 1999 | Compumedics Limited | Vigilance monitoring system |
6577329, | Feb 25 1999 | Tobii AB | Method and system for relevance feedback through gaze tracking and ticker interfaces |
6585521, | Dec 21 2001 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | Video indexing based on viewers' behavior and emotion feedback |
6594521, | Dec 17 1999 | Philips North America LLC | Method for localizing electrical activity in the body |
6598006, | Oct 18 1999 | ADVANCED TELECOMMUNICATIONS RESEARCH INSTITUTE INTERNATIONAL | Data input device using a palatal plate |
6609024, | Nov 12 1998 | Electronics and Telecommunications Research Institute | Method of making a judgment on emotional positivity or negativity by detecting asymmetry of brain waves of left and right cerebral hemispheres |
6648822, | Jul 24 2000 | Sharp Kabushiki Kaisha | Communication apparatus and communication method for outputting an estimate of a patient's mental state |
6652283, | Dec 30 1999 | CEREGO LLC | System apparatus and method for maximizing effectiveness and efficiency of learning retaining and retrieving knowledge and skills |
6654626, | Oct 11 1996 | Covidien LP | Electrode array system for measuring electrophysiological signals |
6662052, | Apr 19 2001 | Neuro and Cardiac Technologies, LLC | Method and system for neuromodulation therapy using external stimulator with wireless communication capabilites |
6665560, | Oct 04 2001 | PENDRAGON NETWORKS LLC | Sleep disconnect safety override for direct human-computer neural interfaces for the control of computer controlled functions |
6678685, | Jan 26 2000 | FAMILYTIME COM, INC | Integrated household management system and method |
6688890, | Feb 09 2001 | QUANTEC GMBH | Device, method and computer program product for measuring a physical or physiological activity by a subject and for assessing the psychosomatic state of the subject |
6708051, | Nov 10 1998 | Compumedics Limited | FMRI compatible electrode and electrode placement techniques |
6712468, | Dec 12 2001 | Techniques for facilitating use of eye tracking data | |
6754524, | Aug 28 2000 | Research Foundation of the City University of New York | Method for detecting deception |
6757556, | May 26 1998 | Ineedmd. com | Electrode sensor |
6788882, | Apr 17 1998 | SYNCHVIEW TECHNOLOGIES, LLC | Systems and methods for storing a plurality of video streams on re-writable random-access media and time-and channel- based retrieval thereof |
6792304, | May 15 1998 | SSPT PROPRIETARY LIMITED | Mass communication assessment system |
6842877, | Dec 18 1998 | Microsoft Technology Licensing, LLC | Contextual responses based on automated learning techniques |
6850252, | Oct 05 1999 | Blanding Hovenweep, LLC; HOFFBERG FAMILY TRUST 1 | Intelligent electronic appliance system and method |
6852875, | May 19 2000 | NUTRASWEET PROPERTY HOLDINGS, INC | Synthesis of N-[N-(3,3-dimethylbutyl)-L-α-aspartyl]-L-phenylalanine 1-methyl ester using oxazolidinone derivatives |
6888457, | Jul 27 2001 | Qualcomm Incorporated | Monitoring of user response to performances |
6904408, | Oct 19 2000 | Bionet method, system and personalized web content manager responsive to browser viewers' psychological preferences, behavioral responses and physiological stress indicators | |
6950698, | Jul 02 2003 | Instrumentarium Corp | Method of positioning electrodes for central nervous system monitoring |
6958710, | Dec 24 2002 | CITIBANK, N A | Universal display media exposure measurement |
6973342, | Mar 02 2000 | ADVANCED NEUROMODULATION SYSTEMS, INC | Flexible bio-probe assembly |
6993380, | Jun 04 2003 | Cleveland Medical Devices, Inc. | Quantitative sleep analysis method and system |
7020508, | Aug 22 2002 | JB IP ACQUISITION LLC | Apparatus for detecting human physiological and contextual information |
7043056, | Mar 08 2001 | SEEING MACHINES LIMITED | Facial image processing system |
7047550, | Jul 03 1997 | Panasonic Intellectual Property Corporation of America | System for processing program information |
7113916, | Sep 07 2001 | SENSORY LOGIC, INC | Method of facial coding monitoring for the purpose of gauging the impact and appeal of commercially-related stimuli |
7120880, | Feb 25 1999 | Tobii AB | Method and system for real-time determination of a subject's interest level to media content |
7130673, | Apr 08 2003 | Instrumentarium Corp | Method of positioning electrodes for central nervous system monitoring and sensing pain reactions of a patient |
7150715, | Feb 05 2001 | Network enabled biofeedback administration | |
7164967, | May 04 2000 | IGUANA ROBOTICS, INC | Biomorphic rhythmic movement controller |
7177675, | Jul 11 2001 | MYND ANALYTICS, INC , A CALIFORNIA CORPORATION | Electroencephalography based systems and methods for selecting therapies and predicting outcomes |
7222071, | Sep 27 2002 | CITIBANK, N A | Audio data receipt/exposure measurement with code monitoring and signature extraction |
7246081, | Sep 07 2001 | SENSORY LOGIC, INC | Method of facial coding monitoring for the purpose of gauging the impact and appeal of commercially-related stimuli |
7249708, | Feb 04 2005 | The Procter & Gamble Company | Household management systems and methods |
7269590, | Jan 29 2004 | Slack Technologies, LLC; SLACK TECHNOLOGIES, INC | Method and system for customizing views of information associated with a social network user |
7272982, | Mar 19 2004 | CITIBANK, N A | Gathering data concerning publication usage |
7286871, | Aug 15 2000 | The Regents of the University of California | Method and apparatus for reducing contamination of an electrical signal |
7340060, | Oct 26 2005 | Black Box Intelligence Limited | System and method for behavioural modelling |
7359894, | Jun 30 2004 | GOOGLE LLC | Methods and systems for requesting and providing information in a social network |
7391835, | Sep 29 2004 | Oracle America, Inc | Optimizing synchronization between monitored computer system signals |
7394385, | Jul 31 2003 | WELLCARE SYSTEMS, INC | Comprehensive monitoring system |
7408460, | Mar 19 2004 | CITIBANK, N A | Gathering data concerning publication usage |
7420464, | Mar 15 2004 | CITIBANK, N A | Methods and systems for gathering market research data inside and outside commercial establishments |
7443292, | Mar 19 2004 | CITIBANK, N A | Gathering data concerning publication usage |
7460827, | Jul 26 2002 | CITIBANK, N A | Radio frequency proximity detection and identification system and method |
7463143, | Mar 15 2004 | CITIBANK, N A | Methods and systems for gathering market research data within commercial establishments |
7463144, | Mar 19 2004 | CITIBANK, N A | Gathering data concerning publication usage |
7471987, | Mar 08 2002 | CITIBANK, N A | Determining location of an audience member having a portable media monitor |
7483835, | Dec 23 2002 | CITIBANK, N A | AD detection using ID code and extracted signature |
7483844, | Feb 20 2002 | Honda Giken Kogyo Kabushiki Kaisha | Computer system for retrieving a product that meets characteristics of a customer |
7496400, | Oct 17 2003 | GE Healthcare Finland Oy | Sensor arrangement |
7548774, | Oct 23 2003 | Vivosonic Inc. | Method and apparatus for the collection of physiological electrical potentials |
7551952, | Oct 26 2005 | Sam Technology, Inc. | EEG electrode headset |
7592908, | Aug 13 2003 | CITIBANK, N A | Universal display exposure monitor using personal locator service |
7614066, | May 03 2002 | Time Warner Cable Enterprises LLC | Use of multiple embedded messages in program signal streams |
7623823, | Aug 31 2004 | CITIBANK, N A | Detecting and measuring exposure to media content items |
7630757, | Jan 06 1997 | Flint Hills Scientific LLC | System for the prediction, rapid detection, warning, prevention, or control of changes in activity states in the brain of a subject |
7636456, | Jan 23 2004 | SONY EUROPE B V | Selectively displaying information based on face detection |
7650793, | Mar 19 2004 | CITIBANK, N A | Gathering data concerning publication usage |
7657523, | Mar 09 2006 | Customerforce.com | Ranking search results presented to on-line users as a function of perspectives of relationships trusted by the users |
7658327, | Oct 03 2005 | TTEC HOLDINGS, INC | Virtual retail assistant |
7689272, | Jun 07 2001 | BRAINWAVE SCIENCE INC | Method for brain fingerprinting, measurement, assessment and analysis of brain function |
7697979, | Oct 18 2002 | Centre National de la Recherche Scientifique | Analysis method and real time medical or cognitive monitoring device based on the analysis of a subject's cerebral electromagnetic activity use of said method for characterizing and differentiating physiological or pathological states |
7698238, | Apr 01 2004 | Sony Deutschland GmbH | Emotion controlled system for processing multimedia data |
7720351, | Apr 04 2005 | STRIPE, INC | Preservation and improvement of television advertising in digital environment |
7729755, | Jun 14 2004 | Cephos Corp. | Questions and control paradigms for detecting deception by measuring brain activity |
7765564, | Aug 29 2003 | CITIBANK, N A | Audio based methods and apparatus for detecting a channel change event |
7774052, | Jun 13 2002 | Compumedics Limited | Methods and apparatus for monitoring consciousness |
7781548, | Mar 02 2004 | Velocys, Inc. | Microchannel polymerization reactor |
7797186, | Oct 18 2005 | DYBUS, DONNELLY ANDREW, MR | Method and system for gathering and recording real-time market survey and other data from radio listeners and television viewers utilizing telephones including wireless cell phones |
7809420, | Jun 25 2003 | Covidien LP | Hat-based oximeter sensor |
7816951, | Oct 10 2008 | National Semiconductor Corporation | Locally boosted top plate sampling for a sampling capacitor |
7840248, | Jan 27 2003 | Compumedics Limited | Online source reconstruction for eeg/meg and ecg/mcg |
7840250, | Nov 13 2001 | Magstim Group, Incorporated | Method for neural current imaging |
7844484, | Oct 11 2000 | BuzzMetrics, Ltd. | System and method for benchmarking electronic message activity |
7865235, | Sep 12 2005 | Emotiv Systems Pty Ltd | Method and system for detecting and classifying the mental state of a subject |
7865394, | Apr 17 2000 | ALTERIAN TECHNOLOGY LIMITED | Multimedia messaging method and system |
7892764, | Nov 21 2006 | Legacy Emanuel Hospital & Health Center | System for seizure suppression |
7895075, | Aug 12 2004 | Cable Television Advertising Group LLC | Method and apparatus for determining an effective media channel to use for advertisement |
7895625, | Dec 24 2003 | Time Warner, Inc | System and method for recommending programming to television viewing communities |
7908133, | Sep 27 2002 | CITIBANK, N A | Gathering research data |
7917366, | Mar 24 2000 | BEYOND VERBAL COMMUNICATION LTD | System and method for determining a personal SHG profile by voice analysis |
7930199, | Jul 21 2006 | SENSORY LOGIC, INC | Method and report assessing consumer reaction to a stimulus by matching eye position with facial coding |
7946974, | Nov 10 2005 | Biomedical device for treating by virtual immersion | |
7962315, | Mar 19 2004 | CITIBANK, N A | Gathering data concerning publication usage |
7974889, | Oct 19 2007 | Social networking interactive shopping system | |
7988557, | Jan 02 2004 | Interactive Productline AB | Method for playing games using brain waves |
8014847, | Dec 13 2001 | MUSC Foundation for Research Development | Systems and methods for detecting deception by measuring brain activity |
8027518, | Jun 25 2007 | Microsoft Technology Licensing, LLC | Automatic configuration of devices based on biometric data |
8055722, | Dec 02 2009 | International Business Machines Corporation | Notification control through brain monitoring of end user concentration |
8060795, | May 15 2008 | International Business Machines Corporation | Solution for automatically incorporating diagnostic data within screen capture images |
8065203, | Aug 01 2007 | SAS Institute Inc. | Computer-implemented systems and methods for product attribute estimations |
8069125, | Dec 20 2007 | The Invention Science Fund I, LLC | Methods and systems for comparing media content |
8073707, | Jun 16 2000 | JB IP ACQUISITION LLC | System for detecting, monitoring, and reporting an individual's physiological or contextual status |
8079054, | Apr 14 2008 | Adobe Inc | Location for secondary content based on data differential |
8082215, | Sep 19 2008 | The Invention Science Fund I, LLC | Acquisition and particular association of inference data indicative of inferred mental states of authoring users |
8086563, | May 23 2008 | The Invention Science Fund I, LLC | Acquisition and particular association of data indicative of an inferred mental state of an authoring user |
8098152, | Aug 13 2003 | CITIBANK, N A | Universal display exposure monitor using personal locator service |
8099159, | Sep 14 2005 | ZYTO CORP | Methods and devices for analyzing and comparing physiological parameter measurements |
8099315, | Jun 05 2007 | Nokia Technologies Oy | Interest profiles for audio and/or video streams |
8103328, | Oct 01 2007 | Quantum Applied Science and Research, Inc.; QUANTUM APPLIED SCIENCE AND RESEARCH, INC | Self-locating sensor mounting apparatus |
8126220, | May 03 2007 | Hewlett-Packard Development Company L.P. | Annotating stimulus based on determined emotional response |
8135606, | Apr 15 2004 | CITIBANK, N A | Gathering data concerning publication usage and exposure to products and/or presence in commercial establishment |
8151298, | Apr 06 2001 | AT&T Corp | Method and system for embedding information into streaming media |
8165916, | Feb 01 1999 | Blanding Hovenweep, LLC; HOFFBERG FAMILY TRUST 1 | Internet appliance system and method |
8179604, | Jul 13 2011 | GOOGLE LLC | Wearable marker for passive interaction |
8196168, | Dec 10 2003 | Time Warner, Inc | Method and apparatus for exchanging preferences for replaying a program on a personal video recorder |
8200775, | Feb 01 2005 | Newsilike Media Group, Inc | Enhanced syndication |
8209224, | Oct 29 2009 | Nielsen Consumer LLC | Intracluster content management using neuro-response priming data |
8229469, | Mar 15 2004 | CITIBANK, N A | Methods and systems for mapping locations of wireless transmitters for use in gathering market research data |
8235725, | Feb 20 2005 | SENSORY LOGIC, INC | Computerized method of assessing consumer reaction to a business stimulus employing facial coding |
8239030, | Jan 06 2010 | EVOKE NEUROSCIENCE, INC | Transcranial stimulation device and method based on electrophysiological testing |
8255267, | Jul 13 2007 | Wahrheit, LLC | System and method for determining relative preferences |
8270814, | Jan 21 2009 | Nielsen Consumer LLC | Methods and apparatus for providing video with embedded media |
8296172, | Sep 05 2006 | Nielsen Consumer LLC | Method and system for determining audience response to a sensory stimulus |
8300526, | Jan 07 2009 | Hitachi, Ltd. | Network relay apparatus and packet distribution method |
8308562, | Apr 29 2008 | LNW GAMING, INC | Biofeedback for a gaming device, such as an electronic gaming machine (EGM) |
8326002, | Aug 13 2009 | SENSORY LOGIC INC | Methods of facial coding scoring for optimally identifying consumers' responses to arrive at effective, incisive, actionable conclusions |
8327395, | Oct 02 2007 | Nielsen Consumer LLC | System providing actionable insights based on physiological responses from viewers of media |
8332883, | Oct 02 2007 | Nielsen Consumer LLC | Providing actionable insights based on physiological responses from viewers of media |
8335715, | Nov 19 2009 | Nielsen Consumer LLC | Advertisement exchange using neuro-response data |
8381244, | Nov 21 2006 | WIDEORBIT OPCO INC | Content injection system and methodology |
8386312, | May 01 2007 | Nielsen Consumer LLC | Neuro-informatics repository system |
8386313, | Aug 28 2007 | Nielsen Consumer LLC | Stimulus placement system using subject neuro-response measurements |
8388165, | Jul 15 2008 | Displaying device and method thereof | |
8392250, | Aug 09 2010 | Nielsen Consumer LLC | Neuro-response evaluated stimulus in virtual reality environments |
8392251, | Aug 09 2010 | Nielsen Consumer LLC | Location aware presentation of stimulus material |
8392253, | May 16 2007 | Nielsen Consumer LLC | Neuro-physiology and neuro-behavioral based stimulus targeting system |
8392254, | Aug 28 2007 | Nielsen Consumer LLC | Consumer experience assessment system |
8392255, | Aug 29 2007 | Nielsen Consumer LLC | Content based selection and meta tagging of advertisement breaks |
8396744, | Aug 25 2010 | Nielsen Consumer LLC | Effective virtual reality environments for presentation of marketing materials |
8442429, | Apr 06 2010 | System for providing audio messages on a mobile device | |
8464288, | Jan 21 2009 | Nielsen Consumer LLC | Methods and apparatus for providing personalized media in video |
8467133, | Feb 28 2010 | Microsoft Technology Licensing, LLC | See-through display with an optical assembly including a wedge-shaped illumination system |
8473345, | Mar 29 2007 | Nielsen Consumer LLC | Protocol generator and presenter device for analysis of marketing and entertainment effectiveness |
8477425, | Feb 28 2010 | Microsoft Technology Licensing, LLC | See-through near-eye display glasses including a partially reflective, partially transmitting optical element |
8484081, | Mar 29 2007 | Nielsen Consumer LLC | Analysis of marketing and entertainment effectiveness using central nervous system, autonomic nervous system, and effector data |
8494610, | Sep 20 2007 | Nielsen Consumer LLC | Analysis of marketing and entertainment effectiveness using magnetoencephalography |
8494905, | Jun 06 2007 | Nielsen Consumer LLC | Audience response analysis using simultaneous electroencephalography (EEG) and functional magnetic resonance imaging (fMRI) |
8533042, | Jul 30 2007 | Nielsen Consumer LLC | Neuro-response stimulus and stimulus attribute resonance estimator |
8543446, | Feb 28 2011 | Nielsen Consumer LLC | Methods and apparatus to predict new product performance metrics |
8548852, | Aug 25 2010 | Nielsen Consumer LLC | Effective virtual reality environments for presentation of marketing materials |
8560530, | May 17 2010 | CITIBANK, N A | Methods, apparatus, and articles of manufacture to rank web site influence |
8561095, | Nov 13 2001 | S I SV EL SOCIETA ITALIANA PER LO SVILUPPO DELL ELETTRONICA S P A | Affective television monitoring and control in response to physiological data |
8600100, | Apr 16 2009 | SENSORY LOGIC, INC | Method of assessing people's self-presentation and actions to evaluate personality type, behavioral tendencies, credibility, motivations and other insights through facial muscle activity and expressions |
8635105, | Aug 28 2007 | Nielsen Consumer LLC | Consumer experience portrayal effectiveness assessment system |
8655428, | May 12 2010 | Nielsen Consumer LLC | Neuro-response data synchronization |
8655437, | Aug 21 2009 | Nielsen Consumer LLC | Analysis of the mirror neuron system for evaluation of stimulus |
8684742, | Apr 19 2010 | Nielsen Consumer LLC | Short imagery task (SIT) research method |
8700009, | Jun 02 2010 | THE VISTA GROUP LLC | Method and apparatus for monitoring emotion in an interactive network |
8762202, | Oct 29 2009 | Nielsen Consumer LLC | Intracluster content management using neuro-response priming data |
8764652, | Mar 08 2007 | Nielsen Consumer LLC | Method and system for measuring and ranking an “engagement” response to audiovisual or interactive media, products, or activities using physiological signals |
8788372, | Mar 08 2004 | SAP SE | Method and system for classifying retail products and services using characteristic-based grouping structures |
8793715, | Dec 18 2007 | Nielsen Consumer LLC | Identifying key media events and modeling causal relationships between key events and reported feelings |
8793727, | Dec 10 2009 | ECHOSTAR UKRAINE, L.L.C. | System and method for selecting audio/video content for presentation to a user in response to monitored user activity |
8856235, | Oct 29 2010 | Meta Platforms, Inc | Inferring user profile attributes from social information |
8874727, | May 31 2010 | CITIBANK, N A | Methods, apparatus, and articles of manufacture to rank users in an online social network |
9000927, | Feb 25 2013 | Sony Interactive Entertainment LLC | System and method for monitoring biometric data |
9021515, | Oct 02 2007 | Nielsen Consumer LLC | Systems and methods to determine media effectiveness |
9032110, | Oct 14 2012 | AFFECTOMATICS LTD | Reducing power consumption of sensor by overriding instructions to measure |
9336535, | May 12 2010 | Nielsen Consumer LLC | Neuro-response data synchronization |
9514436, | Sep 05 2007 | Nielsen Consumer LLC | Method and system for predicting audience viewing behavior |
9514439, | Sep 05 2006 | Nielsen Consumer LLC | Method and system for determining audience response to a sensory stimulus |
9521960, | Oct 31 2007 | Nielsen Consumer LLC | Systems and methods providing en mass collection and centralized processing of physiological responses from viewers |
9557814, | Apr 22 2010 | SONY INTERACTIVE ENTERTAINMENT INC | Biometric interface for a handheld device |
9560984, | Oct 29 2009 | Nielsen Consumer LLC | Analysis of controlled and automatic attention for introduction of stimulus material |
9936250, | May 19 2015 | Nielsen Consumer LLC | Methods and apparatus to adjust content presented to an individual |
20010013009, | |||
20010020236, | |||
20010029468, | |||
20010032140, | |||
20010056225, | |||
20020053076, | |||
20020055857, | |||
20020056087, | |||
20020056124, | |||
20020059577, | |||
20020065826, | |||
20020072952, | |||
20020077534, | |||
20020082902, | |||
20020103429, | |||
20020111796, | |||
20020143627, | |||
20020155878, | |||
20020156842, | |||
20020169665, | |||
20020188216, | |||
20020188217, | |||
20020193670, | |||
20030013981, | |||
20030036955, | |||
20030037333, | |||
20030044050, | |||
20030059750, | |||
20030063222, | |||
20030065524, | |||
20030073921, | |||
20030081834, | |||
20030093784, | |||
20030093792, | |||
20030100998, | |||
20030104865, | |||
20030131351, | |||
20030149344, | |||
20030165270, | |||
20030177488, | |||
20030204412, | |||
20030208754, | |||
20030233278, | |||
20040001616, | |||
20040005143, | |||
20040013398, | |||
20040015608, | |||
20040055448, | |||
20040068431, | |||
20040073129, | |||
20040092809, | |||
20040098298, | |||
20040101212, | |||
20040133081, | |||
20040187167, | |||
20040193068, | |||
20040210159, | |||
20040219184, | |||
20040220483, | |||
20040236623, | |||
20050010475, | |||
20050041951, | |||
20050043646, | |||
20050060312, | |||
20050062637, | |||
20050071462, | |||
20050071865, | |||
20050076359, | |||
20050079474, | |||
20050097594, | |||
20050107716, | |||
20050113649, | |||
20050132401, | |||
20050143629, | |||
20050149964, | |||
20050154290, | |||
20050165766, | |||
20050177058, | |||
20050197590, | |||
20050203798, | |||
20050216243, | |||
20050223237, | |||
20050227233, | |||
20050240956, | |||
20050256905, | |||
20050261980, | |||
20050267798, | |||
20050272017, | |||
20050273017, | |||
20050273802, | |||
20050288954, | |||
20050289582, | |||
20060003732, | |||
20060009702, | |||
20060010470, | |||
20060035707, | |||
20060041548, | |||
20060042483, | |||
20060053110, | |||
20060069663, | |||
20060075003, | |||
20060093998, | |||
20060111044, | |||
20060111644, | |||
20060129458, | |||
20060149337, | |||
20060167376, | |||
20060168613, | |||
20060168630, | |||
20060176289, | |||
20060190822, | |||
20060218046, | |||
20060256133, | |||
20060257834, | |||
20060259360, | |||
20060259371, | |||
20060293921, | |||
20070016096, | |||
20070038516, | |||
20070048707, | |||
20070050256, | |||
20070055169, | |||
20070060830, | |||
20070060831, | |||
20070061720, | |||
20070066874, | |||
20070066915, | |||
20070066916, | |||
20070067007, | |||
20070067305, | |||
20070078700, | |||
20070078706, | |||
20070079331, | |||
20070101360, | |||
20070106170, | |||
20070112460, | |||
20070135727, | |||
20070135728, | |||
20070136753, | |||
20070150916, | |||
20070173733, | |||
20070198510, | |||
20070209047, | |||
20070214121, | |||
20070214471, | |||
20070225585, | |||
20070225674, | |||
20070226760, | |||
20070235716, | |||
20070238945, | |||
20070239713, | |||
20070250846, | |||
20070250901, | |||
20070265507, | |||
20070282566, | |||
20070294132, | |||
20070294705, | |||
20070294706, | |||
20080001600, | |||
20080004940, | |||
20080010110, | |||
20080027345, | |||
20080040740, | |||
20080059997, | |||
20080065468, | |||
20080065721, | |||
20080081961, | |||
20080082019, | |||
20080086356, | |||
20080091463, | |||
20080091512, | |||
20080097854, | |||
20080109840, | |||
20080125110, | |||
20080133724, | |||
20080147488, | |||
20080147742, | |||
20080152300, | |||
20080162182, | |||
20080177197, | |||
20080195471, | |||
20080204273, | |||
20080208072, | |||
20080214902, | |||
20080218472, | |||
20080221400, | |||
20080221472, | |||
20080221969, | |||
20080222670, | |||
20080222671, | |||
20080228077, | |||
20080249865, | |||
20080255949, | |||
20080287821, | |||
20080295126, | |||
20080306398, | |||
20090018996, | |||
20090024049, | |||
20090024447, | |||
20090024448, | |||
20090024449, | |||
20090024475, | |||
20090024747, | |||
20090025023, | |||
20090025024, | |||
20090030287, | |||
20090030303, | |||
20090030717, | |||
20090030762, | |||
20090030780, | |||
20090030930, | |||
20090036755, | |||
20090036756, | |||
20090037575, | |||
20090060240, | |||
20090062629, | |||
20090062679, | |||
20090062680, | |||
20090062681, | |||
20090063255, | |||
20090063256, | |||
20090066722, | |||
20090069652, | |||
20090070219, | |||
20090070798, | |||
20090082643, | |||
20090082689, | |||
20090082692, | |||
20090083129, | |||
20090088610, | |||
20090089830, | |||
20090094286, | |||
20090094627, | |||
20090094628, | |||
20090094629, | |||
20090097689, | |||
20090099873, | |||
20090112077, | |||
20090112117, | |||
20090119154, | |||
20090131764, | |||
20090132441, | |||
20090133047, | |||
20090138356, | |||
20090144780, | |||
20090150919, | |||
20090150920, | |||
20090153328, | |||
20090156907, | |||
20090156955, | |||
20090158308, | |||
20090163777, | |||
20090164132, | |||
20090171164, | |||
20090183193, | |||
20090187467, | |||
20090195392, | |||
20090214060, | |||
20090216611, | |||
20090221928, | |||
20090248484, | |||
20090248496, | |||
20090248594, | |||
20090249223, | |||
20090253996, | |||
20090259137, | |||
20090259509, | |||
20090271294, | |||
20090280215, | |||
20090287108, | |||
20090292587, | |||
20090300672, | |||
20090305006, | |||
20090318773, | |||
20090318826, | |||
20090327068, | |||
20090327907, | |||
20090328089, | |||
20090328122, | |||
20100004977, | |||
20100010370, | |||
20100022821, | |||
20100039618, | |||
20100041962, | |||
20100042012, | |||
20100060300, | |||
20100094702, | |||
20100094869, | |||
20100121716, | |||
20100125219, | |||
20100145176, | |||
20100145215, | |||
20100145217, | |||
20100153175, | |||
20100169153, | |||
20100169162, | |||
20100179881, | |||
20100180029, | |||
20100183279, | |||
20100186031, | |||
20100186032, | |||
20100198042, | |||
20100211439, | |||
20100214318, | |||
20100215289, | |||
20100218208, | |||
20100223094, | |||
20100228604, | |||
20100228614, | |||
20100234752, | |||
20100241580, | |||
20100249538, | |||
20100249636, | |||
20100250325, | |||
20100250347, | |||
20100250458, | |||
20100257023, | |||
20100257052, | |||
20100262477, | |||
20100263005, | |||
20100268540, | |||
20100268573, | |||
20100268720, | |||
20100269127, | |||
20100274152, | |||
20100274153, | |||
20100287152, | |||
20100292998, | |||
20100306030, | |||
20100306120, | |||
20100317988, | |||
20100318507, | |||
20100325660, | |||
20100331661, | |||
20100332283, | |||
20110004089, | |||
20110015503, | |||
20110020778, | |||
20110022965, | |||
20110040155, | |||
20110040202, | |||
20110043759, | |||
20110046473, | |||
20110046502, | |||
20110046503, | |||
20110046504, | |||
20110047035, | |||
20110047121, | |||
20110059422, | |||
20110071874, | |||
20110076942, | |||
20110084795, | |||
20110085700, | |||
20110098593, | |||
20110105937, | |||
20110106621, | |||
20110106750, | |||
20110119124, | |||
20110119129, | |||
20110119130, | |||
20110124977, | |||
20110131274, | |||
20110137894, | |||
20110138326, | |||
20110144519, | |||
20110153390, | |||
20110153391, | |||
20110153414, | |||
20110153423, | |||
20110161095, | |||
20110161163, | |||
20110161790, | |||
20110191142, | |||
20110208515, | |||
20110213670, | |||
20110223571, | |||
20110224569, | |||
20110225021, | |||
20110225043, | |||
20110225049, | |||
20110231240, | |||
20110237923, | |||
20110237971, | |||
20110246574, | |||
20110248729, | |||
20110256520, | |||
20110257502, | |||
20110257937, | |||
20110270620, | |||
20110271312, | |||
20110276504, | |||
20110282231, | |||
20110282232, | |||
20110282749, | |||
20110282880, | |||
20110296004, | |||
20110298706, | |||
20110301431, | |||
20110313849, | |||
20110319975, | |||
20120002848, | |||
20120004899, | |||
20120022391, | |||
20120035428, | |||
20120036004, | |||
20120036005, | |||
20120046993, | |||
20120054018, | |||
20120072289, | |||
20120072845, | |||
20120072936, | |||
20120078065, | |||
20120083668, | |||
20120084139, | |||
20120089552, | |||
20120096363, | |||
20120108995, | |||
20120114305, | |||
20120130800, | |||
20120166252, | |||
20120173701, | |||
20120203363, | |||
20120203559, | |||
20120239407, | |||
20120245978, | |||
20120249797, | |||
20120254909, | |||
20120272256, | |||
20120284332, | |||
20120290409, | |||
20120290637, | |||
20120296699, | |||
20120317198, | |||
20130018949, | |||
20130022948, | |||
20130024272, | |||
20130046577, | |||
20130060125, | |||
20130094722, | |||
20130097715, | |||
20130121591, | |||
20130124365, | |||
20130143185, | |||
20130145385, | |||
20130152506, | |||
20130166373, | |||
20130183646, | |||
20130185140, | |||
20130185141, | |||
20130185142, | |||
20130185144, | |||
20130185145, | |||
20130205314, | |||
20130268279, | |||
20130280682, | |||
20130304540, | |||
20130311132, | |||
20130332259, | |||
20140015002, | |||
20140025620, | |||
20140058828, | |||
20140067466, | |||
20140150002, | |||
20140162225, | |||
20140200416, | |||
20140214335, | |||
20140221866, | |||
20140244345, | |||
20140278914, | |||
20140282646, | |||
20140350431, | |||
20150084860, | |||
20150186923, | |||
20150213019, | |||
20160035060, | |||
20160239139, | |||
20160345060, | |||
20170249466, | |||
20180192126, | |||
EP1087618, | |||
EP1609418, | |||
EP1808102, | |||
GB1374658, | |||
GB2221759, | |||
JP2001147944, | |||
JP2005160805, | |||
JP200551654, | |||
JP2006227994, | |||
JP2006305334, | |||
JP20066355, | |||
KR200422399, | |||
RE34015, | Aug 07 1987 | The Children's Medical Center Corporation | Brain electrical activity mapping |
WO2100241, | |||
WO2102238, | |||
WO1997017774, | |||
WO1997040745, | |||
WO1997041673, | |||
WO2004049225, | |||
WO2008030831, | |||
WO2008055078, | |||
WO2008064431, | |||
WO2008077178, | |||
WO2008109694, | |||
WO2008109699, | |||
WO2008121651, | |||
WO2008137579, | |||
WO2008137581, | |||
WO2008141340, | |||
WO2008154410, | |||
WO2009018374, | |||
WO2009052833, | |||
WO2011055291, | |||
WO2011056679, | |||
WO95018565, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Aug 29 2016 | MURRAY, BRENDAN | THE NIELSEN COMPANY US , LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 045486 | /0454 | |
Sep 20 2016 | LEVINE, BRIAN | THE NIELSEN COMPANY US , LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 045486 | /0454 | |
Dec 01 2016 | MARCI, CARL D | THE NIELSEN COMPANY US , LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 045486 | /0454 | |
Feb 28 2018 | The Nielsen Company (US), LLC | (assignment on the face of the patent) | / | |||
Jun 04 2020 | GRACENOTE MEDIA SERVICES, LLC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NETRATINGS, LLC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | GRACENOTE DIGITAL VENTURES, LLC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | GRACENOTE, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | EXELATE, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | CZT ACN TRADEMARKS, L L C | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | ATHENIAN LEASING CORPORATION | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | ART HOLDING, L L C | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | AFFINNOVA, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | ACNIELSEN ERATINGS COM | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | ACNIELSEN CORPORATION | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | ACN HOLDINGS INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | A C NIELSEN COMPANY, LLC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | A C NIELSEN ARGENTINA S A | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NIELSEN AUDIO, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NIELSEN CONSUMER INSIGHTS, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | VNU INTERNATIONAL B V | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | THE NIELSEN COMPANY B V | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NIELSEN HOLDING AND FINANCE B V | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NMR LICENSING ASSOCIATES, L P | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | VNU MARKETING INFORMATION, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | VIZU CORPORATION | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | THE NIELSEN COMPANY US , LLC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | TNC US HOLDINGS, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | TCG DIVESTITURE INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NMR INVESTING I, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NIELSEN MOBILE, LLC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NIELSEN INTERNATIONAL HOLDINGS, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NIELSEN FINANCE CO | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | NIELSEN CONSUMER NEUROSCIENCE, INC | CITIBANK, N A | CORRECTIVE ASSIGNMENT TO CORRECT THE PATENTS LISTED ON SCHEDULE 1 RECORDED ON 6-9-2020 PREVIOUSLY RECORDED ON REEL 053473 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SUPPLEMENTAL IP SECURITY AGREEMENT | 054066 | /0064 | |
Jun 04 2020 | VNU INTERNATIONAL B V | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | A C NIELSEN COMPANY, LLC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NETRATINGS, LLC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | GRACENOTE MEDIA SERVICES, LLC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | GRACENOTE DIGITAL VENTURES, LLC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | GRACENOTE, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | EXELATE, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | CZT ACN TRADEMARKS, L L C | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | ATHENIAN LEASING CORPORATION | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | ART HOLDING, L L C | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | AFFINNOVA, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | ACNIELSEN ERATINGS COM | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | ACNIELSEN CORPORATION | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | ACN HOLDINGS INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NIELSEN AUDIO, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NIELSEN CONSUMER INSIGHTS, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NIELSEN CONSUMER NEUROSCIENCE, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | THE NIELSEN COMPANY B V | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NIELSEN HOLDING AND FINANCE B V | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NMR LICENSING ASSOCIATES, L P | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | VNU MARKETING INFORMATION, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | VIZU CORPORATION | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | THE NIELSEN COMPANY US , LLC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | TNC US HOLDINGS, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | TCG DIVESTITURE INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NMR INVESTING I, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NIELSEN UK FINANCE I, LLC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NIELSEN MOBILE, LLC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NIELSEN INTERNATIONAL HOLDINGS, INC | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Jun 04 2020 | NIELSEN FINANCE CO | CITIBANK, N A | SUPPLEMENTAL SECURITY AGREEMENT | 053473 | /0001 | |
Feb 09 2021 | THE NIELSEN COMPANY US , LLC | Nielsen Consumer LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 055249 | /0645 | |
Mar 05 2021 | CITIBANK, N A | Nielsen Consumer LLC | PARTIAL RELEASE OF SECURITY INTEREST | 055557 | /0949 | |
Mar 05 2021 | CITIBANK, N A | NIELSEN CONSUMER NEUROSCIENCE, INC | PARTIAL RELEASE OF SECURITY INTEREST | 055557 | /0949 | |
Mar 05 2021 | Nielsen Consumer LLC | BANK OF AMERICA, N A , AS ADMINISTRATIVE AGENT AND COLLATERAL AGENT | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 055742 | /0719 | |
Mar 05 2021 | BYZZER INC | BANK OF AMERICA, N A , AS ADMINISTRATIVE AGENT AND COLLATERAL AGENT | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 055742 | /0719 | |
Oct 11 2022 | CITIBANK, N A | NETRATINGS, LLC | RELEASE REEL 054066 FRAME 0064 | 063605 | /0001 | |
Oct 11 2022 | CITIBANK, N A | THE NIELSEN COMPANY US , LLC | RELEASE REEL 054066 FRAME 0064 | 063605 | /0001 | |
Oct 11 2022 | CITIBANK, N A | GRACENOTE MEDIA SERVICES, LLC | RELEASE REEL 054066 FRAME 0064 | 063605 | /0001 | |
Oct 11 2022 | CITIBANK, N A | GRACENOTE, INC | RELEASE REEL 054066 FRAME 0064 | 063605 | /0001 | |
Oct 11 2022 | CITIBANK, N A | EXELATE, INC | RELEASE REEL 054066 FRAME 0064 | 063605 | /0001 | |
Oct 11 2022 | CITIBANK, N A | A C NIELSEN COMPANY, LLC | RELEASE REEL 054066 FRAME 0064 | 063605 | /0001 | |
Oct 11 2022 | CITIBANK, N A | NETRATINGS, LLC | RELEASE REEL 053473 FRAME 0001 | 063603 | /0001 | |
Oct 11 2022 | CITIBANK, N A | THE NIELSEN COMPANY US , LLC | RELEASE REEL 053473 FRAME 0001 | 063603 | /0001 | |
Oct 11 2022 | CITIBANK, N A | GRACENOTE MEDIA SERVICES, LLC | RELEASE REEL 053473 FRAME 0001 | 063603 | /0001 | |
Oct 11 2022 | CITIBANK, N A | EXELATE, INC | RELEASE REEL 053473 FRAME 0001 | 063603 | /0001 | |
Oct 11 2022 | CITIBANK, N A | A C NIELSEN COMPANY, LLC | RELEASE REEL 053473 FRAME 0001 | 063603 | /0001 | |
Oct 11 2022 | CITIBANK, N A | GRACENOTE, INC | RELEASE REEL 053473 FRAME 0001 | 063603 | /0001 |
Date | Maintenance Fee Events |
Feb 28 2018 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Mar 08 2024 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Sep 08 2023 | 4 years fee payment window open |
Mar 08 2024 | 6 months grace period start (w surcharge) |
Sep 08 2024 | patent expiry (for year 4) |
Sep 08 2026 | 2 years to revive unintentionally abandoned end. (for year 4) |
Sep 08 2027 | 8 years fee payment window open |
Mar 08 2028 | 6 months grace period start (w surcharge) |
Sep 08 2028 | patent expiry (for year 8) |
Sep 08 2030 | 2 years to revive unintentionally abandoned end. (for year 8) |
Sep 08 2031 | 12 years fee payment window open |
Mar 08 2032 | 6 months grace period start (w surcharge) |
Sep 08 2032 | patent expiry (for year 12) |
Sep 08 2034 | 2 years to revive unintentionally abandoned end. (for year 12) |