Posts Tagged ‘ AAPOR

Latent Class Analysis: A Probabilistic Approach to Uncovering Latent Classes of TV Advertisements

Abstract:

In media measurement, examinations of relationships between audience behaviors (e.g., viewership of TV advertisements) and demographics (e.g., age and gender) are common practice. These examinations can follow a variable-centered approach that differentiates variables into dependent and independent variables (e.g., linear regression) or a person-centered approach which groups observations together according to their configural profile of characteristics (e.g., K-means; Morris & Kavussanu, 2008). This paper extends the boundaries of media research by using a latent clustering method — Latent Class Analysis (LCA) for analyzing advertisement viewership data. LCA is a probabilistic mixture modeling technique used to identify a set of discrete latent classes of observations, in this case advertisements, based on the values of a set of categorical indicators. The technique focuses on generating classes of advertisements that share similar patterns of responses and comparing with other classes to subsequently determine how they are differentially related to predictors and outcomes (Wang & Hanges, 2011).

The dataset used in our analysis is characterized by extreme volume (7M rows) and a variety of variable types (variables are a combination of continuous, categorical, and count). Preliminary findings indicate the classification of advertisements into two latent classes as defined by a set of categorical indicators of interest. The most salient features of Class 1 in comparison to Class 2 are a greater likelihood to air within News/Information and Sports TV genres as well as a greater likelihood to air on a weekday as opposed to a weekend or weeknight. We conduct additional analyses to understand how these latent classifications are related to a measure of what the media world refers to as co-viewing. This paper concludes with a discussion of the results and their implications along with recommendations for future research. R code and a sample dataset is provided for interested readers and practitioners.

Recommended Citation:

Craft, M., & Rao, K. (2023). Latent Class Analysis: A Probabilistic Approach to Uncovering Latent Classes of TV Advertisements. American Association for Public Opinion Research, Philadelphia, PA.

Attached Documents/Links:

  • AAPOR 2023 Program
  • For a copy of this presentation, please send me a comment with your email address in the box below or an email to kumarrons at gmail dot com.

Does Ancillary Data about Ancillary Data Help Treat Missing Responses in Ancillary Data?

Abstract:

Consumer-file marketing companies collect and sell customer’s demographic and behavioral information wherever they can: through loyalty cards, website visits, social media posts, etc. These ancillary data are tremendously valuable and have a variety of applications in the field of television and digital media measurement that include targeting households or individuals for marketing campaigns, adjusting for nonresponse bias, and projecting estimates from a sample to an underlying universe. These potential advantages, however, depend on the completeness of the information contained in the data. Today, most applications involving the use of ancillary data exclude incomplete records from the substantive analysis. Exclusion as a missing data handling strategy has been known to result in inaccurate conclusions depending on the nature of the mechanism assumed to have generated the missing data. In this paper, we demonstrate a model-based alternative to exclusion, called multiple imputation (MI), which is used to fill in the missing values. A correctly specified imputation model preserves important characteristics of the data (e.g., variances, correlations, interactions) and can mitigate any bias resulting from exclusion-based missing data handling strategies. Our paper demonstrates MI in the context of a substantive analysis of the relationship between a completely observed measure of the number of advertisement impressions served across television networks and an incomplete measure of the number of children per household. We include likely correlates of the suspected cause of the missing data in the imputation model to evaluate whether their inclusion improves the quality of the imputed values. These correlates are measures pulled from a second, more completely observed, ancillary dataset containing television and digital media audience measurements. This paper concludes with a discussion of our findings, which are likely to have important implications for the use of ancillary data in television and digital media audience measurement.

Recommended Citation:

Craft, M., Raghunath, A., & Rao, K. (2023). Does Ancillary Data about Ancillary Data Help Treat Missing Responses in Ancillary Data? American Association for Public Opinion Research, Philadelphia, PA.

Attached Documents/Links:

  • AAPOR 2023 Program
  • For a copy of this presentation, please send me a comment with your email address in the box below or an email to kumarrons at gmail dot com.

Only for the Young at Heart: Co-Viewing on Mobile Devices and Viewing on the Go?

Abstract:

With the relative ease and accessibility of a variety of content available to users of smartphones and tablets, there has been a subtle behavioral change in how people use these devices. The concept of viewing together or having more than one viewer for a mobile device is a phenomenon referred to as “co-viewing” and is a new area that warrants further investigation. Very little information is available on who is likely to engage in co-viewing behaviors, what types of mobile devices are used, what content is likely to be viewed and if those who engage in this activity / behavior are fundamentally different than those who are less likely—what are the behavioral or demographic differences among those who participate in these activities. Thus the focus here is to examine and provide a baseline understanding around the concept of co-viewing with specific focus of content viewing on the “go” or away from home. Read more

Who Is behind That Screen? Solving the Puzzle of Within-Home Computer Sharing among Household Members

Abstract:

The number of US households with access to computers at home has continued to grow. According to the 2011 Computer and Internet Use report published by US department of Commerce, 77% of US homes have computers in their home, compared to 62% in 2003. Many households, however, do not have multiple computers dedicated to each member living in the house. As such, sharing of computers amongst household members can be a prevalent phenomenon in home computer usage. Understanding this within-house computer sharing phenomenon and identifying the mostly likely person behind the computer screen can be of interest to market researchers and practitioners, particularly those interested in studying effective ways to target online ads based on users, online activities. For survey researchers who are attempting to recruit hard-to-reach individuals like teens and young adults, understanding of computer sharing could help establish contact at times when those individuals are more likely to be behind the computer. Despite its prevalence, within-house computer sharing has barely received any research attention. This study hopes to break through the barriers preventing the light of scientific inquiry into this phenomenon. Read more

Is It Too Much to Ask? The Role of Question Difficulty in Survey Response Accuracy for Measures of Online Behavior

Abstract:

While market research capabilities of online panels have never been greater, the challenges facing these panels in many ways are just as great. Over the past few years, online panels that recruit members using nonprobability/opt-in based methods have come under increased scrutiny and criticism over data quality concerns such as respondent identity and increased satisficing. These concerns have drawn attention to the heart of the issue, which is: the accuracy or truthfulness of data provided by opt-in panel respondents. This issue is of utmost importance given the recently established link between opt-in panel sample and poor survey data quality (see Yeager et. al. 2011). Read more

Evaluation of Alternative Weighting Approaches to Reduce Nonresponse Bias

Abstract:

With declining response rates, surveys increasingly rely on weighting adjustments to correct for potential nonresponse bias. The resulting increased need to improve survey weights faces two key challenges. First, additional auxiliary data are needed to augment the models used to estimate the weights. Depending on the properties of these auxiliary data, nonresponse bias can be reduced, left the same, or even increased. Thus, the second challenge is to be able to evaluate the alternative weights, when the assumption of “different estimates means less bias” may not hold. Ideally, data need to be collected from as many nonrespondents as possible to provide direct estimates of nonresponse bias. Read more

Difficult Data: Comparing the Quality of Behavioral, Recall, and Proxy Data Across Survey Modes

Abstract:

The mode choice literature is rife with evidence on the impact of different survey modes on response rates, respondent cooperation, and data quality. However, insufficient attention has been paid to the quality of “difficult data” provided when respondents cannot choose the mode and thus cannot maximize their comfort with the survey. Here, “difficult data” correspond to questions that are burdensome for respondents to think about – e.g., very specific details on a behavior, on past events, or on the behavior of other persons. Read more

Recruitment and Retention in Multi-Mode Survey Panels

Abstract:

This study builds on a previously published panel recruitment experiment (Rao, Kaminska, and McCutcheon 2010), extending that analysis to an examination of the effectiveness of pre-recruitment factors such as mode and response inducements on three post-recruitment panel participation effects: attrition rates, survey completion rates, and panel data quality. The panel recruitment experiment, conducted with the Gallup Panel, netted 1,282 households with 2,042 panel members. For these recruited members, we collected data on panel participation and retention, and use it for analysis in this study. Read more

Is Past, the Future? Resampling Past Respondents to Improve Current Response Rates

Abstract:

The Nielsen TV Ratings Diary service involves the use of a one-week TV diary survey for measuring TV ratings. While the service has been around for a while, it recently received a sampling makeover to address the diminishing coverage associated with landline random-digit dialing (RDD) surveys. Address-based sampling (ABS) replaced RDD as the sampling methodology for the diary service. Read more

Home or Work or Both? Assessing the Role of Duplication of Website Visitations Using a Online Metered Panel

Abstract:

In this study, for multiple websites, we estimate duplicated audience reach between home and work Internet access locations. By employing a probability-based matching technique, we use metered-panel based data from non-overlapping home and work panels for creating a virtual overlapping home-work panel. Read more